-
Notifications
You must be signed in to change notification settings - Fork 2
Expand file tree
/
Copy pathmain_vis.py
More file actions
77 lines (72 loc) · 2.51 KB
/
main_vis.py
File metadata and controls
77 lines (72 loc) · 2.51 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
# from tools import run_net
from tools import test_net
from utils import parser, dist_utils, misc
from utils.logger import *
from utils.config import *
import time
import os
import torch
from tensorboardX import SummaryWriter
def main():
# args
args = parser.get_args()
# CUDA
args.use_gpu = torch.cuda.is_available()
if args.use_gpu:
torch.backends.cudnn.benchmark = True
# init distributed env first, since logger depends on the dist info.
if args.launcher == "none":
args.distributed = False
else:
args.distributed = True
dist_utils.init_dist(args.launcher)
# re-set gpu_ids with distributed training mode
_, world_size = dist_utils.get_dist_info()
args.world_size = world_size
# logger
timestamp = time.strftime("%Y%m%d_%H%M%S", time.localtime())
log_file = os.path.join(args.experiment_path, f"{timestamp}.log")
logger = get_root_logger(log_file=log_file, name=args.log_name)
# define the tensorboard writer
if not args.test:
if args.local_rank == 0:
train_writer = SummaryWriter(os.path.join(args.tfboard_path, "train"))
val_writer = SummaryWriter(os.path.join(args.tfboard_path, "test"))
else:
train_writer = None
val_writer = None
# config
config = get_config(args, logger=logger)
# batch size
if args.distributed:
assert config.total_bs % world_size == 0
config.dataset.train.others.bs = config.total_bs // world_size
config.dataset.val.others.bs = 1
config.dataset.test.others.bs = 1
else:
config.dataset.train.others.bs = config.total_bs
config.dataset.val.others.bs = 1
config.dataset.test.others.bs = 1
# log
log_args_to_file(args, "args", logger=logger)
log_config_to_file(config, "config", logger=logger)
# exit()
logger.info(f"Distributed training: {args.distributed}")
# set random seeds
if args.seed is not None:
logger.info(
f"Set random seed to {args.seed}, " f"deterministic: {args.deterministic}"
)
misc.set_random_seed(
args.seed + args.local_rank, deterministic=args.deterministic
) # seed + rank, for augmentation
if args.distributed:
assert args.local_rank == torch.distributed.get_rank()
# run
if args.test:
test_net(args, config)
else:
# run_net(args, config, train_writer, val_writer)
raise NotImplementedError
if __name__ == "__main__":
main()