dianecy commited on
Commit
86bdb0c
·
verified ·
1 Parent(s): 4239f14

Upload folder using huggingface_hub

Browse files
config/cris_r50.yaml ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ DATA:
2
+ dataset: refcocog_u
3
+ train_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/train.lmdb
4
+ train_split: train
5
+ val_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
6
+ val_split: val
7
+ mask_root: /home/s1/chaeyunkim/VerbCentric_CY/datasets/masks/refcocog_u
8
+
9
+ # Base Arch
10
+ clip_pretrain: pretrain/RN50.pt
11
+ input_size: 416
12
+ word_len: 22
13
+ word_dim: 1024
14
+ vis_dim: 512
15
+ fpn_in: [512, 1024, 1024]
16
+ fpn_out: [256, 512, 1024]
17
+ sync_bn: True
18
+ # Decoder
19
+ num_layers: 3
20
+ num_head: 8
21
+ dim_ffn: 2048
22
+ dropout: 0.1
23
+ intermediate: False
24
+ # Training Setting
25
+ workers: 32 # data loader workers
26
+ workers_val: 16
27
+ epochs: 50
28
+ milestones: [35]
29
+ start_epoch: 0
30
+ batch_size: 64 # batch size for training
31
+ batch_size_val: 64 # batch size for validation during training, memory and speed tradeoff
32
+ base_lr: 0.0001
33
+ lr_decay: 0.1
34
+ lr_multi: 0.1
35
+ weight_decay: 0.
36
+ max_norm: 0.
37
+ manual_seed: 0
38
+ print_freq: 100
39
+ # Resume & Save
40
+ metric_mode: 'original'
41
+ metric_learning: False
42
+ exp_name: CRIS_R50
43
+ output_folder: exp/refcocog_u_repro
44
+ save_freq: 1
45
+ weight: # path to initial weight (default: none)
46
+ resume: 'latest' # path to latest checkpoint (default: none)
47
+ evaluate: True # evaluate on validation set, extra gpu memory needed and small batch_size_val is recommend
48
+ freeze: True
49
+ Distributed:
50
+ dist_url: tcp://localhost:3681
51
+ dist_backend: 'nccl'
52
+ multiprocessing_distributed: True
53
+ world_size: 1
54
+ rank: 0
55
+ TEST:
56
+ test_split: val-test
57
+ test_lmdb: datasets/lmdb/refcocog_u/val.lmdb
58
+ visualize: False
config/cris_verbonly_b64_nopos.yaml ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ DATA:
3
+ dataset: refcocog_u
4
+ train_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/train.lmdb
5
+ train_split: train
6
+ val_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
7
+ val_split: val
8
+ mask_root: /home/s1/chaeyunkim/VerbCentric_CY/datasets/masks/refcocog_u
9
+
10
+ TRAIN:
11
+ # Base Arch
12
+ clip_pretrain: pretrain/RN50.pt
13
+ input_size: 416
14
+ word_len: 22
15
+ word_dim: 1024
16
+ vis_dim: 512
17
+ fpn_in: [512, 1024, 1024]
18
+ fpn_out: [256, 512, 1024]
19
+ sync_bn: True
20
+ freeze: True
21
+ # Decoder
22
+ num_layers: 3
23
+ num_head: 8
24
+ dim_ffn: 2048
25
+ dropout: 0.1
26
+ intermediate: False
27
+ # Training Setting
28
+ workers: 4 # data loader workers
29
+ workers_val: 4
30
+ epochs: 50
31
+ milestones: [35]
32
+ start_epoch: 0
33
+ batch_size: 32 # batch size for training
34
+ batch_size_val: 32 # batch size for validation during training, memory and speed tradeoff
35
+ base_lr: 0.0001
36
+ lr_decay: 0.1
37
+ lr_multi: 0.1
38
+ weight_decay: 0.
39
+ max_norm: 0.
40
+ manual_seed: 0
41
+ print_freq: 100
42
+ # metric learning args
43
+ metric_learning: True
44
+ # specific metric learning args
45
+ metric_mode : 'hardpos_only_rev' # Choice : ['hardpos_only', 'hardpos_only_rev', 'both']
46
+ exclude_multiobj : True # exclude multiobj (nobj >= 3)
47
+ exclude_pos : True # exclude multiobj w/ positional query expression (nobj >=2 && positional query)
48
+ loss_option : 'ACL_verbonly' # Choice : ['AML_verbonly', 'AML', 'ACL', 'ACL_verbonly']
49
+ metric_loss_weight: 0.1
50
+ hn_prob: 0.0
51
+
52
+ # Angular Margin Contrastive Loss argument
53
+ margin_value : 10
54
+ temperature : 0.05
55
+
56
+ # Resume & Save
57
+ exp_name: CRIS_AML_verbonly_pos25_b32
58
+ output_folder: exp/refcocog_u/exclude_multiobj
59
+ save_freq: 1
60
+ weight: # path to initial weight (default: none)
61
+ resume: "latest" # path to latest checkpoint (default: none)
62
+ evaluate: True # evaluate on validation set, extra gpu memory needed and small batch_size_val is recommend
63
+
64
+ Distributed:
65
+ dist_url: tcp://localhost:2298
66
+ dist_backend: 'nccl'
67
+ multiprocessing_distributed: True
68
+ world_size: 1
69
+ rank: 0
70
+
71
+ TEST:
72
+ test_split: val-test
73
+ test_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
74
+ visualize: False
config/cris_verbonly_b64_nopos_hn0.03.yaml ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ DATA:
3
+ dataset: refcocog_u
4
+ train_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/train.lmdb
5
+ train_split: train
6
+ val_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
7
+ val_split: val
8
+ mask_root: /home/s1/chaeyunkim/VerbCentric_CY/datasets/masks/refcocog_u
9
+
10
+ TRAIN:
11
+ # Base Arch
12
+ clip_pretrain: pretrain/RN50.pt
13
+ input_size: 416
14
+ word_len: 22
15
+ word_dim: 1024
16
+ vis_dim: 512
17
+ fpn_in: [512, 1024, 1024]
18
+ fpn_out: [256, 512, 1024]
19
+ sync_bn: True
20
+ freeze: True
21
+ # Decoder
22
+ num_layers: 3
23
+ num_head: 8
24
+ dim_ffn: 2048
25
+ dropout: 0.1
26
+ intermediate: False
27
+ # Training Setting
28
+ workers: 4 # data loader workers
29
+ workers_val: 4
30
+ epochs: 50
31
+ milestones: [35]
32
+ start_epoch: 0
33
+ batch_size: 32 # batch size for training
34
+ batch_size_val: 32 # batch size for validation during training, memory and speed tradeoff
35
+ base_lr: 0.0001
36
+ lr_decay: 0.1
37
+ lr_multi: 0.1
38
+ weight_decay: 0.
39
+ max_norm: 0.
40
+ manual_seed: 0
41
+ print_freq: 100
42
+ # metric learning args
43
+ metric_learning: True
44
+ # specific metric learning args
45
+ metric_mode : 'hardpos_only_rev' # Choice : ['hardpos_only', 'hardpos_only_rev', 'both']
46
+ exclude_multiobj : True # exclude multiobj (nobj >= 3)
47
+ exclude_pos : True # exclude multiobj w/ positional query expression (nobj >=2 && positional query)
48
+ loss_option : 'ACL_verbonly' # Choice : ['AML_verbonly', 'AML', 'ACL', 'ACL_verbonly']
49
+ metric_loss_weight: 0.1
50
+ hn_prob: 0.0
51
+
52
+ # Angular Margin Contrastive Loss argument
53
+ margin_value : 10
54
+ temperature : 0.05
55
+
56
+ # Resume & Save
57
+ exp_name: CRIS_AML_verbonly_pos25_b32
58
+ output_folder: exp/refcocog_u/exclude_multiobj
59
+ save_freq: 1
60
+ weight: # path to initial weight (default: none)
61
+ resume: "latest" # path to latest checkpoint (default: none)
62
+ evaluate: True # evaluate on validation set, extra gpu memory needed and small batch_size_val is recommend
63
+
64
+ Distributed:
65
+ dist_url: tcp://localhost:2298
66
+ dist_backend: 'nccl'
67
+ multiprocessing_distributed: True
68
+ world_size: 1
69
+ rank: 0
70
+
71
+ TEST:
72
+ test_split: val-test
73
+ test_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
74
+ visualize: False
config/cris_verbonly_b64_nopos_hn0.1.yaml ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ DATA:
3
+ dataset: refcocog_u
4
+ train_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/train.lmdb
5
+ train_split: train
6
+ val_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
7
+ val_split: val
8
+ mask_root: /home/s1/chaeyunkim/VerbCentric_CY/datasets/masks/refcocog_u
9
+
10
+ TRAIN:
11
+ # Base Arch
12
+ clip_pretrain: pretrain/RN50.pt
13
+ input_size: 416
14
+ word_len: 22
15
+ word_dim: 1024
16
+ vis_dim: 512
17
+ fpn_in: [512, 1024, 1024]
18
+ fpn_out: [256, 512, 1024]
19
+ sync_bn: True
20
+ freeze: True
21
+ # Decoder
22
+ num_layers: 3
23
+ num_head: 8
24
+ dim_ffn: 2048
25
+ dropout: 0.1
26
+ intermediate: False
27
+ # Training Setting
28
+ workers: 4 # data loader workers
29
+ workers_val: 4
30
+ epochs: 50
31
+ milestones: [35]
32
+ start_epoch: 0
33
+ batch_size: 32 # batch size for training
34
+ batch_size_val: 32 # batch size for validation during training, memory and speed tradeoff
35
+ base_lr: 0.0001
36
+ lr_decay: 0.1
37
+ lr_multi: 0.1
38
+ weight_decay: 0.
39
+ max_norm: 0.
40
+ manual_seed: 0
41
+ print_freq: 100
42
+ # metric learning args
43
+ metric_learning: True
44
+ # specific metric learning args
45
+ metric_mode : 'hardpos_only_rev' # Choice : ['hardpos_only', 'hardpos_only_rev', 'both']
46
+ exclude_multiobj : True # exclude multiobj (nobj >= 3)
47
+ exclude_pos : True # exclude multiobj w/ positional query expression (nobj >=2 && positional query)
48
+ loss_option : 'ACL_verbonly' # Choice : ['AML_verbonly', 'AML', 'ACL', 'ACL_verbonly']
49
+ metric_loss_weight: 0.1
50
+ hn_prob: 0.0
51
+
52
+ # Angular Margin Contrastive Loss argument
53
+ margin_value : 10
54
+ temperature : 0.05
55
+
56
+ # Resume & Save
57
+ exp_name: CRIS_AML_verbonly_pos25_b32
58
+ output_folder: exp/refcocog_u/exclude_multiobj
59
+ save_freq: 1
60
+ weight: # path to initial weight (default: none)
61
+ resume: "latest" # path to latest checkpoint (default: none)
62
+ evaluate: True # evaluate on validation set, extra gpu memory needed and small batch_size_val is recommend
63
+
64
+ Distributed:
65
+ dist_url: tcp://localhost:2298
66
+ dist_backend: 'nccl'
67
+ multiprocessing_distributed: True
68
+ world_size: 1
69
+ rank: 0
70
+
71
+ TEST:
72
+ test_split: val-test
73
+ test_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
74
+ visualize: False
config/cris_verbonly_b64_nopos_nf.yaml ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ DATA:
3
+ dataset: refcocog_u
4
+ train_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/train.lmdb
5
+ train_split: train
6
+ val_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
7
+ val_split: val
8
+ mask_root: /home/s1/chaeyunkim/VerbCentric_CY/datasets/masks/refcocog_u
9
+
10
+ TRAIN:
11
+ # Base Arch
12
+ clip_pretrain: pretrain/RN50.pt
13
+ input_size: 416
14
+ word_len: 22
15
+ word_dim: 1024
16
+ vis_dim: 512
17
+ fpn_in: [512, 1024, 1024]
18
+ fpn_out: [256, 512, 1024]
19
+ sync_bn: True
20
+ freeze: False
21
+ # Decoder
22
+ num_layers: 3
23
+ num_head: 8
24
+ dim_ffn: 2048
25
+ dropout: 0.1
26
+ intermediate: False
27
+ # Training Setting
28
+ workers: 4 # data loader workers
29
+ workers_val: 4
30
+ epochs: 50
31
+ milestones: [35]
32
+ start_epoch: 0
33
+ batch_size: 32 # batch size for training
34
+ batch_size_val: 32 # batch size for validation during training, memory and speed tradeoff
35
+ base_lr: 0.0001
36
+ lr_decay: 0.1
37
+ lr_multi: 0.1
38
+ weight_decay: 0.
39
+ max_norm: 0.
40
+ manual_seed: 0
41
+ print_freq: 100
42
+ # metric learning args
43
+ metric_learning: True
44
+ # specific metric learning args
45
+ metric_mode : 'hardpos_only' # Choice : ['hardpos_only', 'both']
46
+ exclude_multiobj : True # exclude multiobj (nobj >= 3)
47
+ exclude_pos : True # exclude multiobj w/ positional query expression (nobj >=2 && positional query)
48
+ loss_option : 'ACL_verbonly' # Choice : ['AML_verbonly', 'AML', 'ACL', 'ACL_verbonly']
49
+ metric_loss_weight: 0.1
50
+ hn_prob: 0.0
51
+
52
+ # Angular Margin Contrastive Loss argument
53
+ margin_value : 10
54
+ temperature : 0.05
55
+
56
+ # Resume & Save
57
+ exp_name: CRIS_AML_verbonly_pos25_b32
58
+ output_folder: exp/refcocog_u/exclude_multiobj
59
+ save_freq: 1
60
+ weight: # path to initial weight (default: none)
61
+ resume: "latest" # path to latest checkpoint (default: none)
62
+ evaluate: True # evaluate on validation set, extra gpu memory needed and small batch_size_val is recommend
63
+
64
+ Distributed:
65
+ dist_url: tcp://localhost:2298
66
+ dist_backend: 'nccl'
67
+ multiprocessing_distributed: True
68
+ world_size: 1
69
+ rank: 0
70
+
71
+ TEST:
72
+ test_split: val-test
73
+ test_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
74
+ visualize: False
config/cris_verbonly_b64_nopos_notarget.yaml ADDED
@@ -0,0 +1,75 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ DATA:
3
+ dataset: refcocog_u
4
+ train_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/train.lmdb
5
+ train_split: train
6
+ val_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
7
+ val_split: val
8
+ mask_root: /home/s1/chaeyunkim/VerbCentric_CY/datasets/masks/refcocog_u
9
+
10
+ TRAIN:
11
+ # Base Arch
12
+ clip_pretrain: pretrain/RN50.pt
13
+ input_size: 416
14
+ word_len: 22
15
+ word_dim: 1024
16
+ vis_dim: 512
17
+ fpn_in: [512, 1024, 1024]
18
+ fpn_out: [256, 512, 1024]
19
+ sync_bn: True
20
+ freeze: True
21
+ # Decoder
22
+ num_layers: 3
23
+ num_head: 8
24
+ dim_ffn: 2048
25
+ dropout: 0.1
26
+ intermediate: False
27
+ # Training Setting
28
+ workers: 4 # data loader workers
29
+ workers_val: 4
30
+ epochs: 50
31
+ milestones: [35]
32
+ start_epoch: 0
33
+ batch_size: 32 # batch size for training
34
+ batch_size_val: 32 # batch size for validation during training, memory and speed tradeoff
35
+ base_lr: 0.0001
36
+ lr_decay: 0.1
37
+ lr_multi: 0.1
38
+ weight_decay: 0.
39
+ max_norm: 0.
40
+ manual_seed: 0
41
+ print_freq: 100
42
+ # metric learning args
43
+ metric_learning: True
44
+ # specific metric learning args
45
+ metric_mode : 'hardpos_only_rev' # Choice : ['hardpos_only', 'hardpos_only_rev', 'both']
46
+ exclude_multiobj : True # exclude multiobj (nobj >= 3)
47
+ exclude_pos : True # exclude multiobj w/ positional query expression (nobj >=2 && positional query)
48
+ loss_option : 'ACL_verbonly' # Choice : ['AML_verbonly', 'AML', 'ACL', 'ACL_verbonly']
49
+ metric_loss_weight: 0.1
50
+ hn_prob: 0.0
51
+ hn_celoss: False
52
+
53
+ # Angular Margin Contrastive Loss argument
54
+ margin_value : 10
55
+ temperature : 0.05
56
+
57
+ # Resume & Save
58
+ exp_name: CRIS_AML_verbonly_pos25_b32
59
+ output_folder: exp/refcocog_u/exclude_multiobj
60
+ save_freq: 1
61
+ weight: # path to initial weight (default: none)
62
+ resume: "latest" # path to latest checkpoint (default: none)
63
+ evaluate: True # evaluate on validation set, extra gpu memory needed and small batch_size_val is recommend
64
+
65
+ Distributed:
66
+ dist_url: tcp://localhost:2298
67
+ dist_backend: 'nccl'
68
+ multiprocessing_distributed: True
69
+ world_size: 1
70
+ rank: 0
71
+
72
+ TEST:
73
+ test_split: val-test
74
+ test_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
75
+ visualize: False
config/hardneg/cris_aclvo_hn_b64_nopos.yaml ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ DATA:
3
+ dataset: refcocog_u
4
+ train_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/train.lmdb
5
+ train_split: train
6
+ val_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
7
+ val_split: val
8
+ mask_root: /home/s1/chaeyunkim/VerbCentric_CY/datasets/masks/refcocog_u
9
+
10
+ TRAIN:
11
+ # Base Arch
12
+ clip_pretrain: pretrain/RN50.pt
13
+ input_size: 416
14
+ word_len: 22
15
+ word_dim: 1024
16
+ vis_dim: 512
17
+ fpn_in: [512, 1024, 1024]
18
+ fpn_out: [256, 512, 1024]
19
+ sync_bn: True
20
+ freeze: True
21
+ # Decoder
22
+ num_layers: 3
23
+ num_head: 8
24
+ dim_ffn: 2048
25
+ dropout: 0.1
26
+ intermediate: False
27
+ # Training Setting
28
+ workers: 4 # data loader workers
29
+ workers_val: 4
30
+ epochs: 50
31
+ milestones: [35]
32
+ start_epoch: 0
33
+ batch_size: 32 # batch size for training
34
+ batch_size_val: 32 # batch size for validation during training, memory and speed tradeoff
35
+ base_lr: 0.0001
36
+ lr_decay: 0.1
37
+ lr_multi: 0.1
38
+ weight_decay: 0.
39
+ max_norm: 0.
40
+ manual_seed: 0
41
+ print_freq: 100
42
+ # metric learning args
43
+ metric_learning: True
44
+ # specific metric learning args
45
+ metric_mode : 'hardpos_only' # Choice : ['hardpos_only', 'both']
46
+ exclude_multiobj : True # exclude multiobj (nobj >= 3)
47
+ exclude_pos : True # exclude multiobj w/ positional query expression (nobj >=2 && positional query)
48
+ loss_option : 'ACL_verbonly' # Choice : ['AML_verbonly', 'AML', 'ACL', 'ACL_verbonly']
49
+ metric_loss_weight: 0.1
50
+
51
+ # hard negative related options
52
+ hn_prob: 0.1 # hardneg percentage with regard to hard positive
53
+ acl_hn_weight : 0.1 # hardneg weight in acl loss aggregation
54
+
55
+ # hardneg no target supervision
56
+ hn_celoss : False
57
+
58
+ # Angular Margin Contrastive Loss argument
59
+ margin_value : 10
60
+ temperature : 0.05
61
+
62
+ # Resume & Save
63
+ exp_name: CRIS_AML_verbonly_pos25_b32
64
+ output_folder: exp/refcocog_u/exclude_multiobj
65
+ save_freq: 1
66
+ weight: # path to initial weight (default: none)
67
+ resume: "latest" # path to latest checkpoint (default: none)
68
+ evaluate: True # evaluate on validation set, extra gpu memory needed and small batch_size_val is recommend
69
+
70
+ Distributed:
71
+ dist_url: tcp://localhost:2298
72
+ dist_backend: 'nccl'
73
+ multiprocessing_distributed: True
74
+ world_size: 1
75
+ rank: 0
76
+
77
+ TEST:
78
+ test_split: val-test
79
+ test_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
80
+ visualize: False