Upload folder using huggingface_hub
Browse files- config/cris_r50.yaml +58 -0
- config/cris_verbonly_b64_nopos.yaml +74 -0
- config/cris_verbonly_b64_nopos_hn0.03.yaml +74 -0
- config/cris_verbonly_b64_nopos_hn0.1.yaml +74 -0
- config/cris_verbonly_b64_nopos_nf.yaml +74 -0
- config/cris_verbonly_b64_nopos_notarget.yaml +75 -0
- config/hardneg/cris_aclvo_hn_b64_nopos.yaml +80 -0
config/cris_r50.yaml
ADDED
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
DATA:
|
2 |
+
dataset: refcocog_u
|
3 |
+
train_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/train.lmdb
|
4 |
+
train_split: train
|
5 |
+
val_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
|
6 |
+
val_split: val
|
7 |
+
mask_root: /home/s1/chaeyunkim/VerbCentric_CY/datasets/masks/refcocog_u
|
8 |
+
|
9 |
+
# Base Arch
|
10 |
+
clip_pretrain: pretrain/RN50.pt
|
11 |
+
input_size: 416
|
12 |
+
word_len: 22
|
13 |
+
word_dim: 1024
|
14 |
+
vis_dim: 512
|
15 |
+
fpn_in: [512, 1024, 1024]
|
16 |
+
fpn_out: [256, 512, 1024]
|
17 |
+
sync_bn: True
|
18 |
+
# Decoder
|
19 |
+
num_layers: 3
|
20 |
+
num_head: 8
|
21 |
+
dim_ffn: 2048
|
22 |
+
dropout: 0.1
|
23 |
+
intermediate: False
|
24 |
+
# Training Setting
|
25 |
+
workers: 32 # data loader workers
|
26 |
+
workers_val: 16
|
27 |
+
epochs: 50
|
28 |
+
milestones: [35]
|
29 |
+
start_epoch: 0
|
30 |
+
batch_size: 64 # batch size for training
|
31 |
+
batch_size_val: 64 # batch size for validation during training, memory and speed tradeoff
|
32 |
+
base_lr: 0.0001
|
33 |
+
lr_decay: 0.1
|
34 |
+
lr_multi: 0.1
|
35 |
+
weight_decay: 0.
|
36 |
+
max_norm: 0.
|
37 |
+
manual_seed: 0
|
38 |
+
print_freq: 100
|
39 |
+
# Resume & Save
|
40 |
+
metric_mode: 'original'
|
41 |
+
metric_learning: False
|
42 |
+
exp_name: CRIS_R50
|
43 |
+
output_folder: exp/refcocog_u_repro
|
44 |
+
save_freq: 1
|
45 |
+
weight: # path to initial weight (default: none)
|
46 |
+
resume: 'latest' # path to latest checkpoint (default: none)
|
47 |
+
evaluate: True # evaluate on validation set, extra gpu memory needed and small batch_size_val is recommend
|
48 |
+
freeze: True
|
49 |
+
Distributed:
|
50 |
+
dist_url: tcp://localhost:3681
|
51 |
+
dist_backend: 'nccl'
|
52 |
+
multiprocessing_distributed: True
|
53 |
+
world_size: 1
|
54 |
+
rank: 0
|
55 |
+
TEST:
|
56 |
+
test_split: val-test
|
57 |
+
test_lmdb: datasets/lmdb/refcocog_u/val.lmdb
|
58 |
+
visualize: False
|
config/cris_verbonly_b64_nopos.yaml
ADDED
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
DATA:
|
3 |
+
dataset: refcocog_u
|
4 |
+
train_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/train.lmdb
|
5 |
+
train_split: train
|
6 |
+
val_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
|
7 |
+
val_split: val
|
8 |
+
mask_root: /home/s1/chaeyunkim/VerbCentric_CY/datasets/masks/refcocog_u
|
9 |
+
|
10 |
+
TRAIN:
|
11 |
+
# Base Arch
|
12 |
+
clip_pretrain: pretrain/RN50.pt
|
13 |
+
input_size: 416
|
14 |
+
word_len: 22
|
15 |
+
word_dim: 1024
|
16 |
+
vis_dim: 512
|
17 |
+
fpn_in: [512, 1024, 1024]
|
18 |
+
fpn_out: [256, 512, 1024]
|
19 |
+
sync_bn: True
|
20 |
+
freeze: True
|
21 |
+
# Decoder
|
22 |
+
num_layers: 3
|
23 |
+
num_head: 8
|
24 |
+
dim_ffn: 2048
|
25 |
+
dropout: 0.1
|
26 |
+
intermediate: False
|
27 |
+
# Training Setting
|
28 |
+
workers: 4 # data loader workers
|
29 |
+
workers_val: 4
|
30 |
+
epochs: 50
|
31 |
+
milestones: [35]
|
32 |
+
start_epoch: 0
|
33 |
+
batch_size: 32 # batch size for training
|
34 |
+
batch_size_val: 32 # batch size for validation during training, memory and speed tradeoff
|
35 |
+
base_lr: 0.0001
|
36 |
+
lr_decay: 0.1
|
37 |
+
lr_multi: 0.1
|
38 |
+
weight_decay: 0.
|
39 |
+
max_norm: 0.
|
40 |
+
manual_seed: 0
|
41 |
+
print_freq: 100
|
42 |
+
# metric learning args
|
43 |
+
metric_learning: True
|
44 |
+
# specific metric learning args
|
45 |
+
metric_mode : 'hardpos_only_rev' # Choice : ['hardpos_only', 'hardpos_only_rev', 'both']
|
46 |
+
exclude_multiobj : True # exclude multiobj (nobj >= 3)
|
47 |
+
exclude_pos : True # exclude multiobj w/ positional query expression (nobj >=2 && positional query)
|
48 |
+
loss_option : 'ACL_verbonly' # Choice : ['AML_verbonly', 'AML', 'ACL', 'ACL_verbonly']
|
49 |
+
metric_loss_weight: 0.1
|
50 |
+
hn_prob: 0.0
|
51 |
+
|
52 |
+
# Angular Margin Contrastive Loss argument
|
53 |
+
margin_value : 10
|
54 |
+
temperature : 0.05
|
55 |
+
|
56 |
+
# Resume & Save
|
57 |
+
exp_name: CRIS_AML_verbonly_pos25_b32
|
58 |
+
output_folder: exp/refcocog_u/exclude_multiobj
|
59 |
+
save_freq: 1
|
60 |
+
weight: # path to initial weight (default: none)
|
61 |
+
resume: "latest" # path to latest checkpoint (default: none)
|
62 |
+
evaluate: True # evaluate on validation set, extra gpu memory needed and small batch_size_val is recommend
|
63 |
+
|
64 |
+
Distributed:
|
65 |
+
dist_url: tcp://localhost:2298
|
66 |
+
dist_backend: 'nccl'
|
67 |
+
multiprocessing_distributed: True
|
68 |
+
world_size: 1
|
69 |
+
rank: 0
|
70 |
+
|
71 |
+
TEST:
|
72 |
+
test_split: val-test
|
73 |
+
test_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
|
74 |
+
visualize: False
|
config/cris_verbonly_b64_nopos_hn0.03.yaml
ADDED
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
DATA:
|
3 |
+
dataset: refcocog_u
|
4 |
+
train_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/train.lmdb
|
5 |
+
train_split: train
|
6 |
+
val_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
|
7 |
+
val_split: val
|
8 |
+
mask_root: /home/s1/chaeyunkim/VerbCentric_CY/datasets/masks/refcocog_u
|
9 |
+
|
10 |
+
TRAIN:
|
11 |
+
# Base Arch
|
12 |
+
clip_pretrain: pretrain/RN50.pt
|
13 |
+
input_size: 416
|
14 |
+
word_len: 22
|
15 |
+
word_dim: 1024
|
16 |
+
vis_dim: 512
|
17 |
+
fpn_in: [512, 1024, 1024]
|
18 |
+
fpn_out: [256, 512, 1024]
|
19 |
+
sync_bn: True
|
20 |
+
freeze: True
|
21 |
+
# Decoder
|
22 |
+
num_layers: 3
|
23 |
+
num_head: 8
|
24 |
+
dim_ffn: 2048
|
25 |
+
dropout: 0.1
|
26 |
+
intermediate: False
|
27 |
+
# Training Setting
|
28 |
+
workers: 4 # data loader workers
|
29 |
+
workers_val: 4
|
30 |
+
epochs: 50
|
31 |
+
milestones: [35]
|
32 |
+
start_epoch: 0
|
33 |
+
batch_size: 32 # batch size for training
|
34 |
+
batch_size_val: 32 # batch size for validation during training, memory and speed tradeoff
|
35 |
+
base_lr: 0.0001
|
36 |
+
lr_decay: 0.1
|
37 |
+
lr_multi: 0.1
|
38 |
+
weight_decay: 0.
|
39 |
+
max_norm: 0.
|
40 |
+
manual_seed: 0
|
41 |
+
print_freq: 100
|
42 |
+
# metric learning args
|
43 |
+
metric_learning: True
|
44 |
+
# specific metric learning args
|
45 |
+
metric_mode : 'hardpos_only_rev' # Choice : ['hardpos_only', 'hardpos_only_rev', 'both']
|
46 |
+
exclude_multiobj : True # exclude multiobj (nobj >= 3)
|
47 |
+
exclude_pos : True # exclude multiobj w/ positional query expression (nobj >=2 && positional query)
|
48 |
+
loss_option : 'ACL_verbonly' # Choice : ['AML_verbonly', 'AML', 'ACL', 'ACL_verbonly']
|
49 |
+
metric_loss_weight: 0.1
|
50 |
+
hn_prob: 0.0
|
51 |
+
|
52 |
+
# Angular Margin Contrastive Loss argument
|
53 |
+
margin_value : 10
|
54 |
+
temperature : 0.05
|
55 |
+
|
56 |
+
# Resume & Save
|
57 |
+
exp_name: CRIS_AML_verbonly_pos25_b32
|
58 |
+
output_folder: exp/refcocog_u/exclude_multiobj
|
59 |
+
save_freq: 1
|
60 |
+
weight: # path to initial weight (default: none)
|
61 |
+
resume: "latest" # path to latest checkpoint (default: none)
|
62 |
+
evaluate: True # evaluate on validation set, extra gpu memory needed and small batch_size_val is recommend
|
63 |
+
|
64 |
+
Distributed:
|
65 |
+
dist_url: tcp://localhost:2298
|
66 |
+
dist_backend: 'nccl'
|
67 |
+
multiprocessing_distributed: True
|
68 |
+
world_size: 1
|
69 |
+
rank: 0
|
70 |
+
|
71 |
+
TEST:
|
72 |
+
test_split: val-test
|
73 |
+
test_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
|
74 |
+
visualize: False
|
config/cris_verbonly_b64_nopos_hn0.1.yaml
ADDED
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
DATA:
|
3 |
+
dataset: refcocog_u
|
4 |
+
train_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/train.lmdb
|
5 |
+
train_split: train
|
6 |
+
val_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
|
7 |
+
val_split: val
|
8 |
+
mask_root: /home/s1/chaeyunkim/VerbCentric_CY/datasets/masks/refcocog_u
|
9 |
+
|
10 |
+
TRAIN:
|
11 |
+
# Base Arch
|
12 |
+
clip_pretrain: pretrain/RN50.pt
|
13 |
+
input_size: 416
|
14 |
+
word_len: 22
|
15 |
+
word_dim: 1024
|
16 |
+
vis_dim: 512
|
17 |
+
fpn_in: [512, 1024, 1024]
|
18 |
+
fpn_out: [256, 512, 1024]
|
19 |
+
sync_bn: True
|
20 |
+
freeze: True
|
21 |
+
# Decoder
|
22 |
+
num_layers: 3
|
23 |
+
num_head: 8
|
24 |
+
dim_ffn: 2048
|
25 |
+
dropout: 0.1
|
26 |
+
intermediate: False
|
27 |
+
# Training Setting
|
28 |
+
workers: 4 # data loader workers
|
29 |
+
workers_val: 4
|
30 |
+
epochs: 50
|
31 |
+
milestones: [35]
|
32 |
+
start_epoch: 0
|
33 |
+
batch_size: 32 # batch size for training
|
34 |
+
batch_size_val: 32 # batch size for validation during training, memory and speed tradeoff
|
35 |
+
base_lr: 0.0001
|
36 |
+
lr_decay: 0.1
|
37 |
+
lr_multi: 0.1
|
38 |
+
weight_decay: 0.
|
39 |
+
max_norm: 0.
|
40 |
+
manual_seed: 0
|
41 |
+
print_freq: 100
|
42 |
+
# metric learning args
|
43 |
+
metric_learning: True
|
44 |
+
# specific metric learning args
|
45 |
+
metric_mode : 'hardpos_only_rev' # Choice : ['hardpos_only', 'hardpos_only_rev', 'both']
|
46 |
+
exclude_multiobj : True # exclude multiobj (nobj >= 3)
|
47 |
+
exclude_pos : True # exclude multiobj w/ positional query expression (nobj >=2 && positional query)
|
48 |
+
loss_option : 'ACL_verbonly' # Choice : ['AML_verbonly', 'AML', 'ACL', 'ACL_verbonly']
|
49 |
+
metric_loss_weight: 0.1
|
50 |
+
hn_prob: 0.0
|
51 |
+
|
52 |
+
# Angular Margin Contrastive Loss argument
|
53 |
+
margin_value : 10
|
54 |
+
temperature : 0.05
|
55 |
+
|
56 |
+
# Resume & Save
|
57 |
+
exp_name: CRIS_AML_verbonly_pos25_b32
|
58 |
+
output_folder: exp/refcocog_u/exclude_multiobj
|
59 |
+
save_freq: 1
|
60 |
+
weight: # path to initial weight (default: none)
|
61 |
+
resume: "latest" # path to latest checkpoint (default: none)
|
62 |
+
evaluate: True # evaluate on validation set, extra gpu memory needed and small batch_size_val is recommend
|
63 |
+
|
64 |
+
Distributed:
|
65 |
+
dist_url: tcp://localhost:2298
|
66 |
+
dist_backend: 'nccl'
|
67 |
+
multiprocessing_distributed: True
|
68 |
+
world_size: 1
|
69 |
+
rank: 0
|
70 |
+
|
71 |
+
TEST:
|
72 |
+
test_split: val-test
|
73 |
+
test_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
|
74 |
+
visualize: False
|
config/cris_verbonly_b64_nopos_nf.yaml
ADDED
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
DATA:
|
3 |
+
dataset: refcocog_u
|
4 |
+
train_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/train.lmdb
|
5 |
+
train_split: train
|
6 |
+
val_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
|
7 |
+
val_split: val
|
8 |
+
mask_root: /home/s1/chaeyunkim/VerbCentric_CY/datasets/masks/refcocog_u
|
9 |
+
|
10 |
+
TRAIN:
|
11 |
+
# Base Arch
|
12 |
+
clip_pretrain: pretrain/RN50.pt
|
13 |
+
input_size: 416
|
14 |
+
word_len: 22
|
15 |
+
word_dim: 1024
|
16 |
+
vis_dim: 512
|
17 |
+
fpn_in: [512, 1024, 1024]
|
18 |
+
fpn_out: [256, 512, 1024]
|
19 |
+
sync_bn: True
|
20 |
+
freeze: False
|
21 |
+
# Decoder
|
22 |
+
num_layers: 3
|
23 |
+
num_head: 8
|
24 |
+
dim_ffn: 2048
|
25 |
+
dropout: 0.1
|
26 |
+
intermediate: False
|
27 |
+
# Training Setting
|
28 |
+
workers: 4 # data loader workers
|
29 |
+
workers_val: 4
|
30 |
+
epochs: 50
|
31 |
+
milestones: [35]
|
32 |
+
start_epoch: 0
|
33 |
+
batch_size: 32 # batch size for training
|
34 |
+
batch_size_val: 32 # batch size for validation during training, memory and speed tradeoff
|
35 |
+
base_lr: 0.0001
|
36 |
+
lr_decay: 0.1
|
37 |
+
lr_multi: 0.1
|
38 |
+
weight_decay: 0.
|
39 |
+
max_norm: 0.
|
40 |
+
manual_seed: 0
|
41 |
+
print_freq: 100
|
42 |
+
# metric learning args
|
43 |
+
metric_learning: True
|
44 |
+
# specific metric learning args
|
45 |
+
metric_mode : 'hardpos_only' # Choice : ['hardpos_only', 'both']
|
46 |
+
exclude_multiobj : True # exclude multiobj (nobj >= 3)
|
47 |
+
exclude_pos : True # exclude multiobj w/ positional query expression (nobj >=2 && positional query)
|
48 |
+
loss_option : 'ACL_verbonly' # Choice : ['AML_verbonly', 'AML', 'ACL', 'ACL_verbonly']
|
49 |
+
metric_loss_weight: 0.1
|
50 |
+
hn_prob: 0.0
|
51 |
+
|
52 |
+
# Angular Margin Contrastive Loss argument
|
53 |
+
margin_value : 10
|
54 |
+
temperature : 0.05
|
55 |
+
|
56 |
+
# Resume & Save
|
57 |
+
exp_name: CRIS_AML_verbonly_pos25_b32
|
58 |
+
output_folder: exp/refcocog_u/exclude_multiobj
|
59 |
+
save_freq: 1
|
60 |
+
weight: # path to initial weight (default: none)
|
61 |
+
resume: "latest" # path to latest checkpoint (default: none)
|
62 |
+
evaluate: True # evaluate on validation set, extra gpu memory needed and small batch_size_val is recommend
|
63 |
+
|
64 |
+
Distributed:
|
65 |
+
dist_url: tcp://localhost:2298
|
66 |
+
dist_backend: 'nccl'
|
67 |
+
multiprocessing_distributed: True
|
68 |
+
world_size: 1
|
69 |
+
rank: 0
|
70 |
+
|
71 |
+
TEST:
|
72 |
+
test_split: val-test
|
73 |
+
test_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
|
74 |
+
visualize: False
|
config/cris_verbonly_b64_nopos_notarget.yaml
ADDED
@@ -0,0 +1,75 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
DATA:
|
3 |
+
dataset: refcocog_u
|
4 |
+
train_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/train.lmdb
|
5 |
+
train_split: train
|
6 |
+
val_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
|
7 |
+
val_split: val
|
8 |
+
mask_root: /home/s1/chaeyunkim/VerbCentric_CY/datasets/masks/refcocog_u
|
9 |
+
|
10 |
+
TRAIN:
|
11 |
+
# Base Arch
|
12 |
+
clip_pretrain: pretrain/RN50.pt
|
13 |
+
input_size: 416
|
14 |
+
word_len: 22
|
15 |
+
word_dim: 1024
|
16 |
+
vis_dim: 512
|
17 |
+
fpn_in: [512, 1024, 1024]
|
18 |
+
fpn_out: [256, 512, 1024]
|
19 |
+
sync_bn: True
|
20 |
+
freeze: True
|
21 |
+
# Decoder
|
22 |
+
num_layers: 3
|
23 |
+
num_head: 8
|
24 |
+
dim_ffn: 2048
|
25 |
+
dropout: 0.1
|
26 |
+
intermediate: False
|
27 |
+
# Training Setting
|
28 |
+
workers: 4 # data loader workers
|
29 |
+
workers_val: 4
|
30 |
+
epochs: 50
|
31 |
+
milestones: [35]
|
32 |
+
start_epoch: 0
|
33 |
+
batch_size: 32 # batch size for training
|
34 |
+
batch_size_val: 32 # batch size for validation during training, memory and speed tradeoff
|
35 |
+
base_lr: 0.0001
|
36 |
+
lr_decay: 0.1
|
37 |
+
lr_multi: 0.1
|
38 |
+
weight_decay: 0.
|
39 |
+
max_norm: 0.
|
40 |
+
manual_seed: 0
|
41 |
+
print_freq: 100
|
42 |
+
# metric learning args
|
43 |
+
metric_learning: True
|
44 |
+
# specific metric learning args
|
45 |
+
metric_mode : 'hardpos_only_rev' # Choice : ['hardpos_only', 'hardpos_only_rev', 'both']
|
46 |
+
exclude_multiobj : True # exclude multiobj (nobj >= 3)
|
47 |
+
exclude_pos : True # exclude multiobj w/ positional query expression (nobj >=2 && positional query)
|
48 |
+
loss_option : 'ACL_verbonly' # Choice : ['AML_verbonly', 'AML', 'ACL', 'ACL_verbonly']
|
49 |
+
metric_loss_weight: 0.1
|
50 |
+
hn_prob: 0.0
|
51 |
+
hn_celoss: False
|
52 |
+
|
53 |
+
# Angular Margin Contrastive Loss argument
|
54 |
+
margin_value : 10
|
55 |
+
temperature : 0.05
|
56 |
+
|
57 |
+
# Resume & Save
|
58 |
+
exp_name: CRIS_AML_verbonly_pos25_b32
|
59 |
+
output_folder: exp/refcocog_u/exclude_multiobj
|
60 |
+
save_freq: 1
|
61 |
+
weight: # path to initial weight (default: none)
|
62 |
+
resume: "latest" # path to latest checkpoint (default: none)
|
63 |
+
evaluate: True # evaluate on validation set, extra gpu memory needed and small batch_size_val is recommend
|
64 |
+
|
65 |
+
Distributed:
|
66 |
+
dist_url: tcp://localhost:2298
|
67 |
+
dist_backend: 'nccl'
|
68 |
+
multiprocessing_distributed: True
|
69 |
+
world_size: 1
|
70 |
+
rank: 0
|
71 |
+
|
72 |
+
TEST:
|
73 |
+
test_split: val-test
|
74 |
+
test_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
|
75 |
+
visualize: False
|
config/hardneg/cris_aclvo_hn_b64_nopos.yaml
ADDED
@@ -0,0 +1,80 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
DATA:
|
3 |
+
dataset: refcocog_u
|
4 |
+
train_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/train.lmdb
|
5 |
+
train_split: train
|
6 |
+
val_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
|
7 |
+
val_split: val
|
8 |
+
mask_root: /home/s1/chaeyunkim/VerbCentric_CY/datasets/masks/refcocog_u
|
9 |
+
|
10 |
+
TRAIN:
|
11 |
+
# Base Arch
|
12 |
+
clip_pretrain: pretrain/RN50.pt
|
13 |
+
input_size: 416
|
14 |
+
word_len: 22
|
15 |
+
word_dim: 1024
|
16 |
+
vis_dim: 512
|
17 |
+
fpn_in: [512, 1024, 1024]
|
18 |
+
fpn_out: [256, 512, 1024]
|
19 |
+
sync_bn: True
|
20 |
+
freeze: True
|
21 |
+
# Decoder
|
22 |
+
num_layers: 3
|
23 |
+
num_head: 8
|
24 |
+
dim_ffn: 2048
|
25 |
+
dropout: 0.1
|
26 |
+
intermediate: False
|
27 |
+
# Training Setting
|
28 |
+
workers: 4 # data loader workers
|
29 |
+
workers_val: 4
|
30 |
+
epochs: 50
|
31 |
+
milestones: [35]
|
32 |
+
start_epoch: 0
|
33 |
+
batch_size: 32 # batch size for training
|
34 |
+
batch_size_val: 32 # batch size for validation during training, memory and speed tradeoff
|
35 |
+
base_lr: 0.0001
|
36 |
+
lr_decay: 0.1
|
37 |
+
lr_multi: 0.1
|
38 |
+
weight_decay: 0.
|
39 |
+
max_norm: 0.
|
40 |
+
manual_seed: 0
|
41 |
+
print_freq: 100
|
42 |
+
# metric learning args
|
43 |
+
metric_learning: True
|
44 |
+
# specific metric learning args
|
45 |
+
metric_mode : 'hardpos_only' # Choice : ['hardpos_only', 'both']
|
46 |
+
exclude_multiobj : True # exclude multiobj (nobj >= 3)
|
47 |
+
exclude_pos : True # exclude multiobj w/ positional query expression (nobj >=2 && positional query)
|
48 |
+
loss_option : 'ACL_verbonly' # Choice : ['AML_verbonly', 'AML', 'ACL', 'ACL_verbonly']
|
49 |
+
metric_loss_weight: 0.1
|
50 |
+
|
51 |
+
# hard negative related options
|
52 |
+
hn_prob: 0.1 # hardneg percentage with regard to hard positive
|
53 |
+
acl_hn_weight : 0.1 # hardneg weight in acl loss aggregation
|
54 |
+
|
55 |
+
# hardneg no target supervision
|
56 |
+
hn_celoss : False
|
57 |
+
|
58 |
+
# Angular Margin Contrastive Loss argument
|
59 |
+
margin_value : 10
|
60 |
+
temperature : 0.05
|
61 |
+
|
62 |
+
# Resume & Save
|
63 |
+
exp_name: CRIS_AML_verbonly_pos25_b32
|
64 |
+
output_folder: exp/refcocog_u/exclude_multiobj
|
65 |
+
save_freq: 1
|
66 |
+
weight: # path to initial weight (default: none)
|
67 |
+
resume: "latest" # path to latest checkpoint (default: none)
|
68 |
+
evaluate: True # evaluate on validation set, extra gpu memory needed and small batch_size_val is recommend
|
69 |
+
|
70 |
+
Distributed:
|
71 |
+
dist_url: tcp://localhost:2298
|
72 |
+
dist_backend: 'nccl'
|
73 |
+
multiprocessing_distributed: True
|
74 |
+
world_size: 1
|
75 |
+
rank: 0
|
76 |
+
|
77 |
+
TEST:
|
78 |
+
test_split: val-test
|
79 |
+
test_lmdb: /home/s1/chaeyunkim/VerbCentric_CY/datasets/lmdb/refcocog_u/val.lmdb
|
80 |
+
visualize: False
|