lyhhan commited on
Commit
2b0a667
·
verified ·
1 Parent(s): b645977

Upload 12 files

Browse files
train/act_so100_test/checkpoints/020000/pretrained_model/config.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "act",
3
+ "n_obs_steps": 1,
4
+ "normalization_mapping": {
5
+ "VISUAL": "MEAN_STD",
6
+ "STATE": "MEAN_STD",
7
+ "ACTION": "MEAN_STD"
8
+ },
9
+ "input_features": {
10
+ "observation.state": {
11
+ "type": "STATE",
12
+ "shape": [
13
+ 6
14
+ ]
15
+ },
16
+ "observation.images.laptop": {
17
+ "type": "VISUAL",
18
+ "shape": [
19
+ 3,
20
+ 480,
21
+ 640
22
+ ]
23
+ },
24
+ "observation.images.phone": {
25
+ "type": "VISUAL",
26
+ "shape": [
27
+ 3,
28
+ 480,
29
+ 640
30
+ ]
31
+ }
32
+ },
33
+ "output_features": {
34
+ "action": {
35
+ "type": "ACTION",
36
+ "shape": [
37
+ 6
38
+ ]
39
+ }
40
+ },
41
+ "chunk_size": 100,
42
+ "n_action_steps": 100,
43
+ "vision_backbone": "resnet18",
44
+ "pretrained_backbone_weights": "ResNet18_Weights.IMAGENET1K_V1",
45
+ "replace_final_stride_with_dilation": false,
46
+ "pre_norm": false,
47
+ "dim_model": 512,
48
+ "n_heads": 8,
49
+ "dim_feedforward": 3200,
50
+ "feedforward_activation": "relu",
51
+ "n_encoder_layers": 4,
52
+ "n_decoder_layers": 1,
53
+ "use_vae": true,
54
+ "latent_dim": 32,
55
+ "n_vae_encoder_layers": 4,
56
+ "temporal_ensemble_coeff": null,
57
+ "dropout": 0.1,
58
+ "kl_weight": 10.0,
59
+ "optimizer_lr": 1e-05,
60
+ "optimizer_weight_decay": 0.0001,
61
+ "optimizer_lr_backbone": 1e-05
62
+ }
train/act_so100_test/checkpoints/020000/pretrained_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc38dabaad8b79ae7b27dc6fe16f0fcea35db50e3107cda84e0fa5000efd78e6
3
+ size 206701072
train/act_so100_test/checkpoints/020000/pretrained_model/train_config.json ADDED
@@ -0,0 +1,180 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "repo_id": "lyhhan/so100_test",
4
+ "episodes": null,
5
+ "image_transforms": {
6
+ "enable": false,
7
+ "max_num_transforms": 3,
8
+ "random_order": false,
9
+ "tfs": {
10
+ "brightness": {
11
+ "weight": 1.0,
12
+ "type": "ColorJitter",
13
+ "kwargs": {
14
+ "brightness": [
15
+ 0.8,
16
+ 1.2
17
+ ]
18
+ }
19
+ },
20
+ "contrast": {
21
+ "weight": 1.0,
22
+ "type": "ColorJitter",
23
+ "kwargs": {
24
+ "contrast": [
25
+ 0.8,
26
+ 1.2
27
+ ]
28
+ }
29
+ },
30
+ "saturation": {
31
+ "weight": 1.0,
32
+ "type": "ColorJitter",
33
+ "kwargs": {
34
+ "saturation": [
35
+ 0.5,
36
+ 1.5
37
+ ]
38
+ }
39
+ },
40
+ "hue": {
41
+ "weight": 1.0,
42
+ "type": "ColorJitter",
43
+ "kwargs": {
44
+ "hue": [
45
+ -0.05,
46
+ 0.05
47
+ ]
48
+ }
49
+ },
50
+ "sharpness": {
51
+ "weight": 1.0,
52
+ "type": "SharpnessJitter",
53
+ "kwargs": {
54
+ "sharpness": [
55
+ 0.5,
56
+ 1.5
57
+ ]
58
+ }
59
+ }
60
+ }
61
+ },
62
+ "local_files_only": false,
63
+ "use_imagenet_stats": true,
64
+ "video_backend": "pyav"
65
+ },
66
+ "env": null,
67
+ "policy": {
68
+ "type": "act",
69
+ "n_obs_steps": 1,
70
+ "normalization_mapping": {
71
+ "VISUAL": "MEAN_STD",
72
+ "STATE": "MEAN_STD",
73
+ "ACTION": "MEAN_STD"
74
+ },
75
+ "input_features": {
76
+ "observation.state": {
77
+ "type": "STATE",
78
+ "shape": [
79
+ 6
80
+ ]
81
+ },
82
+ "observation.images.laptop": {
83
+ "type": "VISUAL",
84
+ "shape": [
85
+ 3,
86
+ 480,
87
+ 640
88
+ ]
89
+ },
90
+ "observation.images.phone": {
91
+ "type": "VISUAL",
92
+ "shape": [
93
+ 3,
94
+ 480,
95
+ 640
96
+ ]
97
+ }
98
+ },
99
+ "output_features": {
100
+ "action": {
101
+ "type": "ACTION",
102
+ "shape": [
103
+ 6
104
+ ]
105
+ }
106
+ },
107
+ "chunk_size": 100,
108
+ "n_action_steps": 100,
109
+ "vision_backbone": "resnet18",
110
+ "pretrained_backbone_weights": "ResNet18_Weights.IMAGENET1K_V1",
111
+ "replace_final_stride_with_dilation": false,
112
+ "pre_norm": false,
113
+ "dim_model": 512,
114
+ "n_heads": 8,
115
+ "dim_feedforward": 3200,
116
+ "feedforward_activation": "relu",
117
+ "n_encoder_layers": 4,
118
+ "n_decoder_layers": 1,
119
+ "use_vae": true,
120
+ "latent_dim": 32,
121
+ "n_vae_encoder_layers": 4,
122
+ "temporal_ensemble_coeff": null,
123
+ "dropout": 0.1,
124
+ "kl_weight": 10.0,
125
+ "optimizer_lr": 1e-05,
126
+ "optimizer_weight_decay": 0.0001,
127
+ "optimizer_lr_backbone": 1e-05
128
+ },
129
+ "output_dir": "outputs/train/act_so100_test",
130
+ "job_name": "act_so100_test",
131
+ "resume": false,
132
+ "device": "cuda",
133
+ "use_amp": false,
134
+ "seed": 1000,
135
+ "num_workers": 4,
136
+ "batch_size": 8,
137
+ "eval_freq": 20000,
138
+ "log_freq": 200,
139
+ "save_checkpoint": true,
140
+ "save_freq": 20000,
141
+ "offline": {
142
+ "steps": 100000
143
+ },
144
+ "online": {
145
+ "steps": 0,
146
+ "rollout_n_episodes": 1,
147
+ "rollout_batch_size": 1,
148
+ "steps_between_rollouts": null,
149
+ "sampling_ratio": 0.5,
150
+ "env_seed": null,
151
+ "buffer_capacity": null,
152
+ "buffer_seed_size": 0,
153
+ "do_rollout_async": false
154
+ },
155
+ "use_policy_training_preset": true,
156
+ "optimizer": {
157
+ "type": "adamw",
158
+ "lr": 1e-05,
159
+ "weight_decay": 0.0001,
160
+ "grad_clip_norm": 10.0,
161
+ "betas": [
162
+ 0.9,
163
+ 0.999
164
+ ],
165
+ "eps": 1e-08
166
+ },
167
+ "scheduler": null,
168
+ "eval": {
169
+ "n_episodes": 50,
170
+ "batch_size": 50,
171
+ "use_async_envs": false
172
+ },
173
+ "wandb": {
174
+ "enable": false,
175
+ "disable_artifact": false,
176
+ "project": "lerobot",
177
+ "entity": null,
178
+ "notes": null
179
+ }
180
+ }
train/act_so100_test/checkpoints/020000/training_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29debbb6b8cc071638c2dc5ac9c16581735848ed0713fef5ae9a6a55c305dcb6
3
+ size 412921807
train/act_so100_test/checkpoints/040000/pretrained_model/config.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "act",
3
+ "n_obs_steps": 1,
4
+ "normalization_mapping": {
5
+ "VISUAL": "MEAN_STD",
6
+ "STATE": "MEAN_STD",
7
+ "ACTION": "MEAN_STD"
8
+ },
9
+ "input_features": {
10
+ "observation.state": {
11
+ "type": "STATE",
12
+ "shape": [
13
+ 6
14
+ ]
15
+ },
16
+ "observation.images.laptop": {
17
+ "type": "VISUAL",
18
+ "shape": [
19
+ 3,
20
+ 480,
21
+ 640
22
+ ]
23
+ },
24
+ "observation.images.phone": {
25
+ "type": "VISUAL",
26
+ "shape": [
27
+ 3,
28
+ 480,
29
+ 640
30
+ ]
31
+ }
32
+ },
33
+ "output_features": {
34
+ "action": {
35
+ "type": "ACTION",
36
+ "shape": [
37
+ 6
38
+ ]
39
+ }
40
+ },
41
+ "chunk_size": 100,
42
+ "n_action_steps": 100,
43
+ "vision_backbone": "resnet18",
44
+ "pretrained_backbone_weights": "ResNet18_Weights.IMAGENET1K_V1",
45
+ "replace_final_stride_with_dilation": false,
46
+ "pre_norm": false,
47
+ "dim_model": 512,
48
+ "n_heads": 8,
49
+ "dim_feedforward": 3200,
50
+ "feedforward_activation": "relu",
51
+ "n_encoder_layers": 4,
52
+ "n_decoder_layers": 1,
53
+ "use_vae": true,
54
+ "latent_dim": 32,
55
+ "n_vae_encoder_layers": 4,
56
+ "temporal_ensemble_coeff": null,
57
+ "dropout": 0.1,
58
+ "kl_weight": 10.0,
59
+ "optimizer_lr": 1e-05,
60
+ "optimizer_weight_decay": 0.0001,
61
+ "optimizer_lr_backbone": 1e-05
62
+ }
train/act_so100_test/checkpoints/040000/pretrained_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cbe0dc1a12eb0334c4a098c6cf5d6a9cefe0e7ccccdf516ade5e01745bebc36
3
+ size 206701072
train/act_so100_test/checkpoints/040000/pretrained_model/train_config.json ADDED
@@ -0,0 +1,180 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "repo_id": "lyhhan/so100_test",
4
+ "episodes": null,
5
+ "image_transforms": {
6
+ "enable": false,
7
+ "max_num_transforms": 3,
8
+ "random_order": false,
9
+ "tfs": {
10
+ "brightness": {
11
+ "weight": 1.0,
12
+ "type": "ColorJitter",
13
+ "kwargs": {
14
+ "brightness": [
15
+ 0.8,
16
+ 1.2
17
+ ]
18
+ }
19
+ },
20
+ "contrast": {
21
+ "weight": 1.0,
22
+ "type": "ColorJitter",
23
+ "kwargs": {
24
+ "contrast": [
25
+ 0.8,
26
+ 1.2
27
+ ]
28
+ }
29
+ },
30
+ "saturation": {
31
+ "weight": 1.0,
32
+ "type": "ColorJitter",
33
+ "kwargs": {
34
+ "saturation": [
35
+ 0.5,
36
+ 1.5
37
+ ]
38
+ }
39
+ },
40
+ "hue": {
41
+ "weight": 1.0,
42
+ "type": "ColorJitter",
43
+ "kwargs": {
44
+ "hue": [
45
+ -0.05,
46
+ 0.05
47
+ ]
48
+ }
49
+ },
50
+ "sharpness": {
51
+ "weight": 1.0,
52
+ "type": "SharpnessJitter",
53
+ "kwargs": {
54
+ "sharpness": [
55
+ 0.5,
56
+ 1.5
57
+ ]
58
+ }
59
+ }
60
+ }
61
+ },
62
+ "local_files_only": false,
63
+ "use_imagenet_stats": true,
64
+ "video_backend": "pyav"
65
+ },
66
+ "env": null,
67
+ "policy": {
68
+ "type": "act",
69
+ "n_obs_steps": 1,
70
+ "normalization_mapping": {
71
+ "VISUAL": "MEAN_STD",
72
+ "STATE": "MEAN_STD",
73
+ "ACTION": "MEAN_STD"
74
+ },
75
+ "input_features": {
76
+ "observation.state": {
77
+ "type": "STATE",
78
+ "shape": [
79
+ 6
80
+ ]
81
+ },
82
+ "observation.images.laptop": {
83
+ "type": "VISUAL",
84
+ "shape": [
85
+ 3,
86
+ 480,
87
+ 640
88
+ ]
89
+ },
90
+ "observation.images.phone": {
91
+ "type": "VISUAL",
92
+ "shape": [
93
+ 3,
94
+ 480,
95
+ 640
96
+ ]
97
+ }
98
+ },
99
+ "output_features": {
100
+ "action": {
101
+ "type": "ACTION",
102
+ "shape": [
103
+ 6
104
+ ]
105
+ }
106
+ },
107
+ "chunk_size": 100,
108
+ "n_action_steps": 100,
109
+ "vision_backbone": "resnet18",
110
+ "pretrained_backbone_weights": "ResNet18_Weights.IMAGENET1K_V1",
111
+ "replace_final_stride_with_dilation": false,
112
+ "pre_norm": false,
113
+ "dim_model": 512,
114
+ "n_heads": 8,
115
+ "dim_feedforward": 3200,
116
+ "feedforward_activation": "relu",
117
+ "n_encoder_layers": 4,
118
+ "n_decoder_layers": 1,
119
+ "use_vae": true,
120
+ "latent_dim": 32,
121
+ "n_vae_encoder_layers": 4,
122
+ "temporal_ensemble_coeff": null,
123
+ "dropout": 0.1,
124
+ "kl_weight": 10.0,
125
+ "optimizer_lr": 1e-05,
126
+ "optimizer_weight_decay": 0.0001,
127
+ "optimizer_lr_backbone": 1e-05
128
+ },
129
+ "output_dir": "outputs/train/act_so100_test",
130
+ "job_name": "act_so100_test",
131
+ "resume": false,
132
+ "device": "cuda",
133
+ "use_amp": false,
134
+ "seed": 1000,
135
+ "num_workers": 4,
136
+ "batch_size": 8,
137
+ "eval_freq": 20000,
138
+ "log_freq": 200,
139
+ "save_checkpoint": true,
140
+ "save_freq": 20000,
141
+ "offline": {
142
+ "steps": 100000
143
+ },
144
+ "online": {
145
+ "steps": 0,
146
+ "rollout_n_episodes": 1,
147
+ "rollout_batch_size": 1,
148
+ "steps_between_rollouts": null,
149
+ "sampling_ratio": 0.5,
150
+ "env_seed": null,
151
+ "buffer_capacity": null,
152
+ "buffer_seed_size": 0,
153
+ "do_rollout_async": false
154
+ },
155
+ "use_policy_training_preset": true,
156
+ "optimizer": {
157
+ "type": "adamw",
158
+ "lr": 1e-05,
159
+ "weight_decay": 0.0001,
160
+ "grad_clip_norm": 10.0,
161
+ "betas": [
162
+ 0.9,
163
+ 0.999
164
+ ],
165
+ "eps": 1e-08
166
+ },
167
+ "scheduler": null,
168
+ "eval": {
169
+ "n_episodes": 50,
170
+ "batch_size": 50,
171
+ "use_async_envs": false
172
+ },
173
+ "wandb": {
174
+ "enable": false,
175
+ "disable_artifact": false,
176
+ "project": "lerobot",
177
+ "entity": null,
178
+ "notes": null
179
+ }
180
+ }
train/act_so100_test/checkpoints/040000/training_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93072a350c00fd8fab79ae4f508d9afc325738f96e5855b9f858cef41b5b4c8b
3
+ size 412921807
train/act_so100_test/checkpoints/060000/pretrained_model/config.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "act",
3
+ "n_obs_steps": 1,
4
+ "normalization_mapping": {
5
+ "VISUAL": "MEAN_STD",
6
+ "STATE": "MEAN_STD",
7
+ "ACTION": "MEAN_STD"
8
+ },
9
+ "input_features": {
10
+ "observation.state": {
11
+ "type": "STATE",
12
+ "shape": [
13
+ 6
14
+ ]
15
+ },
16
+ "observation.images.laptop": {
17
+ "type": "VISUAL",
18
+ "shape": [
19
+ 3,
20
+ 480,
21
+ 640
22
+ ]
23
+ },
24
+ "observation.images.phone": {
25
+ "type": "VISUAL",
26
+ "shape": [
27
+ 3,
28
+ 480,
29
+ 640
30
+ ]
31
+ }
32
+ },
33
+ "output_features": {
34
+ "action": {
35
+ "type": "ACTION",
36
+ "shape": [
37
+ 6
38
+ ]
39
+ }
40
+ },
41
+ "chunk_size": 100,
42
+ "n_action_steps": 100,
43
+ "vision_backbone": "resnet18",
44
+ "pretrained_backbone_weights": "ResNet18_Weights.IMAGENET1K_V1",
45
+ "replace_final_stride_with_dilation": false,
46
+ "pre_norm": false,
47
+ "dim_model": 512,
48
+ "n_heads": 8,
49
+ "dim_feedforward": 3200,
50
+ "feedforward_activation": "relu",
51
+ "n_encoder_layers": 4,
52
+ "n_decoder_layers": 1,
53
+ "use_vae": true,
54
+ "latent_dim": 32,
55
+ "n_vae_encoder_layers": 4,
56
+ "temporal_ensemble_coeff": null,
57
+ "dropout": 0.1,
58
+ "kl_weight": 10.0,
59
+ "optimizer_lr": 1e-05,
60
+ "optimizer_weight_decay": 0.0001,
61
+ "optimizer_lr_backbone": 1e-05
62
+ }
train/act_so100_test/checkpoints/060000/pretrained_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17ce679babd29883b4549ffa1bb53dad0cedc96a77a8063b81a5b570a3adfaaa
3
+ size 206701072
train/act_so100_test/checkpoints/060000/pretrained_model/train_config.json ADDED
@@ -0,0 +1,180 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "repo_id": "lyhhan/so100_test",
4
+ "episodes": null,
5
+ "image_transforms": {
6
+ "enable": false,
7
+ "max_num_transforms": 3,
8
+ "random_order": false,
9
+ "tfs": {
10
+ "brightness": {
11
+ "weight": 1.0,
12
+ "type": "ColorJitter",
13
+ "kwargs": {
14
+ "brightness": [
15
+ 0.8,
16
+ 1.2
17
+ ]
18
+ }
19
+ },
20
+ "contrast": {
21
+ "weight": 1.0,
22
+ "type": "ColorJitter",
23
+ "kwargs": {
24
+ "contrast": [
25
+ 0.8,
26
+ 1.2
27
+ ]
28
+ }
29
+ },
30
+ "saturation": {
31
+ "weight": 1.0,
32
+ "type": "ColorJitter",
33
+ "kwargs": {
34
+ "saturation": [
35
+ 0.5,
36
+ 1.5
37
+ ]
38
+ }
39
+ },
40
+ "hue": {
41
+ "weight": 1.0,
42
+ "type": "ColorJitter",
43
+ "kwargs": {
44
+ "hue": [
45
+ -0.05,
46
+ 0.05
47
+ ]
48
+ }
49
+ },
50
+ "sharpness": {
51
+ "weight": 1.0,
52
+ "type": "SharpnessJitter",
53
+ "kwargs": {
54
+ "sharpness": [
55
+ 0.5,
56
+ 1.5
57
+ ]
58
+ }
59
+ }
60
+ }
61
+ },
62
+ "local_files_only": false,
63
+ "use_imagenet_stats": true,
64
+ "video_backend": "pyav"
65
+ },
66
+ "env": null,
67
+ "policy": {
68
+ "type": "act",
69
+ "n_obs_steps": 1,
70
+ "normalization_mapping": {
71
+ "VISUAL": "MEAN_STD",
72
+ "STATE": "MEAN_STD",
73
+ "ACTION": "MEAN_STD"
74
+ },
75
+ "input_features": {
76
+ "observation.state": {
77
+ "type": "STATE",
78
+ "shape": [
79
+ 6
80
+ ]
81
+ },
82
+ "observation.images.laptop": {
83
+ "type": "VISUAL",
84
+ "shape": [
85
+ 3,
86
+ 480,
87
+ 640
88
+ ]
89
+ },
90
+ "observation.images.phone": {
91
+ "type": "VISUAL",
92
+ "shape": [
93
+ 3,
94
+ 480,
95
+ 640
96
+ ]
97
+ }
98
+ },
99
+ "output_features": {
100
+ "action": {
101
+ "type": "ACTION",
102
+ "shape": [
103
+ 6
104
+ ]
105
+ }
106
+ },
107
+ "chunk_size": 100,
108
+ "n_action_steps": 100,
109
+ "vision_backbone": "resnet18",
110
+ "pretrained_backbone_weights": "ResNet18_Weights.IMAGENET1K_V1",
111
+ "replace_final_stride_with_dilation": false,
112
+ "pre_norm": false,
113
+ "dim_model": 512,
114
+ "n_heads": 8,
115
+ "dim_feedforward": 3200,
116
+ "feedforward_activation": "relu",
117
+ "n_encoder_layers": 4,
118
+ "n_decoder_layers": 1,
119
+ "use_vae": true,
120
+ "latent_dim": 32,
121
+ "n_vae_encoder_layers": 4,
122
+ "temporal_ensemble_coeff": null,
123
+ "dropout": 0.1,
124
+ "kl_weight": 10.0,
125
+ "optimizer_lr": 1e-05,
126
+ "optimizer_weight_decay": 0.0001,
127
+ "optimizer_lr_backbone": 1e-05
128
+ },
129
+ "output_dir": "outputs/train/act_so100_test",
130
+ "job_name": "act_so100_test",
131
+ "resume": false,
132
+ "device": "cuda",
133
+ "use_amp": false,
134
+ "seed": 1000,
135
+ "num_workers": 4,
136
+ "batch_size": 8,
137
+ "eval_freq": 20000,
138
+ "log_freq": 200,
139
+ "save_checkpoint": true,
140
+ "save_freq": 20000,
141
+ "offline": {
142
+ "steps": 100000
143
+ },
144
+ "online": {
145
+ "steps": 0,
146
+ "rollout_n_episodes": 1,
147
+ "rollout_batch_size": 1,
148
+ "steps_between_rollouts": null,
149
+ "sampling_ratio": 0.5,
150
+ "env_seed": null,
151
+ "buffer_capacity": null,
152
+ "buffer_seed_size": 0,
153
+ "do_rollout_async": false
154
+ },
155
+ "use_policy_training_preset": true,
156
+ "optimizer": {
157
+ "type": "adamw",
158
+ "lr": 1e-05,
159
+ "weight_decay": 0.0001,
160
+ "grad_clip_norm": 10.0,
161
+ "betas": [
162
+ 0.9,
163
+ 0.999
164
+ ],
165
+ "eps": 1e-08
166
+ },
167
+ "scheduler": null,
168
+ "eval": {
169
+ "n_episodes": 50,
170
+ "batch_size": 50,
171
+ "use_async_envs": false
172
+ },
173
+ "wandb": {
174
+ "enable": false,
175
+ "disable_artifact": false,
176
+ "project": "lerobot",
177
+ "entity": null,
178
+ "notes": null
179
+ }
180
+ }
train/act_so100_test/checkpoints/060000/training_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:718a60e4097bc6ddeba0696f55a48c0be2289265f095bbbd7e02fd96489ba203
3
+ size 412921807