ardaspear commited on
Commit
e4e2234
·
verified ·
1 Parent(s): 1c02efb

Training in progress, step 68, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7509a38f253530a371c4f75f1b5dcbea50b136a890c57a3ee3bc66254766e63f
3
  size 159967880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cbd533dccb77e0785bfd785c08160c33c48af3e1618c6fbf736240de3d65215
3
  size 159967880
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:943bc6dada57397735ca818193ed0b16c7c4f3889837989d8dfbb6097b4a594f
3
  size 81730196
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1dd6641d202d84fd9c4a931b1c22e5c8fda108a40e8f1b7c343a952a7374b7d4
3
  size 81730196
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c543948a535496f0dde46e00d90041e4b64736cb4a715d793f78a309b0450ac5
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24d2bf6b77f8850a276f23712d3998c1486d4cad7b87b73826a884c0e5e13663
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a9fa892ecd236e652150058649b13d9161331d85e374c4bec1d60ad6a556c3a9
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40aaba2d15c70d89033b94214bc8306c06d04f2fbceef0fdedf952808109d707
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.013089509143407122,
5
  "eval_steps": 17,
6
- "global_step": 51,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -158,6 +158,49 @@
158
  "eval_samples_per_second": 13.278,
159
  "eval_steps_per_second": 1.661,
160
  "step": 51
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
161
  }
162
  ],
163
  "logging_steps": 3,
@@ -177,7 +220,7 @@
177
  "attributes": {}
178
  }
179
  },
180
- "total_flos": 7.41634467692544e+16,
181
  "train_batch_size": 8,
182
  "trial_name": null,
183
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.01745267885787616,
5
  "eval_steps": 17,
6
+ "global_step": 68,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
158
  "eval_samples_per_second": 13.278,
159
  "eval_steps_per_second": 1.661,
160
  "step": 51
161
+ },
162
+ {
163
+ "epoch": 0.013859480269489894,
164
+ "grad_norm": 0.3275759220123291,
165
+ "learning_rate": 8.73410738492077e-05,
166
+ "loss": 2.3057,
167
+ "step": 54
168
+ },
169
+ {
170
+ "epoch": 0.014629451395572667,
171
+ "grad_norm": 0.29645541310310364,
172
+ "learning_rate": 8.564642241456986e-05,
173
+ "loss": 2.3326,
174
+ "step": 57
175
+ },
176
+ {
177
+ "epoch": 0.015399422521655439,
178
+ "grad_norm": 0.3317429721355438,
179
+ "learning_rate": 8.386407858128706e-05,
180
+ "loss": 2.3392,
181
+ "step": 60
182
+ },
183
+ {
184
+ "epoch": 0.01616939364773821,
185
+ "grad_norm": 0.3201993405818939,
186
+ "learning_rate": 8.199842702516583e-05,
187
+ "loss": 2.3972,
188
+ "step": 63
189
+ },
190
+ {
191
+ "epoch": 0.01693936477382098,
192
+ "grad_norm": 0.3513089716434479,
193
+ "learning_rate": 8.005405736415126e-05,
194
+ "loss": 2.3498,
195
+ "step": 66
196
+ },
197
+ {
198
+ "epoch": 0.01745267885787616,
199
+ "eval_loss": 2.3205010890960693,
200
+ "eval_runtime": 493.7323,
201
+ "eval_samples_per_second": 13.293,
202
+ "eval_steps_per_second": 1.663,
203
+ "step": 68
204
  }
205
  ],
206
  "logging_steps": 3,
 
220
  "attributes": {}
221
  }
222
  },
223
+ "total_flos": 9.855498037336474e+16,
224
  "train_batch_size": 8,
225
  "trial_name": null,
226
  "trial_params": null