besimray commited on
Commit
ba8a886
·
verified ·
1 Parent(s): 23e9f83

Training in progress, step 160, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:05a9134086f04aaf109815906f18423ab3ce8d6e505044a2e436597e63ac66aa
3
  size 335604696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2737a5b2b664488a6760720bf8f7b0800c4aff5686b549441f71a51dbf2936e
3
  size 335604696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed42d45141484ad4e2d377a32d3d78fbcbe08ce9af63a5bdf9c52376535118da
3
  size 170920084
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89f99cde8e0d4c1a5d284b765033e4de43b51035c42dbac3779173361b1246f6
3
  size 170920084
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:25331c901a2b44ead3ad152cdbe6be884abba42a0b6e921dadb3a67ac6e4e427
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22baeb617fbf458c1c774fdf5681592667b128ead691c19e018ef96010b3aff7
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8fa6a7a6709edeb55cdf23229934c07be2d8aae0e4056fbdb6ff2482d0eb3d3
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac2f052986474c47aa92ee09f81c986cf5cd3c01bde827e5d887e85b2bbda4c2
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 1.533050298690796,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-130",
4
- "epoch": 0.21178962230850687,
5
  "eval_steps": 10,
6
- "global_step": 150,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1185,6 +1185,84 @@
1185
  "eval_samples_per_second": 1.935,
1186
  "eval_steps_per_second": 0.389,
1187
  "step": 150
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1188
  }
1189
  ],
1190
  "logging_steps": 1,
@@ -1199,7 +1277,7 @@
1199
  "early_stopping_threshold": 0.0
1200
  },
1201
  "attributes": {
1202
- "early_stopping_patience_counter": 2
1203
  }
1204
  },
1205
  "TrainerControl": {
@@ -1208,12 +1286,12 @@
1208
  "should_evaluate": false,
1209
  "should_log": false,
1210
  "should_save": true,
1211
- "should_training_stop": false
1212
  },
1213
  "attributes": {}
1214
  }
1215
  },
1216
- "total_flos": 5.59507839123456e+17,
1217
  "train_batch_size": 5,
1218
  "trial_name": null,
1219
  "trial_params": null
 
1
  {
2
  "best_metric": 1.533050298690796,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-130",
4
+ "epoch": 0.22590893046240734,
5
  "eval_steps": 10,
6
+ "global_step": 160,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1185
  "eval_samples_per_second": 1.935,
1186
  "eval_steps_per_second": 0.389,
1187
  "step": 150
1188
+ },
1189
+ {
1190
+ "epoch": 0.21320155312389694,
1191
+ "grad_norm": 0.4451688826084137,
1192
+ "learning_rate": 0.00016184643766056317,
1193
+ "loss": 1.3055,
1194
+ "step": 151
1195
+ },
1196
+ {
1197
+ "epoch": 0.21461348393928698,
1198
+ "grad_norm": 0.4681131839752197,
1199
+ "learning_rate": 0.00016134135287043669,
1200
+ "loss": 1.6771,
1201
+ "step": 152
1202
+ },
1203
+ {
1204
+ "epoch": 0.21602541475467701,
1205
+ "grad_norm": 0.3876592218875885,
1206
+ "learning_rate": 0.00016083374657755134,
1207
+ "loss": 1.945,
1208
+ "step": 153
1209
+ },
1210
+ {
1211
+ "epoch": 0.21743734557006705,
1212
+ "grad_norm": 0.45364028215408325,
1213
+ "learning_rate": 0.00016032363964761363,
1214
+ "loss": 1.6876,
1215
+ "step": 154
1216
+ },
1217
+ {
1218
+ "epoch": 0.21884927638545712,
1219
+ "grad_norm": 0.3172266185283661,
1220
+ "learning_rate": 0.00015981105304912162,
1221
+ "loss": 1.7737,
1222
+ "step": 155
1223
+ },
1224
+ {
1225
+ "epoch": 0.22026120720084716,
1226
+ "grad_norm": 0.3841560184955597,
1227
+ "learning_rate": 0.00015929600785250257,
1228
+ "loss": 1.6019,
1229
+ "step": 156
1230
+ },
1231
+ {
1232
+ "epoch": 0.2216731380162372,
1233
+ "grad_norm": 0.39578792452812195,
1234
+ "learning_rate": 0.00015877852522924732,
1235
+ "loss": 1.8585,
1236
+ "step": 157
1237
+ },
1238
+ {
1239
+ "epoch": 0.22308506883162724,
1240
+ "grad_norm": 0.4143383800983429,
1241
+ "learning_rate": 0.0001582586264510396,
1242
+ "loss": 1.6767,
1243
+ "step": 158
1244
+ },
1245
+ {
1246
+ "epoch": 0.2244969996470173,
1247
+ "grad_norm": 0.43308305740356445,
1248
+ "learning_rate": 0.00015773633288888197,
1249
+ "loss": 1.7304,
1250
+ "step": 159
1251
+ },
1252
+ {
1253
+ "epoch": 0.22590893046240734,
1254
+ "grad_norm": 0.3537898063659668,
1255
+ "learning_rate": 0.00015721166601221698,
1256
+ "loss": 1.0026,
1257
+ "step": 160
1258
+ },
1259
+ {
1260
+ "epoch": 0.22590893046240734,
1261
+ "eval_loss": 1.5331158638000488,
1262
+ "eval_runtime": 385.352,
1263
+ "eval_samples_per_second": 1.936,
1264
+ "eval_steps_per_second": 0.389,
1265
+ "step": 160
1266
  }
1267
  ],
1268
  "logging_steps": 1,
 
1277
  "early_stopping_threshold": 0.0
1278
  },
1279
  "attributes": {
1280
+ "early_stopping_patience_counter": 3
1281
  }
1282
  },
1283
  "TrainerControl": {
 
1286
  "should_evaluate": false,
1287
  "should_log": false,
1288
  "should_save": true,
1289
+ "should_training_stop": true
1290
  },
1291
  "attributes": {}
1292
  }
1293
  },
1294
+ "total_flos": 5.968083617316864e+17,
1295
  "train_batch_size": 5,
1296
  "trial_name": null,
1297
  "trial_params": null