MohamedAhmedAE commited on
Commit
14f01e3
·
verified ·
1 Parent(s): 738dff5

Training in progress, step 21000, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -25,13 +25,13 @@
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
 
28
  "o_proj",
29
- "down_proj",
30
  "q_proj",
31
- "k_proj",
32
- "up_proj",
33
  "v_proj",
34
- "gate_proj"
 
 
35
  ],
36
  "target_parameters": null,
37
  "task_type": "CAUSAL_LM",
 
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
+ "gate_proj",
29
  "o_proj",
 
30
  "q_proj",
 
 
31
  "v_proj",
32
+ "k_proj",
33
+ "down_proj",
34
+ "up_proj"
35
  ],
36
  "target_parameters": null,
37
  "task_type": "CAUSAL_LM",
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:883db002a6734cb265864ebc120f2884e77f649b30dae4a2302eefca4d4137bb
3
  size 389081912
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76d220f6eb62d357cac204cc6ca10e112229c5ea4f345c3e50d0cabaf1919126
3
  size 389081912
last-checkpoint/global_step21000/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc38d7765eee46a46c1cfb655ca1ec93f1839468d5fa79206d9b09ba7266ce4e
3
+ size 1167094117
last-checkpoint/global_step21000/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:094b08b7e988cbbcc8f171789dc77f5095d58244f35d03ba5a4ff32417e44eee
3
+ size 1167094245
last-checkpoint/global_step21000/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3857c1c2fa6aa57562e840fc55be0e457a7e501efdbf0d5d16ac48b30f2a4cb2
3
+ size 1222740115
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step4600
 
1
+ global_step21000
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c29caf33459e77925ce566a510af138891ad4476e5bca606a1c814b77305442
3
  size 14917
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f188a91a21d29457226edd00f5fcf5c04d693fe2d994fee71c314fba23852b9
3
  size 14917
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1d6eccb59d533bc2d9720c2f64f56a9affd86293f6e5cde1146191b5f72a2bbb
3
  size 14917
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d782a09a6bb8b2213c345e44b4eebe638d7afd04a217515bee34b8c83445597c
3
  size 14917
last-checkpoint/special_tokens_map.json CHANGED
@@ -13,5 +13,11 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": "<|eot_id|>"
 
 
 
 
 
 
17
  }
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": {
17
+ "content": "<|eot_id|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
  }
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 0.6345265190702807,
6
  "eval_steps": 500,
7
- "global_step": 4600,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -215,6 +215,744 @@
215
  "mean_token_accuracy": 0.6035999485850334,
216
  "num_tokens": 299295385.0,
217
  "step": 4600
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
218
  }
219
  ],
220
  "logging_steps": 200,
@@ -234,7 +972,7 @@
234
  "attributes": {}
235
  }
236
  },
237
- "total_flos": 5.41112571879909e+18,
238
  "train_batch_size": 4,
239
  "trial_name": null,
240
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 2.8967515001034556,
6
  "eval_steps": 500,
7
+ "global_step": 21000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
215
  "mean_token_accuracy": 0.6035999485850334,
216
  "num_tokens": 299295385.0,
217
  "step": 4600
218
+ },
219
+ {
220
+ "epoch": 0.6621146285950755,
221
+ "grad_norm": 0.2658919095993042,
222
+ "learning_rate": 1.735466960960133e-05,
223
+ "loss": 1.7675,
224
+ "mean_token_accuracy": 0.6052964499592781,
225
+ "num_tokens": 13006009.0,
226
+ "step": 4800
227
+ },
228
+ {
229
+ "epoch": 0.6897027381198704,
230
+ "grad_norm": 0.3054528534412384,
231
+ "learning_rate": 1.7244309559939303e-05,
232
+ "loss": 1.7733,
233
+ "mean_token_accuracy": 0.6038838358223438,
234
+ "num_tokens": 26022591.0,
235
+ "step": 5000
236
+ },
237
+ {
238
+ "epoch": 0.7172908476446651,
239
+ "grad_norm": 0.2557850778102875,
240
+ "learning_rate": 1.7133949510277282e-05,
241
+ "loss": 1.7678,
242
+ "mean_token_accuracy": 0.6045010414719582,
243
+ "num_tokens": 39031786.0,
244
+ "step": 5200
245
+ },
246
+ {
247
+ "epoch": 0.74487895716946,
248
+ "grad_norm": 0.2966762185096741,
249
+ "learning_rate": 1.702358946061526e-05,
250
+ "loss": 1.7645,
251
+ "mean_token_accuracy": 0.6054365809261799,
252
+ "num_tokens": 52042866.0,
253
+ "step": 5400
254
+ },
255
+ {
256
+ "epoch": 0.7724670666942548,
257
+ "grad_norm": 0.2732425630092621,
258
+ "learning_rate": 1.6913229410953235e-05,
259
+ "loss": 1.7647,
260
+ "mean_token_accuracy": 0.6047625686228275,
261
+ "num_tokens": 65053898.0,
262
+ "step": 5600
263
+ },
264
+ {
265
+ "epoch": 0.8000551762190495,
266
+ "grad_norm": 0.27680376172065735,
267
+ "learning_rate": 1.6802869361291214e-05,
268
+ "loss": 1.7642,
269
+ "mean_token_accuracy": 0.6054388484358788,
270
+ "num_tokens": 78051546.0,
271
+ "step": 5800
272
+ },
273
+ {
274
+ "epoch": 0.8276432857438444,
275
+ "grad_norm": 0.2786679267883301,
276
+ "learning_rate": 1.6692509311629192e-05,
277
+ "loss": 1.7653,
278
+ "mean_token_accuracy": 0.6052116820216179,
279
+ "num_tokens": 91069298.0,
280
+ "step": 6000
281
+ },
282
+ {
283
+ "epoch": 0.8552313952686392,
284
+ "grad_norm": 0.2635458707809448,
285
+ "learning_rate": 1.6582149261967168e-05,
286
+ "loss": 1.7648,
287
+ "mean_token_accuracy": 0.6061200417578221,
288
+ "num_tokens": 104059318.0,
289
+ "step": 6200
290
+ },
291
+ {
292
+ "epoch": 0.8828195047934341,
293
+ "grad_norm": 0.2765181362628937,
294
+ "learning_rate": 1.6471789212305146e-05,
295
+ "loss": 1.7667,
296
+ "mean_token_accuracy": 0.6046179200708867,
297
+ "num_tokens": 117087631.0,
298
+ "step": 6400
299
+ },
300
+ {
301
+ "epoch": 0.9104076143182288,
302
+ "grad_norm": 0.28083258867263794,
303
+ "learning_rate": 1.6361429162643124e-05,
304
+ "loss": 1.7612,
305
+ "mean_token_accuracy": 0.6063811640441418,
306
+ "num_tokens": 130092049.0,
307
+ "step": 6600
308
+ },
309
+ {
310
+ "epoch": 0.9379957238430237,
311
+ "grad_norm": 0.2751764953136444,
312
+ "learning_rate": 1.6251069112981103e-05,
313
+ "loss": 1.7509,
314
+ "mean_token_accuracy": 0.6079460391402245,
315
+ "num_tokens": 143094014.0,
316
+ "step": 6800
317
+ },
318
+ {
319
+ "epoch": 0.9655838333678185,
320
+ "grad_norm": 0.2823185920715332,
321
+ "learning_rate": 1.614070906331908e-05,
322
+ "loss": 1.7557,
323
+ "mean_token_accuracy": 0.606988143324852,
324
+ "num_tokens": 156090713.0,
325
+ "step": 7000
326
+ },
327
+ {
328
+ "epoch": 0.9931719428926132,
329
+ "grad_norm": 0.2630833387374878,
330
+ "learning_rate": 1.603034901365706e-05,
331
+ "loss": 1.7554,
332
+ "mean_token_accuracy": 0.6067783555388451,
333
+ "num_tokens": 169115214.0,
334
+ "step": 7200
335
+ },
336
+ {
337
+ "epoch": 1.02082902269122,
338
+ "grad_norm": 0.24643921852111816,
339
+ "learning_rate": 1.5919988963995035e-05,
340
+ "loss": 1.7542,
341
+ "mean_token_accuracy": 0.6083373454146255,
342
+ "num_tokens": 182136269.0,
343
+ "step": 7400
344
+ },
345
+ {
346
+ "epoch": 1.048417132216015,
347
+ "grad_norm": 0.2477455735206604,
348
+ "learning_rate": 1.5809628914333013e-05,
349
+ "loss": 1.7316,
350
+ "mean_token_accuracy": 0.6105767801404,
351
+ "num_tokens": 195134342.0,
352
+ "step": 7600
353
+ },
354
+ {
355
+ "epoch": 1.0760052417408097,
356
+ "grad_norm": 0.2505553066730499,
357
+ "learning_rate": 1.5699268864670992e-05,
358
+ "loss": 1.7412,
359
+ "mean_token_accuracy": 0.6090090696513653,
360
+ "num_tokens": 208152063.0,
361
+ "step": 7800
362
+ },
363
+ {
364
+ "epoch": 1.1035933512656044,
365
+ "grad_norm": 0.25275516510009766,
366
+ "learning_rate": 1.5588908815008967e-05,
367
+ "loss": 1.7389,
368
+ "mean_token_accuracy": 0.6091519020497799,
369
+ "num_tokens": 221135506.0,
370
+ "step": 8000
371
+ },
372
+ {
373
+ "epoch": 1.1311814607903994,
374
+ "grad_norm": 0.256531685590744,
375
+ "learning_rate": 1.5478548765346945e-05,
376
+ "loss": 1.738,
377
+ "mean_token_accuracy": 0.6093683485686779,
378
+ "num_tokens": 234118407.0,
379
+ "step": 8200
380
+ },
381
+ {
382
+ "epoch": 1.1587695703151941,
383
+ "grad_norm": 0.26037585735321045,
384
+ "learning_rate": 1.5368188715684924e-05,
385
+ "loss": 1.7338,
386
+ "mean_token_accuracy": 0.610037949681282,
387
+ "num_tokens": 247128181.0,
388
+ "step": 8400
389
+ },
390
+ {
391
+ "epoch": 1.1863576798399889,
392
+ "grad_norm": 0.2678148150444031,
393
+ "learning_rate": 1.52578286660229e-05,
394
+ "loss": 1.7367,
395
+ "mean_token_accuracy": 0.609309321641922,
396
+ "num_tokens": 260146123.0,
397
+ "step": 8600
398
+ },
399
+ {
400
+ "epoch": 1.2139457893647838,
401
+ "grad_norm": 0.2542251646518707,
402
+ "learning_rate": 1.5147468616360879e-05,
403
+ "loss": 1.7395,
404
+ "mean_token_accuracy": 0.6088805337250233,
405
+ "num_tokens": 273180867.0,
406
+ "step": 8800
407
+ },
408
+ {
409
+ "epoch": 1.2415338988895785,
410
+ "grad_norm": 0.25549325346946716,
411
+ "learning_rate": 1.5037108566698858e-05,
412
+ "loss": 1.7278,
413
+ "mean_token_accuracy": 0.6110722988843917,
414
+ "num_tokens": 286192947.0,
415
+ "step": 9000
416
+ },
417
+ {
418
+ "epoch": 1.2691220084143735,
419
+ "grad_norm": 0.26431170105934143,
420
+ "learning_rate": 1.4926748517036833e-05,
421
+ "loss": 1.7293,
422
+ "mean_token_accuracy": 0.6109644013643265,
423
+ "num_tokens": 299187058.0,
424
+ "step": 9200
425
+ },
426
+ {
427
+ "epoch": 1.2967101179391682,
428
+ "grad_norm": 0.25987866520881653,
429
+ "learning_rate": 1.4816388467374811e-05,
430
+ "loss": 1.7362,
431
+ "mean_token_accuracy": 0.6091528676450253,
432
+ "num_tokens": 312229168.0,
433
+ "step": 9400
434
+ },
435
+ {
436
+ "epoch": 1.324298227463963,
437
+ "grad_norm": 0.2580626308917999,
438
+ "learning_rate": 1.470602841771279e-05,
439
+ "loss": 1.7298,
440
+ "mean_token_accuracy": 0.6105604329705239,
441
+ "num_tokens": 325252754.0,
442
+ "step": 9600
443
+ },
444
+ {
445
+ "epoch": 1.3518863369887577,
446
+ "grad_norm": 0.26252448558807373,
447
+ "learning_rate": 1.4595668368050766e-05,
448
+ "loss": 1.7258,
449
+ "mean_token_accuracy": 0.6110147424042225,
450
+ "num_tokens": 338265857.0,
451
+ "step": 9800
452
+ },
453
+ {
454
+ "epoch": 1.3794744465135527,
455
+ "grad_norm": 0.267373263835907,
456
+ "learning_rate": 1.4485308318388745e-05,
457
+ "loss": 1.73,
458
+ "mean_token_accuracy": 0.6104732736945152,
459
+ "num_tokens": 351268079.0,
460
+ "step": 10000
461
+ },
462
+ {
463
+ "epoch": 1.4070625560383474,
464
+ "grad_norm": 0.2667401134967804,
465
+ "learning_rate": 1.4374948268726723e-05,
466
+ "loss": 1.732,
467
+ "mean_token_accuracy": 0.6098475851118564,
468
+ "num_tokens": 364302227.0,
469
+ "step": 10200
470
+ },
471
+ {
472
+ "epoch": 1.4346506655631424,
473
+ "grad_norm": 0.26889705657958984,
474
+ "learning_rate": 1.4264588219064698e-05,
475
+ "loss": 1.731,
476
+ "mean_token_accuracy": 0.6093322136998176,
477
+ "num_tokens": 377338727.0,
478
+ "step": 10400
479
+ },
480
+ {
481
+ "epoch": 1.462238775087937,
482
+ "grad_norm": 0.2645055055618286,
483
+ "learning_rate": 1.4154228169402677e-05,
484
+ "loss": 1.7287,
485
+ "mean_token_accuracy": 0.6106591558456421,
486
+ "num_tokens": 390354064.0,
487
+ "step": 10600
488
+ },
489
+ {
490
+ "epoch": 1.4898268846127318,
491
+ "grad_norm": 0.268999308347702,
492
+ "learning_rate": 1.4043868119740655e-05,
493
+ "loss": 1.7223,
494
+ "mean_token_accuracy": 0.6116151364147663,
495
+ "num_tokens": 403371674.0,
496
+ "step": 10800
497
+ },
498
+ {
499
+ "epoch": 1.5174149941375266,
500
+ "grad_norm": 0.2641783356666565,
501
+ "learning_rate": 1.3933508070078632e-05,
502
+ "loss": 1.7198,
503
+ "mean_token_accuracy": 0.6120837351679802,
504
+ "num_tokens": 416381735.0,
505
+ "step": 11000
506
+ },
507
+ {
508
+ "epoch": 1.5450031036623215,
509
+ "grad_norm": 0.26654648780822754,
510
+ "learning_rate": 1.382314802041661e-05,
511
+ "loss": 1.7272,
512
+ "mean_token_accuracy": 0.6110439296066761,
513
+ "num_tokens": 429385525.0,
514
+ "step": 11200
515
+ },
516
+ {
517
+ "epoch": 1.5725912131871165,
518
+ "grad_norm": 0.2701609134674072,
519
+ "learning_rate": 1.3712787970754589e-05,
520
+ "loss": 1.7308,
521
+ "mean_token_accuracy": 0.610301011800766,
522
+ "num_tokens": 442382821.0,
523
+ "step": 11400
524
+ },
525
+ {
526
+ "epoch": 1.6001793227119112,
527
+ "grad_norm": 0.26637518405914307,
528
+ "learning_rate": 1.3602427921092564e-05,
529
+ "loss": 1.7314,
530
+ "mean_token_accuracy": 0.6097376590967178,
531
+ "num_tokens": 455417304.0,
532
+ "step": 11600
533
+ },
534
+ {
535
+ "epoch": 1.627767432236706,
536
+ "grad_norm": 0.2647789716720581,
537
+ "learning_rate": 1.3492067871430543e-05,
538
+ "loss": 1.7256,
539
+ "mean_token_accuracy": 0.6110027520358563,
540
+ "num_tokens": 468439793.0,
541
+ "step": 11800
542
+ },
543
+ {
544
+ "epoch": 1.6553555417615007,
545
+ "grad_norm": 0.26624011993408203,
546
+ "learning_rate": 1.3381707821768521e-05,
547
+ "loss": 1.7243,
548
+ "mean_token_accuracy": 0.6117079831659794,
549
+ "num_tokens": 481425220.0,
550
+ "step": 12000
551
+ },
552
+ {
553
+ "epoch": 1.6829436512862956,
554
+ "grad_norm": 0.2641337215900421,
555
+ "learning_rate": 1.3271347772106498e-05,
556
+ "loss": 1.718,
557
+ "mean_token_accuracy": 0.6129114121198654,
558
+ "num_tokens": 494414024.0,
559
+ "step": 12200
560
+ },
561
+ {
562
+ "epoch": 1.7105317608110904,
563
+ "grad_norm": 0.2691522538661957,
564
+ "learning_rate": 1.3160987722444476e-05,
565
+ "loss": 1.7262,
566
+ "mean_token_accuracy": 0.6111969475448131,
567
+ "num_tokens": 507414476.0,
568
+ "step": 12400
569
+ },
570
+ {
571
+ "epoch": 1.7381198703358853,
572
+ "grad_norm": 0.2704949676990509,
573
+ "learning_rate": 1.3050627672782455e-05,
574
+ "loss": 1.7212,
575
+ "mean_token_accuracy": 0.6119842390716076,
576
+ "num_tokens": 520416862.0,
577
+ "step": 12600
578
+ },
579
+ {
580
+ "epoch": 1.76570797986068,
581
+ "grad_norm": 0.2640060782432556,
582
+ "learning_rate": 1.294026762312043e-05,
583
+ "loss": 1.7197,
584
+ "mean_token_accuracy": 0.6121780882775784,
585
+ "num_tokens": 533414023.0,
586
+ "step": 12800
587
+ },
588
+ {
589
+ "epoch": 1.7932960893854748,
590
+ "grad_norm": 0.2656666338443756,
591
+ "learning_rate": 1.2829907573458408e-05,
592
+ "loss": 1.7166,
593
+ "mean_token_accuracy": 0.6125291535258293,
594
+ "num_tokens": 546421754.0,
595
+ "step": 13000
596
+ },
597
+ {
598
+ "epoch": 1.8208841989102695,
599
+ "grad_norm": 0.2837753891944885,
600
+ "learning_rate": 1.2719547523796387e-05,
601
+ "loss": 1.7143,
602
+ "mean_token_accuracy": 0.6129081077873707,
603
+ "num_tokens": 559432463.0,
604
+ "step": 13200
605
+ },
606
+ {
607
+ "epoch": 1.8484723084350645,
608
+ "grad_norm": 0.27201464772224426,
609
+ "learning_rate": 1.2609187474134364e-05,
610
+ "loss": 1.7167,
611
+ "mean_token_accuracy": 0.6130946539342403,
612
+ "num_tokens": 572437690.0,
613
+ "step": 13400
614
+ },
615
+ {
616
+ "epoch": 1.8760604179598594,
617
+ "grad_norm": 0.26749759912490845,
618
+ "learning_rate": 1.2498827424472342e-05,
619
+ "loss": 1.7224,
620
+ "mean_token_accuracy": 0.6114400824904442,
621
+ "num_tokens": 585452377.0,
622
+ "step": 13600
623
+ },
624
+ {
625
+ "epoch": 1.9036485274846542,
626
+ "grad_norm": 0.2753259241580963,
627
+ "learning_rate": 1.238846737481032e-05,
628
+ "loss": 1.7193,
629
+ "mean_token_accuracy": 0.6123683395981788,
630
+ "num_tokens": 598467233.0,
631
+ "step": 13800
632
+ },
633
+ {
634
+ "epoch": 1.931236637009449,
635
+ "grad_norm": 0.2765011489391327,
636
+ "learning_rate": 1.2278107325148296e-05,
637
+ "loss": 1.7185,
638
+ "mean_token_accuracy": 0.6119060663878918,
639
+ "num_tokens": 611505054.0,
640
+ "step": 14000
641
+ },
642
+ {
643
+ "epoch": 1.9588247465342437,
644
+ "grad_norm": 0.29317325353622437,
645
+ "learning_rate": 1.2167747275486274e-05,
646
+ "loss": 1.7164,
647
+ "mean_token_accuracy": 0.6127070464193821,
648
+ "num_tokens": 624515618.0,
649
+ "step": 14200
650
+ },
651
+ {
652
+ "epoch": 1.9864128560590384,
653
+ "grad_norm": 0.27798038721084595,
654
+ "learning_rate": 1.2057387225824253e-05,
655
+ "loss": 1.7148,
656
+ "mean_token_accuracy": 0.6131316345930099,
657
+ "num_tokens": 637507621.0,
658
+ "step": 14400
659
+ },
660
+ {
661
+ "epoch": 2.0139319953100214,
662
+ "grad_norm": 0.28348833322525024,
663
+ "learning_rate": 1.194702717616223e-05,
664
+ "loss": 1.7077,
665
+ "mean_token_accuracy": 0.6135270874900627,
666
+ "num_tokens": 650501333.0,
667
+ "step": 14600
668
+ },
669
+ {
670
+ "epoch": 2.041520104834816,
671
+ "grad_norm": 0.2825703024864197,
672
+ "learning_rate": 1.1836667126500208e-05,
673
+ "loss": 1.7,
674
+ "mean_token_accuracy": 0.6154656463861465,
675
+ "num_tokens": 663491183.0,
676
+ "step": 14800
677
+ },
678
+ {
679
+ "epoch": 2.069108214359611,
680
+ "grad_norm": 0.2799076437950134,
681
+ "learning_rate": 1.1726307076838186e-05,
682
+ "loss": 1.6941,
683
+ "mean_token_accuracy": 0.6160167968273162,
684
+ "num_tokens": 676489125.0,
685
+ "step": 15000
686
+ },
687
+ {
688
+ "epoch": 2.0966963238844056,
689
+ "grad_norm": 0.2782443165779114,
690
+ "learning_rate": 1.1615947027176161e-05,
691
+ "loss": 1.709,
692
+ "mean_token_accuracy": 0.6133450974524021,
693
+ "num_tokens": 689511656.0,
694
+ "step": 15200
695
+ },
696
+ {
697
+ "epoch": 2.124284433409201,
698
+ "grad_norm": 0.2791282832622528,
699
+ "learning_rate": 1.150558697751414e-05,
700
+ "loss": 1.7012,
701
+ "mean_token_accuracy": 0.6146386332809926,
702
+ "num_tokens": 702532680.0,
703
+ "step": 15400
704
+ },
705
+ {
706
+ "epoch": 2.1518725429339955,
707
+ "grad_norm": 0.2790575921535492,
708
+ "learning_rate": 1.1395226927852118e-05,
709
+ "loss": 1.7058,
710
+ "mean_token_accuracy": 0.6137663789093495,
711
+ "num_tokens": 715555115.0,
712
+ "step": 15600
713
+ },
714
+ {
715
+ "epoch": 2.1794606524587903,
716
+ "grad_norm": 0.2901487648487091,
717
+ "learning_rate": 1.1284866878190097e-05,
718
+ "loss": 1.6948,
719
+ "mean_token_accuracy": 0.6157194478809833,
720
+ "num_tokens": 728582765.0,
721
+ "step": 15800
722
+ },
723
+ {
724
+ "epoch": 2.207048761983585,
725
+ "grad_norm": 0.2835938334465027,
726
+ "learning_rate": 1.1174506828528073e-05,
727
+ "loss": 1.698,
728
+ "mean_token_accuracy": 0.615332849919796,
729
+ "num_tokens": 741592804.0,
730
+ "step": 16000
731
+ },
732
+ {
733
+ "epoch": 2.2346368715083798,
734
+ "grad_norm": 0.28976014256477356,
735
+ "learning_rate": 1.1064146778866052e-05,
736
+ "loss": 1.7036,
737
+ "mean_token_accuracy": 0.6142940573394299,
738
+ "num_tokens": 754612801.0,
739
+ "step": 16200
740
+ },
741
+ {
742
+ "epoch": 2.2622249810331745,
743
+ "grad_norm": 0.28366798162460327,
744
+ "learning_rate": 1.095378672920403e-05,
745
+ "loss": 1.6955,
746
+ "mean_token_accuracy": 0.6157566201686859,
747
+ "num_tokens": 767616919.0,
748
+ "step": 16400
749
+ },
750
+ {
751
+ "epoch": 2.2898130905579697,
752
+ "grad_norm": 0.28289395570755005,
753
+ "learning_rate": 1.0843426679542006e-05,
754
+ "loss": 1.6938,
755
+ "mean_token_accuracy": 0.6161167666316032,
756
+ "num_tokens": 780634732.0,
757
+ "step": 16600
758
+ },
759
+ {
760
+ "epoch": 2.3174012000827644,
761
+ "grad_norm": 0.2955986559391022,
762
+ "learning_rate": 1.0733066629879984e-05,
763
+ "loss": 1.6947,
764
+ "mean_token_accuracy": 0.6157604093849659,
765
+ "num_tokens": 793645866.0,
766
+ "step": 16800
767
+ },
768
+ {
769
+ "epoch": 2.344989309607559,
770
+ "grad_norm": 0.2817966341972351,
771
+ "learning_rate": 1.0622706580217962e-05,
772
+ "loss": 1.7028,
773
+ "mean_token_accuracy": 0.6145315931737423,
774
+ "num_tokens": 806653207.0,
775
+ "step": 17000
776
+ },
777
+ {
778
+ "epoch": 2.372577419132354,
779
+ "grad_norm": 0.28311964869499207,
780
+ "learning_rate": 1.051234653055594e-05,
781
+ "loss": 1.696,
782
+ "mean_token_accuracy": 0.615741535872221,
783
+ "num_tokens": 819673655.0,
784
+ "step": 17200
785
+ },
786
+ {
787
+ "epoch": 2.4001655286571486,
788
+ "grad_norm": 0.28408530354499817,
789
+ "learning_rate": 1.0401986480893918e-05,
790
+ "loss": 1.7001,
791
+ "mean_token_accuracy": 0.6146831949055195,
792
+ "num_tokens": 832689455.0,
793
+ "step": 17400
794
+ },
795
+ {
796
+ "epoch": 2.427753638181944,
797
+ "grad_norm": 0.29735735058784485,
798
+ "learning_rate": 1.0291626431231896e-05,
799
+ "loss": 1.6871,
800
+ "mean_token_accuracy": 0.6170836135745048,
801
+ "num_tokens": 845698323.0,
802
+ "step": 17600
803
+ },
804
+ {
805
+ "epoch": 2.4553417477067385,
806
+ "grad_norm": 0.3009927272796631,
807
+ "learning_rate": 1.0181266381569871e-05,
808
+ "loss": 1.6949,
809
+ "mean_token_accuracy": 0.6159922151267528,
810
+ "num_tokens": 858702827.0,
811
+ "step": 17800
812
+ },
813
+ {
814
+ "epoch": 2.4829298572315333,
815
+ "grad_norm": 0.29742059111595154,
816
+ "learning_rate": 1.007090633190785e-05,
817
+ "loss": 1.6968,
818
+ "mean_token_accuracy": 0.6157895983755588,
819
+ "num_tokens": 871701677.0,
820
+ "step": 18000
821
+ },
822
+ {
823
+ "epoch": 2.510517966756328,
824
+ "grad_norm": 0.2822323143482208,
825
+ "learning_rate": 9.960546282245828e-06,
826
+ "loss": 1.7002,
827
+ "mean_token_accuracy": 0.6144832235574722,
828
+ "num_tokens": 884719034.0,
829
+ "step": 18200
830
+ },
831
+ {
832
+ "epoch": 2.5381060762811227,
833
+ "grad_norm": 0.29363512992858887,
834
+ "learning_rate": 9.850186232583807e-06,
835
+ "loss": 1.696,
836
+ "mean_token_accuracy": 0.6158060549199581,
837
+ "num_tokens": 897716370.0,
838
+ "step": 18400
839
+ },
840
+ {
841
+ "epoch": 2.565694185805918,
842
+ "grad_norm": 0.29309144616127014,
843
+ "learning_rate": 9.739826182921783e-06,
844
+ "loss": 1.694,
845
+ "mean_token_accuracy": 0.6163165862858295,
846
+ "num_tokens": 910725195.0,
847
+ "step": 18600
848
+ },
849
+ {
850
+ "epoch": 2.593282295330712,
851
+ "grad_norm": 0.2888060510158539,
852
+ "learning_rate": 9.62946613325976e-06,
853
+ "loss": 1.6957,
854
+ "mean_token_accuracy": 0.6159252269566059,
855
+ "num_tokens": 923732160.0,
856
+ "step": 18800
857
+ },
858
+ {
859
+ "epoch": 2.6208704048555074,
860
+ "grad_norm": 0.28722092509269714,
861
+ "learning_rate": 9.519106083597739e-06,
862
+ "loss": 1.7,
863
+ "mean_token_accuracy": 0.6145291210711002,
864
+ "num_tokens": 936739185.0,
865
+ "step": 19000
866
+ },
867
+ {
868
+ "epoch": 2.648458514380302,
869
+ "grad_norm": 0.29655253887176514,
870
+ "learning_rate": 9.408746033935715e-06,
871
+ "loss": 1.6954,
872
+ "mean_token_accuracy": 0.6159876026213169,
873
+ "num_tokens": 949725573.0,
874
+ "step": 19200
875
+ },
876
+ {
877
+ "epoch": 2.676046623905097,
878
+ "grad_norm": 0.29758745431900024,
879
+ "learning_rate": 9.298385984273694e-06,
880
+ "loss": 1.6958,
881
+ "mean_token_accuracy": 0.616194871366024,
882
+ "num_tokens": 962696837.0,
883
+ "step": 19400
884
+ },
885
+ {
886
+ "epoch": 2.7036347334298916,
887
+ "grad_norm": 0.29457658529281616,
888
+ "learning_rate": 9.188025934611672e-06,
889
+ "loss": 1.692,
890
+ "mean_token_accuracy": 0.6167295287549496,
891
+ "num_tokens": 975699859.0,
892
+ "step": 19600
893
+ },
894
+ {
895
+ "epoch": 2.7312228429546863,
896
+ "grad_norm": 0.29310283064842224,
897
+ "learning_rate": 9.077665884949649e-06,
898
+ "loss": 1.6948,
899
+ "mean_token_accuracy": 0.6162155202031135,
900
+ "num_tokens": 988705637.0,
901
+ "step": 19800
902
+ },
903
+ {
904
+ "epoch": 2.7588109524794815,
905
+ "grad_norm": 0.2922320067882538,
906
+ "learning_rate": 8.967305835287626e-06,
907
+ "loss": 1.6901,
908
+ "mean_token_accuracy": 0.6163058173656464,
909
+ "num_tokens": 1001754945.0,
910
+ "step": 20000
911
+ },
912
+ {
913
+ "epoch": 2.7863990620042762,
914
+ "grad_norm": 0.2901257276535034,
915
+ "learning_rate": 8.856945785625604e-06,
916
+ "loss": 1.6951,
917
+ "mean_token_accuracy": 0.6158974586427212,
918
+ "num_tokens": 1014746336.0,
919
+ "step": 20200
920
+ },
921
+ {
922
+ "epoch": 2.813987171529071,
923
+ "grad_norm": 0.28679758310317993,
924
+ "learning_rate": 8.746585735963581e-06,
925
+ "loss": 1.6858,
926
+ "mean_token_accuracy": 0.6173767544329166,
927
+ "num_tokens": 1027763913.0,
928
+ "step": 20400
929
+ },
930
+ {
931
+ "epoch": 2.8415752810538657,
932
+ "grad_norm": 0.2966858744621277,
933
+ "learning_rate": 8.63622568630156e-06,
934
+ "loss": 1.686,
935
+ "mean_token_accuracy": 0.6175419771671296,
936
+ "num_tokens": 1040766281.0,
937
+ "step": 20600
938
+ },
939
+ {
940
+ "epoch": 2.8691633905786604,
941
+ "grad_norm": 0.300108939409256,
942
+ "learning_rate": 8.525865636639538e-06,
943
+ "loss": 1.694,
944
+ "mean_token_accuracy": 0.6159957525134087,
945
+ "num_tokens": 1053798269.0,
946
+ "step": 20800
947
+ },
948
+ {
949
+ "epoch": 2.8967515001034556,
950
+ "grad_norm": 0.2970072031021118,
951
+ "learning_rate": 8.415505586977515e-06,
952
+ "loss": 1.6881,
953
+ "mean_token_accuracy": 0.6167993746697903,
954
+ "num_tokens": 1066815504.0,
955
+ "step": 21000
956
  }
957
  ],
958
  "logging_steps": 200,
 
972
  "attributes": {}
973
  }
974
  },
975
+ "total_flos": 2.469866946100514e+19,
976
  "train_batch_size": 4,
977
  "trial_name": null,
978
  "trial_params": null