{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.006, "eval_steps": 100, "global_step": 300, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 2e-05, "grad_norm": 1.8784615993499756, "learning_rate": 4e-10, "loss": 1.247075080871582, "memory(GiB)": 45.58, "step": 1, "token_acc": 0.7174163783160323, "train_speed(iter/s)": 0.013609 }, { "epoch": 0.0001, "grad_norm": 1.7829604148864746, "learning_rate": 2e-09, "loss": 1.2329106330871582, "memory(GiB)": 47.68, "step": 5, "token_acc": 0.7190702324418895, "train_speed(iter/s)": 0.056616 }, { "epoch": 0.0002, "grad_norm": 1.8510538339614868, "learning_rate": 4e-09, "loss": 1.2224847793579101, "memory(GiB)": 47.68, "step": 10, "token_acc": 0.7056490003173596, "train_speed(iter/s)": 0.093003 }, { "epoch": 0.0003, "grad_norm": 1.9363449811935425, "learning_rate": 5.6e-09, "loss": 1.2302563667297364, "memory(GiB)": 48.7, "step": 15, "token_acc": 0.7107342772472836, "train_speed(iter/s)": 0.119381 }, { "epoch": 0.0004, "grad_norm": 1.935133695602417, "learning_rate": 7.6e-09, "loss": 1.2335556983947753, "memory(GiB)": 48.7, "step": 20, "token_acc": 0.7361481714016638, "train_speed(iter/s)": 0.137465 }, { "epoch": 0.0005, "grad_norm": 1.7830060720443726, "learning_rate": 9.599999999999998e-09, "loss": 1.2144426345825194, "memory(GiB)": 48.7, "step": 25, "token_acc": 0.7571428571428571, "train_speed(iter/s)": 0.150572 }, { "epoch": 0.0006, "grad_norm": 1.7309236526489258, "learning_rate": 1.1599999999999998e-08, "loss": 1.2357244491577148, "memory(GiB)": 48.7, "step": 30, "token_acc": 0.7153189910979229, "train_speed(iter/s)": 0.162512 }, { "epoch": 0.0007, "grad_norm": 1.7474730014801025, "learning_rate": 1.36e-08, "loss": 1.231928825378418, "memory(GiB)": 48.7, "step": 35, "token_acc": 0.7078455355582737, "train_speed(iter/s)": 0.171663 }, { "epoch": 0.0008, "grad_norm": 1.780007004737854, "learning_rate": 1.5599999999999997e-08, "loss": 1.2193016052246093, "memory(GiB)": 48.7, "step": 40, "token_acc": 0.6932599724896836, "train_speed(iter/s)": 0.178935 }, { "epoch": 0.0009, "grad_norm": 1.723268747329712, "learning_rate": 1.72e-08, "loss": 1.204256820678711, "memory(GiB)": 48.7, "step": 45, "token_acc": 0.6967172137977004, "train_speed(iter/s)": 0.186547 }, { "epoch": 0.001, "grad_norm": 1.7750556468963623, "learning_rate": 1.9199999999999997e-08, "loss": 1.2007877349853515, "memory(GiB)": 48.7, "step": 50, "token_acc": 0.6921988682295878, "train_speed(iter/s)": 0.19125 }, { "epoch": 0.0011, "grad_norm": 1.299729585647583, "learning_rate": 2.1199999999999998e-08, "loss": 0.8069381713867188, "memory(GiB)": 48.7, "step": 55, "token_acc": 0.7952855847688123, "train_speed(iter/s)": 0.195908 }, { "epoch": 0.0012, "grad_norm": 1.3525291681289673, "learning_rate": 2.3199999999999996e-08, "loss": 0.42714319229125974, "memory(GiB)": 48.7, "step": 60, "token_acc": 0.8841448189762796, "train_speed(iter/s)": 0.201 }, { "epoch": 0.0013, "grad_norm": 1.080306053161621, "learning_rate": 2.52e-08, "loss": 0.4192944526672363, "memory(GiB)": 48.7, "step": 65, "token_acc": 0.8929637526652452, "train_speed(iter/s)": 0.204749 }, { "epoch": 0.0014, "grad_norm": 1.177667498588562, "learning_rate": 2.72e-08, "loss": 0.4133622169494629, "memory(GiB)": 48.7, "step": 70, "token_acc": 0.8836150845253576, "train_speed(iter/s)": 0.20853 }, { "epoch": 0.0015, "grad_norm": 1.4200434684753418, "learning_rate": 2.92e-08, "loss": 0.4155101776123047, "memory(GiB)": 48.7, "step": 75, "token_acc": 0.8933333333333333, "train_speed(iter/s)": 0.211872 }, { "epoch": 0.0016, "grad_norm": 1.1931238174438477, "learning_rate": 3.1199999999999995e-08, "loss": 0.40213637351989745, "memory(GiB)": 48.7, "step": 80, "token_acc": 0.9073569482288828, "train_speed(iter/s)": 0.214198 }, { "epoch": 0.0017, "grad_norm": 1.2268942594528198, "learning_rate": 3.32e-08, "loss": 0.4193448543548584, "memory(GiB)": 48.7, "step": 85, "token_acc": 0.890360895386021, "train_speed(iter/s)": 0.217149 }, { "epoch": 0.0018, "grad_norm": 1.168769121170044, "learning_rate": 3.52e-08, "loss": 0.4069235324859619, "memory(GiB)": 48.7, "step": 90, "token_acc": 0.8803038239916187, "train_speed(iter/s)": 0.219373 }, { "epoch": 0.0019, "grad_norm": 1.1890403032302856, "learning_rate": 3.7199999999999996e-08, "loss": 0.4017478942871094, "memory(GiB)": 48.7, "step": 95, "token_acc": 0.8852933438568797, "train_speed(iter/s)": 0.221331 }, { "epoch": 0.002, "grad_norm": 1.0330846309661865, "learning_rate": 3.9199999999999994e-08, "loss": 0.3853747844696045, "memory(GiB)": 48.7, "step": 100, "token_acc": 0.8942406230784997, "train_speed(iter/s)": 0.223547 }, { "epoch": 0.0021, "grad_norm": 0.8882772922515869, "learning_rate": 4.12e-08, "loss": 0.38479089736938477, "memory(GiB)": 48.7, "step": 105, "token_acc": 0.888404744613895, "train_speed(iter/s)": 0.165159 }, { "epoch": 0.0022, "grad_norm": 0.9768219590187073, "learning_rate": 4.32e-08, "loss": 0.3818492889404297, "memory(GiB)": 48.7, "step": 110, "token_acc": 0.8976019604410993, "train_speed(iter/s)": 0.168006 }, { "epoch": 0.0023, "grad_norm": 1.0652796030044556, "learning_rate": 4.5199999999999994e-08, "loss": 0.37880630493164064, "memory(GiB)": 48.7, "step": 115, "token_acc": 0.8944169293111212, "train_speed(iter/s)": 0.170961 }, { "epoch": 0.0024, "grad_norm": 0.9667234420776367, "learning_rate": 4.72e-08, "loss": 0.3773456573486328, "memory(GiB)": 48.7, "step": 120, "token_acc": 0.9041720990873533, "train_speed(iter/s)": 0.17359 }, { "epoch": 0.0025, "grad_norm": 1.1004765033721924, "learning_rate": 4.92e-08, "loss": 0.37454307079315186, "memory(GiB)": 48.7, "step": 125, "token_acc": 0.8971794358871774, "train_speed(iter/s)": 0.17613 }, { "epoch": 0.0026, "grad_norm": 0.9110294580459595, "learning_rate": 5.12e-08, "loss": 0.35935871601104735, "memory(GiB)": 48.7, "step": 130, "token_acc": 0.9002339927674963, "train_speed(iter/s)": 0.178791 }, { "epoch": 0.0027, "grad_norm": 0.8307713270187378, "learning_rate": 5.319999999999999e-08, "loss": 0.3583024501800537, "memory(GiB)": 48.7, "step": 135, "token_acc": 0.9001166861143524, "train_speed(iter/s)": 0.181064 }, { "epoch": 0.0028, "grad_norm": 0.8742683529853821, "learning_rate": 5.52e-08, "loss": 0.34030709266662595, "memory(GiB)": 48.7, "step": 140, "token_acc": 0.8936170212765957, "train_speed(iter/s)": 0.183337 }, { "epoch": 0.0029, "grad_norm": 0.8469884395599365, "learning_rate": 5.7199999999999996e-08, "loss": 0.3286393404006958, "memory(GiB)": 48.7, "step": 145, "token_acc": 0.9112193294624461, "train_speed(iter/s)": 0.185295 }, { "epoch": 0.003, "grad_norm": 0.7903546094894409, "learning_rate": 5.92e-08, "loss": 0.34013702869415285, "memory(GiB)": 48.7, "step": 150, "token_acc": 0.9123177283192633, "train_speed(iter/s)": 0.187277 }, { "epoch": 0.0031, "grad_norm": 0.7647843956947327, "learning_rate": 6.119999999999999e-08, "loss": 0.32967448234558105, "memory(GiB)": 48.7, "step": 155, "token_acc": 0.9061872909698997, "train_speed(iter/s)": 0.189242 }, { "epoch": 0.0032, "grad_norm": 0.7588908076286316, "learning_rate": 6.32e-08, "loss": 0.31885499954223634, "memory(GiB)": 48.7, "step": 160, "token_acc": 0.9148387096774193, "train_speed(iter/s)": 0.190988 }, { "epoch": 0.0033, "grad_norm": 0.7278595566749573, "learning_rate": 6.519999999999999e-08, "loss": 0.3225527048110962, "memory(GiB)": 48.7, "step": 165, "token_acc": 0.9134559535333979, "train_speed(iter/s)": 0.192579 }, { "epoch": 0.0034, "grad_norm": 0.7548233270645142, "learning_rate": 6.719999999999999e-08, "loss": 0.3111454725265503, "memory(GiB)": 48.7, "step": 170, "token_acc": 0.9035423647678315, "train_speed(iter/s)": 0.194372 }, { "epoch": 0.0035, "grad_norm": 0.7382190823554993, "learning_rate": 6.92e-08, "loss": 0.30791687965393066, "memory(GiB)": 48.7, "step": 175, "token_acc": 0.9182072829131652, "train_speed(iter/s)": 0.195886 }, { "epoch": 0.0036, "grad_norm": 0.6201812028884888, "learning_rate": 7.12e-08, "loss": 0.30755660533905027, "memory(GiB)": 48.7, "step": 180, "token_acc": 0.9121319199057715, "train_speed(iter/s)": 0.197365 }, { "epoch": 0.0037, "grad_norm": 0.680054247379303, "learning_rate": 7.32e-08, "loss": 0.3023200511932373, "memory(GiB)": 48.7, "step": 185, "token_acc": 0.9189243427654024, "train_speed(iter/s)": 0.198991 }, { "epoch": 0.0038, "grad_norm": 0.7230331301689148, "learning_rate": 7.52e-08, "loss": 0.29792306423187254, "memory(GiB)": 48.7, "step": 190, "token_acc": 0.9166977960403436, "train_speed(iter/s)": 0.200242 }, { "epoch": 0.0039, "grad_norm": 0.6341392397880554, "learning_rate": 7.72e-08, "loss": 0.2936398983001709, "memory(GiB)": 48.7, "step": 195, "token_acc": 0.916235294117647, "train_speed(iter/s)": 0.20159 }, { "epoch": 0.004, "grad_norm": 0.6513913869857788, "learning_rate": 7.920000000000001e-08, "loss": 0.29084038734436035, "memory(GiB)": 48.7, "step": 200, "token_acc": 0.915651358950328, "train_speed(iter/s)": 0.202853 }, { "epoch": 0.0041, "grad_norm": 0.5849307775497437, "learning_rate": 8.119999999999999e-08, "loss": 0.28313345909118653, "memory(GiB)": 48.7, "step": 205, "token_acc": 0.9201498422712934, "train_speed(iter/s)": 0.174461 }, { "epoch": 0.0042, "grad_norm": 0.5864230990409851, "learning_rate": 8.319999999999999e-08, "loss": 0.27917160987854006, "memory(GiB)": 48.7, "step": 210, "token_acc": 0.9227850926132559, "train_speed(iter/s)": 0.176086 }, { "epoch": 0.0043, "grad_norm": 0.5705774426460266, "learning_rate": 8.52e-08, "loss": 0.27348690032958983, "memory(GiB)": 48.7, "step": 215, "token_acc": 0.9209313047487321, "train_speed(iter/s)": 0.177517 }, { "epoch": 0.0044, "grad_norm": 0.5566577911376953, "learning_rate": 8.72e-08, "loss": 0.2738009214401245, "memory(GiB)": 48.7, "step": 220, "token_acc": 0.9212003454231433, "train_speed(iter/s)": 0.178985 }, { "epoch": 0.0045, "grad_norm": 0.5548610091209412, "learning_rate": 8.919999999999999e-08, "loss": 0.270387601852417, "memory(GiB)": 48.7, "step": 225, "token_acc": 0.9163515016685205, "train_speed(iter/s)": 0.180404 }, { "epoch": 0.0046, "grad_norm": 0.5132172107696533, "learning_rate": 9.12e-08, "loss": 0.2689182758331299, "memory(GiB)": 48.7, "step": 230, "token_acc": 0.9161837748344371, "train_speed(iter/s)": 0.181679 }, { "epoch": 0.0047, "grad_norm": 0.5392292141914368, "learning_rate": 9.32e-08, "loss": 0.2692979335784912, "memory(GiB)": 48.7, "step": 235, "token_acc": 0.9189686924493554, "train_speed(iter/s)": 0.18288 }, { "epoch": 0.0048, "grad_norm": 0.5800856947898865, "learning_rate": 9.52e-08, "loss": 0.2671097755432129, "memory(GiB)": 48.7, "step": 240, "token_acc": 0.9276693455797933, "train_speed(iter/s)": 0.184197 }, { "epoch": 0.0049, "grad_norm": 0.5248638391494751, "learning_rate": 9.72e-08, "loss": 0.2651660442352295, "memory(GiB)": 48.7, "step": 245, "token_acc": 0.9340950826361967, "train_speed(iter/s)": 0.185361 }, { "epoch": 0.005, "grad_norm": 0.4779960811138153, "learning_rate": 9.919999999999999e-08, "loss": 0.2598712921142578, "memory(GiB)": 48.7, "step": 250, "token_acc": 0.9171786120591582, "train_speed(iter/s)": 0.186527 }, { "epoch": 0.0051, "grad_norm": 0.5446517467498779, "learning_rate": 1.0119999999999999e-07, "loss": 0.2589933633804321, "memory(GiB)": 48.7, "step": 255, "token_acc": 0.919714165615805, "train_speed(iter/s)": 0.187704 }, { "epoch": 0.0052, "grad_norm": 0.5014457702636719, "learning_rate": 1.032e-07, "loss": 0.2622577667236328, "memory(GiB)": 48.7, "step": 260, "token_acc": 0.9149246475449684, "train_speed(iter/s)": 0.188846 }, { "epoch": 0.0053, "grad_norm": 0.486289918422699, "learning_rate": 1.052e-07, "loss": 0.2525398969650269, "memory(GiB)": 48.7, "step": 265, "token_acc": 0.9272396543883583, "train_speed(iter/s)": 0.189871 }, { "epoch": 0.0054, "grad_norm": 0.5184631943702698, "learning_rate": 1.072e-07, "loss": 0.2508587121963501, "memory(GiB)": 48.7, "step": 270, "token_acc": 0.9343525179856115, "train_speed(iter/s)": 0.190895 }, { "epoch": 0.0055, "grad_norm": 0.7472315430641174, "learning_rate": 1.092e-07, "loss": 0.3759446620941162, "memory(GiB)": 48.7, "step": 275, "token_acc": 0.8695306284805091, "train_speed(iter/s)": 0.191441 }, { "epoch": 0.0056, "grad_norm": 0.6762179136276245, "learning_rate": 1.1119999999999999e-07, "loss": 0.5864349365234375, "memory(GiB)": 48.7, "step": 280, "token_acc": 0.8324331623008372, "train_speed(iter/s)": 0.191626 }, { "epoch": 0.0057, "grad_norm": 0.5527864694595337, "learning_rate": 1.132e-07, "loss": 0.5735197067260742, "memory(GiB)": 48.7, "step": 285, "token_acc": 0.8262280306444344, "train_speed(iter/s)": 0.19159 }, { "epoch": 0.0058, "grad_norm": 0.46731290221214294, "learning_rate": 1.1519999999999999e-07, "loss": 0.5600314140319824, "memory(GiB)": 48.7, "step": 290, "token_acc": 0.8467968392868804, "train_speed(iter/s)": 0.191504 }, { "epoch": 0.0059, "grad_norm": 0.42628228664398193, "learning_rate": 1.1719999999999999e-07, "loss": 0.5390614986419677, "memory(GiB)": 50.81, "step": 295, "token_acc": 0.8327679356703104, "train_speed(iter/s)": 0.191542 }, { "epoch": 0.006, "grad_norm": 0.3921428322792053, "learning_rate": 1.192e-07, "loss": 0.5223519325256347, "memory(GiB)": 50.81, "step": 300, "token_acc": 0.8428060564559081, "train_speed(iter/s)": 0.191529 } ], "logging_steps": 5, "max_steps": 50000, "num_input_tokens_seen": 0, "num_train_epochs": 9223372036854775807, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 5.573770266594509e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }