phi3m0128-cds-0.5-kendall-onof-ofif-corr-max-2-simpo-max1500-default
/
checkpoint-1450
/trainer_state.json
{ | |
"best_metric": null, | |
"best_model_checkpoint": null, | |
"epoch": 1.5458422174840085, | |
"eval_steps": 50, | |
"global_step": 1450, | |
"is_hyper_param_search": false, | |
"is_local_process_zero": true, | |
"is_world_process_zero": true, | |
"log_history": [ | |
{ | |
"epoch": 0.010660980810234541, | |
"grad_norm": 0.051327500492334366, | |
"learning_rate": 4.999451708687114e-06, | |
"logits/chosen": 14.755006790161133, | |
"logits/rejected": 14.735244750976562, | |
"logps/chosen": -0.29377540946006775, | |
"logps/rejected": -0.30969956517219543, | |
"loss": 0.952, | |
"rewards/accuracies": 0.4375, | |
"rewards/chosen": -0.44066309928894043, | |
"rewards/margins": 0.023886267095804214, | |
"rewards/rejected": -0.46454939246177673, | |
"step": 10 | |
}, | |
{ | |
"epoch": 0.021321961620469083, | |
"grad_norm": 0.04346882924437523, | |
"learning_rate": 4.997807075247147e-06, | |
"logits/chosen": 14.513801574707031, | |
"logits/rejected": 14.946454048156738, | |
"logps/chosen": -0.27995699644088745, | |
"logps/rejected": -0.30138006806373596, | |
"loss": 0.9726, | |
"rewards/accuracies": 0.4124999940395355, | |
"rewards/chosen": -0.4199354648590088, | |
"rewards/margins": 0.03213457390666008, | |
"rewards/rejected": -0.45207005739212036, | |
"step": 20 | |
}, | |
{ | |
"epoch": 0.031982942430703626, | |
"grad_norm": 0.05228634551167488, | |
"learning_rate": 4.9950668210706795e-06, | |
"logits/chosen": 14.266324043273926, | |
"logits/rejected": 14.423965454101562, | |
"logps/chosen": -0.2919609546661377, | |
"logps/rejected": -0.32358455657958984, | |
"loss": 0.9622, | |
"rewards/accuracies": 0.5, | |
"rewards/chosen": -0.43794146180152893, | |
"rewards/margins": 0.047435395419597626, | |
"rewards/rejected": -0.48537683486938477, | |
"step": 30 | |
}, | |
{ | |
"epoch": 0.042643923240938165, | |
"grad_norm": 0.05487598106265068, | |
"learning_rate": 4.9912321481237616e-06, | |
"logits/chosen": 14.965211868286133, | |
"logits/rejected": 15.058088302612305, | |
"logps/chosen": -0.277716726064682, | |
"logps/rejected": -0.3055034577846527, | |
"loss": 0.9403, | |
"rewards/accuracies": 0.4000000059604645, | |
"rewards/chosen": -0.4165751039981842, | |
"rewards/margins": 0.04168009012937546, | |
"rewards/rejected": -0.4582551419734955, | |
"step": 40 | |
}, | |
{ | |
"epoch": 0.053304904051172705, | |
"grad_norm": 0.057255037128925323, | |
"learning_rate": 4.986304738420684e-06, | |
"logits/chosen": 14.539288520812988, | |
"logits/rejected": 15.174041748046875, | |
"logps/chosen": -0.26362231373786926, | |
"logps/rejected": -0.3325727581977844, | |
"loss": 0.9588, | |
"rewards/accuracies": 0.4749999940395355, | |
"rewards/chosen": -0.3954334557056427, | |
"rewards/margins": 0.10342560708522797, | |
"rewards/rejected": -0.49885907769203186, | |
"step": 50 | |
}, | |
{ | |
"epoch": 0.053304904051172705, | |
"eval_logits/chosen": 14.618952751159668, | |
"eval_logits/rejected": 15.176809310913086, | |
"eval_logps/chosen": -0.2685677409172058, | |
"eval_logps/rejected": -0.3283654451370239, | |
"eval_loss": 0.9551004767417908, | |
"eval_rewards/accuracies": 0.5131579041481018, | |
"eval_rewards/chosen": -0.4028516113758087, | |
"eval_rewards/margins": 0.08969658613204956, | |
"eval_rewards/rejected": -0.4925481975078583, | |
"eval_runtime": 21.4453, | |
"eval_samples_per_second": 28.305, | |
"eval_steps_per_second": 3.544, | |
"step": 50 | |
}, | |
{ | |
"epoch": 0.06396588486140725, | |
"grad_norm": 0.05227242782711983, | |
"learning_rate": 4.980286753286196e-06, | |
"logits/chosen": 14.787714004516602, | |
"logits/rejected": 15.379422187805176, | |
"logps/chosen": -0.3143109679222107, | |
"logps/rejected": -0.3425135612487793, | |
"loss": 0.9636, | |
"rewards/accuracies": 0.4625000059604645, | |
"rewards/chosen": -0.4714665412902832, | |
"rewards/margins": 0.042303841561079025, | |
"rewards/rejected": -0.513770341873169, | |
"step": 60 | |
}, | |
{ | |
"epoch": 0.07462686567164178, | |
"grad_norm": 0.0658508762717247, | |
"learning_rate": 4.973180832407471e-06, | |
"logits/chosen": 15.149365425109863, | |
"logits/rejected": 15.115835189819336, | |
"logps/chosen": -0.31501108407974243, | |
"logps/rejected": -0.2854115962982178, | |
"loss": 0.9677, | |
"rewards/accuracies": 0.36250001192092896, | |
"rewards/chosen": -0.47251659631729126, | |
"rewards/margins": -0.04439922422170639, | |
"rewards/rejected": -0.4281173646450043, | |
"step": 70 | |
}, | |
{ | |
"epoch": 0.08528784648187633, | |
"grad_norm": 0.06567618995904922, | |
"learning_rate": 4.964990092676263e-06, | |
"logits/chosen": 15.393908500671387, | |
"logits/rejected": 15.454248428344727, | |
"logps/chosen": -0.31166282296180725, | |
"logps/rejected": -0.3178747594356537, | |
"loss": 0.9609, | |
"rewards/accuracies": 0.42500001192092896, | |
"rewards/chosen": -0.4674941897392273, | |
"rewards/margins": 0.009317949414253235, | |
"rewards/rejected": -0.47681212425231934, | |
"step": 80 | |
}, | |
{ | |
"epoch": 0.09594882729211088, | |
"grad_norm": 0.07566913962364197, | |
"learning_rate": 4.9557181268217225e-06, | |
"logits/chosen": 15.229632377624512, | |
"logits/rejected": 15.477168083190918, | |
"logps/chosen": -0.3294064998626709, | |
"logps/rejected": -0.3528878390789032, | |
"loss": 0.9587, | |
"rewards/accuracies": 0.4749999940395355, | |
"rewards/chosen": -0.49410971999168396, | |
"rewards/margins": 0.03522203490138054, | |
"rewards/rejected": -0.5293318033218384, | |
"step": 90 | |
}, | |
{ | |
"epoch": 0.10660980810234541, | |
"grad_norm": 0.09082464128732681, | |
"learning_rate": 4.9453690018345144e-06, | |
"logits/chosen": 14.481330871582031, | |
"logits/rejected": 15.092982292175293, | |
"logps/chosen": -0.2656436562538147, | |
"logps/rejected": -0.33982905745506287, | |
"loss": 0.9548, | |
"rewards/accuracies": 0.5625, | |
"rewards/chosen": -0.39846545457839966, | |
"rewards/margins": 0.11127817630767822, | |
"rewards/rejected": -0.5097435712814331, | |
"step": 100 | |
}, | |
{ | |
"epoch": 0.10660980810234541, | |
"eval_logits/chosen": 14.7100830078125, | |
"eval_logits/rejected": 15.274725914001465, | |
"eval_logps/chosen": -0.26462864875793457, | |
"eval_logps/rejected": -0.331702321767807, | |
"eval_loss": 0.947841465473175, | |
"eval_rewards/accuracies": 0.5394737124443054, | |
"eval_rewards/chosen": -0.39694297313690186, | |
"eval_rewards/margins": 0.10061051696538925, | |
"eval_rewards/rejected": -0.4975534677505493, | |
"eval_runtime": 21.4421, | |
"eval_samples_per_second": 28.309, | |
"eval_steps_per_second": 3.544, | |
"step": 100 | |
}, | |
{ | |
"epoch": 0.11727078891257996, | |
"grad_norm": 0.20198923349380493, | |
"learning_rate": 4.933947257182901e-06, | |
"logits/chosen": 14.932653427124023, | |
"logits/rejected": 15.476409912109375, | |
"logps/chosen": -0.27830976247787476, | |
"logps/rejected": -0.34150317311286926, | |
"loss": 0.9487, | |
"rewards/accuracies": 0.550000011920929, | |
"rewards/chosen": -0.41746464371681213, | |
"rewards/margins": 0.09479012340307236, | |
"rewards/rejected": -0.5122548341751099, | |
"step": 110 | |
}, | |
{ | |
"epoch": 0.1279317697228145, | |
"grad_norm": 0.31938356161117554, | |
"learning_rate": 4.921457902821578e-06, | |
"logits/chosen": 15.280967712402344, | |
"logits/rejected": 15.5416259765625, | |
"logps/chosen": -0.2816022038459778, | |
"logps/rejected": -0.3262938857078552, | |
"loss": 0.9483, | |
"rewards/accuracies": 0.4749999940395355, | |
"rewards/chosen": -0.4224032461643219, | |
"rewards/margins": 0.06703753769397736, | |
"rewards/rejected": -0.48944082856178284, | |
"step": 120 | |
}, | |
{ | |
"epoch": 0.13859275053304904, | |
"grad_norm": 0.12567812204360962, | |
"learning_rate": 4.907906416994146e-06, | |
"logits/chosen": 14.967382431030273, | |
"logits/rejected": 15.351877212524414, | |
"logps/chosen": -0.3148510456085205, | |
"logps/rejected": -0.3488944172859192, | |
"loss": 0.957, | |
"rewards/accuracies": 0.48750001192092896, | |
"rewards/chosen": -0.47227659821510315, | |
"rewards/margins": 0.05106503888964653, | |
"rewards/rejected": -0.5233416557312012, | |
"step": 130 | |
}, | |
{ | |
"epoch": 0.14925373134328357, | |
"grad_norm": 0.09151162207126617, | |
"learning_rate": 4.893298743830168e-06, | |
"logits/chosen": 14.900466918945312, | |
"logits/rejected": 15.075350761413574, | |
"logps/chosen": -0.2766302227973938, | |
"logps/rejected": -0.312236487865448, | |
"loss": 0.9373, | |
"rewards/accuracies": 0.42500001192092896, | |
"rewards/chosen": -0.4149452745914459, | |
"rewards/margins": 0.05340944975614548, | |
"rewards/rejected": -0.4683547616004944, | |
"step": 140 | |
}, | |
{ | |
"epoch": 0.15991471215351813, | |
"grad_norm": 0.1259378045797348, | |
"learning_rate": 4.8776412907378845e-06, | |
"logits/chosen": 14.528109550476074, | |
"logits/rejected": 14.861102104187012, | |
"logps/chosen": -0.2683579921722412, | |
"logps/rejected": -0.33838269114494324, | |
"loss": 0.9388, | |
"rewards/accuracies": 0.5375000238418579, | |
"rewards/chosen": -0.40253695845603943, | |
"rewards/margins": 0.10503707826137543, | |
"rewards/rejected": -0.5075740218162537, | |
"step": 150 | |
}, | |
{ | |
"epoch": 0.15991471215351813, | |
"eval_logits/chosen": 14.12246036529541, | |
"eval_logits/rejected": 14.733266830444336, | |
"eval_logps/chosen": -0.2611957788467407, | |
"eval_logps/rejected": -0.3492279350757599, | |
"eval_loss": 0.9302574396133423, | |
"eval_rewards/accuracies": 0.5657894611358643, | |
"eval_rewards/chosen": -0.3917936384677887, | |
"eval_rewards/margins": 0.13204820454120636, | |
"eval_rewards/rejected": -0.5238418579101562, | |
"eval_runtime": 21.4406, | |
"eval_samples_per_second": 28.311, | |
"eval_steps_per_second": 3.545, | |
"step": 150 | |
}, | |
{ | |
"epoch": 0.17057569296375266, | |
"grad_norm": 0.11400051414966583, | |
"learning_rate": 4.860940925593703e-06, | |
"logits/chosen": 14.4571533203125, | |
"logits/rejected": 14.769159317016602, | |
"logps/chosen": -0.31032469868659973, | |
"logps/rejected": -0.34650668501853943, | |
"loss": 0.9396, | |
"rewards/accuracies": 0.512499988079071, | |
"rewards/chosen": -0.4654870927333832, | |
"rewards/margins": 0.05427298694849014, | |
"rewards/rejected": -0.519760012626648, | |
"step": 160 | |
}, | |
{ | |
"epoch": 0.1812366737739872, | |
"grad_norm": 0.1102401539683342, | |
"learning_rate": 4.84320497372973e-06, | |
"logits/chosen": 13.959765434265137, | |
"logits/rejected": 14.27458381652832, | |
"logps/chosen": -0.2744378447532654, | |
"logps/rejected": -0.35702812671661377, | |
"loss": 0.9222, | |
"rewards/accuracies": 0.5375000238418579, | |
"rewards/chosen": -0.41165676712989807, | |
"rewards/margins": 0.12388546764850616, | |
"rewards/rejected": -0.5355421900749207, | |
"step": 170 | |
}, | |
{ | |
"epoch": 0.19189765458422176, | |
"grad_norm": 0.14721031486988068, | |
"learning_rate": 4.824441214720629e-06, | |
"logits/chosen": 13.54602336883545, | |
"logits/rejected": 14.076690673828125, | |
"logps/chosen": -0.2713850140571594, | |
"logps/rejected": -0.40618976950645447, | |
"loss": 0.9052, | |
"rewards/accuracies": 0.675000011920929, | |
"rewards/chosen": -0.40707746148109436, | |
"rewards/margins": 0.20220720767974854, | |
"rewards/rejected": -0.6092846989631653, | |
"step": 180 | |
}, | |
{ | |
"epoch": 0.2025586353944563, | |
"grad_norm": 0.1756824553012848, | |
"learning_rate": 4.804657878971252e-06, | |
"logits/chosen": 12.6314697265625, | |
"logits/rejected": 13.246849060058594, | |
"logps/chosen": -0.27216213941574097, | |
"logps/rejected": -0.4351380467414856, | |
"loss": 0.8996, | |
"rewards/accuracies": 0.6875, | |
"rewards/chosen": -0.40824323892593384, | |
"rewards/margins": 0.24446387588977814, | |
"rewards/rejected": -0.652707040309906, | |
"step": 190 | |
}, | |
{ | |
"epoch": 0.21321961620469082, | |
"grad_norm": 0.15476027131080627, | |
"learning_rate": 4.783863644106502e-06, | |
"logits/chosen": 12.40199089050293, | |
"logits/rejected": 12.966108322143555, | |
"logps/chosen": -0.303610622882843, | |
"logps/rejected": -0.423031747341156, | |
"loss": 0.9015, | |
"rewards/accuracies": 0.5375000238418579, | |
"rewards/chosen": -0.4554159641265869, | |
"rewards/margins": 0.17913168668746948, | |
"rewards/rejected": -0.6345476508140564, | |
"step": 200 | |
}, | |
{ | |
"epoch": 0.21321961620469082, | |
"eval_logits/chosen": 11.887229919433594, | |
"eval_logits/rejected": 12.5900239944458, | |
"eval_logps/chosen": -0.269090861082077, | |
"eval_logps/rejected": -0.42408913373947144, | |
"eval_loss": 0.8796805143356323, | |
"eval_rewards/accuracies": 0.6447368264198303, | |
"eval_rewards/chosen": -0.40363630652427673, | |
"eval_rewards/margins": 0.23249731957912445, | |
"eval_rewards/rejected": -0.6361336708068848, | |
"eval_runtime": 21.4455, | |
"eval_samples_per_second": 28.304, | |
"eval_steps_per_second": 3.544, | |
"step": 200 | |
}, | |
{ | |
"epoch": 0.22388059701492538, | |
"grad_norm": 0.18212148547172546, | |
"learning_rate": 4.762067631165049e-06, | |
"logits/chosen": 12.375594139099121, | |
"logits/rejected": 12.701678276062012, | |
"logps/chosen": -0.3136894702911377, | |
"logps/rejected": -0.3944609761238098, | |
"loss": 0.8898, | |
"rewards/accuracies": 0.4625000059604645, | |
"rewards/chosen": -0.47053417563438416, | |
"rewards/margins": 0.12115727365016937, | |
"rewards/rejected": -0.5916914939880371, | |
"step": 210 | |
}, | |
{ | |
"epoch": 0.2345415778251599, | |
"grad_norm": 0.5440058708190918, | |
"learning_rate": 4.7392794005985324e-06, | |
"logits/chosen": 11.23914909362793, | |
"logits/rejected": 11.926396369934082, | |
"logps/chosen": -0.3077571392059326, | |
"logps/rejected": -0.43772149085998535, | |
"loss": 0.8806, | |
"rewards/accuracies": 0.5375000238418579, | |
"rewards/chosen": -0.4616357684135437, | |
"rewards/margins": 0.19494646787643433, | |
"rewards/rejected": -0.656582236289978, | |
"step": 220 | |
}, | |
{ | |
"epoch": 0.24520255863539445, | |
"grad_norm": 0.5628307461738586, | |
"learning_rate": 4.715508948078037e-06, | |
"logits/chosen": 11.177714347839355, | |
"logits/rejected": 11.534266471862793, | |
"logps/chosen": -0.31991320848464966, | |
"logps/rejected": -0.4394511282444, | |
"loss": 0.8778, | |
"rewards/accuracies": 0.550000011920929, | |
"rewards/chosen": -0.4798697829246521, | |
"rewards/margins": 0.17930689454078674, | |
"rewards/rejected": -0.6591767072677612, | |
"step": 230 | |
}, | |
{ | |
"epoch": 0.255863539445629, | |
"grad_norm": 0.40485626459121704, | |
"learning_rate": 4.690766700109659e-06, | |
"logits/chosen": 10.132668495178223, | |
"logits/rejected": 10.29063606262207, | |
"logps/chosen": -0.3195653557777405, | |
"logps/rejected": -0.47949132323265076, | |
"loss": 0.8551, | |
"rewards/accuracies": 0.5874999761581421, | |
"rewards/chosen": -0.47934800386428833, | |
"rewards/margins": 0.23988890647888184, | |
"rewards/rejected": -0.7192369699478149, | |
"step": 240 | |
}, | |
{ | |
"epoch": 0.26652452025586354, | |
"grad_norm": 0.6199322938919067, | |
"learning_rate": 4.665063509461098e-06, | |
"logits/chosen": 8.781888008117676, | |
"logits/rejected": 9.237382888793945, | |
"logps/chosen": -0.3370448052883148, | |
"logps/rejected": -0.610824465751648, | |
"loss": 0.8416, | |
"rewards/accuracies": 0.7250000238418579, | |
"rewards/chosen": -0.505567193031311, | |
"rewards/margins": 0.4106695055961609, | |
"rewards/rejected": -0.9162367582321167, | |
"step": 250 | |
}, | |
{ | |
"epoch": 0.26652452025586354, | |
"eval_logits/chosen": 8.437722206115723, | |
"eval_logits/rejected": 8.843962669372559, | |
"eval_logps/chosen": -0.3058585226535797, | |
"eval_logps/rejected": -0.582990825176239, | |
"eval_loss": 0.8036603331565857, | |
"eval_rewards/accuracies": 0.6447368264198303, | |
"eval_rewards/chosen": -0.4587877094745636, | |
"eval_rewards/margins": 0.4156985878944397, | |
"eval_rewards/rejected": -0.8744862079620361, | |
"eval_runtime": 21.4423, | |
"eval_samples_per_second": 28.308, | |
"eval_steps_per_second": 3.544, | |
"step": 250 | |
}, | |
{ | |
"epoch": 0.2771855010660981, | |
"grad_norm": 0.3213505744934082, | |
"learning_rate": 4.638410650401267e-06, | |
"logits/chosen": 7.914826393127441, | |
"logits/rejected": 8.010818481445312, | |
"logps/chosen": -0.3556877374649048, | |
"logps/rejected": -0.7540119886398315, | |
"loss": 0.7811, | |
"rewards/accuracies": 0.6499999761581421, | |
"rewards/chosen": -0.5335315465927124, | |
"rewards/margins": 0.5974863171577454, | |
"rewards/rejected": -1.1310179233551025, | |
"step": 260 | |
}, | |
{ | |
"epoch": 0.2878464818763326, | |
"grad_norm": 1.0119378566741943, | |
"learning_rate": 4.610819813755038e-06, | |
"logits/chosen": 7.584845542907715, | |
"logits/rejected": 7.812608242034912, | |
"logps/chosen": -0.3649575412273407, | |
"logps/rejected": -0.8042632937431335, | |
"loss": 0.7391, | |
"rewards/accuracies": 0.5249999761581421, | |
"rewards/chosen": -0.5474363565444946, | |
"rewards/margins": 0.6589586734771729, | |
"rewards/rejected": -1.206395149230957, | |
"step": 270 | |
}, | |
{ | |
"epoch": 0.29850746268656714, | |
"grad_norm": 0.5339816808700562, | |
"learning_rate": 4.582303101775249e-06, | |
"logits/chosen": 6.687758445739746, | |
"logits/rejected": 6.233181476593018, | |
"logps/chosen": -0.415935218334198, | |
"logps/rejected": -1.2987438440322876, | |
"loss": 0.7419, | |
"rewards/accuracies": 0.6000000238418579, | |
"rewards/chosen": -0.6239027976989746, | |
"rewards/margins": 1.3242127895355225, | |
"rewards/rejected": -1.9481157064437866, | |
"step": 280 | |
}, | |
{ | |
"epoch": 0.3091684434968017, | |
"grad_norm": 0.3514000475406647, | |
"learning_rate": 4.55287302283426e-06, | |
"logits/chosen": 6.2503981590271, | |
"logits/rejected": 5.798542499542236, | |
"logps/chosen": -0.4319223463535309, | |
"logps/rejected": -1.2257453203201294, | |
"loss": 0.7235, | |
"rewards/accuracies": 0.6000000238418579, | |
"rewards/chosen": -0.6478835344314575, | |
"rewards/margins": 1.1907342672348022, | |
"rewards/rejected": -1.8386180400848389, | |
"step": 290 | |
}, | |
{ | |
"epoch": 0.31982942430703626, | |
"grad_norm": 0.6761008501052856, | |
"learning_rate": 4.522542485937369e-06, | |
"logits/chosen": 4.4480695724487305, | |
"logits/rejected": 4.290585994720459, | |
"logps/chosen": -0.42002564668655396, | |
"logps/rejected": -1.4215493202209473, | |
"loss": 0.7058, | |
"rewards/accuracies": 0.637499988079071, | |
"rewards/chosen": -0.6300384402275085, | |
"rewards/margins": 1.5022855997085571, | |
"rewards/rejected": -2.132323980331421, | |
"step": 300 | |
}, | |
{ | |
"epoch": 0.31982942430703626, | |
"eval_logits/chosen": 4.789332389831543, | |
"eval_logits/rejected": 4.481485366821289, | |
"eval_logps/chosen": -0.4049508571624756, | |
"eval_logps/rejected": -1.395646095275879, | |
"eval_loss": 0.6695442199707031, | |
"eval_rewards/accuracies": 0.6710526347160339, | |
"eval_rewards/chosen": -0.6074262857437134, | |
"eval_rewards/margins": 1.4860429763793945, | |
"eval_rewards/rejected": -2.0934693813323975, | |
"eval_runtime": 21.4397, | |
"eval_samples_per_second": 28.312, | |
"eval_steps_per_second": 3.545, | |
"step": 300 | |
}, | |
{ | |
"epoch": 0.3304904051172708, | |
"grad_norm": 0.44682690501213074, | |
"learning_rate": 4.491324795060491e-06, | |
"logits/chosen": 5.487166404724121, | |
"logits/rejected": 4.501384258270264, | |
"logps/chosen": -0.5215579867362976, | |
"logps/rejected": -1.7223398685455322, | |
"loss": 0.6988, | |
"rewards/accuracies": 0.612500011920929, | |
"rewards/chosen": -0.7823370695114136, | |
"rewards/margins": 1.8011726140975952, | |
"rewards/rejected": -2.5835094451904297, | |
"step": 310 | |
}, | |
{ | |
"epoch": 0.3411513859275053, | |
"grad_norm": 0.41085830330848694, | |
"learning_rate": 4.4592336433146e-06, | |
"logits/chosen": 4.162590026855469, | |
"logits/rejected": 2.876271963119507, | |
"logps/chosen": -0.5402930974960327, | |
"logps/rejected": -1.7925996780395508, | |
"loss": 0.6811, | |
"rewards/accuracies": 0.625, | |
"rewards/chosen": -0.8104397058486938, | |
"rewards/margins": 1.8784599304199219, | |
"rewards/rejected": -2.688899517059326, | |
"step": 320 | |
}, | |
{ | |
"epoch": 0.35181236673773986, | |
"grad_norm": 0.5611584186553955, | |
"learning_rate": 4.426283106939474e-06, | |
"logits/chosen": 4.088540077209473, | |
"logits/rejected": 3.081679582595825, | |
"logps/chosen": -0.541223406791687, | |
"logps/rejected": -1.9464069604873657, | |
"loss": 0.6614, | |
"rewards/accuracies": 0.699999988079071, | |
"rewards/chosen": -0.8118351101875305, | |
"rewards/margins": 2.1077752113342285, | |
"rewards/rejected": -2.919610023498535, | |
"step": 330 | |
}, | |
{ | |
"epoch": 0.3624733475479744, | |
"grad_norm": 4.05828857421875, | |
"learning_rate": 4.3924876391293915e-06, | |
"logits/chosen": 3.3937134742736816, | |
"logits/rejected": 2.4182538986206055, | |
"logps/chosen": -0.6656067967414856, | |
"logps/rejected": -1.5255868434906006, | |
"loss": 0.6583, | |
"rewards/accuracies": 0.574999988079071, | |
"rewards/chosen": -0.9984102249145508, | |
"rewards/margins": 1.2899701595306396, | |
"rewards/rejected": -2.2883803844451904, | |
"step": 340 | |
}, | |
{ | |
"epoch": 0.373134328358209, | |
"grad_norm": 0.8311880230903625, | |
"learning_rate": 4.357862063693486e-06, | |
"logits/chosen": 2.503194570541382, | |
"logits/rejected": 1.5284960269927979, | |
"logps/chosen": -0.6593035459518433, | |
"logps/rejected": -2.211193323135376, | |
"loss": 0.5911, | |
"rewards/accuracies": 0.699999988079071, | |
"rewards/chosen": -0.9889553189277649, | |
"rewards/margins": 2.3278346061706543, | |
"rewards/rejected": -3.3167896270751953, | |
"step": 350 | |
}, | |
{ | |
"epoch": 0.373134328358209, | |
"eval_logits/chosen": 2.556962728500366, | |
"eval_logits/rejected": 1.830418586730957, | |
"eval_logps/chosen": -0.6546408534049988, | |
"eval_logps/rejected": -1.9014692306518555, | |
"eval_loss": 0.5961893200874329, | |
"eval_rewards/accuracies": 0.6842105388641357, | |
"eval_rewards/chosen": -0.9819613099098206, | |
"eval_rewards/margins": 1.8702424764633179, | |
"eval_rewards/rejected": -2.852203845977783, | |
"eval_runtime": 21.4393, | |
"eval_samples_per_second": 28.312, | |
"eval_steps_per_second": 3.545, | |
"step": 350 | |
}, | |
{ | |
"epoch": 0.3837953091684435, | |
"grad_norm": 1.4237236976623535, | |
"learning_rate": 4.322421568553529e-06, | |
"logits/chosen": 3.0001542568206787, | |
"logits/rejected": 1.9715242385864258, | |
"logps/chosen": -0.8050466775894165, | |
"logps/rejected": -2.2938907146453857, | |
"loss": 0.58, | |
"rewards/accuracies": 0.737500011920929, | |
"rewards/chosen": -1.20756995677948, | |
"rewards/margins": 2.2332661151885986, | |
"rewards/rejected": -3.440835952758789, | |
"step": 360 | |
}, | |
{ | |
"epoch": 0.39445628997867804, | |
"grad_norm": 2.2651443481445312, | |
"learning_rate": 4.286181699082008e-06, | |
"logits/chosen": 2.7526040077209473, | |
"logits/rejected": 2.05066180229187, | |
"logps/chosen": -1.6301355361938477, | |
"logps/rejected": -2.9630703926086426, | |
"loss": 0.5823, | |
"rewards/accuracies": 0.75, | |
"rewards/chosen": -2.4452033042907715, | |
"rewards/margins": 1.999402642250061, | |
"rewards/rejected": -4.444605827331543, | |
"step": 370 | |
}, | |
{ | |
"epoch": 0.4051172707889126, | |
"grad_norm": 1.9120367765426636, | |
"learning_rate": 4.249158351283414e-06, | |
"logits/chosen": 1.9757938385009766, | |
"logits/rejected": 1.5915673971176147, | |
"logps/chosen": -2.063323497772217, | |
"logps/rejected": -2.899749755859375, | |
"loss": 0.5675, | |
"rewards/accuracies": 0.737500011920929, | |
"rewards/chosen": -3.094984769821167, | |
"rewards/margins": 1.2546398639678955, | |
"rewards/rejected": -4.3496246337890625, | |
"step": 380 | |
}, | |
{ | |
"epoch": 0.4157782515991471, | |
"grad_norm": 3.0018720626831055, | |
"learning_rate": 4.211367764821722e-06, | |
"logits/chosen": 2.541440486907959, | |
"logits/rejected": 1.7436832189559937, | |
"logps/chosen": -2.279510736465454, | |
"logps/rejected": -3.3447775840759277, | |
"loss": 0.4969, | |
"rewards/accuracies": 0.8500000238418579, | |
"rewards/chosen": -3.4192657470703125, | |
"rewards/margins": 1.5979007482528687, | |
"rewards/rejected": -5.017167091369629, | |
"step": 390 | |
}, | |
{ | |
"epoch": 0.42643923240938164, | |
"grad_norm": 1.9656275510787964, | |
"learning_rate": 4.172826515897146e-06, | |
"logits/chosen": 1.6748476028442383, | |
"logits/rejected": 1.0921740531921387, | |
"logps/chosen": -2.147991180419922, | |
"logps/rejected": -3.380042314529419, | |
"loss": 0.5135, | |
"rewards/accuracies": 0.8374999761581421, | |
"rewards/chosen": -3.221986770629883, | |
"rewards/margins": 1.8480768203735352, | |
"rewards/rejected": -5.07006311416626, | |
"step": 400 | |
}, | |
{ | |
"epoch": 0.42643923240938164, | |
"eval_logits/chosen": 2.210231065750122, | |
"eval_logits/rejected": 1.679926872253418, | |
"eval_logps/chosen": -2.044506788253784, | |
"eval_logps/rejected": -3.713956356048584, | |
"eval_loss": 0.47455134987831116, | |
"eval_rewards/accuracies": 0.9342105388641357, | |
"eval_rewards/chosen": -3.0667598247528076, | |
"eval_rewards/margins": 2.5041754245758057, | |
"eval_rewards/rejected": -5.570935249328613, | |
"eval_runtime": 21.4401, | |
"eval_samples_per_second": 28.311, | |
"eval_steps_per_second": 3.545, | |
"step": 400 | |
}, | |
{ | |
"epoch": 0.43710021321961623, | |
"grad_norm": 2.501361131668091, | |
"learning_rate": 4.133551509975264e-06, | |
"logits/chosen": 1.9820306301116943, | |
"logits/rejected": 1.3992068767547607, | |
"logps/chosen": -2.300197124481201, | |
"logps/rejected": -3.813164472579956, | |
"loss": 0.498, | |
"rewards/accuracies": 0.8500000238418579, | |
"rewards/chosen": -3.4502956867218018, | |
"rewards/margins": 2.2694506645202637, | |
"rewards/rejected": -5.7197465896606445, | |
"step": 410 | |
}, | |
{ | |
"epoch": 0.44776119402985076, | |
"grad_norm": 3.828648090362549, | |
"learning_rate": 4.093559974371725e-06, | |
"logits/chosen": 2.7997095584869385, | |
"logits/rejected": 2.4387598037719727, | |
"logps/chosen": -2.687736749649048, | |
"logps/rejected": -4.425741195678711, | |
"loss": 0.4494, | |
"rewards/accuracies": 0.862500011920929, | |
"rewards/chosen": -4.031605243682861, | |
"rewards/margins": 2.607006788253784, | |
"rewards/rejected": -6.638613224029541, | |
"step": 420 | |
}, | |
{ | |
"epoch": 0.4584221748400853, | |
"grad_norm": 2.635803461074829, | |
"learning_rate": 4.052869450695776e-06, | |
"logits/chosen": 2.942661762237549, | |
"logits/rejected": 2.019963026046753, | |
"logps/chosen": -2.98117733001709, | |
"logps/rejected": -4.717232704162598, | |
"loss": 0.4796, | |
"rewards/accuracies": 0.8500000238418579, | |
"rewards/chosen": -4.471765518188477, | |
"rewards/margins": 2.60408353805542, | |
"rewards/rejected": -7.075850009918213, | |
"step": 430 | |
}, | |
{ | |
"epoch": 0.4690831556503198, | |
"grad_norm": 3.140829086303711, | |
"learning_rate": 4.011497787155938e-06, | |
"logits/chosen": 3.2747459411621094, | |
"logits/rejected": 2.2958083152770996, | |
"logps/chosen": -3.129321575164795, | |
"logps/rejected": -4.921725273132324, | |
"loss": 0.4468, | |
"rewards/accuracies": 0.862500011920929, | |
"rewards/chosen": -4.69398307800293, | |
"rewards/margins": 2.688605785369873, | |
"rewards/rejected": -7.3825883865356445, | |
"step": 440 | |
}, | |
{ | |
"epoch": 0.47974413646055436, | |
"grad_norm": 2.7932240962982178, | |
"learning_rate": 3.969463130731183e-06, | |
"logits/chosen": 2.205420970916748, | |
"logits/rejected": 1.4024155139923096, | |
"logps/chosen": -2.7564563751220703, | |
"logps/rejected": -4.563851356506348, | |
"loss": 0.4073, | |
"rewards/accuracies": 0.887499988079071, | |
"rewards/chosen": -4.1346845626831055, | |
"rewards/margins": 2.711092472076416, | |
"rewards/rejected": -6.8457770347595215, | |
"step": 450 | |
}, | |
{ | |
"epoch": 0.47974413646055436, | |
"eval_logits/chosen": 2.0136826038360596, | |
"eval_logits/rejected": 1.561701774597168, | |
"eval_logps/chosen": -2.7486908435821533, | |
"eval_logps/rejected": -4.690793514251709, | |
"eval_loss": 0.41499289870262146, | |
"eval_rewards/accuracies": 0.9210526347160339, | |
"eval_rewards/chosen": -4.123035907745361, | |
"eval_rewards/margins": 2.913153648376465, | |
"eval_rewards/rejected": -7.036189079284668, | |
"eval_runtime": 21.4387, | |
"eval_samples_per_second": 28.313, | |
"eval_steps_per_second": 3.545, | |
"step": 450 | |
}, | |
{ | |
"epoch": 0.4904051172707889, | |
"grad_norm": 2.7059199810028076, | |
"learning_rate": 3.92678391921108e-06, | |
"logits/chosen": 2.257246494293213, | |
"logits/rejected": 1.6654322147369385, | |
"logps/chosen": -3.389554500579834, | |
"logps/rejected": -5.6951165199279785, | |
"loss": 0.4004, | |
"rewards/accuracies": 0.949999988079071, | |
"rewards/chosen": -5.084332466125488, | |
"rewards/margins": 3.4583427906036377, | |
"rewards/rejected": -8.542675018310547, | |
"step": 460 | |
}, | |
{ | |
"epoch": 0.5010660980810234, | |
"grad_norm": 2.245579719543457, | |
"learning_rate": 3.88347887310836e-06, | |
"logits/chosen": 2.3386971950531006, | |
"logits/rejected": 2.086036205291748, | |
"logps/chosen": -3.2753937244415283, | |
"logps/rejected": -5.4362359046936035, | |
"loss": 0.3976, | |
"rewards/accuracies": 0.875, | |
"rewards/chosen": -4.913090705871582, | |
"rewards/margins": 3.241262912750244, | |
"rewards/rejected": -8.154353141784668, | |
"step": 470 | |
}, | |
{ | |
"epoch": 0.511727078891258, | |
"grad_norm": 2.8131167888641357, | |
"learning_rate": 3.839566987447492e-06, | |
"logits/chosen": 1.6951004266738892, | |
"logits/rejected": 1.3795586824417114, | |
"logps/chosen": -3.2933483123779297, | |
"logps/rejected": -5.050060749053955, | |
"loss": 0.3982, | |
"rewards/accuracies": 0.8999999761581421, | |
"rewards/chosen": -4.940022945404053, | |
"rewards/margins": 2.635068416595459, | |
"rewards/rejected": -7.5750908851623535, | |
"step": 480 | |
}, | |
{ | |
"epoch": 0.5223880597014925, | |
"grad_norm": 2.6465814113616943, | |
"learning_rate": 3.795067523432826e-06, | |
"logits/chosen": 3.136894702911377, | |
"logits/rejected": 2.6332411766052246, | |
"logps/chosen": -4.419378757476807, | |
"logps/rejected": -6.467301845550537, | |
"loss": 0.3974, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -6.629067420959473, | |
"rewards/margins": 3.071885585784912, | |
"rewards/rejected": -9.700952529907227, | |
"step": 490 | |
}, | |
{ | |
"epoch": 0.5330490405117271, | |
"grad_norm": 3.6718053817749023, | |
"learning_rate": 3.7500000000000005e-06, | |
"logits/chosen": 1.681780457496643, | |
"logits/rejected": 1.0038775205612183, | |
"logps/chosen": -3.266970157623291, | |
"logps/rejected": -5.594450950622559, | |
"loss": 0.367, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -4.900455474853516, | |
"rewards/margins": 3.4912209510803223, | |
"rewards/rejected": -8.391676902770996, | |
"step": 500 | |
}, | |
{ | |
"epoch": 0.5330490405117271, | |
"eval_logits/chosen": 2.110192060470581, | |
"eval_logits/rejected": 1.7233155965805054, | |
"eval_logps/chosen": -3.0329930782318115, | |
"eval_logps/rejected": -5.3280930519104, | |
"eval_loss": 0.387028306722641, | |
"eval_rewards/accuracies": 0.9210526347160339, | |
"eval_rewards/chosen": -4.549489498138428, | |
"eval_rewards/margins": 3.4426498413085938, | |
"eval_rewards/rejected": -7.9921393394470215, | |
"eval_runtime": 21.4417, | |
"eval_samples_per_second": 28.309, | |
"eval_steps_per_second": 3.545, | |
"step": 500 | |
}, | |
{ | |
"epoch": 0.5437100213219617, | |
"grad_norm": 3.392418622970581, | |
"learning_rate": 3.7043841852542884e-06, | |
"logits/chosen": 3.0106852054595947, | |
"logits/rejected": 2.309483528137207, | |
"logps/chosen": -3.4964828491210938, | |
"logps/rejected": -5.438345909118652, | |
"loss": 0.3776, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -5.244723796844482, | |
"rewards/margins": 2.9127936363220215, | |
"rewards/rejected": -8.15751838684082, | |
"step": 510 | |
}, | |
{ | |
"epoch": 0.5543710021321961, | |
"grad_norm": 3.551785707473755, | |
"learning_rate": 3.658240087799655e-06, | |
"logits/chosen": 2.8624072074890137, | |
"logits/rejected": 2.425307512283325, | |
"logps/chosen": -3.780029773712158, | |
"logps/rejected": -5.929925918579102, | |
"loss": 0.3999, | |
"rewards/accuracies": 0.887499988079071, | |
"rewards/chosen": -5.670044898986816, | |
"rewards/margins": 3.224843978881836, | |
"rewards/rejected": -8.894887924194336, | |
"step": 520 | |
}, | |
{ | |
"epoch": 0.5650319829424307, | |
"grad_norm": 3.0529866218566895, | |
"learning_rate": 3.611587947962319e-06, | |
"logits/chosen": 2.3145668506622314, | |
"logits/rejected": 1.5088326930999756, | |
"logps/chosen": -3.1259171962738037, | |
"logps/rejected": -5.282050132751465, | |
"loss": 0.3346, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -4.688876152038574, | |
"rewards/margins": 3.234198808670044, | |
"rewards/rejected": -7.923074245452881, | |
"step": 530 | |
}, | |
{ | |
"epoch": 0.5756929637526652, | |
"grad_norm": 1.7146470546722412, | |
"learning_rate": 3.564448228912682e-06, | |
"logits/chosen": 2.5195610523223877, | |
"logits/rejected": 2.286477565765381, | |
"logps/chosen": -4.000255107879639, | |
"logps/rejected": -6.475634574890137, | |
"loss": 0.3675, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -6.000383377075195, | |
"rewards/margins": 3.713069438934326, | |
"rewards/rejected": -9.713452339172363, | |
"step": 540 | |
}, | |
{ | |
"epoch": 0.5863539445628998, | |
"grad_norm": 8.31704330444336, | |
"learning_rate": 3.516841607689501e-06, | |
"logits/chosen": 1.8846759796142578, | |
"logits/rejected": 1.5663390159606934, | |
"logps/chosen": -3.9287109375, | |
"logps/rejected": -6.041600704193115, | |
"loss": 0.3921, | |
"rewards/accuracies": 0.862500011920929, | |
"rewards/chosen": -5.89306640625, | |
"rewards/margins": 3.1693339347839355, | |
"rewards/rejected": -9.062400817871094, | |
"step": 550 | |
}, | |
{ | |
"epoch": 0.5863539445628998, | |
"eval_logits/chosen": 2.0545763969421387, | |
"eval_logits/rejected": 1.7521167993545532, | |
"eval_logps/chosen": -3.2991809844970703, | |
"eval_logps/rejected": -5.87559175491333, | |
"eval_loss": 0.35785165429115295, | |
"eval_rewards/accuracies": 0.9342105388641357, | |
"eval_rewards/chosen": -4.948771953582764, | |
"eval_rewards/margins": 3.8646163940429688, | |
"eval_rewards/rejected": -8.813387870788574, | |
"eval_runtime": 21.439, | |
"eval_samples_per_second": 28.313, | |
"eval_steps_per_second": 3.545, | |
"step": 550 | |
}, | |
{ | |
"epoch": 0.5970149253731343, | |
"grad_norm": 4.491275787353516, | |
"learning_rate": 3.4687889661302577e-06, | |
"logits/chosen": 2.9886631965637207, | |
"logits/rejected": 2.2341508865356445, | |
"logps/chosen": -4.050547122955322, | |
"logps/rejected": -6.301305294036865, | |
"loss": 0.3701, | |
"rewards/accuracies": 0.887499988079071, | |
"rewards/chosen": -6.0758209228515625, | |
"rewards/margins": 3.3761372566223145, | |
"rewards/rejected": -9.451958656311035, | |
"step": 560 | |
}, | |
{ | |
"epoch": 0.6076759061833689, | |
"grad_norm": 2.40407395362854, | |
"learning_rate": 3.4203113817116955e-06, | |
"logits/chosen": 3.0933375358581543, | |
"logits/rejected": 2.4652345180511475, | |
"logps/chosen": -3.9168992042541504, | |
"logps/rejected": -5.933487892150879, | |
"loss": 0.3481, | |
"rewards/accuracies": 0.8999999761581421, | |
"rewards/chosen": -5.875349521636963, | |
"rewards/margins": 3.024883508682251, | |
"rewards/rejected": -8.900232315063477, | |
"step": 570 | |
}, | |
{ | |
"epoch": 0.6183368869936035, | |
"grad_norm": 4.073615550994873, | |
"learning_rate": 3.3714301183045382e-06, | |
"logits/chosen": 2.5948500633239746, | |
"logits/rejected": 2.202971935272217, | |
"logps/chosen": -3.6740527153015137, | |
"logps/rejected": -6.205447196960449, | |
"loss": 0.3571, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -5.511078834533691, | |
"rewards/margins": 3.7970924377441406, | |
"rewards/rejected": -9.308171272277832, | |
"step": 580 | |
}, | |
{ | |
"epoch": 0.6289978678038379, | |
"grad_norm": 2.427555799484253, | |
"learning_rate": 3.3221666168464584e-06, | |
"logits/chosen": 1.6076780557632446, | |
"logits/rejected": 1.0556285381317139, | |
"logps/chosen": -3.609313488006592, | |
"logps/rejected": -5.912892818450928, | |
"loss": 0.3666, | |
"rewards/accuracies": 0.875, | |
"rewards/chosen": -5.413969993591309, | |
"rewards/margins": 3.455368757247925, | |
"rewards/rejected": -8.869338989257812, | |
"step": 590 | |
}, | |
{ | |
"epoch": 0.6396588486140725, | |
"grad_norm": 3.910998821258545, | |
"learning_rate": 3.272542485937369e-06, | |
"logits/chosen": 2.5494394302368164, | |
"logits/rejected": 1.9157211780548096, | |
"logps/chosen": -4.038924217224121, | |
"logps/rejected": -6.580315589904785, | |
"loss": 0.3017, | |
"rewards/accuracies": 0.862500011920929, | |
"rewards/chosen": -6.058385372161865, | |
"rewards/margins": 3.812087297439575, | |
"rewards/rejected": -9.87047290802002, | |
"step": 600 | |
}, | |
{ | |
"epoch": 0.6396588486140725, | |
"eval_logits/chosen": 2.179999589920044, | |
"eval_logits/rejected": 1.918432593345642, | |
"eval_logps/chosen": -3.3114354610443115, | |
"eval_logps/rejected": -6.157403469085693, | |
"eval_loss": 0.3409230411052704, | |
"eval_rewards/accuracies": 0.9210526347160339, | |
"eval_rewards/chosen": -4.967153549194336, | |
"eval_rewards/margins": 4.268952369689941, | |
"eval_rewards/rejected": -9.236105918884277, | |
"eval_runtime": 21.4405, | |
"eval_samples_per_second": 28.311, | |
"eval_steps_per_second": 3.545, | |
"step": 600 | |
}, | |
{ | |
"epoch": 0.650319829424307, | |
"grad_norm": 3.6193580627441406, | |
"learning_rate": 3.222579492361179e-06, | |
"logits/chosen": 2.680032968521118, | |
"logits/rejected": 1.9571218490600586, | |
"logps/chosen": -3.6821212768554688, | |
"logps/rejected": -6.76863956451416, | |
"loss": 0.3429, | |
"rewards/accuracies": 0.9375, | |
"rewards/chosen": -5.523181915283203, | |
"rewards/margins": 4.629776954650879, | |
"rewards/rejected": -10.152959823608398, | |
"step": 610 | |
}, | |
{ | |
"epoch": 0.6609808102345416, | |
"grad_norm": 4.7948808670043945, | |
"learning_rate": 3.1722995515381644e-06, | |
"logits/chosen": 2.570702075958252, | |
"logits/rejected": 2.2683706283569336, | |
"logps/chosen": -4.124785423278809, | |
"logps/rejected": -6.595047950744629, | |
"loss": 0.3355, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -6.187178134918213, | |
"rewards/margins": 3.7053933143615723, | |
"rewards/rejected": -9.892572402954102, | |
"step": 620 | |
}, | |
{ | |
"epoch": 0.6716417910447762, | |
"grad_norm": 2.727231025695801, | |
"learning_rate": 3.121724717912138e-06, | |
"logits/chosen": 2.0630054473876953, | |
"logits/rejected": 1.4909979104995728, | |
"logps/chosen": -3.6443302631378174, | |
"logps/rejected": -6.239049911499023, | |
"loss": 0.3218, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -5.466495513916016, | |
"rewards/margins": 3.8920791149139404, | |
"rewards/rejected": -9.358574867248535, | |
"step": 630 | |
}, | |
{ | |
"epoch": 0.6823027718550106, | |
"grad_norm": 4.138846397399902, | |
"learning_rate": 3.0708771752766397e-06, | |
"logits/chosen": 1.6833502054214478, | |
"logits/rejected": 1.2245066165924072, | |
"logps/chosen": -4.318512916564941, | |
"logps/rejected": -7.222353935241699, | |
"loss": 0.2868, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -6.477769374847412, | |
"rewards/margins": 4.35576057434082, | |
"rewards/rejected": -10.833529472351074, | |
"step": 640 | |
}, | |
{ | |
"epoch": 0.6929637526652452, | |
"grad_norm": 7.172155380249023, | |
"learning_rate": 3.019779227044398e-06, | |
"logits/chosen": 2.5162904262542725, | |
"logits/rejected": 2.0894787311553955, | |
"logps/chosen": -4.13573694229126, | |
"logps/rejected": -7.199030876159668, | |
"loss": 0.3491, | |
"rewards/accuracies": 0.9375, | |
"rewards/chosen": -6.2036051750183105, | |
"rewards/margins": 4.594940185546875, | |
"rewards/rejected": -10.798544883728027, | |
"step": 650 | |
}, | |
{ | |
"epoch": 0.6929637526652452, | |
"eval_logits/chosen": 2.003471612930298, | |
"eval_logits/rejected": 1.8008451461791992, | |
"eval_logps/chosen": -3.2873525619506836, | |
"eval_logps/rejected": -6.338763236999512, | |
"eval_loss": 0.31618499755859375, | |
"eval_rewards/accuracies": 0.9342105388641357, | |
"eval_rewards/chosen": -4.931028842926025, | |
"eval_rewards/margins": 4.5771164894104, | |
"eval_rewards/rejected": -9.508145332336426, | |
"eval_runtime": 21.415, | |
"eval_samples_per_second": 28.345, | |
"eval_steps_per_second": 3.549, | |
"step": 650 | |
}, | |
{ | |
"epoch": 0.7036247334754797, | |
"grad_norm": 3.3068535327911377, | |
"learning_rate": 2.9684532864643123e-06, | |
"logits/chosen": 2.852752208709717, | |
"logits/rejected": 2.3139753341674805, | |
"logps/chosen": -3.9495348930358887, | |
"logps/rejected": -6.509753227233887, | |
"loss": 0.2684, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -5.924302101135254, | |
"rewards/margins": 3.8403282165527344, | |
"rewards/rejected": -9.764630317687988, | |
"step": 660 | |
}, | |
{ | |
"epoch": 0.7142857142857143, | |
"grad_norm": 2.737560987472534, | |
"learning_rate": 2.9169218667902562e-06, | |
"logits/chosen": 3.034759998321533, | |
"logits/rejected": 2.2936713695526123, | |
"logps/chosen": -3.855508327484131, | |
"logps/rejected": -6.571882724761963, | |
"loss": 0.2887, | |
"rewards/accuracies": 0.9375, | |
"rewards/chosen": -5.783262252807617, | |
"rewards/margins": 4.0745625495910645, | |
"rewards/rejected": -9.857824325561523, | |
"step": 670 | |
}, | |
{ | |
"epoch": 0.7249466950959488, | |
"grad_norm": 4.387801647186279, | |
"learning_rate": 2.8652075714060296e-06, | |
"logits/chosen": 1.5638288259506226, | |
"logits/rejected": 1.1110167503356934, | |
"logps/chosen": -3.7569847106933594, | |
"logps/rejected": -6.927552223205566, | |
"loss": 0.3319, | |
"rewards/accuracies": 0.9624999761581421, | |
"rewards/chosen": -5.635476589202881, | |
"rewards/margins": 4.755851745605469, | |
"rewards/rejected": -10.391328811645508, | |
"step": 680 | |
}, | |
{ | |
"epoch": 0.7356076759061834, | |
"grad_norm": 2.357923746109009, | |
"learning_rate": 2.813333083910761e-06, | |
"logits/chosen": 2.058987855911255, | |
"logits/rejected": 1.6591012477874756, | |
"logps/chosen": -4.124215126037598, | |
"logps/rejected": -7.349566459655762, | |
"loss": 0.2362, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -6.186322212219238, | |
"rewards/margins": 4.838027000427246, | |
"rewards/rejected": -11.024351119995117, | |
"step": 690 | |
}, | |
{ | |
"epoch": 0.746268656716418, | |
"grad_norm": 4.507752895355225, | |
"learning_rate": 2.761321158169134e-06, | |
"logits/chosen": 2.295815944671631, | |
"logits/rejected": 1.4525251388549805, | |
"logps/chosen": -3.4264607429504395, | |
"logps/rejected": -6.397761344909668, | |
"loss": 0.3257, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -5.139691352844238, | |
"rewards/margins": 4.45695161819458, | |
"rewards/rejected": -9.596643447875977, | |
"step": 700 | |
}, | |
{ | |
"epoch": 0.746268656716418, | |
"eval_logits/chosen": 1.9169458150863647, | |
"eval_logits/rejected": 1.761539101600647, | |
"eval_logps/chosen": -3.324993848800659, | |
"eval_logps/rejected": -6.471699237823486, | |
"eval_loss": 0.3062504529953003, | |
"eval_rewards/accuracies": 0.9473684430122375, | |
"eval_rewards/chosen": -4.987491607666016, | |
"eval_rewards/margins": 4.720058441162109, | |
"eval_rewards/rejected": -9.707548141479492, | |
"eval_runtime": 21.4456, | |
"eval_samples_per_second": 28.304, | |
"eval_steps_per_second": 3.544, | |
"step": 700 | |
}, | |
{ | |
"epoch": 0.7569296375266524, | |
"grad_norm": 6.864930629730225, | |
"learning_rate": 2.70919460833079e-06, | |
"logits/chosen": 1.8696104288101196, | |
"logits/rejected": 1.4743572473526, | |
"logps/chosen": -3.4536032676696777, | |
"logps/rejected": -6.400083065032959, | |
"loss": 0.3255, | |
"rewards/accuracies": 0.8999999761581421, | |
"rewards/chosen": -5.180405616760254, | |
"rewards/margins": 4.4197187423706055, | |
"rewards/rejected": -9.60012435913086, | |
"step": 710 | |
}, | |
{ | |
"epoch": 0.767590618336887, | |
"grad_norm": 5.012319087982178, | |
"learning_rate": 2.6569762988232838e-06, | |
"logits/chosen": 2.861811399459839, | |
"logits/rejected": 2.266331911087036, | |
"logps/chosen": -4.254661560058594, | |
"logps/rejected": -6.501154899597168, | |
"loss": 0.2845, | |
"rewards/accuracies": 0.8999999761581421, | |
"rewards/chosen": -6.381992816925049, | |
"rewards/margins": 3.3697407245635986, | |
"rewards/rejected": -9.751731872558594, | |
"step": 720 | |
}, | |
{ | |
"epoch": 0.7782515991471215, | |
"grad_norm": 1.460620403289795, | |
"learning_rate": 2.604689134322999e-06, | |
"logits/chosen": 2.917189836502075, | |
"logits/rejected": 2.3850274085998535, | |
"logps/chosen": -3.6072356700897217, | |
"logps/rejected": -6.3972039222717285, | |
"loss": 0.2696, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -5.410854339599609, | |
"rewards/margins": 4.184950828552246, | |
"rewards/rejected": -9.595805168151855, | |
"step": 730 | |
}, | |
{ | |
"epoch": 0.7889125799573561, | |
"grad_norm": 5.478528022766113, | |
"learning_rate": 2.5523560497083927e-06, | |
"logits/chosen": 2.2212367057800293, | |
"logits/rejected": 1.6661183834075928, | |
"logps/chosen": -3.353309154510498, | |
"logps/rejected": -6.527769565582275, | |
"loss": 0.2622, | |
"rewards/accuracies": 0.949999988079071, | |
"rewards/chosen": -5.029964447021484, | |
"rewards/margins": 4.761690616607666, | |
"rewards/rejected": -9.791654586791992, | |
"step": 740 | |
}, | |
{ | |
"epoch": 0.7995735607675906, | |
"grad_norm": 2.7094156742095947, | |
"learning_rate": 2.5e-06, | |
"logits/chosen": 2.9158451557159424, | |
"logits/rejected": 2.286925792694092, | |
"logps/chosen": -4.0971856117248535, | |
"logps/rejected": -6.912562370300293, | |
"loss": 0.3243, | |
"rewards/accuracies": 0.887499988079071, | |
"rewards/chosen": -6.145778179168701, | |
"rewards/margins": 4.2230658531188965, | |
"rewards/rejected": -10.368844985961914, | |
"step": 750 | |
}, | |
{ | |
"epoch": 0.7995735607675906, | |
"eval_logits/chosen": 2.0261385440826416, | |
"eval_logits/rejected": 1.8928862810134888, | |
"eval_logps/chosen": -3.538414478302002, | |
"eval_logps/rejected": -6.846959114074707, | |
"eval_loss": 0.2867603600025177, | |
"eval_rewards/accuracies": 0.9342105388641357, | |
"eval_rewards/chosen": -5.307621479034424, | |
"eval_rewards/margins": 4.962815761566162, | |
"eval_rewards/rejected": -10.270438194274902, | |
"eval_runtime": 21.4446, | |
"eval_samples_per_second": 28.305, | |
"eval_steps_per_second": 3.544, | |
"step": 750 | |
}, | |
{ | |
"epoch": 0.8102345415778252, | |
"grad_norm": 2.477257490158081, | |
"learning_rate": 2.447643950291608e-06, | |
"logits/chosen": 1.8020350933074951, | |
"logits/rejected": 1.3997737169265747, | |
"logps/chosen": -3.34975004196167, | |
"logps/rejected": -6.07710075378418, | |
"loss": 0.2394, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -5.024625301361084, | |
"rewards/margins": 4.091025352478027, | |
"rewards/rejected": -9.11565113067627, | |
"step": 760 | |
}, | |
{ | |
"epoch": 0.8208955223880597, | |
"grad_norm": 4.726482391357422, | |
"learning_rate": 2.3953108656770018e-06, | |
"logits/chosen": 2.757497787475586, | |
"logits/rejected": 2.435739517211914, | |
"logps/chosen": -4.294445991516113, | |
"logps/rejected": -7.038791656494141, | |
"loss": 0.3113, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -6.441669464111328, | |
"rewards/margins": 4.116518497467041, | |
"rewards/rejected": -10.558187484741211, | |
"step": 770 | |
}, | |
{ | |
"epoch": 0.8315565031982942, | |
"grad_norm": 2.759181022644043, | |
"learning_rate": 2.3430237011767166e-06, | |
"logits/chosen": 2.788053274154663, | |
"logits/rejected": 2.1961398124694824, | |
"logps/chosen": -3.792672634124756, | |
"logps/rejected": -7.1101579666137695, | |
"loss": 0.3043, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -5.689009666442871, | |
"rewards/margins": 4.976227760314941, | |
"rewards/rejected": -10.665237426757812, | |
"step": 780 | |
}, | |
{ | |
"epoch": 0.8422174840085288, | |
"grad_norm": 3.382880687713623, | |
"learning_rate": 2.290805391669212e-06, | |
"logits/chosen": 2.3644931316375732, | |
"logits/rejected": 1.838971495628357, | |
"logps/chosen": -3.919532060623169, | |
"logps/rejected": -7.125067710876465, | |
"loss": 0.2825, | |
"rewards/accuracies": 0.9375, | |
"rewards/chosen": -5.879298210144043, | |
"rewards/margins": 4.8083038330078125, | |
"rewards/rejected": -10.687601089477539, | |
"step": 790 | |
}, | |
{ | |
"epoch": 0.8528784648187633, | |
"grad_norm": 3.331052541732788, | |
"learning_rate": 2.238678841830867e-06, | |
"logits/chosen": 2.3921093940734863, | |
"logits/rejected": 1.9997615814208984, | |
"logps/chosen": -3.46040678024292, | |
"logps/rejected": -6.8662567138671875, | |
"loss": 0.3094, | |
"rewards/accuracies": 0.949999988079071, | |
"rewards/chosen": -5.190610885620117, | |
"rewards/margins": 5.108774662017822, | |
"rewards/rejected": -10.299385070800781, | |
"step": 800 | |
}, | |
{ | |
"epoch": 0.8528784648187633, | |
"eval_logits/chosen": 2.198673963546753, | |
"eval_logits/rejected": 2.091115713119507, | |
"eval_logps/chosen": -3.606259346008301, | |
"eval_logps/rejected": -7.107792854309082, | |
"eval_loss": 0.27349653840065, | |
"eval_rewards/accuracies": 0.9473684430122375, | |
"eval_rewards/chosen": -5.409388542175293, | |
"eval_rewards/margins": 5.25230073928833, | |
"eval_rewards/rejected": -10.661689758300781, | |
"eval_runtime": 21.4427, | |
"eval_samples_per_second": 28.308, | |
"eval_steps_per_second": 3.544, | |
"step": 800 | |
}, | |
{ | |
"epoch": 0.8635394456289979, | |
"grad_norm": 1.6861388683319092, | |
"learning_rate": 2.186666916089239e-06, | |
"logits/chosen": 3.2604496479034424, | |
"logits/rejected": 2.558086395263672, | |
"logps/chosen": -3.778765916824341, | |
"logps/rejected": -6.954268455505371, | |
"loss": 0.2923, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -5.668148994445801, | |
"rewards/margins": 4.763254165649414, | |
"rewards/rejected": -10.431403160095215, | |
"step": 810 | |
}, | |
{ | |
"epoch": 0.8742004264392325, | |
"grad_norm": 3.100459575653076, | |
"learning_rate": 2.134792428593971e-06, | |
"logits/chosen": 2.6463425159454346, | |
"logits/rejected": 2.2412638664245605, | |
"logps/chosen": -4.3968634605407715, | |
"logps/rejected": -7.397219181060791, | |
"loss": 0.2767, | |
"rewards/accuracies": 0.862500011920929, | |
"rewards/chosen": -6.595294952392578, | |
"rewards/margins": 4.500533103942871, | |
"rewards/rejected": -11.09582805633545, | |
"step": 820 | |
}, | |
{ | |
"epoch": 0.8848614072494669, | |
"grad_norm": 2.7398672103881836, | |
"learning_rate": 2.0830781332097446e-06, | |
"logits/chosen": 2.8484559059143066, | |
"logits/rejected": 2.408825635910034, | |
"logps/chosen": -3.5274035930633545, | |
"logps/rejected": -6.786914825439453, | |
"loss": 0.2587, | |
"rewards/accuracies": 0.8999999761581421, | |
"rewards/chosen": -5.291105270385742, | |
"rewards/margins": 4.889266014099121, | |
"rewards/rejected": -10.180373191833496, | |
"step": 830 | |
}, | |
{ | |
"epoch": 0.8955223880597015, | |
"grad_norm": 4.221662521362305, | |
"learning_rate": 2.031546713535688e-06, | |
"logits/chosen": 2.5181362628936768, | |
"logits/rejected": 2.1050801277160645, | |
"logps/chosen": -3.7997944355010986, | |
"logps/rejected": -7.4044928550720215, | |
"loss": 0.2985, | |
"rewards/accuracies": 0.9750000238418579, | |
"rewards/chosen": -5.6996917724609375, | |
"rewards/margins": 5.407048225402832, | |
"rewards/rejected": -11.10673999786377, | |
"step": 840 | |
}, | |
{ | |
"epoch": 0.906183368869936, | |
"grad_norm": 4.389741897583008, | |
"learning_rate": 1.9802207729556023e-06, | |
"logits/chosen": 1.7594547271728516, | |
"logits/rejected": 1.5125747919082642, | |
"logps/chosen": -3.7186477184295654, | |
"logps/rejected": -7.241146087646484, | |
"loss": 0.278, | |
"rewards/accuracies": 0.949999988079071, | |
"rewards/chosen": -5.577971935272217, | |
"rewards/margins": 5.28374719619751, | |
"rewards/rejected": -10.861719131469727, | |
"step": 850 | |
}, | |
{ | |
"epoch": 0.906183368869936, | |
"eval_logits/chosen": 2.196138381958008, | |
"eval_logits/rejected": 2.091784715652466, | |
"eval_logps/chosen": -3.610405683517456, | |
"eval_logps/rejected": -7.241917610168457, | |
"eval_loss": 0.2656216025352478, | |
"eval_rewards/accuracies": 0.9473684430122375, | |
"eval_rewards/chosen": -5.415609359741211, | |
"eval_rewards/margins": 5.447267532348633, | |
"eval_rewards/rejected": -10.862876892089844, | |
"eval_runtime": 21.4468, | |
"eval_samples_per_second": 28.303, | |
"eval_steps_per_second": 3.544, | |
"step": 850 | |
}, | |
{ | |
"epoch": 0.9168443496801706, | |
"grad_norm": 4.561154365539551, | |
"learning_rate": 1.9291228247233607e-06, | |
"logits/chosen": 3.5380966663360596, | |
"logits/rejected": 3.1444525718688965, | |
"logps/chosen": -4.1595587730407715, | |
"logps/rejected": -7.205758571624756, | |
"loss": 0.252, | |
"rewards/accuracies": 0.9624999761581421, | |
"rewards/chosen": -6.239338397979736, | |
"rewards/margins": 4.56929874420166, | |
"rewards/rejected": -10.808636665344238, | |
"step": 860 | |
}, | |
{ | |
"epoch": 0.9275053304904051, | |
"grad_norm": 3.8300223350524902, | |
"learning_rate": 1.8782752820878636e-06, | |
"logits/chosen": 2.582847833633423, | |
"logits/rejected": 2.341796398162842, | |
"logps/chosen": -3.9722533226013184, | |
"logps/rejected": -7.717828273773193, | |
"loss": 0.3087, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -5.958379745483398, | |
"rewards/margins": 5.618363380432129, | |
"rewards/rejected": -11.576744079589844, | |
"step": 870 | |
}, | |
{ | |
"epoch": 0.9381663113006397, | |
"grad_norm": 2.60563063621521, | |
"learning_rate": 1.827700448461836e-06, | |
"logits/chosen": 2.616633892059326, | |
"logits/rejected": 2.4994826316833496, | |
"logps/chosen": -3.9796085357666016, | |
"logps/rejected": -7.014098167419434, | |
"loss": 0.2733, | |
"rewards/accuracies": 0.9375, | |
"rewards/chosen": -5.969412803649902, | |
"rewards/margins": 4.551734924316406, | |
"rewards/rejected": -10.521148681640625, | |
"step": 880 | |
}, | |
{ | |
"epoch": 0.9488272921108742, | |
"grad_norm": 2.454512119293213, | |
"learning_rate": 1.7774205076388207e-06, | |
"logits/chosen": 2.3873534202575684, | |
"logits/rejected": 1.922217607498169, | |
"logps/chosen": -4.740939617156982, | |
"logps/rejected": -8.011380195617676, | |
"loss": 0.2365, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -7.1114091873168945, | |
"rewards/margins": 4.9056596755981445, | |
"rewards/rejected": -12.017068862915039, | |
"step": 890 | |
}, | |
{ | |
"epoch": 0.9594882729211087, | |
"grad_norm": 5.833195209503174, | |
"learning_rate": 1.7274575140626318e-06, | |
"logits/chosen": 2.42203950881958, | |
"logits/rejected": 1.9541574716567993, | |
"logps/chosen": -4.811551094055176, | |
"logps/rejected": -7.7176384925842285, | |
"loss": 0.2397, | |
"rewards/accuracies": 0.8999999761581421, | |
"rewards/chosen": -7.2173261642456055, | |
"rewards/margins": 4.359131813049316, | |
"rewards/rejected": -11.576457977294922, | |
"step": 900 | |
}, | |
{ | |
"epoch": 0.9594882729211087, | |
"eval_logits/chosen": 2.22493052482605, | |
"eval_logits/rejected": 2.157407760620117, | |
"eval_logps/chosen": -3.8421285152435303, | |
"eval_logps/rejected": -7.569239616394043, | |
"eval_loss": 0.2578875720500946, | |
"eval_rewards/accuracies": 0.9342105388641357, | |
"eval_rewards/chosen": -5.763192653656006, | |
"eval_rewards/margins": 5.590666770935059, | |
"eval_rewards/rejected": -11.353858947753906, | |
"eval_runtime": 21.4496, | |
"eval_samples_per_second": 28.299, | |
"eval_steps_per_second": 3.543, | |
"step": 900 | |
}, | |
{ | |
"epoch": 0.9701492537313433, | |
"grad_norm": 2.5518722534179688, | |
"learning_rate": 1.677833383153542e-06, | |
"logits/chosen": 2.4641318321228027, | |
"logits/rejected": 2.093945264816284, | |
"logps/chosen": -4.148250102996826, | |
"logps/rejected": -7.6053924560546875, | |
"loss": 0.2498, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -6.22237491607666, | |
"rewards/margins": 5.185714244842529, | |
"rewards/rejected": -11.408089637756348, | |
"step": 910 | |
}, | |
{ | |
"epoch": 0.9808102345415778, | |
"grad_norm": 3.0121946334838867, | |
"learning_rate": 1.6285698816954626e-06, | |
"logits/chosen": 2.5797393321990967, | |
"logits/rejected": 2.4002528190612793, | |
"logps/chosen": -4.170048236846924, | |
"logps/rejected": -7.2330145835876465, | |
"loss": 0.243, | |
"rewards/accuracies": 0.875, | |
"rewards/chosen": -6.255072116851807, | |
"rewards/margins": 4.594450950622559, | |
"rewards/rejected": -10.849523544311523, | |
"step": 920 | |
}, | |
{ | |
"epoch": 0.9914712153518124, | |
"grad_norm": 6.5740532875061035, | |
"learning_rate": 1.5796886182883053e-06, | |
"logits/chosen": 2.631937265396118, | |
"logits/rejected": 2.3605117797851562, | |
"logps/chosen": -4.277863502502441, | |
"logps/rejected": -7.579499244689941, | |
"loss": 0.2851, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -6.416795253753662, | |
"rewards/margins": 4.952453136444092, | |
"rewards/rejected": -11.369248390197754, | |
"step": 930 | |
}, | |
{ | |
"epoch": 1.0021321961620469, | |
"grad_norm": 3.664747953414917, | |
"learning_rate": 1.5312110338697427e-06, | |
"logits/chosen": 2.602048635482788, | |
"logits/rejected": 2.3019237518310547, | |
"logps/chosen": -4.545152187347412, | |
"logps/rejected": -8.159948348999023, | |
"loss": 0.2524, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -6.8177289962768555, | |
"rewards/margins": 5.422194004058838, | |
"rewards/rejected": -12.239922523498535, | |
"step": 940 | |
}, | |
{ | |
"epoch": 1.0127931769722816, | |
"grad_norm": 3.0868308544158936, | |
"learning_rate": 1.4831583923105e-06, | |
"logits/chosen": 2.289614677429199, | |
"logits/rejected": 1.9691753387451172, | |
"logps/chosen": -4.014752388000488, | |
"logps/rejected": -7.477097988128662, | |
"loss": 0.2788, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -6.022129058837891, | |
"rewards/margins": 5.19351863861084, | |
"rewards/rejected": -11.215646743774414, | |
"step": 950 | |
}, | |
{ | |
"epoch": 1.0127931769722816, | |
"eval_logits/chosen": 2.137803316116333, | |
"eval_logits/rejected": 2.122602939605713, | |
"eval_logps/chosen": -3.709460496902466, | |
"eval_logps/rejected": -7.5273919105529785, | |
"eval_loss": 0.2563049793243408, | |
"eval_rewards/accuracies": 0.9473684430122375, | |
"eval_rewards/chosen": -5.56419038772583, | |
"eval_rewards/margins": 5.726898670196533, | |
"eval_rewards/rejected": -11.291089057922363, | |
"eval_runtime": 21.4425, | |
"eval_samples_per_second": 28.308, | |
"eval_steps_per_second": 3.544, | |
"step": 950 | |
}, | |
{ | |
"epoch": 1.023454157782516, | |
"grad_norm": 4.523952960968018, | |
"learning_rate": 1.4355517710873184e-06, | |
"logits/chosen": 3.2274162769317627, | |
"logits/rejected": 2.9835076332092285, | |
"logps/chosen": -4.102515697479248, | |
"logps/rejected": -7.5308661460876465, | |
"loss": 0.2394, | |
"rewards/accuracies": 0.887499988079071, | |
"rewards/chosen": -6.153773307800293, | |
"rewards/margins": 5.142527103424072, | |
"rewards/rejected": -11.296300888061523, | |
"step": 960 | |
}, | |
{ | |
"epoch": 1.0341151385927505, | |
"grad_norm": 5.114956378936768, | |
"learning_rate": 1.388412052037682e-06, | |
"logits/chosen": 2.7468860149383545, | |
"logits/rejected": 2.2672038078308105, | |
"logps/chosen": -3.705359935760498, | |
"logps/rejected": -7.7518720626831055, | |
"loss": 0.2419, | |
"rewards/accuracies": 0.949999988079071, | |
"rewards/chosen": -5.558040142059326, | |
"rewards/margins": 6.069769382476807, | |
"rewards/rejected": -11.627809524536133, | |
"step": 970 | |
}, | |
{ | |
"epoch": 1.044776119402985, | |
"grad_norm": 3.243952512741089, | |
"learning_rate": 1.3417599122003464e-06, | |
"logits/chosen": 3.092737913131714, | |
"logits/rejected": 2.6904711723327637, | |
"logps/chosen": -4.205164909362793, | |
"logps/rejected": -7.69910192489624, | |
"loss": 0.275, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -6.307746887207031, | |
"rewards/margins": 5.240906238555908, | |
"rewards/rejected": -11.548653602600098, | |
"step": 980 | |
}, | |
{ | |
"epoch": 1.0554371002132197, | |
"grad_norm": 4.584798812866211, | |
"learning_rate": 1.2956158147457116e-06, | |
"logits/chosen": 1.9939275979995728, | |
"logits/rejected": 1.555945634841919, | |
"logps/chosen": -3.4762721061706543, | |
"logps/rejected": -7.17560338973999, | |
"loss": 0.2523, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -5.2144083976745605, | |
"rewards/margins": 5.548996448516846, | |
"rewards/rejected": -10.763405799865723, | |
"step": 990 | |
}, | |
{ | |
"epoch": 1.0660980810234542, | |
"grad_norm": 3.283021926879883, | |
"learning_rate": 1.2500000000000007e-06, | |
"logits/chosen": 3.2125916481018066, | |
"logits/rejected": 2.6990933418273926, | |
"logps/chosen": -3.818108320236206, | |
"logps/rejected": -7.657105445861816, | |
"loss": 0.281, | |
"rewards/accuracies": 0.9750000238418579, | |
"rewards/chosen": -5.7271623611450195, | |
"rewards/margins": 5.758495330810547, | |
"rewards/rejected": -11.48565673828125, | |
"step": 1000 | |
}, | |
{ | |
"epoch": 1.0660980810234542, | |
"eval_logits/chosen": 2.120778799057007, | |
"eval_logits/rejected": 2.1084940433502197, | |
"eval_logps/chosen": -3.530414342880249, | |
"eval_logps/rejected": -7.398301124572754, | |
"eval_loss": 0.257539838552475, | |
"eval_rewards/accuracies": 0.9473684430122375, | |
"eval_rewards/chosen": -5.295621395111084, | |
"eval_rewards/margins": 5.801830291748047, | |
"eval_rewards/rejected": -11.097451210021973, | |
"eval_runtime": 21.4442, | |
"eval_samples_per_second": 28.306, | |
"eval_steps_per_second": 3.544, | |
"step": 1000 | |
}, | |
{ | |
"epoch": 1.0767590618336886, | |
"grad_norm": 5.394765377044678, | |
"learning_rate": 1.204932476567175e-06, | |
"logits/chosen": 2.473863363265991, | |
"logits/rejected": 2.0679373741149902, | |
"logps/chosen": -4.0380940437316895, | |
"logps/rejected": -7.006314277648926, | |
"loss": 0.2471, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -6.057140350341797, | |
"rewards/margins": 4.452331066131592, | |
"rewards/rejected": -10.509471893310547, | |
"step": 1010 | |
}, | |
{ | |
"epoch": 1.0874200426439233, | |
"grad_norm": 4.336904525756836, | |
"learning_rate": 1.160433012552508e-06, | |
"logits/chosen": 2.386993408203125, | |
"logits/rejected": 2.060482978820801, | |
"logps/chosen": -3.8012115955352783, | |
"logps/rejected": -7.717658996582031, | |
"loss": 0.2664, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -5.701817512512207, | |
"rewards/margins": 5.874671936035156, | |
"rewards/rejected": -11.576489448547363, | |
"step": 1020 | |
}, | |
{ | |
"epoch": 1.0980810234541578, | |
"grad_norm": 3.5417964458465576, | |
"learning_rate": 1.11652112689164e-06, | |
"logits/chosen": 3.1271402835845947, | |
"logits/rejected": 2.6316893100738525, | |
"logps/chosen": -4.178269386291504, | |
"logps/rejected": -7.592785835266113, | |
"loss": 0.237, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -6.267404556274414, | |
"rewards/margins": 5.121774196624756, | |
"rewards/rejected": -11.389179229736328, | |
"step": 1030 | |
}, | |
{ | |
"epoch": 1.1087420042643923, | |
"grad_norm": 5.173072338104248, | |
"learning_rate": 1.073216080788921e-06, | |
"logits/chosen": 2.7356886863708496, | |
"logits/rejected": 2.5491487979888916, | |
"logps/chosen": -4.352596282958984, | |
"logps/rejected": -7.594882011413574, | |
"loss": 0.2152, | |
"rewards/accuracies": 0.8999999761581421, | |
"rewards/chosen": -6.528894901275635, | |
"rewards/margins": 4.86342716217041, | |
"rewards/rejected": -11.392321586608887, | |
"step": 1040 | |
}, | |
{ | |
"epoch": 1.1194029850746268, | |
"grad_norm": 5.773068904876709, | |
"learning_rate": 1.0305368692688175e-06, | |
"logits/chosen": 2.8095126152038574, | |
"logits/rejected": 2.6465506553649902, | |
"logps/chosen": -4.555963039398193, | |
"logps/rejected": -8.115939140319824, | |
"loss": 0.2337, | |
"rewards/accuracies": 0.887499988079071, | |
"rewards/chosen": -6.833944797515869, | |
"rewards/margins": 5.339962959289551, | |
"rewards/rejected": -12.173909187316895, | |
"step": 1050 | |
}, | |
{ | |
"epoch": 1.1194029850746268, | |
"eval_logits/chosen": 2.186048746109009, | |
"eval_logits/rejected": 2.195524215698242, | |
"eval_logps/chosen": -3.748455286026001, | |
"eval_logps/rejected": -7.740383625030518, | |
"eval_loss": 0.24803213775157928, | |
"eval_rewards/accuracies": 0.9342105388641357, | |
"eval_rewards/chosen": -5.622683525085449, | |
"eval_rewards/margins": 5.987893104553223, | |
"eval_rewards/rejected": -11.610575675964355, | |
"eval_runtime": 21.445, | |
"eval_samples_per_second": 28.305, | |
"eval_steps_per_second": 3.544, | |
"step": 1050 | |
}, | |
{ | |
"epoch": 1.1300639658848615, | |
"grad_norm": 3.7594547271728516, | |
"learning_rate": 9.88502212844063e-07, | |
"logits/chosen": 2.6019225120544434, | |
"logits/rejected": 2.0351009368896484, | |
"logps/chosen": -3.759199857711792, | |
"logps/rejected": -7.5345635414123535, | |
"loss": 0.226, | |
"rewards/accuracies": 0.9750000238418579, | |
"rewards/chosen": -5.638800621032715, | |
"rewards/margins": 5.6630449295043945, | |
"rewards/rejected": -11.301844596862793, | |
"step": 1060 | |
}, | |
{ | |
"epoch": 1.140724946695096, | |
"grad_norm": 4.682251453399658, | |
"learning_rate": 9.471305493042243e-07, | |
"logits/chosen": 2.9025776386260986, | |
"logits/rejected": 2.5837249755859375, | |
"logps/chosen": -4.261716842651367, | |
"logps/rejected": -8.397085189819336, | |
"loss": 0.2819, | |
"rewards/accuracies": 0.8999999761581421, | |
"rewards/chosen": -6.392575263977051, | |
"rewards/margins": 6.2030534744262695, | |
"rewards/rejected": -12.59562873840332, | |
"step": 1070 | |
}, | |
{ | |
"epoch": 1.1513859275053304, | |
"grad_norm": 4.629392147064209, | |
"learning_rate": 9.064400256282757e-07, | |
"logits/chosen": 2.7138123512268066, | |
"logits/rejected": 2.5913774967193604, | |
"logps/chosen": -4.063493251800537, | |
"logps/rejected": -7.581247806549072, | |
"loss": 0.2449, | |
"rewards/accuracies": 0.9624999761581421, | |
"rewards/chosen": -6.095240116119385, | |
"rewards/margins": 5.2766313552856445, | |
"rewards/rejected": -11.371870994567871, | |
"step": 1080 | |
}, | |
{ | |
"epoch": 1.1620469083155651, | |
"grad_norm": 3.096658706665039, | |
"learning_rate": 8.664484900247363e-07, | |
"logits/chosen": 2.294330596923828, | |
"logits/rejected": 1.9075424671173096, | |
"logps/chosen": -4.294546604156494, | |
"logps/rejected": -8.126740455627441, | |
"loss": 0.2137, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -6.441819667816162, | |
"rewards/margins": 5.748291492462158, | |
"rewards/rejected": -12.19011116027832, | |
"step": 1090 | |
}, | |
{ | |
"epoch": 1.1727078891257996, | |
"grad_norm": 7.750303268432617, | |
"learning_rate": 8.271734841028553e-07, | |
"logits/chosen": 2.7251312732696533, | |
"logits/rejected": 1.9645192623138428, | |
"logps/chosen": -4.135927677154541, | |
"logps/rejected": -7.487878322601318, | |
"loss": 0.2413, | |
"rewards/accuracies": 0.9375, | |
"rewards/chosen": -6.203891277313232, | |
"rewards/margins": 5.027926445007324, | |
"rewards/rejected": -11.231817245483398, | |
"step": 1100 | |
}, | |
{ | |
"epoch": 1.1727078891257996, | |
"eval_logits/chosen": 2.114663600921631, | |
"eval_logits/rejected": 2.1521003246307373, | |
"eval_logps/chosen": -3.7643239498138428, | |
"eval_logps/rejected": -7.814781665802002, | |
"eval_loss": 0.24271255731582642, | |
"eval_rewards/accuracies": 0.9342105388641357, | |
"eval_rewards/chosen": -5.646485805511475, | |
"eval_rewards/margins": 6.075687408447266, | |
"eval_rewards/rejected": -11.722171783447266, | |
"eval_runtime": 21.4461, | |
"eval_samples_per_second": 28.304, | |
"eval_steps_per_second": 3.544, | |
"step": 1100 | |
}, | |
{ | |
"epoch": 1.183368869936034, | |
"grad_norm": 3.1332218647003174, | |
"learning_rate": 7.886322351782782e-07, | |
"logits/chosen": 2.4613986015319824, | |
"logits/rejected": 2.4474260807037354, | |
"logps/chosen": -4.631771564483643, | |
"logps/rejected": -8.386178970336914, | |
"loss": 0.2099, | |
"rewards/accuracies": 0.9375, | |
"rewards/chosen": -6.947657108306885, | |
"rewards/margins": 5.631610870361328, | |
"rewards/rejected": -12.579267501831055, | |
"step": 1110 | |
}, | |
{ | |
"epoch": 1.1940298507462686, | |
"grad_norm": 3.8220393657684326, | |
"learning_rate": 7.508416487165862e-07, | |
"logits/chosen": 1.7835853099822998, | |
"logits/rejected": 1.3680188655853271, | |
"logps/chosen": -3.990460157394409, | |
"logps/rejected": -7.8566789627075195, | |
"loss": 0.2527, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -5.985690116882324, | |
"rewards/margins": 5.799327850341797, | |
"rewards/rejected": -11.785018920898438, | |
"step": 1120 | |
}, | |
{ | |
"epoch": 1.2046908315565032, | |
"grad_norm": 4.16413688659668, | |
"learning_rate": 7.138183009179922e-07, | |
"logits/chosen": 2.3430938720703125, | |
"logits/rejected": 1.9356731176376343, | |
"logps/chosen": -3.891129970550537, | |
"logps/rejected": -7.389913082122803, | |
"loss": 0.2229, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -5.836695194244385, | |
"rewards/margins": 5.248173713684082, | |
"rewards/rejected": -11.084868431091309, | |
"step": 1130 | |
}, | |
{ | |
"epoch": 1.2153518123667377, | |
"grad_norm": 5.904109954833984, | |
"learning_rate": 6.775784314464717e-07, | |
"logits/chosen": 2.3378713130950928, | |
"logits/rejected": 2.5543103218078613, | |
"logps/chosen": -4.343329906463623, | |
"logps/rejected": -7.562542915344238, | |
"loss": 0.2684, | |
"rewards/accuracies": 0.887499988079071, | |
"rewards/chosen": -6.5149946212768555, | |
"rewards/margins": 4.828819751739502, | |
"rewards/rejected": -11.3438138961792, | |
"step": 1140 | |
}, | |
{ | |
"epoch": 1.2260127931769722, | |
"grad_norm": 3.845862865447998, | |
"learning_rate": 6.421379363065142e-07, | |
"logits/chosen": 3.033965587615967, | |
"logits/rejected": 2.835658550262451, | |
"logps/chosen": -4.35538911819458, | |
"logps/rejected": -8.115766525268555, | |
"loss": 0.2625, | |
"rewards/accuracies": 0.949999988079071, | |
"rewards/chosen": -6.533083438873291, | |
"rewards/margins": 5.640565395355225, | |
"rewards/rejected": -12.173648834228516, | |
"step": 1150 | |
}, | |
{ | |
"epoch": 1.2260127931769722, | |
"eval_logits/chosen": 2.2132980823516846, | |
"eval_logits/rejected": 2.2383644580841064, | |
"eval_logps/chosen": -3.8321533203125, | |
"eval_logps/rejected": -7.898870944976807, | |
"eval_loss": 0.23817622661590576, | |
"eval_rewards/accuracies": 0.9342105388641357, | |
"eval_rewards/chosen": -5.74822998046875, | |
"eval_rewards/margins": 6.100078105926514, | |
"eval_rewards/rejected": -11.848306655883789, | |
"eval_runtime": 21.4432, | |
"eval_samples_per_second": 28.307, | |
"eval_steps_per_second": 3.544, | |
"step": 1150 | |
}, | |
{ | |
"epoch": 1.236673773987207, | |
"grad_norm": 3.4539601802825928, | |
"learning_rate": 6.075123608706093e-07, | |
"logits/chosen": 2.93648099899292, | |
"logits/rejected": 2.8411507606506348, | |
"logps/chosen": -4.176449775695801, | |
"logps/rejected": -7.605371952056885, | |
"loss": 0.2479, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -6.264674186706543, | |
"rewards/margins": 5.143383979797363, | |
"rewards/rejected": -11.408058166503906, | |
"step": 1160 | |
}, | |
{ | |
"epoch": 1.2473347547974414, | |
"grad_norm": 3.6077513694763184, | |
"learning_rate": 5.737168930605272e-07, | |
"logits/chosen": 3.048224925994873, | |
"logits/rejected": 3.061757802963257, | |
"logps/chosen": -4.667507171630859, | |
"logps/rejected": -8.997530937194824, | |
"loss": 0.2096, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -7.001260280609131, | |
"rewards/margins": 6.4950361251831055, | |
"rewards/rejected": -13.496296882629395, | |
"step": 1170 | |
}, | |
{ | |
"epoch": 1.2579957356076759, | |
"grad_norm": 3.400186061859131, | |
"learning_rate": 5.407663566854008e-07, | |
"logits/chosen": 2.4676003456115723, | |
"logits/rejected": 2.2479636669158936, | |
"logps/chosen": -3.8468101024627686, | |
"logps/rejected": -8.152502059936523, | |
"loss": 0.2542, | |
"rewards/accuracies": 0.925000011920929, | |
"rewards/chosen": -5.7702155113220215, | |
"rewards/margins": 6.458538055419922, | |
"rewards/rejected": -12.228754043579102, | |
"step": 1180 | |
}, | |
{ | |
"epoch": 1.2686567164179103, | |
"grad_norm": 4.976436614990234, | |
"learning_rate": 5.086752049395094e-07, | |
"logits/chosen": 2.905099868774414, | |
"logits/rejected": 2.561856746673584, | |
"logps/chosen": -4.313426971435547, | |
"logps/rejected": -8.205011367797852, | |
"loss": 0.26, | |
"rewards/accuracies": 0.9624999761581421, | |
"rewards/chosen": -6.4701409339904785, | |
"rewards/margins": 5.837375640869141, | |
"rewards/rejected": -12.307516098022461, | |
"step": 1190 | |
}, | |
{ | |
"epoch": 1.279317697228145, | |
"grad_norm": 3.4499430656433105, | |
"learning_rate": 4.774575140626317e-07, | |
"logits/chosen": 2.9027159214019775, | |
"logits/rejected": 2.7010762691497803, | |
"logps/chosen": -4.515786170959473, | |
"logps/rejected": -7.926545143127441, | |
"loss": 0.2364, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -6.773678779602051, | |
"rewards/margins": 5.116138935089111, | |
"rewards/rejected": -11.88981819152832, | |
"step": 1200 | |
}, | |
{ | |
"epoch": 1.279317697228145, | |
"eval_logits/chosen": 2.23946213722229, | |
"eval_logits/rejected": 2.2642405033111572, | |
"eval_logps/chosen": -3.853114128112793, | |
"eval_logps/rejected": -7.962078094482422, | |
"eval_loss": 0.23450209200382233, | |
"eval_rewards/accuracies": 0.9342105388641357, | |
"eval_rewards/chosen": -5.7796711921691895, | |
"eval_rewards/margins": 6.163444519042969, | |
"eval_rewards/rejected": -11.943115234375, | |
"eval_runtime": 21.4408, | |
"eval_samples_per_second": 28.311, | |
"eval_steps_per_second": 3.545, | |
"step": 1200 | |
}, | |
{ | |
"epoch": 1.2899786780383795, | |
"grad_norm": 4.555317401885986, | |
"learning_rate": 4.4712697716573994e-07, | |
"logits/chosen": 2.6336121559143066, | |
"logits/rejected": 2.231733798980713, | |
"logps/chosen": -3.876713514328003, | |
"logps/rejected": -8.170385360717773, | |
"loss": 0.2392, | |
"rewards/accuracies": 0.9624999761581421, | |
"rewards/chosen": -5.815070629119873, | |
"rewards/margins": 6.4405083656311035, | |
"rewards/rejected": -12.255578994750977, | |
"step": 1210 | |
}, | |
{ | |
"epoch": 1.3006396588486142, | |
"grad_norm": 2.7224340438842773, | |
"learning_rate": 4.1769689822475147e-07, | |
"logits/chosen": 1.858135461807251, | |
"logits/rejected": 1.8936878442764282, | |
"logps/chosen": -4.599841117858887, | |
"logps/rejected": -8.262335777282715, | |
"loss": 0.2178, | |
"rewards/accuracies": 0.9624999761581421, | |
"rewards/chosen": -6.899762153625488, | |
"rewards/margins": 5.493741035461426, | |
"rewards/rejected": -12.393503189086914, | |
"step": 1220 | |
}, | |
{ | |
"epoch": 1.3113006396588487, | |
"grad_norm": 4.39373779296875, | |
"learning_rate": 3.891801862449629e-07, | |
"logits/chosen": 2.69350004196167, | |
"logits/rejected": 2.4708731174468994, | |
"logps/chosen": -3.8508377075195312, | |
"logps/rejected": -7.5356645584106445, | |
"loss": 0.2408, | |
"rewards/accuracies": 0.9375, | |
"rewards/chosen": -5.776256084442139, | |
"rewards/margins": 5.5272393226623535, | |
"rewards/rejected": -11.303495407104492, | |
"step": 1230 | |
}, | |
{ | |
"epoch": 1.3219616204690832, | |
"grad_norm": 3.29181170463562, | |
"learning_rate": 3.615893495987335e-07, | |
"logits/chosen": 2.418692111968994, | |
"logits/rejected": 2.2509543895721436, | |
"logps/chosen": -4.329282760620117, | |
"logps/rejected": -8.089945793151855, | |
"loss": 0.2236, | |
"rewards/accuracies": 0.9375, | |
"rewards/chosen": -6.493924140930176, | |
"rewards/margins": 5.640995025634766, | |
"rewards/rejected": -12.134920120239258, | |
"step": 1240 | |
}, | |
{ | |
"epoch": 1.3326226012793176, | |
"grad_norm": 2.411123037338257, | |
"learning_rate": 3.3493649053890325e-07, | |
"logits/chosen": 2.37998628616333, | |
"logits/rejected": 1.8894166946411133, | |
"logps/chosen": -3.7197563648223877, | |
"logps/rejected": -6.7860426902771, | |
"loss": 0.234, | |
"rewards/accuracies": 0.9125000238418579, | |
"rewards/chosen": -5.579634189605713, | |
"rewards/margins": 4.599429607391357, | |
"rewards/rejected": -10.17906379699707, | |
"step": 1250 | |
}, | |
{ | |
"epoch": 1.3326226012793176, | |
"eval_logits/chosen": 2.2661728858947754, | |
"eval_logits/rejected": 2.303562641143799, | |
"eval_logps/chosen": -3.8948957920074463, | |
"eval_logps/rejected": -8.014699935913086, | |
"eval_loss": 0.23487773537635803, | |
"eval_rewards/accuracies": 0.9342105388641357, | |
"eval_rewards/chosen": -5.842343807220459, | |
"eval_rewards/margins": 6.179705619812012, | |
"eval_rewards/rejected": -12.022049903869629, | |
"eval_runtime": 21.4443, | |
"eval_samples_per_second": 28.306, | |
"eval_steps_per_second": 3.544, | |
"step": 1250 | |
}, | |
{ | |
"epoch": 1.3432835820895521, | |
"grad_norm": 4.1479668617248535, | |
"learning_rate": 3.092332998903416e-07, | |
"logits/chosen": 3.07938551902771, | |
"logits/rejected": 2.822139263153076, | |
"logps/chosen": -4.570578575134277, | |
"logps/rejected": -7.377262115478516, | |
"loss": 0.2243, | |
"rewards/accuracies": 0.875, | |
"rewards/chosen": -6.855868339538574, | |
"rewards/margins": 4.210025310516357, | |
"rewards/rejected": -11.065893173217773, | |
"step": 1260 | |
}, | |
{ | |
"epoch": 1.3539445628997868, | |
"grad_norm": 2.2094712257385254, | |
"learning_rate": 2.844910519219632e-07, | |
"logits/chosen": 3.2002041339874268, | |
"logits/rejected": 2.545341730117798, | |
"logps/chosen": -4.465962886810303, | |
"logps/rejected": -8.382864952087402, | |
"loss": 0.2266, | |
"rewards/accuracies": 0.949999988079071, | |
"rewards/chosen": -6.698945045471191, | |
"rewards/margins": 5.8753533363342285, | |
"rewards/rejected": -12.574296951293945, | |
"step": 1270 | |
}, | |
{ | |
"epoch": 1.3646055437100213, | |
"grad_norm": 2.466942548751831, | |
"learning_rate": 2.6072059940146775e-07, | |
"logits/chosen": 2.5502431392669678, | |
"logits/rejected": 2.340998649597168, | |
"logps/chosen": -4.121575355529785, | |
"logps/rejected": -7.655613899230957, | |
"loss": 0.2063, | |
"rewards/accuracies": 0.9750000238418579, | |
"rewards/chosen": -6.182363510131836, | |
"rewards/margins": 5.301057815551758, | |
"rewards/rejected": -11.483420372009277, | |
"step": 1280 | |
}, | |
{ | |
"epoch": 1.375266524520256, | |
"grad_norm": 5.0439276695251465, | |
"learning_rate": 2.3793236883495164e-07, | |
"logits/chosen": 2.9525771141052246, | |
"logits/rejected": 2.2864108085632324, | |
"logps/chosen": -4.728396415710449, | |
"logps/rejected": -8.570863723754883, | |
"loss": 0.2538, | |
"rewards/accuracies": 0.9624999761581421, | |
"rewards/chosen": -7.092595100402832, | |
"rewards/margins": 5.763700008392334, | |
"rewards/rejected": -12.856295585632324, | |
"step": 1290 | |
}, | |
{ | |
"epoch": 1.3859275053304905, | |
"grad_norm": 4.124431610107422, | |
"learning_rate": 2.1613635589349756e-07, | |
"logits/chosen": 3.396380662918091, | |
"logits/rejected": 2.761960983276367, | |
"logps/chosen": -4.269867420196533, | |
"logps/rejected": -7.860057830810547, | |
"loss": 0.2363, | |
"rewards/accuracies": 0.949999988079071, | |
"rewards/chosen": -6.404801368713379, | |
"rewards/margins": 5.3852858543396, | |
"rewards/rejected": -11.790085792541504, | |
"step": 1300 | |
}, | |
{ | |
"epoch": 1.3859275053304905, | |
"eval_logits/chosen": 2.230625867843628, | |
"eval_logits/rejected": 2.2632896900177, | |
"eval_logps/chosen": -3.870891809463501, | |
"eval_logps/rejected": -8.023286819458008, | |
"eval_loss": 0.2328612357378006, | |
"eval_rewards/accuracies": 0.9342105388641357, | |
"eval_rewards/chosen": -5.806337833404541, | |
"eval_rewards/margins": 6.228590965270996, | |
"eval_rewards/rejected": -12.034929275512695, | |
"eval_runtime": 21.4392, | |
"eval_samples_per_second": 28.313, | |
"eval_steps_per_second": 3.545, | |
"step": 1300 | |
}, | |
{ | |
"epoch": 1.396588486140725, | |
"grad_norm": 3.541776180267334, | |
"learning_rate": 1.95342121028749e-07, | |
"logits/chosen": 2.9931178092956543, | |
"logits/rejected": 2.68220591545105, | |
"logps/chosen": -4.966144561767578, | |
"logps/rejected": -8.84475040435791, | |
"loss": 0.2443, | |
"rewards/accuracies": 0.862500011920929, | |
"rewards/chosen": -7.449217319488525, | |
"rewards/margins": 5.817907333374023, | |
"rewards/rejected": -13.267123222351074, | |
"step": 1310 | |
}, | |
{ | |
"epoch": 1.4072494669509594, | |
"grad_norm": 4.5976176261901855, | |
"learning_rate": 1.7555878527937164e-07, | |
"logits/chosen": 2.9437661170959473, | |
"logits/rejected": 2.3923139572143555, | |
"logps/chosen": -4.328519344329834, | |
"logps/rejected": -8.65381145477295, | |
"loss": 0.1976, | |
"rewards/accuracies": 0.949999988079071, | |
"rewards/chosen": -6.492778778076172, | |
"rewards/margins": 6.487936973571777, | |
"rewards/rejected": -12.98071575164795, | |
"step": 1320 | |
}, | |
{ | |
"epoch": 1.417910447761194, | |
"grad_norm": 3.115194320678711, | |
"learning_rate": 1.567950262702714e-07, | |
"logits/chosen": 2.265023946762085, | |
"logits/rejected": 2.229419708251953, | |
"logps/chosen": -4.200360298156738, | |
"logps/rejected": -8.543081283569336, | |
"loss": 0.2462, | |
"rewards/accuracies": 0.949999988079071, | |
"rewards/chosen": -6.300539970397949, | |
"rewards/margins": 6.5140814781188965, | |
"rewards/rejected": -12.814620971679688, | |
"step": 1330 | |
}, | |
{ | |
"epoch": 1.4285714285714286, | |
"grad_norm": 4.518002986907959, | |
"learning_rate": 1.3905907440629752e-07, | |
"logits/chosen": 2.888331651687622, | |
"logits/rejected": 2.4711410999298096, | |
"logps/chosen": -4.383464336395264, | |
"logps/rejected": -8.23735523223877, | |
"loss": 0.2305, | |
"rewards/accuracies": 0.8999999761581421, | |
"rewards/chosen": -6.575196266174316, | |
"rewards/margins": 5.780836582183838, | |
"rewards/rejected": -12.356034278869629, | |
"step": 1340 | |
}, | |
{ | |
"epoch": 1.439232409381663, | |
"grad_norm": 4.919490337371826, | |
"learning_rate": 1.223587092621162e-07, | |
"logits/chosen": 2.230055809020996, | |
"logits/rejected": 1.987587332725525, | |
"logps/chosen": -4.470664978027344, | |
"logps/rejected": -8.079784393310547, | |
"loss": 0.2559, | |
"rewards/accuracies": 0.887499988079071, | |
"rewards/chosen": -6.705998420715332, | |
"rewards/margins": 5.413680076599121, | |
"rewards/rejected": -12.119677543640137, | |
"step": 1350 | |
}, | |
{ | |
"epoch": 1.439232409381663, | |
"eval_logits/chosen": 2.2287845611572266, | |
"eval_logits/rejected": 2.2654802799224854, | |
"eval_logps/chosen": -3.906062126159668, | |
"eval_logps/rejected": -8.057478904724121, | |
"eval_loss": 0.2323991060256958, | |
"eval_rewards/accuracies": 0.9342105388641357, | |
"eval_rewards/chosen": -5.85909366607666, | |
"eval_rewards/margins": 6.227125644683838, | |
"eval_rewards/rejected": -12.08621883392334, | |
"eval_runtime": 21.4379, | |
"eval_samples_per_second": 28.314, | |
"eval_steps_per_second": 3.545, | |
"step": 1350 | |
}, | |
{ | |
"epoch": 1.4498933901918978, | |
"grad_norm": 4.102510929107666, | |
"learning_rate": 1.067012561698319e-07, | |
"logits/chosen": 2.5121119022369385, | |
"logits/rejected": 2.2612366676330566, | |
"logps/chosen": -4.727237701416016, | |
"logps/rejected": -8.042572975158691, | |
"loss": 0.2207, | |
"rewards/accuracies": 0.887499988079071, | |
"rewards/chosen": -7.090857028961182, | |
"rewards/margins": 4.973002910614014, | |
"rewards/rejected": -12.063859939575195, | |
"step": 1360 | |
}, | |
{ | |
"epoch": 1.4605543710021323, | |
"grad_norm": 2.8042705059051514, | |
"learning_rate": 9.209358300585474e-08, | |
"logits/chosen": 1.8767907619476318, | |
"logits/rejected": 1.5723143815994263, | |
"logps/chosen": -4.120102882385254, | |
"logps/rejected": -8.098010063171387, | |
"loss": 0.2425, | |
"rewards/accuracies": 0.949999988079071, | |
"rewards/chosen": -6.180154323577881, | |
"rewards/margins": 5.966861724853516, | |
"rewards/rejected": -12.147015571594238, | |
"step": 1370 | |
}, | |
{ | |
"epoch": 1.4712153518123667, | |
"grad_norm": 3.080106258392334, | |
"learning_rate": 7.854209717842231e-08, | |
"logits/chosen": 2.5300233364105225, | |
"logits/rejected": 2.554988145828247, | |
"logps/chosen": -4.894607067108154, | |
"logps/rejected": -8.75386905670166, | |
"loss": 0.2071, | |
"rewards/accuracies": 0.8999999761581421, | |
"rewards/chosen": -7.341910362243652, | |
"rewards/margins": 5.788893222808838, | |
"rewards/rejected": -13.130803108215332, | |
"step": 1380 | |
}, | |
{ | |
"epoch": 1.4818763326226012, | |
"grad_norm": 2.08298397064209, | |
"learning_rate": 6.605274281709929e-08, | |
"logits/chosen": 2.914356231689453, | |
"logits/rejected": 2.8261847496032715, | |
"logps/chosen": -4.921418190002441, | |
"logps/rejected": -8.554636001586914, | |
"loss": 0.2444, | |
"rewards/accuracies": 0.9375, | |
"rewards/chosen": -7.382127285003662, | |
"rewards/margins": 5.449827671051025, | |
"rewards/rejected": -12.831954956054688, | |
"step": 1390 | |
}, | |
{ | |
"epoch": 1.4925373134328357, | |
"grad_norm": 6.096044540405273, | |
"learning_rate": 5.463099816548578e-08, | |
"logits/chosen": 2.8178741931915283, | |
"logits/rejected": 2.5911364555358887, | |
"logps/chosen": -3.9612858295440674, | |
"logps/rejected": -7.696778297424316, | |
"loss": 0.2212, | |
"rewards/accuracies": 0.949999988079071, | |
"rewards/chosen": -5.941928863525391, | |
"rewards/margins": 5.6032395362854, | |
"rewards/rejected": -11.545166969299316, | |
"step": 1400 | |
}, | |
{ | |
"epoch": 1.4925373134328357, | |
"eval_logits/chosen": 2.2310123443603516, | |
"eval_logits/rejected": 2.2684807777404785, | |
"eval_logps/chosen": -3.877824544906616, | |
"eval_logps/rejected": -8.04324722290039, | |
"eval_loss": 0.23210352659225464, | |
"eval_rewards/accuracies": 0.9342105388641357, | |
"eval_rewards/chosen": -5.816736221313477, | |
"eval_rewards/margins": 6.248134136199951, | |
"eval_rewards/rejected": -12.064870834350586, | |
"eval_runtime": 21.4417, | |
"eval_samples_per_second": 28.309, | |
"eval_steps_per_second": 3.544, | |
"step": 1400 | |
}, | |
{ | |
"epoch": 1.5031982942430704, | |
"grad_norm": 2.979947328567505, | |
"learning_rate": 4.428187317827848e-08, | |
"logits/chosen": 3.1346540451049805, | |
"logits/rejected": 2.842684745788574, | |
"logps/chosen": -4.069863319396973, | |
"logps/rejected": -7.434247016906738, | |
"loss": 0.252, | |
"rewards/accuracies": 0.8999999761581421, | |
"rewards/chosen": -6.104796409606934, | |
"rewards/margins": 5.04657506942749, | |
"rewards/rejected": -11.151371002197266, | |
"step": 1410 | |
}, | |
{ | |
"epoch": 1.5138592750533049, | |
"grad_norm": 6.9334516525268555, | |
"learning_rate": 3.5009907323737826e-08, | |
"logits/chosen": 2.7904608249664307, | |
"logits/rejected": 2.514411449432373, | |
"logps/chosen": -4.571829319000244, | |
"logps/rejected": -8.069104194641113, | |
"loss": 0.2341, | |
"rewards/accuracies": 0.887499988079071, | |
"rewards/chosen": -6.857743740081787, | |
"rewards/margins": 5.245913028717041, | |
"rewards/rejected": -12.103656768798828, | |
"step": 1420 | |
}, | |
{ | |
"epoch": 1.5245202558635396, | |
"grad_norm": 5.216058254241943, | |
"learning_rate": 2.681916759252917e-08, | |
"logits/chosen": 2.4043290615081787, | |
"logits/rejected": 1.943373441696167, | |
"logps/chosen": -4.060954570770264, | |
"logps/rejected": -8.304466247558594, | |
"loss": 0.2172, | |
"rewards/accuracies": 0.949999988079071, | |
"rewards/chosen": -6.091432094573975, | |
"rewards/margins": 6.365268230438232, | |
"rewards/rejected": -12.456700325012207, | |
"step": 1430 | |
}, | |
{ | |
"epoch": 1.535181236673774, | |
"grad_norm": 5.58978271484375, | |
"learning_rate": 1.9713246713805588e-08, | |
"logits/chosen": 2.7537190914154053, | |
"logits/rejected": 2.4206302165985107, | |
"logps/chosen": -4.129147052764893, | |
"logps/rejected": -7.661523342132568, | |
"loss": 0.2221, | |
"rewards/accuracies": 0.9375, | |
"rewards/chosen": -6.193720817565918, | |
"rewards/margins": 5.298563480377197, | |
"rewards/rejected": -11.492284774780273, | |
"step": 1440 | |
}, | |
{ | |
"epoch": 1.5458422174840085, | |
"grad_norm": 2.503553628921509, | |
"learning_rate": 1.3695261579316776e-08, | |
"logits/chosen": 2.6379013061523438, | |
"logits/rejected": 2.386941432952881, | |
"logps/chosen": -4.30028772354126, | |
"logps/rejected": -8.543910026550293, | |
"loss": 0.2011, | |
"rewards/accuracies": 0.9750000238418579, | |
"rewards/chosen": -6.450431823730469, | |
"rewards/margins": 6.3654327392578125, | |
"rewards/rejected": -12.815862655639648, | |
"step": 1450 | |
}, | |
{ | |
"epoch": 1.5458422174840085, | |
"eval_logits/chosen": 2.240720748901367, | |
"eval_logits/rejected": 2.2734382152557373, | |
"eval_logps/chosen": -3.8918349742889404, | |
"eval_logps/rejected": -8.040750503540039, | |
"eval_loss": 0.23077060282230377, | |
"eval_rewards/accuracies": 0.9342105388641357, | |
"eval_rewards/chosen": -5.837751865386963, | |
"eval_rewards/margins": 6.223373889923096, | |
"eval_rewards/rejected": -12.061126708984375, | |
"eval_runtime": 21.4419, | |
"eval_samples_per_second": 28.309, | |
"eval_steps_per_second": 3.544, | |
"step": 1450 | |
} | |
], | |
"logging_steps": 10, | |
"max_steps": 1500, | |
"num_input_tokens_seen": 0, | |
"num_train_epochs": 2, | |
"save_steps": 50, | |
"stateful_callbacks": { | |
"TrainerControl": { | |
"args": { | |
"should_epoch_stop": false, | |
"should_evaluate": false, | |
"should_log": false, | |
"should_save": true, | |
"should_training_stop": false | |
}, | |
"attributes": {} | |
} | |
}, | |
"total_flos": 3.476033789191258e+18, | |
"train_batch_size": 1, | |
"trial_name": null, | |
"trial_params": null | |
} | |