{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9990375360923965, "eval_steps": 500, "global_step": 519, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0019249278152069298, "grad_norm": 9833.949798287254, "learning_rate": 5.769230769230769e-06, "loss": 31.1205, "step": 1 }, { "epoch": 0.009624639076034648, "grad_norm": 3973.4641256325745, "learning_rate": 2.8846153846153845e-05, "loss": 21.4091, "step": 5 }, { "epoch": 0.019249278152069296, "grad_norm": 1204.353307038186, "learning_rate": 5.769230769230769e-05, "loss": 21.7988, "step": 10 }, { "epoch": 0.028873917228103944, "grad_norm": 2719.4113741924452, "learning_rate": 8.653846153846152e-05, "loss": 13.1493, "step": 15 }, { "epoch": 0.03849855630413859, "grad_norm": 1285.5933097188467, "learning_rate": 0.00011538461538461538, "loss": 14.9997, "step": 20 }, { "epoch": 0.04812319538017324, "grad_norm": 1368.1133931918484, "learning_rate": 0.00014423076923076922, "loss": 14.5096, "step": 25 }, { "epoch": 0.05774783445620789, "grad_norm": 1789.116574641265, "learning_rate": 0.00017307692307692304, "loss": 12.9147, "step": 30 }, { "epoch": 0.06737247353224254, "grad_norm": 606.4253426587204, "learning_rate": 0.00020192307692307691, "loss": 10.387, "step": 35 }, { "epoch": 0.07699711260827719, "grad_norm": 676.8551431719706, "learning_rate": 0.00023076923076923076, "loss": 11.1262, "step": 40 }, { "epoch": 0.08662175168431184, "grad_norm": 1231.7385257559686, "learning_rate": 0.0002596153846153846, "loss": 12.1561, "step": 45 }, { "epoch": 0.09624639076034648, "grad_norm": 653.2469828069383, "learning_rate": 0.00028846153846153843, "loss": 14.788, "step": 50 }, { "epoch": 0.10587102983638114, "grad_norm": 936.634056698483, "learning_rate": 0.00029996945395293625, "loss": 16.5564, "step": 55 }, { "epoch": 0.11549566891241578, "grad_norm": 1440.9242545303139, "learning_rate": 0.0002997828287165724, "loss": 15.2974, "step": 60 }, { "epoch": 0.12512030798845045, "grad_norm": 638.446747223475, "learning_rate": 0.00029942675913693153, "loss": 15.8667, "step": 65 }, { "epoch": 0.1347449470644851, "grad_norm": 200.266845272572, "learning_rate": 0.0002989016480237121, "loss": 12.8657, "step": 70 }, { "epoch": 0.14436958614051973, "grad_norm": 89.69963142129608, "learning_rate": 0.0002982080894176895, "loss": 10.8674, "step": 75 }, { "epoch": 0.15399422521655437, "grad_norm": 161.99501564355919, "learning_rate": 0.0002973468679186978, "loss": 10.3407, "step": 80 }, { "epoch": 0.16361886429258904, "grad_norm": 53.94884234688268, "learning_rate": 0.0002963189577980372, "loss": 8.3512, "step": 85 }, { "epoch": 0.17324350336862368, "grad_norm": 62.03416025335345, "learning_rate": 0.0002951255218963139, "loss": 7.7878, "step": 90 }, { "epoch": 0.18286814244465832, "grad_norm": 39.794164043801786, "learning_rate": 0.0002937679103079571, "loss": 7.2552, "step": 95 }, { "epoch": 0.19249278152069296, "grad_norm": 50.8231434503254, "learning_rate": 0.00029224765885390143, "loss": 6.6276, "step": 100 }, { "epoch": 0.20211742059672763, "grad_norm": 69.05369995252939, "learning_rate": 0.0002905664873441643, "loss": 6.6291, "step": 105 }, { "epoch": 0.21174205967276227, "grad_norm": 53.35632399080608, "learning_rate": 0.00028872629763228145, "loss": 6.029, "step": 110 }, { "epoch": 0.22136669874879691, "grad_norm": 31.195519429891153, "learning_rate": 0.0002867291714638035, "loss": 5.6465, "step": 115 }, { "epoch": 0.23099133782483156, "grad_norm": 52.701945055972324, "learning_rate": 0.0002845773681212862, "loss": 5.5662, "step": 120 }, { "epoch": 0.24061597690086622, "grad_norm": 77.35992519654677, "learning_rate": 0.00028227332186843884, "loss": 5.6873, "step": 125 }, { "epoch": 0.2502406159769009, "grad_norm": 51.024242740357835, "learning_rate": 0.0002798196391963229, "loss": 5.5508, "step": 130 }, { "epoch": 0.2598652550529355, "grad_norm": 51.18618353864925, "learning_rate": 0.0002772190958747147, "loss": 5.3892, "step": 135 }, { "epoch": 0.2694898941289702, "grad_norm": 58.56789915561519, "learning_rate": 0.00027447463381196973, "loss": 5.2978, "step": 140 }, { "epoch": 0.2791145332050048, "grad_norm": 23.98785567254054, "learning_rate": 0.0002715893577269389, "loss": 5.0187, "step": 145 }, { "epoch": 0.28873917228103946, "grad_norm": 29.781239631615374, "learning_rate": 0.0002685665316367035, "loss": 5.0355, "step": 150 }, { "epoch": 0.2983638113570741, "grad_norm": 39.95978701958715, "learning_rate": 0.0002654095751641007, "loss": 5.0902, "step": 155 }, { "epoch": 0.30798845043310874, "grad_norm": 52.37392360884158, "learning_rate": 0.00026212205966921786, "loss": 4.9294, "step": 160 }, { "epoch": 0.3176130895091434, "grad_norm": 17.241602171689465, "learning_rate": 0.0002587077042092314, "loss": 4.8938, "step": 165 }, { "epoch": 0.3272377285851781, "grad_norm": 13.689917458219496, "learning_rate": 0.00025517037133116085, "loss": 4.6402, "step": 170 }, { "epoch": 0.3368623676612127, "grad_norm": 20.769636603545532, "learning_rate": 0.0002515140627022976, "loss": 4.7864, "step": 175 }, { "epoch": 0.34648700673724736, "grad_norm": 38.55999150285036, "learning_rate": 0.00024774291458325127, "loss": 4.7515, "step": 180 }, { "epoch": 0.35611164581328203, "grad_norm": 13.304187798246316, "learning_rate": 0.00024386119314873578, "loss": 4.5423, "step": 185 }, { "epoch": 0.36573628488931664, "grad_norm": 59.93010691011633, "learning_rate": 0.00023987328966138704, "loss": 4.5956, "step": 190 }, { "epoch": 0.3753609239653513, "grad_norm": 17.66294429465663, "learning_rate": 0.00023578371550407354, "loss": 4.5076, "step": 195 }, { "epoch": 0.3849855630413859, "grad_norm": 7.3884347256705745, "learning_rate": 0.0002315970970763186, "loss": 4.2684, "step": 200 }, { "epoch": 0.3946102021174206, "grad_norm": 33.762369347498066, "learning_rate": 0.00022731817056060802, "loss": 4.2404, "step": 205 }, { "epoch": 0.40423484119345526, "grad_norm": 20.386360679132945, "learning_rate": 0.00022295177656450404, "loss": 4.294, "step": 210 }, { "epoch": 0.4138594802694899, "grad_norm": 28.236859859544925, "learning_rate": 0.00021850285464462677, "loss": 4.1248, "step": 215 }, { "epoch": 0.42348411934552455, "grad_norm": 39.626801274084244, "learning_rate": 0.0002139764377186976, "loss": 4.0508, "step": 220 }, { "epoch": 0.4331087584215592, "grad_norm": 15.698244933481332, "learning_rate": 0.00020937764637196638, "loss": 3.915, "step": 225 }, { "epoch": 0.44273339749759383, "grad_norm": 32.50383319470333, "learning_rate": 0.00020471168306446336, "loss": 3.9952, "step": 230 }, { "epoch": 0.4523580365736285, "grad_norm": 15.613973762528445, "learning_rate": 0.0001999838262456287, "loss": 4.0167, "step": 235 }, { "epoch": 0.4619826756496631, "grad_norm": 7.560143232774055, "learning_rate": 0.0001951994243829781, "loss": 3.9707, "step": 240 }, { "epoch": 0.4716073147256978, "grad_norm": 51.81465410028492, "learning_rate": 0.00019036388991155846, "loss": 3.7988, "step": 245 }, { "epoch": 0.48123195380173245, "grad_norm": 26.4525846375415, "learning_rate": 0.0001854826931110403, "loss": 3.7806, "step": 250 }, { "epoch": 0.49085659287776706, "grad_norm": 21.59913708750363, "learning_rate": 0.0001805613559173714, "loss": 3.6756, "step": 255 }, { "epoch": 0.5004812319538018, "grad_norm": 9.476011704621481, "learning_rate": 0.0001756054456759944, "loss": 3.5717, "step": 260 }, { "epoch": 0.5101058710298364, "grad_norm": 8.391576754509806, "learning_rate": 0.00017062056884369325, "loss": 3.5459, "step": 265 }, { "epoch": 0.519730510105871, "grad_norm": 28.78324805703396, "learning_rate": 0.0001656123646461951, "loss": 3.4701, "step": 270 }, { "epoch": 0.5293551491819056, "grad_norm": 11.001402915726699, "learning_rate": 0.00016058649869870098, "loss": 3.5334, "step": 275 }, { "epoch": 0.5389797882579404, "grad_norm": 6.8987788835950665, "learning_rate": 0.00015554865659656367, "loss": 3.4547, "step": 280 }, { "epoch": 0.548604427333975, "grad_norm": 9.323168999693845, "learning_rate": 0.00015050453748336224, "loss": 3.3446, "step": 285 }, { "epoch": 0.5582290664100096, "grad_norm": 5.3748827349335775, "learning_rate": 0.00014545984760365, "loss": 3.2687, "step": 290 }, { "epoch": 0.5678537054860443, "grad_norm": 5.367511603419207, "learning_rate": 0.00014042029384766938, "loss": 3.2708, "step": 295 }, { "epoch": 0.5774783445620789, "grad_norm": 12.526033486379177, "learning_rate": 0.00013539157729533678, "loss": 3.2415, "step": 300 }, { "epoch": 0.5871029836381135, "grad_norm": 14.896281107597598, "learning_rate": 0.00013037938676679957, "loss": 3.2106, "step": 305 }, { "epoch": 0.5967276227141483, "grad_norm": 12.127131320650086, "learning_rate": 0.00012538939238686286, "loss": 3.1497, "step": 310 }, { "epoch": 0.6063522617901829, "grad_norm": 8.461594875316374, "learning_rate": 0.0001204272391705654, "loss": 3.0511, "step": 315 }, { "epoch": 0.6159769008662175, "grad_norm": 4.919949282741493, "learning_rate": 0.00011549854063716169, "loss": 3.1097, "step": 320 }, { "epoch": 0.6256015399422522, "grad_norm": 5.0799921762929925, "learning_rate": 0.00011060887245973355, "loss": 3.0301, "step": 325 }, { "epoch": 0.6352261790182868, "grad_norm": 6.47369845719392, "learning_rate": 0.00010576376615761647, "loss": 2.9608, "step": 330 }, { "epoch": 0.6448508180943214, "grad_norm": 7.124047081231151, "learning_rate": 0.00010096870283877523, "loss": 3.0135, "step": 335 }, { "epoch": 0.6544754571703562, "grad_norm": 8.31297997032083, "learning_rate": 9.62291069992085e-05, "loss": 2.9059, "step": 340 }, { "epoch": 0.6641000962463908, "grad_norm": 5.136687672624436, "learning_rate": 9.155034038639637e-05, "loss": 2.8699, "step": 345 }, { "epoch": 0.6737247353224254, "grad_norm": 9.36693811377479, "learning_rate": 8.693769593373337e-05, "loss": 2.7094, "step": 350 }, { "epoch": 0.6833493743984601, "grad_norm": 6.847104207590687, "learning_rate": 8.239639177280888e-05, "loss": 2.7623, "step": 355 }, { "epoch": 0.6929740134744947, "grad_norm": 11.941229453450356, "learning_rate": 7.793156533030761e-05, "loss": 2.7692, "step": 360 }, { "epoch": 0.7025986525505293, "grad_norm": 7.1577287914651935, "learning_rate": 7.354826751620954e-05, "loss": 2.6581, "step": 365 }, { "epoch": 0.7122232916265641, "grad_norm": 9.1635159385854, "learning_rate": 6.925145700986301e-05, "loss": 2.6512, "step": 370 }, { "epoch": 0.7218479307025987, "grad_norm": 7.88033732049449, "learning_rate": 6.504599465039542e-05, "loss": 2.6299, "step": 375 }, { "epoch": 0.7314725697786333, "grad_norm": 11.935678139850591, "learning_rate": 6.093663793780725e-05, "loss": 2.554, "step": 380 }, { "epoch": 0.7410972088546679, "grad_norm": 4.8026362420751285, "learning_rate": 5.692803565096988e-05, "loss": 2.5404, "step": 385 }, { "epoch": 0.7507218479307026, "grad_norm": 7.28758130289109, "learning_rate": 5.302472258861687e-05, "loss": 2.5256, "step": 390 }, { "epoch": 0.7603464870067372, "grad_norm": 4.9526592835021885, "learning_rate": 4.923111443927615e-05, "loss": 2.4819, "step": 395 }, { "epoch": 0.7699711260827719, "grad_norm": 4.8819904543262025, "learning_rate": 4.5551502785948405e-05, "loss": 2.5104, "step": 400 }, { "epoch": 0.7795957651588066, "grad_norm": 6.287198504138435, "learning_rate": 4.199005025118158e-05, "loss": 2.4187, "step": 405 }, { "epoch": 0.7892204042348412, "grad_norm": 6.291582745557658, "learning_rate": 3.855078578803424e-05, "loss": 2.3766, "step": 410 }, { "epoch": 0.7988450433108758, "grad_norm": 3.273436919662468, "learning_rate": 3.5237600122254437e-05, "loss": 2.4711, "step": 415 }, { "epoch": 0.8084696823869105, "grad_norm": 4.223924239880851, "learning_rate": 3.2054241350831046e-05, "loss": 2.3606, "step": 420 }, { "epoch": 0.8180943214629451, "grad_norm": 4.552913809634341, "learning_rate": 2.9004310701895837e-05, "loss": 2.3599, "step": 425 }, { "epoch": 0.8277189605389798, "grad_norm": 4.724368774995479, "learning_rate": 2.6091258460773862e-05, "loss": 2.4209, "step": 430 }, { "epoch": 0.8373435996150145, "grad_norm": 2.1385694187612194, "learning_rate": 2.3318380066789787e-05, "loss": 2.2896, "step": 435 }, { "epoch": 0.8469682386910491, "grad_norm": 3.108099479602538, "learning_rate": 2.0688812385247176e-05, "loss": 2.3179, "step": 440 }, { "epoch": 0.8565928777670837, "grad_norm": 2.870974039675886, "learning_rate": 1.8205530158796505e-05, "loss": 2.2415, "step": 445 }, { "epoch": 0.8662175168431184, "grad_norm": 2.74292535298438, "learning_rate": 1.587134264220778e-05, "loss": 2.3037, "step": 450 }, { "epoch": 0.875842155919153, "grad_norm": 1.9132580940232193, "learning_rate": 1.3688890424353726e-05, "loss": 2.2324, "step": 455 }, { "epoch": 0.8854667949951877, "grad_norm": 1.6426233529700058, "learning_rate": 1.1660642440999196e-05, "loss": 2.2031, "step": 460 }, { "epoch": 0.8950914340712224, "grad_norm": 2.036914499117152, "learning_rate": 9.788893181776297e-06, "loss": 2.2569, "step": 465 }, { "epoch": 0.904716073147257, "grad_norm": 2.3703240916735893, "learning_rate": 8.07576009450408e-06, "loss": 2.1718, "step": 470 }, { "epoch": 0.9143407122232916, "grad_norm": 1.963971475130326, "learning_rate": 6.5231811897903714e-06, "loss": 2.198, "step": 475 }, { "epoch": 0.9239653512993262, "grad_norm": 1.6385713856807373, "learning_rate": 5.13291284862452e-06, "loss": 2.1811, "step": 480 }, { "epoch": 0.933589990375361, "grad_norm": 1.7172361943571977, "learning_rate": 3.906527835442064e-06, "loss": 2.2004, "step": 485 }, { "epoch": 0.9432146294513956, "grad_norm": 1.8271284556111347, "learning_rate": 2.8454135189082684e-06, "loss": 2.2041, "step": 490 }, { "epoch": 0.9528392685274302, "grad_norm": 1.5811216247900128, "learning_rate": 1.950770302434157e-06, "loss": 2.1713, "step": 495 }, { "epoch": 0.9624639076034649, "grad_norm": 1.542394985664717, "learning_rate": 1.223610266200009e-06, "loss": 2.1716, "step": 500 }, { "epoch": 0.9720885466794995, "grad_norm": 1.4288680519545454, "learning_rate": 6.647560222224957e-07, "loss": 2.1115, "step": 505 }, { "epoch": 0.9817131857555341, "grad_norm": 1.6294272207246967, "learning_rate": 2.748397837611105e-07, "loss": 2.1781, "step": 510 }, { "epoch": 0.9913378248315688, "grad_norm": 1.6532836279501169, "learning_rate": 5.430265011625579e-08, "loss": 2.1274, "step": 515 }, { "epoch": 0.9990375360923965, "eval_loss": 5.616799831390381, "eval_runtime": 1.4307, "eval_samples_per_second": 4.194, "eval_steps_per_second": 0.699, "step": 519 }, { "epoch": 0.9990375360923965, "step": 519, "total_flos": 17850320289792.0, "train_loss": 5.281234926800737, "train_runtime": 7794.1138, "train_samples_per_second": 2.133, "train_steps_per_second": 0.067 } ], "logging_steps": 5, "max_steps": 519, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 17850320289792.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }