{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 2181, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0004585052728106373, "grad_norm": 25.35940676221757, "learning_rate": 4.5662100456621004e-08, "loss": 1.4356, "step": 1 }, { "epoch": 0.0022925263640531865, "grad_norm": 23.277460508193656, "learning_rate": 2.2831050228310502e-07, "loss": 1.4178, "step": 5 }, { "epoch": 0.004585052728106373, "grad_norm": 15.396159390081614, "learning_rate": 4.5662100456621004e-07, "loss": 1.3928, "step": 10 }, { "epoch": 0.0068775790921595595, "grad_norm": 9.927996187561872, "learning_rate": 6.849315068493151e-07, "loss": 1.2487, "step": 15 }, { "epoch": 0.009170105456212746, "grad_norm": 8.936136397262343, "learning_rate": 9.132420091324201e-07, "loss": 1.1467, "step": 20 }, { "epoch": 0.011462631820265932, "grad_norm": 3.9419002716272007, "learning_rate": 1.1415525114155251e-06, "loss": 1.0321, "step": 25 }, { "epoch": 0.013755158184319119, "grad_norm": 3.2824292809209212, "learning_rate": 1.3698630136986302e-06, "loss": 0.9911, "step": 30 }, { "epoch": 0.016047684548372305, "grad_norm": 3.198808731865913, "learning_rate": 1.5981735159817353e-06, "loss": 0.9499, "step": 35 }, { "epoch": 0.018340210912425492, "grad_norm": 3.200026153105945, "learning_rate": 1.8264840182648401e-06, "loss": 0.9394, "step": 40 }, { "epoch": 0.02063273727647868, "grad_norm": 3.1015042038551264, "learning_rate": 2.0547945205479454e-06, "loss": 0.9374, "step": 45 }, { "epoch": 0.022925263640531865, "grad_norm": 3.0638884680066116, "learning_rate": 2.2831050228310503e-06, "loss": 0.9366, "step": 50 }, { "epoch": 0.02521779000458505, "grad_norm": 3.1218708697344337, "learning_rate": 2.511415525114155e-06, "loss": 0.9072, "step": 55 }, { "epoch": 0.027510316368638238, "grad_norm": 3.030931859384564, "learning_rate": 2.7397260273972604e-06, "loss": 0.896, "step": 60 }, { "epoch": 0.029802842732691424, "grad_norm": 3.183215428730836, "learning_rate": 2.9680365296803653e-06, "loss": 0.904, "step": 65 }, { "epoch": 0.03209536909674461, "grad_norm": 3.1193991823217884, "learning_rate": 3.1963470319634706e-06, "loss": 0.8992, "step": 70 }, { "epoch": 0.0343878954607978, "grad_norm": 3.1562480345048662, "learning_rate": 3.4246575342465754e-06, "loss": 0.9008, "step": 75 }, { "epoch": 0.036680421824850984, "grad_norm": 3.1106379275365263, "learning_rate": 3.6529680365296803e-06, "loss": 0.8835, "step": 80 }, { "epoch": 0.03897294818890417, "grad_norm": 3.1659334626442455, "learning_rate": 3.881278538812785e-06, "loss": 0.8798, "step": 85 }, { "epoch": 0.04126547455295736, "grad_norm": 3.1010027836059533, "learning_rate": 4.109589041095891e-06, "loss": 0.879, "step": 90 }, { "epoch": 0.04355800091701054, "grad_norm": 3.3519588401192273, "learning_rate": 4.337899543378996e-06, "loss": 0.8615, "step": 95 }, { "epoch": 0.04585052728106373, "grad_norm": 3.049285908948199, "learning_rate": 4.566210045662101e-06, "loss": 0.8529, "step": 100 }, { "epoch": 0.048143053645116916, "grad_norm": 3.109756439871898, "learning_rate": 4.7945205479452054e-06, "loss": 0.8654, "step": 105 }, { "epoch": 0.0504355800091701, "grad_norm": 3.1513505710159335, "learning_rate": 5.02283105022831e-06, "loss": 0.8663, "step": 110 }, { "epoch": 0.05272810637322329, "grad_norm": 3.1767156567086614, "learning_rate": 5.251141552511416e-06, "loss": 0.8613, "step": 115 }, { "epoch": 0.055020632737276476, "grad_norm": 3.453537287264967, "learning_rate": 5.479452054794521e-06, "loss": 0.8771, "step": 120 }, { "epoch": 0.05731315910132966, "grad_norm": 3.013155684535603, "learning_rate": 5.7077625570776266e-06, "loss": 0.8473, "step": 125 }, { "epoch": 0.05960568546538285, "grad_norm": 3.425642520518735, "learning_rate": 5.936073059360731e-06, "loss": 0.8521, "step": 130 }, { "epoch": 0.061898211829436035, "grad_norm": 3.031927176672884, "learning_rate": 6.164383561643836e-06, "loss": 0.84, "step": 135 }, { "epoch": 0.06419073819348922, "grad_norm": 3.239390421336056, "learning_rate": 6.392694063926941e-06, "loss": 0.859, "step": 140 }, { "epoch": 0.06648326455754242, "grad_norm": 3.017820442924467, "learning_rate": 6.621004566210046e-06, "loss": 0.86, "step": 145 }, { "epoch": 0.0687757909215956, "grad_norm": 3.0002036905279503, "learning_rate": 6.849315068493151e-06, "loss": 0.8525, "step": 150 }, { "epoch": 0.07106831728564879, "grad_norm": 3.1828998491124016, "learning_rate": 7.077625570776257e-06, "loss": 0.8433, "step": 155 }, { "epoch": 0.07336084364970197, "grad_norm": 3.087610569097963, "learning_rate": 7.305936073059361e-06, "loss": 0.8361, "step": 160 }, { "epoch": 0.07565337001375516, "grad_norm": 3.115099552868115, "learning_rate": 7.534246575342466e-06, "loss": 0.8436, "step": 165 }, { "epoch": 0.07794589637780834, "grad_norm": 3.1551201699069282, "learning_rate": 7.76255707762557e-06, "loss": 0.8311, "step": 170 }, { "epoch": 0.08023842274186153, "grad_norm": 3.2013023977541617, "learning_rate": 7.990867579908676e-06, "loss": 0.8244, "step": 175 }, { "epoch": 0.08253094910591471, "grad_norm": 3.1031180959674716, "learning_rate": 8.219178082191782e-06, "loss": 0.8362, "step": 180 }, { "epoch": 0.08482347546996791, "grad_norm": 3.056534274967503, "learning_rate": 8.447488584474887e-06, "loss": 0.827, "step": 185 }, { "epoch": 0.08711600183402109, "grad_norm": 2.8738007240926016, "learning_rate": 8.675799086757991e-06, "loss": 0.8264, "step": 190 }, { "epoch": 0.08940852819807428, "grad_norm": 2.9833947743009044, "learning_rate": 8.904109589041097e-06, "loss": 0.8364, "step": 195 }, { "epoch": 0.09170105456212746, "grad_norm": 3.0590617698737606, "learning_rate": 9.132420091324201e-06, "loss": 0.8385, "step": 200 }, { "epoch": 0.09399358092618065, "grad_norm": 2.9544649860589964, "learning_rate": 9.360730593607307e-06, "loss": 0.8306, "step": 205 }, { "epoch": 0.09628610729023383, "grad_norm": 3.156467119939513, "learning_rate": 9.589041095890411e-06, "loss": 0.812, "step": 210 }, { "epoch": 0.09857863365428703, "grad_norm": 3.241792877196348, "learning_rate": 9.817351598173517e-06, "loss": 0.8098, "step": 215 }, { "epoch": 0.1008711600183402, "grad_norm": 3.329896188306964, "learning_rate": 9.999993590241675e-06, "loss": 0.8321, "step": 220 }, { "epoch": 0.1031636863823934, "grad_norm": 2.961456684151267, "learning_rate": 9.999769250425817e-06, "loss": 0.8296, "step": 225 }, { "epoch": 0.10545621274644658, "grad_norm": 3.0123856993460723, "learning_rate": 9.999224439127452e-06, "loss": 0.8223, "step": 230 }, { "epoch": 0.10774873911049977, "grad_norm": 3.1722352404227263, "learning_rate": 9.998359191267488e-06, "loss": 0.8183, "step": 235 }, { "epoch": 0.11004126547455295, "grad_norm": 3.339283823835408, "learning_rate": 9.997173562305937e-06, "loss": 0.812, "step": 240 }, { "epoch": 0.11233379183860615, "grad_norm": 3.051005936600519, "learning_rate": 9.995667628238362e-06, "loss": 0.8159, "step": 245 }, { "epoch": 0.11462631820265932, "grad_norm": 3.621892868476315, "learning_rate": 9.993841485591e-06, "loss": 0.8265, "step": 250 }, { "epoch": 0.11691884456671252, "grad_norm": 3.1501195933267727, "learning_rate": 9.991695251414584e-06, "loss": 0.7829, "step": 255 }, { "epoch": 0.1192113709307657, "grad_norm": 3.2077051728198436, "learning_rate": 9.989229063276829e-06, "loss": 0.8061, "step": 260 }, { "epoch": 0.12150389729481889, "grad_norm": 2.813867856532736, "learning_rate": 9.986443079253628e-06, "loss": 0.8088, "step": 265 }, { "epoch": 0.12379642365887207, "grad_norm": 2.953479405448006, "learning_rate": 9.983337477918904e-06, "loss": 0.8013, "step": 270 }, { "epoch": 0.12608895002292525, "grad_norm": 2.9765536692485752, "learning_rate": 9.979912458333179e-06, "loss": 0.8112, "step": 275 }, { "epoch": 0.12838147638697844, "grad_norm": 2.9261553011693313, "learning_rate": 9.976168240030804e-06, "loss": 0.797, "step": 280 }, { "epoch": 0.13067400275103164, "grad_norm": 2.7549890848982668, "learning_rate": 9.972105063005895e-06, "loss": 0.8047, "step": 285 }, { "epoch": 0.13296652911508483, "grad_norm": 2.783923747108222, "learning_rate": 9.96772318769694e-06, "loss": 0.8045, "step": 290 }, { "epoch": 0.13525905547913802, "grad_norm": 2.922181282361273, "learning_rate": 9.96302289497012e-06, "loss": 0.7891, "step": 295 }, { "epoch": 0.1375515818431912, "grad_norm": 2.8387565382348807, "learning_rate": 9.958004486101293e-06, "loss": 0.7756, "step": 300 }, { "epoch": 0.13984410820724438, "grad_norm": 2.869327340764152, "learning_rate": 9.952668282756692e-06, "loss": 0.8027, "step": 305 }, { "epoch": 0.14213663457129758, "grad_norm": 2.874303723785054, "learning_rate": 9.947014626972298e-06, "loss": 0.7826, "step": 310 }, { "epoch": 0.14442916093535077, "grad_norm": 2.737834462358364, "learning_rate": 9.941043881131928e-06, "loss": 0.7702, "step": 315 }, { "epoch": 0.14672168729940394, "grad_norm": 2.858629644409334, "learning_rate": 9.934756427943996e-06, "loss": 0.7761, "step": 320 }, { "epoch": 0.14901421366345713, "grad_norm": 2.941702373835629, "learning_rate": 9.92815267041699e-06, "loss": 0.7778, "step": 325 }, { "epoch": 0.15130674002751032, "grad_norm": 2.832449171435636, "learning_rate": 9.921233031833639e-06, "loss": 0.7747, "step": 330 }, { "epoch": 0.15359926639156352, "grad_norm": 2.838327247569131, "learning_rate": 9.913997955723777e-06, "loss": 0.7798, "step": 335 }, { "epoch": 0.15589179275561668, "grad_norm": 3.0053878829121357, "learning_rate": 9.90644790583592e-06, "loss": 0.7504, "step": 340 }, { "epoch": 0.15818431911966988, "grad_norm": 2.737407601036532, "learning_rate": 9.898583366107539e-06, "loss": 0.7655, "step": 345 }, { "epoch": 0.16047684548372307, "grad_norm": 3.0259958169837717, "learning_rate": 9.890404840634037e-06, "loss": 0.7582, "step": 350 }, { "epoch": 0.16276937184777626, "grad_norm": 2.804766086619055, "learning_rate": 9.881912853636445e-06, "loss": 0.7747, "step": 355 }, { "epoch": 0.16506189821182943, "grad_norm": 2.7915942235581785, "learning_rate": 9.873107949427815e-06, "loss": 0.7584, "step": 360 }, { "epoch": 0.16735442457588262, "grad_norm": 2.8708773578370588, "learning_rate": 9.863990692378333e-06, "loss": 0.7538, "step": 365 }, { "epoch": 0.16964695093993581, "grad_norm": 2.8372441642155097, "learning_rate": 9.854561666879148e-06, "loss": 0.7457, "step": 370 }, { "epoch": 0.171939477303989, "grad_norm": 2.7820083192682197, "learning_rate": 9.844821477304904e-06, "loss": 0.775, "step": 375 }, { "epoch": 0.17423200366804217, "grad_norm": 2.6780715561867066, "learning_rate": 9.834770747975015e-06, "loss": 0.7442, "step": 380 }, { "epoch": 0.17652453003209537, "grad_norm": 2.7545319149727763, "learning_rate": 9.824410123113634e-06, "loss": 0.7416, "step": 385 }, { "epoch": 0.17881705639614856, "grad_norm": 2.6402444423405225, "learning_rate": 9.813740266808375e-06, "loss": 0.7362, "step": 390 }, { "epoch": 0.18110958276020175, "grad_norm": 2.730909608534738, "learning_rate": 9.802761862967731e-06, "loss": 0.7252, "step": 395 }, { "epoch": 0.18340210912425492, "grad_norm": 2.9284254959639355, "learning_rate": 9.791475615277248e-06, "loss": 0.7453, "step": 400 }, { "epoch": 0.1856946354883081, "grad_norm": 2.790088757652803, "learning_rate": 9.779882247154419e-06, "loss": 0.7344, "step": 405 }, { "epoch": 0.1879871618523613, "grad_norm": 2.725250925456166, "learning_rate": 9.76798250170231e-06, "loss": 0.7246, "step": 410 }, { "epoch": 0.1902796882164145, "grad_norm": 2.667869321574359, "learning_rate": 9.755777141661937e-06, "loss": 0.7193, "step": 415 }, { "epoch": 0.19257221458046767, "grad_norm": 2.5119646512097997, "learning_rate": 9.743266949363368e-06, "loss": 0.7402, "step": 420 }, { "epoch": 0.19486474094452086, "grad_norm": 2.847215415311532, "learning_rate": 9.730452726675583e-06, "loss": 0.7173, "step": 425 }, { "epoch": 0.19715726730857405, "grad_norm": 2.779126735326216, "learning_rate": 9.717335294955078e-06, "loss": 0.7157, "step": 430 }, { "epoch": 0.19944979367262725, "grad_norm": 3.4561646981046454, "learning_rate": 9.703915494993215e-06, "loss": 0.7312, "step": 435 }, { "epoch": 0.2017423200366804, "grad_norm": 2.7730394910581913, "learning_rate": 9.690194186962326e-06, "loss": 0.7335, "step": 440 }, { "epoch": 0.2040348464007336, "grad_norm": 2.859201150645261, "learning_rate": 9.676172250360583e-06, "loss": 0.7383, "step": 445 }, { "epoch": 0.2063273727647868, "grad_norm": 2.9209175577350313, "learning_rate": 9.66185058395563e-06, "loss": 0.7263, "step": 450 }, { "epoch": 0.20861989912884, "grad_norm": 2.704547531489439, "learning_rate": 9.647230105726963e-06, "loss": 0.7143, "step": 455 }, { "epoch": 0.21091242549289316, "grad_norm": 2.670951446360455, "learning_rate": 9.632311752807097e-06, "loss": 0.7307, "step": 460 }, { "epoch": 0.21320495185694635, "grad_norm": 3.2268092839390485, "learning_rate": 9.617096481421498e-06, "loss": 0.6985, "step": 465 }, { "epoch": 0.21549747822099954, "grad_norm": 2.939723635315935, "learning_rate": 9.601585266827288e-06, "loss": 0.7181, "step": 470 }, { "epoch": 0.21779000458505274, "grad_norm": 2.7240300289732082, "learning_rate": 9.58577910325074e-06, "loss": 0.7079, "step": 475 }, { "epoch": 0.2200825309491059, "grad_norm": 2.7348057628577815, "learning_rate": 9.569679003823542e-06, "loss": 0.7063, "step": 480 }, { "epoch": 0.2223750573131591, "grad_norm": 2.6209148336683894, "learning_rate": 9.55328600051787e-06, "loss": 0.7019, "step": 485 }, { "epoch": 0.2246675836772123, "grad_norm": 2.7094717894075093, "learning_rate": 9.536601144080224e-06, "loss": 0.6933, "step": 490 }, { "epoch": 0.22696011004126548, "grad_norm": 2.6005478056383393, "learning_rate": 9.5196255039641e-06, "loss": 0.7008, "step": 495 }, { "epoch": 0.22925263640531865, "grad_norm": 2.9435017052734933, "learning_rate": 9.502360168261424e-06, "loss": 0.7168, "step": 500 }, { "epoch": 0.23154516276937184, "grad_norm": 15.281241231781962, "learning_rate": 9.48480624363281e-06, "loss": 0.6968, "step": 505 }, { "epoch": 0.23383768913342504, "grad_norm": 2.803746155734926, "learning_rate": 9.46696485523664e-06, "loss": 0.7176, "step": 510 }, { "epoch": 0.23613021549747823, "grad_norm": 2.9572910983459275, "learning_rate": 9.448837146656924e-06, "loss": 0.6983, "step": 515 }, { "epoch": 0.2384227418615314, "grad_norm": 2.66575290909559, "learning_rate": 9.430424279830014e-06, "loss": 0.679, "step": 520 }, { "epoch": 0.2407152682255846, "grad_norm": 2.6071015601683056, "learning_rate": 9.411727434970121e-06, "loss": 0.6796, "step": 525 }, { "epoch": 0.24300779458963778, "grad_norm": 2.6190152299969975, "learning_rate": 9.392747810493675e-06, "loss": 0.6922, "step": 530 }, { "epoch": 0.24530032095369098, "grad_norm": 2.9035286162764624, "learning_rate": 9.373486622942494e-06, "loss": 0.6881, "step": 535 }, { "epoch": 0.24759284731774414, "grad_norm": 2.722112266367375, "learning_rate": 9.353945106905822e-06, "loss": 0.691, "step": 540 }, { "epoch": 0.24988537368179733, "grad_norm": 2.8551591177378173, "learning_rate": 9.334124514941185e-06, "loss": 0.6786, "step": 545 }, { "epoch": 0.2521779000458505, "grad_norm": 2.789372421806793, "learning_rate": 9.314026117494116e-06, "loss": 0.6965, "step": 550 }, { "epoch": 0.2544704264099037, "grad_norm": 2.943178087845294, "learning_rate": 9.29365120281671e-06, "loss": 0.6734, "step": 555 }, { "epoch": 0.2567629527739569, "grad_norm": 2.9269593678262678, "learning_rate": 9.273001076885059e-06, "loss": 0.6567, "step": 560 }, { "epoch": 0.2590554791380101, "grad_norm": 2.7577714835234457, "learning_rate": 9.252077063315545e-06, "loss": 0.6628, "step": 565 }, { "epoch": 0.2613480055020633, "grad_norm": 2.595587224144848, "learning_rate": 9.230880503279991e-06, "loss": 0.6593, "step": 570 }, { "epoch": 0.26364053186611647, "grad_norm": 2.6421320876444425, "learning_rate": 9.209412755419703e-06, "loss": 0.6616, "step": 575 }, { "epoch": 0.26593305823016966, "grad_norm": 2.5889083746551487, "learning_rate": 9.18767519575838e-06, "loss": 0.6574, "step": 580 }, { "epoch": 0.26822558459422285, "grad_norm": 2.644361824371662, "learning_rate": 9.165669217613919e-06, "loss": 0.6631, "step": 585 }, { "epoch": 0.27051811095827605, "grad_norm": 2.7328270481402166, "learning_rate": 9.143396231509102e-06, "loss": 0.6591, "step": 590 }, { "epoch": 0.2728106373223292, "grad_norm": 2.6202953814608247, "learning_rate": 9.12085766508119e-06, "loss": 0.6465, "step": 595 }, { "epoch": 0.2751031636863824, "grad_norm": 2.688621083531908, "learning_rate": 9.098054962990415e-06, "loss": 0.6678, "step": 600 }, { "epoch": 0.2773956900504356, "grad_norm": 2.684577688850206, "learning_rate": 9.074989586827375e-06, "loss": 0.6478, "step": 605 }, { "epoch": 0.27968821641448877, "grad_norm": 2.6991742230220708, "learning_rate": 9.05166301501936e-06, "loss": 0.6575, "step": 610 }, { "epoch": 0.28198074277854196, "grad_norm": 2.8422733898390353, "learning_rate": 9.028076742735583e-06, "loss": 0.6606, "step": 615 }, { "epoch": 0.28427326914259515, "grad_norm": 3.3111069999457174, "learning_rate": 9.004232281791341e-06, "loss": 0.6501, "step": 620 }, { "epoch": 0.28656579550664835, "grad_norm": 2.8352207612326676, "learning_rate": 8.980131160551118e-06, "loss": 0.6497, "step": 625 }, { "epoch": 0.28885832187070154, "grad_norm": 2.622577509095012, "learning_rate": 8.955774923830618e-06, "loss": 0.6265, "step": 630 }, { "epoch": 0.2911508482347547, "grad_norm": 2.6180287881898363, "learning_rate": 8.931165132797747e-06, "loss": 0.6397, "step": 635 }, { "epoch": 0.29344337459880787, "grad_norm": 2.7463986227282713, "learning_rate": 8.906303364872545e-06, "loss": 0.6668, "step": 640 }, { "epoch": 0.29573590096286106, "grad_norm": 2.6468423935127254, "learning_rate": 8.881191213626084e-06, "loss": 0.6393, "step": 645 }, { "epoch": 0.29802842732691426, "grad_norm": 2.6005030935816245, "learning_rate": 8.855830288678311e-06, "loss": 0.644, "step": 650 }, { "epoch": 0.30032095369096745, "grad_norm": 2.7192686848560554, "learning_rate": 8.83022221559489e-06, "loss": 0.6479, "step": 655 }, { "epoch": 0.30261348005502064, "grad_norm": 2.673457233400223, "learning_rate": 8.804368635783002e-06, "loss": 0.6384, "step": 660 }, { "epoch": 0.30490600641907384, "grad_norm": 2.850654385793331, "learning_rate": 8.778271206386135e-06, "loss": 0.6456, "step": 665 }, { "epoch": 0.30719853278312703, "grad_norm": 2.6958806241423643, "learning_rate": 8.751931600177863e-06, "loss": 0.6025, "step": 670 }, { "epoch": 0.30949105914718017, "grad_norm": 2.764991202053115, "learning_rate": 8.725351505454631e-06, "loss": 0.6194, "step": 675 }, { "epoch": 0.31178358551123336, "grad_norm": 2.6590991144561906, "learning_rate": 8.69853262592754e-06, "loss": 0.6348, "step": 680 }, { "epoch": 0.31407611187528656, "grad_norm": 2.708732600879308, "learning_rate": 8.671476680613134e-06, "loss": 0.6411, "step": 685 }, { "epoch": 0.31636863823933975, "grad_norm": 2.5456418831079457, "learning_rate": 8.644185403723231e-06, "loss": 0.6138, "step": 690 }, { "epoch": 0.31866116460339294, "grad_norm": 2.903106819651818, "learning_rate": 8.616660544553754e-06, "loss": 0.6237, "step": 695 }, { "epoch": 0.32095369096744614, "grad_norm": 2.7280408027219942, "learning_rate": 8.588903867372607e-06, "loss": 0.6138, "step": 700 }, { "epoch": 0.32324621733149933, "grad_norm": 2.886662280669305, "learning_rate": 8.560917151306594e-06, "loss": 0.6066, "step": 705 }, { "epoch": 0.3255387436955525, "grad_norm": 2.6016420791711994, "learning_rate": 8.53270219022738e-06, "loss": 0.6126, "step": 710 }, { "epoch": 0.32783127005960566, "grad_norm": 2.5696831024854827, "learning_rate": 8.50426079263651e-06, "loss": 0.6191, "step": 715 }, { "epoch": 0.33012379642365886, "grad_norm": 2.789642739261612, "learning_rate": 8.475594781549483e-06, "loss": 0.6171, "step": 720 }, { "epoch": 0.33241632278771205, "grad_norm": 2.662350967821026, "learning_rate": 8.446705994378913e-06, "loss": 0.6262, "step": 725 }, { "epoch": 0.33470884915176524, "grad_norm": 2.749133969632543, "learning_rate": 8.417596282816742e-06, "loss": 0.6084, "step": 730 }, { "epoch": 0.33700137551581844, "grad_norm": 2.8389384155162736, "learning_rate": 8.388267512715565e-06, "loss": 0.6089, "step": 735 }, { "epoch": 0.33929390187987163, "grad_norm": 2.6423715957870115, "learning_rate": 8.358721563969027e-06, "loss": 0.5912, "step": 740 }, { "epoch": 0.3415864282439248, "grad_norm": 2.582427374014035, "learning_rate": 8.328960330391325e-06, "loss": 0.6015, "step": 745 }, { "epoch": 0.343878954607978, "grad_norm": 2.5641005198848763, "learning_rate": 8.298985719595824e-06, "loss": 0.6127, "step": 750 }, { "epoch": 0.34617148097203115, "grad_norm": 2.573968171901929, "learning_rate": 8.268799652872786e-06, "loss": 0.6108, "step": 755 }, { "epoch": 0.34846400733608435, "grad_norm": 2.555840575858041, "learning_rate": 8.23840406506621e-06, "loss": 0.6013, "step": 760 }, { "epoch": 0.35075653370013754, "grad_norm": 2.608505400595271, "learning_rate": 8.207800904449829e-06, "loss": 0.5868, "step": 765 }, { "epoch": 0.35304906006419073, "grad_norm": 2.564041005915397, "learning_rate": 8.176992132602221e-06, "loss": 0.5935, "step": 770 }, { "epoch": 0.3553415864282439, "grad_norm": 2.835188198766609, "learning_rate": 8.145979724281079e-06, "loss": 0.577, "step": 775 }, { "epoch": 0.3576341127922971, "grad_norm": 2.624154236961289, "learning_rate": 8.114765667296628e-06, "loss": 0.5807, "step": 780 }, { "epoch": 0.3599266391563503, "grad_norm": 2.803920892055745, "learning_rate": 8.083351962384234e-06, "loss": 0.5827, "step": 785 }, { "epoch": 0.3622191655204035, "grad_norm": 2.7453769474392438, "learning_rate": 8.051740623076132e-06, "loss": 0.5743, "step": 790 }, { "epoch": 0.36451169188445665, "grad_norm": 2.642012832230722, "learning_rate": 8.019933675572389e-06, "loss": 0.5924, "step": 795 }, { "epoch": 0.36680421824850984, "grad_norm": 2.5959618878893496, "learning_rate": 7.987933158611013e-06, "loss": 0.5765, "step": 800 }, { "epoch": 0.36909674461256303, "grad_norm": 2.6981842811728107, "learning_rate": 7.95574112333729e-06, "loss": 0.5636, "step": 805 }, { "epoch": 0.3713892709766162, "grad_norm": 2.7155825019244246, "learning_rate": 7.923359633172299e-06, "loss": 0.5676, "step": 810 }, { "epoch": 0.3736817973406694, "grad_norm": 2.722727252289237, "learning_rate": 7.890790763680658e-06, "loss": 0.5849, "step": 815 }, { "epoch": 0.3759743237047226, "grad_norm": 2.5941959497564073, "learning_rate": 7.85803660243749e-06, "loss": 0.582, "step": 820 }, { "epoch": 0.3782668500687758, "grad_norm": 2.448527666302428, "learning_rate": 7.8250992488946e-06, "loss": 0.586, "step": 825 }, { "epoch": 0.380559376432829, "grad_norm": 2.786081596311819, "learning_rate": 7.791980814245931e-06, "loss": 0.5547, "step": 830 }, { "epoch": 0.38285190279688214, "grad_norm": 2.6225345564151237, "learning_rate": 7.758683421292217e-06, "loss": 0.5562, "step": 835 }, { "epoch": 0.38514442916093533, "grad_norm": 2.495977821656378, "learning_rate": 7.72520920430493e-06, "loss": 0.5728, "step": 840 }, { "epoch": 0.3874369555249885, "grad_norm": 2.5523314447232535, "learning_rate": 7.691560308889478e-06, "loss": 0.5748, "step": 845 }, { "epoch": 0.3897294818890417, "grad_norm": 2.702511447586494, "learning_rate": 7.657738891847679e-06, "loss": 0.5651, "step": 850 }, { "epoch": 0.3920220082530949, "grad_norm": 2.6729070020445533, "learning_rate": 7.623747121039512e-06, "loss": 0.5716, "step": 855 }, { "epoch": 0.3943145346171481, "grad_norm": 2.7351708064638665, "learning_rate": 7.589587175244162e-06, "loss": 0.565, "step": 860 }, { "epoch": 0.3966070609812013, "grad_norm": 2.5916997954156638, "learning_rate": 7.555261244020371e-06, "loss": 0.5691, "step": 865 }, { "epoch": 0.3988995873452545, "grad_norm": 2.4806248685486407, "learning_rate": 7.520771527566093e-06, "loss": 0.5672, "step": 870 }, { "epoch": 0.40119211370930763, "grad_norm": 2.691711711440267, "learning_rate": 7.486120236577464e-06, "loss": 0.5555, "step": 875 }, { "epoch": 0.4034846400733608, "grad_norm": 2.6506103202422797, "learning_rate": 7.451309592107104e-06, "loss": 0.5548, "step": 880 }, { "epoch": 0.405777166437414, "grad_norm": 2.5210545941984983, "learning_rate": 7.416341825421755e-06, "loss": 0.573, "step": 885 }, { "epoch": 0.4080696928014672, "grad_norm": 2.7103495153803627, "learning_rate": 7.381219177859257e-06, "loss": 0.5428, "step": 890 }, { "epoch": 0.4103622191655204, "grad_norm": 2.5223081344987826, "learning_rate": 7.345943900684896e-06, "loss": 0.5605, "step": 895 }, { "epoch": 0.4126547455295736, "grad_norm": 2.5684242617186364, "learning_rate": 7.310518254947092e-06, "loss": 0.5432, "step": 900 }, { "epoch": 0.4149472718936268, "grad_norm": 2.8905063764239327, "learning_rate": 7.274944511332479e-06, "loss": 0.5355, "step": 905 }, { "epoch": 0.41723979825768, "grad_norm": 2.7288840976281543, "learning_rate": 7.239224950020359e-06, "loss": 0.5583, "step": 910 }, { "epoch": 0.4195323246217332, "grad_norm": 2.573090270715344, "learning_rate": 7.203361860536544e-06, "loss": 0.5528, "step": 915 }, { "epoch": 0.4218248509857863, "grad_norm": 2.7074335935753897, "learning_rate": 7.167357541606613e-06, "loss": 0.5457, "step": 920 }, { "epoch": 0.4241173773498395, "grad_norm": 2.6225623425429614, "learning_rate": 7.131214301008564e-06, "loss": 0.5405, "step": 925 }, { "epoch": 0.4264099037138927, "grad_norm": 2.638186367850455, "learning_rate": 7.094934455424889e-06, "loss": 0.5457, "step": 930 }, { "epoch": 0.4287024300779459, "grad_norm": 2.663625944879504, "learning_rate": 7.058520330294087e-06, "loss": 0.5499, "step": 935 }, { "epoch": 0.4309949564419991, "grad_norm": 2.594656111210185, "learning_rate": 7.021974259661607e-06, "loss": 0.5471, "step": 940 }, { "epoch": 0.4332874828060523, "grad_norm": 2.558300587882855, "learning_rate": 6.985298586030241e-06, "loss": 0.5465, "step": 945 }, { "epoch": 0.4355800091701055, "grad_norm": 2.6435075817238425, "learning_rate": 6.948495660209983e-06, "loss": 0.5331, "step": 950 }, { "epoch": 0.43787253553415867, "grad_norm": 2.494991656905618, "learning_rate": 6.9115678411673345e-06, "loss": 0.5371, "step": 955 }, { "epoch": 0.4401650618982118, "grad_norm": 2.4881542600695643, "learning_rate": 6.8745174958741164e-06, "loss": 0.5329, "step": 960 }, { "epoch": 0.442457588262265, "grad_norm": 2.552409503690461, "learning_rate": 6.837346999155743e-06, "loss": 0.532, "step": 965 }, { "epoch": 0.4447501146263182, "grad_norm": 2.4970182042863445, "learning_rate": 6.800058733539003e-06, "loss": 0.5376, "step": 970 }, { "epoch": 0.4470426409903714, "grad_norm": 2.468594629574796, "learning_rate": 6.762655089099353e-06, "loss": 0.513, "step": 975 }, { "epoch": 0.4493351673544246, "grad_norm": 2.5797501981324453, "learning_rate": 6.725138463307714e-06, "loss": 0.5408, "step": 980 }, { "epoch": 0.4516276937184778, "grad_norm": 2.8482359445979246, "learning_rate": 6.687511260876799e-06, "loss": 0.5189, "step": 985 }, { "epoch": 0.45392022008253097, "grad_norm": 2.6612518014120816, "learning_rate": 6.649775893606982e-06, "loss": 0.5318, "step": 990 }, { "epoch": 0.45621274644658416, "grad_norm": 2.5372082111080347, "learning_rate": 6.611934780231704e-06, "loss": 0.5076, "step": 995 }, { "epoch": 0.4585052728106373, "grad_norm": 2.4460238122171916, "learning_rate": 6.573990346262445e-06, "loss": 0.5028, "step": 1000 }, { "epoch": 0.4607977991746905, "grad_norm": 2.5523381259232747, "learning_rate": 6.535945023833249e-06, "loss": 0.5188, "step": 1005 }, { "epoch": 0.4630903255387437, "grad_norm": 2.6717883324323104, "learning_rate": 6.497801251544833e-06, "loss": 0.5137, "step": 1010 }, { "epoch": 0.4653828519027969, "grad_norm": 2.4441200104866763, "learning_rate": 6.459561474308278e-06, "loss": 0.513, "step": 1015 }, { "epoch": 0.4676753782668501, "grad_norm": 2.4626953473958046, "learning_rate": 6.421228143188325e-06, "loss": 0.5241, "step": 1020 }, { "epoch": 0.46996790463090327, "grad_norm": 2.414799048761899, "learning_rate": 6.382803715246254e-06, "loss": 0.5265, "step": 1025 }, { "epoch": 0.47226043099495646, "grad_norm": 2.661888186403354, "learning_rate": 6.344290653382408e-06, "loss": 0.5122, "step": 1030 }, { "epoch": 0.47455295735900965, "grad_norm": 2.705613301623184, "learning_rate": 6.305691426178316e-06, "loss": 0.5076, "step": 1035 }, { "epoch": 0.4768454837230628, "grad_norm": 2.5901180556298007, "learning_rate": 6.267008507738472e-06, "loss": 0.5309, "step": 1040 }, { "epoch": 0.479138010087116, "grad_norm": 2.5393961483789345, "learning_rate": 6.228244377531747e-06, "loss": 0.506, "step": 1045 }, { "epoch": 0.4814305364511692, "grad_norm": 2.5959034041763154, "learning_rate": 6.189401520232464e-06, "loss": 0.5065, "step": 1050 }, { "epoch": 0.48372306281522237, "grad_norm": 2.6419168193929963, "learning_rate": 6.150482425561135e-06, "loss": 0.5189, "step": 1055 }, { "epoch": 0.48601558917927556, "grad_norm": 2.58024430648069, "learning_rate": 6.11148958812488e-06, "loss": 0.5071, "step": 1060 }, { "epoch": 0.48830811554332876, "grad_norm": 2.4501378891077987, "learning_rate": 6.072425507257528e-06, "loss": 0.5033, "step": 1065 }, { "epoch": 0.49060064190738195, "grad_norm": 2.783006969507733, "learning_rate": 6.033292686859414e-06, "loss": 0.4955, "step": 1070 }, { "epoch": 0.49289316827143514, "grad_norm": 2.428894458608491, "learning_rate": 5.99409363523689e-06, "loss": 0.4973, "step": 1075 }, { "epoch": 0.4951856946354883, "grad_norm": 2.7389561374869342, "learning_rate": 5.9548308649415486e-06, "loss": 0.5051, "step": 1080 }, { "epoch": 0.4974782209995415, "grad_norm": 2.5456232835838124, "learning_rate": 5.91550689260917e-06, "loss": 0.4935, "step": 1085 }, { "epoch": 0.49977074736359467, "grad_norm": 2.6057045786417685, "learning_rate": 5.876124238798424e-06, "loss": 0.501, "step": 1090 }, { "epoch": 0.5020632737276479, "grad_norm": 2.4695060680872873, "learning_rate": 5.836685427829296e-06, "loss": 0.5032, "step": 1095 }, { "epoch": 0.504355800091701, "grad_norm": 2.3783397469941376, "learning_rate": 5.797192987621293e-06, "loss": 0.4985, "step": 1100 }, { "epoch": 0.5066483264557542, "grad_norm": 2.491153548859691, "learning_rate": 5.7576494495314105e-06, "loss": 0.5043, "step": 1105 }, { "epoch": 0.5089408528198074, "grad_norm": 2.6062141152111673, "learning_rate": 5.718057348191874e-06, "loss": 0.4868, "step": 1110 }, { "epoch": 0.5112333791838606, "grad_norm": 2.5012205713207405, "learning_rate": 5.678419221347687e-06, "loss": 0.4979, "step": 1115 }, { "epoch": 0.5135259055479138, "grad_norm": 2.609877005241944, "learning_rate": 5.638737609693953e-06, "loss": 0.495, "step": 1120 }, { "epoch": 0.515818431911967, "grad_norm": 2.684672446431491, "learning_rate": 5.599015056713037e-06, "loss": 0.4823, "step": 1125 }, { "epoch": 0.5181109582760202, "grad_norm": 2.4771534112729228, "learning_rate": 5.559254108511531e-06, "loss": 0.5016, "step": 1130 }, { "epoch": 0.5204034846400734, "grad_norm": 2.46810743209868, "learning_rate": 5.519457313657056e-06, "loss": 0.4896, "step": 1135 }, { "epoch": 0.5226960110041265, "grad_norm": 2.5795208204825983, "learning_rate": 5.479627223014902e-06, "loss": 0.4886, "step": 1140 }, { "epoch": 0.5249885373681797, "grad_norm": 2.434086073989824, "learning_rate": 5.439766389584527e-06, "loss": 0.4865, "step": 1145 }, { "epoch": 0.5272810637322329, "grad_norm": 2.4538097489169934, "learning_rate": 5.399877368335922e-06, "loss": 0.4914, "step": 1150 }, { "epoch": 0.5295735900962861, "grad_norm": 2.5415775013932063, "learning_rate": 5.359962716045836e-06, "loss": 0.4936, "step": 1155 }, { "epoch": 0.5318661164603393, "grad_norm": 2.56697946552087, "learning_rate": 5.3200249911338986e-06, "loss": 0.4894, "step": 1160 }, { "epoch": 0.5341586428243925, "grad_norm": 2.572922499741503, "learning_rate": 5.280066753498632e-06, "loss": 0.4794, "step": 1165 }, { "epoch": 0.5364511691884457, "grad_norm": 2.623599926005301, "learning_rate": 5.240090564353365e-06, "loss": 0.4959, "step": 1170 }, { "epoch": 0.5387436955524989, "grad_norm": 2.4231120561633324, "learning_rate": 5.200098986062072e-06, "loss": 0.4753, "step": 1175 }, { "epoch": 0.5410362219165521, "grad_norm": 2.5196186316057108, "learning_rate": 5.160094581975127e-06, "loss": 0.4783, "step": 1180 }, { "epoch": 0.5433287482806052, "grad_norm": 2.527690400984075, "learning_rate": 5.1200799162650035e-06, "loss": 0.4916, "step": 1185 }, { "epoch": 0.5456212746446584, "grad_norm": 2.6015322908629415, "learning_rate": 5.080057553761917e-06, "loss": 0.4738, "step": 1190 }, { "epoch": 0.5479138010087116, "grad_norm": 2.3467602506879786, "learning_rate": 5.040030059789426e-06, "loss": 0.476, "step": 1195 }, { "epoch": 0.5502063273727648, "grad_norm": 2.570425940808593, "learning_rate": 5e-06, "loss": 0.4903, "step": 1200 }, { "epoch": 0.552498853736818, "grad_norm": 2.5543989632263284, "learning_rate": 4.9599699402105755e-06, "loss": 0.4673, "step": 1205 }, { "epoch": 0.5547913801008711, "grad_norm": 2.5213973685823277, "learning_rate": 4.919942446238085e-06, "loss": 0.4693, "step": 1210 }, { "epoch": 0.5570839064649243, "grad_norm": 2.4952425404718075, "learning_rate": 4.879920083734997e-06, "loss": 0.4692, "step": 1215 }, { "epoch": 0.5593764328289775, "grad_norm": 2.5419193115674776, "learning_rate": 4.839905418024875e-06, "loss": 0.4814, "step": 1220 }, { "epoch": 0.5616689591930307, "grad_norm": 2.558303192571574, "learning_rate": 4.7999010139379295e-06, "loss": 0.4698, "step": 1225 }, { "epoch": 0.5639614855570839, "grad_norm": 2.4678859101946315, "learning_rate": 4.759909435646636e-06, "loss": 0.4896, "step": 1230 }, { "epoch": 0.5662540119211371, "grad_norm": 2.6716519633665783, "learning_rate": 4.719933246501369e-06, "loss": 0.4852, "step": 1235 }, { "epoch": 0.5685465382851903, "grad_norm": 2.4330925797194807, "learning_rate": 4.679975008866103e-06, "loss": 0.4554, "step": 1240 }, { "epoch": 0.5708390646492435, "grad_norm": 2.437937005459216, "learning_rate": 4.640037283954165e-06, "loss": 0.4598, "step": 1245 }, { "epoch": 0.5731315910132967, "grad_norm": 2.413361545021729, "learning_rate": 4.6001226316640804e-06, "loss": 0.4739, "step": 1250 }, { "epoch": 0.5754241173773499, "grad_norm": 2.3552453394422503, "learning_rate": 4.5602336104154745e-06, "loss": 0.4646, "step": 1255 }, { "epoch": 0.5777166437414031, "grad_norm": 2.623470049632146, "learning_rate": 4.520372776985101e-06, "loss": 0.4579, "step": 1260 }, { "epoch": 0.5800091701054562, "grad_norm": 2.4219278336672874, "learning_rate": 4.480542686342946e-06, "loss": 0.4613, "step": 1265 }, { "epoch": 0.5823016964695094, "grad_norm": 2.517369439139374, "learning_rate": 4.440745891488471e-06, "loss": 0.4523, "step": 1270 }, { "epoch": 0.5845942228335625, "grad_norm": 2.501700820037027, "learning_rate": 4.400984943286965e-06, "loss": 0.4671, "step": 1275 }, { "epoch": 0.5868867491976157, "grad_norm": 2.4011689731614605, "learning_rate": 4.361262390306049e-06, "loss": 0.4527, "step": 1280 }, { "epoch": 0.5891792755616689, "grad_norm": 2.5994696717863706, "learning_rate": 4.321580778652316e-06, "loss": 0.4493, "step": 1285 }, { "epoch": 0.5914718019257221, "grad_norm": 2.491956972995198, "learning_rate": 4.2819426518081265e-06, "loss": 0.456, "step": 1290 }, { "epoch": 0.5937643282897753, "grad_norm": 2.4353572047335996, "learning_rate": 4.2423505504685894e-06, "loss": 0.4611, "step": 1295 }, { "epoch": 0.5960568546538285, "grad_norm": 2.4904358458702944, "learning_rate": 4.202807012378707e-06, "loss": 0.4546, "step": 1300 }, { "epoch": 0.5983493810178817, "grad_norm": 2.4617619082762636, "learning_rate": 4.163314572170704e-06, "loss": 0.458, "step": 1305 }, { "epoch": 0.6006419073819349, "grad_norm": 2.354023280982333, "learning_rate": 4.123875761201576e-06, "loss": 0.4433, "step": 1310 }, { "epoch": 0.6029344337459881, "grad_norm": 2.540723557518342, "learning_rate": 4.08449310739083e-06, "loss": 0.4484, "step": 1315 }, { "epoch": 0.6052269601100413, "grad_norm": 2.4043887566981446, "learning_rate": 4.045169135058452e-06, "loss": 0.4416, "step": 1320 }, { "epoch": 0.6075194864740945, "grad_norm": 2.481355244310724, "learning_rate": 4.0059063647631105e-06, "loss": 0.4645, "step": 1325 }, { "epoch": 0.6098120128381477, "grad_norm": 2.499493147862873, "learning_rate": 3.966707313140587e-06, "loss": 0.4542, "step": 1330 }, { "epoch": 0.6121045392022009, "grad_norm": 2.5034183191594477, "learning_rate": 3.927574492742473e-06, "loss": 0.4465, "step": 1335 }, { "epoch": 0.6143970655662541, "grad_norm": 2.450159706952634, "learning_rate": 3.888510411875121e-06, "loss": 0.4451, "step": 1340 }, { "epoch": 0.6166895919303071, "grad_norm": 2.437273107870038, "learning_rate": 3.849517574438866e-06, "loss": 0.4393, "step": 1345 }, { "epoch": 0.6189821182943603, "grad_norm": 2.4867270897195164, "learning_rate": 3.8105984797675364e-06, "loss": 0.4369, "step": 1350 }, { "epoch": 0.6212746446584135, "grad_norm": 2.4474532182002156, "learning_rate": 3.771755622468254e-06, "loss": 0.4459, "step": 1355 }, { "epoch": 0.6235671710224667, "grad_norm": 2.3883568752400737, "learning_rate": 3.7329914922615283e-06, "loss": 0.4414, "step": 1360 }, { "epoch": 0.6258596973865199, "grad_norm": 2.323604786191338, "learning_rate": 3.6943085738216855e-06, "loss": 0.4294, "step": 1365 }, { "epoch": 0.6281522237505731, "grad_norm": 2.5364327673030553, "learning_rate": 3.655709346617593e-06, "loss": 0.4482, "step": 1370 }, { "epoch": 0.6304447501146263, "grad_norm": 2.528211312039227, "learning_rate": 3.6171962847537466e-06, "loss": 0.4483, "step": 1375 }, { "epoch": 0.6327372764786795, "grad_norm": 2.4014535334880533, "learning_rate": 3.5787718568116764e-06, "loss": 0.4479, "step": 1380 }, { "epoch": 0.6350298028427327, "grad_norm": 2.6961239350559687, "learning_rate": 3.540438525691723e-06, "loss": 0.4375, "step": 1385 }, { "epoch": 0.6373223292067859, "grad_norm": 2.4568407427026027, "learning_rate": 3.502198748455169e-06, "loss": 0.4461, "step": 1390 }, { "epoch": 0.6396148555708391, "grad_norm": 2.444432290321262, "learning_rate": 3.464054976166753e-06, "loss": 0.4409, "step": 1395 }, { "epoch": 0.6419073819348923, "grad_norm": 2.3930367223498927, "learning_rate": 3.4260096537375553e-06, "loss": 0.433, "step": 1400 }, { "epoch": 0.6441999082989455, "grad_norm": 2.431394532574176, "learning_rate": 3.3880652197682974e-06, "loss": 0.4229, "step": 1405 }, { "epoch": 0.6464924346629987, "grad_norm": 2.434581693659057, "learning_rate": 3.3502241063930196e-06, "loss": 0.4389, "step": 1410 }, { "epoch": 0.6487849610270519, "grad_norm": 2.3993499417107156, "learning_rate": 3.3124887391232026e-06, "loss": 0.4219, "step": 1415 }, { "epoch": 0.651077487391105, "grad_norm": 2.476740652860741, "learning_rate": 3.2748615366922864e-06, "loss": 0.427, "step": 1420 }, { "epoch": 0.6533700137551581, "grad_norm": 2.507048548706466, "learning_rate": 3.2373449109006476e-06, "loss": 0.4341, "step": 1425 }, { "epoch": 0.6556625401192113, "grad_norm": 2.418497030941838, "learning_rate": 3.1999412664609986e-06, "loss": 0.4329, "step": 1430 }, { "epoch": 0.6579550664832645, "grad_norm": 2.4312888314629144, "learning_rate": 3.162653000844259e-06, "loss": 0.4227, "step": 1435 }, { "epoch": 0.6602475928473177, "grad_norm": 2.353877004261892, "learning_rate": 3.1254825041258852e-06, "loss": 0.4302, "step": 1440 }, { "epoch": 0.6625401192113709, "grad_norm": 2.381814531488306, "learning_rate": 3.0884321588326668e-06, "loss": 0.4376, "step": 1445 }, { "epoch": 0.6648326455754241, "grad_norm": 2.4501307973874287, "learning_rate": 3.051504339790019e-06, "loss": 0.4254, "step": 1450 }, { "epoch": 0.6671251719394773, "grad_norm": 2.459251255110059, "learning_rate": 3.0147014139697596e-06, "loss": 0.4263, "step": 1455 }, { "epoch": 0.6694176983035305, "grad_norm": 2.5254030222294466, "learning_rate": 2.978025740338396e-06, "loss": 0.4195, "step": 1460 }, { "epoch": 0.6717102246675837, "grad_norm": 2.2951603398964235, "learning_rate": 2.9414796697059155e-06, "loss": 0.4129, "step": 1465 }, { "epoch": 0.6740027510316369, "grad_norm": 2.364236291272217, "learning_rate": 2.905065544575114e-06, "loss": 0.4197, "step": 1470 }, { "epoch": 0.6762952773956901, "grad_norm": 2.4601102682369205, "learning_rate": 2.8687856989914393e-06, "loss": 0.4234, "step": 1475 }, { "epoch": 0.6785878037597433, "grad_norm": 2.686432591416178, "learning_rate": 2.8326424583933878e-06, "loss": 0.4223, "step": 1480 }, { "epoch": 0.6808803301237965, "grad_norm": 2.3448228852350788, "learning_rate": 2.796638139463456e-06, "loss": 0.4149, "step": 1485 }, { "epoch": 0.6831728564878496, "grad_norm": 2.317745266155718, "learning_rate": 2.7607750499796426e-06, "loss": 0.4161, "step": 1490 }, { "epoch": 0.6854653828519028, "grad_norm": 2.3719922106424725, "learning_rate": 2.725055488667522e-06, "loss": 0.4275, "step": 1495 }, { "epoch": 0.687757909215956, "grad_norm": 2.4553896347366746, "learning_rate": 2.689481745052908e-06, "loss": 0.3954, "step": 1500 }, { "epoch": 0.6900504355800092, "grad_norm": 2.471280707724599, "learning_rate": 2.6540560993151045e-06, "loss": 0.408, "step": 1505 }, { "epoch": 0.6923429619440623, "grad_norm": 2.375550619652342, "learning_rate": 2.6187808221407433e-06, "loss": 0.4091, "step": 1510 }, { "epoch": 0.6946354883081155, "grad_norm": 2.3794291144670865, "learning_rate": 2.5836581745782474e-06, "loss": 0.4203, "step": 1515 }, { "epoch": 0.6969280146721687, "grad_norm": 2.3959254909604133, "learning_rate": 2.5486904078928954e-06, "loss": 0.4019, "step": 1520 }, { "epoch": 0.6992205410362219, "grad_norm": 2.4572132670378593, "learning_rate": 2.5138797634225358e-06, "loss": 0.4025, "step": 1525 }, { "epoch": 0.7015130674002751, "grad_norm": 2.567664513817577, "learning_rate": 2.4792284724339077e-06, "loss": 0.4096, "step": 1530 }, { "epoch": 0.7038055937643283, "grad_norm": 2.473854002398598, "learning_rate": 2.4447387559796306e-06, "loss": 0.4129, "step": 1535 }, { "epoch": 0.7060981201283815, "grad_norm": 2.2347261984430844, "learning_rate": 2.410412824755839e-06, "loss": 0.4147, "step": 1540 }, { "epoch": 0.7083906464924347, "grad_norm": 2.45007211279529, "learning_rate": 2.3762528789604887e-06, "loss": 0.4159, "step": 1545 }, { "epoch": 0.7106831728564879, "grad_norm": 2.57319881552059, "learning_rate": 2.3422611081523215e-06, "loss": 0.4044, "step": 1550 }, { "epoch": 0.712975699220541, "grad_norm": 2.40694698697041, "learning_rate": 2.3084396911105233e-06, "loss": 0.3888, "step": 1555 }, { "epoch": 0.7152682255845942, "grad_norm": 2.6193951641238837, "learning_rate": 2.274790795695071e-06, "loss": 0.4186, "step": 1560 }, { "epoch": 0.7175607519486474, "grad_norm": 2.3915420788033686, "learning_rate": 2.2413165787077844e-06, "loss": 0.4105, "step": 1565 }, { "epoch": 0.7198532783127006, "grad_norm": 2.4922945082662706, "learning_rate": 2.20801918575407e-06, "loss": 0.41, "step": 1570 }, { "epoch": 0.7221458046767538, "grad_norm": 2.361018492853961, "learning_rate": 2.1749007511054005e-06, "loss": 0.4075, "step": 1575 }, { "epoch": 0.724438331040807, "grad_norm": 2.453915782459234, "learning_rate": 2.1419633975625113e-06, "loss": 0.4123, "step": 1580 }, { "epoch": 0.7267308574048602, "grad_norm": 2.2558599145275458, "learning_rate": 2.109209236319342e-06, "loss": 0.3971, "step": 1585 }, { "epoch": 0.7290233837689133, "grad_norm": 2.3942282865574103, "learning_rate": 2.076640366827703e-06, "loss": 0.4012, "step": 1590 }, { "epoch": 0.7313159101329665, "grad_norm": 2.4100293351001714, "learning_rate": 2.04425887666271e-06, "loss": 0.3926, "step": 1595 }, { "epoch": 0.7336084364970197, "grad_norm": 2.5693096989442927, "learning_rate": 2.0120668413889877e-06, "loss": 0.4021, "step": 1600 }, { "epoch": 0.7359009628610729, "grad_norm": 2.513834629858347, "learning_rate": 1.980066324427613e-06, "loss": 0.3926, "step": 1605 }, { "epoch": 0.7381934892251261, "grad_norm": 2.500502153829468, "learning_rate": 1.9482593769238695e-06, "loss": 0.3932, "step": 1610 }, { "epoch": 0.7404860155891793, "grad_norm": 2.2943690678553827, "learning_rate": 1.916648037615767e-06, "loss": 0.3961, "step": 1615 }, { "epoch": 0.7427785419532325, "grad_norm": 2.4947450845729904, "learning_rate": 1.8852343327033717e-06, "loss": 0.3918, "step": 1620 }, { "epoch": 0.7450710683172856, "grad_norm": 2.475640064869192, "learning_rate": 1.854020275718924e-06, "loss": 0.3953, "step": 1625 }, { "epoch": 0.7473635946813388, "grad_norm": 2.380898479266151, "learning_rate": 1.8230078673977802e-06, "loss": 0.3767, "step": 1630 }, { "epoch": 0.749656121045392, "grad_norm": 2.3124836007659444, "learning_rate": 1.7921990955501705e-06, "loss": 0.386, "step": 1635 }, { "epoch": 0.7519486474094452, "grad_norm": 2.3942291132445375, "learning_rate": 1.7615959349337914e-06, "loss": 0.3964, "step": 1640 }, { "epoch": 0.7542411737734984, "grad_norm": 2.4125792225674614, "learning_rate": 1.731200347127217e-06, "loss": 0.3918, "step": 1645 }, { "epoch": 0.7565337001375516, "grad_norm": 2.4570540617910788, "learning_rate": 1.7010142804041785e-06, "loss": 0.4012, "step": 1650 }, { "epoch": 0.7588262265016048, "grad_norm": 2.3060832536528006, "learning_rate": 1.6710396696086768e-06, "loss": 0.4026, "step": 1655 }, { "epoch": 0.761118752865658, "grad_norm": 2.357410070095031, "learning_rate": 1.6412784360309753e-06, "loss": 0.3876, "step": 1660 }, { "epoch": 0.7634112792297112, "grad_norm": 2.5569987658890434, "learning_rate": 1.611732487284437e-06, "loss": 0.3875, "step": 1665 }, { "epoch": 0.7657038055937643, "grad_norm": 2.5367416684876805, "learning_rate": 1.5824037171832595e-06, "loss": 0.3923, "step": 1670 }, { "epoch": 0.7679963319578175, "grad_norm": 2.370553404803813, "learning_rate": 1.5532940056210882e-06, "loss": 0.3916, "step": 1675 }, { "epoch": 0.7702888583218707, "grad_norm": 2.445473374507484, "learning_rate": 1.524405218450517e-06, "loss": 0.4005, "step": 1680 }, { "epoch": 0.7725813846859239, "grad_norm": 2.416383451707918, "learning_rate": 1.4957392073634912e-06, "loss": 0.385, "step": 1685 }, { "epoch": 0.774873911049977, "grad_norm": 2.4307180782279976, "learning_rate": 1.4672978097726204e-06, "loss": 0.3857, "step": 1690 }, { "epoch": 0.7771664374140302, "grad_norm": 2.4572760495599795, "learning_rate": 1.439082848693406e-06, "loss": 0.3916, "step": 1695 }, { "epoch": 0.7794589637780834, "grad_norm": 2.408412846059606, "learning_rate": 1.4110961326273936e-06, "loss": 0.3908, "step": 1700 }, { "epoch": 0.7817514901421366, "grad_norm": 2.6601098763821596, "learning_rate": 1.3833394554462477e-06, "loss": 0.3859, "step": 1705 }, { "epoch": 0.7840440165061898, "grad_norm": 2.520675032421566, "learning_rate": 1.35581459627677e-06, "loss": 0.3936, "step": 1710 }, { "epoch": 0.786336542870243, "grad_norm": 2.257467358094596, "learning_rate": 1.3285233193868663e-06, "loss": 0.3799, "step": 1715 }, { "epoch": 0.7886290692342962, "grad_norm": 2.327829634660073, "learning_rate": 1.3014673740724615e-06, "loss": 0.3876, "step": 1720 }, { "epoch": 0.7909215955983494, "grad_norm": 2.366347981314184, "learning_rate": 1.2746484945453691e-06, "loss": 0.3829, "step": 1725 }, { "epoch": 0.7932141219624026, "grad_norm": 2.391058577508851, "learning_rate": 1.2480683998221365e-06, "loss": 0.3825, "step": 1730 }, { "epoch": 0.7955066483264558, "grad_norm": 2.470899547865623, "learning_rate": 1.221728793613865e-06, "loss": 0.3895, "step": 1735 }, { "epoch": 0.797799174690509, "grad_norm": 2.399551521415764, "learning_rate": 1.1956313642169974e-06, "loss": 0.3846, "step": 1740 }, { "epoch": 0.8000917010545622, "grad_norm": 2.463312952219633, "learning_rate": 1.1697777844051105e-06, "loss": 0.3788, "step": 1745 }, { "epoch": 0.8023842274186153, "grad_norm": 2.4348320894873092, "learning_rate": 1.1441697113216893e-06, "loss": 0.3803, "step": 1750 }, { "epoch": 0.8046767537826685, "grad_norm": 2.385545108416876, "learning_rate": 1.1188087863739173e-06, "loss": 0.3859, "step": 1755 }, { "epoch": 0.8069692801467216, "grad_norm": 2.4484362721344195, "learning_rate": 1.0936966351274554e-06, "loss": 0.3739, "step": 1760 }, { "epoch": 0.8092618065107748, "grad_norm": 2.4361451039130317, "learning_rate": 1.0688348672022547e-06, "loss": 0.4012, "step": 1765 }, { "epoch": 0.811554332874828, "grad_norm": 2.5671935693516192, "learning_rate": 1.0442250761693829e-06, "loss": 0.3717, "step": 1770 }, { "epoch": 0.8138468592388812, "grad_norm": 2.3910678127476475, "learning_rate": 1.0198688394488837e-06, "loss": 0.3824, "step": 1775 }, { "epoch": 0.8161393856029344, "grad_norm": 2.4337476998865237, "learning_rate": 9.957677182086611e-07, "loss": 0.3754, "step": 1780 }, { "epoch": 0.8184319119669876, "grad_norm": 2.3930303860053055, "learning_rate": 9.719232572644189e-07, "loss": 0.3814, "step": 1785 }, { "epoch": 0.8207244383310408, "grad_norm": 2.4070725664187194, "learning_rate": 9.483369849806401e-07, "loss": 0.3681, "step": 1790 }, { "epoch": 0.823016964695094, "grad_norm": 2.4234654890940277, "learning_rate": 9.250104131726256e-07, "loss": 0.3748, "step": 1795 }, { "epoch": 0.8253094910591472, "grad_norm": 2.4405075201633486, "learning_rate": 9.019450370095867e-07, "loss": 0.3852, "step": 1800 }, { "epoch": 0.8276020174232004, "grad_norm": 2.4157009817816535, "learning_rate": 8.791423349188111e-07, "loss": 0.3738, "step": 1805 }, { "epoch": 0.8298945437872536, "grad_norm": 2.3817117068747695, "learning_rate": 8.566037684908985e-07, "loss": 0.3774, "step": 1810 }, { "epoch": 0.8321870701513068, "grad_norm": 2.643862606121901, "learning_rate": 8.343307823860819e-07, "loss": 0.3747, "step": 1815 }, { "epoch": 0.83447959651536, "grad_norm": 2.415451666660326, "learning_rate": 8.123248042416209e-07, "loss": 0.3807, "step": 1820 }, { "epoch": 0.8367721228794132, "grad_norm": 2.367699275816763, "learning_rate": 7.905872445802976e-07, "loss": 0.3819, "step": 1825 }, { "epoch": 0.8390646492434664, "grad_norm": 2.401428866906129, "learning_rate": 7.691194967200099e-07, "loss": 0.3773, "step": 1830 }, { "epoch": 0.8413571756075194, "grad_norm": 2.3851132017870444, "learning_rate": 7.47922936684457e-07, "loss": 0.3848, "step": 1835 }, { "epoch": 0.8436497019715726, "grad_norm": 2.334920050986847, "learning_rate": 7.269989231149432e-07, "loss": 0.3646, "step": 1840 }, { "epoch": 0.8459422283356258, "grad_norm": 2.302533584527464, "learning_rate": 7.063487971832922e-07, "loss": 0.3719, "step": 1845 }, { "epoch": 0.848234754699679, "grad_norm": 2.4631469089449443, "learning_rate": 6.85973882505886e-07, "loss": 0.3951, "step": 1850 }, { "epoch": 0.8505272810637322, "grad_norm": 2.4860937019904426, "learning_rate": 6.658754850588161e-07, "loss": 0.3877, "step": 1855 }, { "epoch": 0.8528198074277854, "grad_norm": 2.366824744001058, "learning_rate": 6.460548930941801e-07, "loss": 0.3711, "step": 1860 }, { "epoch": 0.8551123337918386, "grad_norm": 2.587488334709295, "learning_rate": 6.265133770575066e-07, "loss": 0.366, "step": 1865 }, { "epoch": 0.8574048601558918, "grad_norm": 2.4606917803825072, "learning_rate": 6.072521895063255e-07, "loss": 0.3818, "step": 1870 }, { "epoch": 0.859697386519945, "grad_norm": 2.4967563072720576, "learning_rate": 5.882725650298787e-07, "loss": 0.3804, "step": 1875 }, { "epoch": 0.8619899128839982, "grad_norm": 2.4902108475668214, "learning_rate": 5.695757201699875e-07, "loss": 0.3751, "step": 1880 }, { "epoch": 0.8642824392480514, "grad_norm": 2.3545990508632713, "learning_rate": 5.511628533430769e-07, "loss": 0.3887, "step": 1885 }, { "epoch": 0.8665749656121046, "grad_norm": 2.4583864322248363, "learning_rate": 5.330351447633603e-07, "loss": 0.3846, "step": 1890 }, { "epoch": 0.8688674919761578, "grad_norm": 2.558178264129578, "learning_rate": 5.151937563671889e-07, "loss": 0.3761, "step": 1895 }, { "epoch": 0.871160018340211, "grad_norm": 2.4125538046249133, "learning_rate": 4.976398317385767e-07, "loss": 0.3789, "step": 1900 }, { "epoch": 0.8734525447042641, "grad_norm": 2.5261586438137718, "learning_rate": 4.803744960358992e-07, "loss": 0.3692, "step": 1905 }, { "epoch": 0.8757450710683173, "grad_norm": 2.5343814063203913, "learning_rate": 4.633988559197761e-07, "loss": 0.3741, "step": 1910 }, { "epoch": 0.8780375974323704, "grad_norm": 2.5455270767430305, "learning_rate": 4.4671399948213233e-07, "loss": 0.3742, "step": 1915 }, { "epoch": 0.8803301237964236, "grad_norm": 2.4299267640638442, "learning_rate": 4.3032099617645874e-07, "loss": 0.3793, "step": 1920 }, { "epoch": 0.8826226501604768, "grad_norm": 2.5350282869807215, "learning_rate": 4.1422089674926113e-07, "loss": 0.3708, "step": 1925 }, { "epoch": 0.88491517652453, "grad_norm": 2.4052098639642745, "learning_rate": 3.984147331727128e-07, "loss": 0.3815, "step": 1930 }, { "epoch": 0.8872077028885832, "grad_norm": 2.440029806154777, "learning_rate": 3.829035185785035e-07, "loss": 0.3559, "step": 1935 }, { "epoch": 0.8895002292526364, "grad_norm": 2.4757422584836783, "learning_rate": 3.676882471929044e-07, "loss": 0.3724, "step": 1940 }, { "epoch": 0.8917927556166896, "grad_norm": 2.405181037542438, "learning_rate": 3.527698942730384e-07, "loss": 0.3678, "step": 1945 }, { "epoch": 0.8940852819807428, "grad_norm": 2.477077740628022, "learning_rate": 3.3814941604437155e-07, "loss": 0.3696, "step": 1950 }, { "epoch": 0.896377808344796, "grad_norm": 2.594645970360135, "learning_rate": 3.2382774963941823e-07, "loss": 0.3689, "step": 1955 }, { "epoch": 0.8986703347088492, "grad_norm": 2.4996924524050526, "learning_rate": 3.0980581303767576e-07, "loss": 0.3641, "step": 1960 }, { "epoch": 0.9009628610729024, "grad_norm": 2.5351766412057364, "learning_rate": 2.9608450500678566e-07, "loss": 0.3736, "step": 1965 }, { "epoch": 0.9032553874369555, "grad_norm": 2.4812985119515374, "learning_rate": 2.826647050449216e-07, "loss": 0.3652, "step": 1970 }, { "epoch": 0.9055479138010087, "grad_norm": 2.4498300583099506, "learning_rate": 2.69547273324417e-07, "loss": 0.3653, "step": 1975 }, { "epoch": 0.9078404401650619, "grad_norm": 2.546961383266402, "learning_rate": 2.5673305063663335e-07, "loss": 0.3723, "step": 1980 }, { "epoch": 0.9101329665291151, "grad_norm": 2.34777660611532, "learning_rate": 2.442228583380646e-07, "loss": 0.3596, "step": 1985 }, { "epoch": 0.9124254928931683, "grad_norm": 2.410870301241545, "learning_rate": 2.3201749829769083e-07, "loss": 0.3783, "step": 1990 }, { "epoch": 0.9147180192572214, "grad_norm": 2.519326020963244, "learning_rate": 2.201177528455828e-07, "loss": 0.3739, "step": 1995 }, { "epoch": 0.9170105456212746, "grad_norm": 2.4872058403028574, "learning_rate": 2.085243847227525e-07, "loss": 0.3768, "step": 2000 }, { "epoch": 0.9193030719853278, "grad_norm": 2.4175176965392544, "learning_rate": 1.9723813703227013e-07, "loss": 0.3794, "step": 2005 }, { "epoch": 0.921595598349381, "grad_norm": 2.514035461894725, "learning_rate": 1.8625973319162605e-07, "loss": 0.3656, "step": 2010 }, { "epoch": 0.9238881247134342, "grad_norm": 2.4532676789082166, "learning_rate": 1.7558987688636675e-07, "loss": 0.361, "step": 2015 }, { "epoch": 0.9261806510774874, "grad_norm": 2.580005311393483, "learning_rate": 1.652292520249865e-07, "loss": 0.369, "step": 2020 }, { "epoch": 0.9284731774415406, "grad_norm": 2.359368965829793, "learning_rate": 1.5517852269509692e-07, "loss": 0.3571, "step": 2025 }, { "epoch": 0.9307657038055938, "grad_norm": 2.4993672807867178, "learning_rate": 1.4543833312085365e-07, "loss": 0.3588, "step": 2030 }, { "epoch": 0.933058230169647, "grad_norm": 2.41149322411576, "learning_rate": 1.360093076216673e-07, "loss": 0.3705, "step": 2035 }, { "epoch": 0.9353507565337001, "grad_norm": 2.474736948512413, "learning_rate": 1.2689205057218602e-07, "loss": 0.361, "step": 2040 }, { "epoch": 0.9376432828977533, "grad_norm": 2.3336360044904736, "learning_rate": 1.1808714636355634e-07, "loss": 0.3568, "step": 2045 }, { "epoch": 0.9399358092618065, "grad_norm": 2.566200023951429, "learning_rate": 1.0959515936596387e-07, "loss": 0.3783, "step": 2050 }, { "epoch": 0.9422283356258597, "grad_norm": 2.5160190954507264, "learning_rate": 1.014166338924627e-07, "loss": 0.372, "step": 2055 }, { "epoch": 0.9445208619899129, "grad_norm": 2.509256348018165, "learning_rate": 9.355209416408051e-08, "loss": 0.3853, "step": 2060 }, { "epoch": 0.9468133883539661, "grad_norm": 2.5224442995349152, "learning_rate": 8.600204427622438e-08, "loss": 0.365, "step": 2065 }, { "epoch": 0.9491059147180193, "grad_norm": 2.4001792608745602, "learning_rate": 7.876696816636276e-08, "loss": 0.3736, "step": 2070 }, { "epoch": 0.9513984410820725, "grad_norm": 2.4422332203602553, "learning_rate": 7.184732958301078e-08, "loss": 0.3651, "step": 2075 }, { "epoch": 0.9536909674461256, "grad_norm": 2.471890892444078, "learning_rate": 6.524357205600518e-08, "loss": 0.3624, "step": 2080 }, { "epoch": 0.9559834938101788, "grad_norm": 2.523417346804641, "learning_rate": 5.895611886807317e-08, "loss": 0.369, "step": 2085 }, { "epoch": 0.958276020174232, "grad_norm": 2.4584360575665776, "learning_rate": 5.2985373027702455e-08, "loss": 0.363, "step": 2090 }, { "epoch": 0.9605685465382852, "grad_norm": 2.467603595232153, "learning_rate": 4.733171724330854e-08, "loss": 0.3814, "step": 2095 }, { "epoch": 0.9628610729023384, "grad_norm": 2.5238201533198072, "learning_rate": 4.19955138987066e-08, "loss": 0.369, "step": 2100 }, { "epoch": 0.9651535992663915, "grad_norm": 2.5600424647957807, "learning_rate": 3.697710502988006e-08, "loss": 0.3652, "step": 2105 }, { "epoch": 0.9674461256304447, "grad_norm": 2.475992842961113, "learning_rate": 3.2276812303060346e-08, "loss": 0.3741, "step": 2110 }, { "epoch": 0.9697386519944979, "grad_norm": 2.4735410644370606, "learning_rate": 2.7894936994106724e-08, "loss": 0.3571, "step": 2115 }, { "epoch": 0.9720311783585511, "grad_norm": 2.384962513457078, "learning_rate": 2.383175996919673e-08, "loss": 0.3654, "step": 2120 }, { "epoch": 0.9743237047226043, "grad_norm": 2.4369560907719414, "learning_rate": 2.008754166682225e-08, "loss": 0.3614, "step": 2125 }, { "epoch": 0.9766162310866575, "grad_norm": 2.334334624814976, "learning_rate": 1.6662522081097308e-08, "loss": 0.3598, "step": 2130 }, { "epoch": 0.9789087574507107, "grad_norm": 2.515966550970349, "learning_rate": 1.3556920746373714e-08, "loss": 0.3539, "step": 2135 }, { "epoch": 0.9812012838147639, "grad_norm": 2.4578356166282704, "learning_rate": 1.0770936723171199e-08, "loss": 0.3684, "step": 2140 }, { "epoch": 0.9834938101788171, "grad_norm": 2.534561019356648, "learning_rate": 8.304748585417077e-09, "loss": 0.3629, "step": 2145 }, { "epoch": 0.9857863365428703, "grad_norm": 2.4815228224834254, "learning_rate": 6.158514409000393e-09, "loss": 0.3617, "step": 2150 }, { "epoch": 0.9880788629069235, "grad_norm": 2.520302407708297, "learning_rate": 4.332371761638921e-09, "loss": 0.3716, "step": 2155 }, { "epoch": 0.9903713892709766, "grad_norm": 2.939805778253569, "learning_rate": 2.8264376940634332e-09, "loss": 0.3685, "step": 2160 }, { "epoch": 0.9926639156350298, "grad_norm": 2.6736093020039484, "learning_rate": 1.640808732513155e-09, "loss": 0.3724, "step": 2165 }, { "epoch": 0.994956441999083, "grad_norm": 2.3884833213363144, "learning_rate": 7.755608725490415e-10, "loss": 0.354, "step": 2170 }, { "epoch": 0.9972489683631361, "grad_norm": 2.378189457774983, "learning_rate": 2.307495741843413e-10, "loss": 0.356, "step": 2175 }, { "epoch": 0.9995414947271893, "grad_norm": 2.543796138070999, "learning_rate": 6.4097583263311725e-12, "loss": 0.3664, "step": 2180 }, { "epoch": 1.0, "eval_runtime": 2.6844, "eval_samples_per_second": 3.725, "eval_steps_per_second": 1.118, "step": 2181 }, { "epoch": 1.0, "step": 2181, "total_flos": 228328514519040.0, "train_loss": 0.0, "train_runtime": 0.0089, "train_samples_per_second": 3927873.864, "train_steps_per_second": 245597.686 } ], "logging_steps": 5, "max_steps": 2181, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 228328514519040.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }