{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9995619798510732, "eval_steps": 500, "global_step": 1141, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0008760402978537013, "grad_norm": 247.023226046774, "learning_rate": 1.7391304347826088e-07, "loss": 8.8281, "step": 1 }, { "epoch": 0.004380201489268506, "grad_norm": 232.10560021267386, "learning_rate": 8.695652173913044e-07, "loss": 8.7637, "step": 5 }, { "epoch": 0.008760402978537012, "grad_norm": 197.45293542519394, "learning_rate": 1.7391304347826088e-06, "loss": 8.5789, "step": 10 }, { "epoch": 0.013140604467805518, "grad_norm": 69.72380158032517, "learning_rate": 2.6086956521739132e-06, "loss": 7.607, "step": 15 }, { "epoch": 0.017520805957074025, "grad_norm": 40.79010585762893, "learning_rate": 3.4782608695652175e-06, "loss": 6.5969, "step": 20 }, { "epoch": 0.021901007446342532, "grad_norm": 31.021399931633685, "learning_rate": 4.347826086956522e-06, "loss": 5.3516, "step": 25 }, { "epoch": 0.026281208935611037, "grad_norm": 15.829514000742273, "learning_rate": 5.2173913043478265e-06, "loss": 4.018, "step": 30 }, { "epoch": 0.030661410424879545, "grad_norm": 12.017912967930332, "learning_rate": 6.086956521739132e-06, "loss": 2.9813, "step": 35 }, { "epoch": 0.03504161191414805, "grad_norm": 2.8893117110423487, "learning_rate": 6.956521739130435e-06, "loss": 2.1303, "step": 40 }, { "epoch": 0.03942181340341656, "grad_norm": 1.4949499944110431, "learning_rate": 7.82608695652174e-06, "loss": 1.8031, "step": 45 }, { "epoch": 0.043802014892685065, "grad_norm": 0.7508052500953509, "learning_rate": 8.695652173913044e-06, "loss": 1.5938, "step": 50 }, { "epoch": 0.04818221638195357, "grad_norm": 0.4902022246315334, "learning_rate": 9.565217391304349e-06, "loss": 1.5084, "step": 55 }, { "epoch": 0.052562417871222074, "grad_norm": 0.42532974187985395, "learning_rate": 1.0434782608695653e-05, "loss": 1.4203, "step": 60 }, { "epoch": 0.05694261936049058, "grad_norm": 0.3463140884611265, "learning_rate": 1.1304347826086957e-05, "loss": 1.3941, "step": 65 }, { "epoch": 0.06132282084975909, "grad_norm": 0.3189213985873913, "learning_rate": 1.2173913043478263e-05, "loss": 1.3566, "step": 70 }, { "epoch": 0.0657030223390276, "grad_norm": 0.2996468490222568, "learning_rate": 1.3043478260869566e-05, "loss": 1.3418, "step": 75 }, { "epoch": 0.0700832238282961, "grad_norm": 0.26712973341018265, "learning_rate": 1.391304347826087e-05, "loss": 1.3201, "step": 80 }, { "epoch": 0.07446342531756461, "grad_norm": 0.2668394062417099, "learning_rate": 1.4782608695652174e-05, "loss": 1.3143, "step": 85 }, { "epoch": 0.07884362680683311, "grad_norm": 0.26274351296407206, "learning_rate": 1.565217391304348e-05, "loss": 1.3031, "step": 90 }, { "epoch": 0.08322382829610162, "grad_norm": 0.25741452098888334, "learning_rate": 1.6521739130434785e-05, "loss": 1.2848, "step": 95 }, { "epoch": 0.08760402978537013, "grad_norm": 0.2441325116229632, "learning_rate": 1.739130434782609e-05, "loss": 1.2732, "step": 100 }, { "epoch": 0.09198423127463863, "grad_norm": 0.24162625924392253, "learning_rate": 1.8260869565217393e-05, "loss": 1.2715, "step": 105 }, { "epoch": 0.09636443276390715, "grad_norm": 0.2519915389487053, "learning_rate": 1.9130434782608697e-05, "loss": 1.2711, "step": 110 }, { "epoch": 0.10074463425317565, "grad_norm": 0.2465793202977791, "learning_rate": 2e-05, "loss": 1.2459, "step": 115 }, { "epoch": 0.10512483574244415, "grad_norm": 0.23497052771629895, "learning_rate": 1.99988280568259e-05, "loss": 1.248, "step": 120 }, { "epoch": 0.10950503723171266, "grad_norm": 0.24553176964100193, "learning_rate": 1.9995312501993765e-05, "loss": 1.2238, "step": 125 }, { "epoch": 0.11388523872098116, "grad_norm": 0.23370960048474732, "learning_rate": 1.998945415950969e-05, "loss": 1.248, "step": 130 }, { "epoch": 0.11826544021024968, "grad_norm": 0.24397346786614016, "learning_rate": 1.9981254402502568e-05, "loss": 1.2063, "step": 135 }, { "epoch": 0.12264564169951818, "grad_norm": 0.27495355893049633, "learning_rate": 1.9970715152902257e-05, "loss": 1.2381, "step": 140 }, { "epoch": 0.12702584318878668, "grad_norm": 0.25282538712904246, "learning_rate": 1.9957838880989076e-05, "loss": 1.2281, "step": 145 }, { "epoch": 0.1314060446780552, "grad_norm": 0.22995731363863653, "learning_rate": 1.9942628604814827e-05, "loss": 1.2051, "step": 150 }, { "epoch": 0.1357862461673237, "grad_norm": 0.23386474611286825, "learning_rate": 1.9925087889495374e-05, "loss": 1.2141, "step": 155 }, { "epoch": 0.1401664476565922, "grad_norm": 0.2309649578398331, "learning_rate": 1.990522084637503e-05, "loss": 1.1822, "step": 160 }, { "epoch": 0.1445466491458607, "grad_norm": 0.24003805632416436, "learning_rate": 1.9883032132062926e-05, "loss": 1.2063, "step": 165 }, { "epoch": 0.14892685063512923, "grad_norm": 0.24338387371805312, "learning_rate": 1.98585269473415e-05, "loss": 1.1879, "step": 170 }, { "epoch": 0.1533070521243977, "grad_norm": 0.23476112374060998, "learning_rate": 1.9831711035947552e-05, "loss": 1.2031, "step": 175 }, { "epoch": 0.15768725361366623, "grad_norm": 0.22915392045241134, "learning_rate": 1.9802590683225945e-05, "loss": 1.1943, "step": 180 }, { "epoch": 0.16206745510293474, "grad_norm": 0.24585515208967154, "learning_rate": 1.97711727146564e-05, "loss": 1.1961, "step": 185 }, { "epoch": 0.16644765659220323, "grad_norm": 0.21471505934692195, "learning_rate": 1.973746449425368e-05, "loss": 1.1883, "step": 190 }, { "epoch": 0.17082785808147175, "grad_norm": 0.22942379356513737, "learning_rate": 1.970147392284154e-05, "loss": 1.2193, "step": 195 }, { "epoch": 0.17520805957074026, "grad_norm": 0.23244007409565076, "learning_rate": 1.9663209436200887e-05, "loss": 1.207, "step": 200 }, { "epoch": 0.17958826106000875, "grad_norm": 0.22422374010740104, "learning_rate": 1.9622680003092503e-05, "loss": 1.174, "step": 205 }, { "epoch": 0.18396846254927726, "grad_norm": 0.2239852274530115, "learning_rate": 1.957989512315489e-05, "loss": 1.1766, "step": 210 }, { "epoch": 0.18834866403854578, "grad_norm": 0.22417592332949882, "learning_rate": 1.953486482467764e-05, "loss": 1.2016, "step": 215 }, { "epoch": 0.1927288655278143, "grad_norm": 0.23718560975504316, "learning_rate": 1.9487599662250945e-05, "loss": 1.1836, "step": 220 }, { "epoch": 0.19710906701708278, "grad_norm": 0.24831579992264016, "learning_rate": 1.9438110714291697e-05, "loss": 1.1885, "step": 225 }, { "epoch": 0.2014892685063513, "grad_norm": 0.22705440610322433, "learning_rate": 1.9386409580446846e-05, "loss": 1.1768, "step": 230 }, { "epoch": 0.2058694699956198, "grad_norm": 0.22232082846727982, "learning_rate": 1.933250837887457e-05, "loss": 1.1695, "step": 235 }, { "epoch": 0.2102496714848883, "grad_norm": 0.20849807972480366, "learning_rate": 1.9276419743403934e-05, "loss": 1.1664, "step": 240 }, { "epoch": 0.2146298729741568, "grad_norm": 0.22884959889218245, "learning_rate": 1.9218156820573618e-05, "loss": 1.1715, "step": 245 }, { "epoch": 0.21901007446342532, "grad_norm": 0.22251029198025443, "learning_rate": 1.9157733266550577e-05, "loss": 1.1664, "step": 250 }, { "epoch": 0.2233902759526938, "grad_norm": 0.21607317255636946, "learning_rate": 1.9095163243929143e-05, "loss": 1.1758, "step": 255 }, { "epoch": 0.22777047744196233, "grad_norm": 0.2275376818247485, "learning_rate": 1.9030461418411498e-05, "loss": 1.173, "step": 260 }, { "epoch": 0.23215067893123084, "grad_norm": 0.2289840446661591, "learning_rate": 1.8963642955370203e-05, "loss": 1.1852, "step": 265 }, { "epoch": 0.23653088042049936, "grad_norm": 0.23768451520157288, "learning_rate": 1.889472351629358e-05, "loss": 1.1551, "step": 270 }, { "epoch": 0.24091108190976784, "grad_norm": 0.22459484609481822, "learning_rate": 1.882371925511488e-05, "loss": 1.1564, "step": 275 }, { "epoch": 0.24529128339903636, "grad_norm": 0.22828454066610995, "learning_rate": 1.875064681442594e-05, "loss": 1.1838, "step": 280 }, { "epoch": 0.24967148488830487, "grad_norm": 0.23814414474493634, "learning_rate": 1.867552332157637e-05, "loss": 1.165, "step": 285 }, { "epoch": 0.25405168637757336, "grad_norm": 0.23695582890788727, "learning_rate": 1.8598366384659113e-05, "loss": 1.1496, "step": 290 }, { "epoch": 0.25843188786684185, "grad_norm": 0.21285197719762172, "learning_rate": 1.851919408838327e-05, "loss": 1.1671, "step": 295 }, { "epoch": 0.2628120893561104, "grad_norm": 0.23402098820515907, "learning_rate": 1.843802498983529e-05, "loss": 1.1662, "step": 300 }, { "epoch": 0.2671922908453789, "grad_norm": 0.21951316193224032, "learning_rate": 1.8354878114129368e-05, "loss": 1.1623, "step": 305 }, { "epoch": 0.2715724923346474, "grad_norm": 0.21226630310736813, "learning_rate": 1.8269772949948185e-05, "loss": 1.1571, "step": 310 }, { "epoch": 0.2759526938239159, "grad_norm": 0.21300490616956957, "learning_rate": 1.8182729444974993e-05, "loss": 1.1729, "step": 315 }, { "epoch": 0.2803328953131844, "grad_norm": 0.2298100232081548, "learning_rate": 1.8093768001218096e-05, "loss": 1.1648, "step": 320 }, { "epoch": 0.28471309680245294, "grad_norm": 0.20939381292392953, "learning_rate": 1.800290947022884e-05, "loss": 1.1633, "step": 325 }, { "epoch": 0.2890932982917214, "grad_norm": 0.21652397528724918, "learning_rate": 1.7910175148214274e-05, "loss": 1.1703, "step": 330 }, { "epoch": 0.2934734997809899, "grad_norm": 0.23693799112996852, "learning_rate": 1.7815586771045535e-05, "loss": 1.1557, "step": 335 }, { "epoch": 0.29785370127025845, "grad_norm": 0.23875037210600733, "learning_rate": 1.771916650916321e-05, "loss": 1.1546, "step": 340 }, { "epoch": 0.30223390275952694, "grad_norm": 0.2086246080725189, "learning_rate": 1.762093696238086e-05, "loss": 1.1414, "step": 345 }, { "epoch": 0.3066141042487954, "grad_norm": 0.2032809446181252, "learning_rate": 1.752092115458784e-05, "loss": 1.1557, "step": 350 }, { "epoch": 0.31099430573806397, "grad_norm": 0.2299671919872682, "learning_rate": 1.7419142528352815e-05, "loss": 1.1584, "step": 355 }, { "epoch": 0.31537450722733246, "grad_norm": 0.2208475255216932, "learning_rate": 1.731562493942904e-05, "loss": 1.1436, "step": 360 }, { "epoch": 0.31975470871660094, "grad_norm": 0.21527409406555012, "learning_rate": 1.721039265116285e-05, "loss": 1.1592, "step": 365 }, { "epoch": 0.3241349102058695, "grad_norm": 0.2075937240152717, "learning_rate": 1.710347032880664e-05, "loss": 1.1332, "step": 370 }, { "epoch": 0.328515111695138, "grad_norm": 0.2155228439936621, "learning_rate": 1.6994883033737582e-05, "loss": 1.1395, "step": 375 }, { "epoch": 0.33289531318440646, "grad_norm": 0.24834287218683154, "learning_rate": 1.688465621758352e-05, "loss": 1.1566, "step": 380 }, { "epoch": 0.337275514673675, "grad_norm": 0.2401994560404927, "learning_rate": 1.6772815716257414e-05, "loss": 1.1582, "step": 385 }, { "epoch": 0.3416557161629435, "grad_norm": 0.22010411419127637, "learning_rate": 1.6659387743901688e-05, "loss": 1.1479, "step": 390 }, { "epoch": 0.346035917652212, "grad_norm": 0.2141467823787157, "learning_rate": 1.6544398886743934e-05, "loss": 1.1424, "step": 395 }, { "epoch": 0.3504161191414805, "grad_norm": 0.20902762397068814, "learning_rate": 1.6427876096865394e-05, "loss": 1.1401, "step": 400 }, { "epoch": 0.354796320630749, "grad_norm": 0.20864014440590625, "learning_rate": 1.6309846685883726e-05, "loss": 1.1404, "step": 405 }, { "epoch": 0.3591765221200175, "grad_norm": 0.22432181453053213, "learning_rate": 1.6190338318551426e-05, "loss": 1.1539, "step": 410 }, { "epoch": 0.36355672360928604, "grad_norm": 0.20427723583571802, "learning_rate": 1.606937900627157e-05, "loss": 1.1402, "step": 415 }, { "epoch": 0.3679369250985545, "grad_norm": 0.20345261931074785, "learning_rate": 1.594699710053223e-05, "loss": 1.1454, "step": 420 }, { "epoch": 0.37231712658782307, "grad_norm": 0.20907937476304458, "learning_rate": 1.5823221286261217e-05, "loss": 1.1282, "step": 425 }, { "epoch": 0.37669732807709155, "grad_norm": 0.2065934957421514, "learning_rate": 1.5698080575102662e-05, "loss": 1.1334, "step": 430 }, { "epoch": 0.38107752956636004, "grad_norm": 0.2161916996533843, "learning_rate": 1.557160429861702e-05, "loss": 1.152, "step": 435 }, { "epoch": 0.3854577310556286, "grad_norm": 0.21804609517909665, "learning_rate": 1.5443822101406066e-05, "loss": 1.1379, "step": 440 }, { "epoch": 0.38983793254489707, "grad_norm": 0.20636895330872523, "learning_rate": 1.531476393416456e-05, "loss": 1.1592, "step": 445 }, { "epoch": 0.39421813403416556, "grad_norm": 0.20437911994063385, "learning_rate": 1.5184460046660139e-05, "loss": 1.1277, "step": 450 }, { "epoch": 0.3985983355234341, "grad_norm": 0.2060999428500134, "learning_rate": 1.50529409806431e-05, "loss": 1.1453, "step": 455 }, { "epoch": 0.4029785370127026, "grad_norm": 0.22533191644861675, "learning_rate": 1.4920237562687784e-05, "loss": 1.1502, "step": 460 }, { "epoch": 0.4073587385019711, "grad_norm": 0.23260594557785327, "learning_rate": 1.478638089696716e-05, "loss": 1.1248, "step": 465 }, { "epoch": 0.4117389399912396, "grad_norm": 0.22947787094553646, "learning_rate": 1.4651402357962368e-05, "loss": 1.1318, "step": 470 }, { "epoch": 0.4161191414805081, "grad_norm": 0.2009264544016685, "learning_rate": 1.4515333583108896e-05, "loss": 1.1543, "step": 475 }, { "epoch": 0.4204993429697766, "grad_norm": 0.20022135464643717, "learning_rate": 1.4378206465381122e-05, "loss": 1.1405, "step": 480 }, { "epoch": 0.42487954445904513, "grad_norm": 0.20457278775552865, "learning_rate": 1.4240053145816968e-05, "loss": 1.1434, "step": 485 }, { "epoch": 0.4292597459483136, "grad_norm": 0.20324616627736056, "learning_rate": 1.4100906005984404e-05, "loss": 1.1479, "step": 490 }, { "epoch": 0.4336399474375821, "grad_norm": 0.21908018025065346, "learning_rate": 1.396079766039157e-05, "loss": 1.1449, "step": 495 }, { "epoch": 0.43802014892685065, "grad_norm": 0.210033035128553, "learning_rate": 1.381976094884232e-05, "loss": 1.132, "step": 500 }, { "epoch": 0.44240035041611914, "grad_norm": 0.2013972756453919, "learning_rate": 1.3677828928738934e-05, "loss": 1.1473, "step": 505 }, { "epoch": 0.4467805519053876, "grad_norm": 0.20820144402446275, "learning_rate": 1.3535034867333838e-05, "loss": 1.1398, "step": 510 }, { "epoch": 0.45116075339465617, "grad_norm": 0.21960485445473304, "learning_rate": 1.3391412233932148e-05, "loss": 1.1527, "step": 515 }, { "epoch": 0.45554095488392465, "grad_norm": 0.21395914429214383, "learning_rate": 1.3246994692046837e-05, "loss": 1.1486, "step": 520 }, { "epoch": 0.45992115637319314, "grad_norm": 0.21230485993764248, "learning_rate": 1.3101816091508389e-05, "loss": 1.1537, "step": 525 }, { "epoch": 0.4643013578624617, "grad_norm": 0.21651837542900804, "learning_rate": 1.2955910460530787e-05, "loss": 1.1496, "step": 530 }, { "epoch": 0.46868155935173017, "grad_norm": 0.20750629246921734, "learning_rate": 1.2809311997735697e-05, "loss": 1.1355, "step": 535 }, { "epoch": 0.4730617608409987, "grad_norm": 0.2039566928608635, "learning_rate": 1.266205506413667e-05, "loss": 1.1375, "step": 540 }, { "epoch": 0.4774419623302672, "grad_norm": 0.20670703932671342, "learning_rate": 1.2514174175085346e-05, "loss": 1.1441, "step": 545 }, { "epoch": 0.4818221638195357, "grad_norm": 0.20394796072197643, "learning_rate": 1.2365703992181425e-05, "loss": 1.1192, "step": 550 }, { "epoch": 0.48620236530880423, "grad_norm": 0.19785902077082673, "learning_rate": 1.2216679315148388e-05, "loss": 1.1303, "step": 555 }, { "epoch": 0.4905825667980727, "grad_norm": 0.20446955932426367, "learning_rate": 1.2067135073676841e-05, "loss": 1.1327, "step": 560 }, { "epoch": 0.4949627682873412, "grad_norm": 0.20819551033939337, "learning_rate": 1.1917106319237386e-05, "loss": 1.1262, "step": 565 }, { "epoch": 0.49934296977660975, "grad_norm": 0.202514436311041, "learning_rate": 1.1766628216864961e-05, "loss": 1.1451, "step": 570 }, { "epoch": 0.5037231712658782, "grad_norm": 0.20270248880014877, "learning_rate": 1.161573603691655e-05, "loss": 1.1096, "step": 575 }, { "epoch": 0.5081033727551467, "grad_norm": 0.19902608440066913, "learning_rate": 1.1464465146804218e-05, "loss": 1.1429, "step": 580 }, { "epoch": 0.5124835742444153, "grad_norm": 0.21488103670405032, "learning_rate": 1.1312851002705383e-05, "loss": 1.1258, "step": 585 }, { "epoch": 0.5168637757336837, "grad_norm": 0.2079533279978145, "learning_rate": 1.1160929141252303e-05, "loss": 1.1326, "step": 590 }, { "epoch": 0.5212439772229522, "grad_norm": 0.2024904670673594, "learning_rate": 1.1008735171202685e-05, "loss": 1.1516, "step": 595 }, { "epoch": 0.5256241787122208, "grad_norm": 0.2009798683547318, "learning_rate": 1.0856304765093391e-05, "loss": 1.126, "step": 600 }, { "epoch": 0.5300043802014893, "grad_norm": 0.20247591682210167, "learning_rate": 1.0703673650879219e-05, "loss": 1.1182, "step": 605 }, { "epoch": 0.5343845816907578, "grad_norm": 0.2040201994565877, "learning_rate": 1.0550877603558656e-05, "loss": 1.1406, "step": 610 }, { "epoch": 0.5387647831800263, "grad_norm": 0.2015454994775207, "learning_rate": 1.0397952436788643e-05, "loss": 1.1166, "step": 615 }, { "epoch": 0.5431449846692948, "grad_norm": 0.19828586396121736, "learning_rate": 1.024493399449025e-05, "loss": 1.1271, "step": 620 }, { "epoch": 0.5475251861585633, "grad_norm": 0.202990185981543, "learning_rate": 1.0091858142447266e-05, "loss": 1.1404, "step": 625 }, { "epoch": 0.5519053876478318, "grad_norm": 0.1953396887915411, "learning_rate": 9.938760759899674e-06, "loss": 1.1282, "step": 630 }, { "epoch": 0.5562855891371004, "grad_norm": 0.19897584069638927, "learning_rate": 9.785677731133972e-06, "loss": 1.116, "step": 635 }, { "epoch": 0.5606657906263688, "grad_norm": 0.19579937491086502, "learning_rate": 9.632644937072277e-06, "loss": 1.1211, "step": 640 }, { "epoch": 0.5650459921156373, "grad_norm": 0.19292547383921327, "learning_rate": 9.479698246862277e-06, "loss": 1.1249, "step": 645 }, { "epoch": 0.5694261936049059, "grad_norm": 0.2035897790907176, "learning_rate": 9.326873509469887e-06, "loss": 1.1222, "step": 650 }, { "epoch": 0.5738063950941743, "grad_norm": 0.20245065608346366, "learning_rate": 9.174206545276678e-06, "loss": 1.1354, "step": 655 }, { "epoch": 0.5781865965834428, "grad_norm": 0.19934269310500388, "learning_rate": 9.021733137683963e-06, "loss": 1.1443, "step": 660 }, { "epoch": 0.5825667980727114, "grad_norm": 0.19598319523008984, "learning_rate": 8.869489024725595e-06, "loss": 1.1326, "step": 665 }, { "epoch": 0.5869469995619798, "grad_norm": 0.1945281080977394, "learning_rate": 8.717509890691369e-06, "loss": 1.1239, "step": 670 }, { "epoch": 0.5913272010512484, "grad_norm": 0.20806507918308362, "learning_rate": 8.565831357763039e-06, "loss": 1.1294, "step": 675 }, { "epoch": 0.5957074025405169, "grad_norm": 0.20251893056570133, "learning_rate": 8.414488977664858e-06, "loss": 1.1239, "step": 680 }, { "epoch": 0.6000876040297853, "grad_norm": 0.1991936009015444, "learning_rate": 8.263518223330698e-06, "loss": 1.1322, "step": 685 }, { "epoch": 0.6044678055190539, "grad_norm": 0.1961163554964441, "learning_rate": 8.112954480589558e-06, "loss": 1.127, "step": 690 }, { "epoch": 0.6088480070083224, "grad_norm": 0.1939932601906298, "learning_rate": 7.962833039871562e-06, "loss": 1.1275, "step": 695 }, { "epoch": 0.6132282084975909, "grad_norm": 0.19487292794675937, "learning_rate": 7.813189087936243e-06, "loss": 1.1222, "step": 700 }, { "epoch": 0.6176084099868594, "grad_norm": 0.19843142347965212, "learning_rate": 7.664057699625215e-06, "loss": 1.1509, "step": 705 }, { "epoch": 0.6219886114761279, "grad_norm": 0.20098071367810383, "learning_rate": 7.515473829640987e-06, "loss": 1.1228, "step": 710 }, { "epoch": 0.6263688129653964, "grad_norm": 0.1972827327140634, "learning_rate": 7.367472304354011e-06, "loss": 1.1268, "step": 715 }, { "epoch": 0.6307490144546649, "grad_norm": 0.20576577812522315, "learning_rate": 7.2200878136397355e-06, "loss": 1.115, "step": 720 }, { "epoch": 0.6351292159439335, "grad_norm": 0.1972932881390162, "learning_rate": 7.073354902747742e-06, "loss": 1.1104, "step": 725 }, { "epoch": 0.6395094174332019, "grad_norm": 0.20208492611746368, "learning_rate": 6.927307964204695e-06, "loss": 1.1356, "step": 730 }, { "epoch": 0.6438896189224704, "grad_norm": 0.19600730585362314, "learning_rate": 6.781981229753145e-06, "loss": 1.1225, "step": 735 }, { "epoch": 0.648269820411739, "grad_norm": 0.19661439451248647, "learning_rate": 6.637408762327972e-06, "loss": 1.1175, "step": 740 }, { "epoch": 0.6526500219010074, "grad_norm": 0.20253732148210266, "learning_rate": 6.4936244480724575e-06, "loss": 1.1224, "step": 745 }, { "epoch": 0.657030223390276, "grad_norm": 0.197391894393827, "learning_rate": 6.350661988395723e-06, "loss": 1.1369, "step": 750 }, { "epoch": 0.6614104248795445, "grad_norm": 0.19396216640485134, "learning_rate": 6.208554892073528e-06, "loss": 1.1332, "step": 755 }, { "epoch": 0.6657906263688129, "grad_norm": 0.19950066877821537, "learning_rate": 6.067336467394169e-06, "loss": 1.1199, "step": 760 }, { "epoch": 0.6701708278580815, "grad_norm": 0.20943729884482393, "learning_rate": 5.927039814351426e-06, "loss": 1.1201, "step": 765 }, { "epoch": 0.67455102934735, "grad_norm": 0.19237438150257924, "learning_rate": 5.787697816886273e-06, "loss": 1.1267, "step": 770 }, { "epoch": 0.6789312308366184, "grad_norm": 0.19252386131062685, "learning_rate": 5.649343135179271e-06, "loss": 1.11, "step": 775 }, { "epoch": 0.683311432325887, "grad_norm": 0.19638246872011564, "learning_rate": 5.512008197995379e-06, "loss": 1.1164, "step": 780 }, { "epoch": 0.6876916338151555, "grad_norm": 0.1877681729558122, "learning_rate": 5.375725195083046e-06, "loss": 1.1257, "step": 785 }, { "epoch": 0.692071835304424, "grad_norm": 0.19133651997309503, "learning_rate": 5.240526069629265e-06, "loss": 1.134, "step": 790 }, { "epoch": 0.6964520367936925, "grad_norm": 0.19745307442946453, "learning_rate": 5.106442510772489e-06, "loss": 1.1412, "step": 795 }, { "epoch": 0.700832238282961, "grad_norm": 0.18826987382961585, "learning_rate": 4.97350594617502e-06, "loss": 1.12, "step": 800 }, { "epoch": 0.7052124397722295, "grad_norm": 0.20198435756216934, "learning_rate": 4.8417475346567635e-06, "loss": 1.1343, "step": 805 }, { "epoch": 0.709592641261498, "grad_norm": 0.19941557715707345, "learning_rate": 4.711198158891909e-06, "loss": 1.1229, "step": 810 }, { "epoch": 0.7139728427507666, "grad_norm": 0.20442006319949774, "learning_rate": 4.581888418170429e-06, "loss": 1.1353, "step": 815 }, { "epoch": 0.718353044240035, "grad_norm": 0.2001055609078045, "learning_rate": 4.453848621225913e-06, "loss": 1.1149, "step": 820 }, { "epoch": 0.7227332457293035, "grad_norm": 0.18925787454519752, "learning_rate": 4.327108779131573e-06, "loss": 1.133, "step": 825 }, { "epoch": 0.7271134472185721, "grad_norm": 0.1863143936478909, "learning_rate": 4.201698598265973e-06, "loss": 1.1162, "step": 830 }, { "epoch": 0.7314936487078406, "grad_norm": 0.1879694577872159, "learning_rate": 4.077647473350201e-06, "loss": 1.126, "step": 835 }, { "epoch": 0.735873850197109, "grad_norm": 0.1919436380192233, "learning_rate": 3.954984480558071e-06, "loss": 1.1195, "step": 840 }, { "epoch": 0.7402540516863776, "grad_norm": 0.20358611131495935, "learning_rate": 3.83373837070101e-06, "loss": 1.1255, "step": 845 }, { "epoch": 0.7446342531756461, "grad_norm": 0.20008378915370192, "learning_rate": 3.7139375624891795e-06, "loss": 1.1412, "step": 850 }, { "epoch": 0.7490144546649146, "grad_norm": 0.19290096180375577, "learning_rate": 3.595610135870472e-06, "loss": 1.1207, "step": 855 }, { "epoch": 0.7533946561541831, "grad_norm": 0.196072459641094, "learning_rate": 3.478783825448869e-06, "loss": 1.1125, "step": 860 }, { "epoch": 0.7577748576434516, "grad_norm": 0.1928642029524879, "learning_rate": 3.3634860139837877e-06, "loss": 1.1229, "step": 865 }, { "epoch": 0.7621550591327201, "grad_norm": 0.20130571120198174, "learning_rate": 3.249743725971849e-06, "loss": 1.1312, "step": 870 }, { "epoch": 0.7665352606219886, "grad_norm": 0.20961643180726597, "learning_rate": 3.1375836213126653e-06, "loss": 1.1232, "step": 875 }, { "epoch": 0.7709154621112572, "grad_norm": 0.18941668671789638, "learning_rate": 3.0270319890600465e-06, "loss": 1.1434, "step": 880 }, { "epoch": 0.7752956636005256, "grad_norm": 0.19381237131051324, "learning_rate": 2.918114741260156e-06, "loss": 1.1439, "step": 885 }, { "epoch": 0.7796758650897941, "grad_norm": 0.1945291975485229, "learning_rate": 2.8108574068780093e-06, "loss": 1.1158, "step": 890 }, { "epoch": 0.7840560665790627, "grad_norm": 0.18927206178983724, "learning_rate": 2.7052851258137936e-06, "loss": 1.1241, "step": 895 }, { "epoch": 0.7884362680683311, "grad_norm": 0.19223522737541443, "learning_rate": 2.601422643010335e-06, "loss": 1.1056, "step": 900 }, { "epoch": 0.7928164695575997, "grad_norm": 0.20045874340313333, "learning_rate": 2.4992943026531935e-06, "loss": 1.1239, "step": 905 }, { "epoch": 0.7971966710468682, "grad_norm": 0.19018510655063894, "learning_rate": 2.3989240424646355e-06, "loss": 1.1189, "step": 910 }, { "epoch": 0.8015768725361366, "grad_norm": 0.19034937418838804, "learning_rate": 2.300335388092929e-06, "loss": 1.1152, "step": 915 }, { "epoch": 0.8059570740254052, "grad_norm": 0.19645900089237364, "learning_rate": 2.2035514475981756e-06, "loss": 1.1313, "step": 920 }, { "epoch": 0.8103372755146737, "grad_norm": 0.18847516841511264, "learning_rate": 2.1085949060360654e-06, "loss": 1.1135, "step": 925 }, { "epoch": 0.8147174770039421, "grad_norm": 0.19132781030558177, "learning_rate": 2.015488020140737e-06, "loss": 1.1045, "step": 930 }, { "epoch": 0.8190976784932107, "grad_norm": 0.18904026009165054, "learning_rate": 1.924252613108073e-06, "loss": 1.1318, "step": 935 }, { "epoch": 0.8234778799824792, "grad_norm": 0.19251546970491693, "learning_rate": 1.8349100694805711e-06, "loss": 1.1243, "step": 940 }, { "epoch": 0.8278580814717477, "grad_norm": 0.19006518497007893, "learning_rate": 1.7474813301350668e-06, "loss": 1.123, "step": 945 }, { "epoch": 0.8322382829610162, "grad_norm": 0.19325212813265874, "learning_rate": 1.661986887374415e-06, "loss": 1.1129, "step": 950 }, { "epoch": 0.8366184844502847, "grad_norm": 0.1889419224748978, "learning_rate": 1.578446780124344e-06, "loss": 1.1103, "step": 955 }, { "epoch": 0.8409986859395532, "grad_norm": 0.19033279856504365, "learning_rate": 1.49688058923654e-06, "loss": 1.1098, "step": 960 }, { "epoch": 0.8453788874288217, "grad_norm": 0.18633846336335902, "learning_rate": 1.4173074328991376e-06, "loss": 1.1287, "step": 965 }, { "epoch": 0.8497590889180903, "grad_norm": 0.19006030021409406, "learning_rate": 1.339745962155613e-06, "loss": 1.1329, "step": 970 }, { "epoch": 0.8541392904073587, "grad_norm": 0.18903779583859964, "learning_rate": 1.2642143565332154e-06, "loss": 1.1316, "step": 975 }, { "epoch": 0.8585194918966272, "grad_norm": 0.19272564813632415, "learning_rate": 1.1907303197818665e-06, "loss": 1.1303, "step": 980 }, { "epoch": 0.8628996933858958, "grad_norm": 0.18878052764179085, "learning_rate": 1.1193110757246251e-06, "loss": 1.1205, "step": 985 }, { "epoch": 0.8672798948751642, "grad_norm": 0.19218035241765347, "learning_rate": 1.0499733642206034e-06, "loss": 1.124, "step": 990 }, { "epoch": 0.8716600963644328, "grad_norm": 0.18852107182367117, "learning_rate": 9.827334372413444e-07, "loss": 1.1301, "step": 995 }, { "epoch": 0.8760402978537013, "grad_norm": 0.19264462588904008, "learning_rate": 9.176070550615379e-07, "loss": 1.1254, "step": 1000 }, { "epoch": 0.8804204993429697, "grad_norm": 0.18440224651228512, "learning_rate": 8.546094825649909e-07, "loss": 1.1189, "step": 1005 }, { "epoch": 0.8848007008322383, "grad_norm": 0.18637643466062778, "learning_rate": 7.937554856667196e-07, "loss": 1.1386, "step": 1010 }, { "epoch": 0.8891809023215068, "grad_norm": 0.19217636415097236, "learning_rate": 7.350593278519824e-07, "loss": 1.1512, "step": 1015 }, { "epoch": 0.8935611038107752, "grad_norm": 0.19868031904477793, "learning_rate": 6.785347668330777e-07, "loss": 1.1385, "step": 1020 }, { "epoch": 0.8979413053000438, "grad_norm": 0.18822804076907482, "learning_rate": 6.241950513246931e-07, "loss": 1.1236, "step": 1025 }, { "epoch": 0.9023215067893123, "grad_norm": 0.18768750481368643, "learning_rate": 5.720529179385659e-07, "loss": 1.1096, "step": 1030 }, { "epoch": 0.9067017082785808, "grad_norm": 0.18565141772239874, "learning_rate": 5.221205881981594e-07, "loss": 1.1197, "step": 1035 }, { "epoch": 0.9110819097678493, "grad_norm": 0.2034853308480849, "learning_rate": 4.7440976567407096e-07, "loss": 1.1371, "step": 1040 }, { "epoch": 0.9154621112571178, "grad_norm": 0.1998255782368389, "learning_rate": 4.2893163324085886e-07, "loss": 1.1115, "step": 1045 }, { "epoch": 0.9198423127463863, "grad_norm": 0.19439905516574385, "learning_rate": 3.856968504558989e-07, "loss": 1.1156, "step": 1050 }, { "epoch": 0.9242225142356548, "grad_norm": 0.18518716914861474, "learning_rate": 3.4471555106090573e-07, "loss": 1.1316, "step": 1055 }, { "epoch": 0.9286027157249234, "grad_norm": 0.18819573805967685, "learning_rate": 3.059973406066963e-07, "loss": 1.1215, "step": 1060 }, { "epoch": 0.9329829172141918, "grad_norm": 0.18672133869392946, "learning_rate": 2.6955129420176193e-07, "loss": 1.1264, "step": 1065 }, { "epoch": 0.9373631187034603, "grad_norm": 0.18848502790832367, "learning_rate": 2.3538595438516442e-07, "loss": 1.1215, "step": 1070 }, { "epoch": 0.9417433201927289, "grad_norm": 0.18909475354670285, "learning_rate": 2.035093291242607e-07, "loss": 1.1324, "step": 1075 }, { "epoch": 0.9461235216819974, "grad_norm": 0.18392034150187955, "learning_rate": 1.7392888993773005e-07, "loss": 1.0989, "step": 1080 }, { "epoch": 0.9505037231712659, "grad_norm": 0.18422655351430084, "learning_rate": 1.466515701443294e-07, "loss": 1.149, "step": 1085 }, { "epoch": 0.9548839246605344, "grad_norm": 0.18472721089921113, "learning_rate": 1.2168376323780652e-07, "loss": 1.1338, "step": 1090 }, { "epoch": 0.9592641261498029, "grad_norm": 0.18618589859950402, "learning_rate": 9.90313213883376e-08, "loss": 1.1488, "step": 1095 }, { "epoch": 0.9636443276390714, "grad_norm": 0.1841956829600993, "learning_rate": 7.86995540708424e-08, "loss": 1.1058, "step": 1100 }, { "epoch": 0.9680245291283399, "grad_norm": 0.18737860582967023, "learning_rate": 6.069322682050516e-08, "loss": 1.1347, "step": 1105 }, { "epoch": 0.9724047306176085, "grad_norm": 0.18327423894477116, "learning_rate": 4.501656011579037e-08, "loss": 1.1223, "step": 1110 }, { "epoch": 0.9767849321068769, "grad_norm": 0.18664413828439025, "learning_rate": 3.167322838920406e-08, "loss": 1.1036, "step": 1115 }, { "epoch": 0.9811651335961454, "grad_norm": 0.18891847587451832, "learning_rate": 2.066635916605386e-08, "loss": 1.1229, "step": 1120 }, { "epoch": 0.985545335085414, "grad_norm": 0.19664758693417972, "learning_rate": 1.1998532331389812e-08, "loss": 1.156, "step": 1125 }, { "epoch": 0.9899255365746824, "grad_norm": 0.18716549597123566, "learning_rate": 5.671779525311394e-09, "loss": 1.1146, "step": 1130 }, { "epoch": 0.994305738063951, "grad_norm": 0.1847881495731737, "learning_rate": 1.6875836667729073e-09, "loss": 1.1207, "step": 1135 }, { "epoch": 0.9986859395532195, "grad_norm": 0.18567353384044155, "learning_rate": 4.687860599927874e-11, "loss": 1.1244, "step": 1140 }, { "epoch": 0.9995619798510732, "step": 1141, "total_flos": 215836199485440.0, "train_loss": 1.3199829850460123, "train_runtime": 8951.0623, "train_samples_per_second": 16.317, "train_steps_per_second": 0.127 } ], "logging_steps": 5, "max_steps": 1141, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 215836199485440.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }