{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9997336884154461, "eval_steps": 500, "global_step": 1877, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0005326231691078562, "grad_norm": 0.5528136747380069, "learning_rate": 1.0638297872340427e-06, "loss": 1.7189, "step": 1 }, { "epoch": 0.002663115845539281, "grad_norm": 0.5731888996610293, "learning_rate": 5.319148936170213e-06, "loss": 1.7582, "step": 5 }, { "epoch": 0.005326231691078562, "grad_norm": 0.5597088455341374, "learning_rate": 1.0638297872340426e-05, "loss": 1.7331, "step": 10 }, { "epoch": 0.007989347536617843, "grad_norm": 0.6373028746703387, "learning_rate": 1.595744680851064e-05, "loss": 1.7083, "step": 15 }, { "epoch": 0.010652463382157125, "grad_norm": 0.6190956100247592, "learning_rate": 2.1276595744680852e-05, "loss": 1.6453, "step": 20 }, { "epoch": 0.013315579227696404, "grad_norm": 0.5346124338518666, "learning_rate": 2.6595744680851064e-05, "loss": 1.5715, "step": 25 }, { "epoch": 0.015978695073235686, "grad_norm": 0.25993843509864667, "learning_rate": 3.191489361702128e-05, "loss": 1.4359, "step": 30 }, { "epoch": 0.018641810918774968, "grad_norm": 0.45126098872878045, "learning_rate": 3.723404255319149e-05, "loss": 1.4122, "step": 35 }, { "epoch": 0.02130492676431425, "grad_norm": 0.16268004246261306, "learning_rate": 4.2553191489361704e-05, "loss": 1.3378, "step": 40 }, { "epoch": 0.023968042609853527, "grad_norm": 0.16755798993478008, "learning_rate": 4.787234042553192e-05, "loss": 1.3315, "step": 45 }, { "epoch": 0.02663115845539281, "grad_norm": 0.14666977328343594, "learning_rate": 5.319148936170213e-05, "loss": 1.3259, "step": 50 }, { "epoch": 0.02929427430093209, "grad_norm": 0.10629659674680719, "learning_rate": 5.851063829787234e-05, "loss": 1.288, "step": 55 }, { "epoch": 0.03195739014647137, "grad_norm": 0.11263809733941188, "learning_rate": 6.382978723404256e-05, "loss": 1.2763, "step": 60 }, { "epoch": 0.03462050599201065, "grad_norm": 0.14445654795383106, "learning_rate": 6.914893617021277e-05, "loss": 1.256, "step": 65 }, { "epoch": 0.037283621837549935, "grad_norm": 0.07679775841947081, "learning_rate": 7.446808510638298e-05, "loss": 1.2481, "step": 70 }, { "epoch": 0.03994673768308921, "grad_norm": 0.07097172061630208, "learning_rate": 7.978723404255319e-05, "loss": 1.2417, "step": 75 }, { "epoch": 0.0426098535286285, "grad_norm": 0.07545648577598915, "learning_rate": 8.510638297872341e-05, "loss": 1.2048, "step": 80 }, { "epoch": 0.045272969374167776, "grad_norm": 0.06657701317469632, "learning_rate": 9.042553191489363e-05, "loss": 1.2138, "step": 85 }, { "epoch": 0.047936085219707054, "grad_norm": 0.08276972517588223, "learning_rate": 9.574468085106384e-05, "loss": 1.2267, "step": 90 }, { "epoch": 0.05059920106524634, "grad_norm": 0.14086227463755532, "learning_rate": 0.00010106382978723406, "loss": 1.2185, "step": 95 }, { "epoch": 0.05326231691078562, "grad_norm": 0.08222057468849275, "learning_rate": 0.00010638297872340425, "loss": 1.2021, "step": 100 }, { "epoch": 0.0559254327563249, "grad_norm": 0.06741247738810993, "learning_rate": 0.00011170212765957446, "loss": 1.1957, "step": 105 }, { "epoch": 0.05858854860186418, "grad_norm": 0.07692927491859038, "learning_rate": 0.00011702127659574468, "loss": 1.1901, "step": 110 }, { "epoch": 0.06125166444740346, "grad_norm": 0.08953054401601632, "learning_rate": 0.0001223404255319149, "loss": 1.2002, "step": 115 }, { "epoch": 0.06391478029294274, "grad_norm": 0.066045987418387, "learning_rate": 0.00012765957446808513, "loss": 1.2086, "step": 120 }, { "epoch": 0.06657789613848203, "grad_norm": 0.06476207146640194, "learning_rate": 0.00013297872340425532, "loss": 1.215, "step": 125 }, { "epoch": 0.0692410119840213, "grad_norm": 0.07334327374343644, "learning_rate": 0.00013829787234042554, "loss": 1.181, "step": 130 }, { "epoch": 0.07190412782956059, "grad_norm": 0.071815699820189, "learning_rate": 0.00014361702127659576, "loss": 1.1795, "step": 135 }, { "epoch": 0.07456724367509987, "grad_norm": 0.08268574123602224, "learning_rate": 0.00014893617021276596, "loss": 1.1892, "step": 140 }, { "epoch": 0.07723035952063914, "grad_norm": 0.07585606469879155, "learning_rate": 0.00015425531914893618, "loss": 1.173, "step": 145 }, { "epoch": 0.07989347536617843, "grad_norm": 0.06900850276332868, "learning_rate": 0.00015957446808510637, "loss": 1.1889, "step": 150 }, { "epoch": 0.08255659121171771, "grad_norm": 0.07702070923317432, "learning_rate": 0.00016489361702127662, "loss": 1.1705, "step": 155 }, { "epoch": 0.085219707057257, "grad_norm": 0.0724505634260966, "learning_rate": 0.00017021276595744682, "loss": 1.1801, "step": 160 }, { "epoch": 0.08788282290279627, "grad_norm": 0.07762044850846143, "learning_rate": 0.000175531914893617, "loss": 1.1863, "step": 165 }, { "epoch": 0.09054593874833555, "grad_norm": 0.0825988729050522, "learning_rate": 0.00018085106382978726, "loss": 1.1687, "step": 170 }, { "epoch": 0.09320905459387484, "grad_norm": 0.08362625552402488, "learning_rate": 0.00018617021276595746, "loss": 1.1875, "step": 175 }, { "epoch": 0.09587217043941411, "grad_norm": 0.07968629276225715, "learning_rate": 0.00019148936170212768, "loss": 1.1629, "step": 180 }, { "epoch": 0.0985352862849534, "grad_norm": 0.09010747938874886, "learning_rate": 0.00019680851063829787, "loss": 1.1682, "step": 185 }, { "epoch": 0.10119840213049268, "grad_norm": 0.10276104499280464, "learning_rate": 0.00019999930805760402, "loss": 1.1618, "step": 190 }, { "epoch": 0.10386151797603196, "grad_norm": 0.07684956533429005, "learning_rate": 0.00019999152381561955, "loss": 1.1902, "step": 195 }, { "epoch": 0.10652463382157124, "grad_norm": 0.09365634557072464, "learning_rate": 0.0001999750910791767, "loss": 1.1673, "step": 200 }, { "epoch": 0.10918774966711052, "grad_norm": 0.08320439951882774, "learning_rate": 0.00019995001126958025, "loss": 1.1845, "step": 205 }, { "epoch": 0.1118508655126498, "grad_norm": 0.08595161594886752, "learning_rate": 0.00019991628655604003, "loss": 1.1444, "step": 210 }, { "epoch": 0.11451398135818908, "grad_norm": 0.07678856693871118, "learning_rate": 0.00019987391985548328, "loss": 1.1724, "step": 215 }, { "epoch": 0.11717709720372836, "grad_norm": 0.07433277571881601, "learning_rate": 0.0001998229148323023, "loss": 1.1469, "step": 220 }, { "epoch": 0.11984021304926765, "grad_norm": 0.11001095605643386, "learning_rate": 0.00019976327589803767, "loss": 1.1383, "step": 225 }, { "epoch": 0.12250332889480692, "grad_norm": 0.0784178760835021, "learning_rate": 0.0001996950082109965, "loss": 1.1818, "step": 230 }, { "epoch": 0.12516644474034622, "grad_norm": 0.08047194112395492, "learning_rate": 0.00019961811767580648, "loss": 1.1445, "step": 235 }, { "epoch": 0.1278295605858855, "grad_norm": 0.0670667235371544, "learning_rate": 0.0001995326109429049, "loss": 1.1741, "step": 240 }, { "epoch": 0.13049267643142476, "grad_norm": 0.07072589446768075, "learning_rate": 0.00019943849540796375, "loss": 1.157, "step": 245 }, { "epoch": 0.13315579227696406, "grad_norm": 0.07466892570841129, "learning_rate": 0.0001993357792112498, "loss": 1.125, "step": 250 }, { "epoch": 0.13581890812250333, "grad_norm": 0.07302104613788317, "learning_rate": 0.0001992244712369207, "loss": 1.1615, "step": 255 }, { "epoch": 0.1384820239680426, "grad_norm": 0.07211635352591637, "learning_rate": 0.00019910458111225646, "loss": 1.1441, "step": 260 }, { "epoch": 0.1411451398135819, "grad_norm": 0.07103357444221702, "learning_rate": 0.00019897611920682677, "loss": 1.1493, "step": 265 }, { "epoch": 0.14380825565912117, "grad_norm": 0.0698227187710226, "learning_rate": 0.00019883909663159424, "loss": 1.1568, "step": 270 }, { "epoch": 0.14647137150466044, "grad_norm": 0.07137557168765225, "learning_rate": 0.0001986935252379532, "loss": 1.171, "step": 275 }, { "epoch": 0.14913448735019974, "grad_norm": 0.07605080544337586, "learning_rate": 0.00019853941761670483, "loss": 1.1623, "step": 280 }, { "epoch": 0.151797603195739, "grad_norm": 0.09532848101140429, "learning_rate": 0.00019837678709696798, "loss": 1.1888, "step": 285 }, { "epoch": 0.15446071904127828, "grad_norm": 0.07485256895909924, "learning_rate": 0.00019820564774502644, "loss": 1.1483, "step": 290 }, { "epoch": 0.15712383488681758, "grad_norm": 0.07483378156117482, "learning_rate": 0.0001980260143631122, "loss": 1.1375, "step": 295 }, { "epoch": 0.15978695073235685, "grad_norm": 0.07954155407009747, "learning_rate": 0.00019783790248812533, "loss": 1.1696, "step": 300 }, { "epoch": 0.16245006657789615, "grad_norm": 0.08502452471103343, "learning_rate": 0.00019764132839029, "loss": 1.168, "step": 305 }, { "epoch": 0.16511318242343542, "grad_norm": 0.08384910068571033, "learning_rate": 0.00019743630907174725, "loss": 1.1659, "step": 310 }, { "epoch": 0.1677762982689747, "grad_norm": 0.06905463640642404, "learning_rate": 0.0001972228622650846, "loss": 1.1612, "step": 315 }, { "epoch": 0.170439414114514, "grad_norm": 0.19257912301232658, "learning_rate": 0.0001970010064318021, "loss": 1.1517, "step": 320 }, { "epoch": 0.17310252996005326, "grad_norm": 0.0793114626498931, "learning_rate": 0.00019677076076071566, "loss": 1.1385, "step": 325 }, { "epoch": 0.17576564580559254, "grad_norm": 0.07393026070000318, "learning_rate": 0.00019653214516629735, "loss": 1.1426, "step": 330 }, { "epoch": 0.17842876165113183, "grad_norm": 0.08179432509124362, "learning_rate": 0.00019628518028695307, "loss": 1.1104, "step": 335 }, { "epoch": 0.1810918774966711, "grad_norm": 0.09735291608528279, "learning_rate": 0.00019602988748323717, "loss": 1.1563, "step": 340 }, { "epoch": 0.18375499334221038, "grad_norm": 0.06743724715009518, "learning_rate": 0.00019576628883600535, "loss": 1.1406, "step": 345 }, { "epoch": 0.18641810918774968, "grad_norm": 0.075326384879952, "learning_rate": 0.00019549440714450444, "loss": 1.1572, "step": 350 }, { "epoch": 0.18908122503328895, "grad_norm": 0.07438689728031705, "learning_rate": 0.00019521426592440072, "loss": 1.1479, "step": 355 }, { "epoch": 0.19174434087882822, "grad_norm": 0.07277611336304127, "learning_rate": 0.00019492588940574586, "loss": 1.1549, "step": 360 }, { "epoch": 0.19440745672436752, "grad_norm": 0.0695324135241875, "learning_rate": 0.0001946293025308813, "loss": 1.1435, "step": 365 }, { "epoch": 0.1970705725699068, "grad_norm": 0.06685927618032904, "learning_rate": 0.00019432453095228076, "loss": 1.1641, "step": 370 }, { "epoch": 0.19973368841544606, "grad_norm": 0.0680367135740568, "learning_rate": 0.00019401160103033174, "loss": 1.1261, "step": 375 }, { "epoch": 0.20239680426098536, "grad_norm": 0.08027336453756874, "learning_rate": 0.00019369053983105532, "loss": 1.1368, "step": 380 }, { "epoch": 0.20505992010652463, "grad_norm": 0.0707161713953054, "learning_rate": 0.00019336137512376532, "loss": 1.1588, "step": 385 }, { "epoch": 0.20772303595206393, "grad_norm": 0.07189527593634382, "learning_rate": 0.00019302413537866642, "loss": 1.1552, "step": 390 }, { "epoch": 0.2103861517976032, "grad_norm": 0.0716934364253126, "learning_rate": 0.0001926788497643916, "loss": 1.1577, "step": 395 }, { "epoch": 0.21304926764314247, "grad_norm": 0.065943892133018, "learning_rate": 0.00019232554814547953, "loss": 1.1203, "step": 400 }, { "epoch": 0.21571238348868177, "grad_norm": 0.07352621386091099, "learning_rate": 0.00019196426107979128, "loss": 1.1266, "step": 405 }, { "epoch": 0.21837549933422104, "grad_norm": 0.07441803674470306, "learning_rate": 0.00019159501981586737, "loss": 1.1432, "step": 410 }, { "epoch": 0.2210386151797603, "grad_norm": 0.07291702193187057, "learning_rate": 0.00019121785629022501, "loss": 1.1344, "step": 415 }, { "epoch": 0.2237017310252996, "grad_norm": 0.07094925179230635, "learning_rate": 0.00019083280312459593, "loss": 1.1137, "step": 420 }, { "epoch": 0.22636484687083888, "grad_norm": 0.07399044805064979, "learning_rate": 0.0001904398936231047, "loss": 1.1533, "step": 425 }, { "epoch": 0.22902796271637815, "grad_norm": 0.07782197426798759, "learning_rate": 0.00019003916176938836, "loss": 1.1458, "step": 430 }, { "epoch": 0.23169107856191745, "grad_norm": 0.06822071830212563, "learning_rate": 0.00018963064222365694, "loss": 1.1448, "step": 435 }, { "epoch": 0.23435419440745672, "grad_norm": 0.06944246120343146, "learning_rate": 0.00018921437031969558, "loss": 1.1577, "step": 440 }, { "epoch": 0.237017310252996, "grad_norm": 0.07108688216307608, "learning_rate": 0.0001887903820618087, "loss": 1.1526, "step": 445 }, { "epoch": 0.2396804260985353, "grad_norm": 0.08455610060485116, "learning_rate": 0.00018835871412170563, "loss": 1.1517, "step": 450 }, { "epoch": 0.24234354194407456, "grad_norm": 0.06664786445884358, "learning_rate": 0.0001879194038353289, "loss": 1.1537, "step": 455 }, { "epoch": 0.24500665778961384, "grad_norm": 0.07089581724112333, "learning_rate": 0.00018747248919962498, "loss": 1.1409, "step": 460 }, { "epoch": 0.24766977363515313, "grad_norm": 0.07242825833109466, "learning_rate": 0.00018701800886925782, "loss": 1.1303, "step": 465 }, { "epoch": 0.25033288948069243, "grad_norm": 0.06598593287452807, "learning_rate": 0.00018655600215326546, "loss": 1.1401, "step": 470 }, { "epoch": 0.2529960053262317, "grad_norm": 0.07020789015379635, "learning_rate": 0.00018608650901166032, "loss": 1.1542, "step": 475 }, { "epoch": 0.255659121171771, "grad_norm": 0.06441793150662321, "learning_rate": 0.0001856095700519726, "loss": 1.1276, "step": 480 }, { "epoch": 0.2583222370173103, "grad_norm": 0.07254719498292789, "learning_rate": 0.0001851252265257384, "loss": 1.1212, "step": 485 }, { "epoch": 0.2609853528628495, "grad_norm": 0.06917909155716108, "learning_rate": 0.0001846335203249316, "loss": 1.1298, "step": 490 }, { "epoch": 0.2636484687083888, "grad_norm": 0.07470942417701212, "learning_rate": 0.00018413449397834051, "loss": 1.1456, "step": 495 }, { "epoch": 0.2663115845539281, "grad_norm": 0.0693858861935873, "learning_rate": 0.00018362819064788956, "loss": 1.1327, "step": 500 }, { "epoch": 0.26897470039946736, "grad_norm": 0.07182079092553902, "learning_rate": 0.00018311465412490608, "loss": 1.1628, "step": 505 }, { "epoch": 0.27163781624500666, "grad_norm": 0.06682954118119949, "learning_rate": 0.00018259392882633265, "loss": 1.1528, "step": 510 }, { "epoch": 0.27430093209054596, "grad_norm": 0.07248673749669132, "learning_rate": 0.00018206605979088542, "loss": 1.156, "step": 515 }, { "epoch": 0.2769640479360852, "grad_norm": 0.06950216959497392, "learning_rate": 0.0001815310926751586, "loss": 1.119, "step": 520 }, { "epoch": 0.2796271637816245, "grad_norm": 0.07067673018407011, "learning_rate": 0.00018098907374967555, "loss": 1.1211, "step": 525 }, { "epoch": 0.2822902796271638, "grad_norm": 0.06820842392733384, "learning_rate": 0.00018044004989488664, "loss": 1.1281, "step": 530 }, { "epoch": 0.28495339547270304, "grad_norm": 0.07418230217074875, "learning_rate": 0.00017988406859711456, "loss": 1.1409, "step": 535 }, { "epoch": 0.28761651131824234, "grad_norm": 0.07009876259688716, "learning_rate": 0.00017932117794444713, "loss": 1.1381, "step": 540 }, { "epoch": 0.29027962716378164, "grad_norm": 0.07129309605598672, "learning_rate": 0.00017875142662257786, "loss": 1.1387, "step": 545 }, { "epoch": 0.2929427430093209, "grad_norm": 0.07830622678131702, "learning_rate": 0.00017817486391059532, "loss": 1.1165, "step": 550 }, { "epoch": 0.2956058588548602, "grad_norm": 0.0709756673443606, "learning_rate": 0.0001775915396767205, "loss": 1.129, "step": 555 }, { "epoch": 0.2982689747003995, "grad_norm": 0.06710174636010342, "learning_rate": 0.00017700150437399405, "loss": 1.1183, "step": 560 }, { "epoch": 0.3009320905459387, "grad_norm": 0.07321620332053846, "learning_rate": 0.0001764048090359121, "loss": 1.1502, "step": 565 }, { "epoch": 0.303595206391478, "grad_norm": 0.07613131980579707, "learning_rate": 0.00017580150527201241, "loss": 1.1322, "step": 570 }, { "epoch": 0.3062583222370173, "grad_norm": 0.07480921806539248, "learning_rate": 0.0001751916452634105, "loss": 1.1269, "step": 575 }, { "epoch": 0.30892143808255657, "grad_norm": 0.07386122393966031, "learning_rate": 0.0001745752817582865, "loss": 1.1528, "step": 580 }, { "epoch": 0.31158455392809586, "grad_norm": 0.07032971968151802, "learning_rate": 0.00017395246806732267, "loss": 1.1642, "step": 585 }, { "epoch": 0.31424766977363516, "grad_norm": 0.07910944378906298, "learning_rate": 0.00017332325805909256, "loss": 1.1328, "step": 590 }, { "epoch": 0.3169107856191744, "grad_norm": 0.06775943214366806, "learning_rate": 0.00017268770615540177, "loss": 1.1142, "step": 595 }, { "epoch": 0.3195739014647137, "grad_norm": 0.0858014191359942, "learning_rate": 0.00017204586732658087, "loss": 1.1393, "step": 600 }, { "epoch": 0.322237017310253, "grad_norm": 0.06968407560583738, "learning_rate": 0.00017139779708673085, "loss": 1.1428, "step": 605 }, { "epoch": 0.3249001331557923, "grad_norm": 0.06812443512073688, "learning_rate": 0.00017074355148892167, "loss": 1.1592, "step": 610 }, { "epoch": 0.32756324900133155, "grad_norm": 0.07150170839574509, "learning_rate": 0.00017008318712034403, "loss": 1.1018, "step": 615 }, { "epoch": 0.33022636484687085, "grad_norm": 0.06906369302490485, "learning_rate": 0.00016941676109741508, "loss": 1.1442, "step": 620 }, { "epoch": 0.33288948069241014, "grad_norm": 0.07869503084625909, "learning_rate": 0.00016874433106083814, "loss": 1.1132, "step": 625 }, { "epoch": 0.3355525965379494, "grad_norm": 0.07767900677929127, "learning_rate": 0.00016806595517061744, "loss": 1.1362, "step": 630 }, { "epoch": 0.3382157123834887, "grad_norm": 0.06780573276986938, "learning_rate": 0.00016738169210102764, "loss": 1.1382, "step": 635 }, { "epoch": 0.340878828229028, "grad_norm": 0.07855904717914698, "learning_rate": 0.00016669160103553884, "loss": 1.1146, "step": 640 }, { "epoch": 0.34354194407456723, "grad_norm": 0.06976051466447154, "learning_rate": 0.00016599574166169782, "loss": 1.1156, "step": 645 }, { "epoch": 0.34620505992010653, "grad_norm": 0.0659729198246215, "learning_rate": 0.0001652941741659655, "loss": 1.1636, "step": 650 }, { "epoch": 0.3488681757656458, "grad_norm": 0.06820327345322129, "learning_rate": 0.00016458695922851125, "loss": 1.1272, "step": 655 }, { "epoch": 0.35153129161118507, "grad_norm": 0.0706423611601847, "learning_rate": 0.0001638741580179645, "loss": 1.15, "step": 660 }, { "epoch": 0.35419440745672437, "grad_norm": 0.07009862917994238, "learning_rate": 0.0001631558321861241, "loss": 1.1133, "step": 665 }, { "epoch": 0.35685752330226367, "grad_norm": 0.10252097837226529, "learning_rate": 0.00016243204386262616, "loss": 1.1275, "step": 670 }, { "epoch": 0.3595206391478029, "grad_norm": 0.0677270369109815, "learning_rate": 0.0001617028556495699, "loss": 1.1463, "step": 675 }, { "epoch": 0.3621837549933422, "grad_norm": 0.07081566637081647, "learning_rate": 0.00016096833061610336, "loss": 1.1557, "step": 680 }, { "epoch": 0.3648468708388815, "grad_norm": 0.07606309640077409, "learning_rate": 0.0001602285322929684, "loss": 1.1279, "step": 685 }, { "epoch": 0.36750998668442075, "grad_norm": 0.06926585358652293, "learning_rate": 0.00015948352466700562, "loss": 1.1058, "step": 690 }, { "epoch": 0.37017310252996005, "grad_norm": 0.0768394516058797, "learning_rate": 0.00015873337217562012, "loss": 1.1451, "step": 695 }, { "epoch": 0.37283621837549935, "grad_norm": 0.07574776045146851, "learning_rate": 0.00015797813970120806, "loss": 1.1529, "step": 700 }, { "epoch": 0.3754993342210386, "grad_norm": 0.08825811667362324, "learning_rate": 0.00015721789256554493, "loss": 1.1427, "step": 705 }, { "epoch": 0.3781624500665779, "grad_norm": 0.07195501325596203, "learning_rate": 0.00015645269652413572, "loss": 1.1348, "step": 710 }, { "epoch": 0.3808255659121172, "grad_norm": 0.07470988656942844, "learning_rate": 0.00015568261776052747, "loss": 1.1389, "step": 715 }, { "epoch": 0.38348868175765644, "grad_norm": 0.07234292608880714, "learning_rate": 0.0001549077228805851, "loss": 1.1265, "step": 720 }, { "epoch": 0.38615179760319573, "grad_norm": 0.07603813138240277, "learning_rate": 0.00015412807890673012, "loss": 1.0975, "step": 725 }, { "epoch": 0.38881491344873503, "grad_norm": 0.06914740850103318, "learning_rate": 0.00015334375327214435, "loss": 1.1656, "step": 730 }, { "epoch": 0.3914780292942743, "grad_norm": 0.0723373355088882, "learning_rate": 0.00015255481381493686, "loss": 1.1235, "step": 735 }, { "epoch": 0.3941411451398136, "grad_norm": 0.07469762097501292, "learning_rate": 0.00015176132877227672, "loss": 1.1401, "step": 740 }, { "epoch": 0.3968042609853529, "grad_norm": 0.06845354027517625, "learning_rate": 0.00015096336677449123, "loss": 1.1299, "step": 745 }, { "epoch": 0.3994673768308921, "grad_norm": 0.07857096344059177, "learning_rate": 0.0001501609968391295, "loss": 1.1362, "step": 750 }, { "epoch": 0.4021304926764314, "grad_norm": 0.07079465135436822, "learning_rate": 0.00014935428836499332, "loss": 1.1268, "step": 755 }, { "epoch": 0.4047936085219707, "grad_norm": 0.07113035589654983, "learning_rate": 0.0001485433111261346, "loss": 1.1357, "step": 760 }, { "epoch": 0.40745672436750996, "grad_norm": 0.0703269232774503, "learning_rate": 0.0001477281352658203, "loss": 1.1239, "step": 765 }, { "epoch": 0.41011984021304926, "grad_norm": 0.07059355929742223, "learning_rate": 0.00014690883129046584, "loss": 1.1442, "step": 770 }, { "epoch": 0.41278295605858856, "grad_norm": 0.07289277380542494, "learning_rate": 0.0001460854700635366, "loss": 1.1267, "step": 775 }, { "epoch": 0.41544607190412786, "grad_norm": 0.06984202720886337, "learning_rate": 0.00014525812279941896, "loss": 1.1258, "step": 780 }, { "epoch": 0.4181091877496671, "grad_norm": 0.07422048925652114, "learning_rate": 0.00014442686105726067, "loss": 1.1193, "step": 785 }, { "epoch": 0.4207723035952064, "grad_norm": 0.0718716738772194, "learning_rate": 0.00014359175673478162, "loss": 1.133, "step": 790 }, { "epoch": 0.4234354194407457, "grad_norm": 0.07204344165616748, "learning_rate": 0.00014275288206205524, "loss": 1.0967, "step": 795 }, { "epoch": 0.42609853528628494, "grad_norm": 0.07353413673583019, "learning_rate": 0.00014191030959526105, "loss": 1.1261, "step": 800 }, { "epoch": 0.42876165113182424, "grad_norm": 0.0707040442967912, "learning_rate": 0.00014106411221040933, "loss": 1.128, "step": 805 }, { "epoch": 0.43142476697736354, "grad_norm": 0.07086259967904554, "learning_rate": 0.00014021436309703765, "loss": 1.107, "step": 810 }, { "epoch": 0.4340878828229028, "grad_norm": 0.06994136097058291, "learning_rate": 0.00013936113575188075, "loss": 1.1221, "step": 815 }, { "epoch": 0.4367509986684421, "grad_norm": 0.06961342073957084, "learning_rate": 0.00013850450397251345, "loss": 1.1208, "step": 820 }, { "epoch": 0.4394141145139814, "grad_norm": 0.07040932769118938, "learning_rate": 0.0001376445418509679, "loss": 1.1208, "step": 825 }, { "epoch": 0.4420772303595206, "grad_norm": 0.07187314857901307, "learning_rate": 0.00013678132376732517, "loss": 1.1267, "step": 830 }, { "epoch": 0.4447403462050599, "grad_norm": 0.07002729221567015, "learning_rate": 0.00013591492438328183, "loss": 1.1421, "step": 835 }, { "epoch": 0.4474034620505992, "grad_norm": 0.07235067324392022, "learning_rate": 0.0001350454186356924, "loss": 1.1191, "step": 840 }, { "epoch": 0.45006657789613846, "grad_norm": 0.07568410556158327, "learning_rate": 0.00013417288173008776, "loss": 1.1123, "step": 845 }, { "epoch": 0.45272969374167776, "grad_norm": 0.07613155957113646, "learning_rate": 0.00013329738913417068, "loss": 1.1137, "step": 850 }, { "epoch": 0.45539280958721706, "grad_norm": 0.06854447943505289, "learning_rate": 0.00013241901657128825, "loss": 1.132, "step": 855 }, { "epoch": 0.4580559254327563, "grad_norm": 0.06884442803824642, "learning_rate": 0.00013153784001388247, "loss": 1.1352, "step": 860 }, { "epoch": 0.4607190412782956, "grad_norm": 0.0818107600119684, "learning_rate": 0.00013065393567691913, "loss": 1.101, "step": 865 }, { "epoch": 0.4633821571238349, "grad_norm": 0.07690425325074156, "learning_rate": 0.00012976738001129606, "loss": 1.1052, "step": 870 }, { "epoch": 0.46604527296937415, "grad_norm": 0.07952080928038242, "learning_rate": 0.00012887824969723034, "loss": 1.1172, "step": 875 }, { "epoch": 0.46870838881491345, "grad_norm": 0.06637081804522467, "learning_rate": 0.00012798662163762635, "loss": 1.1236, "step": 880 }, { "epoch": 0.47137150466045274, "grad_norm": 0.06753657536020181, "learning_rate": 0.00012709257295142422, "loss": 1.1304, "step": 885 }, { "epoch": 0.474034620505992, "grad_norm": 0.07408199495580661, "learning_rate": 0.00012619618096692943, "loss": 1.1523, "step": 890 }, { "epoch": 0.4766977363515313, "grad_norm": 0.07508862680813526, "learning_rate": 0.0001252975232151248, "loss": 1.1158, "step": 895 }, { "epoch": 0.4793608521970706, "grad_norm": 0.07064406721668945, "learning_rate": 0.0001243966774229645, "loss": 1.1334, "step": 900 }, { "epoch": 0.48202396804260983, "grad_norm": 0.0709335857697155, "learning_rate": 0.00012349372150665118, "loss": 1.1104, "step": 905 }, { "epoch": 0.48468708388814913, "grad_norm": 0.06779392856513489, "learning_rate": 0.00012258873356489714, "loss": 1.1299, "step": 910 }, { "epoch": 0.4873501997336884, "grad_norm": 0.07922531663031743, "learning_rate": 0.00012168179187216893, "loss": 1.13, "step": 915 }, { "epoch": 0.49001331557922767, "grad_norm": 0.07185035702343927, "learning_rate": 0.0001207729748719177, "loss": 1.1402, "step": 920 }, { "epoch": 0.49267643142476697, "grad_norm": 0.07162588557593508, "learning_rate": 0.00011986236116979406, "loss": 1.1308, "step": 925 }, { "epoch": 0.49533954727030627, "grad_norm": 0.07242326329471309, "learning_rate": 0.0001189500295268495, "loss": 1.106, "step": 930 }, { "epoch": 0.4980026631158455, "grad_norm": 0.07434286419305244, "learning_rate": 0.0001180360588527242, "loss": 1.119, "step": 935 }, { "epoch": 0.5006657789613849, "grad_norm": 0.07304306593688702, "learning_rate": 0.00011712052819882171, "loss": 1.1503, "step": 940 }, { "epoch": 0.5033288948069241, "grad_norm": 0.07214685709611712, "learning_rate": 0.00011620351675147195, "loss": 1.1095, "step": 945 }, { "epoch": 0.5059920106524634, "grad_norm": 0.07192838129765652, "learning_rate": 0.0001152851038250819, "loss": 1.1451, "step": 950 }, { "epoch": 0.5086551264980027, "grad_norm": 0.06935787206272043, "learning_rate": 0.00011436536885527576, "loss": 1.1251, "step": 955 }, { "epoch": 0.511318242343542, "grad_norm": 0.06801727149157547, "learning_rate": 0.00011344439139202421, "loss": 1.1084, "step": 960 }, { "epoch": 0.5139813581890812, "grad_norm": 0.07024555806497951, "learning_rate": 0.00011252225109276404, "loss": 1.1278, "step": 965 }, { "epoch": 0.5166444740346205, "grad_norm": 0.06796418511383114, "learning_rate": 0.00011159902771550837, "loss": 1.1092, "step": 970 }, { "epoch": 0.5193075898801598, "grad_norm": 0.07219790467805971, "learning_rate": 0.00011067480111194817, "loss": 1.1286, "step": 975 }, { "epoch": 0.521970705725699, "grad_norm": 0.06944360419194191, "learning_rate": 0.00010974965122054579, "loss": 1.1184, "step": 980 }, { "epoch": 0.5246338215712384, "grad_norm": 0.07229326850745169, "learning_rate": 0.00010882365805962083, "loss": 1.1212, "step": 985 }, { "epoch": 0.5272969374167776, "grad_norm": 0.07181125732929394, "learning_rate": 0.00010789690172042912, "loss": 1.1137, "step": 990 }, { "epoch": 0.5299600532623169, "grad_norm": 0.07145294348037948, "learning_rate": 0.00010696946236023567, "loss": 1.1365, "step": 995 }, { "epoch": 0.5326231691078562, "grad_norm": 0.07057067837966788, "learning_rate": 0.00010604142019538135, "loss": 1.1176, "step": 1000 }, { "epoch": 0.5352862849533955, "grad_norm": 0.07461665274600122, "learning_rate": 0.00010511285549434509, "loss": 1.1152, "step": 1005 }, { "epoch": 0.5379494007989347, "grad_norm": 0.07009722058482384, "learning_rate": 0.00010418384857080117, "loss": 1.1117, "step": 1010 }, { "epoch": 0.5406125166444741, "grad_norm": 0.07081680661261375, "learning_rate": 0.00010325447977667263, "loss": 1.1328, "step": 1015 }, { "epoch": 0.5432756324900133, "grad_norm": 0.06980388274242631, "learning_rate": 0.00010232482949518156, "loss": 1.1404, "step": 1020 }, { "epoch": 0.5459387483355526, "grad_norm": 0.06946364728493262, "learning_rate": 0.00010139497813389654, "loss": 1.1127, "step": 1025 }, { "epoch": 0.5486018641810919, "grad_norm": 0.06947172503952885, "learning_rate": 0.00010046500611777798, "loss": 1.0937, "step": 1030 }, { "epoch": 0.5512649800266312, "grad_norm": 0.07414647895551518, "learning_rate": 9.953499388222202e-05, "loss": 1.132, "step": 1035 }, { "epoch": 0.5539280958721704, "grad_norm": 0.07085672498663681, "learning_rate": 9.860502186610349e-05, "loss": 1.0998, "step": 1040 }, { "epoch": 0.5565912117177098, "grad_norm": 0.07198312270884867, "learning_rate": 9.767517050481846e-05, "loss": 1.1263, "step": 1045 }, { "epoch": 0.559254327563249, "grad_norm": 0.07070349541708286, "learning_rate": 9.67455202233274e-05, "loss": 1.1143, "step": 1050 }, { "epoch": 0.5619174434087882, "grad_norm": 0.06990981328605791, "learning_rate": 9.581615142919887e-05, "loss": 1.1168, "step": 1055 }, { "epoch": 0.5645805592543276, "grad_norm": 0.07221018297233557, "learning_rate": 9.488714450565491e-05, "loss": 1.1123, "step": 1060 }, { "epoch": 0.5672436750998668, "grad_norm": 0.06895775963564511, "learning_rate": 9.395857980461867e-05, "loss": 1.1294, "step": 1065 }, { "epoch": 0.5699067909454061, "grad_norm": 0.06904508970279108, "learning_rate": 9.303053763976434e-05, "loss": 1.1179, "step": 1070 }, { "epoch": 0.5725699067909454, "grad_norm": 0.07131791944898686, "learning_rate": 9.210309827957089e-05, "loss": 1.1297, "step": 1075 }, { "epoch": 0.5752330226364847, "grad_norm": 0.07117429268373339, "learning_rate": 9.117634194037922e-05, "loss": 1.1285, "step": 1080 }, { "epoch": 0.5778961384820239, "grad_norm": 0.0720403827517469, "learning_rate": 9.025034877945422e-05, "loss": 1.1418, "step": 1085 }, { "epoch": 0.5805592543275633, "grad_norm": 0.07399424819774852, "learning_rate": 8.932519888805185e-05, "loss": 1.1521, "step": 1090 }, { "epoch": 0.5832223701731025, "grad_norm": 0.06757782738616369, "learning_rate": 8.840097228449165e-05, "loss": 1.1468, "step": 1095 }, { "epoch": 0.5858854860186418, "grad_norm": 0.07513462470349377, "learning_rate": 8.747774890723599e-05, "loss": 1.1008, "step": 1100 }, { "epoch": 0.5885486018641811, "grad_norm": 0.07193663724723076, "learning_rate": 8.655560860797582e-05, "loss": 1.1364, "step": 1105 }, { "epoch": 0.5912117177097204, "grad_norm": 0.07451485853238139, "learning_rate": 8.563463114472425e-05, "loss": 1.1077, "step": 1110 }, { "epoch": 0.5938748335552596, "grad_norm": 0.07208026414944517, "learning_rate": 8.471489617491812e-05, "loss": 1.0828, "step": 1115 }, { "epoch": 0.596537949400799, "grad_norm": 0.06968442188475359, "learning_rate": 8.379648324852808e-05, "loss": 1.0975, "step": 1120 }, { "epoch": 0.5992010652463382, "grad_norm": 0.07690321076998301, "learning_rate": 8.287947180117832e-05, "loss": 1.1149, "step": 1125 }, { "epoch": 0.6018641810918774, "grad_norm": 0.07088556456471248, "learning_rate": 8.196394114727585e-05, "loss": 1.1193, "step": 1130 }, { "epoch": 0.6045272969374168, "grad_norm": 0.07306182012233282, "learning_rate": 8.104997047315048e-05, "loss": 1.1222, "step": 1135 }, { "epoch": 0.607190412782956, "grad_norm": 0.0735136578466246, "learning_rate": 8.013763883020596e-05, "loss": 1.1326, "step": 1140 }, { "epoch": 0.6098535286284953, "grad_norm": 0.0792004084312011, "learning_rate": 7.92270251280823e-05, "loss": 1.1125, "step": 1145 }, { "epoch": 0.6125166444740346, "grad_norm": 0.06995636578434544, "learning_rate": 7.831820812783108e-05, "loss": 1.1397, "step": 1150 }, { "epoch": 0.6151797603195739, "grad_norm": 0.07156672219633958, "learning_rate": 7.741126643510292e-05, "loss": 1.1047, "step": 1155 }, { "epoch": 0.6178428761651131, "grad_norm": 0.06990042451203095, "learning_rate": 7.650627849334881e-05, "loss": 1.0991, "step": 1160 }, { "epoch": 0.6205059920106525, "grad_norm": 0.07175460373797926, "learning_rate": 7.560332257703555e-05, "loss": 1.1179, "step": 1165 }, { "epoch": 0.6231691078561917, "grad_norm": 0.07079860289283964, "learning_rate": 7.470247678487522e-05, "loss": 1.1179, "step": 1170 }, { "epoch": 0.625832223701731, "grad_norm": 0.06955306779443195, "learning_rate": 7.380381903307061e-05, "loss": 1.1261, "step": 1175 }, { "epoch": 0.6284953395472703, "grad_norm": 0.07040121930899221, "learning_rate": 7.290742704857585e-05, "loss": 1.128, "step": 1180 }, { "epoch": 0.6311584553928096, "grad_norm": 0.0696134794381735, "learning_rate": 7.201337836237365e-05, "loss": 1.1006, "step": 1185 }, { "epoch": 0.6338215712383488, "grad_norm": 0.0731972119965864, "learning_rate": 7.112175030276969e-05, "loss": 1.122, "step": 1190 }, { "epoch": 0.6364846870838882, "grad_norm": 0.07600883906312528, "learning_rate": 7.023261998870395e-05, "loss": 1.1054, "step": 1195 }, { "epoch": 0.6391478029294274, "grad_norm": 0.06927208348922254, "learning_rate": 6.934606432308086e-05, "loss": 1.1128, "step": 1200 }, { "epoch": 0.6418109187749668, "grad_norm": 0.06936985978073082, "learning_rate": 6.846215998611757e-05, "loss": 1.118, "step": 1205 }, { "epoch": 0.644474034620506, "grad_norm": 0.07077273319280314, "learning_rate": 6.758098342871174e-05, "loss": 1.1093, "step": 1210 }, { "epoch": 0.6471371504660453, "grad_norm": 0.07079727575348424, "learning_rate": 6.670261086582933e-05, "loss": 1.1231, "step": 1215 }, { "epoch": 0.6498002663115846, "grad_norm": 0.07181688909396321, "learning_rate": 6.582711826991226e-05, "loss": 1.1042, "step": 1220 }, { "epoch": 0.6524633821571239, "grad_norm": 0.07337324643783341, "learning_rate": 6.495458136430765e-05, "loss": 1.1042, "step": 1225 }, { "epoch": 0.6551264980026631, "grad_norm": 0.07212862237245772, "learning_rate": 6.408507561671819e-05, "loss": 1.1509, "step": 1230 }, { "epoch": 0.6577896138482024, "grad_norm": 0.07047120035609092, "learning_rate": 6.321867623267481e-05, "loss": 1.1355, "step": 1235 }, { "epoch": 0.6604527296937417, "grad_norm": 0.07459478385718604, "learning_rate": 6.23554581490321e-05, "loss": 1.1178, "step": 1240 }, { "epoch": 0.6631158455392809, "grad_norm": 0.06825726344400981, "learning_rate": 6.149549602748656e-05, "loss": 1.0862, "step": 1245 }, { "epoch": 0.6657789613848203, "grad_norm": 0.07240193352920372, "learning_rate": 6.063886424811929e-05, "loss": 1.1292, "step": 1250 }, { "epoch": 0.6684420772303595, "grad_norm": 0.07197359550672061, "learning_rate": 5.9785636902962374e-05, "loss": 1.1306, "step": 1255 }, { "epoch": 0.6711051930758988, "grad_norm": 0.07343322507680913, "learning_rate": 5.893588778959067e-05, "loss": 1.1365, "step": 1260 }, { "epoch": 0.6737683089214381, "grad_norm": 0.07842185463764602, "learning_rate": 5.8089690404738925e-05, "loss": 1.1395, "step": 1265 }, { "epoch": 0.6764314247669774, "grad_norm": 0.07270053681642126, "learning_rate": 5.7247117937944786e-05, "loss": 1.1035, "step": 1270 }, { "epoch": 0.6790945406125166, "grad_norm": 0.07124055104083139, "learning_rate": 5.640824326521841e-05, "loss": 1.1121, "step": 1275 }, { "epoch": 0.681757656458056, "grad_norm": 0.07031788560347749, "learning_rate": 5.5573138942739365e-05, "loss": 1.1192, "step": 1280 }, { "epoch": 0.6844207723035952, "grad_norm": 0.07253058219251593, "learning_rate": 5.4741877200581057e-05, "loss": 1.1324, "step": 1285 }, { "epoch": 0.6870838881491345, "grad_norm": 0.0725305584439251, "learning_rate": 5.391452993646342e-05, "loss": 1.1387, "step": 1290 }, { "epoch": 0.6897470039946738, "grad_norm": 0.07021354602161774, "learning_rate": 5.30911687095342e-05, "loss": 1.126, "step": 1295 }, { "epoch": 0.6924101198402131, "grad_norm": 0.07012940672098344, "learning_rate": 5.227186473417971e-05, "loss": 1.1486, "step": 1300 }, { "epoch": 0.6950732356857523, "grad_norm": 0.07232533360594767, "learning_rate": 5.145668887386543e-05, "loss": 1.1111, "step": 1305 }, { "epoch": 0.6977363515312917, "grad_norm": 0.07242821758103567, "learning_rate": 5.064571163500667e-05, "loss": 1.1181, "step": 1310 }, { "epoch": 0.7003994673768309, "grad_norm": 0.07148030530795384, "learning_rate": 4.983900316087051e-05, "loss": 1.0922, "step": 1315 }, { "epoch": 0.7030625832223701, "grad_norm": 0.07277623674879777, "learning_rate": 4.90366332255088e-05, "loss": 1.0985, "step": 1320 }, { "epoch": 0.7057256990679095, "grad_norm": 0.07450043090731064, "learning_rate": 4.823867122772329e-05, "loss": 1.1177, "step": 1325 }, { "epoch": 0.7083888149134487, "grad_norm": 0.07438166449706331, "learning_rate": 4.744518618506319e-05, "loss": 1.1225, "step": 1330 }, { "epoch": 0.711051930758988, "grad_norm": 0.07157294185481793, "learning_rate": 4.665624672785566e-05, "loss": 1.1291, "step": 1335 }, { "epoch": 0.7137150466045273, "grad_norm": 0.07520781395221099, "learning_rate": 4.5871921093269875e-05, "loss": 1.1082, "step": 1340 }, { "epoch": 0.7163781624500666, "grad_norm": 0.07203181974145231, "learning_rate": 4.5092277119414975e-05, "loss": 1.1333, "step": 1345 }, { "epoch": 0.7190412782956058, "grad_norm": 0.07130465564504203, "learning_rate": 4.431738223947252e-05, "loss": 1.0951, "step": 1350 }, { "epoch": 0.7217043941411452, "grad_norm": 0.075489827909183, "learning_rate": 4.35473034758643e-05, "loss": 1.1223, "step": 1355 }, { "epoch": 0.7243675099866844, "grad_norm": 0.07030155738463333, "learning_rate": 4.2782107434455054e-05, "loss": 1.1222, "step": 1360 }, { "epoch": 0.7270306258322237, "grad_norm": 0.07218531177873333, "learning_rate": 4.202186029879195e-05, "loss": 1.1135, "step": 1365 }, { "epoch": 0.729693741677763, "grad_norm": 0.07513760712282253, "learning_rate": 4.12666278243799e-05, "loss": 1.1181, "step": 1370 }, { "epoch": 0.7323568575233023, "grad_norm": 0.07435576640285345, "learning_rate": 4.0516475332994383e-05, "loss": 1.119, "step": 1375 }, { "epoch": 0.7350199733688415, "grad_norm": 0.07125188708419379, "learning_rate": 3.9771467707031615e-05, "loss": 1.1201, "step": 1380 }, { "epoch": 0.7376830892143809, "grad_norm": 0.07352772721979937, "learning_rate": 3.903166938389664e-05, "loss": 1.112, "step": 1385 }, { "epoch": 0.7403462050599201, "grad_norm": 0.07055845215545062, "learning_rate": 3.8297144350430144e-05, "loss": 1.1046, "step": 1390 }, { "epoch": 0.7430093209054593, "grad_norm": 0.07019837722638089, "learning_rate": 3.756795613737388e-05, "loss": 1.1306, "step": 1395 }, { "epoch": 0.7456724367509987, "grad_norm": 0.07239297940522006, "learning_rate": 3.684416781387589e-05, "loss": 1.1184, "step": 1400 }, { "epoch": 0.748335552596538, "grad_norm": 0.07225126596889433, "learning_rate": 3.6125841982035536e-05, "loss": 1.0843, "step": 1405 }, { "epoch": 0.7509986684420772, "grad_norm": 0.07431737331558284, "learning_rate": 3.5413040771488746e-05, "loss": 1.1145, "step": 1410 }, { "epoch": 0.7536617842876165, "grad_norm": 0.07502128782750854, "learning_rate": 3.47058258340345e-05, "loss": 1.1114, "step": 1415 }, { "epoch": 0.7563249001331558, "grad_norm": 0.07207529195587527, "learning_rate": 3.4004258338302195e-05, "loss": 1.116, "step": 1420 }, { "epoch": 0.758988015978695, "grad_norm": 0.07002467859956689, "learning_rate": 3.3308398964461206e-05, "loss": 1.1198, "step": 1425 }, { "epoch": 0.7616511318242344, "grad_norm": 0.07224058742693344, "learning_rate": 3.261830789897241e-05, "loss": 1.1367, "step": 1430 }, { "epoch": 0.7643142476697736, "grad_norm": 0.07150872607987452, "learning_rate": 3.193404482938256e-05, "loss": 1.0982, "step": 1435 }, { "epoch": 0.7669773635153129, "grad_norm": 0.07178998194161153, "learning_rate": 3.1255668939161894e-05, "loss": 1.1301, "step": 1440 }, { "epoch": 0.7696404793608522, "grad_norm": 0.07537933178179766, "learning_rate": 3.058323890258498e-05, "loss": 1.0962, "step": 1445 }, { "epoch": 0.7723035952063915, "grad_norm": 0.07189617812023931, "learning_rate": 2.9916812879655975e-05, "loss": 1.1299, "step": 1450 }, { "epoch": 0.7749667110519307, "grad_norm": 0.07135524342299995, "learning_rate": 2.925644851107835e-05, "loss": 1.1189, "step": 1455 }, { "epoch": 0.7776298268974701, "grad_norm": 0.06926273022163672, "learning_rate": 2.860220291326915e-05, "loss": 1.1068, "step": 1460 }, { "epoch": 0.7802929427430093, "grad_norm": 0.07236268331467403, "learning_rate": 2.7954132673419143e-05, "loss": 1.0981, "step": 1465 }, { "epoch": 0.7829560585885486, "grad_norm": 0.07137745473045948, "learning_rate": 2.7312293844598246e-05, "loss": 1.1045, "step": 1470 }, { "epoch": 0.7856191744340879, "grad_norm": 0.07319916214034358, "learning_rate": 2.6676741940907478e-05, "loss": 1.1281, "step": 1475 }, { "epoch": 0.7882822902796272, "grad_norm": 0.07414585611908868, "learning_rate": 2.6047531932677383e-05, "loss": 1.1225, "step": 1480 }, { "epoch": 0.7909454061251664, "grad_norm": 0.07066978423343756, "learning_rate": 2.542471824171353e-05, "loss": 1.1356, "step": 1485 }, { "epoch": 0.7936085219707057, "grad_norm": 0.07124448208473912, "learning_rate": 2.4808354736589523e-05, "loss": 1.1323, "step": 1490 }, { "epoch": 0.796271637816245, "grad_norm": 0.07192509266254882, "learning_rate": 2.419849472798761e-05, "loss": 1.1386, "step": 1495 }, { "epoch": 0.7989347536617842, "grad_norm": 0.07470536521159465, "learning_rate": 2.359519096408791e-05, "loss": 1.1103, "step": 1500 }, { "epoch": 0.8015978695073236, "grad_norm": 0.07210949951260932, "learning_rate": 2.2998495626005957e-05, "loss": 1.1108, "step": 1505 }, { "epoch": 0.8042609853528628, "grad_norm": 0.07516415250373631, "learning_rate": 2.240846032327949e-05, "loss": 1.1404, "step": 1510 }, { "epoch": 0.8069241011984021, "grad_norm": 0.07560562529629619, "learning_rate": 2.1825136089404718e-05, "loss": 1.0935, "step": 1515 }, { "epoch": 0.8095872170439414, "grad_norm": 0.07195974938474745, "learning_rate": 2.1248573377422155e-05, "loss": 1.1182, "step": 1520 }, { "epoch": 0.8122503328894807, "grad_norm": 0.07250882969384367, "learning_rate": 2.0678822055552906e-05, "loss": 1.1189, "step": 1525 }, { "epoch": 0.8149134487350199, "grad_norm": 0.0721751215640965, "learning_rate": 2.0115931402885458e-05, "loss": 1.1115, "step": 1530 }, { "epoch": 0.8175765645805593, "grad_norm": 0.0753848259347461, "learning_rate": 1.955995010511338e-05, "loss": 1.1348, "step": 1535 }, { "epoch": 0.8202396804260985, "grad_norm": 0.0719207373284397, "learning_rate": 1.901092625032448e-05, "loss": 1.1042, "step": 1540 }, { "epoch": 0.8229027962716379, "grad_norm": 0.07032664869488064, "learning_rate": 1.84689073248414e-05, "loss": 1.1009, "step": 1545 }, { "epoch": 0.8255659121171771, "grad_norm": 0.0700654057925292, "learning_rate": 1.7933940209114597e-05, "loss": 1.1269, "step": 1550 }, { "epoch": 0.8282290279627164, "grad_norm": 0.07325193867745135, "learning_rate": 1.7406071173667372e-05, "loss": 1.1138, "step": 1555 }, { "epoch": 0.8308921438082557, "grad_norm": 0.07059680065497263, "learning_rate": 1.6885345875093918e-05, "loss": 1.1202, "step": 1560 }, { "epoch": 0.833555259653795, "grad_norm": 0.06973843219886788, "learning_rate": 1.6371809352110447e-05, "loss": 1.109, "step": 1565 }, { "epoch": 0.8362183754993342, "grad_norm": 0.07028429615451927, "learning_rate": 1.5865506021659516e-05, "loss": 1.1422, "step": 1570 }, { "epoch": 0.8388814913448736, "grad_norm": 0.07070004444035057, "learning_rate": 1.5366479675068435e-05, "loss": 1.1139, "step": 1575 }, { "epoch": 0.8415446071904128, "grad_norm": 0.06902941716806588, "learning_rate": 1.4874773474261638e-05, "loss": 1.1179, "step": 1580 }, { "epoch": 0.844207723035952, "grad_norm": 0.07227429481260465, "learning_rate": 1.4390429948027428e-05, "loss": 1.1156, "step": 1585 }, { "epoch": 0.8468708388814914, "grad_norm": 0.07067313252269349, "learning_rate": 1.3913490988339718e-05, "loss": 1.1209, "step": 1590 }, { "epoch": 0.8495339547270306, "grad_norm": 0.07249077076436629, "learning_rate": 1.3443997846734535e-05, "loss": 1.1303, "step": 1595 }, { "epoch": 0.8521970705725699, "grad_norm": 0.07121930519374212, "learning_rate": 1.2981991130742211e-05, "loss": 1.1069, "step": 1600 }, { "epoch": 0.8548601864181092, "grad_norm": 0.06967637890151408, "learning_rate": 1.2527510800375043e-05, "loss": 1.1007, "step": 1605 }, { "epoch": 0.8575233022636485, "grad_norm": 0.07205061200000021, "learning_rate": 1.20805961646711e-05, "loss": 1.1199, "step": 1610 }, { "epoch": 0.8601864181091877, "grad_norm": 0.07129006954483187, "learning_rate": 1.1641285878294372e-05, "loss": 1.1054, "step": 1615 }, { "epoch": 0.8628495339547271, "grad_norm": 0.07540152645446788, "learning_rate": 1.1209617938191307e-05, "loss": 1.1032, "step": 1620 }, { "epoch": 0.8655126498002663, "grad_norm": 0.07238275344561401, "learning_rate": 1.0785629680304432e-05, "loss": 1.1246, "step": 1625 }, { "epoch": 0.8681757656458056, "grad_norm": 0.07120538294411066, "learning_rate": 1.0369357776343103e-05, "loss": 1.0932, "step": 1630 }, { "epoch": 0.8708388814913449, "grad_norm": 0.07125849884630578, "learning_rate": 9.960838230611635e-06, "loss": 1.0728, "step": 1635 }, { "epoch": 0.8735019973368842, "grad_norm": 0.07387109327159767, "learning_rate": 9.560106376895306e-06, "loss": 1.1275, "step": 1640 }, { "epoch": 0.8761651131824234, "grad_norm": 0.08867831950654811, "learning_rate": 9.167196875404094e-06, "loss": 1.1134, "step": 1645 }, { "epoch": 0.8788282290279628, "grad_norm": 0.07206488784580595, "learning_rate": 8.782143709775015e-06, "loss": 1.109, "step": 1650 }, { "epoch": 0.881491344873502, "grad_norm": 0.07037650282884113, "learning_rate": 8.40498018413266e-06, "loss": 1.0862, "step": 1655 }, { "epoch": 0.8841544607190412, "grad_norm": 0.07329640323219759, "learning_rate": 8.035738920208714e-06, "loss": 1.1539, "step": 1660 }, { "epoch": 0.8868175765645806, "grad_norm": 0.07238224996992595, "learning_rate": 7.67445185452046e-06, "loss": 1.14, "step": 1665 }, { "epoch": 0.8894806924101198, "grad_norm": 0.0709614207657161, "learning_rate": 7.321150235608399e-06, "loss": 1.1084, "step": 1670 }, { "epoch": 0.8921438082556591, "grad_norm": 0.0726275702018448, "learning_rate": 6.9758646213336165e-06, "loss": 1.1227, "step": 1675 }, { "epoch": 0.8948069241011984, "grad_norm": 0.07467788565919785, "learning_rate": 6.6386248762347004e-06, "loss": 1.1135, "step": 1680 }, { "epoch": 0.8974700399467377, "grad_norm": 0.07203922681266979, "learning_rate": 6.309460168944692e-06, "loss": 1.1071, "step": 1685 }, { "epoch": 0.9001331557922769, "grad_norm": 0.07114745434226347, "learning_rate": 5.988398969668285e-06, "loss": 1.1248, "step": 1690 }, { "epoch": 0.9027962716378163, "grad_norm": 0.07111413913413335, "learning_rate": 5.6754690477192396e-06, "loss": 1.0872, "step": 1695 }, { "epoch": 0.9054593874833555, "grad_norm": 0.07072306177993802, "learning_rate": 5.370697469118713e-06, "loss": 1.0824, "step": 1700 }, { "epoch": 0.9081225033288948, "grad_norm": 0.07185114603062198, "learning_rate": 5.074110594254133e-06, "loss": 1.107, "step": 1705 }, { "epoch": 0.9107856191744341, "grad_norm": 0.07285137048040193, "learning_rate": 4.78573407559928e-06, "loss": 1.1173, "step": 1710 }, { "epoch": 0.9134487350199734, "grad_norm": 0.07013700832744958, "learning_rate": 4.5055928554955665e-06, "loss": 1.116, "step": 1715 }, { "epoch": 0.9161118508655126, "grad_norm": 0.06933918410316768, "learning_rate": 4.233711163994669e-06, "loss": 1.1038, "step": 1720 }, { "epoch": 0.918774966711052, "grad_norm": 0.072978505928099, "learning_rate": 3.970112516762825e-06, "loss": 1.104, "step": 1725 }, { "epoch": 0.9214380825565912, "grad_norm": 0.07096666064833597, "learning_rate": 3.7148197130469576e-06, "loss": 1.1169, "step": 1730 }, { "epoch": 0.9241011984021305, "grad_norm": 0.07024364472390462, "learning_rate": 3.467854833702644e-06, "loss": 1.1051, "step": 1735 }, { "epoch": 0.9267643142476698, "grad_norm": 0.0725948329149083, "learning_rate": 3.229239239284354e-06, "loss": 1.1257, "step": 1740 }, { "epoch": 0.929427430093209, "grad_norm": 0.07215970561357568, "learning_rate": 2.9989935681979164e-06, "loss": 1.1262, "step": 1745 }, { "epoch": 0.9320905459387483, "grad_norm": 0.0737761172543898, "learning_rate": 2.777137734915403e-06, "loss": 1.1091, "step": 1750 }, { "epoch": 0.9347536617842876, "grad_norm": 0.0714550851543958, "learning_rate": 2.563690928252749e-06, "loss": 1.1283, "step": 1755 }, { "epoch": 0.9374167776298269, "grad_norm": 0.07157410935337963, "learning_rate": 2.358671609710017e-06, "loss": 1.1239, "step": 1760 }, { "epoch": 0.9400798934753661, "grad_norm": 0.0715459262435765, "learning_rate": 2.1620975118746835e-06, "loss": 1.1283, "step": 1765 }, { "epoch": 0.9427430093209055, "grad_norm": 0.07286793622379441, "learning_rate": 1.9739856368878096e-06, "loss": 1.1443, "step": 1770 }, { "epoch": 0.9454061251664447, "grad_norm": 0.07254981755374063, "learning_rate": 1.794352254973597e-06, "loss": 1.0752, "step": 1775 }, { "epoch": 0.948069241011984, "grad_norm": 0.07583062173231125, "learning_rate": 1.6232129030320453e-06, "loss": 1.1011, "step": 1780 }, { "epoch": 0.9507323568575233, "grad_norm": 0.07237886602221254, "learning_rate": 1.4605823832951948e-06, "loss": 1.1063, "step": 1785 }, { "epoch": 0.9533954727030626, "grad_norm": 0.07194863306691765, "learning_rate": 1.3064747620468054e-06, "loss": 1.0914, "step": 1790 }, { "epoch": 0.9560585885486018, "grad_norm": 0.07189535177125876, "learning_rate": 1.1609033684057857e-06, "loss": 1.1048, "step": 1795 }, { "epoch": 0.9587217043941412, "grad_norm": 0.07110020697306084, "learning_rate": 1.0238807931732487e-06, "loss": 1.1219, "step": 1800 }, { "epoch": 0.9613848202396804, "grad_norm": 0.07370460890021728, "learning_rate": 8.95418887743571e-07, "loss": 1.1317, "step": 1805 }, { "epoch": 0.9640479360852197, "grad_norm": 0.07207691237434337, "learning_rate": 7.75528763079314e-07, "loss": 1.1126, "step": 1810 }, { "epoch": 0.966711051930759, "grad_norm": 0.07440378502380679, "learning_rate": 6.642207887502027e-07, "loss": 1.1262, "step": 1815 }, { "epoch": 0.9693741677762983, "grad_norm": 0.06937782526053558, "learning_rate": 5.615045920362549e-07, "loss": 1.1191, "step": 1820 }, { "epoch": 0.9720372836218375, "grad_norm": 0.07098237504174736, "learning_rate": 4.673890570951023e-07, "loss": 1.1218, "step": 1825 }, { "epoch": 0.9747003994673769, "grad_norm": 0.07181367245928705, "learning_rate": 3.8188232419352764e-07, "loss": 1.1514, "step": 1830 }, { "epoch": 0.9773635153129161, "grad_norm": 0.07128549274991237, "learning_rate": 3.049917890034837e-07, "loss": 1.0945, "step": 1835 }, { "epoch": 0.9800266311584553, "grad_norm": 0.07161530878264005, "learning_rate": 2.3672410196232675e-07, "loss": 1.1056, "step": 1840 }, { "epoch": 0.9826897470039947, "grad_norm": 0.07090950092701657, "learning_rate": 1.7708516769769924e-07, "loss": 1.1109, "step": 1845 }, { "epoch": 0.9853528628495339, "grad_norm": 0.07140303003446424, "learning_rate": 1.2608014451672702e-07, "loss": 1.1252, "step": 1850 }, { "epoch": 0.9880159786950732, "grad_norm": 0.07103068810909154, "learning_rate": 8.371344395996516e-08, "loss": 1.1255, "step": 1855 }, { "epoch": 0.9906790945406125, "grad_norm": 0.07153307782175972, "learning_rate": 4.998873041975882e-08, "loss": 1.1365, "step": 1860 }, { "epoch": 0.9933422103861518, "grad_norm": 0.06996817906726589, "learning_rate": 2.490892082331886e-08, "loss": 1.1142, "step": 1865 }, { "epoch": 0.996005326231691, "grad_norm": 0.07126895143317796, "learning_rate": 8.476184380468155e-09, "loss": 1.1091, "step": 1870 }, { "epoch": 0.9986684420772304, "grad_norm": 0.0706408297239244, "learning_rate": 6.919423959805826e-10, "loss": 1.1206, "step": 1875 }, { "epoch": 0.9997336884154461, "eval_loss": 1.118857979774475, "eval_runtime": 1652.5253, "eval_samples_per_second": 8.045, "eval_steps_per_second": 0.503, "step": 1877 }, { "epoch": 0.9997336884154461, "step": 1877, "total_flos": 2.979798729936077e+16, "train_loss": 1.1429596533467938, "train_runtime": 55739.5179, "train_samples_per_second": 2.156, "train_steps_per_second": 0.034 } ], "logging_steps": 5, "max_steps": 1877, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.979798729936077e+16, "train_batch_size": 8, "trial_name": null, "trial_params": null }