{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.5780730897009967, "eval_steps": 500, "global_step": 950, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0016611295681063123, "grad_norm": 69.42809295654297, "learning_rate": 5.4054054054054056e-08, "loss": 0.7218, "step": 1 }, { "epoch": 0.0033222591362126247, "grad_norm": 84.8204345703125, "learning_rate": 1.0810810810810811e-07, "loss": 0.7005, "step": 2 }, { "epoch": 0.0049833887043189366, "grad_norm": 62.26895523071289, "learning_rate": 1.6216216216216215e-07, "loss": 0.6448, "step": 3 }, { "epoch": 0.006644518272425249, "grad_norm": 77.79634094238281, "learning_rate": 2.1621621621621622e-07, "loss": 0.6623, "step": 4 }, { "epoch": 0.008305647840531562, "grad_norm": 76.27490234375, "learning_rate": 2.702702702702703e-07, "loss": 0.599, "step": 5 }, { "epoch": 0.009966777408637873, "grad_norm": 76.31900787353516, "learning_rate": 3.243243243243243e-07, "loss": 0.6139, "step": 6 }, { "epoch": 0.011627906976744186, "grad_norm": 72.4638442993164, "learning_rate": 3.783783783783784e-07, "loss": 0.6561, "step": 7 }, { "epoch": 0.013289036544850499, "grad_norm": 45.92433547973633, "learning_rate": 4.3243243243243244e-07, "loss": 0.5786, "step": 8 }, { "epoch": 0.014950166112956811, "grad_norm": 55.887176513671875, "learning_rate": 4.864864864864865e-07, "loss": 0.568, "step": 9 }, { "epoch": 0.016611295681063124, "grad_norm": 34.738040924072266, "learning_rate": 5.405405405405406e-07, "loss": 0.5106, "step": 10 }, { "epoch": 0.018272425249169437, "grad_norm": 23.600587844848633, "learning_rate": 5.945945945945947e-07, "loss": 0.4439, "step": 11 }, { "epoch": 0.019933554817275746, "grad_norm": 22.552448272705078, "learning_rate": 6.486486486486486e-07, "loss": 0.5118, "step": 12 }, { "epoch": 0.02159468438538206, "grad_norm": 17.616506576538086, "learning_rate": 7.027027027027027e-07, "loss": 0.4026, "step": 13 }, { "epoch": 0.023255813953488372, "grad_norm": 21.766603469848633, "learning_rate": 7.567567567567568e-07, "loss": 0.4315, "step": 14 }, { "epoch": 0.024916943521594685, "grad_norm": 21.652666091918945, "learning_rate": 8.108108108108108e-07, "loss": 0.3857, "step": 15 }, { "epoch": 0.026578073089700997, "grad_norm": 29.43855094909668, "learning_rate": 8.648648648648649e-07, "loss": 0.5083, "step": 16 }, { "epoch": 0.02823920265780731, "grad_norm": 27.099016189575195, "learning_rate": 9.18918918918919e-07, "loss": 0.3577, "step": 17 }, { "epoch": 0.029900332225913623, "grad_norm": 33.78343963623047, "learning_rate": 9.72972972972973e-07, "loss": 0.3735, "step": 18 }, { "epoch": 0.03156146179401993, "grad_norm": 37.29399871826172, "learning_rate": 1.0270270270270269e-06, "loss": 0.4251, "step": 19 }, { "epoch": 0.03322259136212625, "grad_norm": 28.423860549926758, "learning_rate": 1.0810810810810812e-06, "loss": 0.4173, "step": 20 }, { "epoch": 0.03488372093023256, "grad_norm": 27.742494583129883, "learning_rate": 1.135135135135135e-06, "loss": 0.3403, "step": 21 }, { "epoch": 0.036544850498338874, "grad_norm": 25.273988723754883, "learning_rate": 1.1891891891891893e-06, "loss": 0.3046, "step": 22 }, { "epoch": 0.03820598006644518, "grad_norm": 29.61273193359375, "learning_rate": 1.2432432432432432e-06, "loss": 0.3323, "step": 23 }, { "epoch": 0.03986710963455149, "grad_norm": 24.798965454101562, "learning_rate": 1.2972972972972972e-06, "loss": 0.2518, "step": 24 }, { "epoch": 0.04152823920265781, "grad_norm": 25.87079620361328, "learning_rate": 1.3513513513513513e-06, "loss": 0.2818, "step": 25 }, { "epoch": 0.04318936877076412, "grad_norm": 32.44166564941406, "learning_rate": 1.4054054054054054e-06, "loss": 0.3059, "step": 26 }, { "epoch": 0.044850498338870434, "grad_norm": 40.5885124206543, "learning_rate": 1.4594594594594594e-06, "loss": 0.3353, "step": 27 }, { "epoch": 0.046511627906976744, "grad_norm": 29.34911346435547, "learning_rate": 1.5135135135135135e-06, "loss": 0.2943, "step": 28 }, { "epoch": 0.04817275747508306, "grad_norm": 34.42272186279297, "learning_rate": 1.5675675675675676e-06, "loss": 0.2938, "step": 29 }, { "epoch": 0.04983388704318937, "grad_norm": 32.65325927734375, "learning_rate": 1.6216216216216216e-06, "loss": 0.3359, "step": 30 }, { "epoch": 0.05149501661129568, "grad_norm": 38.35554504394531, "learning_rate": 1.6756756756756755e-06, "loss": 0.3983, "step": 31 }, { "epoch": 0.053156146179401995, "grad_norm": 27.567651748657227, "learning_rate": 1.7297297297297298e-06, "loss": 0.3107, "step": 32 }, { "epoch": 0.054817275747508304, "grad_norm": 26.41943359375, "learning_rate": 1.7837837837837836e-06, "loss": 0.3461, "step": 33 }, { "epoch": 0.05647840531561462, "grad_norm": 20.003856658935547, "learning_rate": 1.837837837837838e-06, "loss": 0.3131, "step": 34 }, { "epoch": 0.05813953488372093, "grad_norm": 17.432525634765625, "learning_rate": 1.8918918918918918e-06, "loss": 0.2034, "step": 35 }, { "epoch": 0.059800664451827246, "grad_norm": 18.504539489746094, "learning_rate": 1.945945945945946e-06, "loss": 0.2837, "step": 36 }, { "epoch": 0.061461794019933555, "grad_norm": 15.921092987060547, "learning_rate": 2e-06, "loss": 0.2069, "step": 37 }, { "epoch": 0.06312292358803986, "grad_norm": 17.68560791015625, "learning_rate": 1.999996376504091e-06, "loss": 0.2074, "step": 38 }, { "epoch": 0.06478405315614617, "grad_norm": 18.591819763183594, "learning_rate": 1.9999855060426223e-06, "loss": 0.3209, "step": 39 }, { "epoch": 0.0664451827242525, "grad_norm": 21.39113998413086, "learning_rate": 1.9999673886943732e-06, "loss": 0.2902, "step": 40 }, { "epoch": 0.0681063122923588, "grad_norm": 29.161853790283203, "learning_rate": 1.9999420245906396e-06, "loss": 0.306, "step": 41 }, { "epoch": 0.06976744186046512, "grad_norm": 12.444762229919434, "learning_rate": 1.9999094139152346e-06, "loss": 0.2314, "step": 42 }, { "epoch": 0.07142857142857142, "grad_norm": 25.82412338256836, "learning_rate": 1.999869556904488e-06, "loss": 0.3418, "step": 43 }, { "epoch": 0.07308970099667775, "grad_norm": 18.117881774902344, "learning_rate": 1.9998224538472425e-06, "loss": 0.2559, "step": 44 }, { "epoch": 0.07475083056478406, "grad_norm": 17.093626022338867, "learning_rate": 1.999768105084854e-06, "loss": 0.2327, "step": 45 }, { "epoch": 0.07641196013289037, "grad_norm": 19.649778366088867, "learning_rate": 1.999706511011188e-06, "loss": 0.2659, "step": 46 }, { "epoch": 0.07807308970099668, "grad_norm": 20.985107421875, "learning_rate": 1.999637672072616e-06, "loss": 0.3078, "step": 47 }, { "epoch": 0.07973421926910298, "grad_norm": 14.487998962402344, "learning_rate": 1.9995615887680127e-06, "loss": 0.2029, "step": 48 }, { "epoch": 0.08139534883720931, "grad_norm": 12.239130020141602, "learning_rate": 1.9994782616487534e-06, "loss": 0.2548, "step": 49 }, { "epoch": 0.08305647840531562, "grad_norm": 12.82898998260498, "learning_rate": 1.9993876913187095e-06, "loss": 0.2234, "step": 50 }, { "epoch": 0.08471760797342193, "grad_norm": 14.681973457336426, "learning_rate": 1.9992898784342433e-06, "loss": 0.2229, "step": 51 }, { "epoch": 0.08637873754152824, "grad_norm": 19.037193298339844, "learning_rate": 1.9991848237042032e-06, "loss": 0.2545, "step": 52 }, { "epoch": 0.08803986710963455, "grad_norm": 12.438526153564453, "learning_rate": 1.999072527889921e-06, "loss": 0.2018, "step": 53 }, { "epoch": 0.08970099667774087, "grad_norm": 13.439962387084961, "learning_rate": 1.9989529918052027e-06, "loss": 0.1708, "step": 54 }, { "epoch": 0.09136212624584718, "grad_norm": 17.36454200744629, "learning_rate": 1.998826216316326e-06, "loss": 0.265, "step": 55 }, { "epoch": 0.09302325581395349, "grad_norm": 22.255950927734375, "learning_rate": 1.998692202342032e-06, "loss": 0.3088, "step": 56 }, { "epoch": 0.0946843853820598, "grad_norm": 16.178789138793945, "learning_rate": 1.998550950853518e-06, "loss": 0.2184, "step": 57 }, { "epoch": 0.09634551495016612, "grad_norm": 12.416084289550781, "learning_rate": 1.998402462874433e-06, "loss": 0.1486, "step": 58 }, { "epoch": 0.09800664451827243, "grad_norm": 14.603720664978027, "learning_rate": 1.9982467394808674e-06, "loss": 0.2007, "step": 59 }, { "epoch": 0.09966777408637874, "grad_norm": 15.751521110534668, "learning_rate": 1.9980837818013486e-06, "loss": 0.261, "step": 60 }, { "epoch": 0.10132890365448505, "grad_norm": 11.762495040893555, "learning_rate": 1.9979135910168287e-06, "loss": 0.1878, "step": 61 }, { "epoch": 0.10299003322259136, "grad_norm": 11.098094940185547, "learning_rate": 1.997736168360679e-06, "loss": 0.1474, "step": 62 }, { "epoch": 0.10465116279069768, "grad_norm": 17.337011337280273, "learning_rate": 1.9975515151186805e-06, "loss": 0.2041, "step": 63 }, { "epoch": 0.10631229235880399, "grad_norm": 25.151264190673828, "learning_rate": 1.9973596326290133e-06, "loss": 0.254, "step": 64 }, { "epoch": 0.1079734219269103, "grad_norm": 17.115270614624023, "learning_rate": 1.9971605222822486e-06, "loss": 0.1631, "step": 65 }, { "epoch": 0.10963455149501661, "grad_norm": 11.817288398742676, "learning_rate": 1.996954185521337e-06, "loss": 0.1729, "step": 66 }, { "epoch": 0.11129568106312292, "grad_norm": 14.577704429626465, "learning_rate": 1.9967406238415996e-06, "loss": 0.1962, "step": 67 }, { "epoch": 0.11295681063122924, "grad_norm": 18.282690048217773, "learning_rate": 1.996519838790716e-06, "loss": 0.2224, "step": 68 }, { "epoch": 0.11461794019933555, "grad_norm": 14.57044792175293, "learning_rate": 1.996291831968714e-06, "loss": 0.149, "step": 69 }, { "epoch": 0.11627906976744186, "grad_norm": 19.73041343688965, "learning_rate": 1.996056605027956e-06, "loss": 0.2655, "step": 70 }, { "epoch": 0.11794019933554817, "grad_norm": 18.137643814086914, "learning_rate": 1.995814159673132e-06, "loss": 0.229, "step": 71 }, { "epoch": 0.11960132890365449, "grad_norm": 15.201904296875, "learning_rate": 1.995564497661239e-06, "loss": 0.1941, "step": 72 }, { "epoch": 0.1212624584717608, "grad_norm": 15.047739028930664, "learning_rate": 1.9953076208015772e-06, "loss": 0.1623, "step": 73 }, { "epoch": 0.12292358803986711, "grad_norm": 16.15547752380371, "learning_rate": 1.9950435309557302e-06, "loss": 0.1894, "step": 74 }, { "epoch": 0.12458471760797342, "grad_norm": 16.512256622314453, "learning_rate": 1.994772230037556e-06, "loss": 0.1993, "step": 75 }, { "epoch": 0.12624584717607973, "grad_norm": 14.934798240661621, "learning_rate": 1.994493720013169e-06, "loss": 0.1411, "step": 76 }, { "epoch": 0.12790697674418605, "grad_norm": 14.899330139160156, "learning_rate": 1.9942080029009296e-06, "loss": 0.185, "step": 77 }, { "epoch": 0.12956810631229235, "grad_norm": 17.317094802856445, "learning_rate": 1.993915080771427e-06, "loss": 0.1947, "step": 78 }, { "epoch": 0.13122923588039867, "grad_norm": 12.058066368103027, "learning_rate": 1.9936149557474663e-06, "loss": 0.1272, "step": 79 }, { "epoch": 0.132890365448505, "grad_norm": 12.605445861816406, "learning_rate": 1.9933076300040505e-06, "loss": 0.1624, "step": 80 }, { "epoch": 0.1345514950166113, "grad_norm": 15.506592750549316, "learning_rate": 1.9929931057683666e-06, "loss": 0.1833, "step": 81 }, { "epoch": 0.1362126245847176, "grad_norm": 11.711981773376465, "learning_rate": 1.9926713853197696e-06, "loss": 0.165, "step": 82 }, { "epoch": 0.1378737541528239, "grad_norm": 13.009001731872559, "learning_rate": 1.9923424709897644e-06, "loss": 0.2145, "step": 83 }, { "epoch": 0.13953488372093023, "grad_norm": 14.122442245483398, "learning_rate": 1.992006365161991e-06, "loss": 0.1774, "step": 84 }, { "epoch": 0.14119601328903655, "grad_norm": 9.544992446899414, "learning_rate": 1.991663070272206e-06, "loss": 0.1555, "step": 85 }, { "epoch": 0.14285714285714285, "grad_norm": 15.784918785095215, "learning_rate": 1.9913125888082632e-06, "loss": 0.2485, "step": 86 }, { "epoch": 0.14451827242524917, "grad_norm": 14.848572731018066, "learning_rate": 1.9909549233100998e-06, "loss": 0.1986, "step": 87 }, { "epoch": 0.1461794019933555, "grad_norm": 19.732912063598633, "learning_rate": 1.990590076369715e-06, "loss": 0.2783, "step": 88 }, { "epoch": 0.1478405315614618, "grad_norm": 10.916641235351562, "learning_rate": 1.9902180506311514e-06, "loss": 0.1609, "step": 89 }, { "epoch": 0.14950166112956811, "grad_norm": 7.3824782371521, "learning_rate": 1.9898388487904764e-06, "loss": 0.0938, "step": 90 }, { "epoch": 0.1511627906976744, "grad_norm": 20.201457977294922, "learning_rate": 1.989452473595762e-06, "loss": 0.2579, "step": 91 }, { "epoch": 0.15282392026578073, "grad_norm": 11.529504776000977, "learning_rate": 1.989058927847067e-06, "loss": 0.1761, "step": 92 }, { "epoch": 0.15448504983388706, "grad_norm": 12.745220184326172, "learning_rate": 1.9886582143964143e-06, "loss": 0.1881, "step": 93 }, { "epoch": 0.15614617940199335, "grad_norm": 8.80517292022705, "learning_rate": 1.9882503361477703e-06, "loss": 0.1003, "step": 94 }, { "epoch": 0.15780730897009967, "grad_norm": 16.047895431518555, "learning_rate": 1.9878352960570256e-06, "loss": 0.1476, "step": 95 }, { "epoch": 0.15946843853820597, "grad_norm": 13.64163875579834, "learning_rate": 1.987413097131972e-06, "loss": 0.2103, "step": 96 }, { "epoch": 0.1611295681063123, "grad_norm": 10.822211265563965, "learning_rate": 1.9869837424322827e-06, "loss": 0.1423, "step": 97 }, { "epoch": 0.16279069767441862, "grad_norm": 11.291969299316406, "learning_rate": 1.9865472350694867e-06, "loss": 0.1375, "step": 98 }, { "epoch": 0.1644518272425249, "grad_norm": 10.820338249206543, "learning_rate": 1.9861035782069496e-06, "loss": 0.18, "step": 99 }, { "epoch": 0.16611295681063123, "grad_norm": 12.342602729797363, "learning_rate": 1.985652775059849e-06, "loss": 0.123, "step": 100 }, { "epoch": 0.16777408637873753, "grad_norm": 11.896181106567383, "learning_rate": 1.985194828895152e-06, "loss": 0.1846, "step": 101 }, { "epoch": 0.16943521594684385, "grad_norm": 12.606119155883789, "learning_rate": 1.9847297430315903e-06, "loss": 0.1322, "step": 102 }, { "epoch": 0.17109634551495018, "grad_norm": 10.485321044921875, "learning_rate": 1.984257520839637e-06, "loss": 0.1443, "step": 103 }, { "epoch": 0.17275747508305647, "grad_norm": 11.102760314941406, "learning_rate": 1.983778165741483e-06, "loss": 0.1322, "step": 104 }, { "epoch": 0.1744186046511628, "grad_norm": 12.000479698181152, "learning_rate": 1.983291681211011e-06, "loss": 0.1558, "step": 105 }, { "epoch": 0.1760797342192691, "grad_norm": 15.679694175720215, "learning_rate": 1.98279807077377e-06, "loss": 0.1671, "step": 106 }, { "epoch": 0.1777408637873754, "grad_norm": 14.487334251403809, "learning_rate": 1.9822973380069507e-06, "loss": 0.1675, "step": 107 }, { "epoch": 0.17940199335548174, "grad_norm": 12.584924697875977, "learning_rate": 1.9817894865393597e-06, "loss": 0.1733, "step": 108 }, { "epoch": 0.18106312292358803, "grad_norm": 8.71336841583252, "learning_rate": 1.9812745200513923e-06, "loss": 0.1175, "step": 109 }, { "epoch": 0.18272425249169436, "grad_norm": 14.50503158569336, "learning_rate": 1.9807524422750064e-06, "loss": 0.1571, "step": 110 }, { "epoch": 0.18438538205980065, "grad_norm": 12.778355598449707, "learning_rate": 1.9802232569936956e-06, "loss": 0.1631, "step": 111 }, { "epoch": 0.18604651162790697, "grad_norm": 14.556724548339844, "learning_rate": 1.979686968042461e-06, "loss": 0.2194, "step": 112 }, { "epoch": 0.1877076411960133, "grad_norm": 13.824010848999023, "learning_rate": 1.979143579307784e-06, "loss": 0.1673, "step": 113 }, { "epoch": 0.1893687707641196, "grad_norm": 11.80858325958252, "learning_rate": 1.9785930947275985e-06, "loss": 0.1661, "step": 114 }, { "epoch": 0.19102990033222592, "grad_norm": 15.89432144165039, "learning_rate": 1.9780355182912623e-06, "loss": 0.2334, "step": 115 }, { "epoch": 0.19269102990033224, "grad_norm": 11.047170639038086, "learning_rate": 1.977470854039527e-06, "loss": 0.1506, "step": 116 }, { "epoch": 0.19435215946843853, "grad_norm": 12.432543754577637, "learning_rate": 1.9768991060645096e-06, "loss": 0.1775, "step": 117 }, { "epoch": 0.19601328903654486, "grad_norm": 9.709572792053223, "learning_rate": 1.976320278509663e-06, "loss": 0.1365, "step": 118 }, { "epoch": 0.19767441860465115, "grad_norm": 11.285073280334473, "learning_rate": 1.9757343755697456e-06, "loss": 0.1765, "step": 119 }, { "epoch": 0.19933554817275748, "grad_norm": 10.247356414794922, "learning_rate": 1.9751414014907913e-06, "loss": 0.1558, "step": 120 }, { "epoch": 0.2009966777408638, "grad_norm": 14.953573226928711, "learning_rate": 1.974541360570079e-06, "loss": 0.1802, "step": 121 }, { "epoch": 0.2026578073089701, "grad_norm": 14.226633071899414, "learning_rate": 1.9739342571560996e-06, "loss": 0.1492, "step": 122 }, { "epoch": 0.20431893687707642, "grad_norm": 9.25733757019043, "learning_rate": 1.973320095648527e-06, "loss": 0.1563, "step": 123 }, { "epoch": 0.2059800664451827, "grad_norm": 11.132588386535645, "learning_rate": 1.9726988804981845e-06, "loss": 0.1634, "step": 124 }, { "epoch": 0.20764119601328904, "grad_norm": 9.98374080657959, "learning_rate": 1.972070616207013e-06, "loss": 0.16, "step": 125 }, { "epoch": 0.20930232558139536, "grad_norm": 14.084647178649902, "learning_rate": 1.971435307328039e-06, "loss": 0.2341, "step": 126 }, { "epoch": 0.21096345514950166, "grad_norm": 10.617352485656738, "learning_rate": 1.9707929584653408e-06, "loss": 0.1376, "step": 127 }, { "epoch": 0.21262458471760798, "grad_norm": 12.762201309204102, "learning_rate": 1.9701435742740146e-06, "loss": 0.1543, "step": 128 }, { "epoch": 0.21428571428571427, "grad_norm": 11.197344779968262, "learning_rate": 1.9694871594601435e-06, "loss": 0.131, "step": 129 }, { "epoch": 0.2159468438538206, "grad_norm": 24.01343536376953, "learning_rate": 1.9688237187807594e-06, "loss": 0.2743, "step": 130 }, { "epoch": 0.21760797342192692, "grad_norm": 11.367233276367188, "learning_rate": 1.9681532570438117e-06, "loss": 0.1446, "step": 131 }, { "epoch": 0.21926910299003322, "grad_norm": 10.252062797546387, "learning_rate": 1.967475779108131e-06, "loss": 0.1093, "step": 132 }, { "epoch": 0.22093023255813954, "grad_norm": 16.38922691345215, "learning_rate": 1.9667912898833952e-06, "loss": 0.1282, "step": 133 }, { "epoch": 0.22259136212624583, "grad_norm": 14.078662872314453, "learning_rate": 1.966099794330091e-06, "loss": 0.1436, "step": 134 }, { "epoch": 0.22425249169435216, "grad_norm": 12.31657600402832, "learning_rate": 1.9654012974594813e-06, "loss": 0.1442, "step": 135 }, { "epoch": 0.22591362126245848, "grad_norm": 16.30408477783203, "learning_rate": 1.9646958043335675e-06, "loss": 0.1845, "step": 136 }, { "epoch": 0.22757475083056478, "grad_norm": 18.23689842224121, "learning_rate": 1.9639833200650524e-06, "loss": 0.2381, "step": 137 }, { "epoch": 0.2292358803986711, "grad_norm": 14.732518196105957, "learning_rate": 1.9632638498173037e-06, "loss": 0.1844, "step": 138 }, { "epoch": 0.23089700996677742, "grad_norm": 8.49440860748291, "learning_rate": 1.9625373988043164e-06, "loss": 0.1314, "step": 139 }, { "epoch": 0.23255813953488372, "grad_norm": 15.13712215423584, "learning_rate": 1.9618039722906746e-06, "loss": 0.1931, "step": 140 }, { "epoch": 0.23421926910299004, "grad_norm": 18.702835083007812, "learning_rate": 1.961063575591515e-06, "loss": 0.1516, "step": 141 }, { "epoch": 0.23588039867109634, "grad_norm": 6.8990559577941895, "learning_rate": 1.960316214072486e-06, "loss": 0.0791, "step": 142 }, { "epoch": 0.23754152823920266, "grad_norm": 11.587632179260254, "learning_rate": 1.9595618931497105e-06, "loss": 0.1103, "step": 143 }, { "epoch": 0.23920265780730898, "grad_norm": 20.253131866455078, "learning_rate": 1.9588006182897455e-06, "loss": 0.1988, "step": 144 }, { "epoch": 0.24086378737541528, "grad_norm": 15.596393585205078, "learning_rate": 1.958032395009545e-06, "loss": 0.1549, "step": 145 }, { "epoch": 0.2425249169435216, "grad_norm": 10.316424369812012, "learning_rate": 1.9572572288764154e-06, "loss": 0.1564, "step": 146 }, { "epoch": 0.2441860465116279, "grad_norm": 9.298800468444824, "learning_rate": 1.95647512550798e-06, "loss": 0.1488, "step": 147 }, { "epoch": 0.24584717607973422, "grad_norm": 15.457275390625, "learning_rate": 1.955686090572136e-06, "loss": 0.2725, "step": 148 }, { "epoch": 0.24750830564784054, "grad_norm": 13.5877685546875, "learning_rate": 1.9548901297870124e-06, "loss": 0.2401, "step": 149 }, { "epoch": 0.24916943521594684, "grad_norm": 10.776034355163574, "learning_rate": 1.95408724892093e-06, "loss": 0.149, "step": 150 }, { "epoch": 0.25083056478405313, "grad_norm": 13.359519958496094, "learning_rate": 1.9532774537923615e-06, "loss": 0.1533, "step": 151 }, { "epoch": 0.25249169435215946, "grad_norm": 10.034915924072266, "learning_rate": 1.9524607502698843e-06, "loss": 0.1987, "step": 152 }, { "epoch": 0.2541528239202658, "grad_norm": 11.459542274475098, "learning_rate": 1.9516371442721425e-06, "loss": 0.1631, "step": 153 }, { "epoch": 0.2558139534883721, "grad_norm": 9.19011402130127, "learning_rate": 1.9508066417678018e-06, "loss": 0.1789, "step": 154 }, { "epoch": 0.2574750830564784, "grad_norm": 9.150303840637207, "learning_rate": 1.9499692487755076e-06, "loss": 0.1464, "step": 155 }, { "epoch": 0.2591362126245847, "grad_norm": 10.107464790344238, "learning_rate": 1.9491249713638394e-06, "loss": 0.1705, "step": 156 }, { "epoch": 0.260797342192691, "grad_norm": 7.740807056427002, "learning_rate": 1.948273815651269e-06, "loss": 0.1257, "step": 157 }, { "epoch": 0.26245847176079734, "grad_norm": 8.180129051208496, "learning_rate": 1.947415787806115e-06, "loss": 0.09, "step": 158 }, { "epoch": 0.26411960132890366, "grad_norm": 12.505861282348633, "learning_rate": 1.946550894046498e-06, "loss": 0.1999, "step": 159 }, { "epoch": 0.26578073089701, "grad_norm": 16.3990535736084, "learning_rate": 1.9456791406402963e-06, "loss": 0.2219, "step": 160 }, { "epoch": 0.26744186046511625, "grad_norm": 23.716087341308594, "learning_rate": 1.944800533905099e-06, "loss": 0.2294, "step": 161 }, { "epoch": 0.2691029900332226, "grad_norm": 23.61831283569336, "learning_rate": 1.943915080208163e-06, "loss": 0.2393, "step": 162 }, { "epoch": 0.2707641196013289, "grad_norm": 17.86264991760254, "learning_rate": 1.943022785966363e-06, "loss": 0.1994, "step": 163 }, { "epoch": 0.2724252491694352, "grad_norm": 10.839683532714844, "learning_rate": 1.9421236576461487e-06, "loss": 0.0614, "step": 164 }, { "epoch": 0.27408637873754155, "grad_norm": 10.387090682983398, "learning_rate": 1.941217701763495e-06, "loss": 0.1148, "step": 165 }, { "epoch": 0.2757475083056478, "grad_norm": 22.82048797607422, "learning_rate": 1.9403049248838576e-06, "loss": 0.2027, "step": 166 }, { "epoch": 0.27740863787375414, "grad_norm": 12.700292587280273, "learning_rate": 1.9393853336221225e-06, "loss": 0.1516, "step": 167 }, { "epoch": 0.27906976744186046, "grad_norm": 16.656648635864258, "learning_rate": 1.9384589346425605e-06, "loss": 0.2115, "step": 168 }, { "epoch": 0.2807308970099668, "grad_norm": 20.727684020996094, "learning_rate": 1.9375257346587774e-06, "loss": 0.1832, "step": 169 }, { "epoch": 0.2823920265780731, "grad_norm": 12.764060020446777, "learning_rate": 1.936585740433665e-06, "loss": 0.1757, "step": 170 }, { "epoch": 0.2840531561461794, "grad_norm": 17.68997573852539, "learning_rate": 1.9356389587793544e-06, "loss": 0.1494, "step": 171 }, { "epoch": 0.2857142857142857, "grad_norm": 11.008881568908691, "learning_rate": 1.934685396557165e-06, "loss": 0.1242, "step": 172 }, { "epoch": 0.287375415282392, "grad_norm": 8.055181503295898, "learning_rate": 1.9337250606775536e-06, "loss": 0.1221, "step": 173 }, { "epoch": 0.28903654485049834, "grad_norm": 9.147614479064941, "learning_rate": 1.9327579581000665e-06, "loss": 0.1163, "step": 174 }, { "epoch": 0.29069767441860467, "grad_norm": 9.311761856079102, "learning_rate": 1.931784095833289e-06, "loss": 0.1026, "step": 175 }, { "epoch": 0.292358803986711, "grad_norm": 7.291942119598389, "learning_rate": 1.930803480934792e-06, "loss": 0.1058, "step": 176 }, { "epoch": 0.29401993355481726, "grad_norm": 10.639616012573242, "learning_rate": 1.9298161205110838e-06, "loss": 0.1921, "step": 177 }, { "epoch": 0.2956810631229236, "grad_norm": 12.382287979125977, "learning_rate": 1.9288220217175583e-06, "loss": 0.1336, "step": 178 }, { "epoch": 0.2973421926910299, "grad_norm": 9.517138481140137, "learning_rate": 1.92782119175844e-06, "loss": 0.1173, "step": 179 }, { "epoch": 0.29900332225913623, "grad_norm": 10.612733840942383, "learning_rate": 1.9268136378867365e-06, "loss": 0.1215, "step": 180 }, { "epoch": 0.30066445182724255, "grad_norm": 11.6936616897583, "learning_rate": 1.9257993674041814e-06, "loss": 0.1511, "step": 181 }, { "epoch": 0.3023255813953488, "grad_norm": 9.376453399658203, "learning_rate": 1.9247783876611857e-06, "loss": 0.0859, "step": 182 }, { "epoch": 0.30398671096345514, "grad_norm": 15.27945327758789, "learning_rate": 1.9237507060567802e-06, "loss": 0.2082, "step": 183 }, { "epoch": 0.30564784053156147, "grad_norm": 13.213935852050781, "learning_rate": 1.9227163300385662e-06, "loss": 0.1624, "step": 184 }, { "epoch": 0.3073089700996678, "grad_norm": 14.251786231994629, "learning_rate": 1.921675267102657e-06, "loss": 0.1719, "step": 185 }, { "epoch": 0.3089700996677741, "grad_norm": 21.97607421875, "learning_rate": 1.920627524793628e-06, "loss": 0.2718, "step": 186 }, { "epoch": 0.3106312292358804, "grad_norm": 12.712850570678711, "learning_rate": 1.9195731107044594e-06, "loss": 0.1694, "step": 187 }, { "epoch": 0.3122923588039867, "grad_norm": 7.244543075561523, "learning_rate": 1.9185120324764805e-06, "loss": 0.1159, "step": 188 }, { "epoch": 0.313953488372093, "grad_norm": 8.671590805053711, "learning_rate": 1.917444297799317e-06, "loss": 0.1121, "step": 189 }, { "epoch": 0.31561461794019935, "grad_norm": 8.755648612976074, "learning_rate": 1.916369914410834e-06, "loss": 0.1329, "step": 190 }, { "epoch": 0.31727574750830567, "grad_norm": 9.643033027648926, "learning_rate": 1.9152888900970783e-06, "loss": 0.1697, "step": 191 }, { "epoch": 0.31893687707641194, "grad_norm": 10.84849739074707, "learning_rate": 1.9142012326922246e-06, "loss": 0.1537, "step": 192 }, { "epoch": 0.32059800664451826, "grad_norm": 8.020485877990723, "learning_rate": 1.913106950078517e-06, "loss": 0.0776, "step": 193 }, { "epoch": 0.3222591362126246, "grad_norm": 14.385054588317871, "learning_rate": 1.9120060501862128e-06, "loss": 0.1708, "step": 194 }, { "epoch": 0.3239202657807309, "grad_norm": 12.023110389709473, "learning_rate": 1.9108985409935247e-06, "loss": 0.14, "step": 195 }, { "epoch": 0.32558139534883723, "grad_norm": 20.190410614013672, "learning_rate": 1.9097844305265622e-06, "loss": 0.1881, "step": 196 }, { "epoch": 0.3272425249169435, "grad_norm": 13.03503704071045, "learning_rate": 1.908663726859275e-06, "loss": 0.1258, "step": 197 }, { "epoch": 0.3289036544850498, "grad_norm": 12.031476020812988, "learning_rate": 1.907536438113394e-06, "loss": 0.1018, "step": 198 }, { "epoch": 0.33056478405315615, "grad_norm": 12.906877517700195, "learning_rate": 1.9064025724583707e-06, "loss": 0.1026, "step": 199 }, { "epoch": 0.33222591362126247, "grad_norm": 12.325488090515137, "learning_rate": 1.905262138111321e-06, "loss": 0.1089, "step": 200 }, { "epoch": 0.3338870431893688, "grad_norm": 9.743961334228516, "learning_rate": 1.9041151433369623e-06, "loss": 0.0776, "step": 201 }, { "epoch": 0.33554817275747506, "grad_norm": 22.17953109741211, "learning_rate": 1.902961596447557e-06, "loss": 0.2423, "step": 202 }, { "epoch": 0.3372093023255814, "grad_norm": 8.794547080993652, "learning_rate": 1.90180150580285e-06, "loss": 0.1216, "step": 203 }, { "epoch": 0.3388704318936877, "grad_norm": 8.961956977844238, "learning_rate": 1.9006348798100086e-06, "loss": 0.1306, "step": 204 }, { "epoch": 0.34053156146179403, "grad_norm": 12.606147766113281, "learning_rate": 1.8994617269235614e-06, "loss": 0.1489, "step": 205 }, { "epoch": 0.34219269102990035, "grad_norm": 15.009194374084473, "learning_rate": 1.8982820556453384e-06, "loss": 0.1878, "step": 206 }, { "epoch": 0.3438538205980066, "grad_norm": 8.010025024414062, "learning_rate": 1.8970958745244075e-06, "loss": 0.1143, "step": 207 }, { "epoch": 0.34551495016611294, "grad_norm": 9.553529739379883, "learning_rate": 1.8959031921570135e-06, "loss": 0.1271, "step": 208 }, { "epoch": 0.34717607973421927, "grad_norm": 7.424415588378906, "learning_rate": 1.8947040171865157e-06, "loss": 0.0986, "step": 209 }, { "epoch": 0.3488372093023256, "grad_norm": 7.113452434539795, "learning_rate": 1.8934983583033252e-06, "loss": 0.1211, "step": 210 }, { "epoch": 0.3504983388704319, "grad_norm": 8.20436954498291, "learning_rate": 1.8922862242448428e-06, "loss": 0.1339, "step": 211 }, { "epoch": 0.3521594684385382, "grad_norm": 11.628143310546875, "learning_rate": 1.8910676237953931e-06, "loss": 0.2292, "step": 212 }, { "epoch": 0.3538205980066445, "grad_norm": 11.920955657958984, "learning_rate": 1.8898425657861643e-06, "loss": 0.1387, "step": 213 }, { "epoch": 0.3554817275747508, "grad_norm": 7.538236141204834, "learning_rate": 1.8886110590951415e-06, "loss": 0.1015, "step": 214 }, { "epoch": 0.35714285714285715, "grad_norm": 12.414472579956055, "learning_rate": 1.8873731126470434e-06, "loss": 0.1664, "step": 215 }, { "epoch": 0.3588039867109635, "grad_norm": 9.726411819458008, "learning_rate": 1.8861287354132582e-06, "loss": 0.0867, "step": 216 }, { "epoch": 0.36046511627906974, "grad_norm": 17.113433837890625, "learning_rate": 1.8848779364117772e-06, "loss": 0.2126, "step": 217 }, { "epoch": 0.36212624584717606, "grad_norm": 10.119791030883789, "learning_rate": 1.8836207247071307e-06, "loss": 0.1147, "step": 218 }, { "epoch": 0.3637873754152824, "grad_norm": 7.584885597229004, "learning_rate": 1.882357109410322e-06, "loss": 0.0545, "step": 219 }, { "epoch": 0.3654485049833887, "grad_norm": 7.976661682128906, "learning_rate": 1.8810870996787599e-06, "loss": 0.0746, "step": 220 }, { "epoch": 0.36710963455149503, "grad_norm": 16.05130958557129, "learning_rate": 1.8798107047161952e-06, "loss": 0.0874, "step": 221 }, { "epoch": 0.3687707641196013, "grad_norm": 16.22881507873535, "learning_rate": 1.8785279337726517e-06, "loss": 0.2398, "step": 222 }, { "epoch": 0.3704318936877076, "grad_norm": 15.935908317565918, "learning_rate": 1.87723879614436e-06, "loss": 0.1306, "step": 223 }, { "epoch": 0.37209302325581395, "grad_norm": 20.93619155883789, "learning_rate": 1.8759433011736894e-06, "loss": 0.1476, "step": 224 }, { "epoch": 0.37375415282392027, "grad_norm": 8.026069641113281, "learning_rate": 1.8746414582490822e-06, "loss": 0.1122, "step": 225 }, { "epoch": 0.3754152823920266, "grad_norm": 11.861761093139648, "learning_rate": 1.8733332768049826e-06, "loss": 0.1487, "step": 226 }, { "epoch": 0.3770764119601329, "grad_norm": 13.299816131591797, "learning_rate": 1.8720187663217712e-06, "loss": 0.1201, "step": 227 }, { "epoch": 0.3787375415282392, "grad_norm": 12.491223335266113, "learning_rate": 1.870697936325695e-06, "loss": 0.169, "step": 228 }, { "epoch": 0.3803986710963455, "grad_norm": 12.5487060546875, "learning_rate": 1.8693707963887976e-06, "loss": 0.1829, "step": 229 }, { "epoch": 0.38205980066445183, "grad_norm": 14.099169731140137, "learning_rate": 1.8680373561288519e-06, "loss": 0.1489, "step": 230 }, { "epoch": 0.38372093023255816, "grad_norm": 6.0192718505859375, "learning_rate": 1.8666976252092877e-06, "loss": 0.1093, "step": 231 }, { "epoch": 0.3853820598006645, "grad_norm": 8.559925079345703, "learning_rate": 1.8653516133391248e-06, "loss": 0.1205, "step": 232 }, { "epoch": 0.38704318936877075, "grad_norm": 7.076784610748291, "learning_rate": 1.8639993302728998e-06, "loss": 0.1206, "step": 233 }, { "epoch": 0.38870431893687707, "grad_norm": 6.830137729644775, "learning_rate": 1.8626407858105972e-06, "loss": 0.1168, "step": 234 }, { "epoch": 0.3903654485049834, "grad_norm": 10.434679985046387, "learning_rate": 1.8612759897975776e-06, "loss": 0.1848, "step": 235 }, { "epoch": 0.3920265780730897, "grad_norm": 11.184035301208496, "learning_rate": 1.8599049521245066e-06, "loss": 0.1652, "step": 236 }, { "epoch": 0.39368770764119604, "grad_norm": 10.515231132507324, "learning_rate": 1.8585276827272827e-06, "loss": 0.1524, "step": 237 }, { "epoch": 0.3953488372093023, "grad_norm": 7.663083076477051, "learning_rate": 1.8571441915869662e-06, "loss": 0.1314, "step": 238 }, { "epoch": 0.39700996677740863, "grad_norm": 7.605090618133545, "learning_rate": 1.8557544887297061e-06, "loss": 0.1261, "step": 239 }, { "epoch": 0.39867109634551495, "grad_norm": 8.647713661193848, "learning_rate": 1.8543585842266674e-06, "loss": 0.1128, "step": 240 }, { "epoch": 0.4003322259136213, "grad_norm": 13.598785400390625, "learning_rate": 1.8529564881939587e-06, "loss": 0.1606, "step": 241 }, { "epoch": 0.4019933554817276, "grad_norm": 10.329269409179688, "learning_rate": 1.8515482107925587e-06, "loss": 0.1587, "step": 242 }, { "epoch": 0.40365448504983387, "grad_norm": 21.867544174194336, "learning_rate": 1.850133762228242e-06, "loss": 0.0797, "step": 243 }, { "epoch": 0.4053156146179402, "grad_norm": 8.419551849365234, "learning_rate": 1.8487131527515059e-06, "loss": 0.0986, "step": 244 }, { "epoch": 0.4069767441860465, "grad_norm": 17.915225982666016, "learning_rate": 1.8472863926574954e-06, "loss": 0.1774, "step": 245 }, { "epoch": 0.40863787375415284, "grad_norm": 16.778846740722656, "learning_rate": 1.8458534922859294e-06, "loss": 0.2271, "step": 246 }, { "epoch": 0.41029900332225916, "grad_norm": 11.009391784667969, "learning_rate": 1.8444144620210253e-06, "loss": 0.1626, "step": 247 }, { "epoch": 0.4119601328903654, "grad_norm": 14.870036125183105, "learning_rate": 1.8429693122914234e-06, "loss": 0.1578, "step": 248 }, { "epoch": 0.41362126245847175, "grad_norm": 13.489202499389648, "learning_rate": 1.8415180535701125e-06, "loss": 0.1642, "step": 249 }, { "epoch": 0.4152823920265781, "grad_norm": 7.866860866546631, "learning_rate": 1.8400606963743516e-06, "loss": 0.1444, "step": 250 }, { "epoch": 0.4169435215946844, "grad_norm": 11.771472930908203, "learning_rate": 1.8385972512655977e-06, "loss": 0.1617, "step": 251 }, { "epoch": 0.4186046511627907, "grad_norm": 11.763538360595703, "learning_rate": 1.8371277288494245e-06, "loss": 0.1606, "step": 252 }, { "epoch": 0.420265780730897, "grad_norm": 6.283024311065674, "learning_rate": 1.8356521397754493e-06, "loss": 0.1329, "step": 253 }, { "epoch": 0.4219269102990033, "grad_norm": 7.202626705169678, "learning_rate": 1.8341704947372544e-06, "loss": 0.1043, "step": 254 }, { "epoch": 0.42358803986710963, "grad_norm": 9.756691932678223, "learning_rate": 1.8326828044723085e-06, "loss": 0.1469, "step": 255 }, { "epoch": 0.42524916943521596, "grad_norm": 9.103363037109375, "learning_rate": 1.8311890797618913e-06, "loss": 0.1666, "step": 256 }, { "epoch": 0.4269102990033223, "grad_norm": 7.448887825012207, "learning_rate": 1.8296893314310138e-06, "loss": 0.1172, "step": 257 }, { "epoch": 0.42857142857142855, "grad_norm": 9.971731185913086, "learning_rate": 1.8281835703483392e-06, "loss": 0.1809, "step": 258 }, { "epoch": 0.43023255813953487, "grad_norm": 7.47666597366333, "learning_rate": 1.826671807426106e-06, "loss": 0.1046, "step": 259 }, { "epoch": 0.4318936877076412, "grad_norm": 9.57524299621582, "learning_rate": 1.8251540536200482e-06, "loss": 0.1566, "step": 260 }, { "epoch": 0.4335548172757475, "grad_norm": 7.479125022888184, "learning_rate": 1.8236303199293145e-06, "loss": 0.075, "step": 261 }, { "epoch": 0.43521594684385384, "grad_norm": 10.383878707885742, "learning_rate": 1.822100617396391e-06, "loss": 0.1494, "step": 262 }, { "epoch": 0.4368770764119601, "grad_norm": 11.652271270751953, "learning_rate": 1.8205649571070193e-06, "loss": 0.103, "step": 263 }, { "epoch": 0.43853820598006643, "grad_norm": 8.859437942504883, "learning_rate": 1.8190233501901166e-06, "loss": 0.0748, "step": 264 }, { "epoch": 0.44019933554817275, "grad_norm": 16.176801681518555, "learning_rate": 1.8174758078176961e-06, "loss": 0.2191, "step": 265 }, { "epoch": 0.4418604651162791, "grad_norm": 14.912988662719727, "learning_rate": 1.815922341204785e-06, "loss": 0.1775, "step": 266 }, { "epoch": 0.4435215946843854, "grad_norm": 18.418682098388672, "learning_rate": 1.8143629616093419e-06, "loss": 0.2055, "step": 267 }, { "epoch": 0.44518272425249167, "grad_norm": 21.120506286621094, "learning_rate": 1.812797680332179e-06, "loss": 0.205, "step": 268 }, { "epoch": 0.446843853820598, "grad_norm": 22.294757843017578, "learning_rate": 1.811226508716877e-06, "loss": 0.2917, "step": 269 }, { "epoch": 0.4485049833887043, "grad_norm": 19.390213012695312, "learning_rate": 1.8096494581497032e-06, "loss": 0.183, "step": 270 }, { "epoch": 0.45016611295681064, "grad_norm": 5.7725629806518555, "learning_rate": 1.8080665400595302e-06, "loss": 0.0768, "step": 271 }, { "epoch": 0.45182724252491696, "grad_norm": 10.425414085388184, "learning_rate": 1.806477765917753e-06, "loss": 0.1679, "step": 272 }, { "epoch": 0.45348837209302323, "grad_norm": 8.793322563171387, "learning_rate": 1.804883147238204e-06, "loss": 0.1279, "step": 273 }, { "epoch": 0.45514950166112955, "grad_norm": 9.122625350952148, "learning_rate": 1.803282695577072e-06, "loss": 0.1291, "step": 274 }, { "epoch": 0.4568106312292359, "grad_norm": 14.53525161743164, "learning_rate": 1.8016764225328177e-06, "loss": 0.1598, "step": 275 }, { "epoch": 0.4584717607973422, "grad_norm": 10.3401460647583, "learning_rate": 1.8000643397460879e-06, "loss": 0.1736, "step": 276 }, { "epoch": 0.4601328903654485, "grad_norm": 8.310847282409668, "learning_rate": 1.7984464588996339e-06, "loss": 0.1426, "step": 277 }, { "epoch": 0.46179401993355484, "grad_norm": 8.352697372436523, "learning_rate": 1.7968227917182247e-06, "loss": 0.2232, "step": 278 }, { "epoch": 0.4634551495016611, "grad_norm": 7.920416355133057, "learning_rate": 1.7951933499685632e-06, "loss": 0.1649, "step": 279 }, { "epoch": 0.46511627906976744, "grad_norm": 10.16816234588623, "learning_rate": 1.7935581454592002e-06, "loss": 0.1324, "step": 280 }, { "epoch": 0.46677740863787376, "grad_norm": 6.508622646331787, "learning_rate": 1.7919171900404497e-06, "loss": 0.1132, "step": 281 }, { "epoch": 0.4684385382059801, "grad_norm": 6.778066635131836, "learning_rate": 1.790270495604302e-06, "loss": 0.1386, "step": 282 }, { "epoch": 0.4700996677740864, "grad_norm": 8.793827056884766, "learning_rate": 1.7886180740843383e-06, "loss": 0.1468, "step": 283 }, { "epoch": 0.4717607973421927, "grad_norm": 7.901587009429932, "learning_rate": 1.7869599374556438e-06, "loss": 0.1389, "step": 284 }, { "epoch": 0.473421926910299, "grad_norm": 8.337894439697266, "learning_rate": 1.785296097734721e-06, "loss": 0.1333, "step": 285 }, { "epoch": 0.4750830564784053, "grad_norm": 12.676334381103516, "learning_rate": 1.783626566979403e-06, "loss": 0.2014, "step": 286 }, { "epoch": 0.47674418604651164, "grad_norm": 7.533664703369141, "learning_rate": 1.7819513572887655e-06, "loss": 0.1238, "step": 287 }, { "epoch": 0.47840531561461797, "grad_norm": 8.285493850708008, "learning_rate": 1.7802704808030389e-06, "loss": 0.1078, "step": 288 }, { "epoch": 0.48006644518272423, "grad_norm": 8.483846664428711, "learning_rate": 1.778583949703522e-06, "loss": 0.0427, "step": 289 }, { "epoch": 0.48172757475083056, "grad_norm": 7.3853559494018555, "learning_rate": 1.7768917762124915e-06, "loss": 0.0716, "step": 290 }, { "epoch": 0.4833887043189369, "grad_norm": 18.682451248168945, "learning_rate": 1.7751939725931148e-06, "loss": 0.2179, "step": 291 }, { "epoch": 0.4850498338870432, "grad_norm": 14.185220718383789, "learning_rate": 1.773490551149361e-06, "loss": 0.185, "step": 292 }, { "epoch": 0.4867109634551495, "grad_norm": 15.155757904052734, "learning_rate": 1.7717815242259118e-06, "loss": 0.1625, "step": 293 }, { "epoch": 0.4883720930232558, "grad_norm": 15.565550804138184, "learning_rate": 1.7700669042080705e-06, "loss": 0.1547, "step": 294 }, { "epoch": 0.4900332225913621, "grad_norm": 12.508766174316406, "learning_rate": 1.7683467035216749e-06, "loss": 0.1153, "step": 295 }, { "epoch": 0.49169435215946844, "grad_norm": 14.286003112792969, "learning_rate": 1.7666209346330047e-06, "loss": 0.1617, "step": 296 }, { "epoch": 0.49335548172757476, "grad_norm": 12.081841468811035, "learning_rate": 1.7648896100486938e-06, "loss": 0.0842, "step": 297 }, { "epoch": 0.4950166112956811, "grad_norm": 10.268631935119629, "learning_rate": 1.7631527423156367e-06, "loss": 0.0907, "step": 298 }, { "epoch": 0.49667774086378735, "grad_norm": 5.768962383270264, "learning_rate": 1.7614103440209e-06, "loss": 0.077, "step": 299 }, { "epoch": 0.4983388704318937, "grad_norm": 11.941264152526855, "learning_rate": 1.7596624277916292e-06, "loss": 0.1292, "step": 300 }, { "epoch": 0.5, "grad_norm": 14.39963150024414, "learning_rate": 1.7579090062949597e-06, "loss": 0.1434, "step": 301 }, { "epoch": 0.5016611295681063, "grad_norm": 10.610041618347168, "learning_rate": 1.7561500922379225e-06, "loss": 0.1121, "step": 302 }, { "epoch": 0.5033222591362126, "grad_norm": 9.320536613464355, "learning_rate": 1.7543856983673528e-06, "loss": 0.0995, "step": 303 }, { "epoch": 0.5049833887043189, "grad_norm": 14.534180641174316, "learning_rate": 1.7526158374697997e-06, "loss": 0.1487, "step": 304 }, { "epoch": 0.5066445182724253, "grad_norm": 6.3102617263793945, "learning_rate": 1.7508405223714297e-06, "loss": 0.0614, "step": 305 }, { "epoch": 0.5083056478405316, "grad_norm": 13.029088973999023, "learning_rate": 1.7490597659379371e-06, "loss": 0.1273, "step": 306 }, { "epoch": 0.5099667774086378, "grad_norm": 11.671377182006836, "learning_rate": 1.7472735810744493e-06, "loss": 0.1749, "step": 307 }, { "epoch": 0.5116279069767442, "grad_norm": 11.286014556884766, "learning_rate": 1.745481980725433e-06, "loss": 0.1158, "step": 308 }, { "epoch": 0.5132890365448505, "grad_norm": 13.731078147888184, "learning_rate": 1.743684977874602e-06, "loss": 0.0995, "step": 309 }, { "epoch": 0.5149501661129569, "grad_norm": 13.560349464416504, "learning_rate": 1.7418825855448205e-06, "loss": 0.1754, "step": 310 }, { "epoch": 0.5166112956810631, "grad_norm": 15.380659103393555, "learning_rate": 1.7400748167980113e-06, "loss": 0.1893, "step": 311 }, { "epoch": 0.5182724252491694, "grad_norm": 17.06402015686035, "learning_rate": 1.7382616847350597e-06, "loss": 0.172, "step": 312 }, { "epoch": 0.5199335548172758, "grad_norm": 42.22692108154297, "learning_rate": 1.7364432024957193e-06, "loss": 0.1654, "step": 313 }, { "epoch": 0.521594684385382, "grad_norm": 17.076013565063477, "learning_rate": 1.7346193832585153e-06, "loss": 0.1461, "step": 314 }, { "epoch": 0.5232558139534884, "grad_norm": 13.314347267150879, "learning_rate": 1.7327902402406509e-06, "loss": 0.1855, "step": 315 }, { "epoch": 0.5249169435215947, "grad_norm": 14.270951271057129, "learning_rate": 1.7309557866979113e-06, "loss": 0.165, "step": 316 }, { "epoch": 0.526578073089701, "grad_norm": 8.411968231201172, "learning_rate": 1.7291160359245656e-06, "loss": 0.0846, "step": 317 }, { "epoch": 0.5282392026578073, "grad_norm": 8.805373191833496, "learning_rate": 1.727271001253273e-06, "loss": 0.1413, "step": 318 }, { "epoch": 0.5299003322259136, "grad_norm": 9.515275955200195, "learning_rate": 1.7254206960549842e-06, "loss": 0.1666, "step": 319 }, { "epoch": 0.53156146179402, "grad_norm": 9.840771675109863, "learning_rate": 1.7235651337388465e-06, "loss": 0.1051, "step": 320 }, { "epoch": 0.5332225913621262, "grad_norm": 10.479392051696777, "learning_rate": 1.721704327752104e-06, "loss": 0.1611, "step": 321 }, { "epoch": 0.5348837209302325, "grad_norm": 8.246603965759277, "learning_rate": 1.7198382915800032e-06, "loss": 0.1604, "step": 322 }, { "epoch": 0.5365448504983389, "grad_norm": 10.39277172088623, "learning_rate": 1.7179670387456926e-06, "loss": 0.1206, "step": 323 }, { "epoch": 0.5382059800664452, "grad_norm": 6.932436466217041, "learning_rate": 1.7160905828101261e-06, "loss": 0.1056, "step": 324 }, { "epoch": 0.5398671096345515, "grad_norm": 9.80469799041748, "learning_rate": 1.714208937371965e-06, "loss": 0.0961, "step": 325 }, { "epoch": 0.5415282392026578, "grad_norm": 10.003324508666992, "learning_rate": 1.7123221160674781e-06, "loss": 0.142, "step": 326 }, { "epoch": 0.5431893687707641, "grad_norm": 16.17133331298828, "learning_rate": 1.710430132570444e-06, "loss": 0.2312, "step": 327 }, { "epoch": 0.5448504983388704, "grad_norm": 19.575956344604492, "learning_rate": 1.7085330005920514e-06, "loss": 0.2264, "step": 328 }, { "epoch": 0.5465116279069767, "grad_norm": 17.07429313659668, "learning_rate": 1.7066307338808002e-06, "loss": 0.1616, "step": 329 }, { "epoch": 0.5481727574750831, "grad_norm": 6.755792140960693, "learning_rate": 1.7047233462224022e-06, "loss": 0.095, "step": 330 }, { "epoch": 0.5498338870431894, "grad_norm": 8.61792278289795, "learning_rate": 1.7028108514396797e-06, "loss": 0.1088, "step": 331 }, { "epoch": 0.5514950166112956, "grad_norm": 16.29265022277832, "learning_rate": 1.7008932633924671e-06, "loss": 0.2209, "step": 332 }, { "epoch": 0.553156146179402, "grad_norm": 6.72524356842041, "learning_rate": 1.6989705959775088e-06, "loss": 0.0891, "step": 333 }, { "epoch": 0.5548172757475083, "grad_norm": 7.47697114944458, "learning_rate": 1.69704286312836e-06, "loss": 0.0699, "step": 334 }, { "epoch": 0.5564784053156147, "grad_norm": 6.8748064041137695, "learning_rate": 1.695110078815285e-06, "loss": 0.1072, "step": 335 }, { "epoch": 0.5581395348837209, "grad_norm": 12.712291717529297, "learning_rate": 1.6931722570451562e-06, "loss": 0.1672, "step": 336 }, { "epoch": 0.5598006644518272, "grad_norm": 9.507533073425293, "learning_rate": 1.6912294118613515e-06, "loss": 0.0875, "step": 337 }, { "epoch": 0.5614617940199336, "grad_norm": 22.127843856811523, "learning_rate": 1.6892815573436546e-06, "loss": 0.1628, "step": 338 }, { "epoch": 0.5631229235880398, "grad_norm": 16.105968475341797, "learning_rate": 1.6873287076081506e-06, "loss": 0.168, "step": 339 }, { "epoch": 0.5647840531561462, "grad_norm": 17.4877986907959, "learning_rate": 1.6853708768071263e-06, "loss": 0.2376, "step": 340 }, { "epoch": 0.5664451827242525, "grad_norm": 18.06496238708496, "learning_rate": 1.683408079128965e-06, "loss": 0.1721, "step": 341 }, { "epoch": 0.5681063122923588, "grad_norm": 8.012746810913086, "learning_rate": 1.681440328798046e-06, "loss": 0.1208, "step": 342 }, { "epoch": 0.5697674418604651, "grad_norm": 9.295697212219238, "learning_rate": 1.6794676400746387e-06, "loss": 0.0754, "step": 343 }, { "epoch": 0.5714285714285714, "grad_norm": 15.045690536499023, "learning_rate": 1.6774900272548034e-06, "loss": 0.1164, "step": 344 }, { "epoch": 0.5730897009966778, "grad_norm": 6.664107322692871, "learning_rate": 1.6755075046702834e-06, "loss": 0.0714, "step": 345 }, { "epoch": 0.574750830564784, "grad_norm": 13.638888359069824, "learning_rate": 1.6735200866884035e-06, "loss": 0.1313, "step": 346 }, { "epoch": 0.5764119601328903, "grad_norm": 14.996500968933105, "learning_rate": 1.671527787711966e-06, "loss": 0.1683, "step": 347 }, { "epoch": 0.5780730897009967, "grad_norm": 6.490530014038086, "learning_rate": 1.669530622179145e-06, "loss": 0.0478, "step": 348 }, { "epoch": 0.579734219269103, "grad_norm": 12.507638931274414, "learning_rate": 1.6675286045633827e-06, "loss": 0.1066, "step": 349 }, { "epoch": 0.5813953488372093, "grad_norm": 13.5231294631958, "learning_rate": 1.6655217493732847e-06, "loss": 0.1507, "step": 350 }, { "epoch": 0.5830564784053156, "grad_norm": 10.306071281433105, "learning_rate": 1.663510071152514e-06, "loss": 0.0868, "step": 351 }, { "epoch": 0.584717607973422, "grad_norm": 13.600068092346191, "learning_rate": 1.6614935844796863e-06, "loss": 0.1112, "step": 352 }, { "epoch": 0.5863787375415282, "grad_norm": 7.447047710418701, "learning_rate": 1.6594723039682637e-06, "loss": 0.0685, "step": 353 }, { "epoch": 0.5880398671096345, "grad_norm": 17.658233642578125, "learning_rate": 1.65744624426645e-06, "loss": 0.1508, "step": 354 }, { "epoch": 0.5897009966777409, "grad_norm": 9.628240585327148, "learning_rate": 1.6554154200570825e-06, "loss": 0.1017, "step": 355 }, { "epoch": 0.5913621262458472, "grad_norm": 8.991092681884766, "learning_rate": 1.6533798460575284e-06, "loss": 0.1102, "step": 356 }, { "epoch": 0.5930232558139535, "grad_norm": 9.290343284606934, "learning_rate": 1.6513395370195755e-06, "loss": 0.0953, "step": 357 }, { "epoch": 0.5946843853820598, "grad_norm": 12.639936447143555, "learning_rate": 1.6492945077293268e-06, "loss": 0.1731, "step": 358 }, { "epoch": 0.5963455149501661, "grad_norm": 15.398722648620605, "learning_rate": 1.647244773007093e-06, "loss": 0.1704, "step": 359 }, { "epoch": 0.5980066445182725, "grad_norm": 4.541179656982422, "learning_rate": 1.6451903477072845e-06, "loss": 0.0519, "step": 360 }, { "epoch": 0.5996677740863787, "grad_norm": 7.931593418121338, "learning_rate": 1.6431312467183047e-06, "loss": 0.0743, "step": 361 }, { "epoch": 0.6013289036544851, "grad_norm": 10.063234329223633, "learning_rate": 1.6410674849624417e-06, "loss": 0.1202, "step": 362 }, { "epoch": 0.6029900332225914, "grad_norm": 12.936481475830078, "learning_rate": 1.6389990773957605e-06, "loss": 0.1219, "step": 363 }, { "epoch": 0.6046511627906976, "grad_norm": 20.578336715698242, "learning_rate": 1.636926039007993e-06, "loss": 0.1967, "step": 364 }, { "epoch": 0.606312292358804, "grad_norm": 7.607759475708008, "learning_rate": 1.634848384822432e-06, "loss": 0.0589, "step": 365 }, { "epoch": 0.6079734219269103, "grad_norm": 12.115143775939941, "learning_rate": 1.6327661298958203e-06, "loss": 0.1078, "step": 366 }, { "epoch": 0.6096345514950167, "grad_norm": 10.23250675201416, "learning_rate": 1.630679289318242e-06, "loss": 0.0893, "step": 367 }, { "epoch": 0.6112956810631229, "grad_norm": 15.31701946258545, "learning_rate": 1.628587878213014e-06, "loss": 0.2084, "step": 368 }, { "epoch": 0.6129568106312292, "grad_norm": 13.3017578125, "learning_rate": 1.6264919117365756e-06, "loss": 0.1501, "step": 369 }, { "epoch": 0.6146179401993356, "grad_norm": 11.242172241210938, "learning_rate": 1.6243914050783782e-06, "loss": 0.1418, "step": 370 }, { "epoch": 0.6162790697674418, "grad_norm": 15.284845352172852, "learning_rate": 1.6222863734607767e-06, "loss": 0.2246, "step": 371 }, { "epoch": 0.6179401993355482, "grad_norm": 5.70803165435791, "learning_rate": 1.620176832138918e-06, "loss": 0.0742, "step": 372 }, { "epoch": 0.6196013289036545, "grad_norm": 11.78730297088623, "learning_rate": 1.618062796400631e-06, "loss": 0.1321, "step": 373 }, { "epoch": 0.6212624584717608, "grad_norm": 6.646214962005615, "learning_rate": 1.6159442815663149e-06, "loss": 0.1329, "step": 374 }, { "epoch": 0.6229235880398671, "grad_norm": 9.914974212646484, "learning_rate": 1.61382130298883e-06, "loss": 0.0823, "step": 375 }, { "epoch": 0.6245847176079734, "grad_norm": 7.075379848480225, "learning_rate": 1.6116938760533842e-06, "loss": 0.079, "step": 376 }, { "epoch": 0.6262458471760798, "grad_norm": 11.401320457458496, "learning_rate": 1.6095620161774232e-06, "loss": 0.1558, "step": 377 }, { "epoch": 0.627906976744186, "grad_norm": 8.446409225463867, "learning_rate": 1.6074257388105176e-06, "loss": 0.1214, "step": 378 }, { "epoch": 0.6295681063122923, "grad_norm": 21.146482467651367, "learning_rate": 1.6052850594342533e-06, "loss": 0.1599, "step": 379 }, { "epoch": 0.6312292358803987, "grad_norm": 8.168869018554688, "learning_rate": 1.6031399935621153e-06, "loss": 0.0863, "step": 380 }, { "epoch": 0.632890365448505, "grad_norm": 11.66989517211914, "learning_rate": 1.6009905567393783e-06, "loss": 0.1079, "step": 381 }, { "epoch": 0.6345514950166113, "grad_norm": 6.585104465484619, "learning_rate": 1.5988367645429936e-06, "loss": 0.0671, "step": 382 }, { "epoch": 0.6362126245847176, "grad_norm": 7.7533392906188965, "learning_rate": 1.5966786325814757e-06, "loss": 0.0791, "step": 383 }, { "epoch": 0.6378737541528239, "grad_norm": 19.80597496032715, "learning_rate": 1.5945161764947892e-06, "loss": 0.144, "step": 384 }, { "epoch": 0.6395348837209303, "grad_norm": 12.21318531036377, "learning_rate": 1.5923494119542357e-06, "loss": 0.1094, "step": 385 }, { "epoch": 0.6411960132890365, "grad_norm": 13.927626609802246, "learning_rate": 1.5901783546623402e-06, "loss": 0.1036, "step": 386 }, { "epoch": 0.6428571428571429, "grad_norm": 17.715267181396484, "learning_rate": 1.5880030203527367e-06, "loss": 0.1133, "step": 387 }, { "epoch": 0.6445182724252492, "grad_norm": 19.633594512939453, "learning_rate": 1.585823424790056e-06, "loss": 0.1864, "step": 388 }, { "epoch": 0.6461794019933554, "grad_norm": 13.235529899597168, "learning_rate": 1.5836395837698084e-06, "loss": 0.1207, "step": 389 }, { "epoch": 0.6478405315614618, "grad_norm": 10.060149192810059, "learning_rate": 1.5814515131182725e-06, "loss": 0.0723, "step": 390 }, { "epoch": 0.6495016611295681, "grad_norm": 16.433595657348633, "learning_rate": 1.5792592286923778e-06, "loss": 0.176, "step": 391 }, { "epoch": 0.6511627906976745, "grad_norm": 14.24596881866455, "learning_rate": 1.5770627463795924e-06, "loss": 0.1633, "step": 392 }, { "epoch": 0.6528239202657807, "grad_norm": 11.974753379821777, "learning_rate": 1.5748620820978055e-06, "loss": 0.1403, "step": 393 }, { "epoch": 0.654485049833887, "grad_norm": 25.53243637084961, "learning_rate": 1.5726572517952121e-06, "loss": 0.1529, "step": 394 }, { "epoch": 0.6561461794019934, "grad_norm": 11.890317916870117, "learning_rate": 1.5704482714502005e-06, "loss": 0.1004, "step": 395 }, { "epoch": 0.6578073089700996, "grad_norm": 13.248994827270508, "learning_rate": 1.568235157071233e-06, "loss": 0.1605, "step": 396 }, { "epoch": 0.659468438538206, "grad_norm": 10.647342681884766, "learning_rate": 1.5660179246967311e-06, "loss": 0.1145, "step": 397 }, { "epoch": 0.6611295681063123, "grad_norm": 9.527441024780273, "learning_rate": 1.5637965903949597e-06, "loss": 0.1327, "step": 398 }, { "epoch": 0.6627906976744186, "grad_norm": 10.44089412689209, "learning_rate": 1.5615711702639107e-06, "loss": 0.1212, "step": 399 }, { "epoch": 0.6644518272425249, "grad_norm": 7.151495456695557, "learning_rate": 1.559341680431185e-06, "loss": 0.1242, "step": 400 }, { "epoch": 0.6661129568106312, "grad_norm": 13.282041549682617, "learning_rate": 1.5571081370538775e-06, "loss": 0.2158, "step": 401 }, { "epoch": 0.6677740863787376, "grad_norm": 6.182162761688232, "learning_rate": 1.554870556318459e-06, "loss": 0.0821, "step": 402 }, { "epoch": 0.6694352159468439, "grad_norm": 6.321899890899658, "learning_rate": 1.5526289544406583e-06, "loss": 0.1002, "step": 403 }, { "epoch": 0.6710963455149501, "grad_norm": 13.514141082763672, "learning_rate": 1.5503833476653463e-06, "loss": 0.1969, "step": 404 }, { "epoch": 0.6727574750830565, "grad_norm": 9.189848899841309, "learning_rate": 1.5481337522664167e-06, "loss": 0.1388, "step": 405 }, { "epoch": 0.6744186046511628, "grad_norm": 7.638835430145264, "learning_rate": 1.545880184546669e-06, "loss": 0.1531, "step": 406 }, { "epoch": 0.6760797342192691, "grad_norm": 11.591959953308105, "learning_rate": 1.54362266083769e-06, "loss": 0.1512, "step": 407 }, { "epoch": 0.6777408637873754, "grad_norm": 11.338805198669434, "learning_rate": 1.5413611974997354e-06, "loss": 0.1792, "step": 408 }, { "epoch": 0.6794019933554817, "grad_norm": 9.469964981079102, "learning_rate": 1.539095810921612e-06, "loss": 0.1064, "step": 409 }, { "epoch": 0.6810631229235881, "grad_norm": 9.461012840270996, "learning_rate": 1.5368265175205572e-06, "loss": 0.145, "step": 410 }, { "epoch": 0.6827242524916943, "grad_norm": 6.2333903312683105, "learning_rate": 1.534553333742122e-06, "loss": 0.1042, "step": 411 }, { "epoch": 0.6843853820598007, "grad_norm": 9.561917304992676, "learning_rate": 1.5322762760600507e-06, "loss": 0.157, "step": 412 }, { "epoch": 0.686046511627907, "grad_norm": 17.19264030456543, "learning_rate": 1.5299953609761617e-06, "loss": 0.1578, "step": 413 }, { "epoch": 0.6877076411960132, "grad_norm": 13.552623748779297, "learning_rate": 1.527710605020228e-06, "loss": 0.1879, "step": 414 }, { "epoch": 0.6893687707641196, "grad_norm": 10.698836326599121, "learning_rate": 1.525422024749857e-06, "loss": 0.1388, "step": 415 }, { "epoch": 0.6910299003322259, "grad_norm": 8.805788040161133, "learning_rate": 1.5231296367503715e-06, "loss": 0.0887, "step": 416 }, { "epoch": 0.6926910299003323, "grad_norm": 9.94221019744873, "learning_rate": 1.520833457634688e-06, "loss": 0.1246, "step": 417 }, { "epoch": 0.6943521594684385, "grad_norm": 12.409692764282227, "learning_rate": 1.5185335040431988e-06, "loss": 0.1039, "step": 418 }, { "epoch": 0.6960132890365448, "grad_norm": 8.411705017089844, "learning_rate": 1.5162297926436481e-06, "loss": 0.0864, "step": 419 }, { "epoch": 0.6976744186046512, "grad_norm": 12.356369972229004, "learning_rate": 1.5139223401310135e-06, "loss": 0.1329, "step": 420 }, { "epoch": 0.6993355481727574, "grad_norm": 16.476016998291016, "learning_rate": 1.5116111632273847e-06, "loss": 0.1199, "step": 421 }, { "epoch": 0.7009966777408638, "grad_norm": 12.528667449951172, "learning_rate": 1.5092962786818418e-06, "loss": 0.1322, "step": 422 }, { "epoch": 0.7026578073089701, "grad_norm": 16.012834548950195, "learning_rate": 1.506977703270334e-06, "loss": 0.1461, "step": 423 }, { "epoch": 0.7043189368770764, "grad_norm": 8.67282772064209, "learning_rate": 1.5046554537955584e-06, "loss": 0.0871, "step": 424 }, { "epoch": 0.7059800664451827, "grad_norm": 9.403173446655273, "learning_rate": 1.502329547086838e-06, "loss": 0.1064, "step": 425 }, { "epoch": 0.707641196013289, "grad_norm": 10.017657279968262, "learning_rate": 1.5e-06, "loss": 0.0891, "step": 426 }, { "epoch": 0.7093023255813954, "grad_norm": 12.303718566894531, "learning_rate": 1.4976668294172526e-06, "loss": 0.1399, "step": 427 }, { "epoch": 0.7109634551495017, "grad_norm": 10.830999374389648, "learning_rate": 1.4953300522470637e-06, "loss": 0.1286, "step": 428 }, { "epoch": 0.7126245847176079, "grad_norm": 14.537104606628418, "learning_rate": 1.4929896854240394e-06, "loss": 0.1815, "step": 429 }, { "epoch": 0.7142857142857143, "grad_norm": 6.298677444458008, "learning_rate": 1.4906457459087975e-06, "loss": 0.0982, "step": 430 }, { "epoch": 0.7159468438538206, "grad_norm": 7.235755920410156, "learning_rate": 1.4882982506878498e-06, "loss": 0.0952, "step": 431 }, { "epoch": 0.717607973421927, "grad_norm": 12.166476249694824, "learning_rate": 1.4859472167734743e-06, "loss": 0.1752, "step": 432 }, { "epoch": 0.7192691029900332, "grad_norm": 7.841498374938965, "learning_rate": 1.4835926612035944e-06, "loss": 0.1136, "step": 433 }, { "epoch": 0.7209302325581395, "grad_norm": 9.74621868133545, "learning_rate": 1.4812346010416553e-06, "loss": 0.1213, "step": 434 }, { "epoch": 0.7225913621262459, "grad_norm": 6.355194568634033, "learning_rate": 1.4788730533765001e-06, "loss": 0.0953, "step": 435 }, { "epoch": 0.7242524916943521, "grad_norm": 13.128461837768555, "learning_rate": 1.4765080353222446e-06, "loss": 0.1959, "step": 436 }, { "epoch": 0.7259136212624585, "grad_norm": 9.213345527648926, "learning_rate": 1.4741395640181557e-06, "loss": 0.1339, "step": 437 }, { "epoch": 0.7275747508305648, "grad_norm": 8.842706680297852, "learning_rate": 1.4717676566285256e-06, "loss": 0.1312, "step": 438 }, { "epoch": 0.729235880398671, "grad_norm": 10.16006088256836, "learning_rate": 1.4693923303425477e-06, "loss": 0.1049, "step": 439 }, { "epoch": 0.7308970099667774, "grad_norm": 10.167031288146973, "learning_rate": 1.4670136023741924e-06, "loss": 0.139, "step": 440 }, { "epoch": 0.7325581395348837, "grad_norm": 7.213005542755127, "learning_rate": 1.464631489962081e-06, "loss": 0.087, "step": 441 }, { "epoch": 0.7342192691029901, "grad_norm": 5.736367702484131, "learning_rate": 1.4622460103693637e-06, "loss": 0.0763, "step": 442 }, { "epoch": 0.7358803986710963, "grad_norm": 7.528397560119629, "learning_rate": 1.459857180883591e-06, "loss": 0.0728, "step": 443 }, { "epoch": 0.7375415282392026, "grad_norm": 8.150105476379395, "learning_rate": 1.4574650188165908e-06, "loss": 0.1374, "step": 444 }, { "epoch": 0.739202657807309, "grad_norm": 11.9616060256958, "learning_rate": 1.4550695415043418e-06, "loss": 0.142, "step": 445 }, { "epoch": 0.7408637873754153, "grad_norm": 9.410012245178223, "learning_rate": 1.4526707663068488e-06, "loss": 0.1567, "step": 446 }, { "epoch": 0.7425249169435216, "grad_norm": 11.85716438293457, "learning_rate": 1.450268710608016e-06, "loss": 0.1245, "step": 447 }, { "epoch": 0.7441860465116279, "grad_norm": 9.234456062316895, "learning_rate": 1.4478633918155214e-06, "loss": 0.0959, "step": 448 }, { "epoch": 0.7458471760797342, "grad_norm": 22.62066078186035, "learning_rate": 1.4454548273606905e-06, "loss": 0.2426, "step": 449 }, { "epoch": 0.7475083056478405, "grad_norm": 11.37014389038086, "learning_rate": 1.4430430346983698e-06, "loss": 0.1089, "step": 450 }, { "epoch": 0.7491694352159468, "grad_norm": 15.76647663116455, "learning_rate": 1.4406280313068019e-06, "loss": 0.1549, "step": 451 }, { "epoch": 0.7508305647840532, "grad_norm": 8.85522174835205, "learning_rate": 1.4382098346874957e-06, "loss": 0.1208, "step": 452 }, { "epoch": 0.7524916943521595, "grad_norm": 9.161630630493164, "learning_rate": 1.4357884623651031e-06, "loss": 0.132, "step": 453 }, { "epoch": 0.7541528239202658, "grad_norm": 14.351627349853516, "learning_rate": 1.433363931887289e-06, "loss": 0.1408, "step": 454 }, { "epoch": 0.7558139534883721, "grad_norm": 7.650735378265381, "learning_rate": 1.430936260824606e-06, "loss": 0.1038, "step": 455 }, { "epoch": 0.7574750830564784, "grad_norm": 37.2721061706543, "learning_rate": 1.4285054667703666e-06, "loss": 0.1731, "step": 456 }, { "epoch": 0.7591362126245847, "grad_norm": 9.084954261779785, "learning_rate": 1.4260715673405154e-06, "loss": 0.0963, "step": 457 }, { "epoch": 0.760797342192691, "grad_norm": 9.70747184753418, "learning_rate": 1.4236345801735017e-06, "loss": 0.0959, "step": 458 }, { "epoch": 0.7624584717607974, "grad_norm": 13.642250061035156, "learning_rate": 1.421194522930151e-06, "loss": 0.1338, "step": 459 }, { "epoch": 0.7641196013289037, "grad_norm": 7.298130035400391, "learning_rate": 1.4187514132935392e-06, "loss": 0.0824, "step": 460 }, { "epoch": 0.7657807308970099, "grad_norm": 8.563011169433594, "learning_rate": 1.416305268968861e-06, "loss": 0.1036, "step": 461 }, { "epoch": 0.7674418604651163, "grad_norm": 5.615513801574707, "learning_rate": 1.4138561076833047e-06, "loss": 0.0747, "step": 462 }, { "epoch": 0.7691029900332226, "grad_norm": 11.83483600616455, "learning_rate": 1.411403947185922e-06, "loss": 0.1074, "step": 463 }, { "epoch": 0.770764119601329, "grad_norm": 10.225103378295898, "learning_rate": 1.4089488052474997e-06, "loss": 0.1229, "step": 464 }, { "epoch": 0.7724252491694352, "grad_norm": 11.923834800720215, "learning_rate": 1.406490699660432e-06, "loss": 0.1, "step": 465 }, { "epoch": 0.7740863787375415, "grad_norm": 8.985201835632324, "learning_rate": 1.404029648238589e-06, "loss": 0.0827, "step": 466 }, { "epoch": 0.7757475083056479, "grad_norm": 17.72941017150879, "learning_rate": 1.4015656688171916e-06, "loss": 0.1344, "step": 467 }, { "epoch": 0.7774086378737541, "grad_norm": 11.252090454101562, "learning_rate": 1.399098779252677e-06, "loss": 0.1579, "step": 468 }, { "epoch": 0.7790697674418605, "grad_norm": 12.625454902648926, "learning_rate": 1.396628997422575e-06, "loss": 0.1508, "step": 469 }, { "epoch": 0.7807308970099668, "grad_norm": 17.121810913085938, "learning_rate": 1.3941563412253729e-06, "loss": 0.1077, "step": 470 }, { "epoch": 0.782392026578073, "grad_norm": 9.032465934753418, "learning_rate": 1.3916808285803913e-06, "loss": 0.1013, "step": 471 }, { "epoch": 0.7840531561461794, "grad_norm": 9.71914005279541, "learning_rate": 1.3892024774276493e-06, "loss": 0.1332, "step": 472 }, { "epoch": 0.7857142857142857, "grad_norm": 21.92658233642578, "learning_rate": 1.3867213057277378e-06, "loss": 0.281, "step": 473 }, { "epoch": 0.7873754152823921, "grad_norm": 14.790511131286621, "learning_rate": 1.3842373314616878e-06, "loss": 0.1159, "step": 474 }, { "epoch": 0.7890365448504983, "grad_norm": 7.3284077644348145, "learning_rate": 1.38175057263084e-06, "loss": 0.1129, "step": 475 }, { "epoch": 0.7906976744186046, "grad_norm": 8.450250625610352, "learning_rate": 1.379261047256716e-06, "loss": 0.0967, "step": 476 }, { "epoch": 0.792358803986711, "grad_norm": 11.382468223571777, "learning_rate": 1.376768773380885e-06, "loss": 0.131, "step": 477 }, { "epoch": 0.7940199335548173, "grad_norm": 9.446375846862793, "learning_rate": 1.374273769064836e-06, "loss": 0.0924, "step": 478 }, { "epoch": 0.7956810631229236, "grad_norm": 15.360520362854004, "learning_rate": 1.3717760523898449e-06, "loss": 0.1401, "step": 479 }, { "epoch": 0.7973421926910299, "grad_norm": 8.478147506713867, "learning_rate": 1.369275641456844e-06, "loss": 0.1119, "step": 480 }, { "epoch": 0.7990033222591362, "grad_norm": 7.132209777832031, "learning_rate": 1.3667725543862904e-06, "loss": 0.0967, "step": 481 }, { "epoch": 0.8006644518272426, "grad_norm": 7.364678382873535, "learning_rate": 1.364266809318036e-06, "loss": 0.1483, "step": 482 }, { "epoch": 0.8023255813953488, "grad_norm": 6.641203880310059, "learning_rate": 1.3617584244111947e-06, "loss": 0.1094, "step": 483 }, { "epoch": 0.8039867109634552, "grad_norm": 9.827098846435547, "learning_rate": 1.3592474178440114e-06, "loss": 0.1196, "step": 484 }, { "epoch": 0.8056478405315615, "grad_norm": 7.959251880645752, "learning_rate": 1.3567338078137304e-06, "loss": 0.1112, "step": 485 }, { "epoch": 0.8073089700996677, "grad_norm": 10.041277885437012, "learning_rate": 1.3542176125364624e-06, "loss": 0.125, "step": 486 }, { "epoch": 0.8089700996677741, "grad_norm": 7.443721294403076, "learning_rate": 1.3516988502470547e-06, "loss": 0.0793, "step": 487 }, { "epoch": 0.8106312292358804, "grad_norm": 4.981016635894775, "learning_rate": 1.3491775391989567e-06, "loss": 0.0578, "step": 488 }, { "epoch": 0.8122923588039868, "grad_norm": 8.452247619628906, "learning_rate": 1.3466536976640885e-06, "loss": 0.0756, "step": 489 }, { "epoch": 0.813953488372093, "grad_norm": 11.617294311523438, "learning_rate": 1.3441273439327099e-06, "loss": 0.1108, "step": 490 }, { "epoch": 0.8156146179401993, "grad_norm": 13.18224048614502, "learning_rate": 1.3415984963132851e-06, "loss": 0.1833, "step": 491 }, { "epoch": 0.8172757475083057, "grad_norm": 8.125739097595215, "learning_rate": 1.3390671731323526e-06, "loss": 0.0788, "step": 492 }, { "epoch": 0.8189368770764119, "grad_norm": 9.55408763885498, "learning_rate": 1.3365333927343905e-06, "loss": 0.0969, "step": 493 }, { "epoch": 0.8205980066445183, "grad_norm": 14.843066215515137, "learning_rate": 1.3339971734816842e-06, "loss": 0.1476, "step": 494 }, { "epoch": 0.8222591362126246, "grad_norm": 12.41232681274414, "learning_rate": 1.3314585337541947e-06, "loss": 0.1056, "step": 495 }, { "epoch": 0.8239202657807309, "grad_norm": 18.931259155273438, "learning_rate": 1.3289174919494226e-06, "loss": 0.1967, "step": 496 }, { "epoch": 0.8255813953488372, "grad_norm": 19.2910213470459, "learning_rate": 1.3263740664822773e-06, "loss": 0.1126, "step": 497 }, { "epoch": 0.8272425249169435, "grad_norm": 10.538687705993652, "learning_rate": 1.3238282757849428e-06, "loss": 0.1148, "step": 498 }, { "epoch": 0.8289036544850499, "grad_norm": 9.582296371459961, "learning_rate": 1.321280138306743e-06, "loss": 0.0927, "step": 499 }, { "epoch": 0.8305647840531561, "grad_norm": 7.91030740737915, "learning_rate": 1.318729672514009e-06, "loss": 0.0851, "step": 500 }, { "epoch": 0.8322259136212624, "grad_norm": 13.952978134155273, "learning_rate": 1.3161768968899466e-06, "loss": 0.0748, "step": 501 }, { "epoch": 0.8338870431893688, "grad_norm": 6.64433479309082, "learning_rate": 1.3136218299344992e-06, "loss": 0.0811, "step": 502 }, { "epoch": 0.8355481727574751, "grad_norm": 11.787875175476074, "learning_rate": 1.3110644901642162e-06, "loss": 0.1176, "step": 503 }, { "epoch": 0.8372093023255814, "grad_norm": 8.039713859558105, "learning_rate": 1.3085048961121175e-06, "loss": 0.0978, "step": 504 }, { "epoch": 0.8388704318936877, "grad_norm": 21.696443557739258, "learning_rate": 1.3059430663275608e-06, "loss": 0.1458, "step": 505 }, { "epoch": 0.840531561461794, "grad_norm": 7.948482990264893, "learning_rate": 1.3033790193761058e-06, "loss": 0.0772, "step": 506 }, { "epoch": 0.8421926910299004, "grad_norm": 13.564935684204102, "learning_rate": 1.3008127738393793e-06, "loss": 0.1865, "step": 507 }, { "epoch": 0.8438538205980066, "grad_norm": 13.352168083190918, "learning_rate": 1.2982443483149422e-06, "loss": 0.1602, "step": 508 }, { "epoch": 0.845514950166113, "grad_norm": 13.05396842956543, "learning_rate": 1.2956737614161525e-06, "loss": 0.1097, "step": 509 }, { "epoch": 0.8471760797342193, "grad_norm": 11.9678373336792, "learning_rate": 1.293101031772033e-06, "loss": 0.1681, "step": 510 }, { "epoch": 0.8488372093023255, "grad_norm": 9.613672256469727, "learning_rate": 1.2905261780271343e-06, "loss": 0.1004, "step": 511 }, { "epoch": 0.8504983388704319, "grad_norm": 9.589925765991211, "learning_rate": 1.2879492188414008e-06, "loss": 0.0901, "step": 512 }, { "epoch": 0.8521594684385382, "grad_norm": 11.159270286560059, "learning_rate": 1.2853701728900337e-06, "loss": 0.1538, "step": 513 }, { "epoch": 0.8538205980066446, "grad_norm": 11.989233016967773, "learning_rate": 1.2827890588633588e-06, "loss": 0.1468, "step": 514 }, { "epoch": 0.8554817275747508, "grad_norm": 8.866310119628906, "learning_rate": 1.2802058954666873e-06, "loss": 0.1249, "step": 515 }, { "epoch": 0.8571428571428571, "grad_norm": 5.853137493133545, "learning_rate": 1.2776207014201843e-06, "loss": 0.0761, "step": 516 }, { "epoch": 0.8588039867109635, "grad_norm": 10.618156433105469, "learning_rate": 1.2750334954587297e-06, "loss": 0.1087, "step": 517 }, { "epoch": 0.8604651162790697, "grad_norm": 10.3444242477417, "learning_rate": 1.2724442963317835e-06, "loss": 0.1751, "step": 518 }, { "epoch": 0.8621262458471761, "grad_norm": 12.944339752197266, "learning_rate": 1.2698531228032508e-06, "loss": 0.2003, "step": 519 }, { "epoch": 0.8637873754152824, "grad_norm": 9.638875007629395, "learning_rate": 1.2672599936513448e-06, "loss": 0.1065, "step": 520 }, { "epoch": 0.8654485049833887, "grad_norm": 7.887294292449951, "learning_rate": 1.2646649276684514e-06, "loss": 0.1576, "step": 521 }, { "epoch": 0.867109634551495, "grad_norm": 9.495591163635254, "learning_rate": 1.262067943660993e-06, "loss": 0.1919, "step": 522 }, { "epoch": 0.8687707641196013, "grad_norm": 7.008487224578857, "learning_rate": 1.2594690604492904e-06, "loss": 0.1603, "step": 523 }, { "epoch": 0.8704318936877077, "grad_norm": 5.752983093261719, "learning_rate": 1.25686829686743e-06, "loss": 0.116, "step": 524 }, { "epoch": 0.872093023255814, "grad_norm": 6.817920207977295, "learning_rate": 1.2542656717631236e-06, "loss": 0.1433, "step": 525 }, { "epoch": 0.8737541528239202, "grad_norm": 9.408466339111328, "learning_rate": 1.2516612039975742e-06, "loss": 0.1372, "step": 526 }, { "epoch": 0.8754152823920266, "grad_norm": 12.023665428161621, "learning_rate": 1.2490549124453385e-06, "loss": 0.1777, "step": 527 }, { "epoch": 0.8770764119601329, "grad_norm": 7.048364639282227, "learning_rate": 1.2464468159941898e-06, "loss": 0.1261, "step": 528 }, { "epoch": 0.8787375415282392, "grad_norm": 6.778327941894531, "learning_rate": 1.2438369335449822e-06, "loss": 0.1319, "step": 529 }, { "epoch": 0.8803986710963455, "grad_norm": 7.284000873565674, "learning_rate": 1.241225284011512e-06, "loss": 0.1324, "step": 530 }, { "epoch": 0.8820598006644518, "grad_norm": 11.69240951538086, "learning_rate": 1.2386118863203818e-06, "loss": 0.1602, "step": 531 }, { "epoch": 0.8837209302325582, "grad_norm": 9.516483306884766, "learning_rate": 1.2359967594108642e-06, "loss": 0.1768, "step": 532 }, { "epoch": 0.8853820598006644, "grad_norm": 5.771677494049072, "learning_rate": 1.2333799222347617e-06, "loss": 0.1187, "step": 533 }, { "epoch": 0.8870431893687708, "grad_norm": 6.525841236114502, "learning_rate": 1.230761393756272e-06, "loss": 0.1169, "step": 534 }, { "epoch": 0.8887043189368771, "grad_norm": 8.253816604614258, "learning_rate": 1.22814119295185e-06, "loss": 0.1622, "step": 535 }, { "epoch": 0.8903654485049833, "grad_norm": 9.743227005004883, "learning_rate": 1.225519338810069e-06, "loss": 0.1169, "step": 536 }, { "epoch": 0.8920265780730897, "grad_norm": 8.275440216064453, "learning_rate": 1.222895850331485e-06, "loss": 0.1149, "step": 537 }, { "epoch": 0.893687707641196, "grad_norm": 7.478592872619629, "learning_rate": 1.220270746528497e-06, "loss": 0.1055, "step": 538 }, { "epoch": 0.8953488372093024, "grad_norm": 9.597691535949707, "learning_rate": 1.217644046425211e-06, "loss": 0.1186, "step": 539 }, { "epoch": 0.8970099667774086, "grad_norm": 12.220060348510742, "learning_rate": 1.2150157690573017e-06, "loss": 0.105, "step": 540 }, { "epoch": 0.8986710963455149, "grad_norm": 10.828849792480469, "learning_rate": 1.212385933471873e-06, "loss": 0.1782, "step": 541 }, { "epoch": 0.9003322259136213, "grad_norm": 7.446874141693115, "learning_rate": 1.209754558727322e-06, "loss": 0.1279, "step": 542 }, { "epoch": 0.9019933554817275, "grad_norm": 10.593635559082031, "learning_rate": 1.2071216638932e-06, "loss": 0.1389, "step": 543 }, { "epoch": 0.9036544850498339, "grad_norm": 6.1521148681640625, "learning_rate": 1.204487268050074e-06, "loss": 0.08, "step": 544 }, { "epoch": 0.9053156146179402, "grad_norm": 7.610264301300049, "learning_rate": 1.20185139028939e-06, "loss": 0.0942, "step": 545 }, { "epoch": 0.9069767441860465, "grad_norm": 4.965451240539551, "learning_rate": 1.199214049713332e-06, "loss": 0.0551, "step": 546 }, { "epoch": 0.9086378737541528, "grad_norm": 8.87406063079834, "learning_rate": 1.1965752654346853e-06, "loss": 0.1101, "step": 547 }, { "epoch": 0.9102990033222591, "grad_norm": 12.057513236999512, "learning_rate": 1.1939350565766986e-06, "loss": 0.1266, "step": 548 }, { "epoch": 0.9119601328903655, "grad_norm": 6.295923233032227, "learning_rate": 1.1912934422729433e-06, "loss": 0.0802, "step": 549 }, { "epoch": 0.9136212624584718, "grad_norm": 8.021347999572754, "learning_rate": 1.1886504416671768e-06, "loss": 0.1245, "step": 550 }, { "epoch": 0.915282392026578, "grad_norm": 19.368101119995117, "learning_rate": 1.1860060739132034e-06, "loss": 0.1785, "step": 551 }, { "epoch": 0.9169435215946844, "grad_norm": 6.273396968841553, "learning_rate": 1.1833603581747337e-06, "loss": 0.0842, "step": 552 }, { "epoch": 0.9186046511627907, "grad_norm": 11.780827522277832, "learning_rate": 1.180713313625249e-06, "loss": 0.1468, "step": 553 }, { "epoch": 0.920265780730897, "grad_norm": 11.886472702026367, "learning_rate": 1.1780649594478588e-06, "loss": 0.1338, "step": 554 }, { "epoch": 0.9219269102990033, "grad_norm": 15.677685737609863, "learning_rate": 1.1754153148351638e-06, "loss": 0.0981, "step": 555 }, { "epoch": 0.9235880398671097, "grad_norm": 15.455180168151855, "learning_rate": 1.172764398989118e-06, "loss": 0.1265, "step": 556 }, { "epoch": 0.925249169435216, "grad_norm": 11.938597679138184, "learning_rate": 1.1701122311208858e-06, "loss": 0.1704, "step": 557 }, { "epoch": 0.9269102990033222, "grad_norm": 6.007061004638672, "learning_rate": 1.1674588304507066e-06, "loss": 0.0721, "step": 558 }, { "epoch": 0.9285714285714286, "grad_norm": 6.777620792388916, "learning_rate": 1.1648042162077534e-06, "loss": 0.1089, "step": 559 }, { "epoch": 0.9302325581395349, "grad_norm": 6.482328414916992, "learning_rate": 1.1621484076299933e-06, "loss": 0.0737, "step": 560 }, { "epoch": 0.9318936877076412, "grad_norm": 13.06513786315918, "learning_rate": 1.15949142396405e-06, "loss": 0.1332, "step": 561 }, { "epoch": 0.9335548172757475, "grad_norm": 15.244009971618652, "learning_rate": 1.156833284465062e-06, "loss": 0.1236, "step": 562 }, { "epoch": 0.9352159468438538, "grad_norm": 9.853615760803223, "learning_rate": 1.1541740083965445e-06, "loss": 0.0926, "step": 563 }, { "epoch": 0.9368770764119602, "grad_norm": 17.81363868713379, "learning_rate": 1.1515136150302497e-06, "loss": 0.1233, "step": 564 }, { "epoch": 0.9385382059800664, "grad_norm": 17.04298973083496, "learning_rate": 1.1488521236460265e-06, "loss": 0.1688, "step": 565 }, { "epoch": 0.9401993355481728, "grad_norm": 17.051122665405273, "learning_rate": 1.1461895535316807e-06, "loss": 0.1891, "step": 566 }, { "epoch": 0.9418604651162791, "grad_norm": 7.387003421783447, "learning_rate": 1.143525923982837e-06, "loss": 0.1123, "step": 567 }, { "epoch": 0.9435215946843853, "grad_norm": 12.299041748046875, "learning_rate": 1.1408612543027963e-06, "loss": 0.1341, "step": 568 }, { "epoch": 0.9451827242524917, "grad_norm": 10.729867935180664, "learning_rate": 1.138195563802398e-06, "loss": 0.0759, "step": 569 }, { "epoch": 0.946843853820598, "grad_norm": 10.060506820678711, "learning_rate": 1.1355288717998797e-06, "loss": 0.1343, "step": 570 }, { "epoch": 0.9485049833887044, "grad_norm": 5.130414009094238, "learning_rate": 1.1328611976207355e-06, "loss": 0.0635, "step": 571 }, { "epoch": 0.9501661129568106, "grad_norm": 8.048657417297363, "learning_rate": 1.13019256059758e-06, "loss": 0.093, "step": 572 }, { "epoch": 0.9518272425249169, "grad_norm": 11.608697891235352, "learning_rate": 1.1275229800700028e-06, "loss": 0.1367, "step": 573 }, { "epoch": 0.9534883720930233, "grad_norm": 9.293169021606445, "learning_rate": 1.1248524753844324e-06, "loss": 0.1262, "step": 574 }, { "epoch": 0.9551495016611296, "grad_norm": 8.593324661254883, "learning_rate": 1.1221810658939946e-06, "loss": 0.0909, "step": 575 }, { "epoch": 0.9568106312292359, "grad_norm": 7.67872953414917, "learning_rate": 1.1195087709583713e-06, "loss": 0.0845, "step": 576 }, { "epoch": 0.9584717607973422, "grad_norm": 5.514711856842041, "learning_rate": 1.1168356099436628e-06, "loss": 0.0699, "step": 577 }, { "epoch": 0.9601328903654485, "grad_norm": 8.198745727539062, "learning_rate": 1.1141616022222453e-06, "loss": 0.1107, "step": 578 }, { "epoch": 0.9617940199335548, "grad_norm": 10.59040355682373, "learning_rate": 1.1114867671726302e-06, "loss": 0.1206, "step": 579 }, { "epoch": 0.9634551495016611, "grad_norm": 10.234724998474121, "learning_rate": 1.1088111241793258e-06, "loss": 0.0797, "step": 580 }, { "epoch": 0.9651162790697675, "grad_norm": 4.5755486488342285, "learning_rate": 1.1061346926326942e-06, "loss": 0.0451, "step": 581 }, { "epoch": 0.9667774086378738, "grad_norm": 12.668401718139648, "learning_rate": 1.1034574919288137e-06, "loss": 0.0975, "step": 582 }, { "epoch": 0.96843853820598, "grad_norm": 4.996627330780029, "learning_rate": 1.100779541469336e-06, "loss": 0.0908, "step": 583 }, { "epoch": 0.9700996677740864, "grad_norm": 7.567203998565674, "learning_rate": 1.0981008606613454e-06, "loss": 0.0926, "step": 584 }, { "epoch": 0.9717607973421927, "grad_norm": 7.270530700683594, "learning_rate": 1.0954214689172203e-06, "loss": 0.0851, "step": 585 }, { "epoch": 0.973421926910299, "grad_norm": 11.763467788696289, "learning_rate": 1.0927413856544905e-06, "loss": 0.1197, "step": 586 }, { "epoch": 0.9750830564784053, "grad_norm": 10.839402198791504, "learning_rate": 1.0900606302956978e-06, "loss": 0.0805, "step": 587 }, { "epoch": 0.9767441860465116, "grad_norm": 8.311976432800293, "learning_rate": 1.087379222268254e-06, "loss": 0.1222, "step": 588 }, { "epoch": 0.978405315614618, "grad_norm": 16.371187210083008, "learning_rate": 1.0846971810043015e-06, "loss": 0.1755, "step": 589 }, { "epoch": 0.9800664451827242, "grad_norm": 8.993597030639648, "learning_rate": 1.0820145259405713e-06, "loss": 0.1057, "step": 590 }, { "epoch": 0.9817275747508306, "grad_norm": 8.658944129943848, "learning_rate": 1.0793312765182426e-06, "loss": 0.085, "step": 591 }, { "epoch": 0.9833887043189369, "grad_norm": 17.43745994567871, "learning_rate": 1.076647452182802e-06, "loss": 0.1772, "step": 592 }, { "epoch": 0.9850498338870431, "grad_norm": 12.07256031036377, "learning_rate": 1.0739630723839021e-06, "loss": 0.1592, "step": 593 }, { "epoch": 0.9867109634551495, "grad_norm": 14.670401573181152, "learning_rate": 1.0712781565752222e-06, "loss": 0.1656, "step": 594 }, { "epoch": 0.9883720930232558, "grad_norm": 18.60042953491211, "learning_rate": 1.0685927242143246e-06, "loss": 0.142, "step": 595 }, { "epoch": 0.9900332225913622, "grad_norm": 7.408260345458984, "learning_rate": 1.0659067947625161e-06, "loss": 0.1022, "step": 596 }, { "epoch": 0.9916943521594684, "grad_norm": 7.936831951141357, "learning_rate": 1.0632203876847048e-06, "loss": 0.0884, "step": 597 }, { "epoch": 0.9933554817275747, "grad_norm": 11.372370719909668, "learning_rate": 1.0605335224492615e-06, "loss": 0.1498, "step": 598 }, { "epoch": 0.9950166112956811, "grad_norm": 7.789324760437012, "learning_rate": 1.057846218527876e-06, "loss": 0.0774, "step": 599 }, { "epoch": 0.9966777408637874, "grad_norm": 11.030444145202637, "learning_rate": 1.0551584953954182e-06, "loss": 0.1359, "step": 600 }, { "epoch": 0.9983388704318937, "grad_norm": 7.336810111999512, "learning_rate": 1.0524703725297956e-06, "loss": 0.0632, "step": 601 }, { "epoch": 1.0, "grad_norm": 10.567761421203613, "learning_rate": 1.0497818694118122e-06, "loss": 0.1714, "step": 602 }, { "epoch": 1.0016611295681064, "grad_norm": 2.572457790374756, "learning_rate": 1.0470930055250292e-06, "loss": 0.0303, "step": 603 }, { "epoch": 1.0033222591362125, "grad_norm": 2.8890881538391113, "learning_rate": 1.04440380035562e-06, "loss": 0.0319, "step": 604 }, { "epoch": 1.004983388704319, "grad_norm": 2.415273904800415, "learning_rate": 1.041714273392233e-06, "loss": 0.0177, "step": 605 }, { "epoch": 1.0066445182724253, "grad_norm": 2.018181800842285, "learning_rate": 1.0390244441258478e-06, "loss": 0.0181, "step": 606 }, { "epoch": 1.0083056478405317, "grad_norm": 1.576615571975708, "learning_rate": 1.0363343320496357e-06, "loss": 0.0164, "step": 607 }, { "epoch": 1.0099667774086378, "grad_norm": 3.9152066707611084, "learning_rate": 1.033643956658816e-06, "loss": 0.0152, "step": 608 }, { "epoch": 1.0116279069767442, "grad_norm": 9.031929969787598, "learning_rate": 1.0309533374505182e-06, "loss": 0.0461, "step": 609 }, { "epoch": 1.0132890365448506, "grad_norm": 3.230257272720337, "learning_rate": 1.0282624939236367e-06, "loss": 0.017, "step": 610 }, { "epoch": 1.0149501661129567, "grad_norm": 7.940582752227783, "learning_rate": 1.025571445578693e-06, "loss": 0.0407, "step": 611 }, { "epoch": 1.0166112956810631, "grad_norm": 5.057779788970947, "learning_rate": 1.0228802119176927e-06, "loss": 0.0452, "step": 612 }, { "epoch": 1.0182724252491695, "grad_norm": 18.299306869506836, "learning_rate": 1.0201888124439834e-06, "loss": 0.0196, "step": 613 }, { "epoch": 1.0199335548172757, "grad_norm": 5.936789035797119, "learning_rate": 1.017497266662116e-06, "loss": 0.0333, "step": 614 }, { "epoch": 1.021594684385382, "grad_norm": 3.513627767562866, "learning_rate": 1.0148055940776998e-06, "loss": 0.0084, "step": 615 }, { "epoch": 1.0232558139534884, "grad_norm": 0.37478646636009216, "learning_rate": 1.0121138141972648e-06, "loss": 0.0011, "step": 616 }, { "epoch": 1.0249169435215948, "grad_norm": 2.6802878379821777, "learning_rate": 1.0094219465281172e-06, "loss": 0.0082, "step": 617 }, { "epoch": 1.026578073089701, "grad_norm": 4.141627311706543, "learning_rate": 1.0067300105782002e-06, "loss": 0.0133, "step": 618 }, { "epoch": 1.0282392026578073, "grad_norm": 1.266526460647583, "learning_rate": 1.0040380258559518e-06, "loss": 0.0041, "step": 619 }, { "epoch": 1.0299003322259137, "grad_norm": 7.685907363891602, "learning_rate": 1.001346011870163e-06, "loss": 0.0261, "step": 620 }, { "epoch": 1.0315614617940199, "grad_norm": 6.315763473510742, "learning_rate": 9.986539881298368e-07, "loss": 0.0089, "step": 621 }, { "epoch": 1.0332225913621262, "grad_norm": 2.4348950386047363, "learning_rate": 9.959619741440483e-07, "loss": 0.0061, "step": 622 }, { "epoch": 1.0348837209302326, "grad_norm": 12.046463012695312, "learning_rate": 9.932699894217997e-07, "loss": 0.0241, "step": 623 }, { "epoch": 1.0365448504983388, "grad_norm": 7.545411586761475, "learning_rate": 9.905780534718825e-07, "loss": 0.0335, "step": 624 }, { "epoch": 1.0382059800664452, "grad_norm": 0.5908049941062927, "learning_rate": 9.878861858027353e-07, "loss": 0.0015, "step": 625 }, { "epoch": 1.0398671096345515, "grad_norm": 0.7220317125320435, "learning_rate": 9.851944059223001e-07, "loss": 0.0015, "step": 626 }, { "epoch": 1.041528239202658, "grad_norm": 0.47070062160491943, "learning_rate": 9.825027333378842e-07, "loss": 0.001, "step": 627 }, { "epoch": 1.043189368770764, "grad_norm": 1.697333574295044, "learning_rate": 9.798111875560165e-07, "loss": 0.0033, "step": 628 }, { "epoch": 1.0448504983388704, "grad_norm": 0.6291197538375854, "learning_rate": 9.771197880823072e-07, "loss": 0.0008, "step": 629 }, { "epoch": 1.0465116279069768, "grad_norm": 56.27286148071289, "learning_rate": 9.74428554421307e-07, "loss": 0.1389, "step": 630 }, { "epoch": 1.048172757475083, "grad_norm": 9.490167617797852, "learning_rate": 9.717375060763635e-07, "loss": 0.0348, "step": 631 }, { "epoch": 1.0498338870431894, "grad_norm": 12.414752006530762, "learning_rate": 9.690466625494822e-07, "loss": 0.0156, "step": 632 }, { "epoch": 1.0514950166112957, "grad_norm": 11.833617210388184, "learning_rate": 9.663560433411839e-07, "loss": 0.0161, "step": 633 }, { "epoch": 1.053156146179402, "grad_norm": 12.894364356994629, "learning_rate": 9.636656679503647e-07, "loss": 0.0187, "step": 634 }, { "epoch": 1.0548172757475083, "grad_norm": 17.071706771850586, "learning_rate": 9.609755558741523e-07, "loss": 0.096, "step": 635 }, { "epoch": 1.0564784053156147, "grad_norm": 0.9374960064888, "learning_rate": 9.58285726607767e-07, "loss": 0.0029, "step": 636 }, { "epoch": 1.058139534883721, "grad_norm": 20.89844512939453, "learning_rate": 9.555961996443802e-07, "loss": 0.1398, "step": 637 }, { "epoch": 1.0598006644518272, "grad_norm": 14.290300369262695, "learning_rate": 9.52906994474971e-07, "loss": 0.0158, "step": 638 }, { "epoch": 1.0614617940199336, "grad_norm": 38.73701477050781, "learning_rate": 9.502181305881876e-07, "loss": 0.0567, "step": 639 }, { "epoch": 1.06312292358804, "grad_norm": 14.144023895263672, "learning_rate": 9.475296274702043e-07, "loss": 0.0168, "step": 640 }, { "epoch": 1.064784053156146, "grad_norm": 16.016897201538086, "learning_rate": 9.448415046045817e-07, "loss": 0.0747, "step": 641 }, { "epoch": 1.0664451827242525, "grad_norm": 1.7351895570755005, "learning_rate": 9.42153781472124e-07, "loss": 0.0036, "step": 642 }, { "epoch": 1.0681063122923589, "grad_norm": 0.4755818545818329, "learning_rate": 9.394664775507385e-07, "loss": 0.0007, "step": 643 }, { "epoch": 1.069767441860465, "grad_norm": 2.340486526489258, "learning_rate": 9.367796123152951e-07, "loss": 0.0104, "step": 644 }, { "epoch": 1.0714285714285714, "grad_norm": 5.121689796447754, "learning_rate": 9.340932052374839e-07, "loss": 0.006, "step": 645 }, { "epoch": 1.0730897009966778, "grad_norm": 9.980019569396973, "learning_rate": 9.314072757856752e-07, "loss": 0.024, "step": 646 }, { "epoch": 1.0747508305647842, "grad_norm": 12.077081680297852, "learning_rate": 9.287218434247776e-07, "loss": 0.0776, "step": 647 }, { "epoch": 1.0764119601328903, "grad_norm": 10.7062349319458, "learning_rate": 9.260369276160976e-07, "loss": 0.0303, "step": 648 }, { "epoch": 1.0780730897009967, "grad_norm": 7.534442901611328, "learning_rate": 9.233525478171984e-07, "loss": 0.0474, "step": 649 }, { "epoch": 1.079734219269103, "grad_norm": 1.34402334690094, "learning_rate": 9.206687234817574e-07, "loss": 0.0015, "step": 650 }, { "epoch": 1.0813953488372092, "grad_norm": 2.5344579219818115, "learning_rate": 9.179854740594285e-07, "loss": 0.0074, "step": 651 }, { "epoch": 1.0830564784053156, "grad_norm": 4.098090648651123, "learning_rate": 9.153028189956985e-07, "loss": 0.0106, "step": 652 }, { "epoch": 1.084717607973422, "grad_norm": 8.03905200958252, "learning_rate": 9.126207777317457e-07, "loss": 0.0158, "step": 653 }, { "epoch": 1.0863787375415281, "grad_norm": 11.426288604736328, "learning_rate": 9.099393697043023e-07, "loss": 0.024, "step": 654 }, { "epoch": 1.0880398671096345, "grad_norm": 18.223331451416016, "learning_rate": 9.072586143455095e-07, "loss": 0.079, "step": 655 }, { "epoch": 1.089700996677741, "grad_norm": 6.2648844718933105, "learning_rate": 9.045785310827799e-07, "loss": 0.0076, "step": 656 }, { "epoch": 1.0913621262458473, "grad_norm": 22.87317657470703, "learning_rate": 9.018991393386548e-07, "loss": 0.0205, "step": 657 }, { "epoch": 1.0930232558139534, "grad_norm": 9.388856887817383, "learning_rate": 8.99220458530664e-07, "loss": 0.0443, "step": 658 }, { "epoch": 1.0946843853820598, "grad_norm": 4.490157127380371, "learning_rate": 8.965425080711863e-07, "loss": 0.014, "step": 659 }, { "epoch": 1.0963455149501662, "grad_norm": 1.3594952821731567, "learning_rate": 8.938653073673057e-07, "loss": 0.0053, "step": 660 }, { "epoch": 1.0980066445182723, "grad_norm": 8.04047679901123, "learning_rate": 8.911888758206746e-07, "loss": 0.036, "step": 661 }, { "epoch": 1.0996677740863787, "grad_norm": 11.46448040008545, "learning_rate": 8.885132328273699e-07, "loss": 0.0622, "step": 662 }, { "epoch": 1.101328903654485, "grad_norm": 9.573933601379395, "learning_rate": 8.858383977777549e-07, "loss": 0.052, "step": 663 }, { "epoch": 1.1029900332225913, "grad_norm": 13.688606262207031, "learning_rate": 8.831643900563372e-07, "loss": 0.0361, "step": 664 }, { "epoch": 1.1046511627906976, "grad_norm": 2.0683696269989014, "learning_rate": 8.804912290416286e-07, "loss": 0.0057, "step": 665 }, { "epoch": 1.106312292358804, "grad_norm": 2.5601813793182373, "learning_rate": 8.778189341060058e-07, "loss": 0.0032, "step": 666 }, { "epoch": 1.1079734219269104, "grad_norm": 14.53157901763916, "learning_rate": 8.751475246155677e-07, "loss": 0.0506, "step": 667 }, { "epoch": 1.1096345514950166, "grad_norm": 4.818484783172607, "learning_rate": 8.724770199299972e-07, "loss": 0.018, "step": 668 }, { "epoch": 1.111295681063123, "grad_norm": 25.98501205444336, "learning_rate": 8.6980743940242e-07, "loss": 0.188, "step": 669 }, { "epoch": 1.1129568106312293, "grad_norm": 29.606630325317383, "learning_rate": 8.67138802379264e-07, "loss": 0.0453, "step": 670 }, { "epoch": 1.1146179401993355, "grad_norm": 9.709136962890625, "learning_rate": 8.644711282001207e-07, "loss": 0.0377, "step": 671 }, { "epoch": 1.1162790697674418, "grad_norm": 8.595226287841797, "learning_rate": 8.61804436197602e-07, "loss": 0.0385, "step": 672 }, { "epoch": 1.1179401993355482, "grad_norm": 2.2285189628601074, "learning_rate": 8.591387456972039e-07, "loss": 0.0076, "step": 673 }, { "epoch": 1.1196013289036544, "grad_norm": 7.5171990394592285, "learning_rate": 8.564740760171631e-07, "loss": 0.0803, "step": 674 }, { "epoch": 1.1212624584717608, "grad_norm": 12.671175956726074, "learning_rate": 8.53810446468319e-07, "loss": 0.0193, "step": 675 }, { "epoch": 1.1229235880398671, "grad_norm": 6.085766315460205, "learning_rate": 8.511478763539736e-07, "loss": 0.0109, "step": 676 }, { "epoch": 1.1245847176079735, "grad_norm": 8.109288215637207, "learning_rate": 8.484863849697501e-07, "loss": 0.0229, "step": 677 }, { "epoch": 1.1262458471760797, "grad_norm": 11.192323684692383, "learning_rate": 8.458259916034556e-07, "loss": 0.0367, "step": 678 }, { "epoch": 1.127906976744186, "grad_norm": 6.7384796142578125, "learning_rate": 8.43166715534938e-07, "loss": 0.0242, "step": 679 }, { "epoch": 1.1295681063122924, "grad_norm": 6.001007080078125, "learning_rate": 8.405085760359499e-07, "loss": 0.0387, "step": 680 }, { "epoch": 1.1312292358803986, "grad_norm": 4.371544361114502, "learning_rate": 8.378515923700069e-07, "loss": 0.0132, "step": 681 }, { "epoch": 1.132890365448505, "grad_norm": 10.13823413848877, "learning_rate": 8.351957837922467e-07, "loss": 0.0236, "step": 682 }, { "epoch": 1.1345514950166113, "grad_norm": 6.677521228790283, "learning_rate": 8.325411695492936e-07, "loss": 0.0451, "step": 683 }, { "epoch": 1.1362126245847177, "grad_norm": 9.64169692993164, "learning_rate": 8.298877688791143e-07, "loss": 0.0266, "step": 684 }, { "epoch": 1.1378737541528239, "grad_norm": 15.662538528442383, "learning_rate": 8.27235601010882e-07, "loss": 0.0218, "step": 685 }, { "epoch": 1.1395348837209303, "grad_norm": 8.98953628540039, "learning_rate": 8.245846851648363e-07, "loss": 0.031, "step": 686 }, { "epoch": 1.1411960132890366, "grad_norm": 2.669698476791382, "learning_rate": 8.219350405521414e-07, "loss": 0.0055, "step": 687 }, { "epoch": 1.1428571428571428, "grad_norm": 4.246596813201904, "learning_rate": 8.192866863747514e-07, "loss": 0.0153, "step": 688 }, { "epoch": 1.1445182724252492, "grad_norm": 8.687833786010742, "learning_rate": 8.166396418252663e-07, "loss": 0.0387, "step": 689 }, { "epoch": 1.1461794019933556, "grad_norm": 10.32325267791748, "learning_rate": 8.139939260867965e-07, "loss": 0.0539, "step": 690 }, { "epoch": 1.1478405315614617, "grad_norm": 3.5581681728363037, "learning_rate": 8.113495583328232e-07, "loss": 0.0145, "step": 691 }, { "epoch": 1.149501661129568, "grad_norm": 0.5473489165306091, "learning_rate": 8.087065577270569e-07, "loss": 0.0009, "step": 692 }, { "epoch": 1.1511627906976745, "grad_norm": 4.436628818511963, "learning_rate": 8.060649434233016e-07, "loss": 0.0103, "step": 693 }, { "epoch": 1.1528239202657806, "grad_norm": 12.929503440856934, "learning_rate": 8.034247345653147e-07, "loss": 0.0396, "step": 694 }, { "epoch": 1.154485049833887, "grad_norm": 15.860251426696777, "learning_rate": 8.007859502866683e-07, "loss": 0.0702, "step": 695 }, { "epoch": 1.1561461794019934, "grad_norm": 1.6955807209014893, "learning_rate": 7.9814860971061e-07, "loss": 0.0069, "step": 696 }, { "epoch": 1.1578073089700998, "grad_norm": 3.191066265106201, "learning_rate": 7.955127319499257e-07, "loss": 0.0143, "step": 697 }, { "epoch": 1.159468438538206, "grad_norm": 1.4074821472167969, "learning_rate": 7.928783361068001e-07, "loss": 0.0024, "step": 698 }, { "epoch": 1.1611295681063123, "grad_norm": 2.3482460975646973, "learning_rate": 7.90245441272678e-07, "loss": 0.01, "step": 699 }, { "epoch": 1.1627906976744187, "grad_norm": 14.107357025146484, "learning_rate": 7.876140665281271e-07, "loss": 0.0377, "step": 700 }, { "epoch": 1.1644518272425248, "grad_norm": 19.679927825927734, "learning_rate": 7.849842309426983e-07, "loss": 0.0264, "step": 701 }, { "epoch": 1.1661129568106312, "grad_norm": 32.07101058959961, "learning_rate": 7.823559535747885e-07, "loss": 0.1096, "step": 702 }, { "epoch": 1.1677740863787376, "grad_norm": 14.08077621459961, "learning_rate": 7.79729253471503e-07, "loss": 0.1361, "step": 703 }, { "epoch": 1.169435215946844, "grad_norm": 1.9143798351287842, "learning_rate": 7.771041496685149e-07, "loss": 0.0043, "step": 704 }, { "epoch": 1.1710963455149501, "grad_norm": 6.996606349945068, "learning_rate": 7.744806611899309e-07, "loss": 0.0085, "step": 705 }, { "epoch": 1.1727574750830565, "grad_norm": 31.96712875366211, "learning_rate": 7.7185880704815e-07, "loss": 0.0564, "step": 706 }, { "epoch": 1.1744186046511629, "grad_norm": 5.491878986358643, "learning_rate": 7.692386062437279e-07, "loss": 0.0083, "step": 707 }, { "epoch": 1.176079734219269, "grad_norm": 7.473092079162598, "learning_rate": 7.666200777652384e-07, "loss": 0.0191, "step": 708 }, { "epoch": 1.1777408637873754, "grad_norm": 22.00745964050293, "learning_rate": 7.640032405891359e-07, "loss": 0.0485, "step": 709 }, { "epoch": 1.1794019933554818, "grad_norm": 7.070401668548584, "learning_rate": 7.613881136796183e-07, "loss": 0.0273, "step": 710 }, { "epoch": 1.181063122923588, "grad_norm": 6.378243923187256, "learning_rate": 7.587747159884881e-07, "loss": 0.0227, "step": 711 }, { "epoch": 1.1827242524916943, "grad_norm": 3.251873254776001, "learning_rate": 7.561630664550179e-07, "loss": 0.0059, "step": 712 }, { "epoch": 1.1843853820598007, "grad_norm": 0.48984941840171814, "learning_rate": 7.535531840058101e-07, "loss": 0.0006, "step": 713 }, { "epoch": 1.1860465116279069, "grad_norm": 14.683696746826172, "learning_rate": 7.509450875546615e-07, "loss": 0.0288, "step": 714 }, { "epoch": 1.1877076411960132, "grad_norm": 5.808125972747803, "learning_rate": 7.48338796002426e-07, "loss": 0.0081, "step": 715 }, { "epoch": 1.1893687707641196, "grad_norm": 0.9389947056770325, "learning_rate": 7.457343282368763e-07, "loss": 0.0014, "step": 716 }, { "epoch": 1.191029900332226, "grad_norm": 18.770702362060547, "learning_rate": 7.431317031325703e-07, "loss": 0.0237, "step": 717 }, { "epoch": 1.1926910299003322, "grad_norm": 4.809552192687988, "learning_rate": 7.405309395507096e-07, "loss": 0.0111, "step": 718 }, { "epoch": 1.1943521594684385, "grad_norm": 14.019580841064453, "learning_rate": 7.37932056339007e-07, "loss": 0.0242, "step": 719 }, { "epoch": 1.196013289036545, "grad_norm": 1.6724600791931152, "learning_rate": 7.353350723315485e-07, "loss": 0.0013, "step": 720 }, { "epoch": 1.197674418604651, "grad_norm": 24.528261184692383, "learning_rate": 7.327400063486553e-07, "loss": 0.0404, "step": 721 }, { "epoch": 1.1993355481727574, "grad_norm": 9.839111328125, "learning_rate": 7.301468771967495e-07, "loss": 0.022, "step": 722 }, { "epoch": 1.2009966777408638, "grad_norm": 169.1666259765625, "learning_rate": 7.275557036682167e-07, "loss": 0.0955, "step": 723 }, { "epoch": 1.2026578073089702, "grad_norm": 3.6535372734069824, "learning_rate": 7.249665045412702e-07, "loss": 0.0075, "step": 724 }, { "epoch": 1.2043189368770764, "grad_norm": 16.60807228088379, "learning_rate": 7.223792985798156e-07, "loss": 0.0245, "step": 725 }, { "epoch": 1.2059800664451827, "grad_norm": 6.681334018707275, "learning_rate": 7.197941045333125e-07, "loss": 0.0332, "step": 726 }, { "epoch": 1.2076411960132891, "grad_norm": 7.528415203094482, "learning_rate": 7.172109411366416e-07, "loss": 0.0486, "step": 727 }, { "epoch": 1.2093023255813953, "grad_norm": 7.378742694854736, "learning_rate": 7.146298271099665e-07, "loss": 0.0142, "step": 728 }, { "epoch": 1.2109634551495017, "grad_norm": 0.4893551468849182, "learning_rate": 7.120507811585994e-07, "loss": 0.0008, "step": 729 }, { "epoch": 1.212624584717608, "grad_norm": 12.16398811340332, "learning_rate": 7.094738219728656e-07, "loss": 0.0503, "step": 730 }, { "epoch": 1.2142857142857142, "grad_norm": 0.8328701257705688, "learning_rate": 7.068989682279669e-07, "loss": 0.0011, "step": 731 }, { "epoch": 1.2159468438538206, "grad_norm": 5.240982532501221, "learning_rate": 7.043262385838475e-07, "loss": 0.0211, "step": 732 }, { "epoch": 1.217607973421927, "grad_norm": 3.157712936401367, "learning_rate": 7.01755651685058e-07, "loss": 0.0129, "step": 733 }, { "epoch": 1.219269102990033, "grad_norm": 5.299394607543945, "learning_rate": 6.991872261606206e-07, "loss": 0.0166, "step": 734 }, { "epoch": 1.2209302325581395, "grad_norm": 5.783647060394287, "learning_rate": 6.966209806238942e-07, "loss": 0.024, "step": 735 }, { "epoch": 1.2225913621262459, "grad_norm": 29.13331413269043, "learning_rate": 6.940569336724389e-07, "loss": 0.0335, "step": 736 }, { "epoch": 1.2242524916943522, "grad_norm": 8.021717071533203, "learning_rate": 6.914951038878826e-07, "loss": 0.0254, "step": 737 }, { "epoch": 1.2259136212624584, "grad_norm": 18.33582305908203, "learning_rate": 6.889355098357841e-07, "loss": 0.0814, "step": 738 }, { "epoch": 1.2275747508305648, "grad_norm": 3.572312116622925, "learning_rate": 6.863781700655012e-07, "loss": 0.0147, "step": 739 }, { "epoch": 1.2292358803986712, "grad_norm": 2.5561416149139404, "learning_rate": 6.838231031100532e-07, "loss": 0.0034, "step": 740 }, { "epoch": 1.2308970099667773, "grad_norm": 3.5589852333068848, "learning_rate": 6.812703274859905e-07, "loss": 0.0022, "step": 741 }, { "epoch": 1.2325581395348837, "grad_norm": 3.630929946899414, "learning_rate": 6.78719861693257e-07, "loss": 0.0095, "step": 742 }, { "epoch": 1.23421926910299, "grad_norm": 16.794384002685547, "learning_rate": 6.761717242150571e-07, "loss": 0.0761, "step": 743 }, { "epoch": 1.2358803986710964, "grad_norm": 4.896450042724609, "learning_rate": 6.736259335177228e-07, "loss": 0.0075, "step": 744 }, { "epoch": 1.2375415282392026, "grad_norm": 6.048307418823242, "learning_rate": 6.710825080505774e-07, "loss": 0.0125, "step": 745 }, { "epoch": 1.239202657807309, "grad_norm": 12.081902503967285, "learning_rate": 6.685414662458054e-07, "loss": 0.0468, "step": 746 }, { "epoch": 1.2408637873754154, "grad_norm": 30.948566436767578, "learning_rate": 6.660028265183159e-07, "loss": 0.0549, "step": 747 }, { "epoch": 1.2425249169435215, "grad_norm": 5.927905082702637, "learning_rate": 6.634666072656097e-07, "loss": 0.0345, "step": 748 }, { "epoch": 1.244186046511628, "grad_norm": 15.502345085144043, "learning_rate": 6.609328268676476e-07, "loss": 0.0928, "step": 749 }, { "epoch": 1.2458471760797343, "grad_norm": 6.187903881072998, "learning_rate": 6.584015036867149e-07, "loss": 0.0139, "step": 750 }, { "epoch": 1.2475083056478407, "grad_norm": 2.137287139892578, "learning_rate": 6.558726560672898e-07, "loss": 0.0027, "step": 751 }, { "epoch": 1.2491694352159468, "grad_norm": 11.991990089416504, "learning_rate": 6.533463023359114e-07, "loss": 0.0323, "step": 752 }, { "epoch": 1.2508305647840532, "grad_norm": 4.784742832183838, "learning_rate": 6.508224608010435e-07, "loss": 0.0105, "step": 753 }, { "epoch": 1.2524916943521593, "grad_norm": 11.65071964263916, "learning_rate": 6.483011497529456e-07, "loss": 0.0182, "step": 754 }, { "epoch": 1.2541528239202657, "grad_norm": 18.266075134277344, "learning_rate": 6.457823874635376e-07, "loss": 0.0502, "step": 755 }, { "epoch": 1.255813953488372, "grad_norm": 8.90103530883789, "learning_rate": 6.432661921862699e-07, "loss": 0.0364, "step": 756 }, { "epoch": 1.2574750830564785, "grad_norm": 0.34128686785697937, "learning_rate": 6.407525821559888e-07, "loss": 0.001, "step": 757 }, { "epoch": 1.2591362126245846, "grad_norm": 8.536415100097656, "learning_rate": 6.382415755888053e-07, "loss": 0.0258, "step": 758 }, { "epoch": 1.260797342192691, "grad_norm": 0.3508094549179077, "learning_rate": 6.35733190681964e-07, "loss": 0.006, "step": 759 }, { "epoch": 1.2624584717607974, "grad_norm": 5.0073628425598145, "learning_rate": 6.332274456137096e-07, "loss": 0.0098, "step": 760 }, { "epoch": 1.2641196013289036, "grad_norm": 6.868859767913818, "learning_rate": 6.307243585431562e-07, "loss": 0.03, "step": 761 }, { "epoch": 1.26578073089701, "grad_norm": 5.491968154907227, "learning_rate": 6.282239476101549e-07, "loss": 0.004, "step": 762 }, { "epoch": 1.2674418604651163, "grad_norm": 16.57515525817871, "learning_rate": 6.257262309351636e-07, "loss": 0.0318, "step": 763 }, { "epoch": 1.2691029900332227, "grad_norm": 10.209383010864258, "learning_rate": 6.232312266191149e-07, "loss": 0.0263, "step": 764 }, { "epoch": 1.2707641196013288, "grad_norm": 10.865987777709961, "learning_rate": 6.20738952743284e-07, "loss": 0.0335, "step": 765 }, { "epoch": 1.2724252491694352, "grad_norm": 5.695135593414307, "learning_rate": 6.1824942736916e-07, "loss": 0.0256, "step": 766 }, { "epoch": 1.2740863787375416, "grad_norm": 2.8923935890197754, "learning_rate": 6.157626685383123e-07, "loss": 0.0047, "step": 767 }, { "epoch": 1.2757475083056478, "grad_norm": 4.0880842208862305, "learning_rate": 6.13278694272262e-07, "loss": 0.0114, "step": 768 }, { "epoch": 1.2774086378737541, "grad_norm": 22.687896728515625, "learning_rate": 6.107975225723505e-07, "loss": 0.0596, "step": 769 }, { "epoch": 1.2790697674418605, "grad_norm": 0.6194286346435547, "learning_rate": 6.083191714196085e-07, "loss": 0.0012, "step": 770 }, { "epoch": 1.280730897009967, "grad_norm": 21.863033294677734, "learning_rate": 6.058436587746269e-07, "loss": 0.0596, "step": 771 }, { "epoch": 1.282392026578073, "grad_norm": 1.8822462558746338, "learning_rate": 6.033710025774253e-07, "loss": 0.0043, "step": 772 }, { "epoch": 1.2840531561461794, "grad_norm": 6.4909515380859375, "learning_rate": 6.009012207473228e-07, "loss": 0.0152, "step": 773 }, { "epoch": 1.2857142857142856, "grad_norm": 1.9386980533599854, "learning_rate": 5.984343311828086e-07, "loss": 0.0061, "step": 774 }, { "epoch": 1.287375415282392, "grad_norm": 8.743645668029785, "learning_rate": 5.959703517614106e-07, "loss": 0.0134, "step": 775 }, { "epoch": 1.2890365448504983, "grad_norm": 23.125574111938477, "learning_rate": 5.935093003395684e-07, "loss": 0.0258, "step": 776 }, { "epoch": 1.2906976744186047, "grad_norm": 2.1917898654937744, "learning_rate": 5.910511947525001e-07, "loss": 0.0058, "step": 777 }, { "epoch": 1.292358803986711, "grad_norm": 4.327230930328369, "learning_rate": 5.885960528140783e-07, "loss": 0.0111, "step": 778 }, { "epoch": 1.2940199335548173, "grad_norm": 9.585992813110352, "learning_rate": 5.861438923166955e-07, "loss": 0.0214, "step": 779 }, { "epoch": 1.2956810631229236, "grad_norm": 16.866348266601562, "learning_rate": 5.836947310311388e-07, "loss": 0.0145, "step": 780 }, { "epoch": 1.2973421926910298, "grad_norm": 29.587238311767578, "learning_rate": 5.812485867064607e-07, "loss": 0.0262, "step": 781 }, { "epoch": 1.2990033222591362, "grad_norm": 3.231372117996216, "learning_rate": 5.788054770698487e-07, "loss": 0.0076, "step": 782 }, { "epoch": 1.3006644518272426, "grad_norm": 8.976773262023926, "learning_rate": 5.763654198264985e-07, "loss": 0.0182, "step": 783 }, { "epoch": 1.302325581395349, "grad_norm": 9.699551582336426, "learning_rate": 5.739284326594844e-07, "loss": 0.0214, "step": 784 }, { "epoch": 1.303986710963455, "grad_norm": 25.53984832763672, "learning_rate": 5.714945332296333e-07, "loss": 0.0154, "step": 785 }, { "epoch": 1.3056478405315615, "grad_norm": 1.9964032173156738, "learning_rate": 5.690637391753941e-07, "loss": 0.0043, "step": 786 }, { "epoch": 1.3073089700996678, "grad_norm": 28.76060676574707, "learning_rate": 5.666360681127108e-07, "loss": 0.0258, "step": 787 }, { "epoch": 1.308970099667774, "grad_norm": 3.6327102184295654, "learning_rate": 5.642115376348972e-07, "loss": 0.0049, "step": 788 }, { "epoch": 1.3106312292358804, "grad_norm": 1.5518959760665894, "learning_rate": 5.617901653125042e-07, "loss": 0.0064, "step": 789 }, { "epoch": 1.3122923588039868, "grad_norm": 5.097245216369629, "learning_rate": 5.59371968693198e-07, "loss": 0.0039, "step": 790 }, { "epoch": 1.3139534883720931, "grad_norm": 1.6257424354553223, "learning_rate": 5.569569653016299e-07, "loss": 0.002, "step": 791 }, { "epoch": 1.3156146179401993, "grad_norm": 1.6841596364974976, "learning_rate": 5.545451726393097e-07, "loss": 0.0031, "step": 792 }, { "epoch": 1.3172757475083057, "grad_norm": 33.802764892578125, "learning_rate": 5.521366081844787e-07, "loss": 0.0603, "step": 793 }, { "epoch": 1.3189368770764118, "grad_norm": 6.568824291229248, "learning_rate": 5.497312893919837e-07, "loss": 0.0082, "step": 794 }, { "epoch": 1.3205980066445182, "grad_norm": 2.7456605434417725, "learning_rate": 5.47329233693151e-07, "loss": 0.0081, "step": 795 }, { "epoch": 1.3222591362126246, "grad_norm": 27.489286422729492, "learning_rate": 5.449304584956581e-07, "loss": 0.1081, "step": 796 }, { "epoch": 1.323920265780731, "grad_norm": 4.846098899841309, "learning_rate": 5.42534981183409e-07, "loss": 0.0064, "step": 797 }, { "epoch": 1.3255813953488373, "grad_norm": 9.961216926574707, "learning_rate": 5.401428191164092e-07, "loss": 0.0215, "step": 798 }, { "epoch": 1.3272425249169435, "grad_norm": 10.51478099822998, "learning_rate": 5.377539896306363e-07, "loss": 0.0214, "step": 799 }, { "epoch": 1.3289036544850499, "grad_norm": 23.659414291381836, "learning_rate": 5.353685100379189e-07, "loss": 0.0265, "step": 800 }, { "epoch": 1.330564784053156, "grad_norm": 6.710888862609863, "learning_rate": 5.32986397625808e-07, "loss": 0.0251, "step": 801 }, { "epoch": 1.3322259136212624, "grad_norm": 8.738547325134277, "learning_rate": 5.306076696574522e-07, "loss": 0.012, "step": 802 }, { "epoch": 1.3338870431893688, "grad_norm": 16.569536209106445, "learning_rate": 5.282323433714743e-07, "loss": 0.0281, "step": 803 }, { "epoch": 1.3355481727574752, "grad_norm": 15.599034309387207, "learning_rate": 5.258604359818443e-07, "loss": 0.0395, "step": 804 }, { "epoch": 1.3372093023255813, "grad_norm": 11.257906913757324, "learning_rate": 5.234919646777555e-07, "loss": 0.0456, "step": 805 }, { "epoch": 1.3388704318936877, "grad_norm": 9.306239128112793, "learning_rate": 5.211269466235e-07, "loss": 0.0675, "step": 806 }, { "epoch": 1.340531561461794, "grad_norm": 1.2154159545898438, "learning_rate": 5.187653989583445e-07, "loss": 0.0017, "step": 807 }, { "epoch": 1.3421926910299002, "grad_norm": 0.9059697389602661, "learning_rate": 5.164073387964057e-07, "loss": 0.0016, "step": 808 }, { "epoch": 1.3438538205980066, "grad_norm": 0.11658035218715668, "learning_rate": 5.140527832265255e-07, "loss": 0.0002, "step": 809 }, { "epoch": 1.345514950166113, "grad_norm": 22.50667953491211, "learning_rate": 5.117017493121501e-07, "loss": 0.06, "step": 810 }, { "epoch": 1.3471760797342194, "grad_norm": 20.302230834960938, "learning_rate": 5.093542540912024e-07, "loss": 0.0738, "step": 811 }, { "epoch": 1.3488372093023255, "grad_norm": 9.267167091369629, "learning_rate": 5.070103145759605e-07, "loss": 0.011, "step": 812 }, { "epoch": 1.350498338870432, "grad_norm": 13.021448135375977, "learning_rate": 5.046699477529359e-07, "loss": 0.0318, "step": 813 }, { "epoch": 1.352159468438538, "grad_norm": 3.4434173107147217, "learning_rate": 5.023331705827476e-07, "loss": 0.0078, "step": 814 }, { "epoch": 1.3538205980066444, "grad_norm": 3.8246960639953613, "learning_rate": 5.000000000000002e-07, "loss": 0.0071, "step": 815 }, { "epoch": 1.3554817275747508, "grad_norm": 22.36045265197754, "learning_rate": 4.976704529131616e-07, "loss": 0.0593, "step": 816 }, { "epoch": 1.3571428571428572, "grad_norm": 7.905923366546631, "learning_rate": 4.953445462044414e-07, "loss": 0.02, "step": 817 }, { "epoch": 1.3588039867109636, "grad_norm": 13.977715492248535, "learning_rate": 4.930222967296661e-07, "loss": 0.059, "step": 818 }, { "epoch": 1.3604651162790697, "grad_norm": 0.6750720739364624, "learning_rate": 4.90703721318158e-07, "loss": 0.0012, "step": 819 }, { "epoch": 1.3621262458471761, "grad_norm": 2.0304207801818848, "learning_rate": 4.883888367726152e-07, "loss": 0.0036, "step": 820 }, { "epoch": 1.3637873754152823, "grad_norm": 9.32971477508545, "learning_rate": 4.860776598689865e-07, "loss": 0.0328, "step": 821 }, { "epoch": 1.3654485049833887, "grad_norm": 11.44428825378418, "learning_rate": 4.837702073563521e-07, "loss": 0.0399, "step": 822 }, { "epoch": 1.367109634551495, "grad_norm": 15.22606086730957, "learning_rate": 4.81466495956801e-07, "loss": 0.0632, "step": 823 }, { "epoch": 1.3687707641196014, "grad_norm": 5.094400882720947, "learning_rate": 4.791665423653117e-07, "loss": 0.0126, "step": 824 }, { "epoch": 1.3704318936877076, "grad_norm": 11.821290016174316, "learning_rate": 4.768703632496287e-07, "loss": 0.0196, "step": 825 }, { "epoch": 1.372093023255814, "grad_norm": 4.006501197814941, "learning_rate": 4.745779752501429e-07, "loss": 0.0044, "step": 826 }, { "epoch": 1.3737541528239203, "grad_norm": 10.801963806152344, "learning_rate": 4.722893949797724e-07, "loss": 0.023, "step": 827 }, { "epoch": 1.3754152823920265, "grad_norm": 3.6390554904937744, "learning_rate": 4.7000463902383825e-07, "loss": 0.0036, "step": 828 }, { "epoch": 1.3770764119601329, "grad_norm": 1.589859962463379, "learning_rate": 4.677237239399489e-07, "loss": 0.0026, "step": 829 }, { "epoch": 1.3787375415282392, "grad_norm": 8.354605674743652, "learning_rate": 4.654466662578781e-07, "loss": 0.0206, "step": 830 }, { "epoch": 1.3803986710963456, "grad_norm": 6.640425205230713, "learning_rate": 4.631734824794428e-07, "loss": 0.0141, "step": 831 }, { "epoch": 1.3820598006644518, "grad_norm": 13.431829452514648, "learning_rate": 4.6090418907838814e-07, "loss": 0.0374, "step": 832 }, { "epoch": 1.3837209302325582, "grad_norm": 12.409584999084473, "learning_rate": 4.586388025002647e-07, "loss": 0.0198, "step": 833 }, { "epoch": 1.3853820598006645, "grad_norm": 1.4373682737350464, "learning_rate": 4.5637733916231004e-07, "loss": 0.0023, "step": 834 }, { "epoch": 1.3870431893687707, "grad_norm": 12.105127334594727, "learning_rate": 4.541198154533311e-07, "loss": 0.0145, "step": 835 }, { "epoch": 1.388704318936877, "grad_norm": 1.378914475440979, "learning_rate": 4.518662477335834e-07, "loss": 0.0017, "step": 836 }, { "epoch": 1.3903654485049834, "grad_norm": 2.0379269123077393, "learning_rate": 4.4961665233465393e-07, "loss": 0.0026, "step": 837 }, { "epoch": 1.3920265780730898, "grad_norm": 4.553691864013672, "learning_rate": 4.473710455593416e-07, "loss": 0.011, "step": 838 }, { "epoch": 1.393687707641196, "grad_norm": 9.32184886932373, "learning_rate": 4.4512944368154114e-07, "loss": 0.0311, "step": 839 }, { "epoch": 1.3953488372093024, "grad_norm": 3.257469892501831, "learning_rate": 4.4289186294612256e-07, "loss": 0.0049, "step": 840 }, { "epoch": 1.3970099667774085, "grad_norm": 4.196468830108643, "learning_rate": 4.4065831956881494e-07, "loss": 0.0092, "step": 841 }, { "epoch": 1.398671096345515, "grad_norm": 2.6750576496124268, "learning_rate": 4.384288297360894e-07, "loss": 0.0065, "step": 842 }, { "epoch": 1.4003322259136213, "grad_norm": 0.7593079805374146, "learning_rate": 4.362034096050403e-07, "loss": 0.0007, "step": 843 }, { "epoch": 1.4019933554817277, "grad_norm": 0.7584846019744873, "learning_rate": 4.339820753032691e-07, "loss": 0.0016, "step": 844 }, { "epoch": 1.4036544850498338, "grad_norm": 24.843107223510742, "learning_rate": 4.31764842928767e-07, "loss": 0.0216, "step": 845 }, { "epoch": 1.4053156146179402, "grad_norm": 7.665549278259277, "learning_rate": 4.295517285497995e-07, "loss": 0.0099, "step": 846 }, { "epoch": 1.4069767441860466, "grad_norm": 16.56751251220703, "learning_rate": 4.27342748204788e-07, "loss": 0.0097, "step": 847 }, { "epoch": 1.4086378737541527, "grad_norm": 8.430388450622559, "learning_rate": 4.2513791790219467e-07, "loss": 0.0162, "step": 848 }, { "epoch": 1.410299003322259, "grad_norm": 6.689608097076416, "learning_rate": 4.229372536204074e-07, "loss": 0.018, "step": 849 }, { "epoch": 1.4119601328903655, "grad_norm": 13.74167537689209, "learning_rate": 4.207407713076221e-07, "loss": 0.0113, "step": 850 }, { "epoch": 1.4136212624584719, "grad_norm": 13.198443412780762, "learning_rate": 4.185484868817274e-07, "loss": 0.0628, "step": 851 }, { "epoch": 1.415282392026578, "grad_norm": 17.86145782470703, "learning_rate": 4.1636041623019135e-07, "loss": 0.0521, "step": 852 }, { "epoch": 1.4169435215946844, "grad_norm": 25.738597869873047, "learning_rate": 4.14176575209944e-07, "loss": 0.1025, "step": 853 }, { "epoch": 1.4186046511627908, "grad_norm": 0.3076265752315521, "learning_rate": 4.119969796472631e-07, "loss": 0.0003, "step": 854 }, { "epoch": 1.420265780730897, "grad_norm": 17.335969924926758, "learning_rate": 4.098216453376596e-07, "loss": 0.0617, "step": 855 }, { "epoch": 1.4219269102990033, "grad_norm": 6.2664899826049805, "learning_rate": 4.076505880457641e-07, "loss": 0.0253, "step": 856 }, { "epoch": 1.4235880398671097, "grad_norm": 5.2418060302734375, "learning_rate": 4.0548382350521084e-07, "loss": 0.0074, "step": 857 }, { "epoch": 1.425249169435216, "grad_norm": 3.4297940731048584, "learning_rate": 4.033213674185241e-07, "loss": 0.0053, "step": 858 }, { "epoch": 1.4269102990033222, "grad_norm": 9.643595695495605, "learning_rate": 4.011632354570067e-07, "loss": 0.076, "step": 859 }, { "epoch": 1.4285714285714286, "grad_norm": 4.146191596984863, "learning_rate": 3.990094432606218e-07, "loss": 0.0027, "step": 860 }, { "epoch": 1.4302325581395348, "grad_norm": 0.19965682923793793, "learning_rate": 3.96860006437885e-07, "loss": 0.0003, "step": 861 }, { "epoch": 1.4318936877076411, "grad_norm": 19.833518981933594, "learning_rate": 3.9471494056574685e-07, "loss": 0.0622, "step": 862 }, { "epoch": 1.4335548172757475, "grad_norm": 0.17571118474006653, "learning_rate": 3.92574261189482e-07, "loss": 0.0003, "step": 863 }, { "epoch": 1.435215946843854, "grad_norm": 7.566018581390381, "learning_rate": 3.9043798382257697e-07, "loss": 0.0078, "step": 864 }, { "epoch": 1.43687707641196, "grad_norm": 11.348194122314453, "learning_rate": 3.88306123946616e-07, "loss": 0.0235, "step": 865 }, { "epoch": 1.4385382059800664, "grad_norm": 10.6011381149292, "learning_rate": 3.861786970111702e-07, "loss": 0.0313, "step": 866 }, { "epoch": 1.4401993355481728, "grad_norm": 0.5557393431663513, "learning_rate": 3.840557184336849e-07, "loss": 0.0006, "step": 867 }, { "epoch": 1.441860465116279, "grad_norm": 25.956459045410156, "learning_rate": 3.81937203599369e-07, "loss": 0.0449, "step": 868 }, { "epoch": 1.4435215946843853, "grad_norm": 11.637957572937012, "learning_rate": 3.798231678610819e-07, "loss": 0.0237, "step": 869 }, { "epoch": 1.4451827242524917, "grad_norm": 5.680612087249756, "learning_rate": 3.777136265392231e-07, "loss": 0.0055, "step": 870 }, { "epoch": 1.446843853820598, "grad_norm": 4.692670822143555, "learning_rate": 3.7560859492162176e-07, "loss": 0.0038, "step": 871 }, { "epoch": 1.4485049833887043, "grad_norm": 7.798187732696533, "learning_rate": 3.735080882634246e-07, "loss": 0.0065, "step": 872 }, { "epoch": 1.4501661129568106, "grad_norm": 14.402698516845703, "learning_rate": 3.7141212178698576e-07, "loss": 0.0611, "step": 873 }, { "epoch": 1.451827242524917, "grad_norm": 14.016066551208496, "learning_rate": 3.69320710681758e-07, "loss": 0.0834, "step": 874 }, { "epoch": 1.4534883720930232, "grad_norm": 7.999434947967529, "learning_rate": 3.672338701041798e-07, "loss": 0.0183, "step": 875 }, { "epoch": 1.4551495016611296, "grad_norm": 10.387938499450684, "learning_rate": 3.651516151775681e-07, "loss": 0.0428, "step": 876 }, { "epoch": 1.456810631229236, "grad_norm": 15.999258041381836, "learning_rate": 3.630739609920068e-07, "loss": 0.0825, "step": 877 }, { "epoch": 1.4584717607973423, "grad_norm": 7.274929523468018, "learning_rate": 3.6100092260423955e-07, "loss": 0.0211, "step": 878 }, { "epoch": 1.4601328903654485, "grad_norm": 13.323637962341309, "learning_rate": 3.5893251503755815e-07, "loss": 0.0177, "step": 879 }, { "epoch": 1.4617940199335548, "grad_norm": 4.893885135650635, "learning_rate": 3.5686875328169507e-07, "loss": 0.0122, "step": 880 }, { "epoch": 1.463455149501661, "grad_norm": 5.68451452255249, "learning_rate": 3.548096522927154e-07, "loss": 0.0199, "step": 881 }, { "epoch": 1.4651162790697674, "grad_norm": 5.512393951416016, "learning_rate": 3.5275522699290704e-07, "loss": 0.0164, "step": 882 }, { "epoch": 1.4667774086378738, "grad_norm": 18.978796005249023, "learning_rate": 3.5070549227067324e-07, "loss": 0.0334, "step": 883 }, { "epoch": 1.4684385382059801, "grad_norm": 0.4384339153766632, "learning_rate": 3.4866046298042426e-07, "loss": 0.0008, "step": 884 }, { "epoch": 1.4700996677740865, "grad_norm": 3.6679792404174805, "learning_rate": 3.466201539424716e-07, "loss": 0.0035, "step": 885 }, { "epoch": 1.4717607973421927, "grad_norm": 8.465070724487305, "learning_rate": 3.4458457994291757e-07, "loss": 0.0178, "step": 886 }, { "epoch": 1.473421926910299, "grad_norm": 9.443462371826172, "learning_rate": 3.4255375573355014e-07, "loss": 0.0077, "step": 887 }, { "epoch": 1.4750830564784052, "grad_norm": 8.275284767150879, "learning_rate": 3.405276960317366e-07, "loss": 0.0075, "step": 888 }, { "epoch": 1.4767441860465116, "grad_norm": 4.26287317276001, "learning_rate": 3.385064155203138e-07, "loss": 0.0107, "step": 889 }, { "epoch": 1.478405315614618, "grad_norm": 17.598369598388672, "learning_rate": 3.364899288474857e-07, "loss": 0.0305, "step": 890 }, { "epoch": 1.4800664451827243, "grad_norm": 25.065013885498047, "learning_rate": 3.344782506267154e-07, "loss": 0.0558, "step": 891 }, { "epoch": 1.4817275747508305, "grad_norm": 8.171377182006836, "learning_rate": 3.3247139543661705e-07, "loss": 0.0585, "step": 892 }, { "epoch": 1.4833887043189369, "grad_norm": 13.189953804016113, "learning_rate": 3.3046937782085506e-07, "loss": 0.036, "step": 893 }, { "epoch": 1.4850498338870433, "grad_norm": 6.294521331787109, "learning_rate": 3.2847221228803414e-07, "loss": 0.0275, "step": 894 }, { "epoch": 1.4867109634551494, "grad_norm": 2.937936305999756, "learning_rate": 3.264799133115963e-07, "loss": 0.003, "step": 895 }, { "epoch": 1.4883720930232558, "grad_norm": 8.645936012268066, "learning_rate": 3.244924953297167e-07, "loss": 0.0213, "step": 896 }, { "epoch": 1.4900332225913622, "grad_norm": 19.482379913330078, "learning_rate": 3.2250997274519666e-07, "loss": 0.0985, "step": 897 }, { "epoch": 1.4916943521594686, "grad_norm": 2.1434147357940674, "learning_rate": 3.205323599253613e-07, "loss": 0.0075, "step": 898 }, { "epoch": 1.4933554817275747, "grad_norm": 1.688423991203308, "learning_rate": 3.1855967120195413e-07, "loss": 0.0037, "step": 899 }, { "epoch": 1.495016611295681, "grad_norm": 4.679425239562988, "learning_rate": 3.165919208710348e-07, "loss": 0.0176, "step": 900 }, { "epoch": 1.4966777408637872, "grad_norm": 25.044464111328125, "learning_rate": 3.146291231928737e-07, "loss": 0.0764, "step": 901 }, { "epoch": 1.4983388704318936, "grad_norm": 4.9685893058776855, "learning_rate": 3.1267129239184907e-07, "loss": 0.0123, "step": 902 }, { "epoch": 1.5, "grad_norm": 6.012315273284912, "learning_rate": 3.107184426563455e-07, "loss": 0.0124, "step": 903 }, { "epoch": 1.5016611295681064, "grad_norm": 13.312503814697266, "learning_rate": 3.0877058813864854e-07, "loss": 0.0925, "step": 904 }, { "epoch": 1.5033222591362128, "grad_norm": 14.907830238342285, "learning_rate": 3.0682774295484406e-07, "loss": 0.0279, "step": 905 }, { "epoch": 1.504983388704319, "grad_norm": 3.517716646194458, "learning_rate": 3.048899211847148e-07, "loss": 0.0094, "step": 906 }, { "epoch": 1.5066445182724253, "grad_norm": 5.761756896972656, "learning_rate": 3.0295713687164004e-07, "loss": 0.0078, "step": 907 }, { "epoch": 1.5083056478405314, "grad_norm": 8.124967575073242, "learning_rate": 3.010294040224914e-07, "loss": 0.0134, "step": 908 }, { "epoch": 1.5099667774086378, "grad_norm": 14.692148208618164, "learning_rate": 2.9910673660753295e-07, "loss": 0.0122, "step": 909 }, { "epoch": 1.5116279069767442, "grad_norm": 3.8068079948425293, "learning_rate": 2.971891485603203e-07, "loss": 0.0062, "step": 910 }, { "epoch": 1.5132890365448506, "grad_norm": 7.373571395874023, "learning_rate": 2.952766537775978e-07, "loss": 0.0153, "step": 911 }, { "epoch": 1.514950166112957, "grad_norm": 6.404834270477295, "learning_rate": 2.9336926611919953e-07, "loss": 0.0157, "step": 912 }, { "epoch": 1.5166112956810631, "grad_norm": 19.058998107910156, "learning_rate": 2.914669994079486e-07, "loss": 0.0377, "step": 913 }, { "epoch": 1.5182724252491693, "grad_norm": 1.187256097793579, "learning_rate": 2.8956986742955604e-07, "loss": 0.0013, "step": 914 }, { "epoch": 1.5199335548172757, "grad_norm": 19.175989151000977, "learning_rate": 2.876778839325219e-07, "loss": 0.0542, "step": 915 }, { "epoch": 1.521594684385382, "grad_norm": 12.16926097869873, "learning_rate": 2.8579106262803465e-07, "loss": 0.0622, "step": 916 }, { "epoch": 1.5232558139534884, "grad_norm": 10.201384544372559, "learning_rate": 2.839094171898736e-07, "loss": 0.0154, "step": 917 }, { "epoch": 1.5249169435215948, "grad_norm": 17.753768920898438, "learning_rate": 2.8203296125430744e-07, "loss": 0.0435, "step": 918 }, { "epoch": 1.526578073089701, "grad_norm": 9.888166427612305, "learning_rate": 2.801617084199966e-07, "loss": 0.0331, "step": 919 }, { "epoch": 1.5282392026578073, "grad_norm": 1.097863793373108, "learning_rate": 2.782956722478962e-07, "loss": 0.001, "step": 920 }, { "epoch": 1.5299003322259135, "grad_norm": 1.8991848230361938, "learning_rate": 2.764348662611538e-07, "loss": 0.003, "step": 921 }, { "epoch": 1.5315614617940199, "grad_norm": 20.583703994750977, "learning_rate": 2.7457930394501563e-07, "loss": 0.0439, "step": 922 }, { "epoch": 1.5332225913621262, "grad_norm": 6.2648539543151855, "learning_rate": 2.727289987467274e-07, "loss": 0.0226, "step": 923 }, { "epoch": 1.5348837209302326, "grad_norm": 1.496519923210144, "learning_rate": 2.7088396407543435e-07, "loss": 0.0018, "step": 924 }, { "epoch": 1.536544850498339, "grad_norm": 6.547482013702393, "learning_rate": 2.690442133020888e-07, "loss": 0.0212, "step": 925 }, { "epoch": 1.5382059800664452, "grad_norm": 17.5012264251709, "learning_rate": 2.67209759759349e-07, "loss": 0.0462, "step": 926 }, { "epoch": 1.5398671096345515, "grad_norm": 0.6141988635063171, "learning_rate": 2.6538061674148504e-07, "loss": 0.0013, "step": 927 }, { "epoch": 1.5415282392026577, "grad_norm": 1.5886684656143188, "learning_rate": 2.6355679750428084e-07, "loss": 0.0029, "step": 928 }, { "epoch": 1.543189368770764, "grad_norm": 27.549413681030273, "learning_rate": 2.617383152649403e-07, "loss": 0.0891, "step": 929 }, { "epoch": 1.5448504983388704, "grad_norm": 10.47612476348877, "learning_rate": 2.5992518320198874e-07, "loss": 0.0201, "step": 930 }, { "epoch": 1.5465116279069768, "grad_norm": 10.489360809326172, "learning_rate": 2.5811741445517944e-07, "loss": 0.0269, "step": 931 }, { "epoch": 1.5481727574750832, "grad_norm": 3.5623435974121094, "learning_rate": 2.5631502212539815e-07, "loss": 0.0069, "step": 932 }, { "epoch": 1.5498338870431894, "grad_norm": 6.290643215179443, "learning_rate": 2.5451801927456694e-07, "loss": 0.0272, "step": 933 }, { "epoch": 1.5514950166112955, "grad_norm": 11.861459732055664, "learning_rate": 2.5272641892555067e-07, "loss": 0.0365, "step": 934 }, { "epoch": 1.553156146179402, "grad_norm": 1.0134330987930298, "learning_rate": 2.509402340620629e-07, "loss": 0.0022, "step": 935 }, { "epoch": 1.5548172757475083, "grad_norm": 2.671765089035034, "learning_rate": 2.491594776285704e-07, "loss": 0.0024, "step": 936 }, { "epoch": 1.5564784053156147, "grad_norm": 4.3023223876953125, "learning_rate": 2.473841625302006e-07, "loss": 0.0068, "step": 937 }, { "epoch": 1.558139534883721, "grad_norm": 3.784409999847412, "learning_rate": 2.45614301632647e-07, "loss": 0.0088, "step": 938 }, { "epoch": 1.5598006644518272, "grad_norm": 11.982508659362793, "learning_rate": 2.438499077620777e-07, "loss": 0.0185, "step": 939 }, { "epoch": 1.5614617940199336, "grad_norm": 6.5332818031311035, "learning_rate": 2.420909937050405e-07, "loss": 0.0146, "step": 940 }, { "epoch": 1.5631229235880397, "grad_norm": 14.1775484085083, "learning_rate": 2.403375722083707e-07, "loss": 0.0693, "step": 941 }, { "epoch": 1.564784053156146, "grad_norm": 8.041645050048828, "learning_rate": 2.385896559791002e-07, "loss": 0.0071, "step": 942 }, { "epoch": 1.5664451827242525, "grad_norm": 1.4674506187438965, "learning_rate": 2.3684725768436332e-07, "loss": 0.0026, "step": 943 }, { "epoch": 1.5681063122923589, "grad_norm": 10.562248229980469, "learning_rate": 2.3511038995130628e-07, "loss": 0.0332, "step": 944 }, { "epoch": 1.5697674418604652, "grad_norm": 7.5686187744140625, "learning_rate": 2.3337906536699504e-07, "loss": 0.0138, "step": 945 }, { "epoch": 1.5714285714285714, "grad_norm": 6.138749122619629, "learning_rate": 2.316532964783252e-07, "loss": 0.0173, "step": 946 }, { "epoch": 1.5730897009966778, "grad_norm": 9.5950927734375, "learning_rate": 2.2993309579192964e-07, "loss": 0.0328, "step": 947 }, { "epoch": 1.574750830564784, "grad_norm": 8.108460426330566, "learning_rate": 2.2821847577408815e-07, "loss": 0.041, "step": 948 }, { "epoch": 1.5764119601328903, "grad_norm": 43.922672271728516, "learning_rate": 2.2650944885063894e-07, "loss": 0.0735, "step": 949 }, { "epoch": 1.5780730897009967, "grad_norm": 2.7781217098236084, "learning_rate": 2.2480602740688514e-07, "loss": 0.0084, "step": 950 } ], "logging_steps": 1, "max_steps": 1204, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 50, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.0459192604691005e+19, "train_batch_size": 2, "trial_name": null, "trial_params": null }