{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.9766839378238341, "eval_steps": 97, "global_step": 772, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0025906735751295338, "grad_norm": 0.37030652165412903, "learning_rate": 1.0000000000000002e-06, "loss": 1.4535, "step": 1 }, { "epoch": 0.0025906735751295338, "eval_loss": 1.3420950174331665, "eval_runtime": 98.9896, "eval_samples_per_second": 2.526, "eval_steps_per_second": 0.323, "step": 1 }, { "epoch": 0.0051813471502590676, "grad_norm": 0.37870243191719055, "learning_rate": 2.0000000000000003e-06, "loss": 1.3744, "step": 2 }, { "epoch": 0.007772020725388601, "grad_norm": 0.3809736669063568, "learning_rate": 3e-06, "loss": 1.413, "step": 3 }, { "epoch": 0.010362694300518135, "grad_norm": 0.3608606457710266, "learning_rate": 4.000000000000001e-06, "loss": 1.4169, "step": 4 }, { "epoch": 0.012953367875647668, "grad_norm": 0.3465673625469208, "learning_rate": 5e-06, "loss": 1.3166, "step": 5 }, { "epoch": 0.015544041450777202, "grad_norm": 0.36601942777633667, "learning_rate": 6e-06, "loss": 1.4029, "step": 6 }, { "epoch": 0.018134715025906734, "grad_norm": 0.374077171087265, "learning_rate": 7e-06, "loss": 1.4351, "step": 7 }, { "epoch": 0.02072538860103627, "grad_norm": 0.3928413391113281, "learning_rate": 8.000000000000001e-06, "loss": 1.3023, "step": 8 }, { "epoch": 0.023316062176165803, "grad_norm": 0.35695162415504456, "learning_rate": 9e-06, "loss": 1.4251, "step": 9 }, { "epoch": 0.025906735751295335, "grad_norm": 0.38057687878608704, "learning_rate": 1e-05, "loss": 1.4864, "step": 10 }, { "epoch": 0.02849740932642487, "grad_norm": 0.3801506757736206, "learning_rate": 9.999957505845144e-06, "loss": 1.4257, "step": 11 }, { "epoch": 0.031088082901554404, "grad_norm": 0.3910161852836609, "learning_rate": 9.999830024102874e-06, "loss": 1.456, "step": 12 }, { "epoch": 0.03367875647668394, "grad_norm": 0.37710297107696533, "learning_rate": 9.999617556940085e-06, "loss": 1.3441, "step": 13 }, { "epoch": 0.03626943005181347, "grad_norm": 0.4258323311805725, "learning_rate": 9.99932010796822e-06, "loss": 1.3589, "step": 14 }, { "epoch": 0.038860103626943004, "grad_norm": 0.42138487100601196, "learning_rate": 9.998937682243216e-06, "loss": 1.4402, "step": 15 }, { "epoch": 0.04145077720207254, "grad_norm": 0.45730850100517273, "learning_rate": 9.998470286265415e-06, "loss": 1.3916, "step": 16 }, { "epoch": 0.04404145077720207, "grad_norm": 0.4608759582042694, "learning_rate": 9.99791792797946e-06, "loss": 1.4034, "step": 17 }, { "epoch": 0.046632124352331605, "grad_norm": 0.45245274901390076, "learning_rate": 9.997280616774147e-06, "loss": 1.3311, "step": 18 }, { "epoch": 0.04922279792746114, "grad_norm": 0.462851881980896, "learning_rate": 9.996558363482277e-06, "loss": 1.4076, "step": 19 }, { "epoch": 0.05181347150259067, "grad_norm": 0.5080270767211914, "learning_rate": 9.995751180380468e-06, "loss": 1.4116, "step": 20 }, { "epoch": 0.054404145077720206, "grad_norm": 0.44085511565208435, "learning_rate": 9.994859081188944e-06, "loss": 1.399, "step": 21 }, { "epoch": 0.05699481865284974, "grad_norm": 0.40889567136764526, "learning_rate": 9.993882081071307e-06, "loss": 1.3666, "step": 22 }, { "epoch": 0.05958549222797927, "grad_norm": 0.42577946186065674, "learning_rate": 9.992820196634274e-06, "loss": 1.3906, "step": 23 }, { "epoch": 0.06217616580310881, "grad_norm": 0.46674349904060364, "learning_rate": 9.991673445927399e-06, "loss": 1.3446, "step": 24 }, { "epoch": 0.06476683937823834, "grad_norm": 0.48815953731536865, "learning_rate": 9.99044184844276e-06, "loss": 1.3156, "step": 25 }, { "epoch": 0.06735751295336788, "grad_norm": 0.4857616424560547, "learning_rate": 9.989125425114639e-06, "loss": 1.3713, "step": 26 }, { "epoch": 0.06994818652849741, "grad_norm": 0.4946143925189972, "learning_rate": 9.98772419831915e-06, "loss": 1.4438, "step": 27 }, { "epoch": 0.07253886010362694, "grad_norm": 0.47251901030540466, "learning_rate": 9.986238191873874e-06, "loss": 1.3626, "step": 28 }, { "epoch": 0.07512953367875648, "grad_norm": 0.5448423027992249, "learning_rate": 9.984667431037448e-06, "loss": 1.3201, "step": 29 }, { "epoch": 0.07772020725388601, "grad_norm": 0.4412786364555359, "learning_rate": 9.983011942509131e-06, "loss": 1.3581, "step": 30 }, { "epoch": 0.08031088082901554, "grad_norm": 0.44406625628471375, "learning_rate": 9.981271754428361e-06, "loss": 1.2458, "step": 31 }, { "epoch": 0.08290155440414508, "grad_norm": 0.5035629272460938, "learning_rate": 9.979446896374264e-06, "loss": 1.3699, "step": 32 }, { "epoch": 0.08549222797927461, "grad_norm": 0.47425153851509094, "learning_rate": 9.977537399365159e-06, "loss": 1.3558, "step": 33 }, { "epoch": 0.08808290155440414, "grad_norm": 0.5448233485221863, "learning_rate": 9.975543295858035e-06, "loss": 1.3342, "step": 34 }, { "epoch": 0.09067357512953368, "grad_norm": 0.49746477603912354, "learning_rate": 9.973464619747983e-06, "loss": 1.3138, "step": 35 }, { "epoch": 0.09326424870466321, "grad_norm": 0.48222672939300537, "learning_rate": 9.971301406367644e-06, "loss": 1.3991, "step": 36 }, { "epoch": 0.09585492227979274, "grad_norm": 0.5209602117538452, "learning_rate": 9.969053692486582e-06, "loss": 1.2775, "step": 37 }, { "epoch": 0.09844559585492228, "grad_norm": 0.4517224133014679, "learning_rate": 9.966721516310683e-06, "loss": 1.3207, "step": 38 }, { "epoch": 0.10103626943005181, "grad_norm": 0.5346600413322449, "learning_rate": 9.964304917481482e-06, "loss": 1.2108, "step": 39 }, { "epoch": 0.10362694300518134, "grad_norm": 0.5563952326774597, "learning_rate": 9.961803937075516e-06, "loss": 1.2934, "step": 40 }, { "epoch": 0.10621761658031088, "grad_norm": 0.5306499004364014, "learning_rate": 9.959218617603601e-06, "loss": 1.2202, "step": 41 }, { "epoch": 0.10880829015544041, "grad_norm": 0.4907849133014679, "learning_rate": 9.956549003010122e-06, "loss": 1.2261, "step": 42 }, { "epoch": 0.11139896373056994, "grad_norm": 0.5063201189041138, "learning_rate": 9.953795138672291e-06, "loss": 1.2863, "step": 43 }, { "epoch": 0.11398963730569948, "grad_norm": 0.5338860750198364, "learning_rate": 9.950957071399357e-06, "loss": 1.2388, "step": 44 }, { "epoch": 0.11658031088082901, "grad_norm": 0.5024104714393616, "learning_rate": 9.948034849431831e-06, "loss": 1.2573, "step": 45 }, { "epoch": 0.11917098445595854, "grad_norm": 0.5430374145507812, "learning_rate": 9.945028522440654e-06, "loss": 1.1556, "step": 46 }, { "epoch": 0.12176165803108809, "grad_norm": 0.4689851701259613, "learning_rate": 9.941938141526355e-06, "loss": 1.2576, "step": 47 }, { "epoch": 0.12435233160621761, "grad_norm": 0.4983338415622711, "learning_rate": 9.938763759218186e-06, "loss": 1.2247, "step": 48 }, { "epoch": 0.12694300518134716, "grad_norm": 0.47061145305633545, "learning_rate": 9.935505429473221e-06, "loss": 1.2173, "step": 49 }, { "epoch": 0.12953367875647667, "grad_norm": 0.5208876729011536, "learning_rate": 9.93216320767545e-06, "loss": 1.1284, "step": 50 }, { "epoch": 0.13212435233160622, "grad_norm": 0.5343345999717712, "learning_rate": 9.92873715063483e-06, "loss": 1.2766, "step": 51 }, { "epoch": 0.13471502590673576, "grad_norm": 0.5376512408256531, "learning_rate": 9.925227316586316e-06, "loss": 1.3082, "step": 52 }, { "epoch": 0.13730569948186527, "grad_norm": 0.49544206261634827, "learning_rate": 9.921633765188887e-06, "loss": 1.2488, "step": 53 }, { "epoch": 0.13989637305699482, "grad_norm": 0.5503423810005188, "learning_rate": 9.917956557524511e-06, "loss": 1.2837, "step": 54 }, { "epoch": 0.14248704663212436, "grad_norm": 0.6678252816200256, "learning_rate": 9.91419575609712e-06, "loss": 1.2592, "step": 55 }, { "epoch": 0.14507772020725387, "grad_norm": 0.5491769313812256, "learning_rate": 9.910351424831545e-06, "loss": 1.2528, "step": 56 }, { "epoch": 0.14766839378238342, "grad_norm": 0.5200887322425842, "learning_rate": 9.906423629072435e-06, "loss": 1.1433, "step": 57 }, { "epoch": 0.15025906735751296, "grad_norm": 0.5037204623222351, "learning_rate": 9.902412435583127e-06, "loss": 1.204, "step": 58 }, { "epoch": 0.15284974093264247, "grad_norm": 0.62909996509552, "learning_rate": 9.898317912544537e-06, "loss": 1.1012, "step": 59 }, { "epoch": 0.15544041450777202, "grad_norm": 0.5567560791969299, "learning_rate": 9.89414012955398e-06, "loss": 1.2056, "step": 60 }, { "epoch": 0.15803108808290156, "grad_norm": 0.4932163655757904, "learning_rate": 9.889879157624003e-06, "loss": 1.1295, "step": 61 }, { "epoch": 0.16062176165803108, "grad_norm": 0.49116799235343933, "learning_rate": 9.885535069181163e-06, "loss": 1.1767, "step": 62 }, { "epoch": 0.16321243523316062, "grad_norm": 0.5316625833511353, "learning_rate": 9.881107938064806e-06, "loss": 1.2038, "step": 63 }, { "epoch": 0.16580310880829016, "grad_norm": 0.5318853259086609, "learning_rate": 9.876597839525814e-06, "loss": 1.1865, "step": 64 }, { "epoch": 0.16839378238341968, "grad_norm": 0.535534143447876, "learning_rate": 9.872004850225313e-06, "loss": 1.2335, "step": 65 }, { "epoch": 0.17098445595854922, "grad_norm": 0.6383189558982849, "learning_rate": 9.867329048233387e-06, "loss": 1.1985, "step": 66 }, { "epoch": 0.17357512953367876, "grad_norm": 0.5795570611953735, "learning_rate": 9.862570513027736e-06, "loss": 1.1598, "step": 67 }, { "epoch": 0.17616580310880828, "grad_norm": 0.5034987330436707, "learning_rate": 9.857729325492329e-06, "loss": 1.1798, "step": 68 }, { "epoch": 0.17875647668393782, "grad_norm": 0.5522433519363403, "learning_rate": 9.85280556791604e-06, "loss": 1.1093, "step": 69 }, { "epoch": 0.18134715025906736, "grad_norm": 0.5660341382026672, "learning_rate": 9.847799323991234e-06, "loss": 1.1477, "step": 70 }, { "epoch": 0.18393782383419688, "grad_norm": 0.529484748840332, "learning_rate": 9.842710678812352e-06, "loss": 1.1028, "step": 71 }, { "epoch": 0.18652849740932642, "grad_norm": 0.4907700717449188, "learning_rate": 9.837539718874466e-06, "loss": 1.2247, "step": 72 }, { "epoch": 0.18911917098445596, "grad_norm": 0.5624697804450989, "learning_rate": 9.832286532071802e-06, "loss": 1.1997, "step": 73 }, { "epoch": 0.19170984455958548, "grad_norm": 0.5344734787940979, "learning_rate": 9.826951207696258e-06, "loss": 1.1389, "step": 74 }, { "epoch": 0.19430051813471502, "grad_norm": 0.5770263671875, "learning_rate": 9.82153383643587e-06, "loss": 1.2063, "step": 75 }, { "epoch": 0.19689119170984457, "grad_norm": 0.5622824430465698, "learning_rate": 9.816034510373287e-06, "loss": 1.1715, "step": 76 }, { "epoch": 0.19948186528497408, "grad_norm": 0.5856156945228577, "learning_rate": 9.81045332298419e-06, "loss": 1.155, "step": 77 }, { "epoch": 0.20207253886010362, "grad_norm": 0.5373365879058838, "learning_rate": 9.804790369135719e-06, "loss": 1.1258, "step": 78 }, { "epoch": 0.20466321243523317, "grad_norm": 0.5527685284614563, "learning_rate": 9.799045745084848e-06, "loss": 1.1948, "step": 79 }, { "epoch": 0.20725388601036268, "grad_norm": 0.5555474162101746, "learning_rate": 9.793219548476754e-06, "loss": 1.1871, "step": 80 }, { "epoch": 0.20984455958549222, "grad_norm": 0.5365428328514099, "learning_rate": 9.787311878343158e-06, "loss": 1.1591, "step": 81 }, { "epoch": 0.21243523316062177, "grad_norm": 0.5377835631370544, "learning_rate": 9.781322835100639e-06, "loss": 1.1238, "step": 82 }, { "epoch": 0.21502590673575128, "grad_norm": 0.5349002480506897, "learning_rate": 9.77525252054893e-06, "loss": 1.181, "step": 83 }, { "epoch": 0.21761658031088082, "grad_norm": 0.6802361607551575, "learning_rate": 9.769101037869187e-06, "loss": 1.2205, "step": 84 }, { "epoch": 0.22020725388601037, "grad_norm": 0.5526275038719177, "learning_rate": 9.762868491622229e-06, "loss": 1.1173, "step": 85 }, { "epoch": 0.22279792746113988, "grad_norm": 0.6218620538711548, "learning_rate": 9.756554987746777e-06, "loss": 1.1103, "step": 86 }, { "epoch": 0.22538860103626943, "grad_norm": 0.6305224299430847, "learning_rate": 9.750160633557626e-06, "loss": 1.132, "step": 87 }, { "epoch": 0.22797927461139897, "grad_norm": 0.6259211301803589, "learning_rate": 9.743685537743856e-06, "loss": 1.1183, "step": 88 }, { "epoch": 0.23056994818652848, "grad_norm": 0.866215169429779, "learning_rate": 9.737129810366952e-06, "loss": 1.1117, "step": 89 }, { "epoch": 0.23316062176165803, "grad_norm": 0.700805127620697, "learning_rate": 9.730493562858954e-06, "loss": 1.0444, "step": 90 }, { "epoch": 0.23575129533678757, "grad_norm": 0.630706250667572, "learning_rate": 9.72377690802055e-06, "loss": 1.2306, "step": 91 }, { "epoch": 0.23834196891191708, "grad_norm": 0.5979788899421692, "learning_rate": 9.716979960019173e-06, "loss": 1.1418, "step": 92 }, { "epoch": 0.24093264248704663, "grad_norm": 0.496480792760849, "learning_rate": 9.710102834387043e-06, "loss": 1.0891, "step": 93 }, { "epoch": 0.24352331606217617, "grad_norm": 0.5865656137466431, "learning_rate": 9.70314564801922e-06, "loss": 1.0926, "step": 94 }, { "epoch": 0.24611398963730569, "grad_norm": 0.6677915453910828, "learning_rate": 9.696108519171605e-06, "loss": 1.0742, "step": 95 }, { "epoch": 0.24870466321243523, "grad_norm": 0.6176425218582153, "learning_rate": 9.688991567458934e-06, "loss": 1.114, "step": 96 }, { "epoch": 0.25129533678756477, "grad_norm": 0.6433624625205994, "learning_rate": 9.681794913852747e-06, "loss": 1.1304, "step": 97 }, { "epoch": 0.25129533678756477, "eval_loss": 1.0442166328430176, "eval_runtime": 99.046, "eval_samples_per_second": 2.524, "eval_steps_per_second": 0.323, "step": 97 }, { "epoch": 0.2538860103626943, "grad_norm": 0.634186863899231, "learning_rate": 9.67451868067933e-06, "loss": 1.0791, "step": 98 }, { "epoch": 0.25647668393782386, "grad_norm": 0.6801970601081848, "learning_rate": 9.667162991617633e-06, "loss": 1.1209, "step": 99 }, { "epoch": 0.25906735751295334, "grad_norm": 0.5541910529136658, "learning_rate": 9.659727971697173e-06, "loss": 1.1145, "step": 100 }, { "epoch": 0.2616580310880829, "grad_norm": 0.600925624370575, "learning_rate": 9.652213747295906e-06, "loss": 1.1005, "step": 101 }, { "epoch": 0.26424870466321243, "grad_norm": 0.6611819863319397, "learning_rate": 9.644620446138078e-06, "loss": 1.1321, "step": 102 }, { "epoch": 0.266839378238342, "grad_norm": 0.6193287372589111, "learning_rate": 9.636948197292051e-06, "loss": 1.069, "step": 103 }, { "epoch": 0.2694300518134715, "grad_norm": 0.5518503189086914, "learning_rate": 9.629197131168125e-06, "loss": 1.092, "step": 104 }, { "epoch": 0.27202072538860106, "grad_norm": 0.6933501362800598, "learning_rate": 9.621367379516294e-06, "loss": 1.0589, "step": 105 }, { "epoch": 0.27461139896373055, "grad_norm": 0.7061232924461365, "learning_rate": 9.613459075424033e-06, "loss": 1.094, "step": 106 }, { "epoch": 0.2772020725388601, "grad_norm": 0.7254324555397034, "learning_rate": 9.605472353314024e-06, "loss": 1.0386, "step": 107 }, { "epoch": 0.27979274611398963, "grad_norm": 0.5736768841743469, "learning_rate": 9.597407348941865e-06, "loss": 1.0804, "step": 108 }, { "epoch": 0.2823834196891192, "grad_norm": 0.5374739170074463, "learning_rate": 9.589264199393776e-06, "loss": 1.0699, "step": 109 }, { "epoch": 0.2849740932642487, "grad_norm": 0.7706698775291443, "learning_rate": 9.58104304308426e-06, "loss": 1.0498, "step": 110 }, { "epoch": 0.28756476683937826, "grad_norm": 0.7128419876098633, "learning_rate": 9.572744019753753e-06, "loss": 1.1212, "step": 111 }, { "epoch": 0.29015544041450775, "grad_norm": 0.57721346616745, "learning_rate": 9.564367270466247e-06, "loss": 1.0665, "step": 112 }, { "epoch": 0.2927461139896373, "grad_norm": 0.6171419024467468, "learning_rate": 9.555912937606896e-06, "loss": 1.1146, "step": 113 }, { "epoch": 0.29533678756476683, "grad_norm": 0.575703501701355, "learning_rate": 9.54738116487959e-06, "loss": 1.1132, "step": 114 }, { "epoch": 0.2979274611398964, "grad_norm": 0.6700958609580994, "learning_rate": 9.53877209730452e-06, "loss": 1.1184, "step": 115 }, { "epoch": 0.3005181347150259, "grad_norm": 0.7100483179092407, "learning_rate": 9.530085881215705e-06, "loss": 1.1258, "step": 116 }, { "epoch": 0.30310880829015546, "grad_norm": 0.6098718047142029, "learning_rate": 9.52132266425851e-06, "loss": 1.1547, "step": 117 }, { "epoch": 0.30569948186528495, "grad_norm": 0.6828567981719971, "learning_rate": 9.512482595387131e-06, "loss": 1.0376, "step": 118 }, { "epoch": 0.3082901554404145, "grad_norm": 0.6913807988166809, "learning_rate": 9.503565824862076e-06, "loss": 1.1741, "step": 119 }, { "epoch": 0.31088082901554404, "grad_norm": 0.6839683651924133, "learning_rate": 9.494572504247593e-06, "loss": 1.1419, "step": 120 }, { "epoch": 0.3134715025906736, "grad_norm": 0.6861181855201721, "learning_rate": 9.485502786409107e-06, "loss": 1.1121, "step": 121 }, { "epoch": 0.3160621761658031, "grad_norm": 0.7473586201667786, "learning_rate": 9.476356825510613e-06, "loss": 1.1247, "step": 122 }, { "epoch": 0.31865284974093266, "grad_norm": 0.7107284665107727, "learning_rate": 9.467134777012063e-06, "loss": 1.0763, "step": 123 }, { "epoch": 0.32124352331606215, "grad_norm": 0.7920987010002136, "learning_rate": 9.457836797666722e-06, "loss": 1.0304, "step": 124 }, { "epoch": 0.3238341968911917, "grad_norm": 0.9145889282226562, "learning_rate": 9.448463045518498e-06, "loss": 1.2063, "step": 125 }, { "epoch": 0.32642487046632124, "grad_norm": 0.5906431674957275, "learning_rate": 9.439013679899263e-06, "loss": 1.0248, "step": 126 }, { "epoch": 0.3290155440414508, "grad_norm": 0.7043807506561279, "learning_rate": 9.429488861426137e-06, "loss": 1.1032, "step": 127 }, { "epoch": 0.3316062176165803, "grad_norm": 0.6444696187973022, "learning_rate": 9.419888751998768e-06, "loss": 1.0548, "step": 128 }, { "epoch": 0.33419689119170987, "grad_norm": 0.7440561652183533, "learning_rate": 9.410213514796565e-06, "loss": 1.0761, "step": 129 }, { "epoch": 0.33678756476683935, "grad_norm": 0.6268182992935181, "learning_rate": 9.400463314275942e-06, "loss": 1.0308, "step": 130 }, { "epoch": 0.3393782383419689, "grad_norm": 0.7046973705291748, "learning_rate": 9.390638316167513e-06, "loss": 1.061, "step": 131 }, { "epoch": 0.34196891191709844, "grad_norm": 0.6304768323898315, "learning_rate": 9.380738687473274e-06, "loss": 1.0291, "step": 132 }, { "epoch": 0.344559585492228, "grad_norm": 0.7118530869483948, "learning_rate": 9.370764596463764e-06, "loss": 1.0132, "step": 133 }, { "epoch": 0.3471502590673575, "grad_norm": 0.6919021010398865, "learning_rate": 9.360716212675213e-06, "loss": 1.0753, "step": 134 }, { "epoch": 0.34974093264248707, "grad_norm": 0.6322518587112427, "learning_rate": 9.350593706906653e-06, "loss": 1.0591, "step": 135 }, { "epoch": 0.35233160621761656, "grad_norm": 0.9555734395980835, "learning_rate": 9.340397251217009e-06, "loss": 1.0681, "step": 136 }, { "epoch": 0.3549222797927461, "grad_norm": 0.7190225720405579, "learning_rate": 9.330127018922195e-06, "loss": 1.1177, "step": 137 }, { "epoch": 0.35751295336787564, "grad_norm": 0.6554561853408813, "learning_rate": 9.319783184592142e-06, "loss": 1.0129, "step": 138 }, { "epoch": 0.3601036269430052, "grad_norm": 0.6203082203865051, "learning_rate": 9.309365924047853e-06, "loss": 1.0762, "step": 139 }, { "epoch": 0.3626943005181347, "grad_norm": 0.8245846033096313, "learning_rate": 9.298875414358399e-06, "loss": 1.0528, "step": 140 }, { "epoch": 0.36528497409326427, "grad_norm": 0.7231378555297852, "learning_rate": 9.288311833837918e-06, "loss": 1.0407, "step": 141 }, { "epoch": 0.36787564766839376, "grad_norm": 0.7020997405052185, "learning_rate": 9.27767536204258e-06, "loss": 1.1054, "step": 142 }, { "epoch": 0.3704663212435233, "grad_norm": 0.6527167558670044, "learning_rate": 9.266966179767539e-06, "loss": 1.0096, "step": 143 }, { "epoch": 0.37305699481865284, "grad_norm": 0.6411340236663818, "learning_rate": 9.256184469043852e-06, "loss": 1.0547, "step": 144 }, { "epoch": 0.3756476683937824, "grad_norm": 1.0810225009918213, "learning_rate": 9.245330413135395e-06, "loss": 1.0752, "step": 145 }, { "epoch": 0.37823834196891193, "grad_norm": 0.9295507669448853, "learning_rate": 9.23440419653574e-06, "loss": 1.0341, "step": 146 }, { "epoch": 0.38082901554404147, "grad_norm": 0.7355513572692871, "learning_rate": 9.223406004965023e-06, "loss": 1.0998, "step": 147 }, { "epoch": 0.38341968911917096, "grad_norm": 0.8607478141784668, "learning_rate": 9.212336025366789e-06, "loss": 1.026, "step": 148 }, { "epoch": 0.3860103626943005, "grad_norm": 0.7150827646255493, "learning_rate": 9.201194445904804e-06, "loss": 1.071, "step": 149 }, { "epoch": 0.38860103626943004, "grad_norm": 0.688491940498352, "learning_rate": 9.189981455959873e-06, "loss": 0.994, "step": 150 }, { "epoch": 0.3911917098445596, "grad_norm": 0.8984132409095764, "learning_rate": 9.178697246126608e-06, "loss": 1.0318, "step": 151 }, { "epoch": 0.39378238341968913, "grad_norm": 0.6616289019584656, "learning_rate": 9.167342008210191e-06, "loss": 1.0407, "step": 152 }, { "epoch": 0.3963730569948187, "grad_norm": 0.7567489147186279, "learning_rate": 9.15591593522312e-06, "loss": 1.0323, "step": 153 }, { "epoch": 0.39896373056994816, "grad_norm": 0.6091617345809937, "learning_rate": 9.144419221381919e-06, "loss": 1.0361, "step": 154 }, { "epoch": 0.4015544041450777, "grad_norm": 0.721120297908783, "learning_rate": 9.132852062103845e-06, "loss": 1.101, "step": 155 }, { "epoch": 0.40414507772020725, "grad_norm": 0.7634301781654358, "learning_rate": 9.121214654003561e-06, "loss": 1.0285, "step": 156 }, { "epoch": 0.4067357512953368, "grad_norm": 0.6629643440246582, "learning_rate": 9.109507194889793e-06, "loss": 1.0858, "step": 157 }, { "epoch": 0.40932642487046633, "grad_norm": 0.6693721413612366, "learning_rate": 9.097729883761977e-06, "loss": 1.0461, "step": 158 }, { "epoch": 0.4119170984455959, "grad_norm": 0.6852269768714905, "learning_rate": 9.085882920806862e-06, "loss": 1.0298, "step": 159 }, { "epoch": 0.41450777202072536, "grad_norm": 0.6539198756217957, "learning_rate": 9.073966507395123e-06, "loss": 1.0298, "step": 160 }, { "epoch": 0.4170984455958549, "grad_norm": 0.9533804655075073, "learning_rate": 9.061980846077925e-06, "loss": 1.062, "step": 161 }, { "epoch": 0.41968911917098445, "grad_norm": 0.6733927130699158, "learning_rate": 9.049926140583487e-06, "loss": 1.0268, "step": 162 }, { "epoch": 0.422279792746114, "grad_norm": 0.7857198715209961, "learning_rate": 9.037802595813621e-06, "loss": 1.052, "step": 163 }, { "epoch": 0.42487046632124353, "grad_norm": 0.7269684076309204, "learning_rate": 9.025610417840238e-06, "loss": 1.0026, "step": 164 }, { "epoch": 0.4274611398963731, "grad_norm": 0.8214792013168335, "learning_rate": 9.01334981390186e-06, "loss": 1.0533, "step": 165 }, { "epoch": 0.43005181347150256, "grad_norm": 0.6824995875358582, "learning_rate": 9.001020992400086e-06, "loss": 1.0437, "step": 166 }, { "epoch": 0.4326424870466321, "grad_norm": 0.6578338742256165, "learning_rate": 8.988624162896058e-06, "loss": 1.097, "step": 167 }, { "epoch": 0.43523316062176165, "grad_norm": 0.7836469411849976, "learning_rate": 8.976159536106895e-06, "loss": 0.977, "step": 168 }, { "epoch": 0.4378238341968912, "grad_norm": 0.7745511531829834, "learning_rate": 8.963627323902105e-06, "loss": 0.9812, "step": 169 }, { "epoch": 0.44041450777202074, "grad_norm": 0.712700366973877, "learning_rate": 8.951027739299996e-06, "loss": 1.0043, "step": 170 }, { "epoch": 0.4430051813471503, "grad_norm": 0.6669776439666748, "learning_rate": 8.938360996464048e-06, "loss": 1.0269, "step": 171 }, { "epoch": 0.44559585492227977, "grad_norm": 0.693252444267273, "learning_rate": 8.925627310699275e-06, "loss": 0.9963, "step": 172 }, { "epoch": 0.4481865284974093, "grad_norm": 0.741417407989502, "learning_rate": 8.91282689844856e-06, "loss": 1.0292, "step": 173 }, { "epoch": 0.45077720207253885, "grad_norm": 0.8754829168319702, "learning_rate": 8.899959977288988e-06, "loss": 1.0089, "step": 174 }, { "epoch": 0.4533678756476684, "grad_norm": 0.6871668100357056, "learning_rate": 8.887026765928129e-06, "loss": 0.9792, "step": 175 }, { "epoch": 0.45595854922279794, "grad_norm": 0.7657473683357239, "learning_rate": 8.874027484200342e-06, "loss": 1.0297, "step": 176 }, { "epoch": 0.4585492227979275, "grad_norm": 0.7051308751106262, "learning_rate": 8.860962353063022e-06, "loss": 1.0033, "step": 177 }, { "epoch": 0.46113989637305697, "grad_norm": 0.7424786686897278, "learning_rate": 8.847831594592851e-06, "loss": 1.0776, "step": 178 }, { "epoch": 0.4637305699481865, "grad_norm": 0.9562065601348877, "learning_rate": 8.834635431982022e-06, "loss": 0.893, "step": 179 }, { "epoch": 0.46632124352331605, "grad_norm": 0.7324496507644653, "learning_rate": 8.821374089534446e-06, "loss": 0.9691, "step": 180 }, { "epoch": 0.4689119170984456, "grad_norm": 0.6958107948303223, "learning_rate": 8.808047792661941e-06, "loss": 1.0113, "step": 181 }, { "epoch": 0.47150259067357514, "grad_norm": 0.7926040291786194, "learning_rate": 8.794656767880394e-06, "loss": 1.054, "step": 182 }, { "epoch": 0.4740932642487047, "grad_norm": 0.7949718236923218, "learning_rate": 8.781201242805917e-06, "loss": 1.0385, "step": 183 }, { "epoch": 0.47668393782383417, "grad_norm": 0.7993385195732117, "learning_rate": 8.767681446150977e-06, "loss": 0.987, "step": 184 }, { "epoch": 0.4792746113989637, "grad_norm": 0.8635035157203674, "learning_rate": 8.754097607720512e-06, "loss": 1.1012, "step": 185 }, { "epoch": 0.48186528497409326, "grad_norm": 0.7778316140174866, "learning_rate": 8.740449958408006e-06, "loss": 1.0547, "step": 186 }, { "epoch": 0.4844559585492228, "grad_norm": 0.7622816562652588, "learning_rate": 8.726738730191596e-06, "loss": 1.057, "step": 187 }, { "epoch": 0.48704663212435234, "grad_norm": 0.8010302782058716, "learning_rate": 8.7129641561301e-06, "loss": 1.06, "step": 188 }, { "epoch": 0.4896373056994819, "grad_norm": 0.8265953063964844, "learning_rate": 8.699126470359073e-06, "loss": 0.9859, "step": 189 }, { "epoch": 0.49222797927461137, "grad_norm": 0.7856041193008423, "learning_rate": 8.68522590808682e-06, "loss": 1.05, "step": 190 }, { "epoch": 0.4948186528497409, "grad_norm": 0.7639081478118896, "learning_rate": 8.671262705590399e-06, "loss": 1.0018, "step": 191 }, { "epoch": 0.49740932642487046, "grad_norm": 0.8716084361076355, "learning_rate": 8.657237100211604e-06, "loss": 1.0073, "step": 192 }, { "epoch": 0.5, "grad_norm": 0.7365914583206177, "learning_rate": 8.643149330352939e-06, "loss": 1.0448, "step": 193 }, { "epoch": 0.5025906735751295, "grad_norm": 0.7660321593284607, "learning_rate": 8.628999635473547e-06, "loss": 1.0415, "step": 194 }, { "epoch": 0.5025906735751295, "eval_loss": 0.9808682203292847, "eval_runtime": 99.1803, "eval_samples_per_second": 2.521, "eval_steps_per_second": 0.323, "step": 194 }, { "epoch": 0.5051813471502591, "grad_norm": 0.8740248680114746, "learning_rate": 8.61478825608516e-06, "loss": 1.08, "step": 195 }, { "epoch": 0.5077720207253886, "grad_norm": 0.8985514640808105, "learning_rate": 8.600515433748003e-06, "loss": 1.0248, "step": 196 }, { "epoch": 0.5103626943005182, "grad_norm": 0.6852197647094727, "learning_rate": 8.586181411066684e-06, "loss": 1.0513, "step": 197 }, { "epoch": 0.5129533678756477, "grad_norm": 0.8582538366317749, "learning_rate": 8.571786431686074e-06, "loss": 1.0453, "step": 198 }, { "epoch": 0.5155440414507773, "grad_norm": 0.7389397621154785, "learning_rate": 8.557330740287166e-06, "loss": 1.0516, "step": 199 }, { "epoch": 0.5181347150259067, "grad_norm": 0.7529143691062927, "learning_rate": 8.542814582582917e-06, "loss": 1.0294, "step": 200 }, { "epoch": 0.5207253886010362, "grad_norm": 0.828758180141449, "learning_rate": 8.528238205314067e-06, "loss": 1.0357, "step": 201 }, { "epoch": 0.5233160621761658, "grad_norm": 0.841716468334198, "learning_rate": 8.513601856244951e-06, "loss": 1.0827, "step": 202 }, { "epoch": 0.5259067357512953, "grad_norm": 0.807770848274231, "learning_rate": 8.498905784159282e-06, "loss": 1.0939, "step": 203 }, { "epoch": 0.5284974093264249, "grad_norm": 0.7827705144882202, "learning_rate": 8.484150238855921e-06, "loss": 1.0735, "step": 204 }, { "epoch": 0.5310880829015544, "grad_norm": 0.7675580382347107, "learning_rate": 8.469335471144646e-06, "loss": 1.0268, "step": 205 }, { "epoch": 0.533678756476684, "grad_norm": 0.767455518245697, "learning_rate": 8.454461732841864e-06, "loss": 1.0624, "step": 206 }, { "epoch": 0.5362694300518135, "grad_norm": 0.6222347617149353, "learning_rate": 8.439529276766354e-06, "loss": 1.0263, "step": 207 }, { "epoch": 0.538860103626943, "grad_norm": 0.7709267735481262, "learning_rate": 8.424538356734957e-06, "loss": 1.0703, "step": 208 }, { "epoch": 0.5414507772020726, "grad_norm": 0.9357248544692993, "learning_rate": 8.40948922755826e-06, "loss": 1.0299, "step": 209 }, { "epoch": 0.5440414507772021, "grad_norm": 0.7862836718559265, "learning_rate": 8.394382145036277e-06, "loss": 1.017, "step": 210 }, { "epoch": 0.5466321243523317, "grad_norm": 0.8551638722419739, "learning_rate": 8.379217365954089e-06, "loss": 0.9445, "step": 211 }, { "epoch": 0.5492227979274611, "grad_norm": 0.7895774841308594, "learning_rate": 8.363995148077481e-06, "loss": 1.1145, "step": 212 }, { "epoch": 0.5518134715025906, "grad_norm": 0.684273362159729, "learning_rate": 8.348715750148571e-06, "loss": 1.0439, "step": 213 }, { "epoch": 0.5544041450777202, "grad_norm": 0.748660147190094, "learning_rate": 8.333379431881398e-06, "loss": 1.034, "step": 214 }, { "epoch": 0.5569948186528497, "grad_norm": 0.8681594133377075, "learning_rate": 8.317986453957514e-06, "loss": 1.0523, "step": 215 }, { "epoch": 0.5595854922279793, "grad_norm": 0.8759576678276062, "learning_rate": 8.302537078021555e-06, "loss": 1.0167, "step": 216 }, { "epoch": 0.5621761658031088, "grad_norm": 0.8709188103675842, "learning_rate": 8.28703156667679e-06, "loss": 1.0947, "step": 217 }, { "epoch": 0.5647668393782384, "grad_norm": 0.7970831990242004, "learning_rate": 8.271470183480664e-06, "loss": 1.0404, "step": 218 }, { "epoch": 0.5673575129533679, "grad_norm": 0.8916209936141968, "learning_rate": 8.2558531929403e-06, "loss": 1.0681, "step": 219 }, { "epoch": 0.5699481865284974, "grad_norm": 0.7815295457839966, "learning_rate": 8.240180860508027e-06, "loss": 1.0273, "step": 220 }, { "epoch": 0.572538860103627, "grad_norm": 0.8419135808944702, "learning_rate": 8.224453452576855e-06, "loss": 0.974, "step": 221 }, { "epoch": 0.5751295336787565, "grad_norm": 1.0326911211013794, "learning_rate": 8.208671236475945e-06, "loss": 1.0307, "step": 222 }, { "epoch": 0.5777202072538861, "grad_norm": 0.8720448017120361, "learning_rate": 8.192834480466072e-06, "loss": 1.0061, "step": 223 }, { "epoch": 0.5803108808290155, "grad_norm": 0.7192279100418091, "learning_rate": 8.176943453735062e-06, "loss": 0.9383, "step": 224 }, { "epoch": 0.582901554404145, "grad_norm": 0.7928858995437622, "learning_rate": 8.160998426393214e-06, "loss": 1.0538, "step": 225 }, { "epoch": 0.5854922279792746, "grad_norm": 0.9114733934402466, "learning_rate": 8.144999669468714e-06, "loss": 1.0704, "step": 226 }, { "epoch": 0.5880829015544041, "grad_norm": 0.8045042753219604, "learning_rate": 8.12894745490302e-06, "loss": 1.0213, "step": 227 }, { "epoch": 0.5906735751295337, "grad_norm": 0.8414791226387024, "learning_rate": 8.112842055546254e-06, "loss": 1.1058, "step": 228 }, { "epoch": 0.5932642487046632, "grad_norm": 0.8457399606704712, "learning_rate": 8.096683745152544e-06, "loss": 1.0407, "step": 229 }, { "epoch": 0.5958549222797928, "grad_norm": 0.7906298637390137, "learning_rate": 8.080472798375392e-06, "loss": 1.0335, "step": 230 }, { "epoch": 0.5984455958549223, "grad_norm": 0.8662751913070679, "learning_rate": 8.064209490762988e-06, "loss": 1.0809, "step": 231 }, { "epoch": 0.6010362694300518, "grad_norm": 0.8119012117385864, "learning_rate": 8.04789409875354e-06, "loss": 1.0132, "step": 232 }, { "epoch": 0.6036269430051814, "grad_norm": 0.7348366975784302, "learning_rate": 8.031526899670563e-06, "loss": 1.0383, "step": 233 }, { "epoch": 0.6062176165803109, "grad_norm": 0.7896949648857117, "learning_rate": 8.015108171718177e-06, "loss": 1.0656, "step": 234 }, { "epoch": 0.6088082901554405, "grad_norm": 0.8619694113731384, "learning_rate": 7.998638193976366e-06, "loss": 1.0881, "step": 235 }, { "epoch": 0.6113989637305699, "grad_norm": 0.7404602766036987, "learning_rate": 7.982117246396246e-06, "loss": 1.0613, "step": 236 }, { "epoch": 0.6139896373056994, "grad_norm": 0.6466916799545288, "learning_rate": 7.965545609795297e-06, "loss": 1.0333, "step": 237 }, { "epoch": 0.616580310880829, "grad_norm": 0.6092397570610046, "learning_rate": 7.948923565852597e-06, "loss": 1.0161, "step": 238 }, { "epoch": 0.6191709844559585, "grad_norm": 0.7040543556213379, "learning_rate": 7.932251397104031e-06, "loss": 0.9819, "step": 239 }, { "epoch": 0.6217616580310881, "grad_norm": 1.0263625383377075, "learning_rate": 7.915529386937486e-06, "loss": 0.9477, "step": 240 }, { "epoch": 0.6243523316062176, "grad_norm": 0.6552597284317017, "learning_rate": 7.898757819588038e-06, "loss": 0.989, "step": 241 }, { "epoch": 0.6269430051813472, "grad_norm": 0.7315415143966675, "learning_rate": 7.881936980133118e-06, "loss": 1.0219, "step": 242 }, { "epoch": 0.6295336787564767, "grad_norm": 0.8863290548324585, "learning_rate": 7.86506715448767e-06, "loss": 1.0347, "step": 243 }, { "epoch": 0.6321243523316062, "grad_norm": 0.6123449206352234, "learning_rate": 7.848148629399287e-06, "loss": 1.0001, "step": 244 }, { "epoch": 0.6347150259067358, "grad_norm": 0.9395535588264465, "learning_rate": 7.831181692443338e-06, "loss": 1.008, "step": 245 }, { "epoch": 0.6373056994818653, "grad_norm": 0.6824455261230469, "learning_rate": 7.814166632018083e-06, "loss": 1.0588, "step": 246 }, { "epoch": 0.6398963730569949, "grad_norm": 0.79184889793396, "learning_rate": 7.797103737339767e-06, "loss": 1.0284, "step": 247 }, { "epoch": 0.6424870466321243, "grad_norm": 0.8167986273765564, "learning_rate": 7.779993298437704e-06, "loss": 0.9941, "step": 248 }, { "epoch": 0.6450777202072538, "grad_norm": 0.8785682916641235, "learning_rate": 7.762835606149352e-06, "loss": 1.0335, "step": 249 }, { "epoch": 0.6476683937823834, "grad_norm": 0.6932119727134705, "learning_rate": 7.745630952115365e-06, "loss": 1.002, "step": 250 }, { "epoch": 0.6502590673575129, "grad_norm": 0.8167779445648193, "learning_rate": 7.728379628774632e-06, "loss": 0.9496, "step": 251 }, { "epoch": 0.6528497409326425, "grad_norm": 0.819342315196991, "learning_rate": 7.711081929359316e-06, "loss": 1.0228, "step": 252 }, { "epoch": 0.655440414507772, "grad_norm": 0.8480210900306702, "learning_rate": 7.693738147889868e-06, "loss": 1.0217, "step": 253 }, { "epoch": 0.6580310880829016, "grad_norm": 0.8472158908843994, "learning_rate": 7.67634857917002e-06, "loss": 1.0264, "step": 254 }, { "epoch": 0.6606217616580311, "grad_norm": 0.7778662443161011, "learning_rate": 7.658913518781782e-06, "loss": 0.9766, "step": 255 }, { "epoch": 0.6632124352331606, "grad_norm": 0.8037154078483582, "learning_rate": 7.641433263080418e-06, "loss": 1.0129, "step": 256 }, { "epoch": 0.6658031088082902, "grad_norm": 0.778190016746521, "learning_rate": 7.623908109189404e-06, "loss": 1.0401, "step": 257 }, { "epoch": 0.6683937823834197, "grad_norm": 1.0345908403396606, "learning_rate": 7.606338354995381e-06, "loss": 1.0601, "step": 258 }, { "epoch": 0.6709844559585493, "grad_norm": 0.7248339653015137, "learning_rate": 7.588724299143091e-06, "loss": 1.0098, "step": 259 }, { "epoch": 0.6735751295336787, "grad_norm": 0.6873196959495544, "learning_rate": 7.571066241030302e-06, "loss": 1.0574, "step": 260 }, { "epoch": 0.6761658031088082, "grad_norm": 0.7845073938369751, "learning_rate": 7.553364480802715e-06, "loss": 1.0041, "step": 261 }, { "epoch": 0.6787564766839378, "grad_norm": 0.7601717114448547, "learning_rate": 7.5356193193488655e-06, "loss": 0.9021, "step": 262 }, { "epoch": 0.6813471502590673, "grad_norm": 0.8694692850112915, "learning_rate": 7.517831058295013e-06, "loss": 1.0321, "step": 263 }, { "epoch": 0.6839378238341969, "grad_norm": 0.8670944571495056, "learning_rate": 7.500000000000001e-06, "loss": 0.9876, "step": 264 }, { "epoch": 0.6865284974093264, "grad_norm": 0.9318599700927734, "learning_rate": 7.4821264475501325e-06, "loss": 1.0528, "step": 265 }, { "epoch": 0.689119170984456, "grad_norm": 0.9383565187454224, "learning_rate": 7.464210704754009e-06, "loss": 0.9914, "step": 266 }, { "epoch": 0.6917098445595855, "grad_norm": 0.8553853034973145, "learning_rate": 7.446253076137372e-06, "loss": 1.0909, "step": 267 }, { "epoch": 0.694300518134715, "grad_norm": 0.8563231229782104, "learning_rate": 7.4282538669379186e-06, "loss": 1.0639, "step": 268 }, { "epoch": 0.6968911917098446, "grad_norm": 0.8212496042251587, "learning_rate": 7.410213383100126e-06, "loss": 0.9786, "step": 269 }, { "epoch": 0.6994818652849741, "grad_norm": 0.9418748617172241, "learning_rate": 7.3921319312700365e-06, "loss": 1.0663, "step": 270 }, { "epoch": 0.7020725388601037, "grad_norm": 0.7147770524024963, "learning_rate": 7.374009818790058e-06, "loss": 1.0187, "step": 271 }, { "epoch": 0.7046632124352331, "grad_norm": 0.8993127346038818, "learning_rate": 7.355847353693729e-06, "loss": 0.9813, "step": 272 }, { "epoch": 0.7072538860103627, "grad_norm": 0.8212840557098389, "learning_rate": 7.337644844700494e-06, "loss": 0.9602, "step": 273 }, { "epoch": 0.7098445595854922, "grad_norm": 0.8086667656898499, "learning_rate": 7.319402601210448e-06, "loss": 1.1018, "step": 274 }, { "epoch": 0.7124352331606217, "grad_norm": 0.8487491607666016, "learning_rate": 7.301120933299076e-06, "loss": 0.9453, "step": 275 }, { "epoch": 0.7150259067357513, "grad_norm": 1.0112347602844238, "learning_rate": 7.282800151711991e-06, "loss": 0.9753, "step": 276 }, { "epoch": 0.7176165803108808, "grad_norm": 0.7847304940223694, "learning_rate": 7.264440567859645e-06, "loss": 1.0121, "step": 277 }, { "epoch": 0.7202072538860104, "grad_norm": 0.7814239263534546, "learning_rate": 7.246042493812036e-06, "loss": 1.0152, "step": 278 }, { "epoch": 0.7227979274611399, "grad_norm": 0.9602167010307312, "learning_rate": 7.227606242293409e-06, "loss": 1.057, "step": 279 }, { "epoch": 0.7253886010362695, "grad_norm": 0.8330225944519043, "learning_rate": 7.209132126676934e-06, "loss": 1.0834, "step": 280 }, { "epoch": 0.727979274611399, "grad_norm": 0.8452211618423462, "learning_rate": 7.190620460979384e-06, "loss": 1.0382, "step": 281 }, { "epoch": 0.7305699481865285, "grad_norm": 0.7918412089347839, "learning_rate": 7.172071559855792e-06, "loss": 0.9387, "step": 282 }, { "epoch": 0.7331606217616581, "grad_norm": 0.8554849624633789, "learning_rate": 7.153485738594111e-06, "loss": 1.0589, "step": 283 }, { "epoch": 0.7357512953367875, "grad_norm": 0.906253457069397, "learning_rate": 7.134863313109847e-06, "loss": 1.001, "step": 284 }, { "epoch": 0.7383419689119171, "grad_norm": 0.9388474225997925, "learning_rate": 7.116204599940693e-06, "loss": 1.0046, "step": 285 }, { "epoch": 0.7409326424870466, "grad_norm": 0.8858338594436646, "learning_rate": 7.097509916241145e-06, "loss": 1.0284, "step": 286 }, { "epoch": 0.7435233160621761, "grad_norm": 0.8751007914543152, "learning_rate": 7.078779579777122e-06, "loss": 1.0375, "step": 287 }, { "epoch": 0.7461139896373057, "grad_norm": 0.9136319160461426, "learning_rate": 7.060013908920549e-06, "loss": 1.0254, "step": 288 }, { "epoch": 0.7487046632124352, "grad_norm": 0.8126935958862305, "learning_rate": 7.041213222643952e-06, "loss": 1.0091, "step": 289 }, { "epoch": 0.7512953367875648, "grad_norm": 0.8925020098686218, "learning_rate": 7.022377840515047e-06, "loss": 0.9436, "step": 290 }, { "epoch": 0.7538860103626943, "grad_norm": 0.7228291630744934, "learning_rate": 7.003508082691286e-06, "loss": 0.9353, "step": 291 }, { "epoch": 0.7538860103626943, "eval_loss": 0.9649054408073425, "eval_runtime": 99.1248, "eval_samples_per_second": 2.522, "eval_steps_per_second": 0.323, "step": 291 }, { "epoch": 0.7564766839378239, "grad_norm": 0.7875744104385376, "learning_rate": 6.984604269914437e-06, "loss": 1.0115, "step": 292 }, { "epoch": 0.7590673575129534, "grad_norm": 0.8201349377632141, "learning_rate": 6.965666723505118e-06, "loss": 1.0683, "step": 293 }, { "epoch": 0.7616580310880829, "grad_norm": 0.9053636193275452, "learning_rate": 6.94669576535734e-06, "loss": 0.9764, "step": 294 }, { "epoch": 0.7642487046632125, "grad_norm": 0.7280160784721375, "learning_rate": 6.927691717933038e-06, "loss": 1.0246, "step": 295 }, { "epoch": 0.7668393782383419, "grad_norm": 0.756608784198761, "learning_rate": 6.908654904256584e-06, "loss": 1.0057, "step": 296 }, { "epoch": 0.7694300518134715, "grad_norm": 0.8075028657913208, "learning_rate": 6.889585647909303e-06, "loss": 1.0235, "step": 297 }, { "epoch": 0.772020725388601, "grad_norm": 0.7520319223403931, "learning_rate": 6.870484273023967e-06, "loss": 1.007, "step": 298 }, { "epoch": 0.7746113989637305, "grad_norm": 0.8978586792945862, "learning_rate": 6.8513511042792895e-06, "loss": 0.9499, "step": 299 }, { "epoch": 0.7772020725388601, "grad_norm": 0.8076562881469727, "learning_rate": 6.832186466894402e-06, "loss": 0.9844, "step": 300 }, { "epoch": 0.7797927461139896, "grad_norm": 0.9009307026863098, "learning_rate": 6.812990686623335e-06, "loss": 1.0019, "step": 301 }, { "epoch": 0.7823834196891192, "grad_norm": 0.7848314046859741, "learning_rate": 6.793764089749473e-06, "loss": 1.0229, "step": 302 }, { "epoch": 0.7849740932642487, "grad_norm": 0.8111161589622498, "learning_rate": 6.7745070030800075e-06, "loss": 1.0046, "step": 303 }, { "epoch": 0.7875647668393783, "grad_norm": 0.8512140512466431, "learning_rate": 6.755219753940389e-06, "loss": 0.9784, "step": 304 }, { "epoch": 0.7901554404145078, "grad_norm": 0.7778370380401611, "learning_rate": 6.735902670168758e-06, "loss": 1.0444, "step": 305 }, { "epoch": 0.7927461139896373, "grad_norm": 0.8520141839981079, "learning_rate": 6.716556080110374e-06, "loss": 1.036, "step": 306 }, { "epoch": 0.7953367875647669, "grad_norm": 0.9404623508453369, "learning_rate": 6.6971803126120336e-06, "loss": 1.1203, "step": 307 }, { "epoch": 0.7979274611398963, "grad_norm": 0.9082056283950806, "learning_rate": 6.677775697016484e-06, "loss": 1.0798, "step": 308 }, { "epoch": 0.8005181347150259, "grad_norm": 0.7969542145729065, "learning_rate": 6.658342563156821e-06, "loss": 0.9632, "step": 309 }, { "epoch": 0.8031088082901554, "grad_norm": 0.7916974425315857, "learning_rate": 6.638881241350884e-06, "loss": 1.0198, "step": 310 }, { "epoch": 0.805699481865285, "grad_norm": 0.9639979600906372, "learning_rate": 6.619392062395643e-06, "loss": 1.0056, "step": 311 }, { "epoch": 0.8082901554404145, "grad_norm": 0.9673221707344055, "learning_rate": 6.599875357561572e-06, "loss": 0.9875, "step": 312 }, { "epoch": 0.810880829015544, "grad_norm": 1.0038220882415771, "learning_rate": 6.5803314585870225e-06, "loss": 1.1161, "step": 313 }, { "epoch": 0.8134715025906736, "grad_norm": 0.9063149094581604, "learning_rate": 6.560760697672583e-06, "loss": 0.9699, "step": 314 }, { "epoch": 0.8160621761658031, "grad_norm": 0.9444339871406555, "learning_rate": 6.541163407475433e-06, "loss": 1.0099, "step": 315 }, { "epoch": 0.8186528497409327, "grad_norm": 0.869704008102417, "learning_rate": 6.5215399211036815e-06, "loss": 1.0076, "step": 316 }, { "epoch": 0.8212435233160622, "grad_norm": 0.7657658457756042, "learning_rate": 6.50189057211072e-06, "loss": 1.0471, "step": 317 }, { "epoch": 0.8238341968911918, "grad_norm": 0.8859752416610718, "learning_rate": 6.4822156944895375e-06, "loss": 0.9323, "step": 318 }, { "epoch": 0.8264248704663213, "grad_norm": 0.7193804979324341, "learning_rate": 6.462515622667056e-06, "loss": 1.042, "step": 319 }, { "epoch": 0.8290155440414507, "grad_norm": 0.7999783754348755, "learning_rate": 6.442790691498433e-06, "loss": 0.967, "step": 320 }, { "epoch": 0.8316062176165803, "grad_norm": 0.8761709928512573, "learning_rate": 6.423041236261381e-06, "loss": 1.0732, "step": 321 }, { "epoch": 0.8341968911917098, "grad_norm": 1.0846220254898071, "learning_rate": 6.403267592650466e-06, "loss": 0.9741, "step": 322 }, { "epoch": 0.8367875647668394, "grad_norm": 0.6978609561920166, "learning_rate": 6.383470096771396e-06, "loss": 1.0009, "step": 323 }, { "epoch": 0.8393782383419689, "grad_norm": 0.8032934069633484, "learning_rate": 6.363649085135311e-06, "loss": 0.9546, "step": 324 }, { "epoch": 0.8419689119170984, "grad_norm": 1.0093727111816406, "learning_rate": 6.343804894653072e-06, "loss": 1.0592, "step": 325 }, { "epoch": 0.844559585492228, "grad_norm": 0.8561227917671204, "learning_rate": 6.323937862629513e-06, "loss": 1.0551, "step": 326 }, { "epoch": 0.8471502590673575, "grad_norm": 0.7876784801483154, "learning_rate": 6.304048326757735e-06, "loss": 1.0564, "step": 327 }, { "epoch": 0.8497409326424871, "grad_norm": 0.7641052603721619, "learning_rate": 6.2841366251133405e-06, "loss": 0.9427, "step": 328 }, { "epoch": 0.8523316062176166, "grad_norm": 0.8268167972564697, "learning_rate": 6.2642030961487046e-06, "loss": 1.0171, "step": 329 }, { "epoch": 0.8549222797927462, "grad_norm": 0.7793816328048706, "learning_rate": 6.244248078687213e-06, "loss": 0.9475, "step": 330 }, { "epoch": 0.8575129533678757, "grad_norm": 0.7105771899223328, "learning_rate": 6.224271911917508e-06, "loss": 1.0326, "step": 331 }, { "epoch": 0.8601036269430051, "grad_norm": 0.8429663777351379, "learning_rate": 6.204274935387716e-06, "loss": 1.0257, "step": 332 }, { "epoch": 0.8626943005181347, "grad_norm": 0.8630242347717285, "learning_rate": 6.184257488999688e-06, "loss": 1.0037, "step": 333 }, { "epoch": 0.8652849740932642, "grad_norm": 1.0064729452133179, "learning_rate": 6.164219913003208e-06, "loss": 1.0468, "step": 334 }, { "epoch": 0.8678756476683938, "grad_norm": 0.9719647765159607, "learning_rate": 6.14416254799022e-06, "loss": 1.0582, "step": 335 }, { "epoch": 0.8704663212435233, "grad_norm": 0.8877604007720947, "learning_rate": 6.124085734889034e-06, "loss": 0.9288, "step": 336 }, { "epoch": 0.8730569948186528, "grad_norm": 0.8161412477493286, "learning_rate": 6.1039898149585305e-06, "loss": 0.9929, "step": 337 }, { "epoch": 0.8756476683937824, "grad_norm": 0.7888549566268921, "learning_rate": 6.083875129782366e-06, "loss": 1.022, "step": 338 }, { "epoch": 0.8782383419689119, "grad_norm": 0.8845784664154053, "learning_rate": 6.063742021263157e-06, "loss": 0.9448, "step": 339 }, { "epoch": 0.8808290155440415, "grad_norm": 0.791270911693573, "learning_rate": 6.043590831616677e-06, "loss": 0.9783, "step": 340 }, { "epoch": 0.883419689119171, "grad_norm": 0.8304092288017273, "learning_rate": 6.023421903366034e-06, "loss": 1.0105, "step": 341 }, { "epoch": 0.8860103626943006, "grad_norm": 0.7290758490562439, "learning_rate": 6.003235579335851e-06, "loss": 1.011, "step": 342 }, { "epoch": 0.8886010362694301, "grad_norm": 0.9048756957054138, "learning_rate": 5.9830322026464435e-06, "loss": 1.0133, "step": 343 }, { "epoch": 0.8911917098445595, "grad_norm": 0.8403552770614624, "learning_rate": 5.962812116707977e-06, "loss": 1.0031, "step": 344 }, { "epoch": 0.8937823834196891, "grad_norm": 0.8173761963844299, "learning_rate": 5.942575665214634e-06, "loss": 1.007, "step": 345 }, { "epoch": 0.8963730569948186, "grad_norm": 0.8470476865768433, "learning_rate": 5.92232319213878e-06, "loss": 0.9703, "step": 346 }, { "epoch": 0.8989637305699482, "grad_norm": 0.816389262676239, "learning_rate": 5.902055041725105e-06, "loss": 0.9642, "step": 347 }, { "epoch": 0.9015544041450777, "grad_norm": 0.7632498741149902, "learning_rate": 5.8817715584847744e-06, "loss": 1.0025, "step": 348 }, { "epoch": 0.9041450777202072, "grad_norm": 0.8325628638267517, "learning_rate": 5.861473087189584e-06, "loss": 1.0194, "step": 349 }, { "epoch": 0.9067357512953368, "grad_norm": 0.921424150466919, "learning_rate": 5.841159972866085e-06, "loss": 0.9659, "step": 350 }, { "epoch": 0.9093264248704663, "grad_norm": 0.8112065196037292, "learning_rate": 5.820832560789727e-06, "loss": 1.0551, "step": 351 }, { "epoch": 0.9119170984455959, "grad_norm": 0.9384793043136597, "learning_rate": 5.800491196478989e-06, "loss": 0.9915, "step": 352 }, { "epoch": 0.9145077720207254, "grad_norm": 0.6935743689537048, "learning_rate": 5.780136225689505e-06, "loss": 1.0039, "step": 353 }, { "epoch": 0.917098445595855, "grad_norm": 0.7410438060760498, "learning_rate": 5.759767994408188e-06, "loss": 0.8865, "step": 354 }, { "epoch": 0.9196891191709845, "grad_norm": 0.6849777102470398, "learning_rate": 5.739386848847346e-06, "loss": 0.9646, "step": 355 }, { "epoch": 0.9222797927461139, "grad_norm": 0.8872363567352295, "learning_rate": 5.718993135438803e-06, "loss": 1.0218, "step": 356 }, { "epoch": 0.9248704663212435, "grad_norm": 0.8674810528755188, "learning_rate": 5.6985872008280045e-06, "loss": 1.0209, "step": 357 }, { "epoch": 0.927461139896373, "grad_norm": 0.9334598779678345, "learning_rate": 5.678169391868128e-06, "loss": 0.9841, "step": 358 }, { "epoch": 0.9300518134715026, "grad_norm": 0.7986560463905334, "learning_rate": 5.6577400556141906e-06, "loss": 0.9821, "step": 359 }, { "epoch": 0.9326424870466321, "grad_norm": 0.7242634296417236, "learning_rate": 5.637299539317141e-06, "loss": 1.0017, "step": 360 }, { "epoch": 0.9352331606217616, "grad_norm": 0.9153129458427429, "learning_rate": 5.616848190417965e-06, "loss": 0.9975, "step": 361 }, { "epoch": 0.9378238341968912, "grad_norm": 1.0346280336380005, "learning_rate": 5.596386356541779e-06, "loss": 1.0298, "step": 362 }, { "epoch": 0.9404145077720207, "grad_norm": 0.7707669734954834, "learning_rate": 5.575914385491917e-06, "loss": 0.9551, "step": 363 }, { "epoch": 0.9430051813471503, "grad_norm": 0.7947671413421631, "learning_rate": 5.555432625244024e-06, "loss": 0.9803, "step": 364 }, { "epoch": 0.9455958549222798, "grad_norm": 0.8810315132141113, "learning_rate": 5.534941423940135e-06, "loss": 0.9941, "step": 365 }, { "epoch": 0.9481865284974094, "grad_norm": 0.848483681678772, "learning_rate": 5.51444112988276e-06, "loss": 1.0864, "step": 366 }, { "epoch": 0.9507772020725389, "grad_norm": 0.816373884677887, "learning_rate": 5.493932091528972e-06, "loss": 0.9821, "step": 367 }, { "epoch": 0.9533678756476683, "grad_norm": 0.8828781843185425, "learning_rate": 5.473414657484468e-06, "loss": 0.9929, "step": 368 }, { "epoch": 0.9559585492227979, "grad_norm": 0.8907682299613953, "learning_rate": 5.452889176497659e-06, "loss": 0.9888, "step": 369 }, { "epoch": 0.9585492227979274, "grad_norm": 0.9210544228553772, "learning_rate": 5.432355997453729e-06, "loss": 0.9757, "step": 370 }, { "epoch": 0.961139896373057, "grad_norm": 0.9528911709785461, "learning_rate": 5.4118154693687165e-06, "loss": 0.9014, "step": 371 }, { "epoch": 0.9637305699481865, "grad_norm": 0.7835102677345276, "learning_rate": 5.391267941383572e-06, "loss": 0.9736, "step": 372 }, { "epoch": 0.966321243523316, "grad_norm": 0.9772538542747498, "learning_rate": 5.3707137627582315e-06, "loss": 1.0071, "step": 373 }, { "epoch": 0.9689119170984456, "grad_norm": 0.8421518802642822, "learning_rate": 5.350153282865674e-06, "loss": 1.0501, "step": 374 }, { "epoch": 0.9715025906735751, "grad_norm": 0.7764169573783875, "learning_rate": 5.329586851185987e-06, "loss": 1.0491, "step": 375 }, { "epoch": 0.9740932642487047, "grad_norm": 0.7631876468658447, "learning_rate": 5.309014817300422e-06, "loss": 1.0018, "step": 376 }, { "epoch": 0.9766839378238342, "grad_norm": 0.816758394241333, "learning_rate": 5.2884375308854565e-06, "loss": 1.0032, "step": 377 }, { "epoch": 0.9792746113989638, "grad_norm": 0.7835342288017273, "learning_rate": 5.26785534170685e-06, "loss": 0.9686, "step": 378 }, { "epoch": 0.9818652849740933, "grad_norm": 0.8044567704200745, "learning_rate": 5.247268599613696e-06, "loss": 0.9695, "step": 379 }, { "epoch": 0.9844559585492227, "grad_norm": 0.883346438407898, "learning_rate": 5.226677654532476e-06, "loss": 0.9709, "step": 380 }, { "epoch": 0.9870466321243523, "grad_norm": 0.9801338911056519, "learning_rate": 5.206082856461115e-06, "loss": 0.9887, "step": 381 }, { "epoch": 0.9896373056994818, "grad_norm": 0.8825283646583557, "learning_rate": 5.185484555463026e-06, "loss": 0.9649, "step": 382 }, { "epoch": 0.9922279792746114, "grad_norm": 1.0225670337677002, "learning_rate": 5.16488310166117e-06, "loss": 1.0526, "step": 383 }, { "epoch": 0.9948186528497409, "grad_norm": 1.0191495418548584, "learning_rate": 5.1442788452320915e-06, "loss": 0.9138, "step": 384 }, { "epoch": 0.9974093264248705, "grad_norm": 0.8653509616851807, "learning_rate": 5.123672136399975e-06, "loss": 0.9498, "step": 385 }, { "epoch": 1.0, "grad_norm": 0.9232073426246643, "learning_rate": 5.1030633254306935e-06, "loss": 0.9828, "step": 386 }, { "epoch": 1.0025906735751295, "grad_norm": 0.8782436847686768, "learning_rate": 5.082452762625848e-06, "loss": 1.0139, "step": 387 }, { "epoch": 1.005181347150259, "grad_norm": 0.9029590487480164, "learning_rate": 5.061840798316815e-06, "loss": 1.0483, "step": 388 }, { "epoch": 1.005181347150259, "eval_loss": 0.9548383951187134, "eval_runtime": 99.0137, "eval_samples_per_second": 2.525, "eval_steps_per_second": 0.323, "step": 388 }, { "epoch": 1.0077720207253886, "grad_norm": 0.9202731847763062, "learning_rate": 5.041227782858799e-06, "loss": 1.0369, "step": 389 }, { "epoch": 1.0103626943005182, "grad_norm": 0.8768073320388794, "learning_rate": 5.020614066624868e-06, "loss": 1.0371, "step": 390 }, { "epoch": 1.0129533678756477, "grad_norm": 1.0436216592788696, "learning_rate": 5e-06, "loss": 1.0636, "step": 391 }, { "epoch": 1.0155440414507773, "grad_norm": 0.8364196419715881, "learning_rate": 4.979385933375133e-06, "loss": 0.9556, "step": 392 }, { "epoch": 1.0181347150259068, "grad_norm": 0.8371309041976929, "learning_rate": 4.958772217141203e-06, "loss": 0.9272, "step": 393 }, { "epoch": 1.0207253886010363, "grad_norm": 0.7747754454612732, "learning_rate": 4.9381592016831856e-06, "loss": 1.0231, "step": 394 }, { "epoch": 1.0233160621761659, "grad_norm": 0.7783029079437256, "learning_rate": 4.917547237374153e-06, "loss": 0.9855, "step": 395 }, { "epoch": 1.0025906735751295, "grad_norm": 0.9176076650619507, "learning_rate": 4.896936674569309e-06, "loss": 1.0037, "step": 396 }, { "epoch": 1.005181347150259, "grad_norm": 0.9758632183074951, "learning_rate": 4.876327863600026e-06, "loss": 0.987, "step": 397 }, { "epoch": 1.0077720207253886, "grad_norm": 0.8266388177871704, "learning_rate": 4.85572115476791e-06, "loss": 1.0135, "step": 398 }, { "epoch": 1.0103626943005182, "grad_norm": 0.8124467730522156, "learning_rate": 4.83511689833883e-06, "loss": 0.9964, "step": 399 }, { "epoch": 1.0129533678756477, "grad_norm": 1.061855673789978, "learning_rate": 4.814515444536975e-06, "loss": 1.0406, "step": 400 }, { "epoch": 1.0155440414507773, "grad_norm": 0.89823317527771, "learning_rate": 4.793917143538887e-06, "loss": 1.0009, "step": 401 }, { "epoch": 1.0181347150259068, "grad_norm": 0.7810267806053162, "learning_rate": 4.773322345467525e-06, "loss": 1.0269, "step": 402 }, { "epoch": 1.0207253886010363, "grad_norm": 1.0242594480514526, "learning_rate": 4.752731400386306e-06, "loss": 0.9579, "step": 403 }, { "epoch": 1.0233160621761659, "grad_norm": 0.8448682427406311, "learning_rate": 4.732144658293151e-06, "loss": 1.0156, "step": 404 }, { "epoch": 1.0259067357512954, "grad_norm": 1.009443998336792, "learning_rate": 4.711562469114544e-06, "loss": 1.0358, "step": 405 }, { "epoch": 1.028497409326425, "grad_norm": 0.9292247891426086, "learning_rate": 4.690985182699581e-06, "loss": 0.9328, "step": 406 }, { "epoch": 1.0310880829015545, "grad_norm": 0.7662960886955261, "learning_rate": 4.670413148814015e-06, "loss": 0.9747, "step": 407 }, { "epoch": 1.0336787564766838, "grad_norm": 0.8195393681526184, "learning_rate": 4.649846717134327e-06, "loss": 0.9467, "step": 408 }, { "epoch": 1.0362694300518134, "grad_norm": 0.8635261058807373, "learning_rate": 4.62928623724177e-06, "loss": 0.9653, "step": 409 }, { "epoch": 1.038860103626943, "grad_norm": 0.929542064666748, "learning_rate": 4.6087320586164296e-06, "loss": 0.9214, "step": 410 }, { "epoch": 1.0414507772020725, "grad_norm": 0.8770561814308167, "learning_rate": 4.588184530631284e-06, "loss": 0.9743, "step": 411 }, { "epoch": 1.044041450777202, "grad_norm": 0.7577627301216125, "learning_rate": 4.567644002546273e-06, "loss": 1.016, "step": 412 }, { "epoch": 1.0466321243523315, "grad_norm": 0.888802707195282, "learning_rate": 4.547110823502343e-06, "loss": 0.9667, "step": 413 }, { "epoch": 1.049222797927461, "grad_norm": 0.7431952357292175, "learning_rate": 4.526585342515533e-06, "loss": 1.0157, "step": 414 }, { "epoch": 1.0518134715025906, "grad_norm": 0.9357666969299316, "learning_rate": 4.506067908471029e-06, "loss": 1.0433, "step": 415 }, { "epoch": 1.0544041450777202, "grad_norm": 0.9243563413619995, "learning_rate": 4.485558870117241e-06, "loss": 0.9251, "step": 416 }, { "epoch": 1.0569948186528497, "grad_norm": 0.9315341711044312, "learning_rate": 4.465058576059868e-06, "loss": 0.9785, "step": 417 }, { "epoch": 1.0595854922279793, "grad_norm": 1.0324779748916626, "learning_rate": 4.444567374755978e-06, "loss": 0.994, "step": 418 }, { "epoch": 1.0621761658031088, "grad_norm": 0.784785270690918, "learning_rate": 4.424085614508084e-06, "loss": 1.0312, "step": 419 }, { "epoch": 1.0647668393782384, "grad_norm": 0.9518358111381531, "learning_rate": 4.403613643458222e-06, "loss": 1.0396, "step": 420 }, { "epoch": 1.067357512953368, "grad_norm": 1.0345162153244019, "learning_rate": 4.383151809582035e-06, "loss": 1.0877, "step": 421 }, { "epoch": 1.0699481865284974, "grad_norm": 0.8149455189704895, "learning_rate": 4.362700460682861e-06, "loss": 0.9827, "step": 422 }, { "epoch": 1.072538860103627, "grad_norm": 0.8805851936340332, "learning_rate": 4.342259944385811e-06, "loss": 0.9838, "step": 423 }, { "epoch": 1.0751295336787565, "grad_norm": 0.8283563852310181, "learning_rate": 4.321830608131872e-06, "loss": 1.0128, "step": 424 }, { "epoch": 1.077720207253886, "grad_norm": 1.1321141719818115, "learning_rate": 4.301412799171998e-06, "loss": 0.8987, "step": 425 }, { "epoch": 1.0803108808290156, "grad_norm": 0.9323210120201111, "learning_rate": 4.281006864561199e-06, "loss": 0.9618, "step": 426 }, { "epoch": 1.0829015544041452, "grad_norm": 0.7723137140274048, "learning_rate": 4.260613151152655e-06, "loss": 0.9636, "step": 427 }, { "epoch": 1.0854922279792747, "grad_norm": 0.9840835928916931, "learning_rate": 4.240232005591816e-06, "loss": 0.9963, "step": 428 }, { "epoch": 1.0880829015544042, "grad_norm": 0.8220058679580688, "learning_rate": 4.219863774310497e-06, "loss": 0.9901, "step": 429 }, { "epoch": 1.0906735751295338, "grad_norm": 0.9598246216773987, "learning_rate": 4.1995088035210126e-06, "loss": 0.9561, "step": 430 }, { "epoch": 1.093264248704663, "grad_norm": 0.7539969682693481, "learning_rate": 4.179167439210275e-06, "loss": 0.9904, "step": 431 }, { "epoch": 1.0958549222797926, "grad_norm": 0.8430355191230774, "learning_rate": 4.158840027133917e-06, "loss": 1.0678, "step": 432 }, { "epoch": 1.0984455958549222, "grad_norm": 0.9641113877296448, "learning_rate": 4.138526912810418e-06, "loss": 0.9488, "step": 433 }, { "epoch": 1.1010362694300517, "grad_norm": 0.8497964143753052, "learning_rate": 4.1182284415152255e-06, "loss": 1.0023, "step": 434 }, { "epoch": 1.1036269430051813, "grad_norm": 0.8309699892997742, "learning_rate": 4.097944958274898e-06, "loss": 0.9936, "step": 435 }, { "epoch": 1.1062176165803108, "grad_norm": 1.0693992376327515, "learning_rate": 4.077676807861221e-06, "loss": 0.9811, "step": 436 }, { "epoch": 1.1088082901554404, "grad_norm": 0.8500522375106812, "learning_rate": 4.057424334785366e-06, "loss": 0.9753, "step": 437 }, { "epoch": 1.11139896373057, "grad_norm": 0.7945443391799927, "learning_rate": 4.037187883292027e-06, "loss": 0.9833, "step": 438 }, { "epoch": 1.1139896373056994, "grad_norm": 0.7603841423988342, "learning_rate": 4.016967797353558e-06, "loss": 0.9531, "step": 439 }, { "epoch": 1.116580310880829, "grad_norm": 0.9688854217529297, "learning_rate": 3.996764420664149e-06, "loss": 0.9785, "step": 440 }, { "epoch": 1.1191709844559585, "grad_norm": 1.072771668434143, "learning_rate": 3.976578096633969e-06, "loss": 0.9511, "step": 441 }, { "epoch": 1.121761658031088, "grad_norm": 0.8664917349815369, "learning_rate": 3.956409168383325e-06, "loss": 0.9434, "step": 442 }, { "epoch": 1.1243523316062176, "grad_norm": 0.8430099487304688, "learning_rate": 3.936257978736845e-06, "loss": 0.9397, "step": 443 }, { "epoch": 1.1269430051813472, "grad_norm": 0.9500722885131836, "learning_rate": 3.916124870217635e-06, "loss": 1.0198, "step": 444 }, { "epoch": 1.1295336787564767, "grad_norm": 1.0132358074188232, "learning_rate": 3.89601018504147e-06, "loss": 1.0048, "step": 445 }, { "epoch": 1.1321243523316062, "grad_norm": 0.8455902338027954, "learning_rate": 3.875914265110967e-06, "loss": 0.9248, "step": 446 }, { "epoch": 1.1347150259067358, "grad_norm": 0.8625523447990417, "learning_rate": 3.85583745200978e-06, "loss": 1.0251, "step": 447 }, { "epoch": 1.1373056994818653, "grad_norm": 0.9898774027824402, "learning_rate": 3.835780086996794e-06, "loss": 0.998, "step": 448 }, { "epoch": 1.1398963730569949, "grad_norm": 1.021031141281128, "learning_rate": 3.815742511000313e-06, "loss": 1.0321, "step": 449 }, { "epoch": 1.1424870466321244, "grad_norm": 0.8724047541618347, "learning_rate": 3.7957250646122843e-06, "loss": 0.9535, "step": 450 }, { "epoch": 1.145077720207254, "grad_norm": 0.820254921913147, "learning_rate": 3.7757280880824946e-06, "loss": 0.9489, "step": 451 }, { "epoch": 1.1476683937823835, "grad_norm": 1.034761667251587, "learning_rate": 3.755751921312788e-06, "loss": 0.9903, "step": 452 }, { "epoch": 1.150259067357513, "grad_norm": 0.8738400936126709, "learning_rate": 3.735796903851297e-06, "loss": 0.9504, "step": 453 }, { "epoch": 1.1528497409326426, "grad_norm": 0.8705572485923767, "learning_rate": 3.715863374886661e-06, "loss": 0.9571, "step": 454 }, { "epoch": 1.1554404145077721, "grad_norm": 0.8942792415618896, "learning_rate": 3.695951673242267e-06, "loss": 0.932, "step": 455 }, { "epoch": 1.1580310880829017, "grad_norm": 0.9116060733795166, "learning_rate": 3.6760621373704867e-06, "loss": 0.9602, "step": 456 }, { "epoch": 1.160621761658031, "grad_norm": 0.9232982993125916, "learning_rate": 3.6561951053469313e-06, "loss": 0.9944, "step": 457 }, { "epoch": 1.1632124352331605, "grad_norm": 0.6849638223648071, "learning_rate": 3.636350914864689e-06, "loss": 1.014, "step": 458 }, { "epoch": 1.16580310880829, "grad_norm": 1.0841094255447388, "learning_rate": 3.6165299032286055e-06, "loss": 0.8914, "step": 459 }, { "epoch": 1.1683937823834196, "grad_norm": 0.8563059568405151, "learning_rate": 3.5967324073495363e-06, "loss": 0.972, "step": 460 }, { "epoch": 1.1709844559585492, "grad_norm": 1.0777771472930908, "learning_rate": 3.5769587637386206e-06, "loss": 1.0476, "step": 461 }, { "epoch": 1.1735751295336787, "grad_norm": 0.790111780166626, "learning_rate": 3.5572093085015683e-06, "loss": 0.959, "step": 462 }, { "epoch": 1.1761658031088082, "grad_norm": 0.8875803351402283, "learning_rate": 3.537484377332945e-06, "loss": 1.0109, "step": 463 }, { "epoch": 1.1787564766839378, "grad_norm": 0.932518720626831, "learning_rate": 3.5177843055104633e-06, "loss": 1.0012, "step": 464 }, { "epoch": 1.1813471502590673, "grad_norm": 0.776736855506897, "learning_rate": 3.4981094278892813e-06, "loss": 0.9852, "step": 465 }, { "epoch": 1.1839378238341969, "grad_norm": 0.8646918535232544, "learning_rate": 3.4784600788963197e-06, "loss": 1.0376, "step": 466 }, { "epoch": 1.1865284974093264, "grad_norm": 0.9904392957687378, "learning_rate": 3.458836592524569e-06, "loss": 0.9692, "step": 467 }, { "epoch": 1.189119170984456, "grad_norm": 0.8158794641494751, "learning_rate": 3.4392393023274173e-06, "loss": 1.0176, "step": 468 }, { "epoch": 1.1917098445595855, "grad_norm": 0.8880725502967834, "learning_rate": 3.419668541412977e-06, "loss": 0.9946, "step": 469 }, { "epoch": 1.194300518134715, "grad_norm": 1.0934456586837769, "learning_rate": 3.4001246424384294e-06, "loss": 0.9071, "step": 470 }, { "epoch": 1.1968911917098446, "grad_norm": 0.9213566780090332, "learning_rate": 3.380607937604358e-06, "loss": 0.9489, "step": 471 }, { "epoch": 1.1994818652849741, "grad_norm": 0.8582881689071655, "learning_rate": 3.361118758649116e-06, "loss": 0.9756, "step": 472 }, { "epoch": 1.2020725388601037, "grad_norm": 0.9744678735733032, "learning_rate": 3.341657436843181e-06, "loss": 1.0436, "step": 473 }, { "epoch": 1.2046632124352332, "grad_norm": 0.9584829807281494, "learning_rate": 3.322224302983517e-06, "loss": 1.0492, "step": 474 }, { "epoch": 1.2072538860103628, "grad_norm": 0.8227071166038513, "learning_rate": 3.302819687387967e-06, "loss": 0.9419, "step": 475 }, { "epoch": 1.2098445595854923, "grad_norm": 1.1721601486206055, "learning_rate": 3.2834439198896285e-06, "loss": 1.0105, "step": 476 }, { "epoch": 1.2124352331606219, "grad_norm": 0.96018385887146, "learning_rate": 3.264097329831244e-06, "loss": 1.0335, "step": 477 }, { "epoch": 1.2150259067357512, "grad_norm": 0.8224446177482605, "learning_rate": 3.2447802460596124e-06, "loss": 0.9914, "step": 478 }, { "epoch": 1.2176165803108807, "grad_norm": 0.9198357462882996, "learning_rate": 3.2254929969199933e-06, "loss": 0.957, "step": 479 }, { "epoch": 1.2202072538860103, "grad_norm": 1.0417511463165283, "learning_rate": 3.206235910250529e-06, "loss": 1.1192, "step": 480 }, { "epoch": 1.2227979274611398, "grad_norm": 0.9127503037452698, "learning_rate": 3.1870093133766653e-06, "loss": 0.9585, "step": 481 }, { "epoch": 1.2253886010362693, "grad_norm": 0.8656659722328186, "learning_rate": 3.167813533105598e-06, "loss": 0.9106, "step": 482 }, { "epoch": 1.2279792746113989, "grad_norm": 0.9434940218925476, "learning_rate": 3.148648895720714e-06, "loss": 0.9955, "step": 483 }, { "epoch": 1.2305699481865284, "grad_norm": 0.8094484806060791, "learning_rate": 3.1295157269760347e-06, "loss": 1.007, "step": 484 }, { "epoch": 1.233160621761658, "grad_norm": 1.175736904144287, "learning_rate": 3.1104143520906976e-06, "loss": 0.9983, "step": 485 }, { "epoch": 1.233160621761658, "eval_loss": 0.9479498267173767, "eval_runtime": 99.1167, "eval_samples_per_second": 2.522, "eval_steps_per_second": 0.323, "step": 485 }, { "epoch": 1.2357512953367875, "grad_norm": 0.872238278388977, "learning_rate": 3.0913450957434177e-06, "loss": 1.0085, "step": 486 }, { "epoch": 1.238341968911917, "grad_norm": 0.8121694326400757, "learning_rate": 3.0723082820669634e-06, "loss": 0.9683, "step": 487 }, { "epoch": 1.2409326424870466, "grad_norm": 0.8866750597953796, "learning_rate": 3.0533042346426612e-06, "loss": 0.9324, "step": 488 }, { "epoch": 1.2435233160621761, "grad_norm": 0.8889084458351135, "learning_rate": 3.034333276494884e-06, "loss": 0.9622, "step": 489 }, { "epoch": 1.2461139896373057, "grad_norm": 0.9424455761909485, "learning_rate": 3.015395730085565e-06, "loss": 0.9475, "step": 490 }, { "epoch": 1.2487046632124352, "grad_norm": 0.8178656101226807, "learning_rate": 2.9964919173087154e-06, "loss": 0.9323, "step": 491 }, { "epoch": 1.2512953367875648, "grad_norm": 0.8948029279708862, "learning_rate": 2.9776221594849565e-06, "loss": 0.9929, "step": 492 }, { "epoch": 1.2538860103626943, "grad_norm": 0.8116942048072815, "learning_rate": 2.9587867773560488e-06, "loss": 1.0207, "step": 493 }, { "epoch": 1.2564766839378239, "grad_norm": 0.8729721903800964, "learning_rate": 2.9399860910794532e-06, "loss": 1.0275, "step": 494 }, { "epoch": 1.2590673575129534, "grad_norm": 0.977170467376709, "learning_rate": 2.921220420222878e-06, "loss": 1.0432, "step": 495 }, { "epoch": 1.261658031088083, "grad_norm": 0.8280233144760132, "learning_rate": 2.902490083758856e-06, "loss": 0.9642, "step": 496 }, { "epoch": 1.2642487046632125, "grad_norm": 0.9376110434532166, "learning_rate": 2.8837954000593106e-06, "loss": 0.9649, "step": 497 }, { "epoch": 1.266839378238342, "grad_norm": 0.7053350210189819, "learning_rate": 2.8651366868901543e-06, "loss": 0.9851, "step": 498 }, { "epoch": 1.2694300518134716, "grad_norm": 0.8334206938743591, "learning_rate": 2.8465142614058916e-06, "loss": 0.9808, "step": 499 }, { "epoch": 1.2720207253886011, "grad_norm": 0.8525857329368591, "learning_rate": 2.8279284401442085e-06, "loss": 0.9964, "step": 500 }, { "epoch": 1.2746113989637307, "grad_norm": 0.8159149289131165, "learning_rate": 2.809379539020618e-06, "loss": 1.0053, "step": 501 }, { "epoch": 1.2772020725388602, "grad_norm": 0.9699737429618835, "learning_rate": 2.790867873323067e-06, "loss": 0.9218, "step": 502 }, { "epoch": 1.2797927461139897, "grad_norm": 0.8307886123657227, "learning_rate": 2.7723937577065924e-06, "loss": 0.9667, "step": 503 }, { "epoch": 1.2823834196891193, "grad_norm": 0.9945038557052612, "learning_rate": 2.753957506187964e-06, "loss": 1.0314, "step": 504 }, { "epoch": 1.2849740932642488, "grad_norm": 1.1403707265853882, "learning_rate": 2.735559432140358e-06, "loss": 0.9839, "step": 505 }, { "epoch": 1.2875647668393784, "grad_norm": 1.1950939893722534, "learning_rate": 2.7171998482880093e-06, "loss": 1.0106, "step": 506 }, { "epoch": 1.2901554404145077, "grad_norm": 0.9658251404762268, "learning_rate": 2.6988790667009246e-06, "loss": 0.9977, "step": 507 }, { "epoch": 1.2927461139896372, "grad_norm": 0.8657079935073853, "learning_rate": 2.680597398789554e-06, "loss": 0.9296, "step": 508 }, { "epoch": 1.2953367875647668, "grad_norm": 0.9752938151359558, "learning_rate": 2.6623551552995076e-06, "loss": 1.0082, "step": 509 }, { "epoch": 1.2979274611398963, "grad_norm": 0.9197008609771729, "learning_rate": 2.6441526463062727e-06, "loss": 0.9655, "step": 510 }, { "epoch": 1.3005181347150259, "grad_norm": 1.0991517305374146, "learning_rate": 2.6259901812099432e-06, "loss": 0.8592, "step": 511 }, { "epoch": 1.3031088082901554, "grad_norm": 0.9919476509094238, "learning_rate": 2.607868068729966e-06, "loss": 1.014, "step": 512 }, { "epoch": 1.305699481865285, "grad_norm": 0.8860031366348267, "learning_rate": 2.5897866168998754e-06, "loss": 0.9493, "step": 513 }, { "epoch": 1.3082901554404145, "grad_norm": 0.9551827907562256, "learning_rate": 2.571746133062082e-06, "loss": 0.9784, "step": 514 }, { "epoch": 1.310880829015544, "grad_norm": 0.9395473003387451, "learning_rate": 2.5537469238626296e-06, "loss": 0.9847, "step": 515 }, { "epoch": 1.3134715025906736, "grad_norm": 0.9098738431930542, "learning_rate": 2.5357892952459917e-06, "loss": 0.9757, "step": 516 }, { "epoch": 1.3160621761658031, "grad_norm": 1.0057542324066162, "learning_rate": 2.517873552449869e-06, "loss": 1.0406, "step": 517 }, { "epoch": 1.3186528497409327, "grad_norm": 0.9436793327331543, "learning_rate": 2.5000000000000015e-06, "loss": 0.9948, "step": 518 }, { "epoch": 1.3212435233160622, "grad_norm": 0.9765656590461731, "learning_rate": 2.4821689417049898e-06, "loss": 0.9242, "step": 519 }, { "epoch": 1.3238341968911918, "grad_norm": 0.9807612895965576, "learning_rate": 2.4643806806511344e-06, "loss": 0.979, "step": 520 }, { "epoch": 1.3264248704663213, "grad_norm": 1.051147699356079, "learning_rate": 2.4466355191972886e-06, "loss": 0.905, "step": 521 }, { "epoch": 1.3290155440414508, "grad_norm": 0.975245475769043, "learning_rate": 2.4289337589697e-06, "loss": 1.0436, "step": 522 }, { "epoch": 1.3316062176165804, "grad_norm": 1.053765892982483, "learning_rate": 2.4112757008569102e-06, "loss": 1.0155, "step": 523 }, { "epoch": 1.33419689119171, "grad_norm": 0.7797523736953735, "learning_rate": 2.3936616450046207e-06, "loss": 1.0052, "step": 524 }, { "epoch": 1.3367875647668392, "grad_norm": 0.8835865259170532, "learning_rate": 2.376091890810598e-06, "loss": 0.9177, "step": 525 }, { "epoch": 1.3393782383419688, "grad_norm": 0.8106747269630432, "learning_rate": 2.3585667369195815e-06, "loss": 0.9525, "step": 526 }, { "epoch": 1.3419689119170983, "grad_norm": 1.0181241035461426, "learning_rate": 2.341086481218217e-06, "loss": 0.9024, "step": 527 }, { "epoch": 1.3445595854922279, "grad_norm": 0.9752720594406128, "learning_rate": 2.32365142082998e-06, "loss": 1.0024, "step": 528 }, { "epoch": 1.3471502590673574, "grad_norm": 0.8057308793067932, "learning_rate": 2.306261852110132e-06, "loss": 0.9294, "step": 529 }, { "epoch": 1.349740932642487, "grad_norm": 1.02779221534729, "learning_rate": 2.288918070640684e-06, "loss": 0.9539, "step": 530 }, { "epoch": 1.3523316062176165, "grad_norm": 0.8755698800086975, "learning_rate": 2.2716203712253708e-06, "loss": 1.0245, "step": 531 }, { "epoch": 1.354922279792746, "grad_norm": 0.8415625691413879, "learning_rate": 2.254369047884639e-06, "loss": 0.9994, "step": 532 }, { "epoch": 1.3575129533678756, "grad_norm": 1.01762056350708, "learning_rate": 2.2371643938506488e-06, "loss": 0.9763, "step": 533 }, { "epoch": 1.3601036269430051, "grad_norm": 1.0196937322616577, "learning_rate": 2.2200067015622986e-06, "loss": 0.9745, "step": 534 }, { "epoch": 1.3626943005181347, "grad_norm": 1.0955042839050293, "learning_rate": 2.2028962626602346e-06, "loss": 0.9457, "step": 535 }, { "epoch": 1.3652849740932642, "grad_norm": 0.8075233697891235, "learning_rate": 2.185833367981918e-06, "loss": 0.9042, "step": 536 }, { "epoch": 1.3678756476683938, "grad_norm": 0.8643060922622681, "learning_rate": 2.168818307556663e-06, "loss": 0.9541, "step": 537 }, { "epoch": 1.3704663212435233, "grad_norm": 0.7847241759300232, "learning_rate": 2.1518513706007154e-06, "loss": 0.9874, "step": 538 }, { "epoch": 1.3730569948186528, "grad_norm": 0.8477650880813599, "learning_rate": 2.13493284551233e-06, "loss": 1.0541, "step": 539 }, { "epoch": 1.3756476683937824, "grad_norm": 0.993405818939209, "learning_rate": 2.118063019866884e-06, "loss": 1.0024, "step": 540 }, { "epoch": 1.378238341968912, "grad_norm": 0.9656562805175781, "learning_rate": 2.101242180411963e-06, "loss": 0.9929, "step": 541 }, { "epoch": 1.3808290155440415, "grad_norm": 0.88680100440979, "learning_rate": 2.0844706130625146e-06, "loss": 1.0225, "step": 542 }, { "epoch": 1.383419689119171, "grad_norm": 0.9172679781913757, "learning_rate": 2.067748602895969e-06, "loss": 0.9504, "step": 543 }, { "epoch": 1.3860103626943006, "grad_norm": 0.9063828587532043, "learning_rate": 2.0510764341474032e-06, "loss": 0.9287, "step": 544 }, { "epoch": 1.38860103626943, "grad_norm": 0.9309386014938354, "learning_rate": 2.0344543902047043e-06, "loss": 1.015, "step": 545 }, { "epoch": 1.3911917098445596, "grad_norm": 0.9050936102867126, "learning_rate": 2.0178827536037547e-06, "loss": 0.9952, "step": 546 }, { "epoch": 1.3937823834196892, "grad_norm": 1.3205598592758179, "learning_rate": 2.001361806023636e-06, "loss": 0.911, "step": 547 }, { "epoch": 1.3963730569948187, "grad_norm": 0.8794770240783691, "learning_rate": 1.9848918282818242e-06, "loss": 0.9863, "step": 548 }, { "epoch": 1.3989637305699483, "grad_norm": 0.9719308614730835, "learning_rate": 1.968473100329437e-06, "loss": 0.9862, "step": 549 }, { "epoch": 1.4015544041450778, "grad_norm": 0.8657988905906677, "learning_rate": 1.952105901246461e-06, "loss": 1.0129, "step": 550 }, { "epoch": 1.4041450777202074, "grad_norm": 1.1228150129318237, "learning_rate": 1.935790509237013e-06, "loss": 0.9855, "step": 551 }, { "epoch": 1.406735751295337, "grad_norm": 0.838809072971344, "learning_rate": 1.9195272016246105e-06, "loss": 0.9796, "step": 552 }, { "epoch": 1.4093264248704664, "grad_norm": 0.9168699383735657, "learning_rate": 1.9033162548474577e-06, "loss": 0.9694, "step": 553 }, { "epoch": 1.411917098445596, "grad_norm": 0.9007121324539185, "learning_rate": 1.887157944453749e-06, "loss": 0.9737, "step": 554 }, { "epoch": 1.4145077720207253, "grad_norm": 0.8796411156654358, "learning_rate": 1.8710525450969803e-06, "loss": 0.9293, "step": 555 }, { "epoch": 1.4170984455958548, "grad_norm": 0.9988073110580444, "learning_rate": 1.855000330531289e-06, "loss": 0.9632, "step": 556 }, { "epoch": 1.4196891191709844, "grad_norm": 0.9262337684631348, "learning_rate": 1.8390015736067869e-06, "loss": 0.9679, "step": 557 }, { "epoch": 1.422279792746114, "grad_norm": 0.9995079040527344, "learning_rate": 1.823056546264939e-06, "loss": 0.9558, "step": 558 }, { "epoch": 1.4248704663212435, "grad_norm": 0.855469286441803, "learning_rate": 1.8071655195339272e-06, "loss": 1.0535, "step": 559 }, { "epoch": 1.427461139896373, "grad_norm": 0.760111927986145, "learning_rate": 1.7913287635240573e-06, "loss": 0.9714, "step": 560 }, { "epoch": 1.4300518134715026, "grad_norm": 0.9862337708473206, "learning_rate": 1.7755465474231465e-06, "loss": 1.0004, "step": 561 }, { "epoch": 1.432642487046632, "grad_norm": 0.7655313611030579, "learning_rate": 1.7598191394919738e-06, "loss": 0.9812, "step": 562 }, { "epoch": 1.4352331606217616, "grad_norm": 1.0747240781784058, "learning_rate": 1.7441468070597017e-06, "loss": 1.0208, "step": 563 }, { "epoch": 1.4378238341968912, "grad_norm": 0.9578352570533752, "learning_rate": 1.7285298165193388e-06, "loss": 0.9637, "step": 564 }, { "epoch": 1.4404145077720207, "grad_norm": 0.9221131801605225, "learning_rate": 1.7129684333232095e-06, "loss": 0.952, "step": 565 }, { "epoch": 1.4430051813471503, "grad_norm": 1.1166445016860962, "learning_rate": 1.697462921978446e-06, "loss": 1.0394, "step": 566 }, { "epoch": 1.4455958549222798, "grad_norm": 0.8609073758125305, "learning_rate": 1.682013546042488e-06, "loss": 1.0053, "step": 567 }, { "epoch": 1.4481865284974094, "grad_norm": 0.8035356998443604, "learning_rate": 1.6666205681186032e-06, "loss": 1.0051, "step": 568 }, { "epoch": 1.450777202072539, "grad_norm": 0.8815962076187134, "learning_rate": 1.6512842498514315e-06, "loss": 0.9821, "step": 569 }, { "epoch": 1.4533678756476685, "grad_norm": 0.9989993572235107, "learning_rate": 1.6360048519225197e-06, "loss": 0.9804, "step": 570 }, { "epoch": 1.455958549222798, "grad_norm": 0.9338482618331909, "learning_rate": 1.6207826340459131e-06, "loss": 0.9514, "step": 571 }, { "epoch": 1.4585492227979275, "grad_norm": 0.7817990183830261, "learning_rate": 1.6056178549637248e-06, "loss": 1.0398, "step": 572 }, { "epoch": 1.4611398963730569, "grad_norm": 0.9715204834938049, "learning_rate": 1.5905107724417412e-06, "loss": 0.9859, "step": 573 }, { "epoch": 1.4637305699481864, "grad_norm": 0.8073526620864868, "learning_rate": 1.5754616432650443e-06, "loss": 0.9364, "step": 574 }, { "epoch": 1.466321243523316, "grad_norm": 0.908467710018158, "learning_rate": 1.5604707232336457e-06, "loss": 0.97, "step": 575 }, { "epoch": 1.4689119170984455, "grad_norm": 0.7307873964309692, "learning_rate": 1.5455382671581365e-06, "loss": 0.9081, "step": 576 }, { "epoch": 1.471502590673575, "grad_norm": 0.9267300367355347, "learning_rate": 1.5306645288553556e-06, "loss": 0.9719, "step": 577 }, { "epoch": 1.4740932642487046, "grad_norm": 1.200758695602417, "learning_rate": 1.5158497611440792e-06, "loss": 1.0408, "step": 578 }, { "epoch": 1.4766839378238341, "grad_norm": 0.8580562472343445, "learning_rate": 1.5010942158407204e-06, "loss": 0.8879, "step": 579 }, { "epoch": 1.4792746113989637, "grad_norm": 0.7956061959266663, "learning_rate": 1.48639814375505e-06, "loss": 1.0052, "step": 580 }, { "epoch": 1.4818652849740932, "grad_norm": 1.047351360321045, "learning_rate": 1.4717617946859319e-06, "loss": 1.0048, "step": 581 }, { "epoch": 1.4844559585492227, "grad_norm": 0.9025893211364746, "learning_rate": 1.4571854174170847e-06, "loss": 0.9976, "step": 582 }, { "epoch": 1.4844559585492227, "eval_loss": 0.9445139765739441, "eval_runtime": 99.1624, "eval_samples_per_second": 2.521, "eval_steps_per_second": 0.323, "step": 582 }, { "epoch": 1.4870466321243523, "grad_norm": 0.8686254620552063, "learning_rate": 1.4426692597128339e-06, "loss": 0.9141, "step": 583 }, { "epoch": 1.4896373056994818, "grad_norm": 0.8144710659980774, "learning_rate": 1.428213568313927e-06, "loss": 0.9429, "step": 584 }, { "epoch": 1.4922279792746114, "grad_norm": 0.8153760433197021, "learning_rate": 1.4138185889333172e-06, "loss": 0.9564, "step": 585 }, { "epoch": 1.494818652849741, "grad_norm": 0.8345690369606018, "learning_rate": 1.3994845662519985e-06, "loss": 0.9794, "step": 586 }, { "epoch": 1.4974093264248705, "grad_norm": 1.0408412218093872, "learning_rate": 1.3852117439148416e-06, "loss": 0.9961, "step": 587 }, { "epoch": 1.5, "grad_norm": 0.9831898212432861, "learning_rate": 1.3710003645264559e-06, "loss": 1.0319, "step": 588 }, { "epoch": 1.5025906735751295, "grad_norm": 0.9798739552497864, "learning_rate": 1.3568506696470645e-06, "loss": 1.0054, "step": 589 }, { "epoch": 1.505181347150259, "grad_norm": 0.741041898727417, "learning_rate": 1.3427628997883957e-06, "loss": 0.9952, "step": 590 }, { "epoch": 1.5077720207253886, "grad_norm": 0.7825758457183838, "learning_rate": 1.3287372944096021e-06, "loss": 0.993, "step": 591 }, { "epoch": 1.5103626943005182, "grad_norm": 0.8112067580223083, "learning_rate": 1.3147740919131814e-06, "loss": 1.0047, "step": 592 }, { "epoch": 1.5129533678756477, "grad_norm": 0.8506227135658264, "learning_rate": 1.3008735296409281e-06, "loss": 0.9148, "step": 593 }, { "epoch": 1.5155440414507773, "grad_norm": 0.8730190396308899, "learning_rate": 1.2870358438699005e-06, "loss": 0.9805, "step": 594 }, { "epoch": 1.5181347150259068, "grad_norm": 0.8171835541725159, "learning_rate": 1.2732612698084067e-06, "loss": 0.9846, "step": 595 }, { "epoch": 1.5207253886010363, "grad_norm": 0.8341667652130127, "learning_rate": 1.2595500415919948e-06, "loss": 0.9912, "step": 596 }, { "epoch": 1.5233160621761659, "grad_norm": 0.8373621106147766, "learning_rate": 1.245902392279491e-06, "loss": 1.0312, "step": 597 }, { "epoch": 1.5259067357512954, "grad_norm": 1.0364506244659424, "learning_rate": 1.232318553849023e-06, "loss": 0.9417, "step": 598 }, { "epoch": 1.528497409326425, "grad_norm": 0.8765987157821655, "learning_rate": 1.2187987571940846e-06, "loss": 1.0039, "step": 599 }, { "epoch": 1.5310880829015545, "grad_norm": 0.8896673917770386, "learning_rate": 1.2053432321196085e-06, "loss": 0.9867, "step": 600 }, { "epoch": 1.533678756476684, "grad_norm": 0.7940327525138855, "learning_rate": 1.1919522073380614e-06, "loss": 1.0547, "step": 601 }, { "epoch": 1.5362694300518136, "grad_norm": 0.903992772102356, "learning_rate": 1.1786259104655562e-06, "loss": 0.9451, "step": 602 }, { "epoch": 1.5388601036269431, "grad_norm": 0.8643820285797119, "learning_rate": 1.1653645680179792e-06, "loss": 0.8786, "step": 603 }, { "epoch": 1.5414507772020727, "grad_norm": 0.7279991507530212, "learning_rate": 1.1521684054071524e-06, "loss": 0.9964, "step": 604 }, { "epoch": 1.5440414507772022, "grad_norm": 0.8498103022575378, "learning_rate": 1.1390376469369796e-06, "loss": 0.9547, "step": 605 }, { "epoch": 1.5466321243523318, "grad_norm": 0.800202488899231, "learning_rate": 1.1259725157996593e-06, "loss": 0.9599, "step": 606 }, { "epoch": 1.549222797927461, "grad_norm": 0.9869511127471924, "learning_rate": 1.1129732340718702e-06, "loss": 1.0126, "step": 607 }, { "epoch": 1.5518134715025906, "grad_norm": 0.9484280347824097, "learning_rate": 1.1000400227110142e-06, "loss": 0.9711, "step": 608 }, { "epoch": 1.5544041450777202, "grad_norm": 0.9410507082939148, "learning_rate": 1.087173101551439e-06, "loss": 0.9845, "step": 609 }, { "epoch": 1.5569948186528497, "grad_norm": 0.7972487807273865, "learning_rate": 1.0743726893007257e-06, "loss": 0.9653, "step": 610 }, { "epoch": 1.5595854922279793, "grad_norm": 0.9086237549781799, "learning_rate": 1.061639003535952e-06, "loss": 1.002, "step": 611 }, { "epoch": 1.5621761658031088, "grad_norm": 0.9138615131378174, "learning_rate": 1.0489722607000052e-06, "loss": 0.9809, "step": 612 }, { "epoch": 1.5647668393782384, "grad_norm": 0.9875264167785645, "learning_rate": 1.036372676097897e-06, "loss": 0.9163, "step": 613 }, { "epoch": 1.567357512953368, "grad_norm": 0.9866282939910889, "learning_rate": 1.0238404638931077e-06, "loss": 1.067, "step": 614 }, { "epoch": 1.5699481865284974, "grad_norm": 0.9424428939819336, "learning_rate": 1.0113758371039429e-06, "loss": 0.9418, "step": 615 }, { "epoch": 1.572538860103627, "grad_norm": 0.9368731379508972, "learning_rate": 9.989790075999145e-07, "loss": 0.9741, "step": 616 }, { "epoch": 1.5751295336787565, "grad_norm": 0.8245809674263, "learning_rate": 9.866501860981431e-07, "loss": 0.9837, "step": 617 }, { "epoch": 1.577720207253886, "grad_norm": 0.964534342288971, "learning_rate": 9.743895821597638e-07, "loss": 0.961, "step": 618 }, { "epoch": 1.5803108808290154, "grad_norm": 0.9792007803916931, "learning_rate": 9.621974041863813e-07, "loss": 0.898, "step": 619 }, { "epoch": 1.582901554404145, "grad_norm": 0.8622314929962158, "learning_rate": 9.500738594165132e-07, "loss": 1.0085, "step": 620 }, { "epoch": 1.5854922279792745, "grad_norm": 0.936686098575592, "learning_rate": 9.380191539220762e-07, "loss": 0.9979, "step": 621 }, { "epoch": 1.588082901554404, "grad_norm": 1.0908080339431763, "learning_rate": 9.260334926048787e-07, "loss": 0.9885, "step": 622 }, { "epoch": 1.5906735751295336, "grad_norm": 0.8463078737258911, "learning_rate": 9.141170791931386e-07, "loss": 0.9314, "step": 623 }, { "epoch": 1.593264248704663, "grad_norm": 1.073156714439392, "learning_rate": 9.022701162380259e-07, "loss": 1.0001, "step": 624 }, { "epoch": 1.5958549222797926, "grad_norm": 0.8691221475601196, "learning_rate": 8.904928051102074e-07, "loss": 0.9589, "step": 625 }, { "epoch": 1.5984455958549222, "grad_norm": 0.9988768696784973, "learning_rate": 8.787853459964407e-07, "loss": 1.0873, "step": 626 }, { "epoch": 1.6010362694300517, "grad_norm": 1.0085712671279907, "learning_rate": 8.671479378961556e-07, "loss": 0.9401, "step": 627 }, { "epoch": 1.6036269430051813, "grad_norm": 1.0235176086425781, "learning_rate": 8.555807786180814e-07, "loss": 1.0291, "step": 628 }, { "epoch": 1.6062176165803108, "grad_norm": 0.9803045988082886, "learning_rate": 8.4408406477688e-07, "loss": 0.9434, "step": 629 }, { "epoch": 1.6088082901554404, "grad_norm": 0.9053346514701843, "learning_rate": 8.326579917898098e-07, "loss": 0.9941, "step": 630 }, { "epoch": 1.61139896373057, "grad_norm": 0.9022249579429626, "learning_rate": 8.21302753873393e-07, "loss": 0.947, "step": 631 }, { "epoch": 1.6139896373056994, "grad_norm": 1.0404874086380005, "learning_rate": 8.100185440401276e-07, "loss": 1.0749, "step": 632 }, { "epoch": 1.616580310880829, "grad_norm": 0.9517078399658203, "learning_rate": 7.988055540951967e-07, "loss": 0.9732, "step": 633 }, { "epoch": 1.6191709844559585, "grad_norm": 0.9991124868392944, "learning_rate": 7.876639746332132e-07, "loss": 1.0463, "step": 634 }, { "epoch": 1.621761658031088, "grad_norm": 0.9260480403900146, "learning_rate": 7.765939950349776e-07, "loss": 0.9353, "step": 635 }, { "epoch": 1.6243523316062176, "grad_norm": 0.9367355704307556, "learning_rate": 7.655958034642619e-07, "loss": 0.9758, "step": 636 }, { "epoch": 1.6269430051813472, "grad_norm": 0.777481734752655, "learning_rate": 7.54669586864607e-07, "loss": 0.927, "step": 637 }, { "epoch": 1.6295336787564767, "grad_norm": 0.9329550266265869, "learning_rate": 7.43815530956149e-07, "loss": 0.9997, "step": 638 }, { "epoch": 1.6321243523316062, "grad_norm": 0.9558778405189514, "learning_rate": 7.330338202324621e-07, "loss": 0.9486, "step": 639 }, { "epoch": 1.6347150259067358, "grad_norm": 0.9044818878173828, "learning_rate": 7.223246379574206e-07, "loss": 0.9678, "step": 640 }, { "epoch": 1.6373056994818653, "grad_norm": 1.0184038877487183, "learning_rate": 7.116881661620833e-07, "loss": 0.9467, "step": 641 }, { "epoch": 1.6398963730569949, "grad_norm": 0.82332843542099, "learning_rate": 7.011245856416016e-07, "loss": 1.0119, "step": 642 }, { "epoch": 1.6424870466321244, "grad_norm": 1.0882319211959839, "learning_rate": 6.90634075952149e-07, "loss": 1.0505, "step": 643 }, { "epoch": 1.645077720207254, "grad_norm": 0.8696717619895935, "learning_rate": 6.802168154078586e-07, "loss": 1.0155, "step": 644 }, { "epoch": 1.6476683937823835, "grad_norm": 0.8515833616256714, "learning_rate": 6.698729810778065e-07, "loss": 1.0446, "step": 645 }, { "epoch": 1.650259067357513, "grad_norm": 0.9765702486038208, "learning_rate": 6.596027487829915e-07, "loss": 0.9914, "step": 646 }, { "epoch": 1.6528497409326426, "grad_norm": 0.96961510181427, "learning_rate": 6.494062930933497e-07, "loss": 0.9454, "step": 647 }, { "epoch": 1.6554404145077721, "grad_norm": 0.9544724822044373, "learning_rate": 6.392837873247876e-07, "loss": 0.953, "step": 648 }, { "epoch": 1.6580310880829017, "grad_norm": 1.0494675636291504, "learning_rate": 6.292354035362369e-07, "loss": 1.0195, "step": 649 }, { "epoch": 1.6606217616580312, "grad_norm": 1.032488465309143, "learning_rate": 6.192613125267283e-07, "loss": 1.0027, "step": 650 }, { "epoch": 1.6632124352331608, "grad_norm": 1.010940432548523, "learning_rate": 6.093616838324872e-07, "loss": 0.9949, "step": 651 }, { "epoch": 1.6658031088082903, "grad_norm": 0.7320314645767212, "learning_rate": 5.995366857240592e-07, "loss": 0.9881, "step": 652 }, { "epoch": 1.6683937823834198, "grad_norm": 1.1827306747436523, "learning_rate": 5.897864852034368e-07, "loss": 0.9126, "step": 653 }, { "epoch": 1.6709844559585494, "grad_norm": 0.8718465566635132, "learning_rate": 5.801112480012344e-07, "loss": 0.8972, "step": 654 }, { "epoch": 1.6735751295336787, "grad_norm": 0.8429136276245117, "learning_rate": 5.705111385738638e-07, "loss": 1.0019, "step": 655 }, { "epoch": 1.6761658031088082, "grad_norm": 0.9650006294250488, "learning_rate": 5.609863201007382e-07, "loss": 0.992, "step": 656 }, { "epoch": 1.6787564766839378, "grad_norm": 1.091964840888977, "learning_rate": 5.515369544815025e-07, "loss": 1.0321, "step": 657 }, { "epoch": 1.6813471502590673, "grad_norm": 0.7537882924079895, "learning_rate": 5.421632023332779e-07, "loss": 0.9601, "step": 658 }, { "epoch": 1.6839378238341969, "grad_norm": 0.9340682029724121, "learning_rate": 5.328652229879383e-07, "loss": 0.9789, "step": 659 }, { "epoch": 1.6865284974093264, "grad_norm": 0.7913133502006531, "learning_rate": 5.236431744893883e-07, "loss": 0.9062, "step": 660 }, { "epoch": 1.689119170984456, "grad_norm": 0.9730626344680786, "learning_rate": 5.144972135908949e-07, "loss": 0.938, "step": 661 }, { "epoch": 1.6917098445595855, "grad_norm": 0.8744672536849976, "learning_rate": 5.054274957524075e-07, "loss": 1.0014, "step": 662 }, { "epoch": 1.694300518134715, "grad_norm": 0.9841236472129822, "learning_rate": 4.964341751379248e-07, "loss": 0.9094, "step": 663 }, { "epoch": 1.6968911917098446, "grad_norm": 1.0411112308502197, "learning_rate": 4.875174046128684e-07, "loss": 1.011, "step": 664 }, { "epoch": 1.6994818652849741, "grad_norm": 0.9232543706893921, "learning_rate": 4.786773357414926e-07, "loss": 0.9188, "step": 665 }, { "epoch": 1.7020725388601037, "grad_norm": 0.9863471984863281, "learning_rate": 4.6991411878429593e-07, "loss": 0.9577, "step": 666 }, { "epoch": 1.704663212435233, "grad_norm": 0.8515871167182922, "learning_rate": 4.612279026954808e-07, "loss": 0.9712, "step": 667 }, { "epoch": 1.7072538860103625, "grad_norm": 1.0452429056167603, "learning_rate": 4.526188351204103e-07, "loss": 0.9525, "step": 668 }, { "epoch": 1.709844559585492, "grad_norm": 1.039813756942749, "learning_rate": 4.440870623931054e-07, "loss": 0.9677, "step": 669 }, { "epoch": 1.7124352331606216, "grad_norm": 1.0061904191970825, "learning_rate": 4.3563272953375426e-07, "loss": 0.9249, "step": 670 }, { "epoch": 1.7150259067357512, "grad_norm": 0.9659146666526794, "learning_rate": 4.2725598024624783e-07, "loss": 0.932, "step": 671 }, { "epoch": 1.7176165803108807, "grad_norm": 1.100063443183899, "learning_rate": 4.1895695691574146e-07, "loss": 0.9303, "step": 672 }, { "epoch": 1.7202072538860103, "grad_norm": 0.9187989830970764, "learning_rate": 4.1073580060622455e-07, "loss": 0.984, "step": 673 }, { "epoch": 1.7227979274611398, "grad_norm": 0.8058050870895386, "learning_rate": 4.025926510581357e-07, "loss": 0.9469, "step": 674 }, { "epoch": 1.7253886010362693, "grad_norm": 0.9303697347640991, "learning_rate": 3.9452764668597764e-07, "loss": 1.0155, "step": 675 }, { "epoch": 1.7279792746113989, "grad_norm": 0.875150740146637, "learning_rate": 3.8654092457596714e-07, "loss": 0.9498, "step": 676 }, { "epoch": 1.7305699481865284, "grad_norm": 1.206532597541809, "learning_rate": 3.786326204837065e-07, "loss": 1.0035, "step": 677 }, { "epoch": 1.733160621761658, "grad_norm": 0.9007678627967834, "learning_rate": 3.7080286883187713e-07, "loss": 0.9713, "step": 678 }, { "epoch": 1.7357512953367875, "grad_norm": 1.0224071741104126, "learning_rate": 3.6305180270794827e-07, "loss": 0.9525, "step": 679 }, { "epoch": 1.7357512953367875, "eval_loss": 0.9424797892570496, "eval_runtime": 98.9403, "eval_samples_per_second": 2.527, "eval_steps_per_second": 0.323, "step": 679 }, { "epoch": 1.738341968911917, "grad_norm": 0.8955622315406799, "learning_rate": 3.553795538619237e-07, "loss": 0.9495, "step": 680 }, { "epoch": 1.7409326424870466, "grad_norm": 0.9500125050544739, "learning_rate": 3.4778625270409484e-07, "loss": 0.9822, "step": 681 }, { "epoch": 1.7435233160621761, "grad_norm": 0.9673233032226562, "learning_rate": 3.402720283028277e-07, "loss": 0.9392, "step": 682 }, { "epoch": 1.7461139896373057, "grad_norm": 1.0081723928451538, "learning_rate": 3.328370083823679e-07, "loss": 0.9768, "step": 683 }, { "epoch": 1.7487046632124352, "grad_norm": 1.021333932876587, "learning_rate": 3.2548131932067184e-07, "loss": 0.9438, "step": 684 }, { "epoch": 1.7512953367875648, "grad_norm": 0.8333784341812134, "learning_rate": 3.182050861472541e-07, "loss": 1.0067, "step": 685 }, { "epoch": 1.7538860103626943, "grad_norm": 1.0099414587020874, "learning_rate": 3.110084325410667e-07, "loss": 1.0327, "step": 686 }, { "epoch": 1.7564766839378239, "grad_norm": 1.2605355978012085, "learning_rate": 3.03891480828396e-07, "loss": 0.9497, "step": 687 }, { "epoch": 1.7590673575129534, "grad_norm": 0.9966846704483032, "learning_rate": 2.9685435198078095e-07, "loss": 1.0268, "step": 688 }, { "epoch": 1.761658031088083, "grad_norm": 0.920387327671051, "learning_rate": 2.898971656129573e-07, "loss": 0.9526, "step": 689 }, { "epoch": 1.7642487046632125, "grad_norm": 0.9313926100730896, "learning_rate": 2.830200399808286e-07, "loss": 0.9874, "step": 690 }, { "epoch": 1.766839378238342, "grad_norm": 0.9481755495071411, "learning_rate": 2.762230919794506e-07, "loss": 0.9602, "step": 691 }, { "epoch": 1.7694300518134716, "grad_norm": 0.9870350360870361, "learning_rate": 2.6950643714104774e-07, "loss": 0.9758, "step": 692 }, { "epoch": 1.7720207253886011, "grad_norm": 0.8995896577835083, "learning_rate": 2.628701896330482e-07, "loss": 0.9692, "step": 693 }, { "epoch": 1.7746113989637307, "grad_norm": 0.8859801888465881, "learning_rate": 2.563144622561453e-07, "loss": 0.978, "step": 694 }, { "epoch": 1.7772020725388602, "grad_norm": 0.8300227522850037, "learning_rate": 2.498393664423737e-07, "loss": 0.9199, "step": 695 }, { "epoch": 1.7797927461139897, "grad_norm": 0.9782614707946777, "learning_rate": 2.4344501225322557e-07, "loss": 0.9437, "step": 696 }, { "epoch": 1.7823834196891193, "grad_norm": 0.9227774739265442, "learning_rate": 2.3713150837777142e-07, "loss": 0.9153, "step": 697 }, { "epoch": 1.7849740932642488, "grad_norm": 0.854295551776886, "learning_rate": 2.3089896213081553e-07, "loss": 0.8942, "step": 698 }, { "epoch": 1.7875647668393784, "grad_norm": 0.761452853679657, "learning_rate": 2.2474747945107068e-07, "loss": 1.0242, "step": 699 }, { "epoch": 1.790155440414508, "grad_norm": 0.7350905537605286, "learning_rate": 2.1867716489936297e-07, "loss": 1.0056, "step": 700 }, { "epoch": 1.7927461139896375, "grad_norm": 0.9781407117843628, "learning_rate": 2.1268812165684382e-07, "loss": 0.9554, "step": 701 }, { "epoch": 1.795336787564767, "grad_norm": 0.7397287487983704, "learning_rate": 2.0678045152324798e-07, "loss": 1.0214, "step": 702 }, { "epoch": 1.7979274611398963, "grad_norm": 0.9220542311668396, "learning_rate": 2.0095425491515386e-07, "loss": 0.9359, "step": 703 }, { "epoch": 1.8005181347150259, "grad_norm": 0.8612248301506042, "learning_rate": 1.9520963086428258e-07, "loss": 0.9994, "step": 704 }, { "epoch": 1.8031088082901554, "grad_norm": 0.8949776887893677, "learning_rate": 1.8954667701581108e-07, "loss": 0.9721, "step": 705 }, { "epoch": 1.805699481865285, "grad_norm": 1.0004165172576904, "learning_rate": 1.8396548962671456e-07, "loss": 0.9811, "step": 706 }, { "epoch": 1.8082901554404145, "grad_norm": 0.8982537388801575, "learning_rate": 1.7846616356413105e-07, "loss": 0.9457, "step": 707 }, { "epoch": 1.810880829015544, "grad_norm": 0.9139662981033325, "learning_rate": 1.7304879230374328e-07, "loss": 1.0623, "step": 708 }, { "epoch": 1.8134715025906736, "grad_norm": 0.9726191163063049, "learning_rate": 1.677134679281983e-07, "loss": 1.0008, "step": 709 }, { "epoch": 1.8160621761658031, "grad_norm": 0.9855490326881409, "learning_rate": 1.6246028112553603e-07, "loss": 1.0302, "step": 710 }, { "epoch": 1.8186528497409327, "grad_norm": 0.9363789558410645, "learning_rate": 1.5728932118764916e-07, "loss": 0.9876, "step": 711 }, { "epoch": 1.8212435233160622, "grad_norm": 1.0399082899093628, "learning_rate": 1.5220067600876686e-07, "loss": 0.8963, "step": 712 }, { "epoch": 1.8238341968911918, "grad_norm": 0.8987425565719604, "learning_rate": 1.4719443208396078e-07, "loss": 0.9535, "step": 713 }, { "epoch": 1.8264248704663213, "grad_norm": 0.8786448240280151, "learning_rate": 1.422706745076713e-07, "loss": 1.0302, "step": 714 }, { "epoch": 1.8290155440414506, "grad_norm": 1.0369514226913452, "learning_rate": 1.3742948697226533e-07, "loss": 0.9518, "step": 715 }, { "epoch": 1.8316062176165802, "grad_norm": 0.9875585436820984, "learning_rate": 1.3267095176661304e-07, "loss": 0.9448, "step": 716 }, { "epoch": 1.8341968911917097, "grad_norm": 1.0473564863204956, "learning_rate": 1.2799514977468618e-07, "loss": 0.9618, "step": 717 }, { "epoch": 1.8367875647668392, "grad_norm": 1.106512427330017, "learning_rate": 1.2340216047418697e-07, "loss": 1.02, "step": 718 }, { "epoch": 1.8393782383419688, "grad_norm": 0.8493651747703552, "learning_rate": 1.188920619351941e-07, "loss": 0.9718, "step": 719 }, { "epoch": 1.8419689119170983, "grad_norm": 0.8997021317481995, "learning_rate": 1.1446493081883891e-07, "loss": 0.9508, "step": 720 }, { "epoch": 1.8445595854922279, "grad_norm": 0.8545548319816589, "learning_rate": 1.1012084237599808e-07, "loss": 1.0185, "step": 721 }, { "epoch": 1.8471502590673574, "grad_norm": 0.8686257600784302, "learning_rate": 1.0585987044602009e-07, "loss": 0.9757, "step": 722 }, { "epoch": 1.849740932642487, "grad_norm": 0.8913059234619141, "learning_rate": 1.0168208745546416e-07, "loss": 0.9635, "step": 723 }, { "epoch": 1.8523316062176165, "grad_norm": 0.8280147910118103, "learning_rate": 9.758756441687333e-08, "loss": 1.0558, "step": 724 }, { "epoch": 1.854922279792746, "grad_norm": 0.9251644611358643, "learning_rate": 9.357637092756667e-08, "loss": 0.9797, "step": 725 }, { "epoch": 1.8575129533678756, "grad_norm": 0.8681091666221619, "learning_rate": 8.964857516845449e-08, "loss": 0.98, "step": 726 }, { "epoch": 1.8601036269430051, "grad_norm": 0.858220636844635, "learning_rate": 8.580424390288167e-08, "loss": 0.9679, "step": 727 }, { "epoch": 1.8626943005181347, "grad_norm": 1.0083863735198975, "learning_rate": 8.204344247549067e-08, "loss": 0.9765, "step": 728 }, { "epoch": 1.8652849740932642, "grad_norm": 0.7215631604194641, "learning_rate": 7.836623481111416e-08, "loss": 0.9826, "step": 729 }, { "epoch": 1.8678756476683938, "grad_norm": 0.8494441509246826, "learning_rate": 7.47726834136836e-08, "loss": 0.9878, "step": 730 }, { "epoch": 1.8704663212435233, "grad_norm": 0.9821386337280273, "learning_rate": 7.126284936517125e-08, "loss": 0.98, "step": 731 }, { "epoch": 1.8730569948186528, "grad_norm": 0.887163519859314, "learning_rate": 6.783679232455043e-08, "loss": 0.9275, "step": 732 }, { "epoch": 1.8756476683937824, "grad_norm": 0.9071056246757507, "learning_rate": 6.449457052677965e-08, "loss": 0.9906, "step": 733 }, { "epoch": 1.878238341968912, "grad_norm": 0.9382151961326599, "learning_rate": 6.123624078181512e-08, "loss": 1.0063, "step": 734 }, { "epoch": 1.8808290155440415, "grad_norm": 0.8040828704833984, "learning_rate": 5.8061858473645315e-08, "loss": 0.9858, "step": 735 }, { "epoch": 1.883419689119171, "grad_norm": 0.9526659250259399, "learning_rate": 5.4971477559346286e-08, "loss": 0.9893, "step": 736 }, { "epoch": 1.8860103626943006, "grad_norm": 0.8950933218002319, "learning_rate": 5.196515056816898e-08, "loss": 1.0216, "step": 737 }, { "epoch": 1.88860103626943, "grad_norm": 0.8529356718063354, "learning_rate": 4.90429286006433e-08, "loss": 1.0495, "step": 738 }, { "epoch": 1.8911917098445596, "grad_norm": 0.8746633529663086, "learning_rate": 4.620486132770996e-08, "loss": 0.9701, "step": 739 }, { "epoch": 1.8937823834196892, "grad_norm": 0.7848332524299622, "learning_rate": 4.34509969898772e-08, "loss": 0.9551, "step": 740 }, { "epoch": 1.8963730569948187, "grad_norm": 0.844379186630249, "learning_rate": 4.078138239639984e-08, "loss": 0.9392, "step": 741 }, { "epoch": 1.8989637305699483, "grad_norm": 0.9659252762794495, "learning_rate": 3.819606292448541e-08, "loss": 0.9676, "step": 742 }, { "epoch": 1.9015544041450778, "grad_norm": 1.0913461446762085, "learning_rate": 3.569508251851816e-08, "loss": 1.0477, "step": 743 }, { "epoch": 1.9041450777202074, "grad_norm": 1.0511360168457031, "learning_rate": 3.327848368931907e-08, "loss": 0.9345, "step": 744 }, { "epoch": 1.906735751295337, "grad_norm": 0.958997905254364, "learning_rate": 3.0946307513418095e-08, "loss": 1.0417, "step": 745 }, { "epoch": 1.9093264248704664, "grad_norm": 0.8062818050384521, "learning_rate": 2.8698593632357496e-08, "loss": 0.9683, "step": 746 }, { "epoch": 1.911917098445596, "grad_norm": 0.8307934403419495, "learning_rate": 2.653538025201685e-08, "loss": 0.9277, "step": 747 }, { "epoch": 1.9145077720207255, "grad_norm": 0.9197670221328735, "learning_rate": 2.4456704141967437e-08, "loss": 0.9539, "step": 748 }, { "epoch": 1.917098445595855, "grad_norm": 0.885750949382782, "learning_rate": 2.2462600634841624e-08, "loss": 0.9429, "step": 749 }, { "epoch": 1.9196891191709846, "grad_norm": 0.8606031537055969, "learning_rate": 2.0553103625737813e-08, "loss": 1.0091, "step": 750 }, { "epoch": 1.922279792746114, "grad_norm": 0.8774232268333435, "learning_rate": 1.8728245571640324e-08, "loss": 0.958, "step": 751 }, { "epoch": 1.9248704663212435, "grad_norm": 0.9953173398971558, "learning_rate": 1.6988057490868736e-08, "loss": 1.0015, "step": 752 }, { "epoch": 1.927461139896373, "grad_norm": 1.022383213043213, "learning_rate": 1.5332568962553306e-08, "loss": 0.9427, "step": 753 }, { "epoch": 1.9300518134715026, "grad_norm": 1.0359011888504028, "learning_rate": 1.3761808126126486e-08, "loss": 0.9499, "step": 754 }, { "epoch": 1.932642487046632, "grad_norm": 0.8054367899894714, "learning_rate": 1.2275801680851629e-08, "loss": 1.0467, "step": 755 }, { "epoch": 1.9352331606217616, "grad_norm": 0.921036422252655, "learning_rate": 1.0874574885362809e-08, "loss": 0.9384, "step": 756 }, { "epoch": 1.9378238341968912, "grad_norm": 0.8518050909042358, "learning_rate": 9.558151557240158e-09, "loss": 0.9375, "step": 757 }, { "epoch": 1.9404145077720207, "grad_norm": 0.9647573232650757, "learning_rate": 8.32655407260241e-09, "loss": 1.0186, "step": 758 }, { "epoch": 1.9430051813471503, "grad_norm": 0.97606360912323, "learning_rate": 7.179803365726656e-09, "loss": 0.9963, "step": 759 }, { "epoch": 1.9455958549222798, "grad_norm": 1.1201107501983643, "learning_rate": 6.117918928693623e-09, "loss": 0.983, "step": 760 }, { "epoch": 1.9481865284974094, "grad_norm": 0.9616057276725769, "learning_rate": 5.140918811056827e-09, "loss": 0.9794, "step": 761 }, { "epoch": 1.950777202072539, "grad_norm": 0.8792617917060852, "learning_rate": 4.248819619533384e-09, "loss": 0.9646, "step": 762 }, { "epoch": 1.9533678756476682, "grad_norm": 0.8594529032707214, "learning_rate": 3.4416365177236675e-09, "loss": 1.0342, "step": 763 }, { "epoch": 1.9559585492227978, "grad_norm": 0.9270037412643433, "learning_rate": 2.7193832258537447e-09, "loss": 1.0157, "step": 764 }, { "epoch": 1.9585492227979273, "grad_norm": 0.9513561725616455, "learning_rate": 2.08207202054056e-09, "loss": 1.0843, "step": 765 }, { "epoch": 1.9611398963730569, "grad_norm": 0.9862794876098633, "learning_rate": 1.5297137345843261e-09, "loss": 0.9374, "step": 766 }, { "epoch": 1.9637305699481864, "grad_norm": 0.889864444732666, "learning_rate": 1.0623177567847809e-09, "loss": 1.014, "step": 767 }, { "epoch": 1.966321243523316, "grad_norm": 0.850560188293457, "learning_rate": 6.798920317807601e-10, "loss": 1.0113, "step": 768 }, { "epoch": 1.9689119170984455, "grad_norm": 1.0084043741226196, "learning_rate": 3.8244305991530504e-10, "loss": 0.9921, "step": 769 }, { "epoch": 1.971502590673575, "grad_norm": 0.9245330691337585, "learning_rate": 1.6997589712575145e-10, "loss": 1.0003, "step": 770 }, { "epoch": 1.9740932642487046, "grad_norm": 0.9151455163955688, "learning_rate": 4.249415485657604e-11, "loss": 1.0092, "step": 771 }, { "epoch": 1.9766839378238341, "grad_norm": 0.8460503220558167, "learning_rate": 0.0, "loss": 0.9456, "step": 772 } ], "logging_steps": 1, "max_steps": 772, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 193, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.408535403456889e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }