mistral-poe-10p-detach / trainer_state.json
Qin Liu
Model save
ccb265c verified
raw
history blame
126 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 3898,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0,
"grad_norm": 0.18273725037333932,
"learning_rate": 5.128205128205128e-07,
"loss": 0.9609,
"step": 1
},
{
"epoch": 0.0,
"grad_norm": 0.4188407238663635,
"learning_rate": 2.564102564102564e-06,
"loss": 0.8848,
"step": 5
},
{
"epoch": 0.0,
"grad_norm": 0.19777627866810965,
"learning_rate": 5.128205128205128e-06,
"loss": 0.7625,
"step": 10
},
{
"epoch": 0.0,
"grad_norm": 0.22192343921645707,
"learning_rate": 7.692307692307694e-06,
"loss": 0.7876,
"step": 15
},
{
"epoch": 0.01,
"grad_norm": 0.18841857488548644,
"learning_rate": 1.0256410256410256e-05,
"loss": 0.8592,
"step": 20
},
{
"epoch": 0.01,
"grad_norm": 0.33681419457634554,
"learning_rate": 1.282051282051282e-05,
"loss": 0.8162,
"step": 25
},
{
"epoch": 0.01,
"grad_norm": 0.33600938003479874,
"learning_rate": 1.5384615384615387e-05,
"loss": 0.8894,
"step": 30
},
{
"epoch": 0.01,
"grad_norm": 0.3176697356314787,
"learning_rate": 1.794871794871795e-05,
"loss": 0.7155,
"step": 35
},
{
"epoch": 0.01,
"grad_norm": 0.21204726045838998,
"learning_rate": 2.0512820512820512e-05,
"loss": 0.7677,
"step": 40
},
{
"epoch": 0.01,
"grad_norm": 0.19741016497576444,
"learning_rate": 2.307692307692308e-05,
"loss": 0.767,
"step": 45
},
{
"epoch": 0.01,
"grad_norm": 0.2867616395864828,
"learning_rate": 2.564102564102564e-05,
"loss": 0.7459,
"step": 50
},
{
"epoch": 0.01,
"grad_norm": 0.277593018616144,
"learning_rate": 2.8205128205128207e-05,
"loss": 0.716,
"step": 55
},
{
"epoch": 0.02,
"grad_norm": 0.20494371343645457,
"learning_rate": 3.0769230769230774e-05,
"loss": 0.7252,
"step": 60
},
{
"epoch": 0.02,
"grad_norm": 0.2514927327142473,
"learning_rate": 3.3333333333333335e-05,
"loss": 0.7427,
"step": 65
},
{
"epoch": 0.02,
"grad_norm": 0.3569958514232062,
"learning_rate": 3.58974358974359e-05,
"loss": 0.766,
"step": 70
},
{
"epoch": 0.02,
"grad_norm": 0.24879302023855776,
"learning_rate": 3.846153846153846e-05,
"loss": 0.7423,
"step": 75
},
{
"epoch": 0.02,
"grad_norm": 0.25452303418558414,
"learning_rate": 4.1025641025641023e-05,
"loss": 0.6759,
"step": 80
},
{
"epoch": 0.02,
"grad_norm": 0.2839971702285102,
"learning_rate": 4.358974358974359e-05,
"loss": 0.7479,
"step": 85
},
{
"epoch": 0.02,
"grad_norm": 0.30384992603439903,
"learning_rate": 4.615384615384616e-05,
"loss": 0.6552,
"step": 90
},
{
"epoch": 0.02,
"grad_norm": 0.3852050595096914,
"learning_rate": 4.871794871794872e-05,
"loss": 0.6522,
"step": 95
},
{
"epoch": 0.03,
"grad_norm": 0.22298687548607224,
"learning_rate": 5.128205128205128e-05,
"loss": 0.7471,
"step": 100
},
{
"epoch": 0.03,
"grad_norm": 0.2048934487949906,
"learning_rate": 5.384615384615385e-05,
"loss": 0.7084,
"step": 105
},
{
"epoch": 0.03,
"grad_norm": 0.2069689289264241,
"learning_rate": 5.6410256410256414e-05,
"loss": 0.6885,
"step": 110
},
{
"epoch": 0.03,
"grad_norm": 0.24817118260446214,
"learning_rate": 5.897435897435898e-05,
"loss": 0.7736,
"step": 115
},
{
"epoch": 0.03,
"grad_norm": 0.40895048260028827,
"learning_rate": 6.153846153846155e-05,
"loss": 0.8098,
"step": 120
},
{
"epoch": 0.03,
"grad_norm": 0.35594857119791506,
"learning_rate": 6.410256410256412e-05,
"loss": 0.7543,
"step": 125
},
{
"epoch": 0.03,
"grad_norm": 0.21471650478613163,
"learning_rate": 6.666666666666667e-05,
"loss": 0.6914,
"step": 130
},
{
"epoch": 0.03,
"grad_norm": 0.3610348756628403,
"learning_rate": 6.923076923076924e-05,
"loss": 0.7253,
"step": 135
},
{
"epoch": 0.04,
"grad_norm": 0.26836543734793383,
"learning_rate": 7.17948717948718e-05,
"loss": 0.6539,
"step": 140
},
{
"epoch": 0.04,
"grad_norm": 0.2835751411318648,
"learning_rate": 7.435897435897436e-05,
"loss": 0.8097,
"step": 145
},
{
"epoch": 0.04,
"grad_norm": 0.20105648692084657,
"learning_rate": 7.692307692307693e-05,
"loss": 0.7385,
"step": 150
},
{
"epoch": 0.04,
"grad_norm": 0.16414727315119573,
"learning_rate": 7.948717948717948e-05,
"loss": 0.6527,
"step": 155
},
{
"epoch": 0.04,
"grad_norm": 0.24741062361895544,
"learning_rate": 8.205128205128205e-05,
"loss": 0.6857,
"step": 160
},
{
"epoch": 0.04,
"grad_norm": 0.20873983107979052,
"learning_rate": 8.461538461538461e-05,
"loss": 0.7498,
"step": 165
},
{
"epoch": 0.04,
"grad_norm": 0.22699678420428857,
"learning_rate": 8.717948717948718e-05,
"loss": 0.7112,
"step": 170
},
{
"epoch": 0.04,
"grad_norm": 0.19007788693963085,
"learning_rate": 8.974358974358975e-05,
"loss": 0.751,
"step": 175
},
{
"epoch": 0.05,
"grad_norm": 0.18866388272365925,
"learning_rate": 9.230769230769232e-05,
"loss": 0.6833,
"step": 180
},
{
"epoch": 0.05,
"grad_norm": 0.19818865870224087,
"learning_rate": 9.487179487179487e-05,
"loss": 0.6851,
"step": 185
},
{
"epoch": 0.05,
"grad_norm": 0.23302753556690228,
"learning_rate": 9.743589743589744e-05,
"loss": 0.7745,
"step": 190
},
{
"epoch": 0.05,
"grad_norm": 0.18540393542694006,
"learning_rate": 0.0001,
"loss": 0.7403,
"step": 195
},
{
"epoch": 0.05,
"grad_norm": 0.1839673690602772,
"learning_rate": 0.00010256410256410256,
"loss": 0.8231,
"step": 200
},
{
"epoch": 0.05,
"grad_norm": 0.22798420463327215,
"learning_rate": 0.00010512820512820514,
"loss": 0.6674,
"step": 205
},
{
"epoch": 0.05,
"grad_norm": 0.2464229066359474,
"learning_rate": 0.0001076923076923077,
"loss": 0.7846,
"step": 210
},
{
"epoch": 0.06,
"grad_norm": 0.19433765535813324,
"learning_rate": 0.00011025641025641027,
"loss": 0.7632,
"step": 215
},
{
"epoch": 0.06,
"grad_norm": 0.21304653327041365,
"learning_rate": 0.00011282051282051283,
"loss": 0.7541,
"step": 220
},
{
"epoch": 0.06,
"grad_norm": 0.1770514082218146,
"learning_rate": 0.00011538461538461538,
"loss": 0.6967,
"step": 225
},
{
"epoch": 0.06,
"grad_norm": 0.22467577773491165,
"learning_rate": 0.00011794871794871796,
"loss": 0.7131,
"step": 230
},
{
"epoch": 0.06,
"grad_norm": 0.1765044110348834,
"learning_rate": 0.00012051282051282052,
"loss": 0.7442,
"step": 235
},
{
"epoch": 0.06,
"grad_norm": 0.29795271013699604,
"learning_rate": 0.0001230769230769231,
"loss": 0.8024,
"step": 240
},
{
"epoch": 0.06,
"grad_norm": 0.14389624633185258,
"learning_rate": 0.00012564102564102564,
"loss": 0.7101,
"step": 245
},
{
"epoch": 0.06,
"grad_norm": 0.21079245502502064,
"learning_rate": 0.00012820512820512823,
"loss": 0.709,
"step": 250
},
{
"epoch": 0.07,
"grad_norm": 0.17829547036873233,
"learning_rate": 0.00013076923076923077,
"loss": 0.7443,
"step": 255
},
{
"epoch": 0.07,
"grad_norm": 0.2173226668602675,
"learning_rate": 0.00013333333333333334,
"loss": 0.6832,
"step": 260
},
{
"epoch": 0.07,
"grad_norm": 0.26876676153975865,
"learning_rate": 0.0001358974358974359,
"loss": 0.7284,
"step": 265
},
{
"epoch": 0.07,
"grad_norm": 0.15830628083598494,
"learning_rate": 0.00013846153846153847,
"loss": 0.7194,
"step": 270
},
{
"epoch": 0.07,
"grad_norm": 0.19342822085523578,
"learning_rate": 0.00014102564102564104,
"loss": 0.7217,
"step": 275
},
{
"epoch": 0.07,
"grad_norm": 0.18871672612378254,
"learning_rate": 0.0001435897435897436,
"loss": 0.6593,
"step": 280
},
{
"epoch": 0.07,
"grad_norm": 0.21272446268662773,
"learning_rate": 0.00014615384615384615,
"loss": 0.6667,
"step": 285
},
{
"epoch": 0.07,
"grad_norm": 0.2155663432457548,
"learning_rate": 0.00014871794871794872,
"loss": 0.6597,
"step": 290
},
{
"epoch": 0.08,
"grad_norm": 0.44468877213339963,
"learning_rate": 0.00015128205128205128,
"loss": 0.7204,
"step": 295
},
{
"epoch": 0.08,
"grad_norm": 0.2164584239122828,
"learning_rate": 0.00015384615384615385,
"loss": 0.8244,
"step": 300
},
{
"epoch": 0.08,
"grad_norm": 0.1938136691775617,
"learning_rate": 0.00015641025641025642,
"loss": 0.6991,
"step": 305
},
{
"epoch": 0.08,
"grad_norm": 0.1823678116608096,
"learning_rate": 0.00015897435897435896,
"loss": 0.6541,
"step": 310
},
{
"epoch": 0.08,
"grad_norm": 0.1629985226358383,
"learning_rate": 0.00016153846153846155,
"loss": 0.7331,
"step": 315
},
{
"epoch": 0.08,
"grad_norm": 0.19599880954946838,
"learning_rate": 0.0001641025641025641,
"loss": 0.6446,
"step": 320
},
{
"epoch": 0.08,
"grad_norm": 0.23282506824657653,
"learning_rate": 0.0001666666666666667,
"loss": 0.6723,
"step": 325
},
{
"epoch": 0.08,
"grad_norm": 0.20570313944345442,
"learning_rate": 0.00016923076923076923,
"loss": 0.7924,
"step": 330
},
{
"epoch": 0.09,
"grad_norm": 0.18124135373904873,
"learning_rate": 0.0001717948717948718,
"loss": 0.8191,
"step": 335
},
{
"epoch": 0.09,
"grad_norm": 0.18929176365178993,
"learning_rate": 0.00017435897435897436,
"loss": 0.7289,
"step": 340
},
{
"epoch": 0.09,
"grad_norm": 0.19977748525194552,
"learning_rate": 0.00017692307692307693,
"loss": 0.7362,
"step": 345
},
{
"epoch": 0.09,
"grad_norm": 0.1990441990377106,
"learning_rate": 0.0001794871794871795,
"loss": 0.853,
"step": 350
},
{
"epoch": 0.09,
"grad_norm": 0.3087674819281357,
"learning_rate": 0.00018205128205128207,
"loss": 0.6986,
"step": 355
},
{
"epoch": 0.09,
"grad_norm": 0.25912532033253527,
"learning_rate": 0.00018461538461538463,
"loss": 0.6727,
"step": 360
},
{
"epoch": 0.09,
"grad_norm": 0.30170168244659684,
"learning_rate": 0.0001871794871794872,
"loss": 0.7641,
"step": 365
},
{
"epoch": 0.09,
"grad_norm": 0.23159332627603768,
"learning_rate": 0.00018974358974358974,
"loss": 0.7467,
"step": 370
},
{
"epoch": 0.1,
"grad_norm": 0.21437363907196583,
"learning_rate": 0.00019230769230769233,
"loss": 0.689,
"step": 375
},
{
"epoch": 0.1,
"grad_norm": 0.18673889188699844,
"learning_rate": 0.00019487179487179487,
"loss": 0.625,
"step": 380
},
{
"epoch": 0.1,
"grad_norm": 0.24453425122203856,
"learning_rate": 0.00019743589743589744,
"loss": 0.6991,
"step": 385
},
{
"epoch": 0.1,
"grad_norm": 0.19852275666733094,
"learning_rate": 0.0002,
"loss": 0.8059,
"step": 390
},
{
"epoch": 0.1,
"grad_norm": 0.183724259914749,
"learning_rate": 0.00019999899748734544,
"loss": 0.7681,
"step": 395
},
{
"epoch": 0.1,
"grad_norm": 0.2127551390821211,
"learning_rate": 0.00019999598996948235,
"loss": 0.6547,
"step": 400
},
{
"epoch": 0.1,
"grad_norm": 0.2684683587075289,
"learning_rate": 0.00019999097750671223,
"loss": 0.7287,
"step": 405
},
{
"epoch": 0.11,
"grad_norm": 0.2283900397579642,
"learning_rate": 0.00019998396019953624,
"loss": 0.6885,
"step": 410
},
{
"epoch": 0.11,
"grad_norm": 0.24702860686198994,
"learning_rate": 0.00019997493818865318,
"loss": 0.7212,
"step": 415
},
{
"epoch": 0.11,
"grad_norm": 0.241475192547802,
"learning_rate": 0.0001999639116549566,
"loss": 0.5949,
"step": 420
},
{
"epoch": 0.11,
"grad_norm": 0.23533430796243504,
"learning_rate": 0.00019995088081953136,
"loss": 0.7203,
"step": 425
},
{
"epoch": 0.11,
"grad_norm": 0.24956850711444106,
"learning_rate": 0.00019993584594364894,
"loss": 0.6447,
"step": 430
},
{
"epoch": 0.11,
"grad_norm": 0.20131118173728133,
"learning_rate": 0.00019991880732876246,
"loss": 0.699,
"step": 435
},
{
"epoch": 0.11,
"grad_norm": 0.2067287885265624,
"learning_rate": 0.0001998997653165004,
"loss": 0.74,
"step": 440
},
{
"epoch": 0.11,
"grad_norm": 0.20357844001623,
"learning_rate": 0.00019987872028866003,
"loss": 0.8027,
"step": 445
},
{
"epoch": 0.12,
"grad_norm": 0.23984712872623487,
"learning_rate": 0.00019985567266719934,
"loss": 0.7839,
"step": 450
},
{
"epoch": 0.12,
"grad_norm": 0.1897862238822482,
"learning_rate": 0.00019983062291422908,
"loss": 0.7433,
"step": 455
},
{
"epoch": 0.12,
"grad_norm": 0.24095134724957823,
"learning_rate": 0.00019980357153200315,
"loss": 0.7096,
"step": 460
},
{
"epoch": 0.12,
"grad_norm": 0.2852750359187981,
"learning_rate": 0.00019977451906290854,
"loss": 0.7163,
"step": 465
},
{
"epoch": 0.12,
"grad_norm": 0.20980919665053135,
"learning_rate": 0.00019974346608945466,
"loss": 0.7439,
"step": 470
},
{
"epoch": 0.12,
"grad_norm": 0.2295278218977822,
"learning_rate": 0.0001997104132342614,
"loss": 0.6644,
"step": 475
},
{
"epoch": 0.12,
"grad_norm": 0.19640395085838386,
"learning_rate": 0.00019967536116004698,
"loss": 0.8113,
"step": 480
},
{
"epoch": 0.12,
"grad_norm": 0.23742262528964145,
"learning_rate": 0.00019963831056961433,
"loss": 0.6803,
"step": 485
},
{
"epoch": 0.13,
"grad_norm": 0.2209130794031639,
"learning_rate": 0.00019959926220583713,
"loss": 0.7189,
"step": 490
},
{
"epoch": 0.13,
"grad_norm": 0.4158388376129303,
"learning_rate": 0.000199558216851645,
"loss": 0.7337,
"step": 495
},
{
"epoch": 0.13,
"grad_norm": 0.21568480709969076,
"learning_rate": 0.00019951517533000764,
"loss": 0.6068,
"step": 500
},
{
"epoch": 0.13,
"grad_norm": 0.23552741626776288,
"learning_rate": 0.00019947013850391847,
"loss": 0.7098,
"step": 505
},
{
"epoch": 0.13,
"grad_norm": 0.2277845662969724,
"learning_rate": 0.00019942310727637724,
"loss": 0.7695,
"step": 510
},
{
"epoch": 0.13,
"grad_norm": 0.24111051524751667,
"learning_rate": 0.000199374082590372,
"loss": 0.7586,
"step": 515
},
{
"epoch": 0.13,
"grad_norm": 0.19134678351380507,
"learning_rate": 0.00019932306542886009,
"loss": 0.7393,
"step": 520
},
{
"epoch": 0.13,
"grad_norm": 0.23589856188985894,
"learning_rate": 0.0001992700568147485,
"loss": 0.7205,
"step": 525
},
{
"epoch": 0.14,
"grad_norm": 0.252896098878244,
"learning_rate": 0.00019921505781087334,
"loss": 0.7336,
"step": 530
},
{
"epoch": 0.14,
"grad_norm": 0.25248409296825886,
"learning_rate": 0.00019915806951997862,
"loss": 0.7949,
"step": 535
},
{
"epoch": 0.14,
"grad_norm": 0.20137551343484256,
"learning_rate": 0.00019909909308469398,
"loss": 0.7558,
"step": 540
},
{
"epoch": 0.14,
"grad_norm": 0.2111631505276513,
"learning_rate": 0.0001990381296875118,
"loss": 0.6837,
"step": 545
},
{
"epoch": 0.14,
"grad_norm": 0.29184203037550105,
"learning_rate": 0.0001989751805507637,
"loss": 0.8117,
"step": 550
},
{
"epoch": 0.14,
"grad_norm": 0.2206173102779168,
"learning_rate": 0.0001989102469365958,
"loss": 0.8164,
"step": 555
},
{
"epoch": 0.14,
"grad_norm": 0.21507931251783377,
"learning_rate": 0.00019884333014694345,
"loss": 0.7133,
"step": 560
},
{
"epoch": 0.14,
"grad_norm": 0.22239957224386717,
"learning_rate": 0.00019877443152350527,
"loss": 0.7375,
"step": 565
},
{
"epoch": 0.15,
"grad_norm": 0.21591708491473222,
"learning_rate": 0.00019870355244771607,
"loss": 0.6708,
"step": 570
},
{
"epoch": 0.15,
"grad_norm": 0.25918641174746915,
"learning_rate": 0.0001986306943407193,
"loss": 0.7389,
"step": 575
},
{
"epoch": 0.15,
"grad_norm": 0.22262329838643527,
"learning_rate": 0.00019855585866333835,
"loss": 0.7479,
"step": 580
},
{
"epoch": 0.15,
"grad_norm": 0.2449638774409369,
"learning_rate": 0.00019847904691604757,
"loss": 0.7486,
"step": 585
},
{
"epoch": 0.15,
"grad_norm": 0.2485277159299432,
"learning_rate": 0.00019840026063894193,
"loss": 0.6979,
"step": 590
},
{
"epoch": 0.15,
"grad_norm": 0.28798789472500747,
"learning_rate": 0.0001983195014117062,
"loss": 0.7392,
"step": 595
},
{
"epoch": 0.15,
"grad_norm": 0.2560345551346601,
"learning_rate": 0.00019823677085358335,
"loss": 0.6557,
"step": 600
},
{
"epoch": 0.16,
"grad_norm": 0.2503029916293632,
"learning_rate": 0.00019815207062334197,
"loss": 0.684,
"step": 605
},
{
"epoch": 0.16,
"grad_norm": 0.22343460564706522,
"learning_rate": 0.00019806540241924317,
"loss": 0.7469,
"step": 610
},
{
"epoch": 0.16,
"grad_norm": 0.23478030369638528,
"learning_rate": 0.00019797676797900633,
"loss": 0.7651,
"step": 615
},
{
"epoch": 0.16,
"grad_norm": 0.23393987623099682,
"learning_rate": 0.00019788616907977441,
"loss": 0.7356,
"step": 620
},
{
"epoch": 0.16,
"grad_norm": 0.1835277686003351,
"learning_rate": 0.0001977936075380783,
"loss": 0.6319,
"step": 625
},
{
"epoch": 0.16,
"grad_norm": 0.22084991905168924,
"learning_rate": 0.00019769908520980034,
"loss": 0.7451,
"step": 630
},
{
"epoch": 0.16,
"grad_norm": 0.2532435897912756,
"learning_rate": 0.00019760260399013708,
"loss": 0.7217,
"step": 635
},
{
"epoch": 0.16,
"grad_norm": 0.23471511654283184,
"learning_rate": 0.00019750416581356146,
"loss": 0.7114,
"step": 640
},
{
"epoch": 0.17,
"grad_norm": 0.21710442836493216,
"learning_rate": 0.0001974037726537838,
"loss": 0.628,
"step": 645
},
{
"epoch": 0.17,
"grad_norm": 0.267924875976266,
"learning_rate": 0.00019730142652371236,
"loss": 0.6469,
"step": 650
},
{
"epoch": 0.17,
"grad_norm": 0.268474985537624,
"learning_rate": 0.00019719712947541295,
"loss": 0.6829,
"step": 655
},
{
"epoch": 0.17,
"grad_norm": 0.16700886295844172,
"learning_rate": 0.0001970908836000678,
"loss": 0.6877,
"step": 660
},
{
"epoch": 0.17,
"grad_norm": 0.22055090509747258,
"learning_rate": 0.00019698269102793358,
"loss": 0.6525,
"step": 665
},
{
"epoch": 0.17,
"grad_norm": 0.27128839715879177,
"learning_rate": 0.00019687255392829877,
"loss": 0.6582,
"step": 670
},
{
"epoch": 0.17,
"grad_norm": 0.21159901567551387,
"learning_rate": 0.00019676047450944008,
"loss": 0.5732,
"step": 675
},
{
"epoch": 0.17,
"grad_norm": 0.24374137032809187,
"learning_rate": 0.0001966464550185782,
"loss": 0.7634,
"step": 680
},
{
"epoch": 0.18,
"grad_norm": 0.24672679998063676,
"learning_rate": 0.00019653049774183282,
"loss": 0.5845,
"step": 685
},
{
"epoch": 0.18,
"grad_norm": 0.2482930627863346,
"learning_rate": 0.0001964126050041767,
"loss": 0.7651,
"step": 690
},
{
"epoch": 0.18,
"grad_norm": 0.2553369576381814,
"learning_rate": 0.000196292779169389,
"loss": 0.6712,
"step": 695
},
{
"epoch": 0.18,
"grad_norm": 0.30911752230079176,
"learning_rate": 0.0001961710226400081,
"loss": 0.7166,
"step": 700
},
{
"epoch": 0.18,
"grad_norm": 0.31535946743812887,
"learning_rate": 0.00019604733785728317,
"loss": 0.7117,
"step": 705
},
{
"epoch": 0.18,
"grad_norm": 0.28460415510467557,
"learning_rate": 0.00019592172730112544,
"loss": 0.709,
"step": 710
},
{
"epoch": 0.18,
"grad_norm": 0.26288174322385877,
"learning_rate": 0.00019579419349005837,
"loss": 0.7949,
"step": 715
},
{
"epoch": 0.18,
"grad_norm": 0.24918287719559648,
"learning_rate": 0.00019566473898116713,
"loss": 0.7045,
"step": 720
},
{
"epoch": 0.19,
"grad_norm": 0.26861534205040266,
"learning_rate": 0.00019553336637004735,
"loss": 0.7401,
"step": 725
},
{
"epoch": 0.19,
"grad_norm": 0.20350229137993933,
"learning_rate": 0.0001954000782907532,
"loss": 0.7521,
"step": 730
},
{
"epoch": 0.19,
"grad_norm": 0.253527308804326,
"learning_rate": 0.00019526487741574437,
"loss": 0.7833,
"step": 735
},
{
"epoch": 0.19,
"grad_norm": 0.27106014622439245,
"learning_rate": 0.00019512776645583263,
"loss": 0.6898,
"step": 740
},
{
"epoch": 0.19,
"grad_norm": 0.24534894717688374,
"learning_rate": 0.0001949887481601274,
"loss": 0.8102,
"step": 745
},
{
"epoch": 0.19,
"grad_norm": 0.37673623038318094,
"learning_rate": 0.00019484782531598073,
"loss": 0.6929,
"step": 750
},
{
"epoch": 0.19,
"grad_norm": 0.28044355813302363,
"learning_rate": 0.0001947050007489313,
"loss": 0.8396,
"step": 755
},
{
"epoch": 0.19,
"grad_norm": 0.25883167182505523,
"learning_rate": 0.00019456027732264784,
"loss": 0.7169,
"step": 760
},
{
"epoch": 0.2,
"grad_norm": 0.33212091144862455,
"learning_rate": 0.00019441365793887162,
"loss": 0.7562,
"step": 765
},
{
"epoch": 0.2,
"grad_norm": 0.2265256995053789,
"learning_rate": 0.00019426514553735848,
"loss": 0.733,
"step": 770
},
{
"epoch": 0.2,
"grad_norm": 0.1952769250523588,
"learning_rate": 0.00019411474309581958,
"loss": 0.6875,
"step": 775
},
{
"epoch": 0.2,
"grad_norm": 0.23715930446049704,
"learning_rate": 0.00019396245362986197,
"loss": 0.7183,
"step": 780
},
{
"epoch": 0.2,
"grad_norm": 0.26737163458762153,
"learning_rate": 0.00019380828019292798,
"loss": 0.6693,
"step": 785
},
{
"epoch": 0.2,
"grad_norm": 0.3093664057575012,
"learning_rate": 0.00019365222587623405,
"loss": 0.6971,
"step": 790
},
{
"epoch": 0.2,
"grad_norm": 0.22025157297022468,
"learning_rate": 0.00019349429380870873,
"loss": 0.7437,
"step": 795
},
{
"epoch": 0.21,
"grad_norm": 0.23171097491730766,
"learning_rate": 0.00019333448715692995,
"loss": 0.7528,
"step": 800
},
{
"epoch": 0.21,
"grad_norm": 0.23951739750544596,
"learning_rate": 0.0001931728091250615,
"loss": 0.7849,
"step": 805
},
{
"epoch": 0.21,
"grad_norm": 0.24639640582380729,
"learning_rate": 0.00019300926295478884,
"loss": 0.7688,
"step": 810
},
{
"epoch": 0.21,
"grad_norm": 0.24023205632785313,
"learning_rate": 0.00019284385192525405,
"loss": 0.6915,
"step": 815
},
{
"epoch": 0.21,
"grad_norm": 0.2548147235517417,
"learning_rate": 0.0001926765793529902,
"loss": 0.7137,
"step": 820
},
{
"epoch": 0.21,
"grad_norm": 0.24890950536791295,
"learning_rate": 0.00019250744859185468,
"loss": 0.7904,
"step": 825
},
{
"epoch": 0.21,
"grad_norm": 0.31444435610399357,
"learning_rate": 0.00019233646303296205,
"loss": 0.708,
"step": 830
},
{
"epoch": 0.21,
"grad_norm": 0.25866961126213517,
"learning_rate": 0.000192163626104616,
"loss": 0.6808,
"step": 835
},
{
"epoch": 0.22,
"grad_norm": 0.2605403676294042,
"learning_rate": 0.00019198894127224074,
"loss": 0.7549,
"step": 840
},
{
"epoch": 0.22,
"grad_norm": 0.36813656529214944,
"learning_rate": 0.00019181241203831137,
"loss": 0.7715,
"step": 845
},
{
"epoch": 0.22,
"grad_norm": 0.30866872454744043,
"learning_rate": 0.0001916340419422837,
"loss": 0.7488,
"step": 850
},
{
"epoch": 0.22,
"grad_norm": 0.26715773589614006,
"learning_rate": 0.00019145383456052327,
"loss": 0.6933,
"step": 855
},
{
"epoch": 0.22,
"grad_norm": 0.2405309690369464,
"learning_rate": 0.00019127179350623372,
"loss": 0.6964,
"step": 860
},
{
"epoch": 0.22,
"grad_norm": 0.34837890747063266,
"learning_rate": 0.00019108792242938425,
"loss": 0.6903,
"step": 865
},
{
"epoch": 0.22,
"grad_norm": 0.28304812900111975,
"learning_rate": 0.0001909022250166365,
"loss": 0.6982,
"step": 870
},
{
"epoch": 0.22,
"grad_norm": 0.2123369848990679,
"learning_rate": 0.00019071470499127058,
"loss": 0.6929,
"step": 875
},
{
"epoch": 0.23,
"grad_norm": 0.2836396726336964,
"learning_rate": 0.00019052536611311046,
"loss": 0.7434,
"step": 880
},
{
"epoch": 0.23,
"grad_norm": 0.2247103209214737,
"learning_rate": 0.0001903342121784486,
"loss": 0.6816,
"step": 885
},
{
"epoch": 0.23,
"grad_norm": 0.3242567099018861,
"learning_rate": 0.00019014124701996973,
"loss": 0.72,
"step": 890
},
{
"epoch": 0.23,
"grad_norm": 0.2339879402133849,
"learning_rate": 0.00018994647450667413,
"loss": 0.7041,
"step": 895
},
{
"epoch": 0.23,
"grad_norm": 0.22751659387372616,
"learning_rate": 0.00018974989854379996,
"loss": 0.7511,
"step": 900
},
{
"epoch": 0.23,
"grad_norm": 0.26549639698173727,
"learning_rate": 0.00018955152307274507,
"loss": 0.6897,
"step": 905
},
{
"epoch": 0.23,
"grad_norm": 0.21860533117362732,
"learning_rate": 0.00018935135207098785,
"loss": 0.6648,
"step": 910
},
{
"epoch": 0.23,
"grad_norm": 0.2533868012784127,
"learning_rate": 0.00018914938955200754,
"loss": 0.6798,
"step": 915
},
{
"epoch": 0.24,
"grad_norm": 0.24616813985140945,
"learning_rate": 0.00018894563956520374,
"loss": 0.7683,
"step": 920
},
{
"epoch": 0.24,
"grad_norm": 0.2668291016708883,
"learning_rate": 0.0001887401061958153,
"loss": 0.7791,
"step": 925
},
{
"epoch": 0.24,
"grad_norm": 0.2556952354899014,
"learning_rate": 0.00018853279356483826,
"loss": 0.7733,
"step": 930
},
{
"epoch": 0.24,
"grad_norm": 0.2824190700646869,
"learning_rate": 0.00018832370582894334,
"loss": 0.6439,
"step": 935
},
{
"epoch": 0.24,
"grad_norm": 0.3193420311338137,
"learning_rate": 0.00018811284718039256,
"loss": 0.6791,
"step": 940
},
{
"epoch": 0.24,
"grad_norm": 0.25742868616486125,
"learning_rate": 0.00018790022184695523,
"loss": 0.6835,
"step": 945
},
{
"epoch": 0.24,
"grad_norm": 0.31407801057679396,
"learning_rate": 0.00018768583409182305,
"loss": 0.7338,
"step": 950
},
{
"epoch": 0.24,
"grad_norm": 0.2268676531535162,
"learning_rate": 0.00018746968821352483,
"loss": 0.74,
"step": 955
},
{
"epoch": 0.25,
"grad_norm": 0.2706082597185793,
"learning_rate": 0.00018725178854584007,
"loss": 0.7372,
"step": 960
},
{
"epoch": 0.25,
"grad_norm": 0.23942119358366734,
"learning_rate": 0.00018703213945771229,
"loss": 0.7486,
"step": 965
},
{
"epoch": 0.25,
"grad_norm": 0.23415626163520642,
"learning_rate": 0.00018681074535316125,
"loss": 0.8149,
"step": 970
},
{
"epoch": 0.25,
"grad_norm": 0.23625273115645273,
"learning_rate": 0.00018658761067119484,
"loss": 0.6954,
"step": 975
},
{
"epoch": 0.25,
"grad_norm": 0.29616413603025993,
"learning_rate": 0.00018636273988571991,
"loss": 0.6825,
"step": 980
},
{
"epoch": 0.25,
"grad_norm": 0.24204431321093076,
"learning_rate": 0.00018613613750545255,
"loss": 0.6402,
"step": 985
},
{
"epoch": 0.25,
"grad_norm": 0.23843536272675503,
"learning_rate": 0.0001859078080738279,
"loss": 0.7034,
"step": 990
},
{
"epoch": 0.26,
"grad_norm": 0.2422607510213604,
"learning_rate": 0.00018567775616890888,
"loss": 0.7258,
"step": 995
},
{
"epoch": 0.26,
"grad_norm": 0.2367697702106968,
"learning_rate": 0.00018544598640329432,
"loss": 0.7125,
"step": 1000
},
{
"epoch": 0.26,
"grad_norm": 0.21770302111684567,
"learning_rate": 0.00018521250342402672,
"loss": 0.7651,
"step": 1005
},
{
"epoch": 0.26,
"grad_norm": 0.19822970981192417,
"learning_rate": 0.00018497731191249894,
"loss": 0.7705,
"step": 1010
},
{
"epoch": 0.26,
"grad_norm": 0.23080773465076526,
"learning_rate": 0.00018474041658436027,
"loss": 0.7721,
"step": 1015
},
{
"epoch": 0.26,
"grad_norm": 0.22821163739265107,
"learning_rate": 0.000184501822189422,
"loss": 0.7324,
"step": 1020
},
{
"epoch": 0.26,
"grad_norm": 0.22804940526863002,
"learning_rate": 0.0001842615335115621,
"loss": 0.7121,
"step": 1025
},
{
"epoch": 0.26,
"grad_norm": 0.2703519880608756,
"learning_rate": 0.00018401955536862948,
"loss": 0.8063,
"step": 1030
},
{
"epoch": 0.27,
"grad_norm": 0.27550361128710155,
"learning_rate": 0.00018377589261234705,
"loss": 0.7164,
"step": 1035
},
{
"epoch": 0.27,
"grad_norm": 0.19381814425434957,
"learning_rate": 0.0001835305501282148,
"loss": 0.8273,
"step": 1040
},
{
"epoch": 0.27,
"grad_norm": 0.26340620883694543,
"learning_rate": 0.00018328353283541158,
"loss": 0.7239,
"step": 1045
},
{
"epoch": 0.27,
"grad_norm": 0.2620607615738064,
"learning_rate": 0.00018303484568669667,
"loss": 0.6904,
"step": 1050
},
{
"epoch": 0.27,
"grad_norm": 0.15029608182232604,
"learning_rate": 0.00018278449366831035,
"loss": 0.6003,
"step": 1055
},
{
"epoch": 0.27,
"grad_norm": 0.2187384668280594,
"learning_rate": 0.00018253248179987388,
"loss": 0.707,
"step": 1060
},
{
"epoch": 0.27,
"grad_norm": 0.2560638257516859,
"learning_rate": 0.00018227881513428908,
"loss": 0.7375,
"step": 1065
},
{
"epoch": 0.27,
"grad_norm": 0.26865842108524934,
"learning_rate": 0.0001820234987576368,
"loss": 0.7077,
"step": 1070
},
{
"epoch": 0.28,
"grad_norm": 0.24178826048580734,
"learning_rate": 0.00018176653778907492,
"loss": 0.7354,
"step": 1075
},
{
"epoch": 0.28,
"grad_norm": 0.27854735242372025,
"learning_rate": 0.00018150793738073602,
"loss": 0.6523,
"step": 1080
},
{
"epoch": 0.28,
"grad_norm": 0.28792834941972884,
"learning_rate": 0.00018124770271762364,
"loss": 0.6969,
"step": 1085
},
{
"epoch": 0.28,
"grad_norm": 0.30573100271052395,
"learning_rate": 0.00018098583901750867,
"loss": 0.7093,
"step": 1090
},
{
"epoch": 0.28,
"grad_norm": 0.35336463231783916,
"learning_rate": 0.00018072235153082455,
"loss": 0.6441,
"step": 1095
},
{
"epoch": 0.28,
"grad_norm": 0.23753636938939662,
"learning_rate": 0.00018045724554056214,
"loss": 0.7617,
"step": 1100
},
{
"epoch": 0.28,
"grad_norm": 0.2830346076575104,
"learning_rate": 0.0001801905263621636,
"loss": 0.7642,
"step": 1105
},
{
"epoch": 0.28,
"grad_norm": 0.2431998404354548,
"learning_rate": 0.0001799221993434159,
"loss": 0.7183,
"step": 1110
},
{
"epoch": 0.29,
"grad_norm": 0.19371637012307855,
"learning_rate": 0.00017965226986434377,
"loss": 0.6834,
"step": 1115
},
{
"epoch": 0.29,
"grad_norm": 0.371529742091536,
"learning_rate": 0.00017938074333710157,
"loss": 0.7232,
"step": 1120
},
{
"epoch": 0.29,
"grad_norm": 0.26377679420092087,
"learning_rate": 0.00017910762520586485,
"loss": 0.7247,
"step": 1125
},
{
"epoch": 0.29,
"grad_norm": 0.2285136389359427,
"learning_rate": 0.00017883292094672128,
"loss": 0.7205,
"step": 1130
},
{
"epoch": 0.29,
"grad_norm": 0.29796654170976006,
"learning_rate": 0.00017855663606756078,
"loss": 0.6848,
"step": 1135
},
{
"epoch": 0.29,
"grad_norm": 0.2046721683597502,
"learning_rate": 0.00017827877610796514,
"loss": 0.6759,
"step": 1140
},
{
"epoch": 0.29,
"grad_norm": 0.3268765805116052,
"learning_rate": 0.00017799934663909682,
"loss": 0.716,
"step": 1145
},
{
"epoch": 0.3,
"grad_norm": 0.285963694250622,
"learning_rate": 0.00017771835326358743,
"loss": 0.6486,
"step": 1150
},
{
"epoch": 0.3,
"grad_norm": 0.2520717807255334,
"learning_rate": 0.00017743580161542525,
"loss": 0.6561,
"step": 1155
},
{
"epoch": 0.3,
"grad_norm": 0.1981002914068182,
"learning_rate": 0.00017715169735984233,
"loss": 0.6599,
"step": 1160
},
{
"epoch": 0.3,
"grad_norm": 0.26598319836843726,
"learning_rate": 0.00017686604619320093,
"loss": 0.7375,
"step": 1165
},
{
"epoch": 0.3,
"grad_norm": 0.26235482801170973,
"learning_rate": 0.0001765788538428792,
"loss": 0.8063,
"step": 1170
},
{
"epoch": 0.3,
"grad_norm": 0.3377576354866203,
"learning_rate": 0.00017629012606715648,
"loss": 0.7493,
"step": 1175
},
{
"epoch": 0.3,
"grad_norm": 0.3678264436119463,
"learning_rate": 0.00017599986865509767,
"loss": 0.6989,
"step": 1180
},
{
"epoch": 0.3,
"grad_norm": 0.32747709902543076,
"learning_rate": 0.00017570808742643746,
"loss": 0.7948,
"step": 1185
},
{
"epoch": 0.31,
"grad_norm": 0.24912874308295901,
"learning_rate": 0.00017541478823146327,
"loss": 0.7187,
"step": 1190
},
{
"epoch": 0.31,
"grad_norm": 0.2353274982649047,
"learning_rate": 0.00017511997695089822,
"loss": 0.7232,
"step": 1195
},
{
"epoch": 0.31,
"grad_norm": 0.2537606981534416,
"learning_rate": 0.00017482365949578302,
"loss": 0.6607,
"step": 1200
},
{
"epoch": 0.31,
"grad_norm": 0.2372984581891708,
"learning_rate": 0.00017452584180735774,
"loss": 0.732,
"step": 1205
},
{
"epoch": 0.31,
"grad_norm": 0.23094752394219156,
"learning_rate": 0.00017422652985694237,
"loss": 0.7065,
"step": 1210
},
{
"epoch": 0.31,
"grad_norm": 0.21971428715056102,
"learning_rate": 0.00017392572964581725,
"loss": 0.7012,
"step": 1215
},
{
"epoch": 0.31,
"grad_norm": 0.21591121328899532,
"learning_rate": 0.00017362344720510278,
"loss": 0.7083,
"step": 1220
},
{
"epoch": 0.31,
"grad_norm": 0.2779472292134201,
"learning_rate": 0.00017331968859563834,
"loss": 0.7777,
"step": 1225
},
{
"epoch": 0.32,
"grad_norm": 0.37145422903296105,
"learning_rate": 0.00017301445990786102,
"loss": 0.7592,
"step": 1230
},
{
"epoch": 0.32,
"grad_norm": 0.2838257377376314,
"learning_rate": 0.00017270776726168317,
"loss": 0.7147,
"step": 1235
},
{
"epoch": 0.32,
"grad_norm": 0.24195971773139494,
"learning_rate": 0.00017239961680637,
"loss": 0.7442,
"step": 1240
},
{
"epoch": 0.32,
"grad_norm": 0.2531071173762594,
"learning_rate": 0.00017209001472041617,
"loss": 0.7964,
"step": 1245
},
{
"epoch": 0.32,
"grad_norm": 0.2392782350831628,
"learning_rate": 0.0001717789672114218,
"loss": 0.7626,
"step": 1250
},
{
"epoch": 0.32,
"grad_norm": 0.24505477687731997,
"learning_rate": 0.00017146648051596822,
"loss": 0.7101,
"step": 1255
},
{
"epoch": 0.32,
"grad_norm": 0.27488228336035475,
"learning_rate": 0.0001711525608994927,
"loss": 0.6718,
"step": 1260
},
{
"epoch": 0.32,
"grad_norm": 0.1916616192257934,
"learning_rate": 0.00017083721465616306,
"loss": 0.6653,
"step": 1265
},
{
"epoch": 0.33,
"grad_norm": 0.2960082493036361,
"learning_rate": 0.00017052044810875126,
"loss": 0.8172,
"step": 1270
},
{
"epoch": 0.33,
"grad_norm": 0.29672233161501615,
"learning_rate": 0.00017020226760850677,
"loss": 0.695,
"step": 1275
},
{
"epoch": 0.33,
"grad_norm": 0.29971905663286563,
"learning_rate": 0.00016988267953502913,
"loss": 0.6921,
"step": 1280
},
{
"epoch": 0.33,
"grad_norm": 0.1972856751409119,
"learning_rate": 0.0001695616902961401,
"loss": 0.6707,
"step": 1285
},
{
"epoch": 0.33,
"grad_norm": 0.24592016529147168,
"learning_rate": 0.00016923930632775516,
"loss": 0.7049,
"step": 1290
},
{
"epoch": 0.33,
"grad_norm": 0.2362085501756166,
"learning_rate": 0.00016891553409375444,
"loss": 0.7222,
"step": 1295
},
{
"epoch": 0.33,
"grad_norm": 0.2720995150566844,
"learning_rate": 0.00016859038008585326,
"loss": 0.7355,
"step": 1300
},
{
"epoch": 0.33,
"grad_norm": 0.31383124999534473,
"learning_rate": 0.0001682638508234717,
"loss": 0.6735,
"step": 1305
},
{
"epoch": 0.34,
"grad_norm": 0.19810610550474098,
"learning_rate": 0.0001679359528536041,
"loss": 0.6522,
"step": 1310
},
{
"epoch": 0.34,
"grad_norm": 0.31481804318721185,
"learning_rate": 0.0001676066927506878,
"loss": 0.7391,
"step": 1315
},
{
"epoch": 0.34,
"grad_norm": 0.27086869426476023,
"learning_rate": 0.00016727607711647114,
"loss": 0.6897,
"step": 1320
},
{
"epoch": 0.34,
"grad_norm": 0.22378657478817104,
"learning_rate": 0.00016694411257988135,
"loss": 0.6582,
"step": 1325
},
{
"epoch": 0.34,
"grad_norm": 0.2277896285249334,
"learning_rate": 0.00016661080579689132,
"loss": 0.7067,
"step": 1330
},
{
"epoch": 0.34,
"grad_norm": 0.24128076526967204,
"learning_rate": 0.00016627616345038642,
"loss": 0.6632,
"step": 1335
},
{
"epoch": 0.34,
"grad_norm": 0.425397202711577,
"learning_rate": 0.0001659401922500304,
"loss": 0.7114,
"step": 1340
},
{
"epoch": 0.35,
"grad_norm": 0.2401567805733719,
"learning_rate": 0.0001656028989321309,
"loss": 0.6438,
"step": 1345
},
{
"epoch": 0.35,
"grad_norm": 0.2153720205033084,
"learning_rate": 0.00016526429025950424,
"loss": 0.6838,
"step": 1350
},
{
"epoch": 0.35,
"grad_norm": 0.22736402630025615,
"learning_rate": 0.00016492437302134008,
"loss": 0.7221,
"step": 1355
},
{
"epoch": 0.35,
"grad_norm": 0.23982260160339397,
"learning_rate": 0.00016458315403306502,
"loss": 0.7908,
"step": 1360
},
{
"epoch": 0.35,
"grad_norm": 0.27541369809262867,
"learning_rate": 0.00016424064013620615,
"loss": 0.6968,
"step": 1365
},
{
"epoch": 0.35,
"grad_norm": 0.2658031068926381,
"learning_rate": 0.0001638968381982538,
"loss": 0.6988,
"step": 1370
},
{
"epoch": 0.35,
"grad_norm": 0.22486989496520882,
"learning_rate": 0.0001635517551125238,
"loss": 0.81,
"step": 1375
},
{
"epoch": 0.35,
"grad_norm": 0.30674338452959843,
"learning_rate": 0.0001632053977980194,
"loss": 0.7256,
"step": 1380
},
{
"epoch": 0.36,
"grad_norm": 0.30285588055303475,
"learning_rate": 0.0001628577731992924,
"loss": 0.7188,
"step": 1385
},
{
"epoch": 0.36,
"grad_norm": 0.36884764888792976,
"learning_rate": 0.000162508888286304,
"loss": 0.7081,
"step": 1390
},
{
"epoch": 0.36,
"grad_norm": 0.2400495764148457,
"learning_rate": 0.00016215875005428499,
"loss": 0.7678,
"step": 1395
},
{
"epoch": 0.36,
"grad_norm": 0.20873038269420108,
"learning_rate": 0.00016180736552359553,
"loss": 0.7497,
"step": 1400
},
{
"epoch": 0.36,
"grad_norm": 0.2457699326637798,
"learning_rate": 0.0001614547417395844,
"loss": 0.7734,
"step": 1405
},
{
"epoch": 0.36,
"grad_norm": 0.31218574475793187,
"learning_rate": 0.00016110088577244773,
"loss": 0.7162,
"step": 1410
},
{
"epoch": 0.36,
"grad_norm": 0.27454082866628093,
"learning_rate": 0.0001607458047170872,
"loss": 0.745,
"step": 1415
},
{
"epoch": 0.36,
"grad_norm": 0.28758887844401904,
"learning_rate": 0.00016038950569296785,
"loss": 0.6231,
"step": 1420
},
{
"epoch": 0.37,
"grad_norm": 0.21604728751635988,
"learning_rate": 0.00016003199584397528,
"loss": 0.8008,
"step": 1425
},
{
"epoch": 0.37,
"grad_norm": 0.23310840527224116,
"learning_rate": 0.00015967328233827249,
"loss": 0.6987,
"step": 1430
},
{
"epoch": 0.37,
"grad_norm": 0.2568649534565718,
"learning_rate": 0.000159313372368156,
"loss": 0.6466,
"step": 1435
},
{
"epoch": 0.37,
"grad_norm": 0.32113279474719314,
"learning_rate": 0.00015895227314991178,
"loss": 0.6838,
"step": 1440
},
{
"epoch": 0.37,
"grad_norm": 0.296258785686713,
"learning_rate": 0.0001585899919236706,
"loss": 0.6935,
"step": 1445
},
{
"epoch": 0.37,
"grad_norm": 0.23501470975504843,
"learning_rate": 0.00015822653595326275,
"loss": 0.7261,
"step": 1450
},
{
"epoch": 0.37,
"grad_norm": 0.2783256304635712,
"learning_rate": 0.0001578619125260724,
"loss": 0.7512,
"step": 1455
},
{
"epoch": 0.37,
"grad_norm": 0.268618761910228,
"learning_rate": 0.00015749612895289152,
"loss": 0.6482,
"step": 1460
},
{
"epoch": 0.38,
"grad_norm": 0.21978849504732878,
"learning_rate": 0.00015712919256777335,
"loss": 0.7327,
"step": 1465
},
{
"epoch": 0.38,
"grad_norm": 0.23357753091278965,
"learning_rate": 0.00015676111072788527,
"loss": 0.681,
"step": 1470
},
{
"epoch": 0.38,
"grad_norm": 0.271489123229419,
"learning_rate": 0.0001563918908133614,
"loss": 0.7193,
"step": 1475
},
{
"epoch": 0.38,
"grad_norm": 0.31414437765976133,
"learning_rate": 0.00015602154022715435,
"loss": 0.7501,
"step": 1480
},
{
"epoch": 0.38,
"grad_norm": 0.27016203325071253,
"learning_rate": 0.00015565006639488722,
"loss": 0.6972,
"step": 1485
},
{
"epoch": 0.38,
"grad_norm": 0.22532821311440346,
"learning_rate": 0.0001552774767647043,
"loss": 0.6922,
"step": 1490
},
{
"epoch": 0.38,
"grad_norm": 0.27403456862921993,
"learning_rate": 0.000154903778807122,
"loss": 0.6949,
"step": 1495
},
{
"epoch": 0.38,
"grad_norm": 0.27739513562892554,
"learning_rate": 0.0001545289800148789,
"loss": 0.7243,
"step": 1500
},
{
"epoch": 0.39,
"grad_norm": 0.2238572112501968,
"learning_rate": 0.00015415308790278572,
"loss": 0.7353,
"step": 1505
},
{
"epoch": 0.39,
"grad_norm": 0.24964811413280275,
"learning_rate": 0.0001537761100075744,
"loss": 0.7345,
"step": 1510
},
{
"epoch": 0.39,
"grad_norm": 0.31756535201065944,
"learning_rate": 0.00015339805388774714,
"loss": 0.7856,
"step": 1515
},
{
"epoch": 0.39,
"grad_norm": 0.23296901823074231,
"learning_rate": 0.00015301892712342482,
"loss": 0.729,
"step": 1520
},
{
"epoch": 0.39,
"grad_norm": 0.2352427696844548,
"learning_rate": 0.00015263873731619508,
"loss": 0.7662,
"step": 1525
},
{
"epoch": 0.39,
"grad_norm": 0.2761737699751625,
"learning_rate": 0.00015225749208895968,
"loss": 0.7204,
"step": 1530
},
{
"epoch": 0.39,
"grad_norm": 0.3417520503815876,
"learning_rate": 0.000151875199085782,
"loss": 0.7737,
"step": 1535
},
{
"epoch": 0.4,
"grad_norm": 0.2668391012505777,
"learning_rate": 0.0001514918659717335,
"loss": 0.7105,
"step": 1540
},
{
"epoch": 0.4,
"grad_norm": 0.30187672878563326,
"learning_rate": 0.00015110750043274008,
"loss": 0.7326,
"step": 1545
},
{
"epoch": 0.4,
"grad_norm": 0.2242593810622975,
"learning_rate": 0.00015072211017542813,
"loss": 0.7472,
"step": 1550
},
{
"epoch": 0.4,
"grad_norm": 0.29972249141011964,
"learning_rate": 0.0001503357029269698,
"loss": 0.6271,
"step": 1555
},
{
"epoch": 0.4,
"grad_norm": 0.26475447856915474,
"learning_rate": 0.00014994828643492827,
"loss": 0.7354,
"step": 1560
},
{
"epoch": 0.4,
"grad_norm": 0.26450457112724596,
"learning_rate": 0.00014955986846710222,
"loss": 0.7264,
"step": 1565
},
{
"epoch": 0.4,
"grad_norm": 0.3219325804352173,
"learning_rate": 0.00014917045681137026,
"loss": 0.689,
"step": 1570
},
{
"epoch": 0.4,
"grad_norm": 0.26984976027685303,
"learning_rate": 0.00014878005927553456,
"loss": 0.6566,
"step": 1575
},
{
"epoch": 0.41,
"grad_norm": 0.31512358131320956,
"learning_rate": 0.0001483886836871646,
"loss": 0.7099,
"step": 1580
},
{
"epoch": 0.41,
"grad_norm": 0.27177868856475285,
"learning_rate": 0.00014799633789343994,
"loss": 0.7864,
"step": 1585
},
{
"epoch": 0.41,
"grad_norm": 0.2905196107184957,
"learning_rate": 0.00014760302976099304,
"loss": 0.6429,
"step": 1590
},
{
"epoch": 0.41,
"grad_norm": 0.2429940903609081,
"learning_rate": 0.00014720876717575155,
"loss": 0.6899,
"step": 1595
},
{
"epoch": 0.41,
"grad_norm": 0.29122627651097244,
"learning_rate": 0.00014681355804278001,
"loss": 0.6956,
"step": 1600
},
{
"epoch": 0.41,
"grad_norm": 0.214312815706962,
"learning_rate": 0.00014641741028612162,
"loss": 0.641,
"step": 1605
},
{
"epoch": 0.41,
"grad_norm": 0.1858785757380131,
"learning_rate": 0.00014602033184863913,
"loss": 0.7175,
"step": 1610
},
{
"epoch": 0.41,
"grad_norm": 0.2755921853014141,
"learning_rate": 0.00014562233069185572,
"loss": 0.8379,
"step": 1615
},
{
"epoch": 0.42,
"grad_norm": 0.25002568814104187,
"learning_rate": 0.00014522341479579533,
"loss": 0.7584,
"step": 1620
},
{
"epoch": 0.42,
"grad_norm": 0.257559638932229,
"learning_rate": 0.0001448235921588226,
"loss": 0.7739,
"step": 1625
},
{
"epoch": 0.42,
"grad_norm": 0.20960264146716834,
"learning_rate": 0.00014442287079748263,
"loss": 0.7014,
"step": 1630
},
{
"epoch": 0.42,
"grad_norm": 0.2521461373853711,
"learning_rate": 0.00014402125874634012,
"loss": 0.6859,
"step": 1635
},
{
"epoch": 0.42,
"grad_norm": 0.249737216302013,
"learning_rate": 0.00014361876405781832,
"loss": 0.7109,
"step": 1640
},
{
"epoch": 0.42,
"grad_norm": 0.2126565997638845,
"learning_rate": 0.00014321539480203764,
"loss": 0.7759,
"step": 1645
},
{
"epoch": 0.42,
"grad_norm": 0.32895767720771935,
"learning_rate": 0.00014281115906665374,
"loss": 0.7,
"step": 1650
},
{
"epoch": 0.42,
"grad_norm": 0.24271085601266817,
"learning_rate": 0.00014240606495669538,
"loss": 0.6983,
"step": 1655
},
{
"epoch": 0.43,
"grad_norm": 0.3030757588062575,
"learning_rate": 0.00014200012059440207,
"loss": 0.6991,
"step": 1660
},
{
"epoch": 0.43,
"grad_norm": 0.37464984671186286,
"learning_rate": 0.00014159333411906095,
"loss": 0.7526,
"step": 1665
},
{
"epoch": 0.43,
"grad_norm": 0.23935406631920386,
"learning_rate": 0.00014118571368684383,
"loss": 0.7923,
"step": 1670
},
{
"epoch": 0.43,
"grad_norm": 0.296146472400592,
"learning_rate": 0.00014077726747064353,
"loss": 0.6202,
"step": 1675
},
{
"epoch": 0.43,
"grad_norm": 0.23565535010314678,
"learning_rate": 0.00014036800365991008,
"loss": 0.6796,
"step": 1680
},
{
"epoch": 0.43,
"grad_norm": 0.2660771376040614,
"learning_rate": 0.00013995793046048643,
"loss": 0.6981,
"step": 1685
},
{
"epoch": 0.43,
"grad_norm": 0.264009781126536,
"learning_rate": 0.00013954705609444404,
"loss": 0.6901,
"step": 1690
},
{
"epoch": 0.43,
"grad_norm": 0.28853297225085933,
"learning_rate": 0.0001391353887999179,
"loss": 0.7664,
"step": 1695
},
{
"epoch": 0.44,
"grad_norm": 0.24731691813618162,
"learning_rate": 0.00013872293683094152,
"loss": 0.6689,
"step": 1700
},
{
"epoch": 0.44,
"grad_norm": 0.2594833661555708,
"learning_rate": 0.00013830970845728126,
"loss": 0.7701,
"step": 1705
},
{
"epoch": 0.44,
"grad_norm": 0.20644246333852043,
"learning_rate": 0.00013789571196427055,
"loss": 0.6513,
"step": 1710
},
{
"epoch": 0.44,
"grad_norm": 0.2339458014501443,
"learning_rate": 0.00013748095565264384,
"loss": 0.6494,
"step": 1715
},
{
"epoch": 0.44,
"grad_norm": 0.2961333190555473,
"learning_rate": 0.00013706544783837022,
"loss": 0.7054,
"step": 1720
},
{
"epoch": 0.44,
"grad_norm": 0.27743496626408143,
"learning_rate": 0.00013664919685248649,
"loss": 0.7295,
"step": 1725
},
{
"epoch": 0.44,
"grad_norm": 0.27257332981945664,
"learning_rate": 0.00013623221104093025,
"loss": 0.8299,
"step": 1730
},
{
"epoch": 0.45,
"grad_norm": 0.2592802724543621,
"learning_rate": 0.0001358144987643726,
"loss": 0.7277,
"step": 1735
},
{
"epoch": 0.45,
"grad_norm": 0.2238450594222949,
"learning_rate": 0.00013539606839805036,
"loss": 0.7238,
"step": 1740
},
{
"epoch": 0.45,
"grad_norm": 0.2735424814559735,
"learning_rate": 0.00013497692833159828,
"loss": 0.7015,
"step": 1745
},
{
"epoch": 0.45,
"grad_norm": 0.2332151330743781,
"learning_rate": 0.00013455708696888085,
"loss": 0.6969,
"step": 1750
},
{
"epoch": 0.45,
"grad_norm": 0.23230692211371998,
"learning_rate": 0.00013413655272782356,
"loss": 0.7266,
"step": 1755
},
{
"epoch": 0.45,
"grad_norm": 0.23959034038420193,
"learning_rate": 0.00013371533404024438,
"loss": 0.7636,
"step": 1760
},
{
"epoch": 0.45,
"grad_norm": 0.25479290452971226,
"learning_rate": 0.00013329343935168466,
"loss": 0.7075,
"step": 1765
},
{
"epoch": 0.45,
"grad_norm": 0.23415173017919827,
"learning_rate": 0.00013287087712123962,
"loss": 0.7747,
"step": 1770
},
{
"epoch": 0.46,
"grad_norm": 0.2286511684566787,
"learning_rate": 0.000132447655821389,
"loss": 0.6613,
"step": 1775
},
{
"epoch": 0.46,
"grad_norm": 0.2976728303873654,
"learning_rate": 0.00013202378393782692,
"loss": 0.8001,
"step": 1780
},
{
"epoch": 0.46,
"grad_norm": 0.26302357314609043,
"learning_rate": 0.00013159926996929192,
"loss": 0.6637,
"step": 1785
},
{
"epoch": 0.46,
"grad_norm": 0.25180945871071514,
"learning_rate": 0.00013117412242739655,
"loss": 0.6515,
"step": 1790
},
{
"epoch": 0.46,
"grad_norm": 0.22843461927288725,
"learning_rate": 0.00013074834983645657,
"loss": 0.7534,
"step": 1795
},
{
"epoch": 0.46,
"grad_norm": 0.2746800608584564,
"learning_rate": 0.00013032196073332027,
"loss": 0.7448,
"step": 1800
},
{
"epoch": 0.46,
"grad_norm": 0.23358914777160025,
"learning_rate": 0.000129894963667197,
"loss": 0.7462,
"step": 1805
},
{
"epoch": 0.46,
"grad_norm": 0.21563694009001097,
"learning_rate": 0.00012946736719948607,
"loss": 0.7174,
"step": 1810
},
{
"epoch": 0.47,
"grad_norm": 0.25466538845306436,
"learning_rate": 0.00012903917990360485,
"loss": 0.6958,
"step": 1815
},
{
"epoch": 0.47,
"grad_norm": 0.23650763866157107,
"learning_rate": 0.000128610410364817,
"loss": 0.6948,
"step": 1820
},
{
"epoch": 0.47,
"grad_norm": 0.32664640496556885,
"learning_rate": 0.00012818106718006023,
"loss": 0.6132,
"step": 1825
},
{
"epoch": 0.47,
"grad_norm": 0.2195897657124799,
"learning_rate": 0.00012775115895777417,
"loss": 0.7536,
"step": 1830
},
{
"epoch": 0.47,
"grad_norm": 0.23682325415599154,
"learning_rate": 0.0001273206943177274,
"loss": 0.6292,
"step": 1835
},
{
"epoch": 0.47,
"grad_norm": 0.2708829289481542,
"learning_rate": 0.00012688968189084493,
"loss": 0.7273,
"step": 1840
},
{
"epoch": 0.47,
"grad_norm": 0.26405392878632394,
"learning_rate": 0.000126458130319035,
"loss": 0.7254,
"step": 1845
},
{
"epoch": 0.47,
"grad_norm": 0.2710846037228565,
"learning_rate": 0.00012602604825501587,
"loss": 0.7051,
"step": 1850
},
{
"epoch": 0.48,
"grad_norm": 0.18625479643676604,
"learning_rate": 0.00012559344436214223,
"loss": 0.634,
"step": 1855
},
{
"epoch": 0.48,
"grad_norm": 0.23174335299007431,
"learning_rate": 0.00012516032731423165,
"loss": 0.721,
"step": 1860
},
{
"epoch": 0.48,
"grad_norm": 0.2674954048671644,
"learning_rate": 0.00012472670579539055,
"loss": 0.7039,
"step": 1865
},
{
"epoch": 0.48,
"grad_norm": 0.23350249784430052,
"learning_rate": 0.00012429258849984014,
"loss": 0.7119,
"step": 1870
},
{
"epoch": 0.48,
"grad_norm": 0.2307966491739631,
"learning_rate": 0.00012385798413174206,
"loss": 0.7267,
"step": 1875
},
{
"epoch": 0.48,
"grad_norm": 0.2587108210616084,
"learning_rate": 0.00012342290140502388,
"loss": 0.7738,
"step": 1880
},
{
"epoch": 0.48,
"grad_norm": 0.22704947603538148,
"learning_rate": 0.00012298734904320438,
"loss": 0.6771,
"step": 1885
},
{
"epoch": 0.48,
"grad_norm": 0.24005242145344702,
"learning_rate": 0.00012255133577921868,
"loss": 0.7774,
"step": 1890
},
{
"epoch": 0.49,
"grad_norm": 0.21236521326266408,
"learning_rate": 0.00012211487035524305,
"loss": 0.689,
"step": 1895
},
{
"epoch": 0.49,
"grad_norm": 0.26548226590256707,
"learning_rate": 0.0001216779615225197,
"loss": 0.6958,
"step": 1900
},
{
"epoch": 0.49,
"grad_norm": 0.28092405712871943,
"learning_rate": 0.00012124061804118137,
"loss": 0.7061,
"step": 1905
},
{
"epoch": 0.49,
"grad_norm": 0.24730560435280138,
"learning_rate": 0.00012080284868007541,
"loss": 0.6964,
"step": 1910
},
{
"epoch": 0.49,
"grad_norm": 0.1991516776169903,
"learning_rate": 0.00012036466221658847,
"loss": 0.72,
"step": 1915
},
{
"epoch": 0.49,
"grad_norm": 0.29645767275685886,
"learning_rate": 0.0001199260674364699,
"loss": 0.7778,
"step": 1920
},
{
"epoch": 0.49,
"grad_norm": 0.25070174095345915,
"learning_rate": 0.00011948707313365614,
"loss": 0.8114,
"step": 1925
},
{
"epoch": 0.5,
"grad_norm": 0.27676560069954154,
"learning_rate": 0.00011904768811009405,
"loss": 0.7192,
"step": 1930
},
{
"epoch": 0.5,
"grad_norm": 0.3078376640692026,
"learning_rate": 0.00011860792117556454,
"loss": 0.709,
"step": 1935
},
{
"epoch": 0.5,
"grad_norm": 0.2049781937326678,
"learning_rate": 0.00011816778114750593,
"loss": 0.7672,
"step": 1940
},
{
"epoch": 0.5,
"grad_norm": 0.21305453207892233,
"learning_rate": 0.00011772727685083724,
"loss": 0.7744,
"step": 1945
},
{
"epoch": 0.5,
"grad_norm": 0.24046462080108547,
"learning_rate": 0.00011728641711778103,
"loss": 0.7229,
"step": 1950
},
{
"epoch": 0.5,
"grad_norm": 0.21072752361715988,
"learning_rate": 0.00011684521078768658,
"loss": 0.6853,
"step": 1955
},
{
"epoch": 0.5,
"grad_norm": 0.17170386164857987,
"learning_rate": 0.00011640366670685248,
"loss": 0.6348,
"step": 1960
},
{
"epoch": 0.5,
"grad_norm": 0.27104402407445916,
"learning_rate": 0.00011596179372834924,
"loss": 0.8441,
"step": 1965
},
{
"epoch": 0.51,
"grad_norm": 0.24600997987590698,
"learning_rate": 0.00011551960071184195,
"loss": 0.8251,
"step": 1970
},
{
"epoch": 0.51,
"grad_norm": 0.22967416303582927,
"learning_rate": 0.00011507709652341256,
"loss": 0.716,
"step": 1975
},
{
"epoch": 0.51,
"grad_norm": 0.24788639781264693,
"learning_rate": 0.00011463429003538196,
"loss": 0.7307,
"step": 1980
},
{
"epoch": 0.51,
"grad_norm": 0.25315558172885294,
"learning_rate": 0.00011419119012613233,
"loss": 0.7196,
"step": 1985
},
{
"epoch": 0.51,
"grad_norm": 0.24182617769996623,
"learning_rate": 0.000113747805679929,
"loss": 0.8012,
"step": 1990
},
{
"epoch": 0.51,
"grad_norm": 0.2500833847981834,
"learning_rate": 0.00011330414558674234,
"loss": 0.7113,
"step": 1995
},
{
"epoch": 0.51,
"grad_norm": 0.3651534404891642,
"learning_rate": 0.00011286021874206952,
"loss": 0.6693,
"step": 2000
},
{
"epoch": 0.51,
"grad_norm": 0.3146361713006211,
"learning_rate": 0.00011241603404675609,
"loss": 0.7867,
"step": 2005
},
{
"epoch": 0.52,
"grad_norm": 0.2994271238230683,
"learning_rate": 0.00011197160040681762,
"loss": 0.8188,
"step": 2010
},
{
"epoch": 0.52,
"grad_norm": 0.2287392648112236,
"learning_rate": 0.00011152692673326111,
"loss": 0.7414,
"step": 2015
},
{
"epoch": 0.52,
"grad_norm": 0.21561987456802226,
"learning_rate": 0.0001110820219419062,
"loss": 0.743,
"step": 2020
},
{
"epoch": 0.52,
"grad_norm": 0.26286058284503494,
"learning_rate": 0.00011063689495320661,
"loss": 0.709,
"step": 2025
},
{
"epoch": 0.52,
"grad_norm": 0.23831373981952758,
"learning_rate": 0.0001101915546920711,
"loss": 0.7487,
"step": 2030
},
{
"epoch": 0.52,
"grad_norm": 0.4169631396437136,
"learning_rate": 0.00010974601008768461,
"loss": 0.6595,
"step": 2035
},
{
"epoch": 0.52,
"grad_norm": 0.23258437702962542,
"learning_rate": 0.00010930027007332923,
"loss": 0.6659,
"step": 2040
},
{
"epoch": 0.52,
"grad_norm": 0.23047360031102437,
"learning_rate": 0.00010885434358620507,
"loss": 0.7374,
"step": 2045
},
{
"epoch": 0.53,
"grad_norm": 0.2586140902144001,
"learning_rate": 0.00010840823956725103,
"loss": 0.7431,
"step": 2050
},
{
"epoch": 0.53,
"grad_norm": 0.2700901236204115,
"learning_rate": 0.00010796196696096564,
"loss": 0.7076,
"step": 2055
},
{
"epoch": 0.53,
"grad_norm": 0.20102755645586845,
"learning_rate": 0.00010751553471522757,
"loss": 0.7332,
"step": 2060
},
{
"epoch": 0.53,
"grad_norm": 0.25286502232358293,
"learning_rate": 0.00010706895178111634,
"loss": 0.7297,
"step": 2065
},
{
"epoch": 0.53,
"grad_norm": 0.2664571423670369,
"learning_rate": 0.00010662222711273279,
"loss": 0.6364,
"step": 2070
},
{
"epoch": 0.53,
"grad_norm": 0.22487420863698854,
"learning_rate": 0.00010617536966701962,
"loss": 0.7288,
"step": 2075
},
{
"epoch": 0.53,
"grad_norm": 0.24184706211280177,
"learning_rate": 0.00010572838840358168,
"loss": 0.6334,
"step": 2080
},
{
"epoch": 0.53,
"grad_norm": 0.26180250015553935,
"learning_rate": 0.00010528129228450645,
"loss": 0.6579,
"step": 2085
},
{
"epoch": 0.54,
"grad_norm": 0.24611552837710818,
"learning_rate": 0.00010483409027418425,
"loss": 0.7506,
"step": 2090
},
{
"epoch": 0.54,
"grad_norm": 0.22771528096688773,
"learning_rate": 0.00010438679133912858,
"loss": 0.6517,
"step": 2095
},
{
"epoch": 0.54,
"grad_norm": 0.25425984413622493,
"learning_rate": 0.00010393940444779635,
"loss": 0.6997,
"step": 2100
},
{
"epoch": 0.54,
"grad_norm": 0.25948687177354873,
"learning_rate": 0.00010349193857040787,
"loss": 0.7736,
"step": 2105
},
{
"epoch": 0.54,
"grad_norm": 0.29503492872852033,
"learning_rate": 0.00010304440267876727,
"loss": 0.7855,
"step": 2110
},
{
"epoch": 0.54,
"grad_norm": 0.21574528692226247,
"learning_rate": 0.00010259680574608248,
"loss": 0.765,
"step": 2115
},
{
"epoch": 0.54,
"grad_norm": 0.2530425727116828,
"learning_rate": 0.00010214915674678523,
"loss": 0.7705,
"step": 2120
},
{
"epoch": 0.55,
"grad_norm": 0.2355690254358619,
"learning_rate": 0.00010170146465635126,
"loss": 0.7785,
"step": 2125
},
{
"epoch": 0.55,
"grad_norm": 0.2393846506779741,
"learning_rate": 0.00010125373845112034,
"loss": 0.7422,
"step": 2130
},
{
"epoch": 0.55,
"grad_norm": 0.2262608510161545,
"learning_rate": 0.00010080598710811615,
"loss": 0.6922,
"step": 2135
},
{
"epoch": 0.55,
"grad_norm": 0.3256276878966541,
"learning_rate": 0.00010035821960486643,
"loss": 0.7713,
"step": 2140
},
{
"epoch": 0.55,
"grad_norm": 0.2496398007237336,
"learning_rate": 9.9910444919223e-05,
"loss": 0.7058,
"step": 2145
},
{
"epoch": 0.55,
"grad_norm": 0.29531137204475016,
"learning_rate": 9.946267202918157e-05,
"loss": 0.6994,
"step": 2150
},
{
"epoch": 0.55,
"grad_norm": 0.2596775947550526,
"learning_rate": 9.901490991270201e-05,
"loss": 0.6993,
"step": 2155
},
{
"epoch": 0.55,
"grad_norm": 0.5102446180121213,
"learning_rate": 9.856716754752796e-05,
"loss": 0.6543,
"step": 2160
},
{
"epoch": 0.56,
"grad_norm": 0.2754247927387255,
"learning_rate": 9.811945391100724e-05,
"loss": 0.7366,
"step": 2165
},
{
"epoch": 0.56,
"grad_norm": 0.23454396466279584,
"learning_rate": 9.767177797991155e-05,
"loss": 0.7441,
"step": 2170
},
{
"epoch": 0.56,
"grad_norm": 0.22647788058437757,
"learning_rate": 9.722414873025664e-05,
"loss": 0.6849,
"step": 2175
},
{
"epoch": 0.56,
"grad_norm": 0.2210584391132073,
"learning_rate": 9.677657513712221e-05,
"loss": 0.727,
"step": 2180
},
{
"epoch": 0.56,
"grad_norm": 0.24884640250686244,
"learning_rate": 9.632906617447212e-05,
"loss": 0.6939,
"step": 2185
},
{
"epoch": 0.56,
"grad_norm": 0.2189510126025836,
"learning_rate": 9.588163081497427e-05,
"loss": 0.7871,
"step": 2190
},
{
"epoch": 0.56,
"grad_norm": 0.25847268142498553,
"learning_rate": 9.543427802982095e-05,
"loss": 0.765,
"step": 2195
},
{
"epoch": 0.56,
"grad_norm": 0.2887583084675781,
"learning_rate": 9.498701678854865e-05,
"loss": 0.6855,
"step": 2200
},
{
"epoch": 0.57,
"grad_norm": 0.21832636439823108,
"learning_rate": 9.453985605885851e-05,
"loss": 0.732,
"step": 2205
},
{
"epoch": 0.57,
"grad_norm": 0.25875988987504445,
"learning_rate": 9.409280480643628e-05,
"loss": 0.776,
"step": 2210
},
{
"epoch": 0.57,
"grad_norm": 0.2937585368281696,
"learning_rate": 9.364587199477276e-05,
"loss": 0.7234,
"step": 2215
},
{
"epoch": 0.57,
"grad_norm": 0.21327899166987224,
"learning_rate": 9.319906658498389e-05,
"loss": 0.7519,
"step": 2220
},
{
"epoch": 0.57,
"grad_norm": 0.24464572895636677,
"learning_rate": 9.275239753563128e-05,
"loss": 0.7196,
"step": 2225
},
{
"epoch": 0.57,
"grad_norm": 0.22386709524202114,
"learning_rate": 9.230587380254237e-05,
"loss": 0.6901,
"step": 2230
},
{
"epoch": 0.57,
"grad_norm": 0.23753591042587965,
"learning_rate": 9.185950433863107e-05,
"loss": 0.6676,
"step": 2235
},
{
"epoch": 0.57,
"grad_norm": 0.17549702993152236,
"learning_rate": 9.141329809371803e-05,
"loss": 0.7646,
"step": 2240
},
{
"epoch": 0.58,
"grad_norm": 0.23044355517418907,
"learning_rate": 9.096726401435146e-05,
"loss": 0.7121,
"step": 2245
},
{
"epoch": 0.58,
"grad_norm": 0.2037402971541335,
"learning_rate": 9.052141104362748e-05,
"loss": 0.6541,
"step": 2250
},
{
"epoch": 0.58,
"grad_norm": 0.2173788729813436,
"learning_rate": 9.007574812101107e-05,
"loss": 0.7522,
"step": 2255
},
{
"epoch": 0.58,
"grad_norm": 0.24289241576799026,
"learning_rate": 8.963028418215653e-05,
"loss": 0.8154,
"step": 2260
},
{
"epoch": 0.58,
"grad_norm": 0.24110107744789394,
"learning_rate": 8.918502815872865e-05,
"loss": 0.7512,
"step": 2265
},
{
"epoch": 0.58,
"grad_norm": 0.2505142119875894,
"learning_rate": 8.873998897822336e-05,
"loss": 0.7519,
"step": 2270
},
{
"epoch": 0.58,
"grad_norm": 0.2603773293471231,
"learning_rate": 8.829517556378888e-05,
"loss": 0.6786,
"step": 2275
},
{
"epoch": 0.58,
"grad_norm": 0.30552669526483317,
"learning_rate": 8.785059683404672e-05,
"loss": 0.7016,
"step": 2280
},
{
"epoch": 0.59,
"grad_norm": 0.30878912939747605,
"learning_rate": 8.740626170291297e-05,
"loss": 0.7535,
"step": 2285
},
{
"epoch": 0.59,
"grad_norm": 0.28441949312000736,
"learning_rate": 8.696217907941941e-05,
"loss": 0.7209,
"step": 2290
},
{
"epoch": 0.59,
"grad_norm": 0.25705307392892374,
"learning_rate": 8.65183578675351e-05,
"loss": 0.7042,
"step": 2295
},
{
"epoch": 0.59,
"grad_norm": 0.2575869631501832,
"learning_rate": 8.607480696598762e-05,
"loss": 0.6118,
"step": 2300
},
{
"epoch": 0.59,
"grad_norm": 0.3327254376840343,
"learning_rate": 8.563153526808484e-05,
"loss": 0.7738,
"step": 2305
},
{
"epoch": 0.59,
"grad_norm": 0.24348419071615154,
"learning_rate": 8.518855166153644e-05,
"loss": 0.6924,
"step": 2310
},
{
"epoch": 0.59,
"grad_norm": 0.3469772616735954,
"learning_rate": 8.474586502827591e-05,
"loss": 0.7001,
"step": 2315
},
{
"epoch": 0.6,
"grad_norm": 0.19770150428519379,
"learning_rate": 8.43034842442822e-05,
"loss": 0.6706,
"step": 2320
},
{
"epoch": 0.6,
"grad_norm": 0.4391168084334309,
"learning_rate": 8.386141817940213e-05,
"loss": 0.701,
"step": 2325
},
{
"epoch": 0.6,
"grad_norm": 0.24046204644423808,
"learning_rate": 8.341967569717202e-05,
"loss": 0.6511,
"step": 2330
},
{
"epoch": 0.6,
"grad_norm": 0.24121857059524954,
"learning_rate": 8.297826565464057e-05,
"loss": 0.637,
"step": 2335
},
{
"epoch": 0.6,
"grad_norm": 0.289807786292161,
"learning_rate": 8.253719690219079e-05,
"loss": 0.7298,
"step": 2340
},
{
"epoch": 0.6,
"grad_norm": 0.2675093976169691,
"learning_rate": 8.20964782833628e-05,
"loss": 0.6229,
"step": 2345
},
{
"epoch": 0.6,
"grad_norm": 0.32873199218989857,
"learning_rate": 8.165611863467644e-05,
"loss": 0.7254,
"step": 2350
},
{
"epoch": 0.6,
"grad_norm": 0.21463882575401755,
"learning_rate": 8.12161267854542e-05,
"loss": 0.6896,
"step": 2355
},
{
"epoch": 0.61,
"grad_norm": 0.20392542858991738,
"learning_rate": 8.077651155764387e-05,
"loss": 0.6524,
"step": 2360
},
{
"epoch": 0.61,
"grad_norm": 0.21818270130807613,
"learning_rate": 8.033728176564216e-05,
"loss": 0.7482,
"step": 2365
},
{
"epoch": 0.61,
"grad_norm": 0.3236950015797695,
"learning_rate": 7.98984462161175e-05,
"loss": 0.7463,
"step": 2370
},
{
"epoch": 0.61,
"grad_norm": 0.23726714785703115,
"learning_rate": 7.946001370783375e-05,
"loss": 0.7916,
"step": 2375
},
{
"epoch": 0.61,
"grad_norm": 0.2087887063125164,
"learning_rate": 7.902199303147363e-05,
"loss": 0.6875,
"step": 2380
},
{
"epoch": 0.61,
"grad_norm": 0.27106485088624854,
"learning_rate": 7.858439296946263e-05,
"loss": 0.7802,
"step": 2385
},
{
"epoch": 0.61,
"grad_norm": 0.2851539067095835,
"learning_rate": 7.814722229579264e-05,
"loss": 0.6769,
"step": 2390
},
{
"epoch": 0.61,
"grad_norm": 0.24800513231309682,
"learning_rate": 7.771048977584641e-05,
"loss": 0.6148,
"step": 2395
},
{
"epoch": 0.62,
"grad_norm": 0.30123103203277307,
"learning_rate": 7.727420416622144e-05,
"loss": 0.8299,
"step": 2400
},
{
"epoch": 0.62,
"grad_norm": 0.28336701564838623,
"learning_rate": 7.683837421455466e-05,
"loss": 0.6799,
"step": 2405
},
{
"epoch": 0.62,
"grad_norm": 0.24418965636330248,
"learning_rate": 7.640300865934687e-05,
"loss": 0.6924,
"step": 2410
},
{
"epoch": 0.62,
"grad_norm": 0.23926492236991967,
"learning_rate": 7.596811622978765e-05,
"loss": 0.7157,
"step": 2415
},
{
"epoch": 0.62,
"grad_norm": 0.2584954114418789,
"learning_rate": 7.553370564558032e-05,
"loss": 0.7652,
"step": 2420
},
{
"epoch": 0.62,
"grad_norm": 0.29773986167748034,
"learning_rate": 7.509978561676699e-05,
"loss": 0.7581,
"step": 2425
},
{
"epoch": 0.62,
"grad_norm": 0.2368205960863638,
"learning_rate": 7.46663648435541e-05,
"loss": 0.6941,
"step": 2430
},
{
"epoch": 0.62,
"grad_norm": 0.2839556861514916,
"learning_rate": 7.423345201613778e-05,
"loss": 0.677,
"step": 2435
},
{
"epoch": 0.63,
"grad_norm": 0.32721343862085367,
"learning_rate": 7.380105581452987e-05,
"loss": 0.7381,
"step": 2440
},
{
"epoch": 0.63,
"grad_norm": 0.2620811211111583,
"learning_rate": 7.336918490838356e-05,
"loss": 0.7808,
"step": 2445
},
{
"epoch": 0.63,
"grad_norm": 0.23444393543960068,
"learning_rate": 7.293784795681994e-05,
"loss": 0.7252,
"step": 2450
},
{
"epoch": 0.63,
"grad_norm": 0.2457300140925708,
"learning_rate": 7.250705360825392e-05,
"loss": 0.6721,
"step": 2455
},
{
"epoch": 0.63,
"grad_norm": 0.27505166245514767,
"learning_rate": 7.207681050022132e-05,
"loss": 0.7478,
"step": 2460
},
{
"epoch": 0.63,
"grad_norm": 0.2910558326018566,
"learning_rate": 7.16471272592053e-05,
"loss": 0.7808,
"step": 2465
},
{
"epoch": 0.63,
"grad_norm": 0.24862527602067405,
"learning_rate": 7.121801250046363e-05,
"loss": 0.7691,
"step": 2470
},
{
"epoch": 0.63,
"grad_norm": 0.1902191441389871,
"learning_rate": 7.078947482785576e-05,
"loss": 0.7097,
"step": 2475
},
{
"epoch": 0.64,
"grad_norm": 0.23472131108538954,
"learning_rate": 7.036152283367056e-05,
"loss": 0.6912,
"step": 2480
},
{
"epoch": 0.64,
"grad_norm": 0.2189915658804591,
"learning_rate": 6.993416509845376e-05,
"loss": 0.7211,
"step": 2485
},
{
"epoch": 0.64,
"grad_norm": 0.2603896298420689,
"learning_rate": 6.950741019083617e-05,
"loss": 0.7188,
"step": 2490
},
{
"epoch": 0.64,
"grad_norm": 0.2559070868163805,
"learning_rate": 6.908126666736165e-05,
"loss": 0.6925,
"step": 2495
},
{
"epoch": 0.64,
"grad_norm": 0.29850542028796345,
"learning_rate": 6.865574307231575e-05,
"loss": 0.7119,
"step": 2500
},
{
"epoch": 0.64,
"grad_norm": 0.19364511158581588,
"learning_rate": 6.823084793755417e-05,
"loss": 0.685,
"step": 2505
},
{
"epoch": 0.64,
"grad_norm": 0.3367238354963921,
"learning_rate": 6.780658978233199e-05,
"loss": 0.7398,
"step": 2510
},
{
"epoch": 0.65,
"grad_norm": 0.19047235797749595,
"learning_rate": 6.73829771131325e-05,
"loss": 0.7311,
"step": 2515
},
{
"epoch": 0.65,
"grad_norm": 0.3005689166877733,
"learning_rate": 6.696001842349702e-05,
"loss": 0.6747,
"step": 2520
},
{
"epoch": 0.65,
"grad_norm": 0.31211848702866324,
"learning_rate": 6.653772219385427e-05,
"loss": 0.759,
"step": 2525
},
{
"epoch": 0.65,
"grad_norm": 0.25625855783899115,
"learning_rate": 6.611609689135056e-05,
"loss": 0.7853,
"step": 2530
},
{
"epoch": 0.65,
"grad_norm": 0.23106616213334621,
"learning_rate": 6.569515096967991e-05,
"loss": 0.6437,
"step": 2535
},
{
"epoch": 0.65,
"grad_norm": 0.18916281943913288,
"learning_rate": 6.527489286891459e-05,
"loss": 0.5797,
"step": 2540
},
{
"epoch": 0.65,
"grad_norm": 0.21350268536847394,
"learning_rate": 6.485533101533583e-05,
"loss": 0.7013,
"step": 2545
},
{
"epoch": 0.65,
"grad_norm": 0.28763484217532065,
"learning_rate": 6.443647382126509e-05,
"loss": 0.7229,
"step": 2550
},
{
"epoch": 0.66,
"grad_norm": 0.2488901500729856,
"learning_rate": 6.401832968489501e-05,
"loss": 0.686,
"step": 2555
},
{
"epoch": 0.66,
"grad_norm": 0.20556797301856877,
"learning_rate": 6.360090699012145e-05,
"loss": 0.7435,
"step": 2560
},
{
"epoch": 0.66,
"grad_norm": 0.25971455100175256,
"learning_rate": 6.3184214106375e-05,
"loss": 0.6772,
"step": 2565
},
{
"epoch": 0.66,
"grad_norm": 0.2438737947426562,
"learning_rate": 6.27682593884535e-05,
"loss": 0.6351,
"step": 2570
},
{
"epoch": 0.66,
"grad_norm": 0.2239080056853126,
"learning_rate": 6.235305117635428e-05,
"loss": 0.7386,
"step": 2575
},
{
"epoch": 0.66,
"grad_norm": 0.2284401860369392,
"learning_rate": 6.193859779510712e-05,
"loss": 0.6939,
"step": 2580
},
{
"epoch": 0.66,
"grad_norm": 0.24456271598752877,
"learning_rate": 6.152490755460715e-05,
"loss": 0.6706,
"step": 2585
},
{
"epoch": 0.66,
"grad_norm": 0.22235388160369995,
"learning_rate": 6.111198874944845e-05,
"loss": 0.6903,
"step": 2590
},
{
"epoch": 0.67,
"grad_norm": 0.2144202608438619,
"learning_rate": 6.0699849658757545e-05,
"loss": 0.7528,
"step": 2595
},
{
"epoch": 0.67,
"grad_norm": 0.19956268282116124,
"learning_rate": 6.0288498546027536e-05,
"loss": 0.775,
"step": 2600
},
{
"epoch": 0.67,
"grad_norm": 0.22191365952846825,
"learning_rate": 5.987794365895229e-05,
"loss": 0.6759,
"step": 2605
},
{
"epoch": 0.67,
"grad_norm": 0.31000384496295275,
"learning_rate": 5.946819322926127e-05,
"loss": 0.6828,
"step": 2610
},
{
"epoch": 0.67,
"grad_norm": 0.27838526207253456,
"learning_rate": 5.9059255472554195e-05,
"loss": 0.6906,
"step": 2615
},
{
"epoch": 0.67,
"grad_norm": 0.2320736604730188,
"learning_rate": 5.865113858813673e-05,
"loss": 0.6665,
"step": 2620
},
{
"epoch": 0.67,
"grad_norm": 0.282228906687919,
"learning_rate": 5.824385075885557e-05,
"loss": 0.6861,
"step": 2625
},
{
"epoch": 0.67,
"grad_norm": 0.21066270814730526,
"learning_rate": 5.783740015093484e-05,
"loss": 0.7175,
"step": 2630
},
{
"epoch": 0.68,
"grad_norm": 0.2665172474983474,
"learning_rate": 5.743179491381204e-05,
"loss": 0.706,
"step": 2635
},
{
"epoch": 0.68,
"grad_norm": 0.2338869396395331,
"learning_rate": 5.702704317997492e-05,
"loss": 0.6729,
"step": 2640
},
{
"epoch": 0.68,
"grad_norm": 0.21891639712761243,
"learning_rate": 5.662315306479806e-05,
"loss": 0.7053,
"step": 2645
},
{
"epoch": 0.68,
"grad_norm": 0.22963997129503227,
"learning_rate": 5.6220132666380635e-05,
"loss": 0.624,
"step": 2650
},
{
"epoch": 0.68,
"grad_norm": 0.28440467945947445,
"learning_rate": 5.581799006538354e-05,
"loss": 0.7612,
"step": 2655
},
{
"epoch": 0.68,
"grad_norm": 0.29609689514958054,
"learning_rate": 5.541673332486773e-05,
"loss": 0.7444,
"step": 2660
},
{
"epoch": 0.68,
"grad_norm": 0.2650460281999568,
"learning_rate": 5.5016370490132364e-05,
"loss": 0.8484,
"step": 2665
},
{
"epoch": 0.68,
"grad_norm": 0.28809355156027594,
"learning_rate": 5.4616909588553674e-05,
"loss": 0.7871,
"step": 2670
},
{
"epoch": 0.69,
"grad_norm": 0.24812243233821338,
"learning_rate": 5.4218358629423794e-05,
"loss": 0.7977,
"step": 2675
},
{
"epoch": 0.69,
"grad_norm": 0.2920348572251461,
"learning_rate": 5.3820725603790346e-05,
"loss": 0.7394,
"step": 2680
},
{
"epoch": 0.69,
"grad_norm": 0.29200632111876645,
"learning_rate": 5.342401848429615e-05,
"loss": 0.778,
"step": 2685
},
{
"epoch": 0.69,
"grad_norm": 0.2432759115984862,
"learning_rate": 5.30282452250193e-05,
"loss": 0.7885,
"step": 2690
},
{
"epoch": 0.69,
"grad_norm": 0.2838656924088362,
"learning_rate": 5.263341376131382e-05,
"loss": 0.7023,
"step": 2695
},
{
"epoch": 0.69,
"grad_norm": 0.28056038476141837,
"learning_rate": 5.223953200965055e-05,
"loss": 0.7367,
"step": 2700
},
{
"epoch": 0.69,
"grad_norm": 0.22746358276548237,
"learning_rate": 5.1846607867458196e-05,
"loss": 0.8169,
"step": 2705
},
{
"epoch": 0.7,
"grad_norm": 0.2754771874933674,
"learning_rate": 5.145464921296537e-05,
"loss": 0.7762,
"step": 2710
},
{
"epoch": 0.7,
"grad_norm": 0.275520392192173,
"learning_rate": 5.1063663905042225e-05,
"loss": 0.6907,
"step": 2715
},
{
"epoch": 0.7,
"grad_norm": 0.2742206750054377,
"learning_rate": 5.067365978304315e-05,
"loss": 0.6962,
"step": 2720
},
{
"epoch": 0.7,
"grad_norm": 0.3043303805027072,
"learning_rate": 5.0284644666649485e-05,
"loss": 0.6863,
"step": 2725
},
{
"epoch": 0.7,
"grad_norm": 0.26287170160778367,
"learning_rate": 4.9896626355712805e-05,
"loss": 0.6837,
"step": 2730
},
{
"epoch": 0.7,
"grad_norm": 0.23968527436607373,
"learning_rate": 4.95096126300984e-05,
"loss": 0.6932,
"step": 2735
},
{
"epoch": 0.7,
"grad_norm": 0.19323389498485988,
"learning_rate": 4.912361124952948e-05,
"loss": 0.7342,
"step": 2740
},
{
"epoch": 0.7,
"grad_norm": 0.2216783500769926,
"learning_rate": 4.873862995343139e-05,
"loss": 0.6647,
"step": 2745
},
{
"epoch": 0.71,
"grad_norm": 0.26023716724912155,
"learning_rate": 4.835467646077656e-05,
"loss": 0.7357,
"step": 2750
},
{
"epoch": 0.71,
"grad_norm": 0.25899624407357125,
"learning_rate": 4.797175846992964e-05,
"loss": 0.6544,
"step": 2755
},
{
"epoch": 0.71,
"grad_norm": 0.18644116956534368,
"learning_rate": 4.7589883658493296e-05,
"loss": 0.7359,
"step": 2760
},
{
"epoch": 0.71,
"grad_norm": 0.24394521839077538,
"learning_rate": 4.72090596831542e-05,
"loss": 0.7304,
"step": 2765
},
{
"epoch": 0.71,
"grad_norm": 0.3406491454673553,
"learning_rate": 4.682929417952939e-05,
"loss": 0.6508,
"step": 2770
},
{
"epoch": 0.71,
"grad_norm": 0.2508477121094201,
"learning_rate": 4.645059476201333e-05,
"loss": 0.6859,
"step": 2775
},
{
"epoch": 0.71,
"grad_norm": 0.20268926784051644,
"learning_rate": 4.6072969023625165e-05,
"loss": 0.7103,
"step": 2780
},
{
"epoch": 0.71,
"grad_norm": 0.24062467799728976,
"learning_rate": 4.5696424535856574e-05,
"loss": 0.705,
"step": 2785
},
{
"epoch": 0.72,
"grad_norm": 0.26693143779686646,
"learning_rate": 4.532096884851978e-05,
"loss": 0.7319,
"step": 2790
},
{
"epoch": 0.72,
"grad_norm": 0.23409445770349235,
"learning_rate": 4.494660948959645e-05,
"loss": 0.791,
"step": 2795
},
{
"epoch": 0.72,
"grad_norm": 0.2845933689629394,
"learning_rate": 4.457335396508631e-05,
"loss": 0.6062,
"step": 2800
},
{
"epoch": 0.72,
"grad_norm": 0.24119287628680353,
"learning_rate": 4.420120975885723e-05,
"loss": 0.7077,
"step": 2805
},
{
"epoch": 0.72,
"grad_norm": 0.21991354406483815,
"learning_rate": 4.383018433249464e-05,
"loss": 0.6805,
"step": 2810
},
{
"epoch": 0.72,
"grad_norm": 0.2475445630279962,
"learning_rate": 4.346028512515232e-05,
"loss": 0.593,
"step": 2815
},
{
"epoch": 0.72,
"grad_norm": 0.24157860899387687,
"learning_rate": 4.309151955340297e-05,
"loss": 0.7195,
"step": 2820
},
{
"epoch": 0.72,
"grad_norm": 0.2123078817060314,
"learning_rate": 4.272389501108962e-05,
"loss": 0.7566,
"step": 2825
},
{
"epoch": 0.73,
"grad_norm": 0.33254744474875086,
"learning_rate": 4.2357418869177354e-05,
"loss": 0.7517,
"step": 2830
},
{
"epoch": 0.73,
"grad_norm": 0.2450146085609281,
"learning_rate": 4.1992098475605654e-05,
"loss": 0.7792,
"step": 2835
},
{
"epoch": 0.73,
"grad_norm": 0.2704781574806764,
"learning_rate": 4.162794115514078e-05,
"loss": 0.7238,
"step": 2840
},
{
"epoch": 0.73,
"grad_norm": 0.37954117128340364,
"learning_rate": 4.1264954209229254e-05,
"loss": 0.6556,
"step": 2845
},
{
"epoch": 0.73,
"grad_norm": 0.2860982586980259,
"learning_rate": 4.0903144915851174e-05,
"loss": 0.6506,
"step": 2850
},
{
"epoch": 0.73,
"grad_norm": 0.20667223525821882,
"learning_rate": 4.054252052937444e-05,
"loss": 0.6182,
"step": 2855
},
{
"epoch": 0.73,
"grad_norm": 0.2634554788563136,
"learning_rate": 4.018308828040924e-05,
"loss": 0.727,
"step": 2860
},
{
"epoch": 0.73,
"grad_norm": 0.2786956847780988,
"learning_rate": 3.982485537566321e-05,
"loss": 0.7262,
"step": 2865
},
{
"epoch": 0.74,
"grad_norm": 0.18437896768984685,
"learning_rate": 3.946782899779667e-05,
"loss": 0.6169,
"step": 2870
},
{
"epoch": 0.74,
"grad_norm": 0.2622931367541885,
"learning_rate": 3.911201630527894e-05,
"loss": 0.7385,
"step": 2875
},
{
"epoch": 0.74,
"grad_norm": 0.2748261196950285,
"learning_rate": 3.875742443224451e-05,
"loss": 0.6136,
"step": 2880
},
{
"epoch": 0.74,
"grad_norm": 0.28744844039364625,
"learning_rate": 3.84040604883502e-05,
"loss": 0.7284,
"step": 2885
},
{
"epoch": 0.74,
"grad_norm": 0.21857313698448572,
"learning_rate": 3.805193155863247e-05,
"loss": 0.6383,
"step": 2890
},
{
"epoch": 0.74,
"grad_norm": 0.2128180862906637,
"learning_rate": 3.770104470336555e-05,
"loss": 0.7173,
"step": 2895
},
{
"epoch": 0.74,
"grad_norm": 0.2130466930744276,
"learning_rate": 3.7351406957919636e-05,
"loss": 0.7266,
"step": 2900
},
{
"epoch": 0.75,
"grad_norm": 0.30988412023187845,
"learning_rate": 3.700302533262007e-05,
"loss": 0.717,
"step": 2905
},
{
"epoch": 0.75,
"grad_norm": 0.29577601576252954,
"learning_rate": 3.665590681260658e-05,
"loss": 0.6991,
"step": 2910
},
{
"epoch": 0.75,
"grad_norm": 0.29759259924638304,
"learning_rate": 3.631005835769334e-05,
"loss": 0.8029,
"step": 2915
},
{
"epoch": 0.75,
"grad_norm": 0.22770941704354283,
"learning_rate": 3.59654869022294e-05,
"loss": 0.7255,
"step": 2920
},
{
"epoch": 0.75,
"grad_norm": 0.2245089532543722,
"learning_rate": 3.5622199354959677e-05,
"loss": 0.8236,
"step": 2925
},
{
"epoch": 0.75,
"grad_norm": 0.209100399963671,
"learning_rate": 3.5280202598886324e-05,
"loss": 0.6472,
"step": 2930
},
{
"epoch": 0.75,
"grad_norm": 0.22771150718592803,
"learning_rate": 3.4939503491130934e-05,
"loss": 0.6684,
"step": 2935
},
{
"epoch": 0.75,
"grad_norm": 0.2617878485718964,
"learning_rate": 3.4600108862796796e-05,
"loss": 0.7271,
"step": 2940
},
{
"epoch": 0.76,
"grad_norm": 0.2674851467564555,
"learning_rate": 3.426202551883213e-05,
"loss": 0.7587,
"step": 2945
},
{
"epoch": 0.76,
"grad_norm": 0.2200031558966552,
"learning_rate": 3.392526023789349e-05,
"loss": 0.7258,
"step": 2950
},
{
"epoch": 0.76,
"grad_norm": 0.21417253713351433,
"learning_rate": 3.358981977221006e-05,
"loss": 0.6981,
"step": 2955
},
{
"epoch": 0.76,
"grad_norm": 0.2217852576193653,
"learning_rate": 3.325571084744803e-05,
"loss": 0.6972,
"step": 2960
},
{
"epoch": 0.76,
"grad_norm": 0.2984053822842527,
"learning_rate": 3.292294016257598e-05,
"loss": 0.6919,
"step": 2965
},
{
"epoch": 0.76,
"grad_norm": 0.25032718638928686,
"learning_rate": 3.259151438973024e-05,
"loss": 0.7997,
"step": 2970
},
{
"epoch": 0.76,
"grad_norm": 0.26398149994753983,
"learning_rate": 3.2261440174081525e-05,
"loss": 0.7703,
"step": 2975
},
{
"epoch": 0.76,
"grad_norm": 0.2088495888581316,
"learning_rate": 3.1932724133701344e-05,
"loss": 0.5416,
"step": 2980
},
{
"epoch": 0.77,
"grad_norm": 0.2886491177936703,
"learning_rate": 3.160537285942956e-05,
"loss": 0.6853,
"step": 2985
},
{
"epoch": 0.77,
"grad_norm": 0.2594733197020413,
"learning_rate": 3.1279392914742046e-05,
"loss": 0.7261,
"step": 2990
},
{
"epoch": 0.77,
"grad_norm": 0.3017152993780796,
"learning_rate": 3.0954790835619195e-05,
"loss": 0.6988,
"step": 2995
},
{
"epoch": 0.77,
"grad_norm": 0.2613281888272547,
"learning_rate": 3.06315731304148e-05,
"loss": 0.7953,
"step": 3000
},
{
"epoch": 0.77,
"grad_norm": 0.3030649141190901,
"learning_rate": 3.0309746279725748e-05,
"loss": 0.6728,
"step": 3005
},
{
"epoch": 0.77,
"grad_norm": 0.2049675658528854,
"learning_rate": 2.998931673626175e-05,
"loss": 0.6939,
"step": 3010
},
{
"epoch": 0.77,
"grad_norm": 0.23405886262502706,
"learning_rate": 2.9670290924716337e-05,
"loss": 0.7633,
"step": 3015
},
{
"epoch": 0.77,
"grad_norm": 0.16410838860875954,
"learning_rate": 2.935267524163774e-05,
"loss": 0.7258,
"step": 3020
},
{
"epoch": 0.78,
"grad_norm": 0.24028216791671228,
"learning_rate": 2.9036476055300765e-05,
"loss": 0.759,
"step": 3025
},
{
"epoch": 0.78,
"grad_norm": 0.2512891253559569,
"learning_rate": 2.872169970557913e-05,
"loss": 0.7622,
"step": 3030
},
{
"epoch": 0.78,
"grad_norm": 0.24326580414626328,
"learning_rate": 2.8408352503818357e-05,
"loss": 0.7471,
"step": 3035
},
{
"epoch": 0.78,
"grad_norm": 0.30684996416868315,
"learning_rate": 2.8096440732709083e-05,
"loss": 0.7234,
"step": 3040
},
{
"epoch": 0.78,
"grad_norm": 0.2610371672143647,
"learning_rate": 2.778597064616133e-05,
"loss": 0.7114,
"step": 3045
},
{
"epoch": 0.78,
"grad_norm": 0.28649543075043066,
"learning_rate": 2.7476948469178887e-05,
"loss": 0.6914,
"step": 3050
},
{
"epoch": 0.78,
"grad_norm": 0.24940516209657004,
"learning_rate": 2.716938039773459e-05,
"loss": 0.754,
"step": 3055
},
{
"epoch": 0.79,
"grad_norm": 0.31813491613101524,
"learning_rate": 2.6863272598646106e-05,
"loss": 0.6058,
"step": 3060
},
{
"epoch": 0.79,
"grad_norm": 0.21911314508163085,
"learning_rate": 2.6558631209452323e-05,
"loss": 0.7083,
"step": 3065
},
{
"epoch": 0.79,
"grad_norm": 0.24723577107070233,
"learning_rate": 2.625546233829016e-05,
"loss": 0.6785,
"step": 3070
},
{
"epoch": 0.79,
"grad_norm": 0.26215344220728737,
"learning_rate": 2.5953772063772252e-05,
"loss": 0.7434,
"step": 3075
},
{
"epoch": 0.79,
"grad_norm": 0.2182580130319897,
"learning_rate": 2.5653566434864928e-05,
"loss": 0.7029,
"step": 3080
},
{
"epoch": 0.79,
"grad_norm": 0.2589107810453318,
"learning_rate": 2.5354851470767037e-05,
"loss": 0.702,
"step": 3085
},
{
"epoch": 0.79,
"grad_norm": 0.25236658924095434,
"learning_rate": 2.5057633160789184e-05,
"loss": 0.7096,
"step": 3090
},
{
"epoch": 0.79,
"grad_norm": 0.2751653091880903,
"learning_rate": 2.476191746423373e-05,
"loss": 0.7791,
"step": 3095
},
{
"epoch": 0.8,
"grad_norm": 0.27791939731022475,
"learning_rate": 2.446771031027527e-05,
"loss": 0.7008,
"step": 3100
},
{
"epoch": 0.8,
"grad_norm": 0.2337931562400226,
"learning_rate": 2.4175017597841666e-05,
"loss": 0.7951,
"step": 3105
},
{
"epoch": 0.8,
"grad_norm": 0.23060409799544296,
"learning_rate": 2.3883845195495878e-05,
"loss": 0.7475,
"step": 3110
},
{
"epoch": 0.8,
"grad_norm": 0.19515481997034956,
"learning_rate": 2.3594198941318248e-05,
"loss": 0.677,
"step": 3115
},
{
"epoch": 0.8,
"grad_norm": 0.21013466271107667,
"learning_rate": 2.330608464278953e-05,
"loss": 0.7343,
"step": 3120
},
{
"epoch": 0.8,
"grad_norm": 0.2655018550539882,
"learning_rate": 2.3019508076674267e-05,
"loss": 0.6745,
"step": 3125
},
{
"epoch": 0.8,
"grad_norm": 0.22554453658559034,
"learning_rate": 2.273447498890521e-05,
"loss": 0.8374,
"step": 3130
},
{
"epoch": 0.8,
"grad_norm": 0.290174375117781,
"learning_rate": 2.2450991094467865e-05,
"loss": 0.668,
"step": 3135
},
{
"epoch": 0.81,
"grad_norm": 0.22779449704015578,
"learning_rate": 2.2169062077286075e-05,
"loss": 0.6899,
"step": 3140
},
{
"epoch": 0.81,
"grad_norm": 0.25494436011601035,
"learning_rate": 2.1888693590107946e-05,
"loss": 0.7013,
"step": 3145
},
{
"epoch": 0.81,
"grad_norm": 0.14266544623444502,
"learning_rate": 2.1609891254392678e-05,
"loss": 0.6311,
"step": 3150
},
{
"epoch": 0.81,
"grad_norm": 0.23298437086921395,
"learning_rate": 2.1332660660197602e-05,
"loss": 0.6729,
"step": 3155
},
{
"epoch": 0.81,
"grad_norm": 0.30998975291365666,
"learning_rate": 2.1057007366066373e-05,
"loss": 0.777,
"step": 3160
},
{
"epoch": 0.81,
"grad_norm": 0.24570574953745566,
"learning_rate": 2.0782936898917204e-05,
"loss": 0.6747,
"step": 3165
},
{
"epoch": 0.81,
"grad_norm": 0.20435326171988946,
"learning_rate": 2.0510454753932395e-05,
"loss": 0.6928,
"step": 3170
},
{
"epoch": 0.81,
"grad_norm": 0.24135761451835352,
"learning_rate": 2.0239566394447874e-05,
"loss": 0.7221,
"step": 3175
},
{
"epoch": 0.82,
"grad_norm": 0.21829756764053368,
"learning_rate": 1.9970277251843862e-05,
"loss": 0.7179,
"step": 3180
},
{
"epoch": 0.82,
"grad_norm": 0.19424951966772022,
"learning_rate": 1.9702592725435786e-05,
"loss": 0.7273,
"step": 3185
},
{
"epoch": 0.82,
"grad_norm": 0.234887739233565,
"learning_rate": 1.9436518182366158e-05,
"loss": 0.7391,
"step": 3190
},
{
"epoch": 0.82,
"grad_norm": 0.26397014493832816,
"learning_rate": 1.9172058957496876e-05,
"loss": 0.7456,
"step": 3195
},
{
"epoch": 0.82,
"grad_norm": 0.19484126439669286,
"learning_rate": 1.8909220353302392e-05,
"loss": 0.6979,
"step": 3200
},
{
"epoch": 0.82,
"grad_norm": 0.3263727199356017,
"learning_rate": 1.8648007639763176e-05,
"loss": 0.7104,
"step": 3205
},
{
"epoch": 0.82,
"grad_norm": 0.2490196028039222,
"learning_rate": 1.838842605426031e-05,
"loss": 0.6245,
"step": 3210
},
{
"epoch": 0.82,
"grad_norm": 0.22590753983736134,
"learning_rate": 1.813048080147025e-05,
"loss": 0.6477,
"step": 3215
},
{
"epoch": 0.83,
"grad_norm": 0.1934318577669605,
"learning_rate": 1.7874177053260598e-05,
"loss": 0.6484,
"step": 3220
},
{
"epoch": 0.83,
"grad_norm": 0.19824954761719257,
"learning_rate": 1.7619519948586348e-05,
"loss": 0.6,
"step": 3225
},
{
"epoch": 0.83,
"grad_norm": 0.29596182070597254,
"learning_rate": 1.736651459338695e-05,
"loss": 0.6995,
"step": 3230
},
{
"epoch": 0.83,
"grad_norm": 0.2552348994159813,
"learning_rate": 1.711516606048377e-05,
"loss": 0.7123,
"step": 3235
},
{
"epoch": 0.83,
"grad_norm": 0.23821982266920058,
"learning_rate": 1.6865479389478545e-05,
"loss": 0.7231,
"step": 3240
},
{
"epoch": 0.83,
"grad_norm": 0.20223828918906173,
"learning_rate": 1.6617459586652196e-05,
"loss": 0.673,
"step": 3245
},
{
"epoch": 0.83,
"grad_norm": 0.21541619069004347,
"learning_rate": 1.6371111624864543e-05,
"loss": 0.6788,
"step": 3250
},
{
"epoch": 0.84,
"grad_norm": 0.23132279393395067,
"learning_rate": 1.612644044345456e-05,
"loss": 0.7001,
"step": 3255
},
{
"epoch": 0.84,
"grad_norm": 0.3104102493541718,
"learning_rate": 1.5883450948141377e-05,
"loss": 0.6592,
"step": 3260
},
{
"epoch": 0.84,
"grad_norm": 0.19498096199332454,
"learning_rate": 1.564214801092585e-05,
"loss": 0.764,
"step": 3265
},
{
"epoch": 0.84,
"grad_norm": 0.22350807889337018,
"learning_rate": 1.540253646999299e-05,
"loss": 0.7285,
"step": 3270
},
{
"epoch": 0.84,
"grad_norm": 0.2530056423195899,
"learning_rate": 1.5164621129614787e-05,
"loss": 0.6813,
"step": 3275
},
{
"epoch": 0.84,
"grad_norm": 0.24407289762340453,
"learning_rate": 1.4928406760054059e-05,
"loss": 0.702,
"step": 3280
},
{
"epoch": 0.84,
"grad_norm": 0.27667414480460156,
"learning_rate": 1.4693898097468662e-05,
"loss": 0.7285,
"step": 3285
},
{
"epoch": 0.84,
"grad_norm": 0.24681628743295808,
"learning_rate": 1.4461099843816684e-05,
"loss": 0.6954,
"step": 3290
},
{
"epoch": 0.85,
"grad_norm": 0.24873679256421594,
"learning_rate": 1.4230016666761981e-05,
"loss": 0.7655,
"step": 3295
},
{
"epoch": 0.85,
"grad_norm": 0.22862446264265862,
"learning_rate": 1.4000653199580782e-05,
"loss": 0.7687,
"step": 3300
},
{
"epoch": 0.85,
"grad_norm": 0.3046659102738475,
"learning_rate": 1.3773014041068633e-05,
"loss": 0.7156,
"step": 3305
},
{
"epoch": 0.85,
"grad_norm": 0.22407782938978063,
"learning_rate": 1.3547103755448287e-05,
"loss": 0.7858,
"step": 3310
},
{
"epoch": 0.85,
"grad_norm": 0.19036374903254158,
"learning_rate": 1.3322926872278097e-05,
"loss": 0.7766,
"step": 3315
},
{
"epoch": 0.85,
"grad_norm": 0.26699406565117034,
"learning_rate": 1.3100487886361379e-05,
"loss": 0.7457,
"step": 3320
},
{
"epoch": 0.85,
"grad_norm": 0.22344849327126515,
"learning_rate": 1.2879791257656049e-05,
"loss": 0.7227,
"step": 3325
},
{
"epoch": 0.85,
"grad_norm": 0.1919000639308728,
"learning_rate": 1.266084141118542e-05,
"loss": 0.683,
"step": 3330
},
{
"epoch": 0.86,
"grad_norm": 0.2126312569959486,
"learning_rate": 1.2443642736949258e-05,
"loss": 0.702,
"step": 3335
},
{
"epoch": 0.86,
"grad_norm": 0.21328716751976873,
"learning_rate": 1.2228199589835999e-05,
"loss": 0.7592,
"step": 3340
},
{
"epoch": 0.86,
"grad_norm": 0.25817417734261355,
"learning_rate": 1.2014516289535249e-05,
"loss": 0.7101,
"step": 3345
},
{
"epoch": 0.86,
"grad_norm": 0.24654221507857982,
"learning_rate": 1.1802597120451286e-05,
"loss": 0.7651,
"step": 3350
},
{
"epoch": 0.86,
"grad_norm": 0.2729024564998873,
"learning_rate": 1.1592446331617068e-05,
"loss": 0.6934,
"step": 3355
},
{
"epoch": 0.86,
"grad_norm": 0.27603586653087275,
"learning_rate": 1.1384068136609105e-05,
"loss": 0.7848,
"step": 3360
},
{
"epoch": 0.86,
"grad_norm": 0.3079236794395131,
"learning_rate": 1.1177466713462926e-05,
"loss": 0.6581,
"step": 3365
},
{
"epoch": 0.86,
"grad_norm": 0.31502255340495816,
"learning_rate": 1.0972646204589377e-05,
"loss": 0.6583,
"step": 3370
},
{
"epoch": 0.87,
"grad_norm": 0.22084726894868312,
"learning_rate": 1.076961071669148e-05,
"loss": 0.7294,
"step": 3375
},
{
"epoch": 0.87,
"grad_norm": 0.2472008055830039,
"learning_rate": 1.0568364320682178e-05,
"loss": 0.7834,
"step": 3380
},
{
"epoch": 0.87,
"grad_norm": 0.2277343730480719,
"learning_rate": 1.0368911051602636e-05,
"loss": 0.66,
"step": 3385
},
{
"epoch": 0.87,
"grad_norm": 0.2415285274184652,
"learning_rate": 1.0171254908541372e-05,
"loss": 0.6748,
"step": 3390
},
{
"epoch": 0.87,
"grad_norm": 0.2861895095719399,
"learning_rate": 9.975399854554068e-06,
"loss": 0.6721,
"step": 3395
},
{
"epoch": 0.87,
"grad_norm": 0.25395148971387527,
"learning_rate": 9.781349816584162e-06,
"loss": 0.7267,
"step": 3400
},
{
"epoch": 0.87,
"grad_norm": 0.21070182161070172,
"learning_rate": 9.589108685383975e-06,
"loss": 0.7106,
"step": 3405
},
{
"epoch": 0.87,
"grad_norm": 0.25593439329555395,
"learning_rate": 9.398680315436903e-06,
"loss": 0.6957,
"step": 3410
},
{
"epoch": 0.88,
"grad_norm": 0.21752295344145742,
"learning_rate": 9.210068524879923e-06,
"loss": 0.685,
"step": 3415
},
{
"epoch": 0.88,
"grad_norm": 0.2446368196142467,
"learning_rate": 9.023277095427173e-06,
"loss": 0.7026,
"step": 3420
},
{
"epoch": 0.88,
"grad_norm": 0.23687207505265248,
"learning_rate": 8.838309772294085e-06,
"loss": 0.744,
"step": 3425
},
{
"epoch": 0.88,
"grad_norm": 0.25129247742502786,
"learning_rate": 8.655170264122303e-06,
"loss": 0.731,
"step": 3430
},
{
"epoch": 0.88,
"grad_norm": 0.22617739846256776,
"learning_rate": 8.473862242905339e-06,
"loss": 0.7936,
"step": 3435
},
{
"epoch": 0.88,
"grad_norm": 0.19703347689106585,
"learning_rate": 8.294389343914899e-06,
"loss": 0.6508,
"step": 3440
},
{
"epoch": 0.88,
"grad_norm": 0.2290192280255233,
"learning_rate": 8.11675516562802e-06,
"loss": 0.7213,
"step": 3445
},
{
"epoch": 0.89,
"grad_norm": 0.22631388422212603,
"learning_rate": 7.940963269654922e-06,
"loss": 0.8072,
"step": 3450
},
{
"epoch": 0.89,
"grad_norm": 0.23898504344416246,
"learning_rate": 7.767017180667645e-06,
"loss": 0.7186,
"step": 3455
},
{
"epoch": 0.89,
"grad_norm": 0.1771416341491615,
"learning_rate": 7.594920386329252e-06,
"loss": 0.6843,
"step": 3460
},
{
"epoch": 0.89,
"grad_norm": 0.32232301037200284,
"learning_rate": 7.4246763372240795e-06,
"loss": 0.6092,
"step": 3465
},
{
"epoch": 0.89,
"grad_norm": 0.20983698372327894,
"learning_rate": 7.256288446788362e-06,
"loss": 0.6873,
"step": 3470
},
{
"epoch": 0.89,
"grad_norm": 0.23642454930535492,
"learning_rate": 7.089760091241937e-06,
"loss": 0.7668,
"step": 3475
},
{
"epoch": 0.89,
"grad_norm": 0.19342866826333685,
"learning_rate": 6.925094609520455e-06,
"loss": 0.6985,
"step": 3480
},
{
"epoch": 0.89,
"grad_norm": 0.1659312022620069,
"learning_rate": 6.762295303208532e-06,
"loss": 0.6082,
"step": 3485
},
{
"epoch": 0.9,
"grad_norm": 0.23314155105131054,
"learning_rate": 6.601365436473439e-06,
"loss": 0.7625,
"step": 3490
},
{
"epoch": 0.9,
"grad_norm": 0.2266096665634237,
"learning_rate": 6.44230823599975e-06,
"loss": 0.7175,
"step": 3495
},
{
"epoch": 0.9,
"grad_norm": 0.285619796115372,
"learning_rate": 6.2851268909245865e-06,
"loss": 0.6727,
"step": 3500
},
{
"epoch": 0.9,
"grad_norm": 0.2499473433608083,
"learning_rate": 6.1298245527737e-06,
"loss": 0.7572,
"step": 3505
},
{
"epoch": 0.9,
"grad_norm": 0.23560223011085285,
"learning_rate": 5.976404335398256e-06,
"loss": 0.6292,
"step": 3510
},
{
"epoch": 0.9,
"grad_norm": 0.33344906781754363,
"learning_rate": 5.824869314912473e-06,
"loss": 0.7019,
"step": 3515
},
{
"epoch": 0.9,
"grad_norm": 0.19032010411212455,
"learning_rate": 5.675222529631841e-06,
"loss": 0.7159,
"step": 3520
},
{
"epoch": 0.9,
"grad_norm": 0.33161459892073875,
"learning_rate": 5.527466980012297e-06,
"loss": 0.7218,
"step": 3525
},
{
"epoch": 0.91,
"grad_norm": 0.2327886817133746,
"learning_rate": 5.381605628590003e-06,
"loss": 0.7328,
"step": 3530
},
{
"epoch": 0.91,
"grad_norm": 0.22800725916467116,
"learning_rate": 5.237641399921955e-06,
"loss": 0.6795,
"step": 3535
},
{
"epoch": 0.91,
"grad_norm": 0.23265993183122352,
"learning_rate": 5.095577180527378e-06,
"loss": 0.752,
"step": 3540
},
{
"epoch": 0.91,
"grad_norm": 0.21740715842870284,
"learning_rate": 4.9554158188298445e-06,
"loss": 0.7713,
"step": 3545
},
{
"epoch": 0.91,
"grad_norm": 0.2575708746269016,
"learning_rate": 4.817160125100106e-06,
"loss": 0.7631,
"step": 3550
},
{
"epoch": 0.91,
"grad_norm": 0.26658430974820374,
"learning_rate": 4.680812871399854e-06,
"loss": 0.7095,
"step": 3555
},
{
"epoch": 0.91,
"grad_norm": 0.22923585645582018,
"learning_rate": 4.546376791525975e-06,
"loss": 0.7197,
"step": 3560
},
{
"epoch": 0.91,
"grad_norm": 0.20839075381158542,
"learning_rate": 4.413854580955945e-06,
"loss": 0.717,
"step": 3565
},
{
"epoch": 0.92,
"grad_norm": 0.2126973684335578,
"learning_rate": 4.2832488967935795e-06,
"loss": 0.7439,
"step": 3570
},
{
"epoch": 0.92,
"grad_norm": 0.22046801778519598,
"learning_rate": 4.154562357715952e-06,
"loss": 0.7287,
"step": 3575
},
{
"epoch": 0.92,
"grad_norm": 0.22198088834095264,
"learning_rate": 4.02779754392072e-06,
"loss": 0.6862,
"step": 3580
},
{
"epoch": 0.92,
"grad_norm": 0.2562953238617015,
"learning_rate": 3.90295699707447e-06,
"loss": 0.705,
"step": 3585
},
{
"epoch": 0.92,
"grad_norm": 0.2832586699922064,
"learning_rate": 3.780043220261764e-06,
"loss": 0.7525,
"step": 3590
},
{
"epoch": 0.92,
"grad_norm": 0.23818590199709755,
"learning_rate": 3.659058677934957e-06,
"loss": 0.7018,
"step": 3595
},
{
"epoch": 0.92,
"grad_norm": 0.3095559851055113,
"learning_rate": 3.540005795864709e-06,
"loss": 0.6073,
"step": 3600
},
{
"epoch": 0.92,
"grad_norm": 0.1861043245337588,
"learning_rate": 3.422886961091476e-06,
"loss": 0.6862,
"step": 3605
},
{
"epoch": 0.93,
"grad_norm": 0.2807414585458434,
"learning_rate": 3.3077045218775192e-06,
"loss": 0.7819,
"step": 3610
},
{
"epoch": 0.93,
"grad_norm": 0.23756038931466225,
"learning_rate": 3.194460787659892e-06,
"loss": 0.7211,
"step": 3615
},
{
"epoch": 0.93,
"grad_norm": 0.2375925655429517,
"learning_rate": 3.0831580290041184e-06,
"loss": 0.7611,
"step": 3620
},
{
"epoch": 0.93,
"grad_norm": 0.35878418981357185,
"learning_rate": 2.973798477558709e-06,
"loss": 0.7034,
"step": 3625
},
{
"epoch": 0.93,
"grad_norm": 0.18014791198441923,
"learning_rate": 2.8663843260103074e-06,
"loss": 0.6713,
"step": 3630
},
{
"epoch": 0.93,
"grad_norm": 0.2066769742143976,
"learning_rate": 2.76091772803988e-06,
"loss": 0.7165,
"step": 3635
},
{
"epoch": 0.93,
"grad_norm": 0.29548750089547526,
"learning_rate": 2.6574007982793857e-06,
"loss": 0.7431,
"step": 3640
},
{
"epoch": 0.94,
"grad_norm": 0.276827175378835,
"learning_rate": 2.555835612269475e-06,
"loss": 0.665,
"step": 3645
},
{
"epoch": 0.94,
"grad_norm": 0.21719297673422058,
"learning_rate": 2.456224206417812e-06,
"loss": 0.697,
"step": 3650
},
{
"epoch": 0.94,
"grad_norm": 0.24117198209251042,
"learning_rate": 2.35856857795832e-06,
"loss": 0.7668,
"step": 3655
},
{
"epoch": 0.94,
"grad_norm": 0.4419996258477192,
"learning_rate": 2.262870684911045e-06,
"loss": 0.6807,
"step": 3660
},
{
"epoch": 0.94,
"grad_norm": 0.2922290003913108,
"learning_rate": 2.169132446042976e-06,
"loss": 0.6772,
"step": 3665
},
{
"epoch": 0.94,
"grad_norm": 0.21578507137937,
"learning_rate": 2.0773557408295343e-06,
"loss": 0.7143,
"step": 3670
},
{
"epoch": 0.94,
"grad_norm": 0.22699656079084274,
"learning_rate": 1.9875424094168647e-06,
"loss": 0.6936,
"step": 3675
},
{
"epoch": 0.94,
"grad_norm": 0.24001514084329242,
"learning_rate": 1.8996942525850047e-06,
"loss": 0.6977,
"step": 3680
},
{
"epoch": 0.95,
"grad_norm": 0.26514502359960324,
"learning_rate": 1.813813031711742e-06,
"loss": 0.701,
"step": 3685
},
{
"epoch": 0.95,
"grad_norm": 0.23244532178437396,
"learning_rate": 1.7299004687372665e-06,
"loss": 0.6792,
"step": 3690
},
{
"epoch": 0.95,
"grad_norm": 0.2517442477838529,
"learning_rate": 1.6479582461297217e-06,
"loss": 0.6781,
"step": 3695
},
{
"epoch": 0.95,
"grad_norm": 0.22695769814956426,
"learning_rate": 1.5679880068514174e-06,
"loss": 0.6858,
"step": 3700
},
{
"epoch": 0.95,
"grad_norm": 0.21710844280108782,
"learning_rate": 1.4899913543258814e-06,
"loss": 0.7148,
"step": 3705
},
{
"epoch": 0.95,
"grad_norm": 0.2967274268521252,
"learning_rate": 1.4139698524057165e-06,
"loss": 0.781,
"step": 3710
},
{
"epoch": 0.95,
"grad_norm": 0.30352069911343654,
"learning_rate": 1.3399250253413154e-06,
"loss": 0.7702,
"step": 3715
},
{
"epoch": 0.95,
"grad_norm": 0.2929106725336177,
"learning_rate": 1.2678583577501624e-06,
"loss": 0.6759,
"step": 3720
},
{
"epoch": 0.96,
"grad_norm": 0.29851324638490223,
"learning_rate": 1.1977712945872243e-06,
"loss": 0.7205,
"step": 3725
},
{
"epoch": 0.96,
"grad_norm": 0.26915452333889833,
"learning_rate": 1.1296652411158182e-06,
"loss": 0.7057,
"step": 3730
},
{
"epoch": 0.96,
"grad_norm": 0.2245310916127946,
"learning_rate": 1.0635415628795665e-06,
"loss": 0.6926,
"step": 3735
},
{
"epoch": 0.96,
"grad_norm": 0.2178134240444238,
"learning_rate": 9.994015856749527e-07,
"loss": 0.7104,
"step": 3740
},
{
"epoch": 0.96,
"grad_norm": 0.18742301371018372,
"learning_rate": 9.372465955247544e-07,
"loss": 0.6046,
"step": 3745
},
{
"epoch": 0.96,
"grad_norm": 0.2282365317937403,
"learning_rate": 8.770778386522627e-07,
"loss": 0.7643,
"step": 3750
},
{
"epoch": 0.96,
"grad_norm": 0.23411248148329084,
"learning_rate": 8.188965214562804e-07,
"loss": 0.7397,
"step": 3755
},
{
"epoch": 0.96,
"grad_norm": 0.25053318378159295,
"learning_rate": 7.627038104869199e-07,
"loss": 0.6724,
"step": 3760
},
{
"epoch": 0.97,
"grad_norm": 0.27403052281802975,
"learning_rate": 7.085008324222763e-07,
"loss": 0.6412,
"step": 3765
},
{
"epoch": 0.97,
"grad_norm": 0.2635701878097459,
"learning_rate": 6.562886740457797e-07,
"loss": 0.7377,
"step": 3770
},
{
"epoch": 0.97,
"grad_norm": 0.21017227480655776,
"learning_rate": 6.060683822244117e-07,
"loss": 0.6977,
"step": 3775
},
{
"epoch": 0.97,
"grad_norm": 0.3453902794038148,
"learning_rate": 5.578409638877457e-07,
"loss": 0.5952,
"step": 3780
},
{
"epoch": 0.97,
"grad_norm": 0.24555784325804006,
"learning_rate": 5.116073860077064e-07,
"loss": 0.6711,
"step": 3785
},
{
"epoch": 0.97,
"grad_norm": 0.29390073833065394,
"learning_rate": 4.6736857557925227e-07,
"loss": 0.691,
"step": 3790
},
{
"epoch": 0.97,
"grad_norm": 0.2782701779800116,
"learning_rate": 4.2512541960171294e-07,
"loss": 0.7768,
"step": 3795
},
{
"epoch": 0.97,
"grad_norm": 0.37197636711074117,
"learning_rate": 3.8487876506106966e-07,
"loss": 0.7539,
"step": 3800
},
{
"epoch": 0.98,
"grad_norm": 0.23773581142305195,
"learning_rate": 3.466294189129249e-07,
"loss": 0.6583,
"step": 3805
},
{
"epoch": 0.98,
"grad_norm": 0.2623487880042545,
"learning_rate": 3.1037814806634815e-07,
"loss": 0.7085,
"step": 3810
},
{
"epoch": 0.98,
"grad_norm": 0.21088747078936257,
"learning_rate": 2.7612567936849964e-07,
"loss": 0.6627,
"step": 3815
},
{
"epoch": 0.98,
"grad_norm": 0.19741270913916117,
"learning_rate": 2.43872699590042e-07,
"loss": 0.7,
"step": 3820
},
{
"epoch": 0.98,
"grad_norm": 0.2518577876286286,
"learning_rate": 2.136198554113844e-07,
"loss": 0.7888,
"step": 3825
},
{
"epoch": 0.98,
"grad_norm": 0.25499260076898056,
"learning_rate": 1.8536775340970425e-07,
"loss": 0.7371,
"step": 3830
},
{
"epoch": 0.98,
"grad_norm": 0.22301016570777385,
"learning_rate": 1.591169600468123e-07,
"loss": 0.7354,
"step": 3835
},
{
"epoch": 0.99,
"grad_norm": 0.24054514645205372,
"learning_rate": 1.348680016577397e-07,
"loss": 0.7868,
"step": 3840
},
{
"epoch": 0.99,
"grad_norm": 0.24969353601595792,
"learning_rate": 1.126213644402463e-07,
"loss": 0.7333,
"step": 3845
},
{
"epoch": 0.99,
"grad_norm": 0.22761401425681646,
"learning_rate": 9.237749444505062e-08,
"loss": 0.6615,
"step": 3850
},
{
"epoch": 0.99,
"grad_norm": 0.24426703862468954,
"learning_rate": 7.413679756684832e-08,
"loss": 0.7324,
"step": 3855
},
{
"epoch": 0.99,
"grad_norm": 0.2867096795033602,
"learning_rate": 5.7899639536251883e-08,
"loss": 0.7088,
"step": 3860
},
{
"epoch": 0.99,
"grad_norm": 0.26451528734107205,
"learning_rate": 4.366634591237428e-08,
"loss": 0.7354,
"step": 3865
},
{
"epoch": 0.99,
"grad_norm": 0.2064960678251034,
"learning_rate": 3.143720207635648e-08,
"loss": 0.674,
"step": 3870
},
{
"epoch": 0.99,
"grad_norm": 0.32104009231527136,
"learning_rate": 2.1212453225627482e-08,
"loss": 0.7268,
"step": 3875
},
{
"epoch": 1.0,
"grad_norm": 0.28886214510321495,
"learning_rate": 1.299230436898613e-08,
"loss": 0.782,
"step": 3880
},
{
"epoch": 1.0,
"grad_norm": 0.25342207402007316,
"learning_rate": 6.776920322515423e-09,
"loss": 0.6615,
"step": 3885
},
{
"epoch": 1.0,
"grad_norm": 0.2610844053652662,
"learning_rate": 2.566425706218567e-09,
"loss": 0.6916,
"step": 3890
},
{
"epoch": 1.0,
"grad_norm": 0.21290106643942017,
"learning_rate": 3.609049415764787e-10,
"loss": 0.7345,
"step": 3895
},
{
"epoch": 1.0,
"eval_loss": NaN,
"eval_runtime": 2056.4641,
"eval_samples_per_second": 3.371,
"eval_steps_per_second": 0.843,
"step": 3898
},
{
"epoch": 1.0,
"step": 3898,
"total_flos": 1.0023612803710976e+16,
"train_loss": 0.7175084921933248,
"train_runtime": 21380.3358,
"train_samples_per_second": 2.917,
"train_steps_per_second": 0.182
}
],
"logging_steps": 5,
"max_steps": 3898,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"total_flos": 1.0023612803710976e+16,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}