|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 20583, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.004858378273332362, |
|
"grad_norm": 66137.25, |
|
"learning_rate": 4.9757081086333386e-05, |
|
"loss": 0.5598, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.009716756546664724, |
|
"grad_norm": 60222.27734375, |
|
"learning_rate": 4.951416217266677e-05, |
|
"loss": 0.5419, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.014575134819997084, |
|
"grad_norm": 67718.328125, |
|
"learning_rate": 4.927124325900015e-05, |
|
"loss": 0.5491, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.019433513093329448, |
|
"grad_norm": 73856.1015625, |
|
"learning_rate": 4.902832434533353e-05, |
|
"loss": 0.5496, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.024291891366661808, |
|
"grad_norm": 60560.8359375, |
|
"learning_rate": 4.878540543166691e-05, |
|
"loss": 0.5381, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.029150269639994168, |
|
"grad_norm": 76622.125, |
|
"learning_rate": 4.8542486518000295e-05, |
|
"loss": 0.5397, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.03400864791332653, |
|
"grad_norm": 55521.8203125, |
|
"learning_rate": 4.829956760433367e-05, |
|
"loss": 0.5339, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.038867026186658896, |
|
"grad_norm": 59273.4921875, |
|
"learning_rate": 4.8056648690667055e-05, |
|
"loss": 0.5342, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.043725404459991256, |
|
"grad_norm": 63513.421875, |
|
"learning_rate": 4.781372977700044e-05, |
|
"loss": 0.5247, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.048583782733323616, |
|
"grad_norm": 61088.77734375, |
|
"learning_rate": 4.757081086333382e-05, |
|
"loss": 0.5302, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.053442161006655976, |
|
"grad_norm": 71392.75, |
|
"learning_rate": 4.7327891949667204e-05, |
|
"loss": 0.5261, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.058300539279988337, |
|
"grad_norm": 72375.1328125, |
|
"learning_rate": 4.708497303600059e-05, |
|
"loss": 0.5153, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.0631589175533207, |
|
"grad_norm": 65540.171875, |
|
"learning_rate": 4.6842054122333964e-05, |
|
"loss": 0.5068, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.06801729582665306, |
|
"grad_norm": 63439.60546875, |
|
"learning_rate": 4.659913520866735e-05, |
|
"loss": 0.5087, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.07287567409998542, |
|
"grad_norm": 56190.2265625, |
|
"learning_rate": 4.635621629500073e-05, |
|
"loss": 0.5074, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.07773405237331779, |
|
"grad_norm": 72489.0234375, |
|
"learning_rate": 4.6113297381334114e-05, |
|
"loss": 0.5036, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.08259243064665014, |
|
"grad_norm": 55422.61328125, |
|
"learning_rate": 4.58703784676675e-05, |
|
"loss": 0.4979, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.08745080891998251, |
|
"grad_norm": 58258.125, |
|
"learning_rate": 4.562745955400088e-05, |
|
"loss": 0.5016, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.09230918719331486, |
|
"grad_norm": 57958.55859375, |
|
"learning_rate": 4.5384540640334264e-05, |
|
"loss": 0.508, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.09716756546664723, |
|
"grad_norm": 52961.26953125, |
|
"learning_rate": 4.514162172666764e-05, |
|
"loss": 0.4843, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.1020259437399796, |
|
"grad_norm": 57415.234375, |
|
"learning_rate": 4.489870281300102e-05, |
|
"loss": 0.4943, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.10688432201331195, |
|
"grad_norm": 49461.50390625, |
|
"learning_rate": 4.46557838993344e-05, |
|
"loss": 0.4817, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.11174270028664432, |
|
"grad_norm": 51018.04296875, |
|
"learning_rate": 4.441286498566778e-05, |
|
"loss": 0.4825, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.11660107855997667, |
|
"grad_norm": 52488.46484375, |
|
"learning_rate": 4.4169946072001166e-05, |
|
"loss": 0.4786, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.12145945683330904, |
|
"grad_norm": 61512.5078125, |
|
"learning_rate": 4.392702715833455e-05, |
|
"loss": 0.4775, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.1263178351066414, |
|
"grad_norm": 56490.4296875, |
|
"learning_rate": 4.368410824466793e-05, |
|
"loss": 0.4813, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.13117621337997376, |
|
"grad_norm": 63387.01171875, |
|
"learning_rate": 4.3441189331001316e-05, |
|
"loss": 0.4744, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.13603459165330611, |
|
"grad_norm": 62635.20703125, |
|
"learning_rate": 4.31982704173347e-05, |
|
"loss": 0.4758, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.1408929699266385, |
|
"grad_norm": 52577.96875, |
|
"learning_rate": 4.2955351503668076e-05, |
|
"loss": 0.4867, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.14575134819997085, |
|
"grad_norm": 54731.5, |
|
"learning_rate": 4.271243259000146e-05, |
|
"loss": 0.4658, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.1506097264733032, |
|
"grad_norm": 52055.84765625, |
|
"learning_rate": 4.246951367633484e-05, |
|
"loss": 0.4737, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.15546810474663558, |
|
"grad_norm": 48978.90234375, |
|
"learning_rate": 4.2226594762668225e-05, |
|
"loss": 0.4765, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.16032648301996794, |
|
"grad_norm": 54819.38671875, |
|
"learning_rate": 4.198367584900161e-05, |
|
"loss": 0.466, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.1651848612933003, |
|
"grad_norm": 58637.75390625, |
|
"learning_rate": 4.1740756935334985e-05, |
|
"loss": 0.4725, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.17004323956663267, |
|
"grad_norm": 55842.4609375, |
|
"learning_rate": 4.149783802166837e-05, |
|
"loss": 0.4724, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.17490161783996502, |
|
"grad_norm": 52083.8125, |
|
"learning_rate": 4.125491910800175e-05, |
|
"loss": 0.4796, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.17975999611329738, |
|
"grad_norm": 56764.9296875, |
|
"learning_rate": 4.1012000194335135e-05, |
|
"loss": 0.4736, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.18461837438662973, |
|
"grad_norm": 50524.82421875, |
|
"learning_rate": 4.076908128066851e-05, |
|
"loss": 0.4627, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.1894767526599621, |
|
"grad_norm": 50128.8359375, |
|
"learning_rate": 4.0526162367001894e-05, |
|
"loss": 0.4717, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.19433513093329446, |
|
"grad_norm": 72415.296875, |
|
"learning_rate": 4.028324345333528e-05, |
|
"loss": 0.456, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.19919350920662682, |
|
"grad_norm": 57324.921875, |
|
"learning_rate": 4.004032453966866e-05, |
|
"loss": 0.4505, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.2040518874799592, |
|
"grad_norm": 60287.41796875, |
|
"learning_rate": 3.9797405626002044e-05, |
|
"loss": 0.4441, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.20891026575329155, |
|
"grad_norm": 65722.1171875, |
|
"learning_rate": 3.955448671233543e-05, |
|
"loss": 0.4426, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.2137686440266239, |
|
"grad_norm": 57532.5625, |
|
"learning_rate": 3.931156779866881e-05, |
|
"loss": 0.4532, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.2186270222999563, |
|
"grad_norm": 60035.2265625, |
|
"learning_rate": 3.906864888500219e-05, |
|
"loss": 0.443, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.22348540057328864, |
|
"grad_norm": 50842.09375, |
|
"learning_rate": 3.882572997133557e-05, |
|
"loss": 0.4527, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.228343778846621, |
|
"grad_norm": 53098.82421875, |
|
"learning_rate": 3.858281105766895e-05, |
|
"loss": 0.441, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.23320215711995335, |
|
"grad_norm": 52847.52734375, |
|
"learning_rate": 3.833989214400233e-05, |
|
"loss": 0.4437, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.23806053539328573, |
|
"grad_norm": 55201.1953125, |
|
"learning_rate": 3.809697323033571e-05, |
|
"loss": 0.4534, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.24291891366661808, |
|
"grad_norm": 55352.77734375, |
|
"learning_rate": 3.7854054316669096e-05, |
|
"loss": 0.4559, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.24777729193995043, |
|
"grad_norm": 77840.125, |
|
"learning_rate": 3.761113540300248e-05, |
|
"loss": 0.4368, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.2526356702132828, |
|
"grad_norm": 69213.84375, |
|
"learning_rate": 3.736821648933586e-05, |
|
"loss": 0.4428, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.25749404848661517, |
|
"grad_norm": 61722.49609375, |
|
"learning_rate": 3.7125297575669246e-05, |
|
"loss": 0.444, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.2623524267599475, |
|
"grad_norm": 56948.3125, |
|
"learning_rate": 3.688237866200262e-05, |
|
"loss": 0.4423, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.2672108050332799, |
|
"grad_norm": 50691.4453125, |
|
"learning_rate": 3.6639459748336006e-05, |
|
"loss": 0.4371, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.27206918330661223, |
|
"grad_norm": 48622.8671875, |
|
"learning_rate": 3.639654083466939e-05, |
|
"loss": 0.4459, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.27692756157994464, |
|
"grad_norm": 58408.26171875, |
|
"learning_rate": 3.615362192100277e-05, |
|
"loss": 0.4215, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.281785939853277, |
|
"grad_norm": 48251.1953125, |
|
"learning_rate": 3.5910703007336156e-05, |
|
"loss": 0.4204, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.28664431812660934, |
|
"grad_norm": 73270.4453125, |
|
"learning_rate": 3.566778409366954e-05, |
|
"loss": 0.4421, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.2915026963999417, |
|
"grad_norm": 52048.3828125, |
|
"learning_rate": 3.5424865180002915e-05, |
|
"loss": 0.4438, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.29636107467327405, |
|
"grad_norm": 57475.203125, |
|
"learning_rate": 3.51819462663363e-05, |
|
"loss": 0.418, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.3012194529466064, |
|
"grad_norm": 55124.11328125, |
|
"learning_rate": 3.493902735266968e-05, |
|
"loss": 0.4347, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.3060778312199388, |
|
"grad_norm": 62276.3671875, |
|
"learning_rate": 3.469610843900306e-05, |
|
"loss": 0.4192, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.31093620949327117, |
|
"grad_norm": 52723.7421875, |
|
"learning_rate": 3.445318952533644e-05, |
|
"loss": 0.4413, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.3157945877666035, |
|
"grad_norm": 53744.6484375, |
|
"learning_rate": 3.4210270611669825e-05, |
|
"loss": 0.4324, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.32065296603993587, |
|
"grad_norm": 44454.984375, |
|
"learning_rate": 3.396735169800321e-05, |
|
"loss": 0.4269, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.3255113443132682, |
|
"grad_norm": 52088.671875, |
|
"learning_rate": 3.372443278433659e-05, |
|
"loss": 0.4317, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.3303697225866006, |
|
"grad_norm": 50359.18359375, |
|
"learning_rate": 3.3481513870669974e-05, |
|
"loss": 0.4102, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.33522810085993293, |
|
"grad_norm": 54712.453125, |
|
"learning_rate": 3.323859495700336e-05, |
|
"loss": 0.4194, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.34008647913326534, |
|
"grad_norm": 49144.39453125, |
|
"learning_rate": 3.2995676043336734e-05, |
|
"loss": 0.414, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.3449448574065977, |
|
"grad_norm": 48121.30078125, |
|
"learning_rate": 3.275275712967012e-05, |
|
"loss": 0.4203, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.34980323567993005, |
|
"grad_norm": 60245.484375, |
|
"learning_rate": 3.25098382160035e-05, |
|
"loss": 0.4227, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.3546616139532624, |
|
"grad_norm": 53339.56640625, |
|
"learning_rate": 3.226691930233688e-05, |
|
"loss": 0.4235, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.35951999222659475, |
|
"grad_norm": 61334.859375, |
|
"learning_rate": 3.202400038867026e-05, |
|
"loss": 0.4318, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.3643783704999271, |
|
"grad_norm": 61368.46484375, |
|
"learning_rate": 3.1781081475003643e-05, |
|
"loss": 0.4223, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.36923674877325946, |
|
"grad_norm": 67382.265625, |
|
"learning_rate": 3.153816256133703e-05, |
|
"loss": 0.4053, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.37409512704659187, |
|
"grad_norm": 56739.12890625, |
|
"learning_rate": 3.129524364767041e-05, |
|
"loss": 0.4121, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.3789535053199242, |
|
"grad_norm": 51203.3203125, |
|
"learning_rate": 3.105232473400379e-05, |
|
"loss": 0.432, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.3838118835932566, |
|
"grad_norm": 58815.6484375, |
|
"learning_rate": 3.080940582033717e-05, |
|
"loss": 0.405, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.38867026186658893, |
|
"grad_norm": 59941.88671875, |
|
"learning_rate": 3.056648690667055e-05, |
|
"loss": 0.425, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.3935286401399213, |
|
"grad_norm": 53543.46484375, |
|
"learning_rate": 3.0323567993003936e-05, |
|
"loss": 0.4103, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.39838701841325364, |
|
"grad_norm": 55630.40625, |
|
"learning_rate": 3.008064907933732e-05, |
|
"loss": 0.3964, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.40324539668658604, |
|
"grad_norm": 52717.6953125, |
|
"learning_rate": 2.9837730165670703e-05, |
|
"loss": 0.3989, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.4081037749599184, |
|
"grad_norm": 50914.84375, |
|
"learning_rate": 2.9594811252004086e-05, |
|
"loss": 0.4097, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.41296215323325075, |
|
"grad_norm": 61387.0390625, |
|
"learning_rate": 2.9351892338337466e-05, |
|
"loss": 0.4051, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.4178205315065831, |
|
"grad_norm": 45076.01953125, |
|
"learning_rate": 2.910897342467085e-05, |
|
"loss": 0.4012, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.42267890977991546, |
|
"grad_norm": 54953.07421875, |
|
"learning_rate": 2.8866054511004225e-05, |
|
"loss": 0.4202, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.4275372880532478, |
|
"grad_norm": 59405.3828125, |
|
"learning_rate": 2.862313559733761e-05, |
|
"loss": 0.3931, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.43239566632658016, |
|
"grad_norm": 61979.93359375, |
|
"learning_rate": 2.8380216683670992e-05, |
|
"loss": 0.3987, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.4372540445999126, |
|
"grad_norm": 46517.14453125, |
|
"learning_rate": 2.813729777000437e-05, |
|
"loss": 0.4135, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.4421124228732449, |
|
"grad_norm": 57222.25, |
|
"learning_rate": 2.7894378856337755e-05, |
|
"loss": 0.4048, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.4469708011465773, |
|
"grad_norm": 60419.09375, |
|
"learning_rate": 2.7651459942671138e-05, |
|
"loss": 0.4008, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.45182917941990963, |
|
"grad_norm": 42610.984375, |
|
"learning_rate": 2.740854102900452e-05, |
|
"loss": 0.3975, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.456687557693242, |
|
"grad_norm": 58822.0234375, |
|
"learning_rate": 2.71656221153379e-05, |
|
"loss": 0.3923, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.46154593596657434, |
|
"grad_norm": 47564.5234375, |
|
"learning_rate": 2.6922703201671284e-05, |
|
"loss": 0.391, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.4664043142399067, |
|
"grad_norm": 48113.20703125, |
|
"learning_rate": 2.6679784288004668e-05, |
|
"loss": 0.4064, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.4712626925132391, |
|
"grad_norm": 49079.30078125, |
|
"learning_rate": 2.6436865374338048e-05, |
|
"loss": 0.4077, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.47612107078657145, |
|
"grad_norm": 53815.7890625, |
|
"learning_rate": 2.619394646067143e-05, |
|
"loss": 0.3844, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.4809794490599038, |
|
"grad_norm": 45962.41796875, |
|
"learning_rate": 2.5951027547004814e-05, |
|
"loss": 0.3925, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.48583782733323616, |
|
"grad_norm": 50792.91015625, |
|
"learning_rate": 2.570810863333819e-05, |
|
"loss": 0.3838, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.4906962056065685, |
|
"grad_norm": 44864.3046875, |
|
"learning_rate": 2.5465189719671574e-05, |
|
"loss": 0.3924, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.49555458387990087, |
|
"grad_norm": 50477.6015625, |
|
"learning_rate": 2.5222270806004954e-05, |
|
"loss": 0.3875, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.5004129621532333, |
|
"grad_norm": 46136.0859375, |
|
"learning_rate": 2.4979351892338337e-05, |
|
"loss": 0.3956, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.5052713404265656, |
|
"grad_norm": 57640.59765625, |
|
"learning_rate": 2.473643297867172e-05, |
|
"loss": 0.38, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.510129718699898, |
|
"grad_norm": 60448.453125, |
|
"learning_rate": 2.4493514065005103e-05, |
|
"loss": 0.3925, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.5149880969732303, |
|
"grad_norm": 43668.453125, |
|
"learning_rate": 2.4250595151338483e-05, |
|
"loss": 0.3966, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.5198464752465627, |
|
"grad_norm": 51975.83203125, |
|
"learning_rate": 2.4007676237671866e-05, |
|
"loss": 0.399, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.524704853519895, |
|
"grad_norm": 60277.078125, |
|
"learning_rate": 2.376475732400525e-05, |
|
"loss": 0.3968, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.5295632317932274, |
|
"grad_norm": 47762.875, |
|
"learning_rate": 2.3521838410338633e-05, |
|
"loss": 0.3937, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.5344216100665597, |
|
"grad_norm": 62721.125, |
|
"learning_rate": 2.327891949667201e-05, |
|
"loss": 0.3871, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.5392799883398921, |
|
"grad_norm": 41137.8515625, |
|
"learning_rate": 2.3036000583005392e-05, |
|
"loss": 0.3814, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.5441383666132245, |
|
"grad_norm": 50033.2734375, |
|
"learning_rate": 2.2793081669338776e-05, |
|
"loss": 0.3987, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.5489967448865569, |
|
"grad_norm": 50574.140625, |
|
"learning_rate": 2.255016275567216e-05, |
|
"loss": 0.3741, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.5538551231598893, |
|
"grad_norm": 46732.82421875, |
|
"learning_rate": 2.230724384200554e-05, |
|
"loss": 0.3775, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.5587135014332216, |
|
"grad_norm": 57961.8671875, |
|
"learning_rate": 2.2064324928338922e-05, |
|
"loss": 0.3826, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.563571879706554, |
|
"grad_norm": 43375.69140625, |
|
"learning_rate": 2.1821406014672302e-05, |
|
"loss": 0.3823, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.5684302579798863, |
|
"grad_norm": 49148.40234375, |
|
"learning_rate": 2.1578487101005685e-05, |
|
"loss": 0.3782, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 0.5732886362532187, |
|
"grad_norm": 60182.703125, |
|
"learning_rate": 2.133556818733907e-05, |
|
"loss": 0.3804, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.578147014526551, |
|
"grad_norm": 47590.46484375, |
|
"learning_rate": 2.1092649273672448e-05, |
|
"loss": 0.3974, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 0.5830053927998834, |
|
"grad_norm": 65701.859375, |
|
"learning_rate": 2.084973036000583e-05, |
|
"loss": 0.3798, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.5878637710732157, |
|
"grad_norm": 43476.69140625, |
|
"learning_rate": 2.0606811446339215e-05, |
|
"loss": 0.378, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 0.5927221493465481, |
|
"grad_norm": 43104.83203125, |
|
"learning_rate": 2.0363892532672595e-05, |
|
"loss": 0.3718, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 0.5975805276198805, |
|
"grad_norm": 58024.0546875, |
|
"learning_rate": 2.0120973619005974e-05, |
|
"loss": 0.3837, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 0.6024389058932128, |
|
"grad_norm": 56071.4765625, |
|
"learning_rate": 1.9878054705339358e-05, |
|
"loss": 0.3789, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 0.6072972841665452, |
|
"grad_norm": 56979.65625, |
|
"learning_rate": 1.963513579167274e-05, |
|
"loss": 0.3726, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.6121556624398776, |
|
"grad_norm": 48315.94140625, |
|
"learning_rate": 1.9392216878006124e-05, |
|
"loss": 0.3697, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 0.61701404071321, |
|
"grad_norm": 48739.546875, |
|
"learning_rate": 1.9149297964339504e-05, |
|
"loss": 0.3849, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 0.6218724189865423, |
|
"grad_norm": 49214.71875, |
|
"learning_rate": 1.8906379050672887e-05, |
|
"loss": 0.3646, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 0.6267307972598747, |
|
"grad_norm": 46501.2265625, |
|
"learning_rate": 1.866346013700627e-05, |
|
"loss": 0.3693, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 0.631589175533207, |
|
"grad_norm": 69907.9609375, |
|
"learning_rate": 1.842054122333965e-05, |
|
"loss": 0.3737, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.6364475538065394, |
|
"grad_norm": 47973.73046875, |
|
"learning_rate": 1.817762230967303e-05, |
|
"loss": 0.3875, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 0.6413059320798717, |
|
"grad_norm": 58501.1796875, |
|
"learning_rate": 1.7934703396006413e-05, |
|
"loss": 0.3778, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 0.6461643103532041, |
|
"grad_norm": 61768.96875, |
|
"learning_rate": 1.7691784482339797e-05, |
|
"loss": 0.384, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 0.6510226886265364, |
|
"grad_norm": 46427.80859375, |
|
"learning_rate": 1.744886556867318e-05, |
|
"loss": 0.365, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 0.6558810668998688, |
|
"grad_norm": 58254.9453125, |
|
"learning_rate": 1.720594665500656e-05, |
|
"loss": 0.3641, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.6607394451732012, |
|
"grad_norm": 49971.00390625, |
|
"learning_rate": 1.696302774133994e-05, |
|
"loss": 0.363, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 0.6655978234465335, |
|
"grad_norm": 56874.12890625, |
|
"learning_rate": 1.6720108827673323e-05, |
|
"loss": 0.3795, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 0.6704562017198659, |
|
"grad_norm": 51813.33984375, |
|
"learning_rate": 1.6477189914006706e-05, |
|
"loss": 0.3809, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 0.6753145799931982, |
|
"grad_norm": 51388.640625, |
|
"learning_rate": 1.6234271000340086e-05, |
|
"loss": 0.3649, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 0.6801729582665307, |
|
"grad_norm": 50291.19921875, |
|
"learning_rate": 1.599135208667347e-05, |
|
"loss": 0.3609, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.685031336539863, |
|
"grad_norm": 57660.2578125, |
|
"learning_rate": 1.5748433173006852e-05, |
|
"loss": 0.3719, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 0.6898897148131954, |
|
"grad_norm": 67932.59375, |
|
"learning_rate": 1.5505514259340236e-05, |
|
"loss": 0.3685, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 0.6947480930865277, |
|
"grad_norm": 54271.5234375, |
|
"learning_rate": 1.5262595345673615e-05, |
|
"loss": 0.376, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 0.6996064713598601, |
|
"grad_norm": 45938.90234375, |
|
"learning_rate": 1.5019676432006995e-05, |
|
"loss": 0.3555, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 0.7044648496331924, |
|
"grad_norm": 65276.1328125, |
|
"learning_rate": 1.4776757518340378e-05, |
|
"loss": 0.382, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.7093232279065248, |
|
"grad_norm": 52780.53125, |
|
"learning_rate": 1.453383860467376e-05, |
|
"loss": 0.3627, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 0.7141816061798572, |
|
"grad_norm": 66279.4375, |
|
"learning_rate": 1.4290919691007143e-05, |
|
"loss": 0.3726, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 0.7190399844531895, |
|
"grad_norm": 48593.3359375, |
|
"learning_rate": 1.4048000777340525e-05, |
|
"loss": 0.3623, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 0.7238983627265219, |
|
"grad_norm": 47858.83203125, |
|
"learning_rate": 1.3805081863673908e-05, |
|
"loss": 0.3605, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 0.7287567409998542, |
|
"grad_norm": 55828.08203125, |
|
"learning_rate": 1.3562162950007288e-05, |
|
"loss": 0.3559, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.7336151192731866, |
|
"grad_norm": 50614.34765625, |
|
"learning_rate": 1.331924403634067e-05, |
|
"loss": 0.3534, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 0.7384734975465189, |
|
"grad_norm": 48779.75, |
|
"learning_rate": 1.3076325122674053e-05, |
|
"loss": 0.3593, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 0.7433318758198514, |
|
"grad_norm": 50229.625, |
|
"learning_rate": 1.2833406209007434e-05, |
|
"loss": 0.3594, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 0.7481902540931837, |
|
"grad_norm": 47616.875, |
|
"learning_rate": 1.2590487295340816e-05, |
|
"loss": 0.3417, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 0.7530486323665161, |
|
"grad_norm": 51565.13671875, |
|
"learning_rate": 1.2347568381674197e-05, |
|
"loss": 0.3556, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.7579070106398484, |
|
"grad_norm": 159462.1875, |
|
"learning_rate": 1.210464946800758e-05, |
|
"loss": 0.3641, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 0.7627653889131808, |
|
"grad_norm": 54541.4375, |
|
"learning_rate": 1.1861730554340962e-05, |
|
"loss": 0.362, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 0.7676237671865132, |
|
"grad_norm": 60987.12109375, |
|
"learning_rate": 1.1618811640674344e-05, |
|
"loss": 0.3573, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 0.7724821454598455, |
|
"grad_norm": 51971.9765625, |
|
"learning_rate": 1.1375892727007725e-05, |
|
"loss": 0.3595, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 0.7773405237331779, |
|
"grad_norm": 53555.78515625, |
|
"learning_rate": 1.1132973813341108e-05, |
|
"loss": 0.3617, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.7821989020065102, |
|
"grad_norm": 50965.8984375, |
|
"learning_rate": 1.0890054899674488e-05, |
|
"loss": 0.356, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 0.7870572802798426, |
|
"grad_norm": 52491.86328125, |
|
"learning_rate": 1.0647135986007871e-05, |
|
"loss": 0.3442, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 0.7919156585531749, |
|
"grad_norm": 46180.28515625, |
|
"learning_rate": 1.0404217072341253e-05, |
|
"loss": 0.3618, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 0.7967740368265073, |
|
"grad_norm": 43963.23828125, |
|
"learning_rate": 1.0161298158674636e-05, |
|
"loss": 0.3451, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 0.8016324150998396, |
|
"grad_norm": 50149.4140625, |
|
"learning_rate": 9.918379245008016e-06, |
|
"loss": 0.3515, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.8064907933731721, |
|
"grad_norm": 64846.9609375, |
|
"learning_rate": 9.6754603313414e-06, |
|
"loss": 0.3593, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 0.8113491716465044, |
|
"grad_norm": 59240.46484375, |
|
"learning_rate": 9.432541417674781e-06, |
|
"loss": 0.345, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 0.8162075499198368, |
|
"grad_norm": 51653.4765625, |
|
"learning_rate": 9.189622504008162e-06, |
|
"loss": 0.3486, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 0.8210659281931691, |
|
"grad_norm": 56696.0703125, |
|
"learning_rate": 8.946703590341544e-06, |
|
"loss": 0.3514, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 0.8259243064665015, |
|
"grad_norm": 44659.8515625, |
|
"learning_rate": 8.703784676674927e-06, |
|
"loss": 0.3538, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.8307826847398339, |
|
"grad_norm": 62480.4453125, |
|
"learning_rate": 8.460865763008307e-06, |
|
"loss": 0.3553, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 0.8356410630131662, |
|
"grad_norm": 49702.1796875, |
|
"learning_rate": 8.21794684934169e-06, |
|
"loss": 0.355, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 0.8404994412864986, |
|
"grad_norm": 51918.6171875, |
|
"learning_rate": 7.975027935675072e-06, |
|
"loss": 0.3516, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 0.8453578195598309, |
|
"grad_norm": 70460.1875, |
|
"learning_rate": 7.732109022008453e-06, |
|
"loss": 0.3485, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 0.8502161978331633, |
|
"grad_norm": 39541.1328125, |
|
"learning_rate": 7.489190108341836e-06, |
|
"loss": 0.3682, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.8550745761064956, |
|
"grad_norm": 55487.01171875, |
|
"learning_rate": 7.246271194675218e-06, |
|
"loss": 0.3634, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 0.859932954379828, |
|
"grad_norm": 61263.55859375, |
|
"learning_rate": 7.0033522810086005e-06, |
|
"loss": 0.3661, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 0.8647913326531603, |
|
"grad_norm": 51811.1875, |
|
"learning_rate": 6.760433367341981e-06, |
|
"loss": 0.3545, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 0.8696497109264928, |
|
"grad_norm": 58568.05078125, |
|
"learning_rate": 6.517514453675364e-06, |
|
"loss": 0.3511, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 0.8745080891998251, |
|
"grad_norm": 42418.73828125, |
|
"learning_rate": 6.274595540008746e-06, |
|
"loss": 0.3462, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.8793664674731575, |
|
"grad_norm": 66980.625, |
|
"learning_rate": 6.0316766263421275e-06, |
|
"loss": 0.3429, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 0.8842248457464899, |
|
"grad_norm": 51797.6875, |
|
"learning_rate": 5.788757712675509e-06, |
|
"loss": 0.3457, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 0.8890832240198222, |
|
"grad_norm": 51049.6640625, |
|
"learning_rate": 5.5458387990088915e-06, |
|
"loss": 0.3462, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 0.8939416022931546, |
|
"grad_norm": 53216.921875, |
|
"learning_rate": 5.302919885342273e-06, |
|
"loss": 0.3482, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 0.8987999805664869, |
|
"grad_norm": 52944.14453125, |
|
"learning_rate": 5.0600009716756545e-06, |
|
"loss": 0.3429, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.9036583588398193, |
|
"grad_norm": 55617.85546875, |
|
"learning_rate": 4.817082058009037e-06, |
|
"loss": 0.3496, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 0.9085167371131516, |
|
"grad_norm": 51810.25390625, |
|
"learning_rate": 4.5741631443424185e-06, |
|
"loss": 0.3406, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 0.913375115386484, |
|
"grad_norm": 62037.01171875, |
|
"learning_rate": 4.331244230675801e-06, |
|
"loss": 0.3393, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 0.9182334936598163, |
|
"grad_norm": 55568.06640625, |
|
"learning_rate": 4.088325317009182e-06, |
|
"loss": 0.3412, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 0.9230918719331487, |
|
"grad_norm": 55022.734375, |
|
"learning_rate": 3.845406403342564e-06, |
|
"loss": 0.3297, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.927950250206481, |
|
"grad_norm": 59076.30078125, |
|
"learning_rate": 3.6024874896759463e-06, |
|
"loss": 0.3311, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 0.9328086284798134, |
|
"grad_norm": 55583.625, |
|
"learning_rate": 3.3595685760093283e-06, |
|
"loss": 0.3334, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 0.9376670067531458, |
|
"grad_norm": 44107.91015625, |
|
"learning_rate": 3.1166496623427103e-06, |
|
"loss": 0.3287, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 0.9425253850264782, |
|
"grad_norm": 43654.265625, |
|
"learning_rate": 2.8737307486760922e-06, |
|
"loss": 0.3501, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 0.9473837632998106, |
|
"grad_norm": 53803.21875, |
|
"learning_rate": 2.630811835009474e-06, |
|
"loss": 0.3418, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.9522421415731429, |
|
"grad_norm": 67202.1953125, |
|
"learning_rate": 2.3878929213428557e-06, |
|
"loss": 0.3343, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 0.9571005198464753, |
|
"grad_norm": 57509.7734375, |
|
"learning_rate": 2.1449740076762377e-06, |
|
"loss": 0.3397, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 0.9619588981198076, |
|
"grad_norm": 45354.30859375, |
|
"learning_rate": 1.9020550940096197e-06, |
|
"loss": 0.3467, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 0.96681727639314, |
|
"grad_norm": 51548.08984375, |
|
"learning_rate": 1.6591361803430016e-06, |
|
"loss": 0.3619, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 0.9716756546664723, |
|
"grad_norm": 65316.8515625, |
|
"learning_rate": 1.4162172666763834e-06, |
|
"loss": 0.3216, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.9765340329398047, |
|
"grad_norm": 48864.44921875, |
|
"learning_rate": 1.1732983530097653e-06, |
|
"loss": 0.3303, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 0.981392411213137, |
|
"grad_norm": 48878.0, |
|
"learning_rate": 9.303794393431472e-07, |
|
"loss": 0.3492, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 0.9862507894864694, |
|
"grad_norm": 43658.50390625, |
|
"learning_rate": 6.874605256765293e-07, |
|
"loss": 0.3306, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 0.9911091677598017, |
|
"grad_norm": 58298.01171875, |
|
"learning_rate": 4.4454161200991114e-07, |
|
"loss": 0.3366, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 0.9959675460331341, |
|
"grad_norm": 49216.18359375, |
|
"learning_rate": 2.01622698343293e-07, |
|
"loss": 0.3288, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 20583, |
|
"total_flos": 3.1397459461543035e+18, |
|
"train_loss": 0.4062042611626282, |
|
"train_runtime": 286226.4367, |
|
"train_samples_per_second": 0.863, |
|
"train_steps_per_second": 0.072 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 20583, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 20583, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.1397459461543035e+18, |
|
"train_batch_size": 12, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|