|
{ |
|
"best_metric": 1.1806877851486206, |
|
"best_model_checkpoint": "model_training/reprover/checkpoints-by_file-09-07-16-31/checkpoint-450", |
|
"epoch": 7.309644670050761, |
|
"eval_steps": 25, |
|
"global_step": 450, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.08121827411167512, |
|
"grad_norm": 24.31340217590332, |
|
"learning_rate": 5.102040816326531e-05, |
|
"loss": 12.4632, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.16243654822335024, |
|
"grad_norm": 7.6981658935546875, |
|
"learning_rate": 0.00010204081632653062, |
|
"loss": 7.3045, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.2436548223350254, |
|
"grad_norm": 0.584043562412262, |
|
"learning_rate": 0.00015306122448979594, |
|
"loss": 3.8878, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.3248730964467005, |
|
"grad_norm": 0.229707270860672, |
|
"learning_rate": 0.00020408163265306123, |
|
"loss": 3.6345, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.40609137055837563, |
|
"grad_norm": 0.2534404695034027, |
|
"learning_rate": 0.00025510204081632655, |
|
"loss": 3.508, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.40609137055837563, |
|
"eval_loss": 3.2811102867126465, |
|
"eval_runtime": 86.0155, |
|
"eval_samples_per_second": 11.626, |
|
"eval_steps_per_second": 1.453, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.4873096446700508, |
|
"grad_norm": 2.226205587387085, |
|
"learning_rate": 0.0003061224489795919, |
|
"loss": 3.4572, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.5685279187817259, |
|
"grad_norm": 2.194638729095459, |
|
"learning_rate": 0.00035714285714285714, |
|
"loss": 2.504, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.649746192893401, |
|
"grad_norm": 0.7849161624908447, |
|
"learning_rate": 0.00040816326530612246, |
|
"loss": 1.9298, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.7309644670050761, |
|
"grad_norm": 0.5836678147315979, |
|
"learning_rate": 0.0004591836734693878, |
|
"loss": 1.7631, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.8121827411167513, |
|
"grad_norm": 0.4033905565738678, |
|
"learning_rate": 0.0004999935985425297, |
|
"loss": 1.6774, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.8121827411167513, |
|
"eval_loss": 1.367048740386963, |
|
"eval_runtime": 85.8218, |
|
"eval_samples_per_second": 11.652, |
|
"eval_steps_per_second": 1.457, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.8934010152284264, |
|
"grad_norm": 0.3479253947734833, |
|
"learning_rate": 0.0004997695819512612, |
|
"loss": 1.6005, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.9746192893401016, |
|
"grad_norm": 0.4574773907661438, |
|
"learning_rate": 0.0004992258202402822, |
|
"loss": 1.5389, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.0558375634517767, |
|
"grad_norm": 0.2955416738986969, |
|
"learning_rate": 0.0004983630095117843, |
|
"loss": 1.4855, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.1370558375634519, |
|
"grad_norm": 0.235136017203331, |
|
"learning_rate": 0.0004971822543018662, |
|
"loss": 1.4427, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.218274111675127, |
|
"grad_norm": 0.2502729594707489, |
|
"learning_rate": 0.0004956850661665511, |
|
"loss": 1.462, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.218274111675127, |
|
"eval_loss": 1.2605030536651611, |
|
"eval_runtime": 85.8507, |
|
"eval_samples_per_second": 11.648, |
|
"eval_steps_per_second": 1.456, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.299492385786802, |
|
"grad_norm": 0.46766167879104614, |
|
"learning_rate": 0.0004938733617467517, |
|
"loss": 1.4421, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.380710659898477, |
|
"grad_norm": 0.2875135540962219, |
|
"learning_rate": 0.0004917494603146632, |
|
"loss": 1.4042, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 1.4619289340101522, |
|
"grad_norm": 0.5056741833686829, |
|
"learning_rate": 0.0004893160808047222, |
|
"loss": 1.3901, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.5431472081218274, |
|
"grad_norm": 0.17907263338565826, |
|
"learning_rate": 0.00048657633833293557, |
|
"loss": 1.3829, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 1.6243654822335025, |
|
"grad_norm": 0.25355765223503113, |
|
"learning_rate": 0.0004835337402090316, |
|
"loss": 1.3592, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.6243654822335025, |
|
"eval_loss": 1.2396838665008545, |
|
"eval_runtime": 85.75, |
|
"eval_samples_per_second": 11.662, |
|
"eval_steps_per_second": 1.458, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.7055837563451777, |
|
"grad_norm": 0.1514243483543396, |
|
"learning_rate": 0.0004801921814465414, |
|
"loss": 1.379, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 1.7868020304568528, |
|
"grad_norm": 0.14222057163715363, |
|
"learning_rate": 0.00047655593977655674, |
|
"loss": 1.3385, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.868020304568528, |
|
"grad_norm": 0.38187268376350403, |
|
"learning_rate": 0.0004726296701715489, |
|
"loss": 1.339, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 1.9492385786802031, |
|
"grad_norm": 0.42166340351104736, |
|
"learning_rate": 0.00046841839888625623, |
|
"loss": 1.4193, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.030456852791878, |
|
"grad_norm": 0.3789190948009491, |
|
"learning_rate": 0.0004639275170232734, |
|
"loss": 1.3529, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.030456852791878, |
|
"eval_loss": 1.2485618591308594, |
|
"eval_runtime": 85.8636, |
|
"eval_samples_per_second": 11.646, |
|
"eval_steps_per_second": 1.456, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.1116751269035534, |
|
"grad_norm": 0.3297707736492157, |
|
"learning_rate": 0.0004591627736315743, |
|
"loss": 1.3754, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 2.1928934010152283, |
|
"grad_norm": 0.6720882058143616, |
|
"learning_rate": 0.0004541302683468084, |
|
"loss": 1.3633, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 2.2741116751269037, |
|
"grad_norm": 0.423013299703598, |
|
"learning_rate": 0.0004488364435827881, |
|
"loss": 1.3739, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 2.3553299492385786, |
|
"grad_norm": 0.3096015751361847, |
|
"learning_rate": 0.00044328807628416644, |
|
"loss": 1.3734, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 2.436548223350254, |
|
"grad_norm": 0.25548315048217773, |
|
"learning_rate": 0.0004374922692508611, |
|
"loss": 1.354, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.436548223350254, |
|
"eval_loss": 1.2303425073623657, |
|
"eval_runtime": 85.8621, |
|
"eval_samples_per_second": 11.647, |
|
"eval_steps_per_second": 1.456, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.517766497461929, |
|
"grad_norm": 0.2693212330341339, |
|
"learning_rate": 0.0004314564420453311, |
|
"loss": 1.3311, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 2.598984771573604, |
|
"grad_norm": 0.2327384501695633, |
|
"learning_rate": 0.0004251883214943475, |
|
"loss": 1.3367, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 2.6802030456852792, |
|
"grad_norm": 0.20477798581123352, |
|
"learning_rate": 0.0004186959317974155, |
|
"loss": 1.3718, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 2.761421319796954, |
|
"grad_norm": 0.6859282851219177, |
|
"learning_rate": 0.00041198758425451266, |
|
"loss": 1.3351, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 2.8426395939086295, |
|
"grad_norm": 0.4204564392566681, |
|
"learning_rate": 0.00040507186662629185, |
|
"loss": 1.3633, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 2.8426395939086295, |
|
"eval_loss": 1.2046416997909546, |
|
"eval_runtime": 85.7507, |
|
"eval_samples_per_second": 11.662, |
|
"eval_steps_per_second": 1.458, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 2.9238578680203045, |
|
"grad_norm": 0.4428386390209198, |
|
"learning_rate": 0.0003979576321403705, |
|
"loss": 1.2933, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 3.00507614213198, |
|
"grad_norm": 0.5064650774002075, |
|
"learning_rate": 0.0003906539881577793, |
|
"loss": 1.3117, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 3.0862944162436547, |
|
"grad_norm": 0.19443967938423157, |
|
"learning_rate": 0.0003831702845140801, |
|
"loss": 1.3285, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 3.16751269035533, |
|
"grad_norm": 0.17194636166095734, |
|
"learning_rate": 0.00037551610155007613, |
|
"loss": 1.2587, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 3.248730964467005, |
|
"grad_norm": 0.4715280830860138, |
|
"learning_rate": 0.00036770123784744027, |
|
"loss": 1.3266, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 3.248730964467005, |
|
"eval_loss": 1.209210991859436, |
|
"eval_runtime": 85.8547, |
|
"eval_samples_per_second": 11.648, |
|
"eval_steps_per_second": 1.456, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 3.3299492385786804, |
|
"grad_norm": 0.5037745237350464, |
|
"learning_rate": 0.00035973569768495855, |
|
"loss": 1.3085, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 3.4111675126903553, |
|
"grad_norm": 0.4525098204612732, |
|
"learning_rate": 0.0003516296782314491, |
|
"loss": 1.2944, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 3.4923857868020303, |
|
"grad_norm": 0.30172044038772583, |
|
"learning_rate": 0.00034339355649175095, |
|
"loss": 1.3206, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 3.5736040609137056, |
|
"grad_norm": 0.8903113603591919, |
|
"learning_rate": 0.00033503787602249364, |
|
"loss": 1.3404, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 3.6548223350253806, |
|
"grad_norm": 0.5123381614685059, |
|
"learning_rate": 0.00032657333343465356, |
|
"loss": 1.321, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 3.6548223350253806, |
|
"eval_loss": 1.200507640838623, |
|
"eval_runtime": 85.8377, |
|
"eval_samples_per_second": 11.65, |
|
"eval_steps_per_second": 1.456, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 3.736040609137056, |
|
"grad_norm": 0.19661971926689148, |
|
"learning_rate": 0.0003180107647001769, |
|
"loss": 1.3276, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 3.817258883248731, |
|
"grad_norm": 0.34714922308921814, |
|
"learning_rate": 0.0003093611312801979, |
|
"loss": 1.2869, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 3.8984771573604062, |
|
"grad_norm": 0.3150271773338318, |
|
"learning_rate": 0.00030063550609261025, |
|
"loss": 1.2738, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 3.979695431472081, |
|
"grad_norm": 0.24157479405403137, |
|
"learning_rate": 0.000291845059336957, |
|
"loss": 1.3107, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 4.060913705583756, |
|
"grad_norm": 0.2525878846645355, |
|
"learning_rate": 0.0002830010441947834, |
|
"loss": 1.2679, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 4.060913705583756, |
|
"eval_loss": 1.196422815322876, |
|
"eval_runtime": 85.8149, |
|
"eval_samples_per_second": 11.653, |
|
"eval_steps_per_second": 1.457, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 4.1421319796954315, |
|
"grad_norm": 0.49697160720825195, |
|
"learning_rate": 0.00027411478242376017, |
|
"loss": 1.3141, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 4.223350253807107, |
|
"grad_norm": 0.3860456943511963, |
|
"learning_rate": 0.00026519764986401774, |
|
"loss": 1.3039, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 4.304568527918782, |
|
"grad_norm": 0.2929648160934448, |
|
"learning_rate": 0.000256261061875247, |
|
"loss": 1.2864, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 4.385786802030457, |
|
"grad_norm": 0.49687063694000244, |
|
"learning_rate": 0.0002473164587232079, |
|
"loss": 1.3065, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 4.467005076142132, |
|
"grad_norm": 0.26030004024505615, |
|
"learning_rate": 0.0002383752909343547, |
|
"loss": 1.299, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 4.467005076142132, |
|
"eval_loss": 1.1841660737991333, |
|
"eval_runtime": 85.8048, |
|
"eval_samples_per_second": 11.654, |
|
"eval_steps_per_second": 1.457, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 4.548223350253807, |
|
"grad_norm": 0.1852976679801941, |
|
"learning_rate": 0.0002294490046373259, |
|
"loss": 1.2506, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 4.629441624365482, |
|
"grad_norm": 0.41372132301330566, |
|
"learning_rate": 0.00022054902691006405, |
|
"loss": 1.3174, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 4.710659898477157, |
|
"grad_norm": 0.18286170065402985, |
|
"learning_rate": 0.00021168675115132315, |
|
"loss": 1.3335, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 4.791878172588833, |
|
"grad_norm": 0.1263434737920761, |
|
"learning_rate": 0.00020287352249529153, |
|
"loss": 1.3019, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 4.873096446700508, |
|
"grad_norm": 0.13266858458518982, |
|
"learning_rate": 0.00019412062328800044, |
|
"loss": 1.2709, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 4.873096446700508, |
|
"eval_loss": 1.1873031854629517, |
|
"eval_runtime": 85.8269, |
|
"eval_samples_per_second": 11.651, |
|
"eval_steps_per_second": 1.456, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 4.9543147208121825, |
|
"grad_norm": 0.1231088787317276, |
|
"learning_rate": 0.000185439258644112, |
|
"loss": 1.3102, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 5.035532994923858, |
|
"grad_norm": 0.16395896673202515, |
|
"learning_rate": 0.00017684054210257517, |
|
"loss": 1.3088, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 5.116751269035533, |
|
"grad_norm": 0.14288240671157837, |
|
"learning_rate": 0.00016833548139951395, |
|
"loss": 1.3205, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 5.197969543147208, |
|
"grad_norm": 0.16564807295799255, |
|
"learning_rate": 0.0001599349643765599, |
|
"loss": 1.2845, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 5.279187817258883, |
|
"grad_norm": 0.133669912815094, |
|
"learning_rate": 0.0001516497450426686, |
|
"loss": 1.3105, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 5.279187817258883, |
|
"eval_loss": 1.183776617050171, |
|
"eval_runtime": 85.7228, |
|
"eval_samples_per_second": 11.666, |
|
"eval_steps_per_second": 1.458, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 5.3604060913705585, |
|
"grad_norm": 0.050202421844005585, |
|
"learning_rate": 0.00014349042980726362, |
|
"loss": 1.2825, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 5.441624365482234, |
|
"grad_norm": 0.08104140311479568, |
|
"learning_rate": 0.0001354674639023318, |
|
"loss": 1.2975, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 5.522842639593908, |
|
"grad_norm": 0.045941274613142014, |
|
"learning_rate": 0.00012759111801085066, |
|
"loss": 1.2848, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 5.604060913705584, |
|
"grad_norm": 0.10393151640892029, |
|
"learning_rate": 0.00011987147511866788, |
|
"loss": 1.3075, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 5.685279187817259, |
|
"grad_norm": 0.05486277490854263, |
|
"learning_rate": 0.00011231841760666186, |
|
"loss": 1.2337, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 5.685279187817259, |
|
"eval_loss": 1.1837860345840454, |
|
"eval_runtime": 85.8492, |
|
"eval_samples_per_second": 11.648, |
|
"eval_steps_per_second": 1.456, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 5.7664974619289335, |
|
"grad_norm": 0.06317424029111862, |
|
"learning_rate": 0.0001049416145997094, |
|
"loss": 1.2696, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 5.847715736040609, |
|
"grad_norm": 0.06952204555273056, |
|
"learning_rate": 9.775050958865584e-05, |
|
"loss": 1.2792, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 5.928934010152284, |
|
"grad_norm": 0.05977623909711838, |
|
"learning_rate": 9.075430834113152e-05, |
|
"loss": 1.2672, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 6.01015228426396, |
|
"grad_norm": 0.1008043885231018, |
|
"learning_rate": 8.396196711669335e-05, |
|
"loss": 1.3045, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 6.091370558375634, |
|
"grad_norm": 0.06693598628044128, |
|
"learning_rate": 7.738218120137671e-05, |
|
"loss": 1.2753, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 6.091370558375634, |
|
"eval_loss": 1.1829280853271484, |
|
"eval_runtime": 85.8364, |
|
"eval_samples_per_second": 11.65, |
|
"eval_steps_per_second": 1.456, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 6.1725888324873095, |
|
"grad_norm": 0.10474739223718643, |
|
"learning_rate": 7.102337377633394e-05, |
|
"loss": 1.2936, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 6.253807106598985, |
|
"grad_norm": 0.05388178676366806, |
|
"learning_rate": 6.489368513481228e-05, |
|
"loss": 1.2439, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 6.33502538071066, |
|
"grad_norm": 0.09186024218797684, |
|
"learning_rate": 5.9000962261273136e-05, |
|
"loss": 1.3222, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 6.416243654822335, |
|
"grad_norm": 0.07748736441135406, |
|
"learning_rate": 5.3352748785993164e-05, |
|
"loss": 1.2731, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 6.49746192893401, |
|
"grad_norm": 0.07806465774774551, |
|
"learning_rate": 4.795627532800806e-05, |
|
"loss": 1.2849, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 6.49746192893401, |
|
"eval_loss": 1.1841729879379272, |
|
"eval_runtime": 85.8225, |
|
"eval_samples_per_second": 11.652, |
|
"eval_steps_per_second": 1.456, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 6.5786802030456855, |
|
"grad_norm": 0.05898735672235489, |
|
"learning_rate": 4.281845023876074e-05, |
|
"loss": 1.2866, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 6.659898477157361, |
|
"grad_norm": 0.052701685577631, |
|
"learning_rate": 3.794585075830329e-05, |
|
"loss": 1.2848, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 6.741116751269035, |
|
"grad_norm": 0.04117720574140549, |
|
"learning_rate": 3.334471459537497e-05, |
|
"loss": 1.2859, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 6.822335025380711, |
|
"grad_norm": 0.04338378086686134, |
|
"learning_rate": 2.902093194213526e-05, |
|
"loss": 1.2934, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 6.903553299492386, |
|
"grad_norm": 0.05017438903450966, |
|
"learning_rate": 2.4980037933772488e-05, |
|
"loss": 1.2508, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 6.903553299492386, |
|
"eval_loss": 1.1836540699005127, |
|
"eval_runtime": 85.78, |
|
"eval_samples_per_second": 11.658, |
|
"eval_steps_per_second": 1.457, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 6.9847715736040605, |
|
"grad_norm": 0.04826565086841583, |
|
"learning_rate": 2.122720556264357e-05, |
|
"loss": 1.2958, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 7.065989847715736, |
|
"grad_norm": 0.05151581019163132, |
|
"learning_rate": 1.776723905601438e-05, |
|
"loss": 1.279, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 7.147208121827411, |
|
"grad_norm": 0.06657218188047409, |
|
"learning_rate": 1.4604567725877926e-05, |
|
"loss": 1.2756, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 7.228426395939087, |
|
"grad_norm": 0.05859878286719322, |
|
"learning_rate": 1.1743240298725116e-05, |
|
"loss": 1.3084, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 7.309644670050761, |
|
"grad_norm": 0.051177285611629486, |
|
"learning_rate": 9.18691973252539e-06, |
|
"loss": 1.296, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 7.309644670050761, |
|
"eval_loss": 1.1806877851486206, |
|
"eval_runtime": 85.7014, |
|
"eval_samples_per_second": 11.668, |
|
"eval_steps_per_second": 1.459, |
|
"step": 450 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 488, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 8, |
|
"save_steps": 25, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.089179686194381e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|