|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 70.0, |
|
"eval_steps": 500, |
|
"global_step": 91910, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.004999955274106163, |
|
"loss": 9.6382, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 0.004999821098024974, |
|
"loss": 9.5189, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.00022625107915810782, |
|
"eval_loss": 9.40969181060791, |
|
"eval_runtime": 6168.4952, |
|
"eval_samples_per_second": 27.228, |
|
"eval_steps_per_second": 1.702, |
|
"step": 1313 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.0049995974765573515, |
|
"loss": 9.363, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 0.00499928441770463, |
|
"loss": 9.2724, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 0.00499888193266828, |
|
"loss": 9.1635, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.0006430293828704117, |
|
"eval_loss": 9.041797637939453, |
|
"eval_runtime": 4719.6922, |
|
"eval_samples_per_second": 35.586, |
|
"eval_steps_per_second": 2.224, |
|
"step": 2626 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 0.004998390035849502, |
|
"loss": 8.8459, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 0.0049978087448487194, |
|
"loss": 8.3432, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.006466017683308029, |
|
"eval_loss": 7.845377445220947, |
|
"eval_runtime": 2624.5595, |
|
"eval_samples_per_second": 63.994, |
|
"eval_steps_per_second": 4.0, |
|
"step": 3939 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 0.004997138080464938, |
|
"loss": 7.8952, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 0.004996378066695007, |
|
"loss": 7.4467, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 0.004995528730732766, |
|
"loss": 6.9913, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.04887023309815129, |
|
"eval_loss": 6.311844825744629, |
|
"eval_runtime": 2630.4579, |
|
"eval_samples_per_second": 63.85, |
|
"eval_steps_per_second": 3.991, |
|
"step": 5252 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 0.0049945901029680604, |
|
"loss": 6.4357, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 0.004993562216985665, |
|
"loss": 5.9619, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 0.004992445109564073, |
|
"loss": 5.5048, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_accuracy": 0.1423000208389152, |
|
"eval_loss": 4.949326515197754, |
|
"eval_runtime": 2650.8854, |
|
"eval_samples_per_second": 63.358, |
|
"eval_steps_per_second": 3.96, |
|
"step": 6565 |
|
}, |
|
{ |
|
"epoch": 5.33, |
|
"learning_rate": 0.004991238820674189, |
|
"loss": 4.982, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 5.71, |
|
"learning_rate": 0.00498994339347789, |
|
"loss": 4.6895, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_accuracy": 0.24502396475246346, |
|
"eval_loss": 3.960129499435425, |
|
"eval_runtime": 2617.6761, |
|
"eval_samples_per_second": 64.162, |
|
"eval_steps_per_second": 4.01, |
|
"step": 7878 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"learning_rate": 0.004988558874326489, |
|
"loss": 4.3479, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 6.47, |
|
"learning_rate": 0.004987085312759071, |
|
"loss": 3.9972, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 6.85, |
|
"learning_rate": 0.004985522761500722, |
|
"loss": 3.8881, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_accuracy": 0.31358994968890475, |
|
"eval_loss": 3.4186415672302246, |
|
"eval_runtime": 2616.4405, |
|
"eval_samples_per_second": 64.192, |
|
"eval_steps_per_second": 4.012, |
|
"step": 9191 |
|
}, |
|
{ |
|
"epoch": 7.24, |
|
"learning_rate": 0.004983871276460643, |
|
"loss": 3.6025, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 7.62, |
|
"learning_rate": 0.0049821309167301505, |
|
"loss": 3.4698, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 0.004980301744580561, |
|
"loss": 3.391, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_accuracy": 0.37655919740406657, |
|
"eval_loss": 2.9798192977905273, |
|
"eval_runtime": 2644.3802, |
|
"eval_samples_per_second": 63.514, |
|
"eval_steps_per_second": 3.97, |
|
"step": 10504 |
|
}, |
|
{ |
|
"epoch": 8.38, |
|
"learning_rate": 0.004978383825460959, |
|
"loss": 3.1113, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 8.76, |
|
"learning_rate": 0.004976377227995865, |
|
"loss": 3.0887, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_accuracy": 0.422124973951356, |
|
"eval_loss": 2.705430030822754, |
|
"eval_runtime": 2622.7588, |
|
"eval_samples_per_second": 64.038, |
|
"eval_steps_per_second": 4.003, |
|
"step": 11817 |
|
}, |
|
{ |
|
"epoch": 9.14, |
|
"learning_rate": 0.00497428202398277, |
|
"loss": 2.925, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 9.52, |
|
"learning_rate": 0.004972098288389572, |
|
"loss": 2.7977, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 9.9, |
|
"learning_rate": 0.004969826099351891, |
|
"loss": 2.7935, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_accuracy": 0.45515167753267244, |
|
"eval_loss": 2.501314401626587, |
|
"eval_runtime": 2679.1466, |
|
"eval_samples_per_second": 62.69, |
|
"eval_steps_per_second": 3.918, |
|
"step": 13130 |
|
}, |
|
{ |
|
"epoch": 10.28, |
|
"learning_rate": 0.004967465538170278, |
|
"loss": 2.5817, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 10.66, |
|
"learning_rate": 0.004965016689307298, |
|
"loss": 2.5629, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_accuracy": 0.4803846268345688, |
|
"eval_loss": 2.3581297397613525, |
|
"eval_runtime": 2615.9499, |
|
"eval_samples_per_second": 64.204, |
|
"eval_steps_per_second": 4.013, |
|
"step": 14443 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 0.004962479640384515, |
|
"loss": 2.5549, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 11.42, |
|
"learning_rate": 0.004959854482179353, |
|
"loss": 2.3342, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 11.81, |
|
"learning_rate": 0.004957141308621851, |
|
"loss": 2.3777, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_accuracy": 0.48086689887172157, |
|
"eval_loss": 2.3543388843536377, |
|
"eval_runtime": 2613.6118, |
|
"eval_samples_per_second": 64.262, |
|
"eval_steps_per_second": 4.017, |
|
"step": 15756 |
|
}, |
|
{ |
|
"epoch": 12.19, |
|
"learning_rate": 0.004954340216791298, |
|
"loss": 2.1546, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 12.57, |
|
"learning_rate": 0.004951451306912763, |
|
"loss": 2.2162, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 12.95, |
|
"learning_rate": 0.0049484746823535075, |
|
"loss": 2.2264, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_accuracy": 0.5178887201929088, |
|
"eval_loss": 2.163175344467163, |
|
"eval_runtime": 5970.7626, |
|
"eval_samples_per_second": 28.13, |
|
"eval_steps_per_second": 1.758, |
|
"step": 17069 |
|
}, |
|
{ |
|
"epoch": 13.33, |
|
"learning_rate": 0.004945410449619286, |
|
"loss": 2.035, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 13.71, |
|
"learning_rate": 0.0049422587183505385, |
|
"loss": 2.0932, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_accuracy": 0.5219076538358489, |
|
"eval_loss": 2.136173725128174, |
|
"eval_runtime": 2759.3303, |
|
"eval_samples_per_second": 60.868, |
|
"eval_steps_per_second": 3.805, |
|
"step": 18382 |
|
}, |
|
{ |
|
"epoch": 14.09, |
|
"learning_rate": 0.004939019601318461, |
|
"loss": 2.0545, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 14.47, |
|
"learning_rate": 0.004935693214420979, |
|
"loss": 1.9224, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 14.85, |
|
"learning_rate": 0.004932279676678594, |
|
"loss": 1.9667, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_accuracy": 0.5591200023815903, |
|
"eval_loss": 1.956689715385437, |
|
"eval_runtime": 2781.1993, |
|
"eval_samples_per_second": 60.389, |
|
"eval_steps_per_second": 3.775, |
|
"step": 19695 |
|
}, |
|
{ |
|
"epoch": 15.23, |
|
"learning_rate": 0.004928779110230126, |
|
"loss": 1.8635, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 15.61, |
|
"learning_rate": 0.004925191640328348, |
|
"loss": 1.8396, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 15.99, |
|
"learning_rate": 0.004921517395335496, |
|
"loss": 1.8788, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_accuracy": 0.5609538269179244, |
|
"eval_loss": 1.9347097873687744, |
|
"eval_runtime": 2696.6196, |
|
"eval_samples_per_second": 62.284, |
|
"eval_steps_per_second": 3.893, |
|
"step": 21008 |
|
}, |
|
{ |
|
"epoch": 16.37, |
|
"learning_rate": 0.004917756506718685, |
|
"loss": 1.7225, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 16.76, |
|
"learning_rate": 0.004913909109045198, |
|
"loss": 1.7705, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_accuracy": 0.5684439284332112, |
|
"eval_loss": 1.948270559310913, |
|
"eval_runtime": 2619.7155, |
|
"eval_samples_per_second": 64.112, |
|
"eval_steps_per_second": 4.007, |
|
"step": 22321 |
|
}, |
|
{ |
|
"epoch": 17.14, |
|
"learning_rate": 0.004909975339977676, |
|
"loss": 1.7215, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 17.52, |
|
"learning_rate": 0.004905955340269188, |
|
"loss": 1.6354, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 17.9, |
|
"learning_rate": 0.004901849253758198, |
|
"loss": 1.7089, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_accuracy": 0.5790598672263404, |
|
"eval_loss": 1.8927983045578003, |
|
"eval_runtime": 2625.6089, |
|
"eval_samples_per_second": 63.968, |
|
"eval_steps_per_second": 3.998, |
|
"step": 23634 |
|
}, |
|
{ |
|
"epoch": 18.28, |
|
"learning_rate": 0.004897657227363419, |
|
"loss": 1.579, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 18.66, |
|
"learning_rate": 0.004893379411078551, |
|
"loss": 1.6068, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_accuracy": 0.5855497008127177, |
|
"eval_loss": 1.843549370765686, |
|
"eval_runtime": 2690.4067, |
|
"eval_samples_per_second": 62.427, |
|
"eval_steps_per_second": 3.902, |
|
"step": 24947 |
|
}, |
|
{ |
|
"epoch": 19.04, |
|
"learning_rate": 0.004889015957966921, |
|
"loss": 1.6159, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 19.42, |
|
"learning_rate": 0.004884567024156001, |
|
"loss": 1.4949, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 19.8, |
|
"learning_rate": 0.004880032768831825, |
|
"loss": 1.5572, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_accuracy": 0.5879670149742491, |
|
"eval_loss": 1.840835690498352, |
|
"eval_runtime": 2628.5762, |
|
"eval_samples_per_second": 63.896, |
|
"eval_steps_per_second": 3.994, |
|
"step": 26260 |
|
}, |
|
{ |
|
"epoch": 20.18, |
|
"learning_rate": 0.004875413354233289, |
|
"loss": 1.4928, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 20.56, |
|
"learning_rate": 0.0048707089456463525, |
|
"loss": 1.4466, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 20.94, |
|
"learning_rate": 0.004865919711398117, |
|
"loss": 1.4938, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_accuracy": 0.6109672233634009, |
|
"eval_loss": 1.7412927150726318, |
|
"eval_runtime": 2614.3175, |
|
"eval_samples_per_second": 64.244, |
|
"eval_steps_per_second": 4.016, |
|
"step": 27573 |
|
}, |
|
{ |
|
"epoch": 21.33, |
|
"learning_rate": 0.0048610458228508105, |
|
"loss": 1.3796, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 21.71, |
|
"learning_rate": 0.004856087454395648, |
|
"loss": 1.4182, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_accuracy": 0.6154565210919591, |
|
"eval_loss": 1.7196179628372192, |
|
"eval_runtime": 2640.6742, |
|
"eval_samples_per_second": 63.603, |
|
"eval_steps_per_second": 3.975, |
|
"step": 28886 |
|
}, |
|
{ |
|
"epoch": 22.09, |
|
"learning_rate": 0.0048510447834466, |
|
"loss": 1.4096, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 22.47, |
|
"learning_rate": 0.0048459179904340385, |
|
"loss": 1.3308, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 22.85, |
|
"learning_rate": 0.004840707258798284, |
|
"loss": 1.3784, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_accuracy": 0.6238218570450418, |
|
"eval_loss": 1.7105211019515991, |
|
"eval_runtime": 2885.4428, |
|
"eval_samples_per_second": 58.208, |
|
"eval_steps_per_second": 3.638, |
|
"step": 30199 |
|
}, |
|
{ |
|
"epoch": 23.23, |
|
"learning_rate": 0.004835412774983039, |
|
"loss": 1.3104, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 23.61, |
|
"learning_rate": 0.004830034728428723, |
|
"loss": 1.301, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 23.99, |
|
"learning_rate": 0.004824573311565684, |
|
"loss": 1.3578, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_accuracy": 0.6176059063439612, |
|
"eval_loss": 1.7758694887161255, |
|
"eval_runtime": 2617.3025, |
|
"eval_samples_per_second": 64.171, |
|
"eval_steps_per_second": 4.011, |
|
"step": 31512 |
|
}, |
|
{ |
|
"epoch": 24.37, |
|
"learning_rate": 0.004819028719807326, |
|
"loss": 1.2176, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 24.75, |
|
"learning_rate": 0.004813401151543105, |
|
"loss": 1.2763, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_accuracy": 0.6218749069691286, |
|
"eval_loss": 1.7364755868911743, |
|
"eval_runtime": 5925.4842, |
|
"eval_samples_per_second": 28.345, |
|
"eval_steps_per_second": 1.772, |
|
"step": 32825 |
|
}, |
|
{ |
|
"epoch": 25.13, |
|
"learning_rate": 0.004807690808131437, |
|
"loss": 1.2551, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 25.51, |
|
"learning_rate": 0.004801897893892493, |
|
"loss": 1.1861, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 25.89, |
|
"learning_rate": 0.004796022616100888, |
|
"loss": 1.2484, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_accuracy": 0.6199220029174481, |
|
"eval_loss": 1.7483400106430054, |
|
"eval_runtime": 2684.7282, |
|
"eval_samples_per_second": 62.559, |
|
"eval_steps_per_second": 3.91, |
|
"step": 34138 |
|
}, |
|
{ |
|
"epoch": 26.28, |
|
"learning_rate": 0.004790065184978262, |
|
"loss": 1.1726, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 26.66, |
|
"learning_rate": 0.004784025813685761, |
|
"loss": 1.1936, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_accuracy": 0.6313953142210711, |
|
"eval_loss": 1.700264573097229, |
|
"eval_runtime": 2640.722, |
|
"eval_samples_per_second": 63.602, |
|
"eval_steps_per_second": 3.975, |
|
"step": 35451 |
|
}, |
|
{ |
|
"epoch": 27.04, |
|
"learning_rate": 0.004777904718316407, |
|
"loss": 1.2231, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 27.42, |
|
"learning_rate": 0.00477170211788737, |
|
"loss": 1.1121, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 27.8, |
|
"learning_rate": 0.00476541823433213, |
|
"loss": 1.1499, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_accuracy": 0.6247209073859069, |
|
"eval_loss": 1.7398767471313477, |
|
"eval_runtime": 2691.471, |
|
"eval_samples_per_second": 62.403, |
|
"eval_steps_per_second": 3.9, |
|
"step": 36764 |
|
}, |
|
{ |
|
"epoch": 28.18, |
|
"learning_rate": 0.004759053292492532, |
|
"loss": 1.1329, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 28.56, |
|
"learning_rate": 0.004752607520110748, |
|
"loss": 1.1121, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 28.94, |
|
"learning_rate": 0.004746081147821121, |
|
"loss": 1.1418, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_accuracy": 0.6316751510821351, |
|
"eval_loss": 1.7091107368469238, |
|
"eval_runtime": 2632.2426, |
|
"eval_samples_per_second": 63.807, |
|
"eval_steps_per_second": 3.988, |
|
"step": 38077 |
|
}, |
|
{ |
|
"epoch": 29.32, |
|
"learning_rate": 0.0047394744091419205, |
|
"loss": 1.0493, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 29.7, |
|
"learning_rate": 0.004732787540466979, |
|
"loss": 1.0895, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_accuracy": 0.6383078800869281, |
|
"eval_loss": 1.7165930271148682, |
|
"eval_runtime": 2645.0605, |
|
"eval_samples_per_second": 63.498, |
|
"eval_steps_per_second": 3.969, |
|
"step": 39390 |
|
}, |
|
{ |
|
"epoch": 30.08, |
|
"learning_rate": 0.004726020781057241, |
|
"loss": 1.0905, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 30.46, |
|
"learning_rate": 0.004719174373032195, |
|
"loss": 1.0191, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 30.85, |
|
"learning_rate": 0.0047122485613612175, |
|
"loss": 1.0706, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_accuracy": 0.6373611979399244, |
|
"eval_loss": 1.7384090423583984, |
|
"eval_runtime": 2658.4828, |
|
"eval_samples_per_second": 63.177, |
|
"eval_steps_per_second": 3.949, |
|
"step": 40703 |
|
}, |
|
{ |
|
"epoch": 31.23, |
|
"learning_rate": 0.004705243593854802, |
|
"loss": 1.0205, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 31.61, |
|
"learning_rate": 0.004698159721155694, |
|
"loss": 1.0072, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 31.99, |
|
"learning_rate": 0.004690997196729925, |
|
"loss": 1.0541, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_accuracy": 0.640868089666875, |
|
"eval_loss": 1.7335774898529053, |
|
"eval_runtime": 2595.2091, |
|
"eval_samples_per_second": 64.717, |
|
"eval_steps_per_second": 4.045, |
|
"step": 42016 |
|
}, |
|
{ |
|
"epoch": 32.37, |
|
"learning_rate": 0.004683756276857741, |
|
"loss": 0.9544, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 32.75, |
|
"learning_rate": 0.004676437220624432, |
|
"loss": 1.0013, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_accuracy": 0.6451489982435772, |
|
"eval_loss": 1.7185142040252686, |
|
"eval_runtime": 2605.1367, |
|
"eval_samples_per_second": 64.471, |
|
"eval_steps_per_second": 4.03, |
|
"step": 43329 |
|
}, |
|
{ |
|
"epoch": 33.13, |
|
"learning_rate": 0.004669040289911064, |
|
"loss": 0.9986, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 33.51, |
|
"learning_rate": 0.004661565749385107, |
|
"loss": 0.9399, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 33.89, |
|
"learning_rate": 0.004654013866490966, |
|
"loss": 0.9811, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_accuracy": 0.6478640111934745, |
|
"eval_loss": 1.7246044874191284, |
|
"eval_runtime": 2595.1453, |
|
"eval_samples_per_second": 64.719, |
|
"eval_steps_per_second": 4.045, |
|
"step": 44642 |
|
}, |
|
{ |
|
"epoch": 34.27, |
|
"learning_rate": 0.004646384911440411, |
|
"loss": 0.914, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 34.65, |
|
"learning_rate": 0.004638679157202909, |
|
"loss": 0.9447, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"eval_accuracy": 0.6540263761126492, |
|
"eval_loss": 1.7245060205459595, |
|
"eval_runtime": 2605.4725, |
|
"eval_samples_per_second": 64.462, |
|
"eval_steps_per_second": 4.029, |
|
"step": 45955 |
|
}, |
|
{ |
|
"epoch": 35.03, |
|
"learning_rate": 0.0005810494104962897, |
|
"loss": 0.8424, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 35.42, |
|
"learning_rate": 0.0005391326120998649, |
|
"loss": 0.692, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 35.8, |
|
"learning_rate": 0.0004986016371937035, |
|
"loss": 0.6587, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_accuracy": 0.7018546634515197, |
|
"eval_loss": 1.5848796367645264, |
|
"eval_runtime": 4946.5461, |
|
"eval_samples_per_second": 33.954, |
|
"eval_steps_per_second": 2.122, |
|
"step": 47268 |
|
}, |
|
{ |
|
"epoch": 36.18, |
|
"learning_rate": 0.0004594851306405101, |
|
"loss": 0.6234, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 36.56, |
|
"learning_rate": 0.0004218107376415739, |
|
"loss": 0.6053, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 36.94, |
|
"learning_rate": 0.00038560508419879514, |
|
"loss": 0.6044, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"eval_accuracy": 0.706195111785895, |
|
"eval_loss": 1.6146132946014404, |
|
"eval_runtime": 2680.1847, |
|
"eval_samples_per_second": 62.665, |
|
"eval_steps_per_second": 3.917, |
|
"step": 48581 |
|
}, |
|
{ |
|
"epoch": 37.32, |
|
"learning_rate": 0.0003508937582970281, |
|
"loss": 0.5763, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 37.7, |
|
"learning_rate": 0.0003177012918200253, |
|
"loss": 0.572, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"eval_accuracy": 0.7081241999345063, |
|
"eval_loss": 1.6583054065704346, |
|
"eval_runtime": 2714.09, |
|
"eval_samples_per_second": 61.883, |
|
"eval_steps_per_second": 3.868, |
|
"step": 49894 |
|
}, |
|
{ |
|
"epoch": 38.08, |
|
"learning_rate": 0.0002860511432127716, |
|
"loss": 0.5693, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 38.46, |
|
"learning_rate": 0.00025596568090246547, |
|
"loss": 0.5492, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 38.84, |
|
"learning_rate": 0.00022746616748985177, |
|
"loss": 0.545, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"eval_accuracy": 0.7086898276324015, |
|
"eval_loss": 1.6993166208267212, |
|
"eval_runtime": 8190.9805, |
|
"eval_samples_per_second": 20.505, |
|
"eval_steps_per_second": 1.282, |
|
"step": 51207 |
|
}, |
|
{ |
|
"epoch": 39.22, |
|
"learning_rate": 0.0002005727447220948, |
|
"loss": 0.5409, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 39.6, |
|
"learning_rate": 0.00017530441925779654, |
|
"loss": 0.5351, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 39.98, |
|
"learning_rate": 0.00015167904923422986, |
|
"loss": 0.5341, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_accuracy": 0.7105593760233396, |
|
"eval_loss": 1.7077844142913818, |
|
"eval_runtime": 33225.1715, |
|
"eval_samples_per_second": 5.055, |
|
"eval_steps_per_second": 0.316, |
|
"step": 52520 |
|
}, |
|
{ |
|
"epoch": 40.37, |
|
"learning_rate": 0.0001297133316462798, |
|
"loss": 0.5184, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 40.75, |
|
"learning_rate": 0.00010942279054600657, |
|
"loss": 0.5284, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 41.0, |
|
"eval_accuracy": 0.7104641124110624, |
|
"eval_loss": 1.7240744829177856, |
|
"eval_runtime": 3401.4075, |
|
"eval_samples_per_second": 49.378, |
|
"eval_steps_per_second": 3.086, |
|
"step": 53833 |
|
}, |
|
{ |
|
"epoch": 41.13, |
|
"learning_rate": 9.082176607117799e-05, |
|
"loss": 0.5239, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 41.51, |
|
"learning_rate": 7.392340431051969e-05, |
|
"loss": 0.5152, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 41.89, |
|
"learning_rate": 5.87396480128477e-05, |
|
"loss": 0.5186, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 42.0, |
|
"eval_accuracy": 0.7111666815516061, |
|
"eval_loss": 1.7408066987991333, |
|
"eval_runtime": 4964.1178, |
|
"eval_samples_per_second": 33.834, |
|
"eval_steps_per_second": 2.115, |
|
"step": 55146 |
|
}, |
|
{ |
|
"epoch": 42.27, |
|
"learning_rate": 4.528122814664942e-05, |
|
"loss": 0.5113, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 42.65, |
|
"learning_rate": 3.35576563160786e-05, |
|
"loss": 0.506, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 43.0, |
|
"eval_accuracy": 0.7105772379506415, |
|
"eval_loss": 1.7486644983291626, |
|
"eval_runtime": 6355.0593, |
|
"eval_samples_per_second": 26.429, |
|
"eval_steps_per_second": 1.652, |
|
"step": 56459 |
|
}, |
|
{ |
|
"epoch": 43.03, |
|
"learning_rate": 2.357721803872476e-05, |
|
"loss": 0.5137, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 43.41, |
|
"learning_rate": 1.5346966889905334e-05, |
|
"loss": 0.5062, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 43.79, |
|
"learning_rate": 8.872719517621642e-06, |
|
"loss": 0.5043, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 44.0, |
|
"eval_accuracy": 0.7108689827632402, |
|
"eval_loss": 1.7547060251235962, |
|
"eval_runtime": 4684.6858, |
|
"eval_samples_per_second": 35.852, |
|
"eval_steps_per_second": 2.241, |
|
"step": 57772 |
|
}, |
|
{ |
|
"epoch": 44.17, |
|
"learning_rate": 4.159051531700819e-06, |
|
"loss": 0.5084, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 44.55, |
|
"learning_rate": 1.2092942700298037e-06, |
|
"loss": 0.5043, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 44.94, |
|
"learning_rate": 2.5532444165032865e-08, |
|
"loss": 0.5094, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"eval_accuracy": 0.711053556012027, |
|
"eval_loss": 1.7535769939422607, |
|
"eval_runtime": 2625.3482, |
|
"eval_samples_per_second": 63.974, |
|
"eval_steps_per_second": 3.999, |
|
"step": 59085 |
|
}, |
|
{ |
|
"epoch": 45.32, |
|
"learning_rate": 0.0007032209920746526, |
|
"loss": 0.5386, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 45.7, |
|
"learning_rate": 0.0006689210221610515, |
|
"loss": 0.5547, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 46.0, |
|
"eval_accuracy": 0.7068619570718347, |
|
"eval_loss": 1.7073622941970825, |
|
"eval_runtime": 6607.5308, |
|
"eval_samples_per_second": 25.419, |
|
"eval_steps_per_second": 1.589, |
|
"step": 60398 |
|
}, |
|
{ |
|
"epoch": 46.08, |
|
"learning_rate": 0.0006353490004690119, |
|
"loss": 0.553, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 46.46, |
|
"learning_rate": 0.0006025182736066154, |
|
"loss": 0.5329, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 46.84, |
|
"learning_rate": 0.000570441893479062, |
|
"loss": 0.5391, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 47.0, |
|
"eval_accuracy": 0.7089637105176982, |
|
"eval_loss": 1.740086317062378, |
|
"eval_runtime": 5830.6066, |
|
"eval_samples_per_second": 28.806, |
|
"eval_steps_per_second": 1.8, |
|
"step": 61711 |
|
}, |
|
{ |
|
"epoch": 47.22, |
|
"learning_rate": 0.0005391326120998649, |
|
"loss": 0.5284, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 47.6, |
|
"learning_rate": 0.0005086028765212663, |
|
"loss": 0.5223, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 47.98, |
|
"learning_rate": 0.00047886482388589615, |
|
"loss": 0.5253, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 48.0, |
|
"eval_accuracy": 0.7092852252091334, |
|
"eval_loss": 1.7769759893417358, |
|
"eval_runtime": 2617.8614, |
|
"eval_samples_per_second": 64.157, |
|
"eval_steps_per_second": 4.01, |
|
"step": 63024 |
|
}, |
|
{ |
|
"epoch": 48.36, |
|
"learning_rate": 0.0004499302766016297, |
|
"loss": 0.5041, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 48.74, |
|
"learning_rate": 0.0004218107376415739, |
|
"loss": 0.5066, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 49.0, |
|
"eval_accuracy": 0.7102319073561371, |
|
"eval_loss": 1.813459873199463, |
|
"eval_runtime": 2608.5743, |
|
"eval_samples_per_second": 64.386, |
|
"eval_steps_per_second": 4.024, |
|
"step": 64337 |
|
}, |
|
{ |
|
"epoch": 49.12, |
|
"learning_rate": 0.00039451738597103935, |
|
"loss": 0.5057, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 49.5, |
|
"learning_rate": 0.00036806107210333034, |
|
"loss": 0.4897, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 49.89, |
|
"learning_rate": 0.00034245231378610494, |
|
"loss": 0.495, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"eval_accuracy": 0.7110476020362597, |
|
"eval_loss": 1.8451627492904663, |
|
"eval_runtime": 2646.4398, |
|
"eval_samples_per_second": 63.465, |
|
"eval_steps_per_second": 3.967, |
|
"step": 65650 |
|
}, |
|
{ |
|
"epoch": 50.27, |
|
"learning_rate": 0.0003177012918200253, |
|
"loss": 0.4867, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 50.65, |
|
"learning_rate": 0.0002938178460113711, |
|
"loss": 0.4813, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 51.0, |
|
"eval_accuracy": 0.7107320413205919, |
|
"eval_loss": 1.8846126794815063, |
|
"eval_runtime": 5243.8791, |
|
"eval_samples_per_second": 32.029, |
|
"eval_steps_per_second": 2.002, |
|
"step": 66963 |
|
}, |
|
{ |
|
"epoch": 51.03, |
|
"learning_rate": 0.00027081147126020445, |
|
"loss": 0.4829, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 51.41, |
|
"learning_rate": 0.00024869131378566, |
|
"loss": 0.466, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 51.79, |
|
"learning_rate": 0.00022746616748985177, |
|
"loss": 0.4704, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 52.0, |
|
"eval_accuracy": 0.7123693846566045, |
|
"eval_loss": 1.8989039659500122, |
|
"eval_runtime": 2607.2421, |
|
"eval_samples_per_second": 64.419, |
|
"eval_steps_per_second": 4.026, |
|
"step": 68276 |
|
}, |
|
{ |
|
"epoch": 52.17, |
|
"learning_rate": 0.00020714447046184416, |
|
"loss": 0.47, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 52.55, |
|
"learning_rate": 0.0001877343016230762, |
|
"loss": 0.465, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 52.93, |
|
"learning_rate": 0.0001692433775155744, |
|
"loss": 0.4689, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 53.0, |
|
"eval_accuracy": 0.7131850793367271, |
|
"eval_loss": 1.931063175201416, |
|
"eval_runtime": 2614.8439, |
|
"eval_samples_per_second": 64.231, |
|
"eval_steps_per_second": 4.015, |
|
"step": 69589 |
|
}, |
|
{ |
|
"epoch": 53.31, |
|
"learning_rate": 0.00015167904923422986, |
|
"loss": 0.4572, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 53.69, |
|
"learning_rate": 0.0001350482995043595, |
|
"loss": 0.4611, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 54.0, |
|
"eval_accuracy": 0.7130659998213807, |
|
"eval_loss": 1.93539559841156, |
|
"eval_runtime": 2604.5499, |
|
"eval_samples_per_second": 64.485, |
|
"eval_steps_per_second": 4.031, |
|
"step": 70902 |
|
}, |
|
{ |
|
"epoch": 54.07, |
|
"learning_rate": 0.00011935773990571164, |
|
"loss": 0.4579, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 54.46, |
|
"learning_rate": 0.00010461360824402521, |
|
"loss": 0.4554, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 54.84, |
|
"learning_rate": 9.082176607117799e-05, |
|
"loss": 0.4547, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 55.0, |
|
"eval_accuracy": 0.7133160668036082, |
|
"eval_loss": 1.974073052406311, |
|
"eval_runtime": 2574.37, |
|
"eval_samples_per_second": 65.241, |
|
"eval_steps_per_second": 4.078, |
|
"step": 72215 |
|
}, |
|
{ |
|
"epoch": 55.22, |
|
"learning_rate": 7.798769635491976e-05, |
|
"loss": 0.4497, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 55.6, |
|
"learning_rate": 6.611650129910801e-05, |
|
"loss": 0.4458, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 55.98, |
|
"learning_rate": 5.521290031532178e-05, |
|
"loss": 0.4481, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 56.0, |
|
"eval_accuracy": 0.713071953797148, |
|
"eval_loss": 1.9898993968963623, |
|
"eval_runtime": 2622.4841, |
|
"eval_samples_per_second": 64.044, |
|
"eval_steps_per_second": 4.003, |
|
"step": 73528 |
|
}, |
|
{ |
|
"epoch": 56.36, |
|
"learning_rate": 0.00045401414969093167, |
|
"loss": 0.4607, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 56.74, |
|
"learning_rate": 0.0004297610695623985, |
|
"loss": 0.4709, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 57.0, |
|
"eval_accuracy": 0.7104283885564586, |
|
"eval_loss": 1.9411875009536743, |
|
"eval_runtime": 3269.1549, |
|
"eval_samples_per_second": 51.376, |
|
"eval_steps_per_second": 3.211, |
|
"step": 74841 |
|
}, |
|
{ |
|
"epoch": 57.12, |
|
"learning_rate": 0.0004061126673698029, |
|
"loss": 0.4736, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 57.5, |
|
"learning_rate": 0.0003830758503675777, |
|
"loss": 0.4601, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 57.88, |
|
"learning_rate": 0.00036065734717759815, |
|
"loss": 0.4647, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 58.0, |
|
"eval_accuracy": 0.7097913131493555, |
|
"eval_loss": 1.9707293510437012, |
|
"eval_runtime": 2595.9104, |
|
"eval_samples_per_second": 64.7, |
|
"eval_steps_per_second": 4.044, |
|
"step": 76154 |
|
}, |
|
{ |
|
"epoch": 58.26, |
|
"learning_rate": 0.00033886370582388, |
|
"loss": 0.4562, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 58.64, |
|
"learning_rate": 0.0003177012918200253, |
|
"loss": 0.4566, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 59.0, |
|
"eval_accuracy": 0.7116251376856896, |
|
"eval_loss": 2.0150792598724365, |
|
"eval_runtime": 2589.1949, |
|
"eval_samples_per_second": 64.868, |
|
"eval_steps_per_second": 4.055, |
|
"step": 77467 |
|
}, |
|
{ |
|
"epoch": 59.03, |
|
"learning_rate": 0.0002971762863099739, |
|
"loss": 0.4535, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 59.41, |
|
"learning_rate": 0.0002772946842626087, |
|
"loss": 0.448, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 59.79, |
|
"learning_rate": 0.000258062292720736, |
|
"loss": 0.4511, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 60.0, |
|
"eval_accuracy": 0.7113929326307642, |
|
"eval_loss": 2.0363075733184814, |
|
"eval_runtime": 2640.6575, |
|
"eval_samples_per_second": 63.603, |
|
"eval_steps_per_second": 3.976, |
|
"step": 78780 |
|
}, |
|
{ |
|
"epoch": 60.17, |
|
"learning_rate": 0.00023948472910495817, |
|
"loss": 0.4446, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 60.55, |
|
"learning_rate": 0.0002215674195729317, |
|
"loss": 0.4403, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 60.93, |
|
"learning_rate": 0.0002043155974344868, |
|
"loss": 0.4423, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 61.0, |
|
"eval_accuracy": 0.7112143133577447, |
|
"eval_loss": 2.0710268020629883, |
|
"eval_runtime": 2593.0457, |
|
"eval_samples_per_second": 64.771, |
|
"eval_steps_per_second": 4.049, |
|
"step": 80093 |
|
}, |
|
{ |
|
"epoch": 61.31, |
|
"learning_rate": 0.0001877343016230762, |
|
"loss": 0.4304, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 61.69, |
|
"learning_rate": 0.00017182837522399742, |
|
"loss": 0.4356, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 62.0, |
|
"eval_accuracy": 0.7116251376856896, |
|
"eval_loss": 2.061058521270752, |
|
"eval_runtime": 2582.6346, |
|
"eval_samples_per_second": 65.032, |
|
"eval_steps_per_second": 4.065, |
|
"step": 81406 |
|
}, |
|
{ |
|
"epoch": 62.07, |
|
"learning_rate": 0.00015660246405981793, |
|
"loss": 0.4332, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 62.45, |
|
"learning_rate": 0.00014206101533341926, |
|
"loss": 0.43, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 62.83, |
|
"learning_rate": 0.0001282082763290507, |
|
"loss": 0.4272, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 63.0, |
|
"eval_accuracy": 0.7117918490071745, |
|
"eval_loss": 2.089148998260498, |
|
"eval_runtime": 2586.4124, |
|
"eval_samples_per_second": 64.937, |
|
"eval_steps_per_second": 4.059, |
|
"step": 82719 |
|
}, |
|
{ |
|
"epoch": 63.21, |
|
"learning_rate": 0.00011504829317177723, |
|
"loss": 0.4259, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 63.59, |
|
"learning_rate": 0.00010258490964568406, |
|
"loss": 0.4271, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 63.98, |
|
"learning_rate": 9.082176607117799e-05, |
|
"loss": 0.4254, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 64.0, |
|
"eval_accuracy": 0.7123812926081391, |
|
"eval_loss": 2.0879101753234863, |
|
"eval_runtime": 2610.5128, |
|
"eval_samples_per_second": 64.338, |
|
"eval_steps_per_second": 4.021, |
|
"step": 84032 |
|
}, |
|
{ |
|
"epoch": 64.36, |
|
"learning_rate": 7.976229824171872e-05, |
|
"loss": 0.4249, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 64.74, |
|
"learning_rate": 6.940973642028724e-05, |
|
"loss": 0.4221, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 65.0, |
|
"eval_accuracy": 0.7130957697002174, |
|
"eval_loss": 2.1167192459106445, |
|
"eval_runtime": 2625.6926, |
|
"eval_samples_per_second": 63.966, |
|
"eval_steps_per_second": 3.998, |
|
"step": 85345 |
|
}, |
|
{ |
|
"epoch": 65.12, |
|
"learning_rate": 5.976710439588779e-05, |
|
"loss": 0.4227, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 65.5, |
|
"learning_rate": 5.083721860035529e-05, |
|
"loss": 0.421, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 65.88, |
|
"learning_rate": 4.262268728572882e-05, |
|
"loss": 0.4189, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 66.0, |
|
"eval_accuracy": 0.7129111964514304, |
|
"eval_loss": 2.1362690925598145, |
|
"eval_runtime": 2775.8163, |
|
"eval_samples_per_second": 60.507, |
|
"eval_steps_per_second": 3.782, |
|
"step": 86658 |
|
}, |
|
{ |
|
"epoch": 66.26, |
|
"learning_rate": 3.512590976242902e-05, |
|
"loss": 0.4192, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 66.64, |
|
"learning_rate": 2.834907569846612e-05, |
|
"loss": 0.4219, |
|
"step": 87500 |
|
}, |
|
{ |
|
"epoch": 67.0, |
|
"eval_accuracy": 0.712958828257569, |
|
"eval_loss": 2.1355273723602295, |
|
"eval_runtime": 2581.0691, |
|
"eval_samples_per_second": 65.072, |
|
"eval_steps_per_second": 4.067, |
|
"step": 87971 |
|
}, |
|
{ |
|
"epoch": 67.02, |
|
"learning_rate": 2.2294164479878255e-05, |
|
"loss": 0.4197, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 67.4, |
|
"learning_rate": 1.6962944632589183e-05, |
|
"loss": 0.4157, |
|
"step": 88500 |
|
}, |
|
{ |
|
"epoch": 67.78, |
|
"learning_rate": 1.2356973305856779e-05, |
|
"loss": 0.4149, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 68.0, |
|
"eval_accuracy": 0.7132267571670984, |
|
"eval_loss": 2.1466283798217773, |
|
"eval_runtime": 2769.7004, |
|
"eval_samples_per_second": 60.64, |
|
"eval_steps_per_second": 3.79, |
|
"step": 89284 |
|
}, |
|
{ |
|
"epoch": 68.16, |
|
"learning_rate": 8.477595817457106e-06, |
|
"loss": 0.4138, |
|
"step": 89500 |
|
}, |
|
{ |
|
"epoch": 68.55, |
|
"learning_rate": 5.325945260744314e-06, |
|
"loss": 0.4147, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 68.93, |
|
"learning_rate": 2.90294217369369e-06, |
|
"loss": 0.4125, |
|
"step": 90500 |
|
}, |
|
{ |
|
"epoch": 69.0, |
|
"eval_accuracy": 0.7131434015063559, |
|
"eval_loss": 2.1478145122528076, |
|
"eval_runtime": 2585.5905, |
|
"eval_samples_per_second": 64.958, |
|
"eval_steps_per_second": 4.06, |
|
"step": 90597 |
|
}, |
|
{ |
|
"epoch": 69.31, |
|
"learning_rate": 1.2092942700298037e-06, |
|
"loss": 0.4146, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 69.69, |
|
"learning_rate": 2.4549623251851615e-07, |
|
"loss": 0.4162, |
|
"step": 91500 |
|
}, |
|
{ |
|
"epoch": 70.0, |
|
"eval_accuracy": 0.7132267571670984, |
|
"eval_loss": 2.1484131813049316, |
|
"eval_runtime": 2581.5568, |
|
"eval_samples_per_second": 65.06, |
|
"eval_steps_per_second": 4.067, |
|
"step": 91910 |
|
}, |
|
{ |
|
"epoch": 70.0, |
|
"step": 91910, |
|
"total_flos": 3.2073215097814647e+21, |
|
"train_loss": 0.08673005339085323, |
|
"train_runtime": 125919.3774, |
|
"train_samples_per_second": 373.471, |
|
"train_steps_per_second": 0.73 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 91910, |
|
"num_train_epochs": 70, |
|
"save_steps": 500, |
|
"total_flos": 3.2073215097814647e+21, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|