|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 11.0, |
|
"global_step": 221540, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.909609120521173e-05, |
|
"loss": 2.4473, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.796530039811799e-05, |
|
"loss": 1.8064, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.683473579442635e-05, |
|
"loss": 1.6957, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.570417119073471e-05, |
|
"loss": 1.6302, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 9.457383279044517e-05, |
|
"loss": 1.5838, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 9.344304198335143e-05, |
|
"loss": 1.552, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 9.23127035830619e-05, |
|
"loss": 1.5235, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 9.118236518277235e-05, |
|
"loss": 1.5019, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 9.005157437567862e-05, |
|
"loss": 1.4807, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 8.892100977198697e-05, |
|
"loss": 1.4627, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 8.779044516829533e-05, |
|
"loss": 1.449, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 8.665965436120159e-05, |
|
"loss": 1.4371, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 8.552931596091206e-05, |
|
"loss": 1.4229, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 8.439852515381831e-05, |
|
"loss": 1.4127, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"learning_rate": 8.326773434672458e-05, |
|
"loss": 1.4034, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 8.213716974303293e-05, |
|
"loss": 1.394, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 8.100683134274339e-05, |
|
"loss": 1.3822, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 7.987581433224755e-05, |
|
"loss": 1.3746, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 7.874524972855593e-05, |
|
"loss": 1.3659, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"learning_rate": 7.761468512486428e-05, |
|
"loss": 1.3603, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"learning_rate": 7.648412052117264e-05, |
|
"loss": 1.3509, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 5.46, |
|
"learning_rate": 7.53533297140789e-05, |
|
"loss": 1.3439, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 5.71, |
|
"learning_rate": 7.422276511038726e-05, |
|
"loss": 1.3389, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 5.96, |
|
"learning_rate": 7.309220050669564e-05, |
|
"loss": 1.3318, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 6.21, |
|
"learning_rate": 7.196163590300398e-05, |
|
"loss": 1.3233, |
|
"step": 125000 |
|
}, |
|
{ |
|
"epoch": 6.45, |
|
"learning_rate": 7.083107129931235e-05, |
|
"loss": 1.3169, |
|
"step": 130000 |
|
}, |
|
{ |
|
"epoch": 6.7, |
|
"learning_rate": 6.97002804922186e-05, |
|
"loss": 1.3104, |
|
"step": 135000 |
|
}, |
|
{ |
|
"epoch": 6.95, |
|
"learning_rate": 6.856994209192906e-05, |
|
"loss": 1.3064, |
|
"step": 140000 |
|
}, |
|
{ |
|
"epoch": 7.2, |
|
"learning_rate": 6.743937748823743e-05, |
|
"loss": 1.3007, |
|
"step": 145000 |
|
}, |
|
{ |
|
"epoch": 7.45, |
|
"learning_rate": 6.630858668114369e-05, |
|
"loss": 1.2962, |
|
"step": 150000 |
|
}, |
|
{ |
|
"epoch": 7.7, |
|
"learning_rate": 6.517779587404995e-05, |
|
"loss": 1.2915, |
|
"step": 155000 |
|
}, |
|
{ |
|
"epoch": 7.94, |
|
"learning_rate": 6.40470050669562e-05, |
|
"loss": 1.2877, |
|
"step": 160000 |
|
}, |
|
{ |
|
"epoch": 8.19, |
|
"learning_rate": 6.291644046326457e-05, |
|
"loss": 1.2783, |
|
"step": 165000 |
|
}, |
|
{ |
|
"epoch": 8.44, |
|
"learning_rate": 6.178564965617082e-05, |
|
"loss": 1.2741, |
|
"step": 170000 |
|
}, |
|
{ |
|
"epoch": 8.69, |
|
"learning_rate": 6.065485884907709e-05, |
|
"loss": 1.2705, |
|
"step": 175000 |
|
}, |
|
{ |
|
"epoch": 8.94, |
|
"learning_rate": 5.9524746652189655e-05, |
|
"loss": 1.267, |
|
"step": 180000 |
|
}, |
|
{ |
|
"epoch": 9.19, |
|
"learning_rate": 5.839395584509592e-05, |
|
"loss": 1.26, |
|
"step": 185000 |
|
}, |
|
{ |
|
"epoch": 9.43, |
|
"learning_rate": 5.7263391241404275e-05, |
|
"loss": 1.2558, |
|
"step": 190000 |
|
}, |
|
{ |
|
"epoch": 9.68, |
|
"learning_rate": 5.613260043431053e-05, |
|
"loss": 1.2531, |
|
"step": 195000 |
|
}, |
|
{ |
|
"epoch": 9.93, |
|
"learning_rate": 5.5002035830618894e-05, |
|
"loss": 1.2492, |
|
"step": 200000 |
|
}, |
|
{ |
|
"epoch": 10.18, |
|
"learning_rate": 5.387124502352515e-05, |
|
"loss": 1.2432, |
|
"step": 205000 |
|
}, |
|
{ |
|
"epoch": 10.43, |
|
"learning_rate": 5.274113282663772e-05, |
|
"loss": 1.2391, |
|
"step": 210000 |
|
}, |
|
{ |
|
"epoch": 10.68, |
|
"learning_rate": 5.1610342019543975e-05, |
|
"loss": 1.2347, |
|
"step": 215000 |
|
}, |
|
{ |
|
"epoch": 10.92, |
|
"learning_rate": 5.047977741585234e-05, |
|
"loss": 1.231, |
|
"step": 220000 |
|
} |
|
], |
|
"max_steps": 443080, |
|
"num_train_epochs": 22, |
|
"total_flos": 7.46536345693738e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|