|
{ |
|
"best_metric": 0.5126128792762756, |
|
"best_model_checkpoint": "./llama3/30-08-24-Weni-Pipeline_test_Experiment with SFT and Llama3 70b-2_max_steps-1362_batch_8_2024-08-30/checkpoint-300", |
|
"epoch": 1.3201320132013201, |
|
"eval_steps": 100, |
|
"global_step": 300, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.04400440044004401, |
|
"grad_norm": 0.5568628907203674, |
|
"learning_rate": 7.5e-05, |
|
"loss": 2.0875, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.08800880088008801, |
|
"grad_norm": 0.2537558972835541, |
|
"learning_rate": 0.00015, |
|
"loss": 0.9378, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.132013201320132, |
|
"grad_norm": 0.24558919668197632, |
|
"learning_rate": 0.000225, |
|
"loss": 0.7, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.17601760176017603, |
|
"grad_norm": 0.13937097787857056, |
|
"learning_rate": 0.0003, |
|
"loss": 0.6298, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.22002200220022003, |
|
"grad_norm": 0.1871194988489151, |
|
"learning_rate": 0.00029995764763563235, |
|
"loss": 0.6321, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.264026402640264, |
|
"grad_norm": 0.14626263082027435, |
|
"learning_rate": 0.00029983061445883305, |
|
"loss": 0.6403, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.30803080308030806, |
|
"grad_norm": 0.12049665302038193, |
|
"learning_rate": 0.0002996189722050073, |
|
"loss": 0.5998, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.35203520352035206, |
|
"grad_norm": 0.13617923855781555, |
|
"learning_rate": 0.0002993228403881531, |
|
"loss": 0.5942, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.39603960396039606, |
|
"grad_norm": 0.1271793246269226, |
|
"learning_rate": 0.00029894238623337174, |
|
"loss": 0.5647, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.44004400440044006, |
|
"grad_norm": 0.18757876753807068, |
|
"learning_rate": 0.00029847782458243663, |
|
"loss": 0.5619, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.44004400440044006, |
|
"eval_accuracy": 1.0, |
|
"eval_f1": 1.0, |
|
"eval_f1_macro": 1.0, |
|
"eval_f1_micro": 1.0, |
|
"eval_loss": 0.5742923021316528, |
|
"eval_precision": 1.0, |
|
"eval_precision_macro": 1.0, |
|
"eval_precision_micro": 1.0, |
|
"eval_recall": 1.0, |
|
"eval_recall_macro": 1.0, |
|
"eval_recall_micro": 1.0, |
|
"eval_runtime": 90.5857, |
|
"eval_samples_per_second": 4.46, |
|
"eval_steps_per_second": 1.115, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.48404840484048406, |
|
"grad_norm": 0.14132679998874664, |
|
"learning_rate": 0.00029792941777247184, |
|
"loss": 0.5584, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.528052805280528, |
|
"grad_norm": 0.15474887192249298, |
|
"learning_rate": 0.0002972974754878111, |
|
"loss": 0.5752, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.5720572057205721, |
|
"grad_norm": 0.13014496862888336, |
|
"learning_rate": 0.0002965823545851199, |
|
"loss": 0.5565, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.6160616061606161, |
|
"grad_norm": 0.12456662207841873, |
|
"learning_rate": 0.00029578445889187865, |
|
"loss": 0.5722, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.6600660066006601, |
|
"grad_norm": 0.12824317812919617, |
|
"learning_rate": 0.00029490423897834234, |
|
"loss": 0.523, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.7040704070407041, |
|
"grad_norm": 0.14279119670391083, |
|
"learning_rate": 0.0002939421919031044, |
|
"loss": 0.5523, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.7480748074807481, |
|
"grad_norm": 0.11781885474920273, |
|
"learning_rate": 0.00029289886093240847, |
|
"loss": 0.5291, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.7920792079207921, |
|
"grad_norm": 0.1608349233865738, |
|
"learning_rate": 0.0002917748352333667, |
|
"loss": 0.5417, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.8360836083608361, |
|
"grad_norm": 0.13777320086956024, |
|
"learning_rate": 0.0002905707495412589, |
|
"loss": 0.4967, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.8800880088008801, |
|
"grad_norm": 0.21577192842960358, |
|
"learning_rate": 0.00028928728380109764, |
|
"loss": 0.6545, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.8800880088008801, |
|
"eval_accuracy": 1.0, |
|
"eval_f1": 1.0, |
|
"eval_f1_macro": 1.0, |
|
"eval_f1_micro": 1.0, |
|
"eval_loss": 0.6772989630699158, |
|
"eval_precision": 1.0, |
|
"eval_precision_macro": 1.0, |
|
"eval_precision_micro": 1.0, |
|
"eval_recall": 1.0, |
|
"eval_recall_macro": 1.0, |
|
"eval_recall_micro": 1.0, |
|
"eval_runtime": 90.2067, |
|
"eval_samples_per_second": 4.479, |
|
"eval_steps_per_second": 1.12, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.9240924092409241, |
|
"grad_norm": 3.4556286334991455, |
|
"learning_rate": 0.00028792516278366547, |
|
"loss": 2.5144, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.9680968096809681, |
|
"grad_norm": 0.13046959042549133, |
|
"learning_rate": 0.00028648515567623764, |
|
"loss": 0.6004, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.012101210121012, |
|
"grad_norm": 0.39528125524520874, |
|
"learning_rate": 0.0002849680756482235, |
|
"loss": 0.5174, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.056105610561056, |
|
"grad_norm": 0.23764920234680176, |
|
"learning_rate": 0.00028337477939197135, |
|
"loss": 0.4065, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.1001100110011002, |
|
"grad_norm": 0.14821326732635498, |
|
"learning_rate": 0.0002817061666389958, |
|
"loss": 0.4425, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.1441144114411441, |
|
"grad_norm": 0.15903742611408234, |
|
"learning_rate": 0.0002799631796519007, |
|
"loss": 0.4107, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.188118811881188, |
|
"grad_norm": 0.16189326345920563, |
|
"learning_rate": 0.00027814680269228574, |
|
"loss": 0.4215, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.2321232123212322, |
|
"grad_norm": 0.20622508227825165, |
|
"learning_rate": 0.00027625806146493523, |
|
"loss": 0.3968, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.2761276127612762, |
|
"grad_norm": 0.17622467875480652, |
|
"learning_rate": 0.0002742980225386045, |
|
"loss": 0.4419, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.3201320132013201, |
|
"grad_norm": 0.12472284585237503, |
|
"learning_rate": 0.0002722677927437307, |
|
"loss": 0.396, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.3201320132013201, |
|
"eval_accuracy": 1.0, |
|
"eval_f1": 1.0, |
|
"eval_f1_macro": 1.0, |
|
"eval_f1_micro": 1.0, |
|
"eval_loss": 0.5126128792762756, |
|
"eval_precision": 1.0, |
|
"eval_precision_macro": 1.0, |
|
"eval_precision_micro": 1.0, |
|
"eval_recall": 1.0, |
|
"eval_recall_macro": 1.0, |
|
"eval_recall_micro": 1.0, |
|
"eval_runtime": 90.5135, |
|
"eval_samples_per_second": 4.463, |
|
"eval_steps_per_second": 1.116, |
|
"step": 300 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 1362, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 6, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 4.182761586777129e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|