|
{ |
|
"best_metric": 1.1286808252334595, |
|
"best_model_checkpoint": "./mistral/22-04-24-Weni-WeniGPT-Agents-Mistral-1.0.11-SFT_Experiment with SFT and a new tokenizer configuration for chat template of mistral-2_max_steps-330_batch_8_2024-04-22_ppid_9/checkpoint-90", |
|
"epoch": 3.2142857142857144, |
|
"eval_steps": 30, |
|
"global_step": 180, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.17857142857142858, |
|
"grad_norm": 0.45132189989089966, |
|
"learning_rate": 0.00019999521087449523, |
|
"loss": 1.5157, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.35714285714285715, |
|
"grad_norm": 0.25432512164115906, |
|
"learning_rate": 0.00019942107065112286, |
|
"loss": 1.1691, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.5357142857142857, |
|
"grad_norm": 0.30843719840049744, |
|
"learning_rate": 0.0001978954027238763, |
|
"loss": 1.0466, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.5357142857142857, |
|
"eval_loss": 1.1547143459320068, |
|
"eval_runtime": 36.25, |
|
"eval_samples_per_second": 1.269, |
|
"eval_steps_per_second": 0.331, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.7142857142857143, |
|
"grad_norm": 0.35677382349967957, |
|
"learning_rate": 0.00019543280877920072, |
|
"loss": 1.0075, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.8928571428571429, |
|
"grad_norm": 0.3124666213989258, |
|
"learning_rate": 0.00019205685752689177, |
|
"loss": 0.9662, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.0714285714285714, |
|
"grad_norm": 0.3985981047153473, |
|
"learning_rate": 0.00018779985913140924, |
|
"loss": 0.8783, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.0714285714285714, |
|
"eval_loss": 1.12884521484375, |
|
"eval_runtime": 36.2483, |
|
"eval_samples_per_second": 1.269, |
|
"eval_steps_per_second": 0.331, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.40197721123695374, |
|
"learning_rate": 0.0001827025559814854, |
|
"loss": 0.7525, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.4285714285714286, |
|
"grad_norm": 0.4119006395339966, |
|
"learning_rate": 0.0001768137327575751, |
|
"loss": 0.6832, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.6071428571428572, |
|
"grad_norm": 0.3348449766635895, |
|
"learning_rate": 0.00017018974952906884, |
|
"loss": 0.6778, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.6071428571428572, |
|
"eval_loss": 1.1286808252334595, |
|
"eval_runtime": 36.2515, |
|
"eval_samples_per_second": 1.269, |
|
"eval_steps_per_second": 0.331, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.7857142857142856, |
|
"grad_norm": 0.49415308237075806, |
|
"learning_rate": 0.0001628940023498477, |
|
"loss": 0.682, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.9642857142857144, |
|
"grad_norm": 0.28949180245399475, |
|
"learning_rate": 0.00015499631651465085, |
|
"loss": 0.544, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 2.142857142857143, |
|
"grad_norm": 0.4506300687789917, |
|
"learning_rate": 0.00014657227828320635, |
|
"loss": 0.4609, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.142857142857143, |
|
"eval_loss": 1.1452628374099731, |
|
"eval_runtime": 36.2339, |
|
"eval_samples_per_second": 1.27, |
|
"eval_steps_per_second": 0.331, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.3214285714285716, |
|
"grad_norm": 0.49903905391693115, |
|
"learning_rate": 0.000137702511467984, |
|
"loss": 0.5106, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 0.422453910112381, |
|
"learning_rate": 0.0001284719058091194, |
|
"loss": 0.4149, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 2.678571428571429, |
|
"grad_norm": 0.536363959312439, |
|
"learning_rate": 0.00011896880452149077, |
|
"loss": 0.3946, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.678571428571429, |
|
"eval_loss": 1.1766327619552612, |
|
"eval_runtime": 36.2209, |
|
"eval_samples_per_second": 1.27, |
|
"eval_steps_per_second": 0.331, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.857142857142857, |
|
"grad_norm": 0.5005641579627991, |
|
"learning_rate": 0.00010928415878967781, |
|
"loss": 0.4512, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 3.0357142857142856, |
|
"grad_norm": 0.4409847557544708, |
|
"learning_rate": 9.951065730286553e-05, |
|
"loss": 0.3759, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 3.2142857142857144, |
|
"grad_norm": 0.33086544275283813, |
|
"learning_rate": 8.974183916063968e-05, |
|
"loss": 0.2669, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 3.2142857142857144, |
|
"eval_loss": 1.212720274925232, |
|
"eval_runtime": 36.2554, |
|
"eval_samples_per_second": 1.269, |
|
"eval_steps_per_second": 0.331, |
|
"step": 180 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 330, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 6, |
|
"save_steps": 90, |
|
"total_flos": 1.7291593631399936e+17, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|