{ "best_metric": 1.1614402532577515, "best_model_checkpoint": "./mistral/22-04-24-Weni-WeniGPT-Agents-Mistral-1.0.12-SFT_Experiment with SFT and a new tokenizer configuration for chat template of mistral-2_max_steps-330_batch_8_2024-04-22_ppid_9/checkpoint-90", "epoch": 1.6071428571428572, "eval_steps": 30, "global_step": 90, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.17857142857142858, "grad_norm": 0.7089748978614807, "learning_rate": 0.00019999521087449523, "loss": 1.366, "step": 10 }, { "epoch": 0.35714285714285715, "grad_norm": 0.47184231877326965, "learning_rate": 0.00019942107065112286, "loss": 1.0042, "step": 20 }, { "epoch": 0.5357142857142857, "grad_norm": 0.4505631625652313, "learning_rate": 0.0001978954027238763, "loss": 0.8828, "step": 30 }, { "epoch": 0.5357142857142857, "eval_loss": 1.1399036645889282, "eval_runtime": 38.6638, "eval_samples_per_second": 1.19, "eval_steps_per_second": 0.31, "step": 30 }, { "epoch": 0.7142857142857143, "grad_norm": 0.4148388206958771, "learning_rate": 0.00019543280877920072, "loss": 0.806, "step": 40 }, { "epoch": 0.8928571428571429, "grad_norm": 0.3701043725013733, "learning_rate": 0.00019205685752689177, "loss": 0.771, "step": 50 }, { "epoch": 1.0714285714285714, "grad_norm": 0.39483487606048584, "learning_rate": 0.00018779985913140924, "loss": 0.6609, "step": 60 }, { "epoch": 1.0714285714285714, "eval_loss": 1.1470385789871216, "eval_runtime": 38.654, "eval_samples_per_second": 1.19, "eval_steps_per_second": 0.31, "step": 60 }, { "epoch": 1.25, "grad_norm": 0.40163591504096985, "learning_rate": 0.0001827025559814854, "loss": 0.4089, "step": 70 }, { "epoch": 1.4285714285714286, "grad_norm": 0.3985687792301178, "learning_rate": 0.0001768137327575751, "loss": 0.3764, "step": 80 }, { "epoch": 1.6071428571428572, "grad_norm": 0.3154538869857788, "learning_rate": 0.00017018974952906884, "loss": 0.3918, "step": 90 }, { "epoch": 1.6071428571428572, "eval_loss": 1.1614402532577515, "eval_runtime": 38.6493, "eval_samples_per_second": 1.19, "eval_steps_per_second": 0.31, "step": 90 } ], "logging_steps": 10, "max_steps": 330, "num_input_tokens_seen": 0, "num_train_epochs": 6, "save_steps": 90, "total_flos": 9.259625757016064e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }