{ "best_metric": 1.3635696172714233, "best_model_checkpoint": "./llama3/21-04-24-Weni-WeniGPT-Agents-Llama3-1.0.9-SFT_Experiment with SFT and Llama3 and updates in requirements-2_max_steps-330_batch_8_2024-04-21_ppid_9/checkpoint-90", "epoch": 3.2142857142857144, "eval_steps": 30, "global_step": 180, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.17857142857142858, "grad_norm": 0.6299236416816711, "learning_rate": 0.00015555555555555556, "loss": 1.9266, "step": 10 }, { "epoch": 0.35714285714285715, "grad_norm": 0.4194231629371643, "learning_rate": 0.00019969365006623072, "loss": 1.5439, "step": 20 }, { "epoch": 0.5357142857142857, "grad_norm": 0.44722992181777954, "learning_rate": 0.00019845231970029773, "loss": 1.3488, "step": 30 }, { "epoch": 0.5357142857142857, "eval_loss": 1.3763371706008911, "eval_runtime": 4.9516, "eval_samples_per_second": 9.29, "eval_steps_per_second": 2.423, "step": 30 }, { "epoch": 0.7142857142857143, "grad_norm": 0.4362790286540985, "learning_rate": 0.00019626873324023915, "loss": 1.3256, "step": 40 }, { "epoch": 0.8928571428571429, "grad_norm": 0.435635507106781, "learning_rate": 0.00019316378910323102, "loss": 1.2663, "step": 50 }, { "epoch": 1.0714285714285714, "grad_norm": 0.5550065636634827, "learning_rate": 0.00018916720373012426, "loss": 1.1798, "step": 60 }, { "epoch": 1.0714285714285714, "eval_loss": 1.3427196741104126, "eval_runtime": 4.9453, "eval_samples_per_second": 9.302, "eval_steps_per_second": 2.427, "step": 60 }, { "epoch": 1.25, "grad_norm": 0.571713387966156, "learning_rate": 0.00018431722717876384, "loss": 1.0592, "step": 70 }, { "epoch": 1.4285714285714286, "grad_norm": 0.5714108943939209, "learning_rate": 0.0001786602770447513, "loss": 0.98, "step": 80 }, { "epoch": 1.6071428571428572, "grad_norm": 0.5294617414474487, "learning_rate": 0.00017225049421328023, "loss": 0.9664, "step": 90 }, { "epoch": 1.6071428571428572, "eval_loss": 1.3635696172714233, "eval_runtime": 4.9445, "eval_samples_per_second": 9.303, "eval_steps_per_second": 2.427, "step": 90 }, { "epoch": 1.7857142857142856, "grad_norm": 0.8601852655410767, "learning_rate": 0.00016514922469380337, "loss": 0.9626, "step": 100 }, { "epoch": 1.9642857142857144, "grad_norm": 0.6094381213188171, "learning_rate": 0.00015742443249672827, "loss": 0.8046, "step": 110 }, { "epoch": 2.142857142857143, "grad_norm": 0.7323458790779114, "learning_rate": 0.00014915004917131344, "loss": 0.7042, "step": 120 }, { "epoch": 2.142857142857143, "eval_loss": 1.3712576627731323, "eval_runtime": 4.9472, "eval_samples_per_second": 9.298, "eval_steps_per_second": 2.426, "step": 120 }, { "epoch": 2.3214285714285716, "grad_norm": 0.7422382831573486, "learning_rate": 0.00014040526623013318, "loss": 0.7782, "step": 130 }, { "epoch": 2.5, "grad_norm": 0.6805801391601562, "learning_rate": 0.000131273777232092, "loss": 0.6468, "step": 140 }, { "epoch": 2.678571428571429, "grad_norm": 0.873374342918396, "learning_rate": 0.00012184297677777463, "loss": 0.6411, "step": 150 }, { "epoch": 2.678571428571429, "eval_loss": 1.4060558080673218, "eval_runtime": 4.9458, "eval_samples_per_second": 9.301, "eval_steps_per_second": 2.426, "step": 150 }, { "epoch": 2.857142857142857, "grad_norm": 0.9634660482406616, "learning_rate": 0.00011220312408329318, "loss": 0.7025, "step": 160 }, { "epoch": 3.0357142857142856, "grad_norm": 0.8289043307304382, "learning_rate": 0.00010244647913780272, "loss": 0.6012, "step": 170 }, { "epoch": 3.2142857142857144, "grad_norm": 0.5140141844749451, "learning_rate": 9.266641971224963e-05, "loss": 0.4672, "step": 180 }, { "epoch": 3.2142857142857144, "eval_loss": 1.4097956418991089, "eval_runtime": 4.9459, "eval_samples_per_second": 9.301, "eval_steps_per_second": 2.426, "step": 180 } ], "logging_steps": 10, "max_steps": 330, "num_input_tokens_seen": 0, "num_train_epochs": 6, "save_steps": 90, "total_flos": 1.4827719499553178e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }