|
{ |
|
"best_metric": 0.3619329333305359, |
|
"best_model_checkpoint": "./zephyr/08-04-24-Weni-WeniGPT-Agents-Zephyr-1.0.17-KTO_Hyperparameter search, altering lora params for KTO task.-2_max_steps-145_batch_16_2024-04-08_ppid_9/checkpoint-100", |
|
"epoch": 0.684931506849315, |
|
"eval_steps": 50, |
|
"global_step": 100, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 6.565266132354736, |
|
"learning_rate": 0.00018142857142857142, |
|
"loss": 0.3847, |
|
"step": 20, |
|
"train/kl": 3.3989148139953613, |
|
"train/logps/chosen": -298.30405479753523, |
|
"train/logps/rejected": -317.74027826544943, |
|
"train/rewards/chosen": 0.4325446276597574, |
|
"train/rewards/margins": 3.3660581404318783, |
|
"train/rewards/rejected": -2.933513512772121 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 2.64931583404541, |
|
"learning_rate": 0.00015285714285714287, |
|
"loss": 0.4039, |
|
"step": 40, |
|
"train/kl": 0.9900484085083008, |
|
"train/logps/chosen": -305.07652368012424, |
|
"train/logps/rejected": -347.4229805424528, |
|
"train/rewards/chosen": -2.13509221995099, |
|
"train/rewards/margins": 5.288080256955457, |
|
"train/rewards/rejected": -7.423172476906447 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"eval/kl": 0.7344650030136108, |
|
"eval/logps/chosen": -329.9394806338028, |
|
"eval/logps/rejected": -444.9535700158228, |
|
"eval/rewards/chosen": -4.62774013465559, |
|
"eval/rewards/margins": 13.430626471851532, |
|
"eval/rewards/rejected": -18.05836660650712, |
|
"eval_loss": 0.37672173976898193, |
|
"eval_runtime": 140.2349, |
|
"eval_samples_per_second": 2.139, |
|
"eval_steps_per_second": 0.535, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 8.288690567016602, |
|
"learning_rate": 0.00012714285714285714, |
|
"loss": 0.3602, |
|
"step": 60, |
|
"train/kl": 2.266563653945923, |
|
"train/logps/chosen": -318.9802876655629, |
|
"train/logps/rejected": -473.3688517011834, |
|
"train/rewards/chosen": -3.7534725366049253, |
|
"train/rewards/margins": 16.50871138073975, |
|
"train/rewards/rejected": -20.262183917344675 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 4.295706748962402, |
|
"learning_rate": 9.857142857142858e-05, |
|
"loss": 0.3496, |
|
"step": 80, |
|
"train/kl": 0.8752914667129517, |
|
"train/logps/chosen": -335.9194670376712, |
|
"train/logps/rejected": -405.97503591954023, |
|
"train/rewards/chosen": -3.1786450947800726, |
|
"train/rewards/margins": 7.811195007141622, |
|
"train/rewards/rejected": -10.989840101921695 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 5.850632667541504, |
|
"learning_rate": 7e-05, |
|
"loss": 0.302, |
|
"step": 100, |
|
"train/kl": 0.4592212736606598, |
|
"train/logps/chosen": -277.03286903782896, |
|
"train/logps/rejected": -392.9672154017857, |
|
"train/rewards/chosen": 0.3360620799817537, |
|
"train/rewards/margins": 9.165781260134283, |
|
"train/rewards/rejected": -8.82971918015253 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"eval/kl": 3.9467480182647705, |
|
"eval/logps/chosen": -265.8084286971831, |
|
"eval/logps/rejected": -297.73909711234177, |
|
"eval/rewards/chosen": 1.7853647151463468, |
|
"eval/rewards/margins": 5.122287276212407, |
|
"eval/rewards/rejected": -3.3369225610660602, |
|
"eval_loss": 0.3619329333305359, |
|
"eval_runtime": 140.261, |
|
"eval_samples_per_second": 2.139, |
|
"eval_steps_per_second": 0.535, |
|
"step": 100 |
|
} |
|
], |
|
"logging_steps": 20, |
|
"max_steps": 145, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"total_flos": 0.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|