|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.971563981042654, |
|
"eval_steps": 100, |
|
"global_step": 104, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 139.638709617328, |
|
"learning_rate": 4.545454545454545e-08, |
|
"logits/chosen": 111.16130065917969, |
|
"logits/rejected": 86.8372802734375, |
|
"logps/chosen": -326.8536071777344, |
|
"logps/rejected": -329.15960693359375, |
|
"loss": 0.6931, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.0, |
|
"rewards/margins": 0.0, |
|
"rewards/rejected": 0.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 141.5345140695996, |
|
"learning_rate": 4.545454545454545e-07, |
|
"logits/chosen": 110.37065124511719, |
|
"logits/rejected": 133.2639923095703, |
|
"logps/chosen": -350.3541259765625, |
|
"logps/rejected": -434.3558349609375, |
|
"loss": 0.7191, |
|
"rewards/accuracies": 0.4722222089767456, |
|
"rewards/chosen": 0.13274627923965454, |
|
"rewards/margins": 0.07573667168617249, |
|
"rewards/rejected": 0.05700961872935295, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 123.71909837085582, |
|
"learning_rate": 4.885348141000122e-07, |
|
"logits/chosen": 117.74342346191406, |
|
"logits/rejected": 128.52548217773438, |
|
"logps/chosen": -333.21240234375, |
|
"logps/rejected": -410.2923889160156, |
|
"loss": 0.6097, |
|
"rewards/accuracies": 0.7124999761581421, |
|
"rewards/chosen": 0.11470325291156769, |
|
"rewards/margins": 0.7479402422904968, |
|
"rewards/rejected": -0.6332370042800903, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 111.89651526533274, |
|
"learning_rate": 4.5025027361734613e-07, |
|
"logits/chosen": 114.44095611572266, |
|
"logits/rejected": 119.11683654785156, |
|
"logps/chosen": -399.1412048339844, |
|
"logps/rejected": -474.2645568847656, |
|
"loss": 0.596, |
|
"rewards/accuracies": 0.7250000238418579, |
|
"rewards/chosen": -1.7276217937469482, |
|
"rewards/margins": 1.0803521871566772, |
|
"rewards/rejected": -2.807974100112915, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 102.67088507130228, |
|
"learning_rate": 3.893311157806091e-07, |
|
"logits/chosen": 116.33101654052734, |
|
"logits/rejected": 111.0595703125, |
|
"logps/chosen": -428.7275390625, |
|
"logps/rejected": -464.0934143066406, |
|
"loss": 0.5343, |
|
"rewards/accuracies": 0.7250000238418579, |
|
"rewards/chosen": -2.2770252227783203, |
|
"rewards/margins": 0.9522085189819336, |
|
"rewards/rejected": -3.229233503341675, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 130.9996197198566, |
|
"learning_rate": 3.126631330646801e-07, |
|
"logits/chosen": 123.2393569946289, |
|
"logits/rejected": 124.50789642333984, |
|
"logps/chosen": -438.548095703125, |
|
"logps/rejected": -474.1234436035156, |
|
"loss": 0.5138, |
|
"rewards/accuracies": 0.762499988079071, |
|
"rewards/chosen": -2.3258581161499023, |
|
"rewards/margins": 1.3220884799957275, |
|
"rewards/rejected": -3.647946834564209, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 56.950942870641875, |
|
"learning_rate": 2.2891223348923882e-07, |
|
"logits/chosen": 122.619384765625, |
|
"logits/rejected": 126.1447525024414, |
|
"logps/chosen": -414.3634338378906, |
|
"logps/rejected": -468.19586181640625, |
|
"loss": 0.2724, |
|
"rewards/accuracies": 0.893750011920929, |
|
"rewards/chosen": -2.3773388862609863, |
|
"rewards/margins": 2.358515501022339, |
|
"rewards/rejected": -4.735854148864746, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 52.820355390804025, |
|
"learning_rate": 1.4754491880085317e-07, |
|
"logits/chosen": 117.16709899902344, |
|
"logits/rejected": 118.9737319946289, |
|
"logps/chosen": -387.70526123046875, |
|
"logps/rejected": -511.97503662109375, |
|
"loss": 0.1936, |
|
"rewards/accuracies": 0.9437500238418579, |
|
"rewards/chosen": -2.4186935424804688, |
|
"rewards/margins": 2.5914835929870605, |
|
"rewards/rejected": -5.010177135467529, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 51.657826972971314, |
|
"learning_rate": 7.775827023107834e-08, |
|
"logits/chosen": 124.15473937988281, |
|
"logits/rejected": 125.7086181640625, |
|
"logps/chosen": -446.75421142578125, |
|
"logps/rejected": -543.6109619140625, |
|
"loss": 0.1779, |
|
"rewards/accuracies": 0.981249988079071, |
|
"rewards/chosen": -2.316882848739624, |
|
"rewards/margins": 2.962496757507324, |
|
"rewards/rejected": -5.279379844665527, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 86.34373603352554, |
|
"learning_rate": 2.7440387297912122e-08, |
|
"logits/chosen": 107.07579040527344, |
|
"logits/rejected": 111.74522399902344, |
|
"logps/chosen": -425.4237365722656, |
|
"logps/rejected": -509.67718505859375, |
|
"loss": 0.1765, |
|
"rewards/accuracies": 0.9437500238418579, |
|
"rewards/chosen": -2.749206066131592, |
|
"rewards/margins": 3.0597147941589355, |
|
"rewards/rejected": -5.8089213371276855, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 51.66215546933828, |
|
"learning_rate": 2.27878296044029e-09, |
|
"logits/chosen": 123.38490295410156, |
|
"logits/rejected": 113.675537109375, |
|
"logps/chosen": -439.7268981933594, |
|
"logps/rejected": -550.8162841796875, |
|
"loss": 0.1923, |
|
"rewards/accuracies": 0.9624999761581421, |
|
"rewards/chosen": -2.560769557952881, |
|
"rewards/margins": 3.2135703563690186, |
|
"rewards/rejected": -5.77433967590332, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"eval_logits/chosen": 92.73604583740234, |
|
"eval_logits/rejected": 86.38631439208984, |
|
"eval_logps/chosen": -431.5707092285156, |
|
"eval_logps/rejected": -459.1661682128906, |
|
"eval_loss": 0.4735770523548126, |
|
"eval_rewards/accuracies": 0.75, |
|
"eval_rewards/chosen": -3.4575202465057373, |
|
"eval_rewards/margins": 1.4980329275131226, |
|
"eval_rewards/rejected": -4.9555535316467285, |
|
"eval_runtime": 50.3064, |
|
"eval_samples_per_second": 14.909, |
|
"eval_steps_per_second": 0.477, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"step": 104, |
|
"total_flos": 0.0, |
|
"train_loss": 0.38887147261546207, |
|
"train_runtime": 1183.8142, |
|
"train_samples_per_second": 11.404, |
|
"train_steps_per_second": 0.088 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 104, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 500, |
|
"total_flos": 0.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|