{ "best_metric": 0.74, "best_model_checkpoint": "distilhubert-finetuned-not-a-word2/run-16/checkpoint-96", "epoch": 3.0, "eval_steps": 500, "global_step": 144, "is_hyper_param_search": true, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.1, "grad_norm": 1.2602354288101196, "learning_rate": 6.163228300166189e-06, "loss": 0.7018, "step": 5 }, { "epoch": 0.21, "grad_norm": 1.1513686180114746, "learning_rate": 1.2326456600332378e-05, "loss": 0.6815, "step": 10 }, { "epoch": 0.31, "grad_norm": 1.21480131149292, "learning_rate": 1.8489684900498564e-05, "loss": 0.6818, "step": 15 }, { "epoch": 0.42, "grad_norm": 0.8887011408805847, "learning_rate": 2.4652913200664756e-05, "loss": 0.6556, "step": 20 }, { "epoch": 0.52, "grad_norm": 0.8671120405197144, "learning_rate": 3.081614150083095e-05, "loss": 0.6284, "step": 25 }, { "epoch": 0.62, "grad_norm": 0.6801092028617859, "learning_rate": 3.697936980099713e-05, "loss": 0.6166, "step": 30 }, { "epoch": 0.73, "grad_norm": 2.9359517097473145, "learning_rate": 4.314259810116332e-05, "loss": 0.5407, "step": 35 }, { "epoch": 0.83, "grad_norm": 1.6087604761123657, "learning_rate": 4.930582640132951e-05, "loss": 0.6755, "step": 40 }, { "epoch": 0.94, "grad_norm": 3.7170913219451904, "learning_rate": 5.54690547014957e-05, "loss": 0.6359, "step": 45 }, { "epoch": 1.0, "eval_f1": 0.7326732673267327, "eval_loss": 0.7614002227783203, "eval_runtime": 1.3801, "eval_samples_per_second": 46.373, "eval_steps_per_second": 5.797, "step": 48 }, { "epoch": 1.04, "grad_norm": 4.872888565063477, "learning_rate": 5.889307042381025e-05, "loss": 0.7115, "step": 50 }, { "epoch": 1.15, "grad_norm": 0.9400704503059387, "learning_rate": 5.820826727934734e-05, "loss": 0.7024, "step": 55 }, { "epoch": 1.25, "grad_norm": 2.040548086166382, "learning_rate": 5.752346413488443e-05, "loss": 0.4111, "step": 60 }, { "epoch": 1.35, "grad_norm": 0.7833982706069946, "learning_rate": 5.6838660990421516e-05, "loss": 0.5595, "step": 65 }, { "epoch": 1.46, "grad_norm": 1.4219136238098145, "learning_rate": 5.615385784595861e-05, "loss": 0.4831, "step": 70 }, { "epoch": 1.56, "grad_norm": Infinity, "learning_rate": 5.560601533038828e-05, "loss": 0.7837, "step": 75 }, { "epoch": 1.67, "grad_norm": 2.069016218185425, "learning_rate": 5.492121218592537e-05, "loss": 0.62, "step": 80 }, { "epoch": 1.77, "grad_norm": 1.6181252002716064, "learning_rate": 5.423640904146246e-05, "loss": 0.5911, "step": 85 }, { "epoch": 1.88, "grad_norm": 1.8916351795196533, "learning_rate": 5.355160589699955e-05, "loss": 0.5905, "step": 90 }, { "epoch": 1.98, "grad_norm": 3.4919614791870117, "learning_rate": 5.286680275253664e-05, "loss": 0.5972, "step": 95 }, { "epoch": 2.0, "eval_f1": 0.74, "eval_loss": 0.6621341705322266, "eval_runtime": 1.3734, "eval_samples_per_second": 46.601, "eval_steps_per_second": 5.825, "step": 96 }, { "epoch": 2.08, "grad_norm": 1.276584267616272, "learning_rate": 5.231896023696631e-05, "loss": 0.5767, "step": 100 }, { "epoch": 2.19, "grad_norm": 1.1047593355178833, "learning_rate": 5.163415709250341e-05, "loss": 0.4715, "step": 105 }, { "epoch": 2.29, "grad_norm": 1.551737666130066, "learning_rate": 5.0949353948040496e-05, "loss": 0.5687, "step": 110 }, { "epoch": 2.4, "grad_norm": 1.3372799158096313, "learning_rate": 5.0264550803577585e-05, "loss": 0.5399, "step": 115 }, { "epoch": 2.5, "grad_norm": 4.043832778930664, "learning_rate": 4.9716708288007255e-05, "loss": 0.5178, "step": 120 }, { "epoch": 2.6, "grad_norm": 5.607926845550537, "learning_rate": 4.9031905143544343e-05, "loss": 0.4201, "step": 125 }, { "epoch": 2.71, "grad_norm": 10.094707489013672, "learning_rate": 4.834710199908144e-05, "loss": 0.6339, "step": 130 }, { "epoch": 2.81, "grad_norm": 2.790902614593506, "learning_rate": 4.766229885461853e-05, "loss": 0.3564, "step": 135 }, { "epoch": 2.92, "grad_norm": 8.226430892944336, "learning_rate": 4.6977495710155616e-05, "loss": 0.2918, "step": 140 }, { "epoch": 3.0, "eval_f1": 0.72, "eval_loss": 0.8946094512939453, "eval_runtime": 1.3813, "eval_samples_per_second": 46.334, "eval_steps_per_second": 5.792, "step": 144 } ], "logging_steps": 5, "max_steps": 480, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "total_flos": 2143403497915440.0, "train_batch_size": 4, "trial_name": null, "trial_params": { "learning_rate": 5.916699168159541e-05, "per_device_train_batch_size": 4 } }