{ "best_metric": 0.8089887640449438, "best_model_checkpoint": "distilhubert-finetuned-not-a-word2/run-16/checkpoint-240", "epoch": 5.0, "eval_steps": 500, "global_step": 240, "is_hyper_param_search": true, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.1, "grad_norm": 1.2602354288101196, "learning_rate": 6.163228300166189e-06, "loss": 0.7018, "step": 5 }, { "epoch": 0.21, "grad_norm": 1.1513686180114746, "learning_rate": 1.2326456600332378e-05, "loss": 0.6815, "step": 10 }, { "epoch": 0.31, "grad_norm": 1.21480131149292, "learning_rate": 1.8489684900498564e-05, "loss": 0.6818, "step": 15 }, { "epoch": 0.42, "grad_norm": 0.8887011408805847, "learning_rate": 2.4652913200664756e-05, "loss": 0.6556, "step": 20 }, { "epoch": 0.52, "grad_norm": 0.8671120405197144, "learning_rate": 3.081614150083095e-05, "loss": 0.6284, "step": 25 }, { "epoch": 0.62, "grad_norm": 0.6801092028617859, "learning_rate": 3.697936980099713e-05, "loss": 0.6166, "step": 30 }, { "epoch": 0.73, "grad_norm": 2.9359517097473145, "learning_rate": 4.314259810116332e-05, "loss": 0.5407, "step": 35 }, { "epoch": 0.83, "grad_norm": 1.6087604761123657, "learning_rate": 4.930582640132951e-05, "loss": 0.6755, "step": 40 }, { "epoch": 0.94, "grad_norm": 3.7170913219451904, "learning_rate": 5.54690547014957e-05, "loss": 0.6359, "step": 45 }, { "epoch": 1.0, "eval_f1": 0.7326732673267327, "eval_loss": 0.7614002227783203, "eval_runtime": 1.3801, "eval_samples_per_second": 46.373, "eval_steps_per_second": 5.797, "step": 48 }, { "epoch": 1.04, "grad_norm": 4.872888565063477, "learning_rate": 5.889307042381025e-05, "loss": 0.7115, "step": 50 }, { "epoch": 1.15, "grad_norm": 0.9400704503059387, "learning_rate": 5.820826727934734e-05, "loss": 0.7024, "step": 55 }, { "epoch": 1.25, "grad_norm": 2.040548086166382, "learning_rate": 5.752346413488443e-05, "loss": 0.4111, "step": 60 }, { "epoch": 1.35, "grad_norm": 0.7833982706069946, "learning_rate": 5.6838660990421516e-05, "loss": 0.5595, "step": 65 }, { "epoch": 1.46, "grad_norm": 1.4219136238098145, "learning_rate": 5.615385784595861e-05, "loss": 0.4831, "step": 70 }, { "epoch": 1.56, "grad_norm": Infinity, "learning_rate": 5.560601533038828e-05, "loss": 0.7837, "step": 75 }, { "epoch": 1.67, "grad_norm": 2.069016218185425, "learning_rate": 5.492121218592537e-05, "loss": 0.62, "step": 80 }, { "epoch": 1.77, "grad_norm": 1.6181252002716064, "learning_rate": 5.423640904146246e-05, "loss": 0.5911, "step": 85 }, { "epoch": 1.88, "grad_norm": 1.8916351795196533, "learning_rate": 5.355160589699955e-05, "loss": 0.5905, "step": 90 }, { "epoch": 1.98, "grad_norm": 3.4919614791870117, "learning_rate": 5.286680275253664e-05, "loss": 0.5972, "step": 95 }, { "epoch": 2.0, "eval_f1": 0.74, "eval_loss": 0.6621341705322266, "eval_runtime": 1.3734, "eval_samples_per_second": 46.601, "eval_steps_per_second": 5.825, "step": 96 }, { "epoch": 2.08, "grad_norm": 1.276584267616272, "learning_rate": 5.231896023696631e-05, "loss": 0.5767, "step": 100 }, { "epoch": 2.19, "grad_norm": 1.1047593355178833, "learning_rate": 5.163415709250341e-05, "loss": 0.4715, "step": 105 }, { "epoch": 2.29, "grad_norm": 1.551737666130066, "learning_rate": 5.0949353948040496e-05, "loss": 0.5687, "step": 110 }, { "epoch": 2.4, "grad_norm": 1.3372799158096313, "learning_rate": 5.0264550803577585e-05, "loss": 0.5399, "step": 115 }, { "epoch": 2.5, "grad_norm": 4.043832778930664, "learning_rate": 4.9716708288007255e-05, "loss": 0.5178, "step": 120 }, { "epoch": 2.6, "grad_norm": 5.607926845550537, "learning_rate": 4.9031905143544343e-05, "loss": 0.4201, "step": 125 }, { "epoch": 2.71, "grad_norm": 10.094707489013672, "learning_rate": 4.834710199908144e-05, "loss": 0.6339, "step": 130 }, { "epoch": 2.81, "grad_norm": 2.790902614593506, "learning_rate": 4.766229885461853e-05, "loss": 0.3564, "step": 135 }, { "epoch": 2.92, "grad_norm": 8.226430892944336, "learning_rate": 4.6977495710155616e-05, "loss": 0.2918, "step": 140 }, { "epoch": 3.0, "eval_f1": 0.72, "eval_loss": 0.8946094512939453, "eval_runtime": 1.3813, "eval_samples_per_second": 46.334, "eval_steps_per_second": 5.792, "step": 144 }, { "epoch": 3.02, "grad_norm": 3.521246910095215, "learning_rate": 4.6292692565692705e-05, "loss": 0.4787, "step": 145 }, { "epoch": 3.12, "grad_norm": Infinity, "learning_rate": 4.574485005012238e-05, "loss": 0.3851, "step": 150 }, { "epoch": 3.23, "grad_norm": 3.6489925384521484, "learning_rate": 4.506004690565947e-05, "loss": 0.2152, "step": 155 }, { "epoch": 3.33, "grad_norm": 9.47097396850586, "learning_rate": 4.451220439008914e-05, "loss": 0.6503, "step": 160 }, { "epoch": 3.44, "grad_norm": 1.4761028289794922, "learning_rate": 4.382740124562623e-05, "loss": 0.177, "step": 165 }, { "epoch": 3.54, "grad_norm": 15.07903003692627, "learning_rate": 4.314259810116332e-05, "loss": 0.3834, "step": 170 }, { "epoch": 3.65, "grad_norm": 98.74039459228516, "learning_rate": 4.245779495670041e-05, "loss": 0.816, "step": 175 }, { "epoch": 3.75, "grad_norm": 0.6765172481536865, "learning_rate": 4.17729918122375e-05, "loss": 0.3177, "step": 180 }, { "epoch": 3.85, "grad_norm": 1.5212448835372925, "learning_rate": 4.108818866777459e-05, "loss": 0.3593, "step": 185 }, { "epoch": 3.96, "grad_norm": 27.127269744873047, "learning_rate": 4.040338552331168e-05, "loss": 0.3041, "step": 190 }, { "epoch": 4.0, "eval_f1": 0.7692307692307693, "eval_loss": 1.0041449069976807, "eval_runtime": 1.3873, "eval_samples_per_second": 46.134, "eval_steps_per_second": 5.767, "step": 192 }, { "epoch": 4.06, "grad_norm": 0.7140729427337646, "learning_rate": 3.9718582378848774e-05, "loss": 0.2064, "step": 195 }, { "epoch": 4.17, "grad_norm": 4.367918491363525, "learning_rate": 3.903377923438586e-05, "loss": 0.365, "step": 200 }, { "epoch": 4.27, "grad_norm": 0.4135078489780426, "learning_rate": 3.834897608992295e-05, "loss": 0.3944, "step": 205 }, { "epoch": 4.38, "grad_norm": 88.720458984375, "learning_rate": 3.766417294546004e-05, "loss": 0.3774, "step": 210 }, { "epoch": 4.48, "grad_norm": 1.5230985879898071, "learning_rate": 3.697936980099713e-05, "loss": 0.0844, "step": 215 }, { "epoch": 4.58, "grad_norm": Infinity, "learning_rate": 3.6431527285426805e-05, "loss": 0.3702, "step": 220 }, { "epoch": 4.69, "grad_norm": 39.243804931640625, "learning_rate": 3.5746724140963894e-05, "loss": 0.3381, "step": 225 }, { "epoch": 4.79, "grad_norm": 0.23127637803554535, "learning_rate": 3.506192099650098e-05, "loss": 0.03, "step": 230 }, { "epoch": 4.9, "grad_norm": 23.879899978637695, "learning_rate": 3.451407848093066e-05, "loss": 0.1969, "step": 235 }, { "epoch": 5.0, "grad_norm": 0.2918482720851898, "learning_rate": 3.382927533646775e-05, "loss": 0.2145, "step": 240 }, { "epoch": 5.0, "eval_f1": 0.8089887640449438, "eval_loss": 1.1479806900024414, "eval_runtime": 1.3695, "eval_samples_per_second": 46.731, "eval_steps_per_second": 5.841, "step": 240 } ], "logging_steps": 5, "max_steps": 480, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "total_flos": 3675891927575280.0, "train_batch_size": 4, "trial_name": null, "trial_params": { "learning_rate": 5.916699168159541e-05, "per_device_train_batch_size": 4 } }