colerobertson's picture
Training in progress, epoch 1
d89240d verified
raw
history blame
9.41 kB
{
"best_metric": 0.7551020408163265,
"best_model_checkpoint": "distilhubert-finetuned-not-a-word2/run-0/checkpoint-240",
"epoch": 5.0,
"eval_steps": 500,
"global_step": 240,
"is_hyper_param_search": true,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.1,
"grad_norm": 1.2701194286346436,
"learning_rate": 1.7573936619349767e-06,
"loss": 0.7024,
"step": 5
},
{
"epoch": 0.21,
"grad_norm": 1.218135952949524,
"learning_rate": 3.5147873238699533e-06,
"loss": 0.6932,
"step": 10
},
{
"epoch": 0.31,
"grad_norm": 1.5910921096801758,
"learning_rate": 5.2721809858049295e-06,
"loss": 0.6916,
"step": 15
},
{
"epoch": 0.42,
"grad_norm": 1.0592715740203857,
"learning_rate": 7.029574647739907e-06,
"loss": 0.6854,
"step": 20
},
{
"epoch": 0.52,
"grad_norm": 1.132177472114563,
"learning_rate": 8.786968309674883e-06,
"loss": 0.6719,
"step": 25
},
{
"epoch": 0.62,
"grad_norm": 0.9851714372634888,
"learning_rate": 1.0544361971609859e-05,
"loss": 0.6633,
"step": 30
},
{
"epoch": 0.73,
"grad_norm": 1.5592440366744995,
"learning_rate": 1.2301755633544835e-05,
"loss": 0.6288,
"step": 35
},
{
"epoch": 0.83,
"grad_norm": 1.8793619871139526,
"learning_rate": 1.4059149295479813e-05,
"loss": 0.6561,
"step": 40
},
{
"epoch": 0.94,
"grad_norm": 1.987308382987976,
"learning_rate": 1.581654295741479e-05,
"loss": 0.6343,
"step": 45
},
{
"epoch": 1.0,
"eval_f1": 0.7326732673267327,
"eval_loss": 0.6868629455566406,
"eval_runtime": 1.3443,
"eval_samples_per_second": 47.609,
"eval_steps_per_second": 5.951,
"step": 48
},
{
"epoch": 1.04,
"grad_norm": 3.1652989387512207,
"learning_rate": 1.6792872769600888e-05,
"loss": 0.6712,
"step": 50
},
{
"epoch": 1.15,
"grad_norm": 0.9990371465682983,
"learning_rate": 1.6597606807163667e-05,
"loss": 0.6657,
"step": 55
},
{
"epoch": 1.25,
"grad_norm": 1.8210965394973755,
"learning_rate": 1.640234084472645e-05,
"loss": 0.476,
"step": 60
},
{
"epoch": 1.35,
"grad_norm": 0.5364782810211182,
"learning_rate": 1.6207074882289228e-05,
"loss": 0.5768,
"step": 65
},
{
"epoch": 1.46,
"grad_norm": 1.6772537231445312,
"learning_rate": 1.6011808919852007e-05,
"loss": 0.5039,
"step": 70
},
{
"epoch": 1.56,
"grad_norm": 1.0453933477401733,
"learning_rate": 1.581654295741479e-05,
"loss": 0.7182,
"step": 75
},
{
"epoch": 1.67,
"grad_norm": 1.621067762374878,
"learning_rate": 1.562127699497757e-05,
"loss": 0.6178,
"step": 80
},
{
"epoch": 1.77,
"grad_norm": 1.7878597974777222,
"learning_rate": 1.542601103254035e-05,
"loss": 0.6192,
"step": 85
},
{
"epoch": 1.88,
"grad_norm": 1.6824157238006592,
"learning_rate": 1.5230745070103131e-05,
"loss": 0.6109,
"step": 90
},
{
"epoch": 1.98,
"grad_norm": 0.7989784479141235,
"learning_rate": 1.5035479107665912e-05,
"loss": 0.6367,
"step": 95
},
{
"epoch": 2.0,
"eval_f1": 0.7326732673267327,
"eval_loss": 0.7258987426757812,
"eval_runtime": 1.3947,
"eval_samples_per_second": 45.888,
"eval_steps_per_second": 5.736,
"step": 96
},
{
"epoch": 2.08,
"grad_norm": 0.6805263757705688,
"learning_rate": 1.484021314522869e-05,
"loss": 0.6271,
"step": 100
},
{
"epoch": 2.19,
"grad_norm": 0.6469630002975464,
"learning_rate": 1.4644947182791471e-05,
"loss": 0.5085,
"step": 105
},
{
"epoch": 2.29,
"grad_norm": 0.9272030591964722,
"learning_rate": 1.4449681220354252e-05,
"loss": 0.6377,
"step": 110
},
{
"epoch": 2.4,
"grad_norm": 0.7515610456466675,
"learning_rate": 1.4254415257917033e-05,
"loss": 0.655,
"step": 115
},
{
"epoch": 2.5,
"grad_norm": 2.1927835941314697,
"learning_rate": 1.4059149295479813e-05,
"loss": 0.5797,
"step": 120
},
{
"epoch": 2.6,
"grad_norm": 1.0289098024368286,
"learning_rate": 1.3863883333042594e-05,
"loss": 0.5291,
"step": 125
},
{
"epoch": 2.71,
"grad_norm": 1.7173566818237305,
"learning_rate": 1.3668617370605374e-05,
"loss": 0.6829,
"step": 130
},
{
"epoch": 2.81,
"grad_norm": 0.8574482202529907,
"learning_rate": 1.3473351408168155e-05,
"loss": 0.4509,
"step": 135
},
{
"epoch": 2.92,
"grad_norm": 1.1971725225448608,
"learning_rate": 1.3278085445730936e-05,
"loss": 0.4369,
"step": 140
},
{
"epoch": 3.0,
"eval_f1": 0.7326732673267327,
"eval_loss": 0.7368831634521484,
"eval_runtime": 1.4074,
"eval_samples_per_second": 45.475,
"eval_steps_per_second": 5.684,
"step": 144
},
{
"epoch": 3.02,
"grad_norm": 2.6530747413635254,
"learning_rate": 1.3082819483293713e-05,
"loss": 0.6167,
"step": 145
},
{
"epoch": 3.12,
"grad_norm": 2.0653493404388428,
"learning_rate": 1.2887553520856493e-05,
"loss": 0.5266,
"step": 150
},
{
"epoch": 3.23,
"grad_norm": 1.5351653099060059,
"learning_rate": 1.2692287558419274e-05,
"loss": 0.5224,
"step": 155
},
{
"epoch": 3.33,
"grad_norm": 1.5351920127868652,
"learning_rate": 1.2497021595982055e-05,
"loss": 0.622,
"step": 160
},
{
"epoch": 3.44,
"grad_norm": 1.5447399616241455,
"learning_rate": 1.2301755633544835e-05,
"loss": 0.4725,
"step": 165
},
{
"epoch": 3.54,
"grad_norm": 4.063059329986572,
"learning_rate": 1.2106489671107616e-05,
"loss": 0.4089,
"step": 170
},
{
"epoch": 3.65,
"grad_norm": 2.517096519470215,
"learning_rate": 1.1911223708670397e-05,
"loss": 0.579,
"step": 175
},
{
"epoch": 3.75,
"grad_norm": 2.847086191177368,
"learning_rate": 1.1755010938720622e-05,
"loss": 0.4687,
"step": 180
},
{
"epoch": 3.85,
"grad_norm": 1.7841814756393433,
"learning_rate": 1.1559744976283403e-05,
"loss": 0.4846,
"step": 185
},
{
"epoch": 3.96,
"grad_norm": 3.5284550189971924,
"learning_rate": 1.1364479013846184e-05,
"loss": 0.5628,
"step": 190
},
{
"epoch": 4.0,
"eval_f1": 0.74,
"eval_loss": 0.6948537826538086,
"eval_runtime": 1.3673,
"eval_samples_per_second": 46.807,
"eval_steps_per_second": 5.851,
"step": 192
},
{
"epoch": 4.06,
"grad_norm": 1.6642357110977173,
"learning_rate": 1.1169213051408962e-05,
"loss": 0.5529,
"step": 195
},
{
"epoch": 4.17,
"grad_norm": 3.0695078372955322,
"learning_rate": 1.0973947088971741e-05,
"loss": 0.3534,
"step": 200
},
{
"epoch": 4.27,
"grad_norm": 2.5939130783081055,
"learning_rate": 1.0778681126534522e-05,
"loss": 0.4534,
"step": 205
},
{
"epoch": 4.38,
"grad_norm": 9.777613639831543,
"learning_rate": 1.0583415164097303e-05,
"loss": 0.5345,
"step": 210
},
{
"epoch": 4.48,
"grad_norm": 4.490750312805176,
"learning_rate": 1.0388149201660083e-05,
"loss": 0.4854,
"step": 215
},
{
"epoch": 4.58,
"grad_norm": 3.7475597858428955,
"learning_rate": 1.0192883239222864e-05,
"loss": 0.4174,
"step": 220
},
{
"epoch": 4.69,
"grad_norm": 1.7532075643539429,
"learning_rate": 9.997617276785644e-06,
"loss": 0.4379,
"step": 225
},
{
"epoch": 4.79,
"grad_norm": 1.4538604021072388,
"learning_rate": 9.802351314348425e-06,
"loss": 0.3269,
"step": 230
},
{
"epoch": 4.9,
"grad_norm": 3.058314085006714,
"learning_rate": 9.607085351911206e-06,
"loss": 0.3553,
"step": 235
},
{
"epoch": 5.0,
"grad_norm": 5.367821216583252,
"learning_rate": 9.45087258196143e-06,
"loss": 0.4753,
"step": 240
},
{
"epoch": 5.0,
"eval_f1": 0.7551020408163265,
"eval_loss": 0.7308206558227539,
"eval_runtime": 1.3546,
"eval_samples_per_second": 47.246,
"eval_steps_per_second": 5.906,
"step": 240
}
],
"logging_steps": 5,
"max_steps": 480,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 500,
"total_flos": 3654362860415712.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": {
"learning_rate": 1.6870979154575775e-05,
"per_device_train_batch_size": 4
}
}