|
{
|
|
"best_metric": 18.436363636363637,
|
|
"best_model_checkpoint": "galgame-whisper\\checkpoint-24500",
|
|
"epoch": 13.0,
|
|
"eval_steps": 500,
|
|
"global_step": 26624,
|
|
"is_hyper_param_search": false,
|
|
"is_local_process_zero": true,
|
|
"is_world_process_zero": true,
|
|
"log_history": [
|
|
{
|
|
"epoch": 0.01220703125,
|
|
"grad_norm": 11.14741039276123,
|
|
"learning_rate": 4.600000000000001e-06,
|
|
"loss": 1.7805,
|
|
"step": 25
|
|
},
|
|
{
|
|
"epoch": 0.0244140625,
|
|
"grad_norm": 6.63244104385376,
|
|
"learning_rate": 9.600000000000001e-06,
|
|
"loss": 1.3099,
|
|
"step": 50
|
|
},
|
|
{
|
|
"epoch": 0.03662109375,
|
|
"grad_norm": 4.578396320343018,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.9171,
|
|
"step": 75
|
|
},
|
|
{
|
|
"epoch": 0.048828125,
|
|
"grad_norm": 5.572494029998779,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.6639,
|
|
"step": 100
|
|
},
|
|
{
|
|
"epoch": 0.048828125,
|
|
"eval_cer": 25.6,
|
|
"eval_loss": 0.593059778213501,
|
|
"eval_normalized_cer": 20.427455815865187,
|
|
"eval_runtime": 123.1903,
|
|
"eval_samples_per_second": 1.039,
|
|
"eval_steps_per_second": 0.13,
|
|
"step": 100
|
|
},
|
|
{
|
|
"epoch": 0.06103515625,
|
|
"grad_norm": 5.427572250366211,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.6159,
|
|
"step": 125
|
|
},
|
|
{
|
|
"epoch": 0.0732421875,
|
|
"grad_norm": 4.667797565460205,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5685,
|
|
"step": 150
|
|
},
|
|
{
|
|
"epoch": 0.08544921875,
|
|
"grad_norm": 4.875271797180176,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5758,
|
|
"step": 175
|
|
},
|
|
{
|
|
"epoch": 0.09765625,
|
|
"grad_norm": 4.109724044799805,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5772,
|
|
"step": 200
|
|
},
|
|
{
|
|
"epoch": 0.09765625,
|
|
"eval_cer": 25.418181818181818,
|
|
"eval_loss": 0.567274808883667,
|
|
"eval_normalized_cer": 20.34525277435265,
|
|
"eval_runtime": 140.399,
|
|
"eval_samples_per_second": 0.912,
|
|
"eval_steps_per_second": 0.114,
|
|
"step": 200
|
|
},
|
|
{
|
|
"epoch": 0.10986328125,
|
|
"grad_norm": 4.04423189163208,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5723,
|
|
"step": 225
|
|
},
|
|
{
|
|
"epoch": 0.1220703125,
|
|
"grad_norm": 4.566165447235107,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.6474,
|
|
"step": 250
|
|
},
|
|
{
|
|
"epoch": 0.13427734375,
|
|
"grad_norm": 4.996761798858643,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.611,
|
|
"step": 275
|
|
},
|
|
{
|
|
"epoch": 0.146484375,
|
|
"grad_norm": 4.392452716827393,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5462,
|
|
"step": 300
|
|
},
|
|
{
|
|
"epoch": 0.146484375,
|
|
"eval_cer": 22.21818181818182,
|
|
"eval_loss": 0.5370786190032959,
|
|
"eval_normalized_cer": 17.673653925195232,
|
|
"eval_runtime": 134.3037,
|
|
"eval_samples_per_second": 0.953,
|
|
"eval_steps_per_second": 0.119,
|
|
"step": 300
|
|
},
|
|
{
|
|
"epoch": 0.15869140625,
|
|
"grad_norm": 3.719304084777832,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5986,
|
|
"step": 325
|
|
},
|
|
{
|
|
"epoch": 0.1708984375,
|
|
"grad_norm": 11.704819679260254,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.6941,
|
|
"step": 350
|
|
},
|
|
{
|
|
"epoch": 0.18310546875,
|
|
"grad_norm": 5.188851833343506,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5486,
|
|
"step": 375
|
|
},
|
|
{
|
|
"epoch": 0.1953125,
|
|
"grad_norm": 3.8557848930358887,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.6085,
|
|
"step": 400
|
|
},
|
|
{
|
|
"epoch": 0.1953125,
|
|
"eval_cer": 21.854545454545455,
|
|
"eval_loss": 0.5271878838539124,
|
|
"eval_normalized_cer": 17.221537196876284,
|
|
"eval_runtime": 131.3514,
|
|
"eval_samples_per_second": 0.974,
|
|
"eval_steps_per_second": 0.122,
|
|
"step": 400
|
|
},
|
|
{
|
|
"epoch": 0.20751953125,
|
|
"grad_norm": 3.7432358264923096,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.6387,
|
|
"step": 425
|
|
},
|
|
{
|
|
"epoch": 0.2197265625,
|
|
"grad_norm": 3.6960973739624023,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5481,
|
|
"step": 450
|
|
},
|
|
{
|
|
"epoch": 0.23193359375,
|
|
"grad_norm": 5.312159061431885,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5565,
|
|
"step": 475
|
|
},
|
|
{
|
|
"epoch": 0.244140625,
|
|
"grad_norm": 5.376567363739014,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5262,
|
|
"step": 500
|
|
},
|
|
{
|
|
"epoch": 0.244140625,
|
|
"eval_cer": 24.509090909090908,
|
|
"eval_loss": 0.5245115756988525,
|
|
"eval_normalized_cer": 19.482120838471022,
|
|
"eval_runtime": 136.6132,
|
|
"eval_samples_per_second": 0.937,
|
|
"eval_steps_per_second": 0.117,
|
|
"step": 500
|
|
},
|
|
{
|
|
"epoch": 0.25634765625,
|
|
"grad_norm": 3.251194715499878,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5751,
|
|
"step": 525
|
|
},
|
|
{
|
|
"epoch": 0.2685546875,
|
|
"grad_norm": 3.575839042663574,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5391,
|
|
"step": 550
|
|
},
|
|
{
|
|
"epoch": 0.28076171875,
|
|
"grad_norm": 4.032149314880371,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5314,
|
|
"step": 575
|
|
},
|
|
{
|
|
"epoch": 0.29296875,
|
|
"grad_norm": 4.987039566040039,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.576,
|
|
"step": 600
|
|
},
|
|
{
|
|
"epoch": 0.29296875,
|
|
"eval_cer": 22.69090909090909,
|
|
"eval_loss": 0.5259984135627747,
|
|
"eval_normalized_cer": 18.207973695026716,
|
|
"eval_runtime": 127.1247,
|
|
"eval_samples_per_second": 1.007,
|
|
"eval_steps_per_second": 0.126,
|
|
"step": 600
|
|
},
|
|
{
|
|
"epoch": 0.30517578125,
|
|
"grad_norm": 4.042311191558838,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5671,
|
|
"step": 625
|
|
},
|
|
{
|
|
"epoch": 0.3173828125,
|
|
"grad_norm": 5.254710674285889,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5525,
|
|
"step": 650
|
|
},
|
|
{
|
|
"epoch": 0.32958984375,
|
|
"grad_norm": 3.7549564838409424,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5489,
|
|
"step": 675
|
|
},
|
|
{
|
|
"epoch": 0.341796875,
|
|
"grad_norm": 4.815126419067383,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5844,
|
|
"step": 700
|
|
},
|
|
{
|
|
"epoch": 0.341796875,
|
|
"eval_cer": 26.581818181818186,
|
|
"eval_loss": 0.5143883228302002,
|
|
"eval_normalized_cer": 21.98931360460337,
|
|
"eval_runtime": 125.4245,
|
|
"eval_samples_per_second": 1.021,
|
|
"eval_steps_per_second": 0.128,
|
|
"step": 700
|
|
},
|
|
{
|
|
"epoch": 0.35400390625,
|
|
"grad_norm": 3.240372657775879,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5684,
|
|
"step": 725
|
|
},
|
|
{
|
|
"epoch": 0.3662109375,
|
|
"grad_norm": 4.83163595199585,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5157,
|
|
"step": 750
|
|
},
|
|
{
|
|
"epoch": 0.37841796875,
|
|
"grad_norm": 3.209136724472046,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5348,
|
|
"step": 775
|
|
},
|
|
{
|
|
"epoch": 0.390625,
|
|
"grad_norm": 3.9407966136932373,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5689,
|
|
"step": 800
|
|
},
|
|
{
|
|
"epoch": 0.390625,
|
|
"eval_cer": 23.200000000000003,
|
|
"eval_loss": 0.5060851573944092,
|
|
"eval_normalized_cer": 18.372379778051787,
|
|
"eval_runtime": 132.0722,
|
|
"eval_samples_per_second": 0.969,
|
|
"eval_steps_per_second": 0.121,
|
|
"step": 800
|
|
},
|
|
{
|
|
"epoch": 0.40283203125,
|
|
"grad_norm": 5.201825141906738,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5981,
|
|
"step": 825
|
|
},
|
|
{
|
|
"epoch": 0.4150390625,
|
|
"grad_norm": 4.420695781707764,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.6028,
|
|
"step": 850
|
|
},
|
|
{
|
|
"epoch": 0.42724609375,
|
|
"grad_norm": 4.9300856590271,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5807,
|
|
"step": 875
|
|
},
|
|
{
|
|
"epoch": 0.439453125,
|
|
"grad_norm": 4.098893165588379,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5731,
|
|
"step": 900
|
|
},
|
|
{
|
|
"epoch": 0.439453125,
|
|
"eval_cer": 22.0,
|
|
"eval_loss": 0.4978080987930298,
|
|
"eval_normalized_cer": 17.7147554459515,
|
|
"eval_runtime": 131.4521,
|
|
"eval_samples_per_second": 0.974,
|
|
"eval_steps_per_second": 0.122,
|
|
"step": 900
|
|
},
|
|
{
|
|
"epoch": 0.45166015625,
|
|
"grad_norm": 3.947633981704712,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4836,
|
|
"step": 925
|
|
},
|
|
{
|
|
"epoch": 0.4638671875,
|
|
"grad_norm": 3.02842378616333,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5248,
|
|
"step": 950
|
|
},
|
|
{
|
|
"epoch": 0.47607421875,
|
|
"grad_norm": 5.237096309661865,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5299,
|
|
"step": 975
|
|
},
|
|
{
|
|
"epoch": 0.48828125,
|
|
"grad_norm": 4.260601997375488,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5058,
|
|
"step": 1000
|
|
},
|
|
{
|
|
"epoch": 0.48828125,
|
|
"eval_cer": 23.599999999999998,
|
|
"eval_loss": 0.494682252407074,
|
|
"eval_normalized_cer": 18.988902589395806,
|
|
"eval_runtime": 132.3816,
|
|
"eval_samples_per_second": 0.967,
|
|
"eval_steps_per_second": 0.121,
|
|
"step": 1000
|
|
},
|
|
{
|
|
"epoch": 0.50048828125,
|
|
"grad_norm": 4.105225086212158,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4992,
|
|
"step": 1025
|
|
},
|
|
{
|
|
"epoch": 0.5126953125,
|
|
"grad_norm": 4.636739730834961,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5517,
|
|
"step": 1050
|
|
},
|
|
{
|
|
"epoch": 0.52490234375,
|
|
"grad_norm": 4.676263809204102,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.511,
|
|
"step": 1075
|
|
},
|
|
{
|
|
"epoch": 0.537109375,
|
|
"grad_norm": 4.2934465408325195,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5556,
|
|
"step": 1100
|
|
},
|
|
{
|
|
"epoch": 0.537109375,
|
|
"eval_cer": 23.454545454545457,
|
|
"eval_loss": 0.4881322383880615,
|
|
"eval_normalized_cer": 19.399917796958487,
|
|
"eval_runtime": 148.1276,
|
|
"eval_samples_per_second": 0.864,
|
|
"eval_steps_per_second": 0.054,
|
|
"step": 1100
|
|
},
|
|
{
|
|
"epoch": 0.54931640625,
|
|
"grad_norm": 3.7905611991882324,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5309,
|
|
"step": 1125
|
|
},
|
|
{
|
|
"epoch": 0.5615234375,
|
|
"grad_norm": 3.6019198894500732,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4845,
|
|
"step": 1150
|
|
},
|
|
{
|
|
"epoch": 0.57373046875,
|
|
"grad_norm": 3.77414870262146,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.491,
|
|
"step": 1175
|
|
},
|
|
{
|
|
"epoch": 0.5859375,
|
|
"grad_norm": 3.13875412940979,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5033,
|
|
"step": 1200
|
|
},
|
|
{
|
|
"epoch": 0.5859375,
|
|
"eval_cer": 23.81818181818182,
|
|
"eval_loss": 0.48237502574920654,
|
|
"eval_normalized_cer": 19.31771475544595,
|
|
"eval_runtime": 167.3993,
|
|
"eval_samples_per_second": 0.765,
|
|
"eval_steps_per_second": 0.048,
|
|
"step": 1200
|
|
},
|
|
{
|
|
"epoch": 0.59814453125,
|
|
"grad_norm": 5.719494819641113,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5085,
|
|
"step": 1225
|
|
},
|
|
{
|
|
"epoch": 0.6103515625,
|
|
"grad_norm": 3.9417619705200195,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4764,
|
|
"step": 1250
|
|
},
|
|
{
|
|
"epoch": 0.62255859375,
|
|
"grad_norm": 4.024901390075684,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4834,
|
|
"step": 1275
|
|
},
|
|
{
|
|
"epoch": 0.634765625,
|
|
"grad_norm": 4.798065185546875,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5041,
|
|
"step": 1300
|
|
},
|
|
{
|
|
"epoch": 0.634765625,
|
|
"eval_cer": 21.672727272727276,
|
|
"eval_loss": 0.47094789147377014,
|
|
"eval_normalized_cer": 17.632552404438965,
|
|
"eval_runtime": 167.9645,
|
|
"eval_samples_per_second": 0.762,
|
|
"eval_steps_per_second": 0.048,
|
|
"step": 1300
|
|
},
|
|
{
|
|
"epoch": 0.64697265625,
|
|
"grad_norm": 3.018202781677246,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5181,
|
|
"step": 1325
|
|
},
|
|
{
|
|
"epoch": 0.6591796875,
|
|
"grad_norm": 4.9327311515808105,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5419,
|
|
"step": 1350
|
|
},
|
|
{
|
|
"epoch": 0.67138671875,
|
|
"grad_norm": 4.631186485290527,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4995,
|
|
"step": 1375
|
|
},
|
|
{
|
|
"epoch": 0.68359375,
|
|
"grad_norm": 5.249241828918457,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5324,
|
|
"step": 1400
|
|
},
|
|
{
|
|
"epoch": 0.68359375,
|
|
"eval_cer": 21.01818181818182,
|
|
"eval_loss": 0.4812983274459839,
|
|
"eval_normalized_cer": 16.605014385532264,
|
|
"eval_runtime": 164.7775,
|
|
"eval_samples_per_second": 0.777,
|
|
"eval_steps_per_second": 0.049,
|
|
"step": 1400
|
|
},
|
|
{
|
|
"epoch": 0.69580078125,
|
|
"grad_norm": 3.599472761154175,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5059,
|
|
"step": 1425
|
|
},
|
|
{
|
|
"epoch": 0.7080078125,
|
|
"grad_norm": 2.960740089416504,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.474,
|
|
"step": 1450
|
|
},
|
|
{
|
|
"epoch": 0.72021484375,
|
|
"grad_norm": 5.463324546813965,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4646,
|
|
"step": 1475
|
|
},
|
|
{
|
|
"epoch": 0.732421875,
|
|
"grad_norm": 2.3991730213165283,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5374,
|
|
"step": 1500
|
|
},
|
|
{
|
|
"epoch": 0.732421875,
|
|
"eval_cer": 21.30909090909091,
|
|
"eval_loss": 0.4700665771961212,
|
|
"eval_normalized_cer": 17.057131113851213,
|
|
"eval_runtime": 168.2533,
|
|
"eval_samples_per_second": 0.761,
|
|
"eval_steps_per_second": 0.048,
|
|
"step": 1500
|
|
},
|
|
{
|
|
"epoch": 0.74462890625,
|
|
"grad_norm": 3.7515509128570557,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4979,
|
|
"step": 1525
|
|
},
|
|
{
|
|
"epoch": 0.7568359375,
|
|
"grad_norm": 3.6961734294891357,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5254,
|
|
"step": 1550
|
|
},
|
|
{
|
|
"epoch": 0.76904296875,
|
|
"grad_norm": 5.233846187591553,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5757,
|
|
"step": 1575
|
|
},
|
|
{
|
|
"epoch": 0.78125,
|
|
"grad_norm": 3.659038543701172,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4875,
|
|
"step": 1600
|
|
},
|
|
{
|
|
"epoch": 0.78125,
|
|
"eval_cer": 20.872727272727275,
|
|
"eval_loss": 0.4664301574230194,
|
|
"eval_normalized_cer": 16.48170982326346,
|
|
"eval_runtime": 170.1628,
|
|
"eval_samples_per_second": 0.752,
|
|
"eval_steps_per_second": 0.047,
|
|
"step": 1600
|
|
},
|
|
{
|
|
"epoch": 0.79345703125,
|
|
"grad_norm": 3.4504241943359375,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4794,
|
|
"step": 1625
|
|
},
|
|
{
|
|
"epoch": 0.8056640625,
|
|
"grad_norm": 3.2377774715423584,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5338,
|
|
"step": 1650
|
|
},
|
|
{
|
|
"epoch": 0.81787109375,
|
|
"grad_norm": 3.6238794326782227,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4747,
|
|
"step": 1675
|
|
},
|
|
{
|
|
"epoch": 0.830078125,
|
|
"grad_norm": 3.8324549198150635,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5155,
|
|
"step": 1700
|
|
},
|
|
{
|
|
"epoch": 0.830078125,
|
|
"eval_cer": 22.254545454545454,
|
|
"eval_loss": 0.4597916305065155,
|
|
"eval_normalized_cer": 18.12577065351418,
|
|
"eval_runtime": 166.0402,
|
|
"eval_samples_per_second": 0.771,
|
|
"eval_steps_per_second": 0.048,
|
|
"step": 1700
|
|
},
|
|
{
|
|
"epoch": 0.84228515625,
|
|
"grad_norm": 4.575786113739014,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.515,
|
|
"step": 1725
|
|
},
|
|
{
|
|
"epoch": 0.8544921875,
|
|
"grad_norm": 3.636530637741089,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4976,
|
|
"step": 1750
|
|
},
|
|
{
|
|
"epoch": 0.86669921875,
|
|
"grad_norm": 4.921797752380371,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5211,
|
|
"step": 1775
|
|
},
|
|
{
|
|
"epoch": 0.87890625,
|
|
"grad_norm": 3.620969533920288,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4824,
|
|
"step": 1800
|
|
},
|
|
{
|
|
"epoch": 0.87890625,
|
|
"eval_cer": 26.181818181818183,
|
|
"eval_loss": 0.46164897084236145,
|
|
"eval_normalized_cer": 21.537196876284423,
|
|
"eval_runtime": 150.2,
|
|
"eval_samples_per_second": 0.852,
|
|
"eval_steps_per_second": 0.053,
|
|
"step": 1800
|
|
},
|
|
{
|
|
"epoch": 0.89111328125,
|
|
"grad_norm": 4.432845592498779,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5138,
|
|
"step": 1825
|
|
},
|
|
{
|
|
"epoch": 0.9033203125,
|
|
"grad_norm": 3.5764822959899902,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5413,
|
|
"step": 1850
|
|
},
|
|
{
|
|
"epoch": 0.91552734375,
|
|
"grad_norm": 4.142552852630615,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4824,
|
|
"step": 1875
|
|
},
|
|
{
|
|
"epoch": 0.927734375,
|
|
"grad_norm": 3.680915594100952,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5295,
|
|
"step": 1900
|
|
},
|
|
{
|
|
"epoch": 0.927734375,
|
|
"eval_cer": 20.327272727272728,
|
|
"eval_loss": 0.45038798451423645,
|
|
"eval_normalized_cer": 15.906288532675708,
|
|
"eval_runtime": 173.3587,
|
|
"eval_samples_per_second": 0.738,
|
|
"eval_steps_per_second": 0.046,
|
|
"step": 1900
|
|
},
|
|
{
|
|
"epoch": 0.93994140625,
|
|
"grad_norm": 2.5720534324645996,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5137,
|
|
"step": 1925
|
|
},
|
|
{
|
|
"epoch": 0.9521484375,
|
|
"grad_norm": 2.782604455947876,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5086,
|
|
"step": 1950
|
|
},
|
|
{
|
|
"epoch": 0.96435546875,
|
|
"grad_norm": 4.596216201782227,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.525,
|
|
"step": 1975
|
|
},
|
|
{
|
|
"epoch": 0.9765625,
|
|
"grad_norm": 4.53118371963501,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5004,
|
|
"step": 2000
|
|
},
|
|
{
|
|
"epoch": 0.9765625,
|
|
"eval_cer": 22.145454545454545,
|
|
"eval_loss": 0.4644533395767212,
|
|
"eval_normalized_cer": 17.26263871763255,
|
|
"eval_runtime": 170.2375,
|
|
"eval_samples_per_second": 0.752,
|
|
"eval_steps_per_second": 0.047,
|
|
"step": 2000
|
|
},
|
|
{
|
|
"epoch": 0.98876953125,
|
|
"grad_norm": 3.8377697467803955,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5364,
|
|
"step": 2025
|
|
},
|
|
{
|
|
"epoch": 1.0009765625,
|
|
"grad_norm": 6.334020614624023,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.619,
|
|
"step": 2050
|
|
},
|
|
{
|
|
"epoch": 1.01318359375,
|
|
"grad_norm": 4.159861087799072,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5219,
|
|
"step": 2075
|
|
},
|
|
{
|
|
"epoch": 1.025390625,
|
|
"grad_norm": 3.607752799987793,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.529,
|
|
"step": 2100
|
|
},
|
|
{
|
|
"epoch": 1.03759765625,
|
|
"grad_norm": 5.036830425262451,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5207,
|
|
"step": 2125
|
|
},
|
|
{
|
|
"epoch": 1.0498046875,
|
|
"grad_norm": 3.520869016647339,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5213,
|
|
"step": 2150
|
|
},
|
|
{
|
|
"epoch": 1.06201171875,
|
|
"grad_norm": 4.612700939178467,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4954,
|
|
"step": 2175
|
|
},
|
|
{
|
|
"epoch": 1.07421875,
|
|
"grad_norm": 4.156740188598633,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5406,
|
|
"step": 2200
|
|
},
|
|
{
|
|
"epoch": 1.08642578125,
|
|
"grad_norm": 3.5834848880767822,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5097,
|
|
"step": 2225
|
|
},
|
|
{
|
|
"epoch": 1.0986328125,
|
|
"grad_norm": 3.4885971546173096,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4756,
|
|
"step": 2250
|
|
},
|
|
{
|
|
"epoch": 1.11083984375,
|
|
"grad_norm": 5.202791213989258,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5375,
|
|
"step": 2275
|
|
},
|
|
{
|
|
"epoch": 1.123046875,
|
|
"grad_norm": 2.762514114379883,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4792,
|
|
"step": 2300
|
|
},
|
|
{
|
|
"epoch": 1.13525390625,
|
|
"grad_norm": 2.9590845108032227,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4671,
|
|
"step": 2325
|
|
},
|
|
{
|
|
"epoch": 1.1474609375,
|
|
"grad_norm": 2.85101318359375,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4856,
|
|
"step": 2350
|
|
},
|
|
{
|
|
"epoch": 1.15966796875,
|
|
"grad_norm": 3.444801092147827,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5393,
|
|
"step": 2375
|
|
},
|
|
{
|
|
"epoch": 1.171875,
|
|
"grad_norm": 4.466598033905029,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5304,
|
|
"step": 2400
|
|
},
|
|
{
|
|
"epoch": 1.18408203125,
|
|
"grad_norm": 3.2998430728912354,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5008,
|
|
"step": 2425
|
|
},
|
|
{
|
|
"epoch": 1.1962890625,
|
|
"grad_norm": 3.405848264694214,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4788,
|
|
"step": 2450
|
|
},
|
|
{
|
|
"epoch": 1.20849609375,
|
|
"grad_norm": 3.5294339656829834,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4954,
|
|
"step": 2475
|
|
},
|
|
{
|
|
"epoch": 1.220703125,
|
|
"grad_norm": 3.5182113647460938,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5109,
|
|
"step": 2500
|
|
},
|
|
{
|
|
"epoch": 1.220703125,
|
|
"eval_cer": 22.8,
|
|
"eval_loss": 0.449319064617157,
|
|
"eval_normalized_cer": 18.495684340320594,
|
|
"eval_runtime": 122.2523,
|
|
"eval_samples_per_second": 1.047,
|
|
"eval_steps_per_second": 0.065,
|
|
"step": 2500
|
|
},
|
|
{
|
|
"epoch": 1.23291015625,
|
|
"grad_norm": 4.339564800262451,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5178,
|
|
"step": 2525
|
|
},
|
|
{
|
|
"epoch": 1.2451171875,
|
|
"grad_norm": 3.9555203914642334,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4995,
|
|
"step": 2550
|
|
},
|
|
{
|
|
"epoch": 1.25732421875,
|
|
"grad_norm": 2.8713884353637695,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4892,
|
|
"step": 2575
|
|
},
|
|
{
|
|
"epoch": 1.26953125,
|
|
"grad_norm": 3.2301995754241943,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4916,
|
|
"step": 2600
|
|
},
|
|
{
|
|
"epoch": 1.28173828125,
|
|
"grad_norm": Infinity,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5373,
|
|
"step": 2625
|
|
},
|
|
{
|
|
"epoch": 1.2939453125,
|
|
"grad_norm": 3.3091487884521484,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4517,
|
|
"step": 2650
|
|
},
|
|
{
|
|
"epoch": 1.30615234375,
|
|
"grad_norm": 5.547422409057617,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5115,
|
|
"step": 2675
|
|
},
|
|
{
|
|
"epoch": 1.318359375,
|
|
"grad_norm": 4.931210517883301,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4886,
|
|
"step": 2700
|
|
},
|
|
{
|
|
"epoch": 1.33056640625,
|
|
"grad_norm": 4.160281658172607,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4653,
|
|
"step": 2725
|
|
},
|
|
{
|
|
"epoch": 1.3427734375,
|
|
"grad_norm": 3.172577381134033,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5092,
|
|
"step": 2750
|
|
},
|
|
{
|
|
"epoch": 1.35498046875,
|
|
"grad_norm": 4.536301612854004,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5189,
|
|
"step": 2775
|
|
},
|
|
{
|
|
"epoch": 1.3671875,
|
|
"grad_norm": 4.744750499725342,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4882,
|
|
"step": 2800
|
|
},
|
|
{
|
|
"epoch": 1.37939453125,
|
|
"grad_norm": 4.030979633331299,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4538,
|
|
"step": 2825
|
|
},
|
|
{
|
|
"epoch": 1.3916015625,
|
|
"grad_norm": 4.93550443649292,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4778,
|
|
"step": 2850
|
|
},
|
|
{
|
|
"epoch": 1.40380859375,
|
|
"grad_norm": 3.9617207050323486,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4542,
|
|
"step": 2875
|
|
},
|
|
{
|
|
"epoch": 1.416015625,
|
|
"grad_norm": 2.711639642715454,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.501,
|
|
"step": 2900
|
|
},
|
|
{
|
|
"epoch": 1.42822265625,
|
|
"grad_norm": 3.6887452602386475,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4909,
|
|
"step": 2925
|
|
},
|
|
{
|
|
"epoch": 1.4404296875,
|
|
"grad_norm": 4.250792026519775,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4814,
|
|
"step": 2950
|
|
},
|
|
{
|
|
"epoch": 1.45263671875,
|
|
"grad_norm": 3.804023265838623,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4922,
|
|
"step": 2975
|
|
},
|
|
{
|
|
"epoch": 1.46484375,
|
|
"grad_norm": 4.579716205596924,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4816,
|
|
"step": 3000
|
|
},
|
|
{
|
|
"epoch": 1.46484375,
|
|
"eval_cer": 21.163636363636364,
|
|
"eval_loss": 0.4456084370613098,
|
|
"eval_normalized_cer": 16.6872174270448,
|
|
"eval_runtime": 143.2684,
|
|
"eval_samples_per_second": 0.893,
|
|
"eval_steps_per_second": 0.056,
|
|
"step": 3000
|
|
},
|
|
{
|
|
"epoch": 1.47705078125,
|
|
"grad_norm": 6.303469657897949,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5534,
|
|
"step": 3025
|
|
},
|
|
{
|
|
"epoch": 1.4892578125,
|
|
"grad_norm": 3.4602832794189453,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5466,
|
|
"step": 3050
|
|
},
|
|
{
|
|
"epoch": 1.50146484375,
|
|
"grad_norm": 5.2441205978393555,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4971,
|
|
"step": 3075
|
|
},
|
|
{
|
|
"epoch": 1.513671875,
|
|
"grad_norm": 4.630413055419922,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4965,
|
|
"step": 3100
|
|
},
|
|
{
|
|
"epoch": 1.52587890625,
|
|
"grad_norm": 4.705837726593018,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4555,
|
|
"step": 3125
|
|
},
|
|
{
|
|
"epoch": 1.5380859375,
|
|
"grad_norm": 5.804379463195801,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5324,
|
|
"step": 3150
|
|
},
|
|
{
|
|
"epoch": 1.55029296875,
|
|
"grad_norm": 4.1124701499938965,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4799,
|
|
"step": 3175
|
|
},
|
|
{
|
|
"epoch": 1.5625,
|
|
"grad_norm": 2.9505362510681152,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4615,
|
|
"step": 3200
|
|
},
|
|
{
|
|
"epoch": 1.57470703125,
|
|
"grad_norm": 4.232094764709473,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5032,
|
|
"step": 3225
|
|
},
|
|
{
|
|
"epoch": 1.5869140625,
|
|
"grad_norm": 3.584272861480713,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.429,
|
|
"step": 3250
|
|
},
|
|
{
|
|
"epoch": 1.59912109375,
|
|
"grad_norm": 4.316075801849365,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5184,
|
|
"step": 3275
|
|
},
|
|
{
|
|
"epoch": 1.611328125,
|
|
"grad_norm": 2.490178108215332,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5072,
|
|
"step": 3300
|
|
},
|
|
{
|
|
"epoch": 1.62353515625,
|
|
"grad_norm": 4.204127311706543,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4874,
|
|
"step": 3325
|
|
},
|
|
{
|
|
"epoch": 1.6357421875,
|
|
"grad_norm": 3.575812339782715,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4741,
|
|
"step": 3350
|
|
},
|
|
{
|
|
"epoch": 1.64794921875,
|
|
"grad_norm": 3.3744544982910156,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4931,
|
|
"step": 3375
|
|
},
|
|
{
|
|
"epoch": 1.66015625,
|
|
"grad_norm": 3.7953217029571533,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.434,
|
|
"step": 3400
|
|
},
|
|
{
|
|
"epoch": 1.67236328125,
|
|
"grad_norm": 3.54185152053833,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.449,
|
|
"step": 3425
|
|
},
|
|
{
|
|
"epoch": 1.6845703125,
|
|
"grad_norm": 3.5956945419311523,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4836,
|
|
"step": 3450
|
|
},
|
|
{
|
|
"epoch": 1.69677734375,
|
|
"grad_norm": 3.323965072631836,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5173,
|
|
"step": 3475
|
|
},
|
|
{
|
|
"epoch": 1.708984375,
|
|
"grad_norm": 3.515545606613159,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4816,
|
|
"step": 3500
|
|
},
|
|
{
|
|
"epoch": 1.708984375,
|
|
"eval_cer": 22.472727272727273,
|
|
"eval_loss": 0.44189023971557617,
|
|
"eval_normalized_cer": 18.12577065351418,
|
|
"eval_runtime": 125.2883,
|
|
"eval_samples_per_second": 1.022,
|
|
"eval_steps_per_second": 0.064,
|
|
"step": 3500
|
|
},
|
|
{
|
|
"epoch": 1.72119140625,
|
|
"grad_norm": 2.902378797531128,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4416,
|
|
"step": 3525
|
|
},
|
|
{
|
|
"epoch": 1.7333984375,
|
|
"grad_norm": 5.24100399017334,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4477,
|
|
"step": 3550
|
|
},
|
|
{
|
|
"epoch": 1.74560546875,
|
|
"grad_norm": 2.840927839279175,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4683,
|
|
"step": 3575
|
|
},
|
|
{
|
|
"epoch": 1.7578125,
|
|
"grad_norm": 3.2699358463287354,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4806,
|
|
"step": 3600
|
|
},
|
|
{
|
|
"epoch": 1.77001953125,
|
|
"grad_norm": 3.1949098110198975,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4652,
|
|
"step": 3625
|
|
},
|
|
{
|
|
"epoch": 1.7822265625,
|
|
"grad_norm": 2.1296651363372803,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.473,
|
|
"step": 3650
|
|
},
|
|
{
|
|
"epoch": 1.79443359375,
|
|
"grad_norm": 2.817379951477051,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5153,
|
|
"step": 3675
|
|
},
|
|
{
|
|
"epoch": 1.806640625,
|
|
"grad_norm": 3.289232015609741,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4662,
|
|
"step": 3700
|
|
},
|
|
{
|
|
"epoch": 1.81884765625,
|
|
"grad_norm": 5.639113426208496,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4773,
|
|
"step": 3725
|
|
},
|
|
{
|
|
"epoch": 1.8310546875,
|
|
"grad_norm": 3.2733285427093506,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4468,
|
|
"step": 3750
|
|
},
|
|
{
|
|
"epoch": 1.84326171875,
|
|
"grad_norm": 3.131972312927246,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5031,
|
|
"step": 3775
|
|
},
|
|
{
|
|
"epoch": 1.85546875,
|
|
"grad_norm": 3.908076524734497,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5018,
|
|
"step": 3800
|
|
},
|
|
{
|
|
"epoch": 1.86767578125,
|
|
"grad_norm": 3.57212233543396,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4862,
|
|
"step": 3825
|
|
},
|
|
{
|
|
"epoch": 1.8798828125,
|
|
"grad_norm": 3.8183035850524902,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4953,
|
|
"step": 3850
|
|
},
|
|
{
|
|
"epoch": 1.89208984375,
|
|
"grad_norm": 3.292840003967285,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4726,
|
|
"step": 3875
|
|
},
|
|
{
|
|
"epoch": 1.904296875,
|
|
"grad_norm": 2.816706418991089,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4925,
|
|
"step": 3900
|
|
},
|
|
{
|
|
"epoch": 1.91650390625,
|
|
"grad_norm": 4.149738311767578,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4868,
|
|
"step": 3925
|
|
},
|
|
{
|
|
"epoch": 1.9287109375,
|
|
"grad_norm": 3.7483160495758057,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4371,
|
|
"step": 3950
|
|
},
|
|
{
|
|
"epoch": 1.94091796875,
|
|
"grad_norm": 3.3911263942718506,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4279,
|
|
"step": 3975
|
|
},
|
|
{
|
|
"epoch": 1.953125,
|
|
"grad_norm": 4.811086654663086,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4482,
|
|
"step": 4000
|
|
},
|
|
{
|
|
"epoch": 1.953125,
|
|
"eval_cer": 23.054545454545455,
|
|
"eval_loss": 0.43437713384628296,
|
|
"eval_normalized_cer": 18.66009042334566,
|
|
"eval_runtime": 152.761,
|
|
"eval_samples_per_second": 0.838,
|
|
"eval_steps_per_second": 0.052,
|
|
"step": 4000
|
|
},
|
|
{
|
|
"epoch": 1.96533203125,
|
|
"grad_norm": 2.9040095806121826,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.448,
|
|
"step": 4025
|
|
},
|
|
{
|
|
"epoch": 1.9775390625,
|
|
"grad_norm": 3.5257699489593506,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4533,
|
|
"step": 4050
|
|
},
|
|
{
|
|
"epoch": 1.98974609375,
|
|
"grad_norm": 3.5330731868743896,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4978,
|
|
"step": 4075
|
|
},
|
|
{
|
|
"epoch": 2.001953125,
|
|
"grad_norm": 3.301609992980957,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4408,
|
|
"step": 4100
|
|
},
|
|
{
|
|
"epoch": 2.01416015625,
|
|
"grad_norm": 2.8477797508239746,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4858,
|
|
"step": 4125
|
|
},
|
|
{
|
|
"epoch": 2.0263671875,
|
|
"grad_norm": 3.097527503967285,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.455,
|
|
"step": 4150
|
|
},
|
|
{
|
|
"epoch": 2.03857421875,
|
|
"grad_norm": 3.6209845542907715,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4461,
|
|
"step": 4175
|
|
},
|
|
{
|
|
"epoch": 2.05078125,
|
|
"grad_norm": 3.182161331176758,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4472,
|
|
"step": 4200
|
|
},
|
|
{
|
|
"epoch": 2.06298828125,
|
|
"grad_norm": 4.14016056060791,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4556,
|
|
"step": 4225
|
|
},
|
|
{
|
|
"epoch": 2.0751953125,
|
|
"grad_norm": 3.5136237144470215,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4817,
|
|
"step": 4250
|
|
},
|
|
{
|
|
"epoch": 2.08740234375,
|
|
"grad_norm": 4.494429111480713,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4399,
|
|
"step": 4275
|
|
},
|
|
{
|
|
"epoch": 2.099609375,
|
|
"grad_norm": 4.786192417144775,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4755,
|
|
"step": 4300
|
|
},
|
|
{
|
|
"epoch": 2.11181640625,
|
|
"grad_norm": 4.873692512512207,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4622,
|
|
"step": 4325
|
|
},
|
|
{
|
|
"epoch": 2.1240234375,
|
|
"grad_norm": 3.9967992305755615,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4796,
|
|
"step": 4350
|
|
},
|
|
{
|
|
"epoch": 2.13623046875,
|
|
"grad_norm": 4.429341793060303,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4564,
|
|
"step": 4375
|
|
},
|
|
{
|
|
"epoch": 2.1484375,
|
|
"grad_norm": 3.952096939086914,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4753,
|
|
"step": 4400
|
|
},
|
|
{
|
|
"epoch": 2.16064453125,
|
|
"grad_norm": 4.080462455749512,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4625,
|
|
"step": 4425
|
|
},
|
|
{
|
|
"epoch": 2.1728515625,
|
|
"grad_norm": 3.9949817657470703,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4935,
|
|
"step": 4450
|
|
},
|
|
{
|
|
"epoch": 2.18505859375,
|
|
"grad_norm": 3.152660846710205,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5341,
|
|
"step": 4475
|
|
},
|
|
{
|
|
"epoch": 2.197265625,
|
|
"grad_norm": 3.485163927078247,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4524,
|
|
"step": 4500
|
|
},
|
|
{
|
|
"epoch": 2.197265625,
|
|
"eval_cer": 22.98181818181818,
|
|
"eval_loss": 0.4256882667541504,
|
|
"eval_normalized_cer": 18.29017673653925,
|
|
"eval_runtime": 133.4464,
|
|
"eval_samples_per_second": 0.959,
|
|
"eval_steps_per_second": 0.06,
|
|
"step": 4500
|
|
},
|
|
{
|
|
"epoch": 2.20947265625,
|
|
"grad_norm": 3.786696195602417,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4827,
|
|
"step": 4525
|
|
},
|
|
{
|
|
"epoch": 2.2216796875,
|
|
"grad_norm": 3.4594826698303223,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.505,
|
|
"step": 4550
|
|
},
|
|
{
|
|
"epoch": 2.23388671875,
|
|
"grad_norm": 3.2194244861602783,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.465,
|
|
"step": 4575
|
|
},
|
|
{
|
|
"epoch": 2.24609375,
|
|
"grad_norm": 3.4851391315460205,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4484,
|
|
"step": 4600
|
|
},
|
|
{
|
|
"epoch": 2.25830078125,
|
|
"grad_norm": 4.488097667694092,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4357,
|
|
"step": 4625
|
|
},
|
|
{
|
|
"epoch": 2.2705078125,
|
|
"grad_norm": 3.071812868118286,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4411,
|
|
"step": 4650
|
|
},
|
|
{
|
|
"epoch": 2.28271484375,
|
|
"grad_norm": 4.025106430053711,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4436,
|
|
"step": 4675
|
|
},
|
|
{
|
|
"epoch": 2.294921875,
|
|
"grad_norm": 3.1780993938446045,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4645,
|
|
"step": 4700
|
|
},
|
|
{
|
|
"epoch": 2.30712890625,
|
|
"grad_norm": 4.2200446128845215,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5211,
|
|
"step": 4725
|
|
},
|
|
{
|
|
"epoch": 2.3193359375,
|
|
"grad_norm": 3.622480869293213,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4763,
|
|
"step": 4750
|
|
},
|
|
{
|
|
"epoch": 2.33154296875,
|
|
"grad_norm": 3.0763843059539795,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4876,
|
|
"step": 4775
|
|
},
|
|
{
|
|
"epoch": 2.34375,
|
|
"grad_norm": 3.9623701572418213,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4683,
|
|
"step": 4800
|
|
},
|
|
{
|
|
"epoch": 2.35595703125,
|
|
"grad_norm": 4.627608776092529,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.426,
|
|
"step": 4825
|
|
},
|
|
{
|
|
"epoch": 2.3681640625,
|
|
"grad_norm": 4.717302322387695,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4368,
|
|
"step": 4850
|
|
},
|
|
{
|
|
"epoch": 2.38037109375,
|
|
"grad_norm": 4.21370792388916,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5348,
|
|
"step": 4875
|
|
},
|
|
{
|
|
"epoch": 2.392578125,
|
|
"grad_norm": 4.240349292755127,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4766,
|
|
"step": 4900
|
|
},
|
|
{
|
|
"epoch": 2.40478515625,
|
|
"grad_norm": 4.513136386871338,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.421,
|
|
"step": 4925
|
|
},
|
|
{
|
|
"epoch": 2.4169921875,
|
|
"grad_norm": 3.5849685668945312,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4599,
|
|
"step": 4950
|
|
},
|
|
{
|
|
"epoch": 2.42919921875,
|
|
"grad_norm": 3.148627519607544,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4812,
|
|
"step": 4975
|
|
},
|
|
{
|
|
"epoch": 2.44140625,
|
|
"grad_norm": 3.3476321697235107,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5048,
|
|
"step": 5000
|
|
},
|
|
{
|
|
"epoch": 2.44140625,
|
|
"eval_cer": 23.78181818181818,
|
|
"eval_loss": 0.42172056436538696,
|
|
"eval_normalized_cer": 18.7422934648582,
|
|
"eval_runtime": 133.6191,
|
|
"eval_samples_per_second": 0.958,
|
|
"eval_steps_per_second": 0.06,
|
|
"step": 5000
|
|
},
|
|
{
|
|
"epoch": 2.45361328125,
|
|
"grad_norm": 3.0917651653289795,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4876,
|
|
"step": 5025
|
|
},
|
|
{
|
|
"epoch": 2.4658203125,
|
|
"grad_norm": 4.9287309646606445,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4631,
|
|
"step": 5050
|
|
},
|
|
{
|
|
"epoch": 2.47802734375,
|
|
"grad_norm": 3.54841947555542,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4752,
|
|
"step": 5075
|
|
},
|
|
{
|
|
"epoch": 2.490234375,
|
|
"grad_norm": 3.062551975250244,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4232,
|
|
"step": 5100
|
|
},
|
|
{
|
|
"epoch": 2.50244140625,
|
|
"grad_norm": 5.846357822418213,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4819,
|
|
"step": 5125
|
|
},
|
|
{
|
|
"epoch": 2.5146484375,
|
|
"grad_norm": 5.39830207824707,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4574,
|
|
"step": 5150
|
|
},
|
|
{
|
|
"epoch": 2.52685546875,
|
|
"grad_norm": 3.695359468460083,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5137,
|
|
"step": 5175
|
|
},
|
|
{
|
|
"epoch": 2.5390625,
|
|
"grad_norm": 3.8433547019958496,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4097,
|
|
"step": 5200
|
|
},
|
|
{
|
|
"epoch": 2.55126953125,
|
|
"grad_norm": 3.887317419052124,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.481,
|
|
"step": 5225
|
|
},
|
|
{
|
|
"epoch": 2.5634765625,
|
|
"grad_norm": 3.9261627197265625,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4513,
|
|
"step": 5250
|
|
},
|
|
{
|
|
"epoch": 2.57568359375,
|
|
"grad_norm": 2.480574369430542,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4372,
|
|
"step": 5275
|
|
},
|
|
{
|
|
"epoch": 2.587890625,
|
|
"grad_norm": 4.399667263031006,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4252,
|
|
"step": 5300
|
|
},
|
|
{
|
|
"epoch": 2.60009765625,
|
|
"grad_norm": 3.6141607761383057,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4369,
|
|
"step": 5325
|
|
},
|
|
{
|
|
"epoch": 2.6123046875,
|
|
"grad_norm": 2.9269521236419678,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4987,
|
|
"step": 5350
|
|
},
|
|
{
|
|
"epoch": 2.62451171875,
|
|
"grad_norm": 3.667206048965454,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.495,
|
|
"step": 5375
|
|
},
|
|
{
|
|
"epoch": 2.63671875,
|
|
"grad_norm": 5.493015289306641,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4439,
|
|
"step": 5400
|
|
},
|
|
{
|
|
"epoch": 2.64892578125,
|
|
"grad_norm": 3.328899383544922,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4539,
|
|
"step": 5425
|
|
},
|
|
{
|
|
"epoch": 2.6611328125,
|
|
"grad_norm": 2.529545783996582,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4698,
|
|
"step": 5450
|
|
},
|
|
{
|
|
"epoch": 2.67333984375,
|
|
"grad_norm": 3.669677495956421,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3939,
|
|
"step": 5475
|
|
},
|
|
{
|
|
"epoch": 2.685546875,
|
|
"grad_norm": 3.070936441421509,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4721,
|
|
"step": 5500
|
|
},
|
|
{
|
|
"epoch": 2.685546875,
|
|
"eval_cer": 23.01818181818182,
|
|
"eval_loss": 0.4226231276988983,
|
|
"eval_normalized_cer": 18.33127825729552,
|
|
"eval_runtime": 126.3723,
|
|
"eval_samples_per_second": 1.013,
|
|
"eval_steps_per_second": 0.063,
|
|
"step": 5500
|
|
},
|
|
{
|
|
"epoch": 2.69775390625,
|
|
"grad_norm": 4.265174865722656,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5078,
|
|
"step": 5525
|
|
},
|
|
{
|
|
"epoch": 2.7099609375,
|
|
"grad_norm": 5.314130783081055,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.452,
|
|
"step": 5550
|
|
},
|
|
{
|
|
"epoch": 2.72216796875,
|
|
"grad_norm": 4.048317909240723,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4419,
|
|
"step": 5575
|
|
},
|
|
{
|
|
"epoch": 2.734375,
|
|
"grad_norm": 3.8373677730560303,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4236,
|
|
"step": 5600
|
|
},
|
|
{
|
|
"epoch": 2.74658203125,
|
|
"grad_norm": 3.880476713180542,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4224,
|
|
"step": 5625
|
|
},
|
|
{
|
|
"epoch": 2.7587890625,
|
|
"grad_norm": 4.03834867477417,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4762,
|
|
"step": 5650
|
|
},
|
|
{
|
|
"epoch": 2.77099609375,
|
|
"grad_norm": 3.321204900741577,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4823,
|
|
"step": 5675
|
|
},
|
|
{
|
|
"epoch": 2.783203125,
|
|
"grad_norm": 3.013662099838257,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4839,
|
|
"step": 5700
|
|
},
|
|
{
|
|
"epoch": 2.79541015625,
|
|
"grad_norm": 4.0579705238342285,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4754,
|
|
"step": 5725
|
|
},
|
|
{
|
|
"epoch": 2.8076171875,
|
|
"grad_norm": 3.929385185241699,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4123,
|
|
"step": 5750
|
|
},
|
|
{
|
|
"epoch": 2.81982421875,
|
|
"grad_norm": 4.071752071380615,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4989,
|
|
"step": 5775
|
|
},
|
|
{
|
|
"epoch": 2.83203125,
|
|
"grad_norm": 3.5751779079437256,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4507,
|
|
"step": 5800
|
|
},
|
|
{
|
|
"epoch": 2.84423828125,
|
|
"grad_norm": 3.4296460151672363,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.522,
|
|
"step": 5825
|
|
},
|
|
{
|
|
"epoch": 2.8564453125,
|
|
"grad_norm": 2.707711696624756,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4347,
|
|
"step": 5850
|
|
},
|
|
{
|
|
"epoch": 2.86865234375,
|
|
"grad_norm": 2.0897769927978516,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4971,
|
|
"step": 5875
|
|
},
|
|
{
|
|
"epoch": 2.880859375,
|
|
"grad_norm": 2.9398937225341797,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4507,
|
|
"step": 5900
|
|
},
|
|
{
|
|
"epoch": 2.89306640625,
|
|
"grad_norm": 3.4962351322174072,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4662,
|
|
"step": 5925
|
|
},
|
|
{
|
|
"epoch": 2.9052734375,
|
|
"grad_norm": 3.520770311355591,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4784,
|
|
"step": 5950
|
|
},
|
|
{
|
|
"epoch": 2.91748046875,
|
|
"grad_norm": 3.4182958602905273,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4952,
|
|
"step": 5975
|
|
},
|
|
{
|
|
"epoch": 2.9296875,
|
|
"grad_norm": 2.9308042526245117,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4321,
|
|
"step": 6000
|
|
},
|
|
{
|
|
"epoch": 2.9296875,
|
|
"eval_cer": 20.472727272727273,
|
|
"eval_loss": 0.4113434851169586,
|
|
"eval_normalized_cer": 15.577476366625564,
|
|
"eval_runtime": 105.0617,
|
|
"eval_samples_per_second": 1.218,
|
|
"eval_steps_per_second": 0.076,
|
|
"step": 6000
|
|
},
|
|
{
|
|
"epoch": 2.94189453125,
|
|
"grad_norm": 2.3675143718719482,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5268,
|
|
"step": 6025
|
|
},
|
|
{
|
|
"epoch": 2.9541015625,
|
|
"grad_norm": 3.0600383281707764,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4166,
|
|
"step": 6050
|
|
},
|
|
{
|
|
"epoch": 2.96630859375,
|
|
"grad_norm": 3.9981579780578613,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4758,
|
|
"step": 6075
|
|
},
|
|
{
|
|
"epoch": 2.978515625,
|
|
"grad_norm": 4.047635555267334,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4376,
|
|
"step": 6100
|
|
},
|
|
{
|
|
"epoch": 2.99072265625,
|
|
"grad_norm": 3.3930447101593018,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4391,
|
|
"step": 6125
|
|
},
|
|
{
|
|
"epoch": 3.0029296875,
|
|
"grad_norm": 2.583280563354492,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3899,
|
|
"step": 6150
|
|
},
|
|
{
|
|
"epoch": 3.01513671875,
|
|
"grad_norm": 5.198780059814453,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4538,
|
|
"step": 6175
|
|
},
|
|
{
|
|
"epoch": 3.02734375,
|
|
"grad_norm": 5.145614147186279,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5109,
|
|
"step": 6200
|
|
},
|
|
{
|
|
"epoch": 3.03955078125,
|
|
"grad_norm": 4.1711320877075195,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4124,
|
|
"step": 6225
|
|
},
|
|
{
|
|
"epoch": 3.0517578125,
|
|
"grad_norm": 3.6686923503875732,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4233,
|
|
"step": 6250
|
|
},
|
|
{
|
|
"epoch": 3.06396484375,
|
|
"grad_norm": 3.542102575302124,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5235,
|
|
"step": 6275
|
|
},
|
|
{
|
|
"epoch": 3.076171875,
|
|
"grad_norm": 2.960315227508545,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4599,
|
|
"step": 6300
|
|
},
|
|
{
|
|
"epoch": 3.08837890625,
|
|
"grad_norm": 3.370656728744507,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4668,
|
|
"step": 6325
|
|
},
|
|
{
|
|
"epoch": 3.1005859375,
|
|
"grad_norm": 4.451176643371582,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4846,
|
|
"step": 6350
|
|
},
|
|
{
|
|
"epoch": 3.11279296875,
|
|
"grad_norm": 3.628671646118164,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4636,
|
|
"step": 6375
|
|
},
|
|
{
|
|
"epoch": 3.125,
|
|
"grad_norm": 3.664491653442383,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4843,
|
|
"step": 6400
|
|
},
|
|
{
|
|
"epoch": 3.13720703125,
|
|
"grad_norm": 3.1877737045288086,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5361,
|
|
"step": 6425
|
|
},
|
|
{
|
|
"epoch": 3.1494140625,
|
|
"grad_norm": 3.7835752964019775,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4808,
|
|
"step": 6450
|
|
},
|
|
{
|
|
"epoch": 3.16162109375,
|
|
"grad_norm": 3.494187355041504,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4401,
|
|
"step": 6475
|
|
},
|
|
{
|
|
"epoch": 3.173828125,
|
|
"grad_norm": 2.837970495223999,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4454,
|
|
"step": 6500
|
|
},
|
|
{
|
|
"epoch": 3.173828125,
|
|
"eval_cer": 21.709090909090907,
|
|
"eval_loss": 0.40472567081451416,
|
|
"eval_normalized_cer": 16.152897657213316,
|
|
"eval_runtime": 133.6937,
|
|
"eval_samples_per_second": 0.957,
|
|
"eval_steps_per_second": 0.06,
|
|
"step": 6500
|
|
},
|
|
{
|
|
"epoch": 3.18603515625,
|
|
"grad_norm": 3.5152645111083984,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4798,
|
|
"step": 6525
|
|
},
|
|
{
|
|
"epoch": 3.1982421875,
|
|
"grad_norm": 3.201646327972412,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4576,
|
|
"step": 6550
|
|
},
|
|
{
|
|
"epoch": 3.21044921875,
|
|
"grad_norm": 3.704028606414795,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4428,
|
|
"step": 6575
|
|
},
|
|
{
|
|
"epoch": 3.22265625,
|
|
"grad_norm": 4.099301815032959,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4787,
|
|
"step": 6600
|
|
},
|
|
{
|
|
"epoch": 3.23486328125,
|
|
"grad_norm": 3.0543127059936523,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4761,
|
|
"step": 6625
|
|
},
|
|
{
|
|
"epoch": 3.2470703125,
|
|
"grad_norm": 3.4532554149627686,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4559,
|
|
"step": 6650
|
|
},
|
|
{
|
|
"epoch": 3.25927734375,
|
|
"grad_norm": 2.8139491081237793,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4705,
|
|
"step": 6675
|
|
},
|
|
{
|
|
"epoch": 3.271484375,
|
|
"grad_norm": 3.291689157485962,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4738,
|
|
"step": 6700
|
|
},
|
|
{
|
|
"epoch": 3.28369140625,
|
|
"grad_norm": 3.0359091758728027,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3979,
|
|
"step": 6725
|
|
},
|
|
{
|
|
"epoch": 3.2958984375,
|
|
"grad_norm": 4.325143337249756,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4309,
|
|
"step": 6750
|
|
},
|
|
{
|
|
"epoch": 3.30810546875,
|
|
"grad_norm": 3.864635705947876,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4905,
|
|
"step": 6775
|
|
},
|
|
{
|
|
"epoch": 3.3203125,
|
|
"grad_norm": 3.659311532974243,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5003,
|
|
"step": 6800
|
|
},
|
|
{
|
|
"epoch": 3.33251953125,
|
|
"grad_norm": 3.6954805850982666,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4197,
|
|
"step": 6825
|
|
},
|
|
{
|
|
"epoch": 3.3447265625,
|
|
"grad_norm": 2.9357662200927734,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.425,
|
|
"step": 6850
|
|
},
|
|
{
|
|
"epoch": 3.35693359375,
|
|
"grad_norm": 2.9969289302825928,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4883,
|
|
"step": 6875
|
|
},
|
|
{
|
|
"epoch": 3.369140625,
|
|
"grad_norm": 3.333348035812378,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4687,
|
|
"step": 6900
|
|
},
|
|
{
|
|
"epoch": 3.38134765625,
|
|
"grad_norm": 4.444482803344727,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4324,
|
|
"step": 6925
|
|
},
|
|
{
|
|
"epoch": 3.3935546875,
|
|
"grad_norm": 2.3350095748901367,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4476,
|
|
"step": 6950
|
|
},
|
|
{
|
|
"epoch": 3.40576171875,
|
|
"grad_norm": 3.4752862453460693,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4601,
|
|
"step": 6975
|
|
},
|
|
{
|
|
"epoch": 3.41796875,
|
|
"grad_norm": 3.1359448432922363,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4894,
|
|
"step": 7000
|
|
},
|
|
{
|
|
"epoch": 3.41796875,
|
|
"eval_cer": 20.363636363636363,
|
|
"eval_loss": 0.41460537910461426,
|
|
"eval_normalized_cer": 15.988491574188245,
|
|
"eval_runtime": 131.3226,
|
|
"eval_samples_per_second": 0.975,
|
|
"eval_steps_per_second": 0.061,
|
|
"step": 7000
|
|
},
|
|
{
|
|
"epoch": 3.43017578125,
|
|
"grad_norm": 3.0270822048187256,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4288,
|
|
"step": 7025
|
|
},
|
|
{
|
|
"epoch": 3.4423828125,
|
|
"grad_norm": 3.487501382827759,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.473,
|
|
"step": 7050
|
|
},
|
|
{
|
|
"epoch": 3.45458984375,
|
|
"grad_norm": 3.1691272258758545,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4516,
|
|
"step": 7075
|
|
},
|
|
{
|
|
"epoch": 3.466796875,
|
|
"grad_norm": 3.07665753364563,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.429,
|
|
"step": 7100
|
|
},
|
|
{
|
|
"epoch": 3.47900390625,
|
|
"grad_norm": 3.697643756866455,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.479,
|
|
"step": 7125
|
|
},
|
|
{
|
|
"epoch": 3.4912109375,
|
|
"grad_norm": 3.787980556488037,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4837,
|
|
"step": 7150
|
|
},
|
|
{
|
|
"epoch": 3.50341796875,
|
|
"grad_norm": 3.792851686477661,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4339,
|
|
"step": 7175
|
|
},
|
|
{
|
|
"epoch": 3.515625,
|
|
"grad_norm": 4.371828079223633,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4364,
|
|
"step": 7200
|
|
},
|
|
{
|
|
"epoch": 3.52783203125,
|
|
"grad_norm": 2.8231394290924072,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4776,
|
|
"step": 7225
|
|
},
|
|
{
|
|
"epoch": 3.5400390625,
|
|
"grad_norm": 3.1332738399505615,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4158,
|
|
"step": 7250
|
|
},
|
|
{
|
|
"epoch": 3.55224609375,
|
|
"grad_norm": 3.6352245807647705,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4797,
|
|
"step": 7275
|
|
},
|
|
{
|
|
"epoch": 3.564453125,
|
|
"grad_norm": 4.453933238983154,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4355,
|
|
"step": 7300
|
|
},
|
|
{
|
|
"epoch": 3.57666015625,
|
|
"grad_norm": 2.9594037532806396,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4696,
|
|
"step": 7325
|
|
},
|
|
{
|
|
"epoch": 3.5888671875,
|
|
"grad_norm": 5.014936447143555,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4622,
|
|
"step": 7350
|
|
},
|
|
{
|
|
"epoch": 3.60107421875,
|
|
"grad_norm": 3.2860615253448486,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4536,
|
|
"step": 7375
|
|
},
|
|
{
|
|
"epoch": 3.61328125,
|
|
"grad_norm": 5.162105083465576,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4456,
|
|
"step": 7400
|
|
},
|
|
{
|
|
"epoch": 3.62548828125,
|
|
"grad_norm": 3.706555128097534,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4843,
|
|
"step": 7425
|
|
},
|
|
{
|
|
"epoch": 3.6376953125,
|
|
"grad_norm": 2.490443706512451,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4284,
|
|
"step": 7450
|
|
},
|
|
{
|
|
"epoch": 3.64990234375,
|
|
"grad_norm": 3.4704225063323975,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4562,
|
|
"step": 7475
|
|
},
|
|
{
|
|
"epoch": 3.662109375,
|
|
"grad_norm": 3.8659839630126953,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4603,
|
|
"step": 7500
|
|
},
|
|
{
|
|
"epoch": 3.662109375,
|
|
"eval_cer": 23.78181818181818,
|
|
"eval_loss": 0.4173641800880432,
|
|
"eval_normalized_cer": 19.35881627620222,
|
|
"eval_runtime": 134.5888,
|
|
"eval_samples_per_second": 0.951,
|
|
"eval_steps_per_second": 0.059,
|
|
"step": 7500
|
|
},
|
|
{
|
|
"epoch": 3.67431640625,
|
|
"grad_norm": 3.280311346054077,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4501,
|
|
"step": 7525
|
|
},
|
|
{
|
|
"epoch": 3.6865234375,
|
|
"grad_norm": 4.28645133972168,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4437,
|
|
"step": 7550
|
|
},
|
|
{
|
|
"epoch": 3.69873046875,
|
|
"grad_norm": 3.8637635707855225,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4302,
|
|
"step": 7575
|
|
},
|
|
{
|
|
"epoch": 3.7109375,
|
|
"grad_norm": 3.35193133354187,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4298,
|
|
"step": 7600
|
|
},
|
|
{
|
|
"epoch": 3.72314453125,
|
|
"grad_norm": 3.156805992126465,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4294,
|
|
"step": 7625
|
|
},
|
|
{
|
|
"epoch": 3.7353515625,
|
|
"grad_norm": 3.2977190017700195,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4471,
|
|
"step": 7650
|
|
},
|
|
{
|
|
"epoch": 3.74755859375,
|
|
"grad_norm": 3.3117992877960205,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4024,
|
|
"step": 7675
|
|
},
|
|
{
|
|
"epoch": 3.759765625,
|
|
"grad_norm": 4.135869979858398,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.431,
|
|
"step": 7700
|
|
},
|
|
{
|
|
"epoch": 3.77197265625,
|
|
"grad_norm": 3.1641712188720703,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4494,
|
|
"step": 7725
|
|
},
|
|
{
|
|
"epoch": 3.7841796875,
|
|
"grad_norm": 3.1952223777770996,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4415,
|
|
"step": 7750
|
|
},
|
|
{
|
|
"epoch": 3.79638671875,
|
|
"grad_norm": 3.2214698791503906,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4409,
|
|
"step": 7775
|
|
},
|
|
{
|
|
"epoch": 3.80859375,
|
|
"grad_norm": 3.14152193069458,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4407,
|
|
"step": 7800
|
|
},
|
|
{
|
|
"epoch": 3.82080078125,
|
|
"grad_norm": 3.8332841396331787,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4438,
|
|
"step": 7825
|
|
},
|
|
{
|
|
"epoch": 3.8330078125,
|
|
"grad_norm": 3.2300031185150146,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4377,
|
|
"step": 7850
|
|
},
|
|
{
|
|
"epoch": 3.84521484375,
|
|
"grad_norm": 3.9044997692108154,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4376,
|
|
"step": 7875
|
|
},
|
|
{
|
|
"epoch": 3.857421875,
|
|
"grad_norm": 2.9638853073120117,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4605,
|
|
"step": 7900
|
|
},
|
|
{
|
|
"epoch": 3.86962890625,
|
|
"grad_norm": 3.6924691200256348,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4632,
|
|
"step": 7925
|
|
},
|
|
{
|
|
"epoch": 3.8818359375,
|
|
"grad_norm": 2.571397542953491,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.439,
|
|
"step": 7950
|
|
},
|
|
{
|
|
"epoch": 3.89404296875,
|
|
"grad_norm": 2.900402069091797,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4431,
|
|
"step": 7975
|
|
},
|
|
{
|
|
"epoch": 3.90625,
|
|
"grad_norm": 3.216660737991333,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4479,
|
|
"step": 8000
|
|
},
|
|
{
|
|
"epoch": 3.90625,
|
|
"eval_cer": 21.527272727272727,
|
|
"eval_loss": 0.40407997369766235,
|
|
"eval_normalized_cer": 16.974928072338678,
|
|
"eval_runtime": 135.4416,
|
|
"eval_samples_per_second": 0.945,
|
|
"eval_steps_per_second": 0.059,
|
|
"step": 8000
|
|
},
|
|
{
|
|
"epoch": 3.91845703125,
|
|
"grad_norm": 3.2848970890045166,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4152,
|
|
"step": 8025
|
|
},
|
|
{
|
|
"epoch": 3.9306640625,
|
|
"grad_norm": 2.901078939437866,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4292,
|
|
"step": 8050
|
|
},
|
|
{
|
|
"epoch": 3.94287109375,
|
|
"grad_norm": 3.3636679649353027,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4624,
|
|
"step": 8075
|
|
},
|
|
{
|
|
"epoch": 3.955078125,
|
|
"grad_norm": 4.242199420928955,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4427,
|
|
"step": 8100
|
|
},
|
|
{
|
|
"epoch": 3.96728515625,
|
|
"grad_norm": 3.422555446624756,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4161,
|
|
"step": 8125
|
|
},
|
|
{
|
|
"epoch": 3.9794921875,
|
|
"grad_norm": 2.359015703201294,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.422,
|
|
"step": 8150
|
|
},
|
|
{
|
|
"epoch": 3.99169921875,
|
|
"grad_norm": 3.1188254356384277,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4346,
|
|
"step": 8175
|
|
},
|
|
{
|
|
"epoch": 4.00390625,
|
|
"grad_norm": 3.1094157695770264,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4511,
|
|
"step": 8200
|
|
},
|
|
{
|
|
"epoch": 4.01611328125,
|
|
"grad_norm": 3.269327163696289,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4548,
|
|
"step": 8225
|
|
},
|
|
{
|
|
"epoch": 4.0283203125,
|
|
"grad_norm": 2.589052677154541,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4267,
|
|
"step": 8250
|
|
},
|
|
{
|
|
"epoch": 4.04052734375,
|
|
"grad_norm": 3.032090187072754,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4194,
|
|
"step": 8275
|
|
},
|
|
{
|
|
"epoch": 4.052734375,
|
|
"grad_norm": 3.1409432888031006,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4423,
|
|
"step": 8300
|
|
},
|
|
{
|
|
"epoch": 4.06494140625,
|
|
"grad_norm": 4.070183753967285,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4171,
|
|
"step": 8325
|
|
},
|
|
{
|
|
"epoch": 4.0771484375,
|
|
"grad_norm": 4.6292619705200195,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4689,
|
|
"step": 8350
|
|
},
|
|
{
|
|
"epoch": 4.08935546875,
|
|
"grad_norm": 3.5778212547302246,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4341,
|
|
"step": 8375
|
|
},
|
|
{
|
|
"epoch": 4.1015625,
|
|
"grad_norm": 2.5463714599609375,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4193,
|
|
"step": 8400
|
|
},
|
|
{
|
|
"epoch": 4.11376953125,
|
|
"grad_norm": 2.498852491378784,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4656,
|
|
"step": 8425
|
|
},
|
|
{
|
|
"epoch": 4.1259765625,
|
|
"grad_norm": 2.5242180824279785,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4491,
|
|
"step": 8450
|
|
},
|
|
{
|
|
"epoch": 4.13818359375,
|
|
"grad_norm": 4.114852428436279,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4202,
|
|
"step": 8475
|
|
},
|
|
{
|
|
"epoch": 4.150390625,
|
|
"grad_norm": 2.866490125656128,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4523,
|
|
"step": 8500
|
|
},
|
|
{
|
|
"epoch": 4.150390625,
|
|
"eval_cer": 19.70909090909091,
|
|
"eval_loss": 0.40719008445739746,
|
|
"eval_normalized_cer": 15.371968762844226,
|
|
"eval_runtime": 136.9983,
|
|
"eval_samples_per_second": 0.934,
|
|
"eval_steps_per_second": 0.058,
|
|
"step": 8500
|
|
},
|
|
{
|
|
"epoch": 4.16259765625,
|
|
"grad_norm": 4.2479448318481445,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4742,
|
|
"step": 8525
|
|
},
|
|
{
|
|
"epoch": 4.1748046875,
|
|
"grad_norm": 2.736863851547241,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4366,
|
|
"step": 8550
|
|
},
|
|
{
|
|
"epoch": 4.18701171875,
|
|
"grad_norm": 3.3167812824249268,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4509,
|
|
"step": 8575
|
|
},
|
|
{
|
|
"epoch": 4.19921875,
|
|
"grad_norm": 2.8058440685272217,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4179,
|
|
"step": 8600
|
|
},
|
|
{
|
|
"epoch": 4.21142578125,
|
|
"grad_norm": 3.10642409324646,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4799,
|
|
"step": 8625
|
|
},
|
|
{
|
|
"epoch": 4.2236328125,
|
|
"grad_norm": 3.9041507244110107,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4788,
|
|
"step": 8650
|
|
},
|
|
{
|
|
"epoch": 4.23583984375,
|
|
"grad_norm": 4.171154022216797,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4534,
|
|
"step": 8675
|
|
},
|
|
{
|
|
"epoch": 4.248046875,
|
|
"grad_norm": 2.6511294841766357,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4395,
|
|
"step": 8700
|
|
},
|
|
{
|
|
"epoch": 4.26025390625,
|
|
"grad_norm": 3.9899098873138428,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4421,
|
|
"step": 8725
|
|
},
|
|
{
|
|
"epoch": 4.2724609375,
|
|
"grad_norm": 2.973851442337036,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4166,
|
|
"step": 8750
|
|
},
|
|
{
|
|
"epoch": 4.28466796875,
|
|
"grad_norm": 3.789973497390747,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4312,
|
|
"step": 8775
|
|
},
|
|
{
|
|
"epoch": 4.296875,
|
|
"grad_norm": 4.157674789428711,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4571,
|
|
"step": 8800
|
|
},
|
|
{
|
|
"epoch": 4.30908203125,
|
|
"grad_norm": 4.191178321838379,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4222,
|
|
"step": 8825
|
|
},
|
|
{
|
|
"epoch": 4.3212890625,
|
|
"grad_norm": 2.899761915206909,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4174,
|
|
"step": 8850
|
|
},
|
|
{
|
|
"epoch": 4.33349609375,
|
|
"grad_norm": 2.9615023136138916,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4684,
|
|
"step": 8875
|
|
},
|
|
{
|
|
"epoch": 4.345703125,
|
|
"grad_norm": 2.9529151916503906,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4079,
|
|
"step": 8900
|
|
},
|
|
{
|
|
"epoch": 4.35791015625,
|
|
"grad_norm": 3.4950997829437256,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4239,
|
|
"step": 8925
|
|
},
|
|
{
|
|
"epoch": 4.3701171875,
|
|
"grad_norm": 4.151655673980713,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4737,
|
|
"step": 8950
|
|
},
|
|
{
|
|
"epoch": 4.38232421875,
|
|
"grad_norm": 3.2207164764404297,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.468,
|
|
"step": 8975
|
|
},
|
|
{
|
|
"epoch": 4.39453125,
|
|
"grad_norm": 4.433598041534424,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4742,
|
|
"step": 9000
|
|
},
|
|
{
|
|
"epoch": 4.39453125,
|
|
"eval_cer": 20.836363636363636,
|
|
"eval_loss": 0.40726813673973083,
|
|
"eval_normalized_cer": 16.563912864775997,
|
|
"eval_runtime": 148.3773,
|
|
"eval_samples_per_second": 0.863,
|
|
"eval_steps_per_second": 0.054,
|
|
"step": 9000
|
|
},
|
|
{
|
|
"epoch": 4.40673828125,
|
|
"grad_norm": 2.8774595260620117,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4355,
|
|
"step": 9025
|
|
},
|
|
{
|
|
"epoch": 4.4189453125,
|
|
"grad_norm": 3.043325185775757,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4267,
|
|
"step": 9050
|
|
},
|
|
{
|
|
"epoch": 4.43115234375,
|
|
"grad_norm": 3.7946414947509766,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4424,
|
|
"step": 9075
|
|
},
|
|
{
|
|
"epoch": 4.443359375,
|
|
"grad_norm": 2.7561936378479004,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4054,
|
|
"step": 9100
|
|
},
|
|
{
|
|
"epoch": 4.45556640625,
|
|
"grad_norm": 3.3554115295410156,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4442,
|
|
"step": 9125
|
|
},
|
|
{
|
|
"epoch": 4.4677734375,
|
|
"grad_norm": 3.1039364337921143,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4289,
|
|
"step": 9150
|
|
},
|
|
{
|
|
"epoch": 4.47998046875,
|
|
"grad_norm": 2.8910741806030273,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4103,
|
|
"step": 9175
|
|
},
|
|
{
|
|
"epoch": 4.4921875,
|
|
"grad_norm": 3.005373001098633,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3944,
|
|
"step": 9200
|
|
},
|
|
{
|
|
"epoch": 4.50439453125,
|
|
"grad_norm": 2.7834503650665283,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4235,
|
|
"step": 9225
|
|
},
|
|
{
|
|
"epoch": 4.5166015625,
|
|
"grad_norm": 3.2540178298950195,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4617,
|
|
"step": 9250
|
|
},
|
|
{
|
|
"epoch": 4.52880859375,
|
|
"grad_norm": 3.1600005626678467,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4445,
|
|
"step": 9275
|
|
},
|
|
{
|
|
"epoch": 4.541015625,
|
|
"grad_norm": 3.318638801574707,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4164,
|
|
"step": 9300
|
|
},
|
|
{
|
|
"epoch": 4.55322265625,
|
|
"grad_norm": 4.483547210693359,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.468,
|
|
"step": 9325
|
|
},
|
|
{
|
|
"epoch": 4.5654296875,
|
|
"grad_norm": 3.946882963180542,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4349,
|
|
"step": 9350
|
|
},
|
|
{
|
|
"epoch": 4.57763671875,
|
|
"grad_norm": 2.6997079849243164,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4483,
|
|
"step": 9375
|
|
},
|
|
{
|
|
"epoch": 4.58984375,
|
|
"grad_norm": 5.617648601531982,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.477,
|
|
"step": 9400
|
|
},
|
|
{
|
|
"epoch": 4.60205078125,
|
|
"grad_norm": 3.01735520362854,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4198,
|
|
"step": 9425
|
|
},
|
|
{
|
|
"epoch": 4.6142578125,
|
|
"grad_norm": 3.301004648208618,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.466,
|
|
"step": 9450
|
|
},
|
|
{
|
|
"epoch": 4.62646484375,
|
|
"grad_norm": 3.2421813011169434,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4366,
|
|
"step": 9475
|
|
},
|
|
{
|
|
"epoch": 4.638671875,
|
|
"grad_norm": 4.813492774963379,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4737,
|
|
"step": 9500
|
|
},
|
|
{
|
|
"epoch": 4.638671875,
|
|
"eval_cer": 21.78181818181818,
|
|
"eval_loss": 0.40351322293281555,
|
|
"eval_normalized_cer": 16.892725030826142,
|
|
"eval_runtime": 136.3812,
|
|
"eval_samples_per_second": 0.939,
|
|
"eval_steps_per_second": 0.059,
|
|
"step": 9500
|
|
},
|
|
{
|
|
"epoch": 4.65087890625,
|
|
"grad_norm": 3.3861124515533447,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.447,
|
|
"step": 9525
|
|
},
|
|
{
|
|
"epoch": 4.6630859375,
|
|
"grad_norm": 3.244462013244629,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4422,
|
|
"step": 9550
|
|
},
|
|
{
|
|
"epoch": 4.67529296875,
|
|
"grad_norm": 3.824782133102417,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4431,
|
|
"step": 9575
|
|
},
|
|
{
|
|
"epoch": 4.6875,
|
|
"grad_norm": 3.746281862258911,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.48,
|
|
"step": 9600
|
|
},
|
|
{
|
|
"epoch": 4.69970703125,
|
|
"grad_norm": 4.126583099365234,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4643,
|
|
"step": 9625
|
|
},
|
|
{
|
|
"epoch": 4.7119140625,
|
|
"grad_norm": 3.3953585624694824,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4826,
|
|
"step": 9650
|
|
},
|
|
{
|
|
"epoch": 4.72412109375,
|
|
"grad_norm": 3.077698230743408,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4138,
|
|
"step": 9675
|
|
},
|
|
{
|
|
"epoch": 4.736328125,
|
|
"grad_norm": 4.628427982330322,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4102,
|
|
"step": 9700
|
|
},
|
|
{
|
|
"epoch": 4.74853515625,
|
|
"grad_norm": 3.1975038051605225,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4374,
|
|
"step": 9725
|
|
},
|
|
{
|
|
"epoch": 4.7607421875,
|
|
"grad_norm": 3.606600522994995,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4095,
|
|
"step": 9750
|
|
},
|
|
{
|
|
"epoch": 4.77294921875,
|
|
"grad_norm": 3.183274507522583,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4759,
|
|
"step": 9775
|
|
},
|
|
{
|
|
"epoch": 4.78515625,
|
|
"grad_norm": 3.52425479888916,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4429,
|
|
"step": 9800
|
|
},
|
|
{
|
|
"epoch": 4.79736328125,
|
|
"grad_norm": 3.2504448890686035,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4271,
|
|
"step": 9825
|
|
},
|
|
{
|
|
"epoch": 4.8095703125,
|
|
"grad_norm": 3.7762293815612793,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4164,
|
|
"step": 9850
|
|
},
|
|
{
|
|
"epoch": 4.82177734375,
|
|
"grad_norm": 2.5560717582702637,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4622,
|
|
"step": 9875
|
|
},
|
|
{
|
|
"epoch": 4.833984375,
|
|
"grad_norm": 2.2996389865875244,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4728,
|
|
"step": 9900
|
|
},
|
|
{
|
|
"epoch": 4.84619140625,
|
|
"grad_norm": 4.10749626159668,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4954,
|
|
"step": 9925
|
|
},
|
|
{
|
|
"epoch": 4.8583984375,
|
|
"grad_norm": 3.2446017265319824,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.485,
|
|
"step": 9950
|
|
},
|
|
{
|
|
"epoch": 4.87060546875,
|
|
"grad_norm": 2.7468326091766357,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4026,
|
|
"step": 9975
|
|
},
|
|
{
|
|
"epoch": 4.8828125,
|
|
"grad_norm": 4.006360054016113,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4785,
|
|
"step": 10000
|
|
},
|
|
{
|
|
"epoch": 4.8828125,
|
|
"eval_cer": 21.490909090909092,
|
|
"eval_loss": 0.39705541729927063,
|
|
"eval_normalized_cer": 16.48170982326346,
|
|
"eval_runtime": 135.3957,
|
|
"eval_samples_per_second": 0.945,
|
|
"eval_steps_per_second": 0.059,
|
|
"step": 10000
|
|
},
|
|
{
|
|
"epoch": 4.89501953125,
|
|
"grad_norm": 1.9145240783691406,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4676,
|
|
"step": 10025
|
|
},
|
|
{
|
|
"epoch": 4.9072265625,
|
|
"grad_norm": 3.474172830581665,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4759,
|
|
"step": 10050
|
|
},
|
|
{
|
|
"epoch": 4.91943359375,
|
|
"grad_norm": 4.628629684448242,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4614,
|
|
"step": 10075
|
|
},
|
|
{
|
|
"epoch": 4.931640625,
|
|
"grad_norm": 3.4078354835510254,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4194,
|
|
"step": 10100
|
|
},
|
|
{
|
|
"epoch": 4.94384765625,
|
|
"grad_norm": 2.336055278778076,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4508,
|
|
"step": 10125
|
|
},
|
|
{
|
|
"epoch": 4.9560546875,
|
|
"grad_norm": 3.3985817432403564,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5009,
|
|
"step": 10150
|
|
},
|
|
{
|
|
"epoch": 4.96826171875,
|
|
"grad_norm": 3.0444228649139404,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4485,
|
|
"step": 10175
|
|
},
|
|
{
|
|
"epoch": 4.98046875,
|
|
"grad_norm": 3.737316131591797,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4723,
|
|
"step": 10200
|
|
},
|
|
{
|
|
"epoch": 4.99267578125,
|
|
"grad_norm": 2.8594186305999756,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4411,
|
|
"step": 10225
|
|
},
|
|
{
|
|
"epoch": 5.0048828125,
|
|
"grad_norm": 3.403825044631958,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3681,
|
|
"step": 10250
|
|
},
|
|
{
|
|
"epoch": 5.01708984375,
|
|
"grad_norm": 2.896949291229248,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4386,
|
|
"step": 10275
|
|
},
|
|
{
|
|
"epoch": 5.029296875,
|
|
"grad_norm": 3.3506815433502197,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4113,
|
|
"step": 10300
|
|
},
|
|
{
|
|
"epoch": 5.04150390625,
|
|
"grad_norm": 4.682892322540283,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4773,
|
|
"step": 10325
|
|
},
|
|
{
|
|
"epoch": 5.0537109375,
|
|
"grad_norm": 4.2950439453125,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4289,
|
|
"step": 10350
|
|
},
|
|
{
|
|
"epoch": 5.06591796875,
|
|
"grad_norm": 3.2204432487487793,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.423,
|
|
"step": 10375
|
|
},
|
|
{
|
|
"epoch": 5.078125,
|
|
"grad_norm": 4.998103618621826,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4115,
|
|
"step": 10400
|
|
},
|
|
{
|
|
"epoch": 5.09033203125,
|
|
"grad_norm": 2.7820868492126465,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4685,
|
|
"step": 10425
|
|
},
|
|
{
|
|
"epoch": 5.1025390625,
|
|
"grad_norm": 3.236323356628418,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4255,
|
|
"step": 10450
|
|
},
|
|
{
|
|
"epoch": 5.11474609375,
|
|
"grad_norm": 3.809432029724121,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4964,
|
|
"step": 10475
|
|
},
|
|
{
|
|
"epoch": 5.126953125,
|
|
"grad_norm": 5.030189514160156,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4664,
|
|
"step": 10500
|
|
},
|
|
{
|
|
"epoch": 5.126953125,
|
|
"eval_cer": 22.581818181818182,
|
|
"eval_loss": 0.4009712040424347,
|
|
"eval_normalized_cer": 18.33127825729552,
|
|
"eval_runtime": 135.5914,
|
|
"eval_samples_per_second": 0.944,
|
|
"eval_steps_per_second": 0.059,
|
|
"step": 10500
|
|
},
|
|
{
|
|
"epoch": 5.13916015625,
|
|
"grad_norm": 2.855731248855591,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4279,
|
|
"step": 10525
|
|
},
|
|
{
|
|
"epoch": 5.1513671875,
|
|
"grad_norm": 3.264129161834717,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4094,
|
|
"step": 10550
|
|
},
|
|
{
|
|
"epoch": 5.16357421875,
|
|
"grad_norm": 3.299495220184326,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4343,
|
|
"step": 10575
|
|
},
|
|
{
|
|
"epoch": 5.17578125,
|
|
"grad_norm": 3.564100503921509,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4514,
|
|
"step": 10600
|
|
},
|
|
{
|
|
"epoch": 5.18798828125,
|
|
"grad_norm": 2.6773531436920166,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4512,
|
|
"step": 10625
|
|
},
|
|
{
|
|
"epoch": 5.2001953125,
|
|
"grad_norm": 3.721902370452881,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4326,
|
|
"step": 10650
|
|
},
|
|
{
|
|
"epoch": 5.21240234375,
|
|
"grad_norm": 2.919142961502075,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4466,
|
|
"step": 10675
|
|
},
|
|
{
|
|
"epoch": 5.224609375,
|
|
"grad_norm": 3.0286405086517334,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3802,
|
|
"step": 10700
|
|
},
|
|
{
|
|
"epoch": 5.23681640625,
|
|
"grad_norm": 4.255770206451416,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4289,
|
|
"step": 10725
|
|
},
|
|
{
|
|
"epoch": 5.2490234375,
|
|
"grad_norm": 4.174068450927734,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4699,
|
|
"step": 10750
|
|
},
|
|
{
|
|
"epoch": 5.26123046875,
|
|
"grad_norm": 2.8515052795410156,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.464,
|
|
"step": 10775
|
|
},
|
|
{
|
|
"epoch": 5.2734375,
|
|
"grad_norm": 2.487253427505493,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4222,
|
|
"step": 10800
|
|
},
|
|
{
|
|
"epoch": 5.28564453125,
|
|
"grad_norm": 2.592994213104248,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3916,
|
|
"step": 10825
|
|
},
|
|
{
|
|
"epoch": 5.2978515625,
|
|
"grad_norm": 3.5928993225097656,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4278,
|
|
"step": 10850
|
|
},
|
|
{
|
|
"epoch": 5.31005859375,
|
|
"grad_norm": 3.690007448196411,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4369,
|
|
"step": 10875
|
|
},
|
|
{
|
|
"epoch": 5.322265625,
|
|
"grad_norm": 3.200507879257202,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.391,
|
|
"step": 10900
|
|
},
|
|
{
|
|
"epoch": 5.33447265625,
|
|
"grad_norm": 2.6833174228668213,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4602,
|
|
"step": 10925
|
|
},
|
|
{
|
|
"epoch": 5.3466796875,
|
|
"grad_norm": 4.44938325881958,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4648,
|
|
"step": 10950
|
|
},
|
|
{
|
|
"epoch": 5.35888671875,
|
|
"grad_norm": 3.660412073135376,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4193,
|
|
"step": 10975
|
|
},
|
|
{
|
|
"epoch": 5.37109375,
|
|
"grad_norm": 3.3475520610809326,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4497,
|
|
"step": 11000
|
|
},
|
|
{
|
|
"epoch": 5.37109375,
|
|
"eval_cer": 23.381818181818183,
|
|
"eval_loss": 0.40254130959510803,
|
|
"eval_normalized_cer": 18.824496506370735,
|
|
"eval_runtime": 134.1905,
|
|
"eval_samples_per_second": 0.954,
|
|
"eval_steps_per_second": 0.06,
|
|
"step": 11000
|
|
},
|
|
{
|
|
"epoch": 5.38330078125,
|
|
"grad_norm": 2.6810033321380615,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4332,
|
|
"step": 11025
|
|
},
|
|
{
|
|
"epoch": 5.3955078125,
|
|
"grad_norm": 3.203681707382202,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4225,
|
|
"step": 11050
|
|
},
|
|
{
|
|
"epoch": 5.40771484375,
|
|
"grad_norm": 2.916377305984497,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4133,
|
|
"step": 11075
|
|
},
|
|
{
|
|
"epoch": 5.419921875,
|
|
"grad_norm": 3.2195537090301514,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4842,
|
|
"step": 11100
|
|
},
|
|
{
|
|
"epoch": 5.43212890625,
|
|
"grad_norm": 3.1595821380615234,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4547,
|
|
"step": 11125
|
|
},
|
|
{
|
|
"epoch": 5.4443359375,
|
|
"grad_norm": 3.207057237625122,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.394,
|
|
"step": 11150
|
|
},
|
|
{
|
|
"epoch": 5.45654296875,
|
|
"grad_norm": 2.9330296516418457,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3902,
|
|
"step": 11175
|
|
},
|
|
{
|
|
"epoch": 5.46875,
|
|
"grad_norm": 2.4040415287017822,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4565,
|
|
"step": 11200
|
|
},
|
|
{
|
|
"epoch": 5.48095703125,
|
|
"grad_norm": 5.465662479400635,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4368,
|
|
"step": 11225
|
|
},
|
|
{
|
|
"epoch": 5.4931640625,
|
|
"grad_norm": 3.320988178253174,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4346,
|
|
"step": 11250
|
|
},
|
|
{
|
|
"epoch": 5.50537109375,
|
|
"grad_norm": 2.748600959777832,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4632,
|
|
"step": 11275
|
|
},
|
|
{
|
|
"epoch": 5.517578125,
|
|
"grad_norm": 2.1236891746520996,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4194,
|
|
"step": 11300
|
|
},
|
|
{
|
|
"epoch": 5.52978515625,
|
|
"grad_norm": 3.5435776710510254,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4293,
|
|
"step": 11325
|
|
},
|
|
{
|
|
"epoch": 5.5419921875,
|
|
"grad_norm": 3.9727423191070557,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4373,
|
|
"step": 11350
|
|
},
|
|
{
|
|
"epoch": 5.55419921875,
|
|
"grad_norm": 4.122025966644287,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3914,
|
|
"step": 11375
|
|
},
|
|
{
|
|
"epoch": 5.56640625,
|
|
"grad_norm": 3.523620128631592,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4883,
|
|
"step": 11400
|
|
},
|
|
{
|
|
"epoch": 5.57861328125,
|
|
"grad_norm": 3.5819413661956787,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4592,
|
|
"step": 11425
|
|
},
|
|
{
|
|
"epoch": 5.5908203125,
|
|
"grad_norm": 3.022115468978882,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4231,
|
|
"step": 11450
|
|
},
|
|
{
|
|
"epoch": 5.60302734375,
|
|
"grad_norm": 3.9213130474090576,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4391,
|
|
"step": 11475
|
|
},
|
|
{
|
|
"epoch": 5.615234375,
|
|
"grad_norm": 2.55260968208313,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4375,
|
|
"step": 11500
|
|
},
|
|
{
|
|
"epoch": 5.615234375,
|
|
"eval_cer": 23.745454545454546,
|
|
"eval_loss": 0.3965916633605957,
|
|
"eval_normalized_cer": 20.098643649815042,
|
|
"eval_runtime": 136.6172,
|
|
"eval_samples_per_second": 0.937,
|
|
"eval_steps_per_second": 0.059,
|
|
"step": 11500
|
|
},
|
|
{
|
|
"epoch": 5.62744140625,
|
|
"grad_norm": 4.3812479972839355,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4251,
|
|
"step": 11525
|
|
},
|
|
{
|
|
"epoch": 5.6396484375,
|
|
"grad_norm": 3.3777589797973633,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4634,
|
|
"step": 11550
|
|
},
|
|
{
|
|
"epoch": 5.65185546875,
|
|
"grad_norm": 3.418990135192871,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.453,
|
|
"step": 11575
|
|
},
|
|
{
|
|
"epoch": 5.6640625,
|
|
"grad_norm": 3.4627134799957275,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4335,
|
|
"step": 11600
|
|
},
|
|
{
|
|
"epoch": 5.67626953125,
|
|
"grad_norm": 3.955687999725342,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4648,
|
|
"step": 11625
|
|
},
|
|
{
|
|
"epoch": 5.6884765625,
|
|
"grad_norm": 3.9896974563598633,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.468,
|
|
"step": 11650
|
|
},
|
|
{
|
|
"epoch": 5.70068359375,
|
|
"grad_norm": 4.082157611846924,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4801,
|
|
"step": 11675
|
|
},
|
|
{
|
|
"epoch": 5.712890625,
|
|
"grad_norm": 2.509510040283203,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4414,
|
|
"step": 11700
|
|
},
|
|
{
|
|
"epoch": 5.72509765625,
|
|
"grad_norm": 4.054114818572998,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4341,
|
|
"step": 11725
|
|
},
|
|
{
|
|
"epoch": 5.7373046875,
|
|
"grad_norm": 3.3360326290130615,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4294,
|
|
"step": 11750
|
|
},
|
|
{
|
|
"epoch": 5.74951171875,
|
|
"grad_norm": 3.023287773132324,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4177,
|
|
"step": 11775
|
|
},
|
|
{
|
|
"epoch": 5.76171875,
|
|
"grad_norm": 2.7048487663269043,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4284,
|
|
"step": 11800
|
|
},
|
|
{
|
|
"epoch": 5.77392578125,
|
|
"grad_norm": 3.7769124507904053,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4297,
|
|
"step": 11825
|
|
},
|
|
{
|
|
"epoch": 5.7861328125,
|
|
"grad_norm": 2.583249092102051,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4328,
|
|
"step": 11850
|
|
},
|
|
{
|
|
"epoch": 5.79833984375,
|
|
"grad_norm": 4.198108196258545,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4499,
|
|
"step": 11875
|
|
},
|
|
{
|
|
"epoch": 5.810546875,
|
|
"grad_norm": 2.8993310928344727,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4081,
|
|
"step": 11900
|
|
},
|
|
{
|
|
"epoch": 5.82275390625,
|
|
"grad_norm": 3.5087976455688477,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4716,
|
|
"step": 11925
|
|
},
|
|
{
|
|
"epoch": 5.8349609375,
|
|
"grad_norm": 3.527189254760742,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4229,
|
|
"step": 11950
|
|
},
|
|
{
|
|
"epoch": 5.84716796875,
|
|
"grad_norm": 2.4347596168518066,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.448,
|
|
"step": 11975
|
|
},
|
|
{
|
|
"epoch": 5.859375,
|
|
"grad_norm": 3.4706249237060547,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4526,
|
|
"step": 12000
|
|
},
|
|
{
|
|
"epoch": 5.859375,
|
|
"eval_cer": 24.87272727272727,
|
|
"eval_loss": 0.3955676555633545,
|
|
"eval_normalized_cer": 20.263049732840116,
|
|
"eval_runtime": 133.1253,
|
|
"eval_samples_per_second": 0.961,
|
|
"eval_steps_per_second": 0.06,
|
|
"step": 12000
|
|
},
|
|
{
|
|
"epoch": 5.87158203125,
|
|
"grad_norm": 3.7815253734588623,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4,
|
|
"step": 12025
|
|
},
|
|
{
|
|
"epoch": 5.8837890625,
|
|
"grad_norm": 4.76561164855957,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4176,
|
|
"step": 12050
|
|
},
|
|
{
|
|
"epoch": 5.89599609375,
|
|
"grad_norm": 3.329918622970581,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4232,
|
|
"step": 12075
|
|
},
|
|
{
|
|
"epoch": 5.908203125,
|
|
"grad_norm": 4.160305023193359,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4123,
|
|
"step": 12100
|
|
},
|
|
{
|
|
"epoch": 5.92041015625,
|
|
"grad_norm": 3.11075758934021,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4468,
|
|
"step": 12125
|
|
},
|
|
{
|
|
"epoch": 5.9326171875,
|
|
"grad_norm": 3.1409554481506348,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4624,
|
|
"step": 12150
|
|
},
|
|
{
|
|
"epoch": 5.94482421875,
|
|
"grad_norm": 2.9979705810546875,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4369,
|
|
"step": 12175
|
|
},
|
|
{
|
|
"epoch": 5.95703125,
|
|
"grad_norm": 3.538670301437378,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4299,
|
|
"step": 12200
|
|
},
|
|
{
|
|
"epoch": 5.96923828125,
|
|
"grad_norm": 3.377985954284668,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3944,
|
|
"step": 12225
|
|
},
|
|
{
|
|
"epoch": 5.9814453125,
|
|
"grad_norm": 3.6021058559417725,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4314,
|
|
"step": 12250
|
|
},
|
|
{
|
|
"epoch": 5.99365234375,
|
|
"grad_norm": 3.2901623249053955,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4104,
|
|
"step": 12275
|
|
},
|
|
{
|
|
"epoch": 6.005859375,
|
|
"grad_norm": 4.384257793426514,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5273,
|
|
"step": 12300
|
|
},
|
|
{
|
|
"epoch": 6.01806640625,
|
|
"grad_norm": 3.564580202102661,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4292,
|
|
"step": 12325
|
|
},
|
|
{
|
|
"epoch": 6.0302734375,
|
|
"grad_norm": 3.4145617485046387,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4183,
|
|
"step": 12350
|
|
},
|
|
{
|
|
"epoch": 6.04248046875,
|
|
"grad_norm": 3.1783013343811035,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4193,
|
|
"step": 12375
|
|
},
|
|
{
|
|
"epoch": 6.0546875,
|
|
"grad_norm": 3.3418772220611572,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4635,
|
|
"step": 12400
|
|
},
|
|
{
|
|
"epoch": 6.06689453125,
|
|
"grad_norm": 3.057255744934082,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4068,
|
|
"step": 12425
|
|
},
|
|
{
|
|
"epoch": 6.0791015625,
|
|
"grad_norm": 3.5321204662323,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3614,
|
|
"step": 12450
|
|
},
|
|
{
|
|
"epoch": 6.09130859375,
|
|
"grad_norm": 3.536223888397217,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4195,
|
|
"step": 12475
|
|
},
|
|
{
|
|
"epoch": 6.103515625,
|
|
"grad_norm": 2.386425018310547,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4545,
|
|
"step": 12500
|
|
},
|
|
{
|
|
"epoch": 6.103515625,
|
|
"eval_cer": 26.47272727272727,
|
|
"eval_loss": 0.38523948192596436,
|
|
"eval_normalized_cer": 20.550760378133994,
|
|
"eval_runtime": 117.0609,
|
|
"eval_samples_per_second": 1.093,
|
|
"eval_steps_per_second": 0.068,
|
|
"step": 12500
|
|
},
|
|
{
|
|
"epoch": 6.11572265625,
|
|
"grad_norm": 2.7935492992401123,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4149,
|
|
"step": 12525
|
|
},
|
|
{
|
|
"epoch": 6.1279296875,
|
|
"grad_norm": 3.3430776596069336,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4718,
|
|
"step": 12550
|
|
},
|
|
{
|
|
"epoch": 6.14013671875,
|
|
"grad_norm": 2.8501319885253906,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.467,
|
|
"step": 12575
|
|
},
|
|
{
|
|
"epoch": 6.15234375,
|
|
"grad_norm": 3.3480148315429688,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4245,
|
|
"step": 12600
|
|
},
|
|
{
|
|
"epoch": 6.16455078125,
|
|
"grad_norm": 3.101698160171509,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4185,
|
|
"step": 12625
|
|
},
|
|
{
|
|
"epoch": 6.1767578125,
|
|
"grad_norm": 2.7815115451812744,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.446,
|
|
"step": 12650
|
|
},
|
|
{
|
|
"epoch": 6.18896484375,
|
|
"grad_norm": 2.8936548233032227,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4136,
|
|
"step": 12675
|
|
},
|
|
{
|
|
"epoch": 6.201171875,
|
|
"grad_norm": 4.183928966522217,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4384,
|
|
"step": 12700
|
|
},
|
|
{
|
|
"epoch": 6.21337890625,
|
|
"grad_norm": 3.2840652465820312,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4782,
|
|
"step": 12725
|
|
},
|
|
{
|
|
"epoch": 6.2255859375,
|
|
"grad_norm": 3.9204938411712646,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.432,
|
|
"step": 12750
|
|
},
|
|
{
|
|
"epoch": 6.23779296875,
|
|
"grad_norm": 4.316317081451416,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4644,
|
|
"step": 12775
|
|
},
|
|
{
|
|
"epoch": 6.25,
|
|
"grad_norm": 3.1822116374969482,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4331,
|
|
"step": 12800
|
|
},
|
|
{
|
|
"epoch": 6.26220703125,
|
|
"grad_norm": 2.9394965171813965,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4263,
|
|
"step": 12825
|
|
},
|
|
{
|
|
"epoch": 6.2744140625,
|
|
"grad_norm": 3.419806718826294,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4219,
|
|
"step": 12850
|
|
},
|
|
{
|
|
"epoch": 6.28662109375,
|
|
"grad_norm": 3.1678640842437744,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4425,
|
|
"step": 12875
|
|
},
|
|
{
|
|
"epoch": 6.298828125,
|
|
"grad_norm": 4.1609697341918945,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4355,
|
|
"step": 12900
|
|
},
|
|
{
|
|
"epoch": 6.31103515625,
|
|
"grad_norm": 3.895420789718628,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4633,
|
|
"step": 12925
|
|
},
|
|
{
|
|
"epoch": 6.3232421875,
|
|
"grad_norm": 3.244763135910034,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4219,
|
|
"step": 12950
|
|
},
|
|
{
|
|
"epoch": 6.33544921875,
|
|
"grad_norm": 4.2628984451293945,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.432,
|
|
"step": 12975
|
|
},
|
|
{
|
|
"epoch": 6.34765625,
|
|
"grad_norm": 4.046095848083496,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4199,
|
|
"step": 13000
|
|
},
|
|
{
|
|
"epoch": 6.34765625,
|
|
"eval_cer": 20.909090909090907,
|
|
"eval_loss": 0.38916099071502686,
|
|
"eval_normalized_cer": 16.31730374023839,
|
|
"eval_runtime": 128.9363,
|
|
"eval_samples_per_second": 0.993,
|
|
"eval_steps_per_second": 0.062,
|
|
"step": 13000
|
|
},
|
|
{
|
|
"epoch": 6.35986328125,
|
|
"grad_norm": 3.0475289821624756,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4105,
|
|
"step": 13025
|
|
},
|
|
{
|
|
"epoch": 6.3720703125,
|
|
"grad_norm": 3.288393974304199,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4627,
|
|
"step": 13050
|
|
},
|
|
{
|
|
"epoch": 6.38427734375,
|
|
"grad_norm": 4.572826862335205,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4259,
|
|
"step": 13075
|
|
},
|
|
{
|
|
"epoch": 6.396484375,
|
|
"grad_norm": 3.2055158615112305,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4124,
|
|
"step": 13100
|
|
},
|
|
{
|
|
"epoch": 6.40869140625,
|
|
"grad_norm": 3.237438917160034,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4525,
|
|
"step": 13125
|
|
},
|
|
{
|
|
"epoch": 6.4208984375,
|
|
"grad_norm": 3.7672817707061768,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4218,
|
|
"step": 13150
|
|
},
|
|
{
|
|
"epoch": 6.43310546875,
|
|
"grad_norm": 2.8958332538604736,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4501,
|
|
"step": 13175
|
|
},
|
|
{
|
|
"epoch": 6.4453125,
|
|
"grad_norm": 3.3539748191833496,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4294,
|
|
"step": 13200
|
|
},
|
|
{
|
|
"epoch": 6.45751953125,
|
|
"grad_norm": 3.745314836502075,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4038,
|
|
"step": 13225
|
|
},
|
|
{
|
|
"epoch": 6.4697265625,
|
|
"grad_norm": 4.028524398803711,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4313,
|
|
"step": 13250
|
|
},
|
|
{
|
|
"epoch": 6.48193359375,
|
|
"grad_norm": 3.058047294616699,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4215,
|
|
"step": 13275
|
|
},
|
|
{
|
|
"epoch": 6.494140625,
|
|
"grad_norm": 2.5536253452301025,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4306,
|
|
"step": 13300
|
|
},
|
|
{
|
|
"epoch": 6.50634765625,
|
|
"grad_norm": 2.894327402114868,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4211,
|
|
"step": 13325
|
|
},
|
|
{
|
|
"epoch": 6.5185546875,
|
|
"grad_norm": 4.001802444458008,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4212,
|
|
"step": 13350
|
|
},
|
|
{
|
|
"epoch": 6.53076171875,
|
|
"grad_norm": 3.1187326908111572,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4371,
|
|
"step": 13375
|
|
},
|
|
{
|
|
"epoch": 6.54296875,
|
|
"grad_norm": 2.8355751037597656,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4356,
|
|
"step": 13400
|
|
},
|
|
{
|
|
"epoch": 6.55517578125,
|
|
"grad_norm": 4.038667678833008,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4009,
|
|
"step": 13425
|
|
},
|
|
{
|
|
"epoch": 6.5673828125,
|
|
"grad_norm": 3.5054690837860107,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4561,
|
|
"step": 13450
|
|
},
|
|
{
|
|
"epoch": 6.57958984375,
|
|
"grad_norm": 3.1248323917388916,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4244,
|
|
"step": 13475
|
|
},
|
|
{
|
|
"epoch": 6.591796875,
|
|
"grad_norm": 3.1242165565490723,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4456,
|
|
"step": 13500
|
|
},
|
|
{
|
|
"epoch": 6.591796875,
|
|
"eval_cer": 22.545454545454547,
|
|
"eval_loss": 0.3937221169471741,
|
|
"eval_normalized_cer": 17.79695848746404,
|
|
"eval_runtime": 131.9352,
|
|
"eval_samples_per_second": 0.97,
|
|
"eval_steps_per_second": 0.061,
|
|
"step": 13500
|
|
},
|
|
{
|
|
"epoch": 6.60400390625,
|
|
"grad_norm": 2.5615017414093018,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4451,
|
|
"step": 13525
|
|
},
|
|
{
|
|
"epoch": 6.6162109375,
|
|
"grad_norm": 3.3750381469726562,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4233,
|
|
"step": 13550
|
|
},
|
|
{
|
|
"epoch": 6.62841796875,
|
|
"grad_norm": 3.462449073791504,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4426,
|
|
"step": 13575
|
|
},
|
|
{
|
|
"epoch": 6.640625,
|
|
"grad_norm": 2.785858154296875,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4073,
|
|
"step": 13600
|
|
},
|
|
{
|
|
"epoch": 6.65283203125,
|
|
"grad_norm": 2.7007555961608887,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.438,
|
|
"step": 13625
|
|
},
|
|
{
|
|
"epoch": 6.6650390625,
|
|
"grad_norm": 2.3689684867858887,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4226,
|
|
"step": 13650
|
|
},
|
|
{
|
|
"epoch": 6.67724609375,
|
|
"grad_norm": 3.3723831176757812,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4279,
|
|
"step": 13675
|
|
},
|
|
{
|
|
"epoch": 6.689453125,
|
|
"grad_norm": 2.9226441383361816,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4242,
|
|
"step": 13700
|
|
},
|
|
{
|
|
"epoch": 6.70166015625,
|
|
"grad_norm": 2.7876479625701904,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4328,
|
|
"step": 13725
|
|
},
|
|
{
|
|
"epoch": 6.7138671875,
|
|
"grad_norm": 2.7610867023468018,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4412,
|
|
"step": 13750
|
|
},
|
|
{
|
|
"epoch": 6.72607421875,
|
|
"grad_norm": 2.8302724361419678,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4526,
|
|
"step": 13775
|
|
},
|
|
{
|
|
"epoch": 6.73828125,
|
|
"grad_norm": 3.1445977687835693,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4544,
|
|
"step": 13800
|
|
},
|
|
{
|
|
"epoch": 6.75048828125,
|
|
"grad_norm": 3.8864002227783203,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3965,
|
|
"step": 13825
|
|
},
|
|
{
|
|
"epoch": 6.7626953125,
|
|
"grad_norm": 3.2896971702575684,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4182,
|
|
"step": 13850
|
|
},
|
|
{
|
|
"epoch": 6.77490234375,
|
|
"grad_norm": 4.539292335510254,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4413,
|
|
"step": 13875
|
|
},
|
|
{
|
|
"epoch": 6.787109375,
|
|
"grad_norm": 3.2103347778320312,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4346,
|
|
"step": 13900
|
|
},
|
|
{
|
|
"epoch": 6.79931640625,
|
|
"grad_norm": 3.929469585418701,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4181,
|
|
"step": 13925
|
|
},
|
|
{
|
|
"epoch": 6.8115234375,
|
|
"grad_norm": 3.3086280822753906,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3981,
|
|
"step": 13950
|
|
},
|
|
{
|
|
"epoch": 6.82373046875,
|
|
"grad_norm": 3.101243734359741,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.503,
|
|
"step": 13975
|
|
},
|
|
{
|
|
"epoch": 6.8359375,
|
|
"grad_norm": 3.602478265762329,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4186,
|
|
"step": 14000
|
|
},
|
|
{
|
|
"epoch": 6.8359375,
|
|
"eval_cer": 22.472727272727273,
|
|
"eval_loss": 0.3899889588356018,
|
|
"eval_normalized_cer": 18.33127825729552,
|
|
"eval_runtime": 144.5925,
|
|
"eval_samples_per_second": 0.885,
|
|
"eval_steps_per_second": 0.055,
|
|
"step": 14000
|
|
},
|
|
{
|
|
"epoch": 6.84814453125,
|
|
"grad_norm": 3.0590033531188965,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4268,
|
|
"step": 14025
|
|
},
|
|
{
|
|
"epoch": 6.8603515625,
|
|
"grad_norm": 4.508790969848633,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4063,
|
|
"step": 14050
|
|
},
|
|
{
|
|
"epoch": 6.87255859375,
|
|
"grad_norm": 3.374329090118408,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4482,
|
|
"step": 14075
|
|
},
|
|
{
|
|
"epoch": 6.884765625,
|
|
"grad_norm": 2.569153308868408,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4477,
|
|
"step": 14100
|
|
},
|
|
{
|
|
"epoch": 6.89697265625,
|
|
"grad_norm": 2.5716049671173096,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4282,
|
|
"step": 14125
|
|
},
|
|
{
|
|
"epoch": 6.9091796875,
|
|
"grad_norm": 2.990652561187744,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4409,
|
|
"step": 14150
|
|
},
|
|
{
|
|
"epoch": 6.92138671875,
|
|
"grad_norm": 2.772998094558716,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3622,
|
|
"step": 14175
|
|
},
|
|
{
|
|
"epoch": 6.93359375,
|
|
"grad_norm": 3.0832324028015137,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4017,
|
|
"step": 14200
|
|
},
|
|
{
|
|
"epoch": 6.94580078125,
|
|
"grad_norm": 2.2825822830200195,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4341,
|
|
"step": 14225
|
|
},
|
|
{
|
|
"epoch": 6.9580078125,
|
|
"grad_norm": 2.8773045539855957,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4512,
|
|
"step": 14250
|
|
},
|
|
{
|
|
"epoch": 6.97021484375,
|
|
"grad_norm": 2.9340527057647705,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4004,
|
|
"step": 14275
|
|
},
|
|
{
|
|
"epoch": 6.982421875,
|
|
"grad_norm": 3.3255691528320312,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3775,
|
|
"step": 14300
|
|
},
|
|
{
|
|
"epoch": 6.99462890625,
|
|
"grad_norm": 3.9721550941467285,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4239,
|
|
"step": 14325
|
|
},
|
|
{
|
|
"epoch": 7.0068359375,
|
|
"grad_norm": 2.075686454772949,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3713,
|
|
"step": 14350
|
|
},
|
|
{
|
|
"epoch": 7.01904296875,
|
|
"grad_norm": 3.62715744972229,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4379,
|
|
"step": 14375
|
|
},
|
|
{
|
|
"epoch": 7.03125,
|
|
"grad_norm": 2.931847333908081,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3971,
|
|
"step": 14400
|
|
},
|
|
{
|
|
"epoch": 7.04345703125,
|
|
"grad_norm": 3.470655679702759,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4529,
|
|
"step": 14425
|
|
},
|
|
{
|
|
"epoch": 7.0556640625,
|
|
"grad_norm": 3.3625166416168213,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4197,
|
|
"step": 14450
|
|
},
|
|
{
|
|
"epoch": 7.06787109375,
|
|
"grad_norm": 3.200965404510498,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4553,
|
|
"step": 14475
|
|
},
|
|
{
|
|
"epoch": 7.080078125,
|
|
"grad_norm": 3.83722186088562,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4601,
|
|
"step": 14500
|
|
},
|
|
{
|
|
"epoch": 7.080078125,
|
|
"eval_cer": 20.727272727272727,
|
|
"eval_loss": 0.3850056231021881,
|
|
"eval_normalized_cer": 16.399506781750926,
|
|
"eval_runtime": 135.6446,
|
|
"eval_samples_per_second": 0.944,
|
|
"eval_steps_per_second": 0.059,
|
|
"step": 14500
|
|
},
|
|
{
|
|
"epoch": 7.09228515625,
|
|
"grad_norm": 4.523788928985596,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4101,
|
|
"step": 14525
|
|
},
|
|
{
|
|
"epoch": 7.1044921875,
|
|
"grad_norm": 3.117316722869873,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3924,
|
|
"step": 14550
|
|
},
|
|
{
|
|
"epoch": 7.11669921875,
|
|
"grad_norm": 2.9018518924713135,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.421,
|
|
"step": 14575
|
|
},
|
|
{
|
|
"epoch": 7.12890625,
|
|
"grad_norm": 4.131909370422363,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4576,
|
|
"step": 14600
|
|
},
|
|
{
|
|
"epoch": 7.14111328125,
|
|
"grad_norm": 3.206359386444092,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4068,
|
|
"step": 14625
|
|
},
|
|
{
|
|
"epoch": 7.1533203125,
|
|
"grad_norm": 4.659736156463623,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3906,
|
|
"step": 14650
|
|
},
|
|
{
|
|
"epoch": 7.16552734375,
|
|
"grad_norm": 3.376612901687622,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4197,
|
|
"step": 14675
|
|
},
|
|
{
|
|
"epoch": 7.177734375,
|
|
"grad_norm": 3.268672227859497,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4122,
|
|
"step": 14700
|
|
},
|
|
{
|
|
"epoch": 7.18994140625,
|
|
"grad_norm": 3.0294389724731445,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4154,
|
|
"step": 14725
|
|
},
|
|
{
|
|
"epoch": 7.2021484375,
|
|
"grad_norm": 3.547778367996216,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4375,
|
|
"step": 14750
|
|
},
|
|
{
|
|
"epoch": 7.21435546875,
|
|
"grad_norm": 2.7753546237945557,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4212,
|
|
"step": 14775
|
|
},
|
|
{
|
|
"epoch": 7.2265625,
|
|
"grad_norm": 3.724691152572632,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4074,
|
|
"step": 14800
|
|
},
|
|
{
|
|
"epoch": 7.23876953125,
|
|
"grad_norm": 2.489899158477783,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4094,
|
|
"step": 14825
|
|
},
|
|
{
|
|
"epoch": 7.2509765625,
|
|
"grad_norm": 3.542677879333496,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4034,
|
|
"step": 14850
|
|
},
|
|
{
|
|
"epoch": 7.26318359375,
|
|
"grad_norm": 3.15748929977417,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4553,
|
|
"step": 14875
|
|
},
|
|
{
|
|
"epoch": 7.275390625,
|
|
"grad_norm": 3.477710008621216,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4394,
|
|
"step": 14900
|
|
},
|
|
{
|
|
"epoch": 7.28759765625,
|
|
"grad_norm": 4.362137794494629,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4311,
|
|
"step": 14925
|
|
},
|
|
{
|
|
"epoch": 7.2998046875,
|
|
"grad_norm": 2.3837854862213135,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4179,
|
|
"step": 14950
|
|
},
|
|
{
|
|
"epoch": 7.31201171875,
|
|
"grad_norm": 3.1717562675476074,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4504,
|
|
"step": 14975
|
|
},
|
|
{
|
|
"epoch": 7.32421875,
|
|
"grad_norm": 2.988471269607544,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4357,
|
|
"step": 15000
|
|
},
|
|
{
|
|
"epoch": 7.32421875,
|
|
"eval_cer": 23.599999999999998,
|
|
"eval_loss": 0.38635486364364624,
|
|
"eval_normalized_cer": 20.098643649815042,
|
|
"eval_runtime": 132.5727,
|
|
"eval_samples_per_second": 0.966,
|
|
"eval_steps_per_second": 0.06,
|
|
"step": 15000
|
|
},
|
|
{
|
|
"epoch": 7.33642578125,
|
|
"grad_norm": 3.9990575313568115,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4263,
|
|
"step": 15025
|
|
},
|
|
{
|
|
"epoch": 7.3486328125,
|
|
"grad_norm": 2.6988320350646973,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4311,
|
|
"step": 15050
|
|
},
|
|
{
|
|
"epoch": 7.36083984375,
|
|
"grad_norm": 3.537297487258911,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4552,
|
|
"step": 15075
|
|
},
|
|
{
|
|
"epoch": 7.373046875,
|
|
"grad_norm": 2.706217050552368,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4365,
|
|
"step": 15100
|
|
},
|
|
{
|
|
"epoch": 7.38525390625,
|
|
"grad_norm": 3.2613677978515625,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4512,
|
|
"step": 15125
|
|
},
|
|
{
|
|
"epoch": 7.3974609375,
|
|
"grad_norm": 3.039867877960205,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4485,
|
|
"step": 15150
|
|
},
|
|
{
|
|
"epoch": 7.40966796875,
|
|
"grad_norm": 3.6644127368927,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4218,
|
|
"step": 15175
|
|
},
|
|
{
|
|
"epoch": 7.421875,
|
|
"grad_norm": 3.3461201190948486,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4292,
|
|
"step": 15200
|
|
},
|
|
{
|
|
"epoch": 7.43408203125,
|
|
"grad_norm": 2.8256707191467285,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4205,
|
|
"step": 15225
|
|
},
|
|
{
|
|
"epoch": 7.4462890625,
|
|
"grad_norm": 3.3773632049560547,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4155,
|
|
"step": 15250
|
|
},
|
|
{
|
|
"epoch": 7.45849609375,
|
|
"grad_norm": 2.6571528911590576,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4567,
|
|
"step": 15275
|
|
},
|
|
{
|
|
"epoch": 7.470703125,
|
|
"grad_norm": 3.0483086109161377,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4293,
|
|
"step": 15300
|
|
},
|
|
{
|
|
"epoch": 7.48291015625,
|
|
"grad_norm": 3.6026933193206787,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.426,
|
|
"step": 15325
|
|
},
|
|
{
|
|
"epoch": 7.4951171875,
|
|
"grad_norm": 4.094547748565674,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4562,
|
|
"step": 15350
|
|
},
|
|
{
|
|
"epoch": 7.50732421875,
|
|
"grad_norm": 3.6871438026428223,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4438,
|
|
"step": 15375
|
|
},
|
|
{
|
|
"epoch": 7.51953125,
|
|
"grad_norm": 3.5841362476348877,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3874,
|
|
"step": 15400
|
|
},
|
|
{
|
|
"epoch": 7.53173828125,
|
|
"grad_norm": 3.3375258445739746,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4234,
|
|
"step": 15425
|
|
},
|
|
{
|
|
"epoch": 7.5439453125,
|
|
"grad_norm": 4.966141223907471,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4329,
|
|
"step": 15450
|
|
},
|
|
{
|
|
"epoch": 7.55615234375,
|
|
"grad_norm": 2.2988944053649902,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4377,
|
|
"step": 15475
|
|
},
|
|
{
|
|
"epoch": 7.568359375,
|
|
"grad_norm": 3.4808993339538574,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4375,
|
|
"step": 15500
|
|
},
|
|
{
|
|
"epoch": 7.568359375,
|
|
"eval_cer": 23.854545454545452,
|
|
"eval_loss": 0.38098400831222534,
|
|
"eval_normalized_cer": 19.399917796958487,
|
|
"eval_runtime": 137.4032,
|
|
"eval_samples_per_second": 0.932,
|
|
"eval_steps_per_second": 0.058,
|
|
"step": 15500
|
|
},
|
|
{
|
|
"epoch": 7.58056640625,
|
|
"grad_norm": 4.054482936859131,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3687,
|
|
"step": 15525
|
|
},
|
|
{
|
|
"epoch": 7.5927734375,
|
|
"grad_norm": 2.874267578125,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4033,
|
|
"step": 15550
|
|
},
|
|
{
|
|
"epoch": 7.60498046875,
|
|
"grad_norm": 2.669652223587036,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4262,
|
|
"step": 15575
|
|
},
|
|
{
|
|
"epoch": 7.6171875,
|
|
"grad_norm": 3.2882497310638428,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4163,
|
|
"step": 15600
|
|
},
|
|
{
|
|
"epoch": 7.62939453125,
|
|
"grad_norm": 4.546947479248047,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4668,
|
|
"step": 15625
|
|
},
|
|
{
|
|
"epoch": 7.6416015625,
|
|
"grad_norm": 2.2762033939361572,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4313,
|
|
"step": 15650
|
|
},
|
|
{
|
|
"epoch": 7.65380859375,
|
|
"grad_norm": 3.4660661220550537,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4181,
|
|
"step": 15675
|
|
},
|
|
{
|
|
"epoch": 7.666015625,
|
|
"grad_norm": 3.2215335369110107,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4114,
|
|
"step": 15700
|
|
},
|
|
{
|
|
"epoch": 7.67822265625,
|
|
"grad_norm": 3.1161818504333496,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4599,
|
|
"step": 15725
|
|
},
|
|
{
|
|
"epoch": 7.6904296875,
|
|
"grad_norm": 3.0069620609283447,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4498,
|
|
"step": 15750
|
|
},
|
|
{
|
|
"epoch": 7.70263671875,
|
|
"grad_norm": 2.84161114692688,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4252,
|
|
"step": 15775
|
|
},
|
|
{
|
|
"epoch": 7.71484375,
|
|
"grad_norm": 4.6650590896606445,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4297,
|
|
"step": 15800
|
|
},
|
|
{
|
|
"epoch": 7.72705078125,
|
|
"grad_norm": 2.6274569034576416,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3753,
|
|
"step": 15825
|
|
},
|
|
{
|
|
"epoch": 7.7392578125,
|
|
"grad_norm": 3.5292246341705322,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4168,
|
|
"step": 15850
|
|
},
|
|
{
|
|
"epoch": 7.75146484375,
|
|
"grad_norm": 2.8617982864379883,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4809,
|
|
"step": 15875
|
|
},
|
|
{
|
|
"epoch": 7.763671875,
|
|
"grad_norm": 3.669074773788452,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4613,
|
|
"step": 15900
|
|
},
|
|
{
|
|
"epoch": 7.77587890625,
|
|
"grad_norm": 2.994079113006592,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3902,
|
|
"step": 15925
|
|
},
|
|
{
|
|
"epoch": 7.7880859375,
|
|
"grad_norm": 2.6492769718170166,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3756,
|
|
"step": 15950
|
|
},
|
|
{
|
|
"epoch": 7.80029296875,
|
|
"grad_norm": 3.2743895053863525,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4476,
|
|
"step": 15975
|
|
},
|
|
{
|
|
"epoch": 7.8125,
|
|
"grad_norm": 3.279031276702881,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4563,
|
|
"step": 16000
|
|
},
|
|
{
|
|
"epoch": 7.8125,
|
|
"eval_cer": 23.381818181818183,
|
|
"eval_loss": 0.38140711188316345,
|
|
"eval_normalized_cer": 18.90669954788327,
|
|
"eval_runtime": 135.0002,
|
|
"eval_samples_per_second": 0.948,
|
|
"eval_steps_per_second": 0.059,
|
|
"step": 16000
|
|
},
|
|
{
|
|
"epoch": 7.82470703125,
|
|
"grad_norm": 2.7851359844207764,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4344,
|
|
"step": 16025
|
|
},
|
|
{
|
|
"epoch": 7.8369140625,
|
|
"grad_norm": 3.965728282928467,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4328,
|
|
"step": 16050
|
|
},
|
|
{
|
|
"epoch": 7.84912109375,
|
|
"grad_norm": 4.293801307678223,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4593,
|
|
"step": 16075
|
|
},
|
|
{
|
|
"epoch": 7.861328125,
|
|
"grad_norm": 3.0265841484069824,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4039,
|
|
"step": 16100
|
|
},
|
|
{
|
|
"epoch": 7.87353515625,
|
|
"grad_norm": 3.9985835552215576,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3911,
|
|
"step": 16125
|
|
},
|
|
{
|
|
"epoch": 7.8857421875,
|
|
"grad_norm": 4.016017436981201,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.458,
|
|
"step": 16150
|
|
},
|
|
{
|
|
"epoch": 7.89794921875,
|
|
"grad_norm": 3.816105842590332,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4125,
|
|
"step": 16175
|
|
},
|
|
{
|
|
"epoch": 7.91015625,
|
|
"grad_norm": 2.5083696842193604,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3769,
|
|
"step": 16200
|
|
},
|
|
{
|
|
"epoch": 7.92236328125,
|
|
"grad_norm": 3.251554012298584,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4705,
|
|
"step": 16225
|
|
},
|
|
{
|
|
"epoch": 7.9345703125,
|
|
"grad_norm": 3.370077610015869,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4565,
|
|
"step": 16250
|
|
},
|
|
{
|
|
"epoch": 7.94677734375,
|
|
"grad_norm": 2.996699810028076,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.461,
|
|
"step": 16275
|
|
},
|
|
{
|
|
"epoch": 7.958984375,
|
|
"grad_norm": 4.30794620513916,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4035,
|
|
"step": 16300
|
|
},
|
|
{
|
|
"epoch": 7.97119140625,
|
|
"grad_norm": 3.3062832355499268,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4067,
|
|
"step": 16325
|
|
},
|
|
{
|
|
"epoch": 7.9833984375,
|
|
"grad_norm": 2.4719371795654297,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4868,
|
|
"step": 16350
|
|
},
|
|
{
|
|
"epoch": 7.99560546875,
|
|
"grad_norm": 3.1429994106292725,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4341,
|
|
"step": 16375
|
|
},
|
|
{
|
|
"epoch": 8.0078125,
|
|
"grad_norm": 4.408067226409912,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4874,
|
|
"step": 16400
|
|
},
|
|
{
|
|
"epoch": 8.02001953125,
|
|
"grad_norm": 3.3049044609069824,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4783,
|
|
"step": 16425
|
|
},
|
|
{
|
|
"epoch": 8.0322265625,
|
|
"grad_norm": 2.1289191246032715,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3918,
|
|
"step": 16450
|
|
},
|
|
{
|
|
"epoch": 8.04443359375,
|
|
"grad_norm": 3.2784934043884277,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3953,
|
|
"step": 16475
|
|
},
|
|
{
|
|
"epoch": 8.056640625,
|
|
"grad_norm": 2.8000524044036865,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4145,
|
|
"step": 16500
|
|
},
|
|
{
|
|
"epoch": 8.056640625,
|
|
"eval_cer": 19.236363636363638,
|
|
"eval_loss": 0.3803868889808655,
|
|
"eval_normalized_cer": 15.207562679819153,
|
|
"eval_runtime": 142.7418,
|
|
"eval_samples_per_second": 0.897,
|
|
"eval_steps_per_second": 0.056,
|
|
"step": 16500
|
|
},
|
|
{
|
|
"epoch": 8.06884765625,
|
|
"grad_norm": 4.039844989776611,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4458,
|
|
"step": 16525
|
|
},
|
|
{
|
|
"epoch": 8.0810546875,
|
|
"grad_norm": 3.5788333415985107,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4211,
|
|
"step": 16550
|
|
},
|
|
{
|
|
"epoch": 8.09326171875,
|
|
"grad_norm": 2.7754452228546143,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3947,
|
|
"step": 16575
|
|
},
|
|
{
|
|
"epoch": 8.10546875,
|
|
"grad_norm": 3.6685307025909424,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4371,
|
|
"step": 16600
|
|
},
|
|
{
|
|
"epoch": 8.11767578125,
|
|
"grad_norm": 4.283354759216309,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4196,
|
|
"step": 16625
|
|
},
|
|
{
|
|
"epoch": 8.1298828125,
|
|
"grad_norm": 3.37106990814209,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4286,
|
|
"step": 16650
|
|
},
|
|
{
|
|
"epoch": 8.14208984375,
|
|
"grad_norm": 3.3839213848114014,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4081,
|
|
"step": 16675
|
|
},
|
|
{
|
|
"epoch": 8.154296875,
|
|
"grad_norm": 2.9974148273468018,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4218,
|
|
"step": 16700
|
|
},
|
|
{
|
|
"epoch": 8.16650390625,
|
|
"grad_norm": 2.861589193344116,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3866,
|
|
"step": 16725
|
|
},
|
|
{
|
|
"epoch": 8.1787109375,
|
|
"grad_norm": 3.7804067134857178,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.402,
|
|
"step": 16750
|
|
},
|
|
{
|
|
"epoch": 8.19091796875,
|
|
"grad_norm": 4.124424457550049,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4198,
|
|
"step": 16775
|
|
},
|
|
{
|
|
"epoch": 8.203125,
|
|
"grad_norm": 3.785266876220703,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4744,
|
|
"step": 16800
|
|
},
|
|
{
|
|
"epoch": 8.21533203125,
|
|
"grad_norm": 2.816866636276245,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4509,
|
|
"step": 16825
|
|
},
|
|
{
|
|
"epoch": 8.2275390625,
|
|
"grad_norm": 2.5543909072875977,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4379,
|
|
"step": 16850
|
|
},
|
|
{
|
|
"epoch": 8.23974609375,
|
|
"grad_norm": 3.561476707458496,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4109,
|
|
"step": 16875
|
|
},
|
|
{
|
|
"epoch": 8.251953125,
|
|
"grad_norm": 2.819086790084839,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4265,
|
|
"step": 16900
|
|
},
|
|
{
|
|
"epoch": 8.26416015625,
|
|
"grad_norm": 3.327193021774292,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3561,
|
|
"step": 16925
|
|
},
|
|
{
|
|
"epoch": 8.2763671875,
|
|
"grad_norm": 2.7188267707824707,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.405,
|
|
"step": 16950
|
|
},
|
|
{
|
|
"epoch": 8.28857421875,
|
|
"grad_norm": 4.3515543937683105,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4097,
|
|
"step": 16975
|
|
},
|
|
{
|
|
"epoch": 8.30078125,
|
|
"grad_norm": 3.3475663661956787,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3892,
|
|
"step": 17000
|
|
},
|
|
{
|
|
"epoch": 8.30078125,
|
|
"eval_cer": 22.581818181818182,
|
|
"eval_loss": 0.38389354944229126,
|
|
"eval_normalized_cer": 18.16687217427045,
|
|
"eval_runtime": 140.8271,
|
|
"eval_samples_per_second": 0.909,
|
|
"eval_steps_per_second": 0.057,
|
|
"step": 17000
|
|
},
|
|
{
|
|
"epoch": 8.31298828125,
|
|
"grad_norm": 3.1067376136779785,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.402,
|
|
"step": 17025
|
|
},
|
|
{
|
|
"epoch": 8.3251953125,
|
|
"grad_norm": 4.873087406158447,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4059,
|
|
"step": 17050
|
|
},
|
|
{
|
|
"epoch": 8.33740234375,
|
|
"grad_norm": 3.115588665008545,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4014,
|
|
"step": 17075
|
|
},
|
|
{
|
|
"epoch": 8.349609375,
|
|
"grad_norm": 4.50831413269043,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4337,
|
|
"step": 17100
|
|
},
|
|
{
|
|
"epoch": 8.36181640625,
|
|
"grad_norm": 2.399965763092041,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4034,
|
|
"step": 17125
|
|
},
|
|
{
|
|
"epoch": 8.3740234375,
|
|
"grad_norm": 3.1204044818878174,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4344,
|
|
"step": 17150
|
|
},
|
|
{
|
|
"epoch": 8.38623046875,
|
|
"grad_norm": 3.457709789276123,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4311,
|
|
"step": 17175
|
|
},
|
|
{
|
|
"epoch": 8.3984375,
|
|
"grad_norm": 2.5090980529785156,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4631,
|
|
"step": 17200
|
|
},
|
|
{
|
|
"epoch": 8.41064453125,
|
|
"grad_norm": 2.9333646297454834,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4329,
|
|
"step": 17225
|
|
},
|
|
{
|
|
"epoch": 8.4228515625,
|
|
"grad_norm": 2.8422629833221436,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3823,
|
|
"step": 17250
|
|
},
|
|
{
|
|
"epoch": 8.43505859375,
|
|
"grad_norm": 2.7649292945861816,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4538,
|
|
"step": 17275
|
|
},
|
|
{
|
|
"epoch": 8.447265625,
|
|
"grad_norm": 3.3747799396514893,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.416,
|
|
"step": 17300
|
|
},
|
|
{
|
|
"epoch": 8.45947265625,
|
|
"grad_norm": 4.490537643432617,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3744,
|
|
"step": 17325
|
|
},
|
|
{
|
|
"epoch": 8.4716796875,
|
|
"grad_norm": 3.027010202407837,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3874,
|
|
"step": 17350
|
|
},
|
|
{
|
|
"epoch": 8.48388671875,
|
|
"grad_norm": 2.616095542907715,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3695,
|
|
"step": 17375
|
|
},
|
|
{
|
|
"epoch": 8.49609375,
|
|
"grad_norm": 3.389639377593994,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4259,
|
|
"step": 17400
|
|
},
|
|
{
|
|
"epoch": 8.50830078125,
|
|
"grad_norm": 4.086320400238037,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4047,
|
|
"step": 17425
|
|
},
|
|
{
|
|
"epoch": 8.5205078125,
|
|
"grad_norm": 2.4645962715148926,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4357,
|
|
"step": 17450
|
|
},
|
|
{
|
|
"epoch": 8.53271484375,
|
|
"grad_norm": 2.2121071815490723,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4381,
|
|
"step": 17475
|
|
},
|
|
{
|
|
"epoch": 8.544921875,
|
|
"grad_norm": 2.978116512298584,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3699,
|
|
"step": 17500
|
|
},
|
|
{
|
|
"epoch": 8.544921875,
|
|
"eval_cer": 20.0,
|
|
"eval_loss": 0.382442444562912,
|
|
"eval_normalized_cer": 16.276202219482123,
|
|
"eval_runtime": 146.3607,
|
|
"eval_samples_per_second": 0.875,
|
|
"eval_steps_per_second": 0.055,
|
|
"step": 17500
|
|
},
|
|
{
|
|
"epoch": 8.55712890625,
|
|
"grad_norm": 3.6570897102355957,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4495,
|
|
"step": 17525
|
|
},
|
|
{
|
|
"epoch": 8.5693359375,
|
|
"grad_norm": 3.3602068424224854,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4132,
|
|
"step": 17550
|
|
},
|
|
{
|
|
"epoch": 8.58154296875,
|
|
"grad_norm": 3.4608097076416016,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4641,
|
|
"step": 17575
|
|
},
|
|
{
|
|
"epoch": 8.59375,
|
|
"grad_norm": 3.122091293334961,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4325,
|
|
"step": 17600
|
|
},
|
|
{
|
|
"epoch": 8.60595703125,
|
|
"grad_norm": 4.238564491271973,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4002,
|
|
"step": 17625
|
|
},
|
|
{
|
|
"epoch": 8.6181640625,
|
|
"grad_norm": 3.2410008907318115,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4678,
|
|
"step": 17650
|
|
},
|
|
{
|
|
"epoch": 8.63037109375,
|
|
"grad_norm": 2.8786799907684326,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3947,
|
|
"step": 17675
|
|
},
|
|
{
|
|
"epoch": 8.642578125,
|
|
"grad_norm": 2.8009581565856934,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4154,
|
|
"step": 17700
|
|
},
|
|
{
|
|
"epoch": 8.65478515625,
|
|
"grad_norm": 3.148653507232666,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3856,
|
|
"step": 17725
|
|
},
|
|
{
|
|
"epoch": 8.6669921875,
|
|
"grad_norm": 3.803799867630005,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4285,
|
|
"step": 17750
|
|
},
|
|
{
|
|
"epoch": 8.67919921875,
|
|
"grad_norm": 3.944755792617798,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4598,
|
|
"step": 17775
|
|
},
|
|
{
|
|
"epoch": 8.69140625,
|
|
"grad_norm": 3.1085519790649414,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4379,
|
|
"step": 17800
|
|
},
|
|
{
|
|
"epoch": 8.70361328125,
|
|
"grad_norm": 5.655707836151123,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4121,
|
|
"step": 17825
|
|
},
|
|
{
|
|
"epoch": 8.7158203125,
|
|
"grad_norm": 2.179668664932251,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4212,
|
|
"step": 17850
|
|
},
|
|
{
|
|
"epoch": 8.72802734375,
|
|
"grad_norm": 3.687593460083008,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4643,
|
|
"step": 17875
|
|
},
|
|
{
|
|
"epoch": 8.740234375,
|
|
"grad_norm": 2.9461138248443604,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4126,
|
|
"step": 17900
|
|
},
|
|
{
|
|
"epoch": 8.75244140625,
|
|
"grad_norm": 3.524925470352173,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4407,
|
|
"step": 17925
|
|
},
|
|
{
|
|
"epoch": 8.7646484375,
|
|
"grad_norm": 4.609610080718994,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.42,
|
|
"step": 17950
|
|
},
|
|
{
|
|
"epoch": 8.77685546875,
|
|
"grad_norm": 4.5199079513549805,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3921,
|
|
"step": 17975
|
|
},
|
|
{
|
|
"epoch": 8.7890625,
|
|
"grad_norm": 3.358597993850708,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.441,
|
|
"step": 18000
|
|
},
|
|
{
|
|
"epoch": 8.7890625,
|
|
"eval_cer": 21.854545454545455,
|
|
"eval_loss": 0.38204386830329895,
|
|
"eval_normalized_cer": 16.358405260994658,
|
|
"eval_runtime": 141.6466,
|
|
"eval_samples_per_second": 0.904,
|
|
"eval_steps_per_second": 0.056,
|
|
"step": 18000
|
|
},
|
|
{
|
|
"epoch": 8.80126953125,
|
|
"grad_norm": 3.380112648010254,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3832,
|
|
"step": 18025
|
|
},
|
|
{
|
|
"epoch": 8.8134765625,
|
|
"grad_norm": 3.249871015548706,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4449,
|
|
"step": 18050
|
|
},
|
|
{
|
|
"epoch": 8.82568359375,
|
|
"grad_norm": 3.3038084506988525,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4919,
|
|
"step": 18075
|
|
},
|
|
{
|
|
"epoch": 8.837890625,
|
|
"grad_norm": 3.6768815517425537,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.42,
|
|
"step": 18100
|
|
},
|
|
{
|
|
"epoch": 8.85009765625,
|
|
"grad_norm": 2.7637171745300293,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4275,
|
|
"step": 18125
|
|
},
|
|
{
|
|
"epoch": 8.8623046875,
|
|
"grad_norm": 2.896497964859009,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3999,
|
|
"step": 18150
|
|
},
|
|
{
|
|
"epoch": 8.87451171875,
|
|
"grad_norm": 2.759514570236206,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4459,
|
|
"step": 18175
|
|
},
|
|
{
|
|
"epoch": 8.88671875,
|
|
"grad_norm": 3.696629762649536,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.5002,
|
|
"step": 18200
|
|
},
|
|
{
|
|
"epoch": 8.89892578125,
|
|
"grad_norm": 2.6874115467071533,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4143,
|
|
"step": 18225
|
|
},
|
|
{
|
|
"epoch": 8.9111328125,
|
|
"grad_norm": 3.066502809524536,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3943,
|
|
"step": 18250
|
|
},
|
|
{
|
|
"epoch": 8.92333984375,
|
|
"grad_norm": 3.2830264568328857,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3878,
|
|
"step": 18275
|
|
},
|
|
{
|
|
"epoch": 8.935546875,
|
|
"grad_norm": 3.5021724700927734,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4332,
|
|
"step": 18300
|
|
},
|
|
{
|
|
"epoch": 8.94775390625,
|
|
"grad_norm": 4.337588310241699,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4256,
|
|
"step": 18325
|
|
},
|
|
{
|
|
"epoch": 8.9599609375,
|
|
"grad_norm": 3.243098258972168,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4593,
|
|
"step": 18350
|
|
},
|
|
{
|
|
"epoch": 8.97216796875,
|
|
"grad_norm": 2.650510787963867,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3951,
|
|
"step": 18375
|
|
},
|
|
{
|
|
"epoch": 8.984375,
|
|
"grad_norm": 3.37617564201355,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4143,
|
|
"step": 18400
|
|
},
|
|
{
|
|
"epoch": 8.99658203125,
|
|
"grad_norm": 3.7461116313934326,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4147,
|
|
"step": 18425
|
|
},
|
|
{
|
|
"epoch": 9.0087890625,
|
|
"grad_norm": 4.165828704833984,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4538,
|
|
"step": 18450
|
|
},
|
|
{
|
|
"epoch": 9.02099609375,
|
|
"grad_norm": 3.3768396377563477,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4128,
|
|
"step": 18475
|
|
},
|
|
{
|
|
"epoch": 9.033203125,
|
|
"grad_norm": 3.006622314453125,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.427,
|
|
"step": 18500
|
|
},
|
|
{
|
|
"epoch": 9.033203125,
|
|
"eval_cer": 21.054545454545455,
|
|
"eval_loss": 0.3800523281097412,
|
|
"eval_normalized_cer": 16.85162351006987,
|
|
"eval_runtime": 92.9904,
|
|
"eval_samples_per_second": 1.376,
|
|
"eval_steps_per_second": 0.086,
|
|
"step": 18500
|
|
},
|
|
{
|
|
"epoch": 9.04541015625,
|
|
"grad_norm": 2.577570915222168,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4233,
|
|
"step": 18525
|
|
},
|
|
{
|
|
"epoch": 9.0576171875,
|
|
"grad_norm": 3.2769463062286377,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4209,
|
|
"step": 18550
|
|
},
|
|
{
|
|
"epoch": 9.06982421875,
|
|
"grad_norm": 3.296905040740967,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4576,
|
|
"step": 18575
|
|
},
|
|
{
|
|
"epoch": 9.08203125,
|
|
"grad_norm": 2.849393367767334,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4033,
|
|
"step": 18600
|
|
},
|
|
{
|
|
"epoch": 9.09423828125,
|
|
"grad_norm": 4.293519496917725,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4604,
|
|
"step": 18625
|
|
},
|
|
{
|
|
"epoch": 9.1064453125,
|
|
"grad_norm": 3.5369253158569336,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.444,
|
|
"step": 18650
|
|
},
|
|
{
|
|
"epoch": 9.11865234375,
|
|
"grad_norm": 4.512628555297852,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4008,
|
|
"step": 18675
|
|
},
|
|
{
|
|
"epoch": 9.130859375,
|
|
"grad_norm": 3.0092594623565674,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4219,
|
|
"step": 18700
|
|
},
|
|
{
|
|
"epoch": 9.14306640625,
|
|
"grad_norm": 3.3797607421875,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4102,
|
|
"step": 18725
|
|
},
|
|
{
|
|
"epoch": 9.1552734375,
|
|
"grad_norm": 3.1966211795806885,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4152,
|
|
"step": 18750
|
|
},
|
|
{
|
|
"epoch": 9.16748046875,
|
|
"grad_norm": 3.016399383544922,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4277,
|
|
"step": 18775
|
|
},
|
|
{
|
|
"epoch": 9.1796875,
|
|
"grad_norm": 3.313685417175293,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3932,
|
|
"step": 18800
|
|
},
|
|
{
|
|
"epoch": 9.19189453125,
|
|
"grad_norm": 2.8252785205841064,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4041,
|
|
"step": 18825
|
|
},
|
|
{
|
|
"epoch": 9.2041015625,
|
|
"grad_norm": 2.935676097869873,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4107,
|
|
"step": 18850
|
|
},
|
|
{
|
|
"epoch": 9.21630859375,
|
|
"grad_norm": 2.280316114425659,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4701,
|
|
"step": 18875
|
|
},
|
|
{
|
|
"epoch": 9.228515625,
|
|
"grad_norm": 2.718478202819824,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4174,
|
|
"step": 18900
|
|
},
|
|
{
|
|
"epoch": 9.24072265625,
|
|
"grad_norm": 4.804378986358643,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3947,
|
|
"step": 18925
|
|
},
|
|
{
|
|
"epoch": 9.2529296875,
|
|
"grad_norm": 4.070915222167969,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4158,
|
|
"step": 18950
|
|
},
|
|
{
|
|
"epoch": 9.26513671875,
|
|
"grad_norm": 3.2516276836395264,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4379,
|
|
"step": 18975
|
|
},
|
|
{
|
|
"epoch": 9.27734375,
|
|
"grad_norm": 3.5097761154174805,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3767,
|
|
"step": 19000
|
|
},
|
|
{
|
|
"epoch": 9.27734375,
|
|
"eval_cer": 24.254545454545454,
|
|
"eval_loss": 0.36628666520118713,
|
|
"eval_normalized_cer": 18.701191944101932,
|
|
"eval_runtime": 93.7861,
|
|
"eval_samples_per_second": 1.365,
|
|
"eval_steps_per_second": 0.085,
|
|
"step": 19000
|
|
},
|
|
{
|
|
"epoch": 9.28955078125,
|
|
"grad_norm": 4.052979469299316,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.446,
|
|
"step": 19025
|
|
},
|
|
{
|
|
"epoch": 9.3017578125,
|
|
"grad_norm": 1.887149691581726,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3933,
|
|
"step": 19050
|
|
},
|
|
{
|
|
"epoch": 9.31396484375,
|
|
"grad_norm": 3.2341785430908203,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4361,
|
|
"step": 19075
|
|
},
|
|
{
|
|
"epoch": 9.326171875,
|
|
"grad_norm": 2.683950185775757,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4107,
|
|
"step": 19100
|
|
},
|
|
{
|
|
"epoch": 9.33837890625,
|
|
"grad_norm": 3.6661105155944824,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.396,
|
|
"step": 19125
|
|
},
|
|
{
|
|
"epoch": 9.3505859375,
|
|
"grad_norm": 2.379519462585449,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3875,
|
|
"step": 19150
|
|
},
|
|
{
|
|
"epoch": 9.36279296875,
|
|
"grad_norm": 2.952665090560913,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4471,
|
|
"step": 19175
|
|
},
|
|
{
|
|
"epoch": 9.375,
|
|
"grad_norm": 3.1039767265319824,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4052,
|
|
"step": 19200
|
|
},
|
|
{
|
|
"epoch": 9.38720703125,
|
|
"grad_norm": 3.941380023956299,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3908,
|
|
"step": 19225
|
|
},
|
|
{
|
|
"epoch": 9.3994140625,
|
|
"grad_norm": 2.511928081512451,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4215,
|
|
"step": 19250
|
|
},
|
|
{
|
|
"epoch": 9.41162109375,
|
|
"grad_norm": 2.6684021949768066,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.409,
|
|
"step": 19275
|
|
},
|
|
{
|
|
"epoch": 9.423828125,
|
|
"grad_norm": 3.9744958877563477,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4778,
|
|
"step": 19300
|
|
},
|
|
{
|
|
"epoch": 9.43603515625,
|
|
"grad_norm": 2.601891279220581,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4059,
|
|
"step": 19325
|
|
},
|
|
{
|
|
"epoch": 9.4482421875,
|
|
"grad_norm": 3.2740561962127686,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4299,
|
|
"step": 19350
|
|
},
|
|
{
|
|
"epoch": 9.46044921875,
|
|
"grad_norm": 4.9015889167785645,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4427,
|
|
"step": 19375
|
|
},
|
|
{
|
|
"epoch": 9.47265625,
|
|
"grad_norm": 2.882230281829834,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3959,
|
|
"step": 19400
|
|
},
|
|
{
|
|
"epoch": 9.48486328125,
|
|
"grad_norm": 4.404541969299316,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.414,
|
|
"step": 19425
|
|
},
|
|
{
|
|
"epoch": 9.4970703125,
|
|
"grad_norm": 3.5113275051116943,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3937,
|
|
"step": 19450
|
|
},
|
|
{
|
|
"epoch": 9.50927734375,
|
|
"grad_norm": 2.4990580081939697,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4262,
|
|
"step": 19475
|
|
},
|
|
{
|
|
"epoch": 9.521484375,
|
|
"grad_norm": 3.3028454780578613,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4519,
|
|
"step": 19500
|
|
},
|
|
{
|
|
"epoch": 9.521484375,
|
|
"eval_cer": 20.472727272727273,
|
|
"eval_loss": 0.37869974970817566,
|
|
"eval_normalized_cer": 15.454171804356761,
|
|
"eval_runtime": 97.3092,
|
|
"eval_samples_per_second": 1.315,
|
|
"eval_steps_per_second": 0.082,
|
|
"step": 19500
|
|
},
|
|
{
|
|
"epoch": 9.53369140625,
|
|
"grad_norm": 3.5164365768432617,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4119,
|
|
"step": 19525
|
|
},
|
|
{
|
|
"epoch": 9.5458984375,
|
|
"grad_norm": 3.049210786819458,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4198,
|
|
"step": 19550
|
|
},
|
|
{
|
|
"epoch": 9.55810546875,
|
|
"grad_norm": 2.659877300262451,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4047,
|
|
"step": 19575
|
|
},
|
|
{
|
|
"epoch": 9.5703125,
|
|
"grad_norm": 2.617429733276367,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.446,
|
|
"step": 19600
|
|
},
|
|
{
|
|
"epoch": 9.58251953125,
|
|
"grad_norm": 3.912100076675415,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4169,
|
|
"step": 19625
|
|
},
|
|
{
|
|
"epoch": 9.5947265625,
|
|
"grad_norm": 2.6492602825164795,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4215,
|
|
"step": 19650
|
|
},
|
|
{
|
|
"epoch": 9.60693359375,
|
|
"grad_norm": 3.404269218444824,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3782,
|
|
"step": 19675
|
|
},
|
|
{
|
|
"epoch": 9.619140625,
|
|
"grad_norm": 2.4062957763671875,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4342,
|
|
"step": 19700
|
|
},
|
|
{
|
|
"epoch": 9.63134765625,
|
|
"grad_norm": 2.835569143295288,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.372,
|
|
"step": 19725
|
|
},
|
|
{
|
|
"epoch": 9.6435546875,
|
|
"grad_norm": 2.3654327392578125,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4095,
|
|
"step": 19750
|
|
},
|
|
{
|
|
"epoch": 9.65576171875,
|
|
"grad_norm": 2.8574941158294678,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4225,
|
|
"step": 19775
|
|
},
|
|
{
|
|
"epoch": 9.66796875,
|
|
"grad_norm": 3.122192859649658,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4372,
|
|
"step": 19800
|
|
},
|
|
{
|
|
"epoch": 9.68017578125,
|
|
"grad_norm": 2.591271162033081,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4523,
|
|
"step": 19825
|
|
},
|
|
{
|
|
"epoch": 9.6923828125,
|
|
"grad_norm": 3.0938827991485596,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4055,
|
|
"step": 19850
|
|
},
|
|
{
|
|
"epoch": 9.70458984375,
|
|
"grad_norm": 3.643186569213867,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4066,
|
|
"step": 19875
|
|
},
|
|
{
|
|
"epoch": 9.716796875,
|
|
"grad_norm": 3.9704155921936035,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4615,
|
|
"step": 19900
|
|
},
|
|
{
|
|
"epoch": 9.72900390625,
|
|
"grad_norm": 2.1930134296417236,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4453,
|
|
"step": 19925
|
|
},
|
|
{
|
|
"epoch": 9.7412109375,
|
|
"grad_norm": 3.3762011528015137,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3818,
|
|
"step": 19950
|
|
},
|
|
{
|
|
"epoch": 9.75341796875,
|
|
"grad_norm": 3.172435998916626,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4118,
|
|
"step": 19975
|
|
},
|
|
{
|
|
"epoch": 9.765625,
|
|
"grad_norm": 3.6243398189544678,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4184,
|
|
"step": 20000
|
|
},
|
|
{
|
|
"epoch": 9.765625,
|
|
"eval_cer": 21.236363636363638,
|
|
"eval_loss": 0.37908729910850525,
|
|
"eval_normalized_cer": 16.85162351006987,
|
|
"eval_runtime": 97.9066,
|
|
"eval_samples_per_second": 1.307,
|
|
"eval_steps_per_second": 0.082,
|
|
"step": 20000
|
|
},
|
|
{
|
|
"epoch": 9.77783203125,
|
|
"grad_norm": 3.5656485557556152,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4162,
|
|
"step": 20025
|
|
},
|
|
{
|
|
"epoch": 9.7900390625,
|
|
"grad_norm": 3.304313898086548,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3687,
|
|
"step": 20050
|
|
},
|
|
{
|
|
"epoch": 9.80224609375,
|
|
"grad_norm": 2.4108333587646484,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.364,
|
|
"step": 20075
|
|
},
|
|
{
|
|
"epoch": 9.814453125,
|
|
"grad_norm": 3.621305465698242,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4071,
|
|
"step": 20100
|
|
},
|
|
{
|
|
"epoch": 9.82666015625,
|
|
"grad_norm": 3.3533174991607666,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4386,
|
|
"step": 20125
|
|
},
|
|
{
|
|
"epoch": 9.8388671875,
|
|
"grad_norm": 2.8671376705169678,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4299,
|
|
"step": 20150
|
|
},
|
|
{
|
|
"epoch": 9.85107421875,
|
|
"grad_norm": 3.2498745918273926,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4011,
|
|
"step": 20175
|
|
},
|
|
{
|
|
"epoch": 9.86328125,
|
|
"grad_norm": 3.473806381225586,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.441,
|
|
"step": 20200
|
|
},
|
|
{
|
|
"epoch": 9.87548828125,
|
|
"grad_norm": 2.80523419380188,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4404,
|
|
"step": 20225
|
|
},
|
|
{
|
|
"epoch": 9.8876953125,
|
|
"grad_norm": 2.332698345184326,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4269,
|
|
"step": 20250
|
|
},
|
|
{
|
|
"epoch": 9.89990234375,
|
|
"grad_norm": 2.8714382648468018,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4263,
|
|
"step": 20275
|
|
},
|
|
{
|
|
"epoch": 9.912109375,
|
|
"grad_norm": 3.2583768367767334,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4206,
|
|
"step": 20300
|
|
},
|
|
{
|
|
"epoch": 9.92431640625,
|
|
"grad_norm": 3.4748706817626953,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4348,
|
|
"step": 20325
|
|
},
|
|
{
|
|
"epoch": 9.9365234375,
|
|
"grad_norm": 2.696352243423462,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4237,
|
|
"step": 20350
|
|
},
|
|
{
|
|
"epoch": 9.94873046875,
|
|
"grad_norm": 3.5071768760681152,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4521,
|
|
"step": 20375
|
|
},
|
|
{
|
|
"epoch": 9.9609375,
|
|
"grad_norm": 3.7230043411254883,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4387,
|
|
"step": 20400
|
|
},
|
|
{
|
|
"epoch": 9.97314453125,
|
|
"grad_norm": 4.502689361572266,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4303,
|
|
"step": 20425
|
|
},
|
|
{
|
|
"epoch": 9.9853515625,
|
|
"grad_norm": 2.849353790283203,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3952,
|
|
"step": 20450
|
|
},
|
|
{
|
|
"epoch": 9.99755859375,
|
|
"grad_norm": 3.3098607063293457,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3903,
|
|
"step": 20475
|
|
},
|
|
{
|
|
"epoch": 10.009765625,
|
|
"grad_norm": 2.970693349838257,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3752,
|
|
"step": 20500
|
|
},
|
|
{
|
|
"epoch": 10.009765625,
|
|
"eval_cer": 24.581818181818182,
|
|
"eval_loss": 0.3705242872238159,
|
|
"eval_normalized_cer": 19.31771475544595,
|
|
"eval_runtime": 88.6677,
|
|
"eval_samples_per_second": 1.444,
|
|
"eval_steps_per_second": 0.09,
|
|
"step": 20500
|
|
},
|
|
{
|
|
"epoch": 10.02197265625,
|
|
"grad_norm": 2.900347948074341,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4193,
|
|
"step": 20525
|
|
},
|
|
{
|
|
"epoch": 10.0341796875,
|
|
"grad_norm": 2.864896297454834,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4096,
|
|
"step": 20550
|
|
},
|
|
{
|
|
"epoch": 10.04638671875,
|
|
"grad_norm": 2.738795518875122,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4447,
|
|
"step": 20575
|
|
},
|
|
{
|
|
"epoch": 10.05859375,
|
|
"grad_norm": 2.6308414936065674,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3794,
|
|
"step": 20600
|
|
},
|
|
{
|
|
"epoch": 10.07080078125,
|
|
"grad_norm": 2.4290077686309814,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4009,
|
|
"step": 20625
|
|
},
|
|
{
|
|
"epoch": 10.0830078125,
|
|
"grad_norm": 2.612316846847534,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3893,
|
|
"step": 20650
|
|
},
|
|
{
|
|
"epoch": 10.09521484375,
|
|
"grad_norm": 3.2632391452789307,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3999,
|
|
"step": 20675
|
|
},
|
|
{
|
|
"epoch": 10.107421875,
|
|
"grad_norm": 2.5368306636810303,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4016,
|
|
"step": 20700
|
|
},
|
|
{
|
|
"epoch": 10.11962890625,
|
|
"grad_norm": 3.5920157432556152,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4465,
|
|
"step": 20725
|
|
},
|
|
{
|
|
"epoch": 10.1318359375,
|
|
"grad_norm": 2.3094773292541504,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4074,
|
|
"step": 20750
|
|
},
|
|
{
|
|
"epoch": 10.14404296875,
|
|
"grad_norm": 2.4848082065582275,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.381,
|
|
"step": 20775
|
|
},
|
|
{
|
|
"epoch": 10.15625,
|
|
"grad_norm": 3.432950496673584,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3846,
|
|
"step": 20800
|
|
},
|
|
{
|
|
"epoch": 10.16845703125,
|
|
"grad_norm": 3.441004514694214,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4343,
|
|
"step": 20825
|
|
},
|
|
{
|
|
"epoch": 10.1806640625,
|
|
"grad_norm": 1.9125665426254272,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3822,
|
|
"step": 20850
|
|
},
|
|
{
|
|
"epoch": 10.19287109375,
|
|
"grad_norm": 2.5689966678619385,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4069,
|
|
"step": 20875
|
|
},
|
|
{
|
|
"epoch": 10.205078125,
|
|
"grad_norm": 2.4284005165100098,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4481,
|
|
"step": 20900
|
|
},
|
|
{
|
|
"epoch": 10.21728515625,
|
|
"grad_norm": 2.9357855319976807,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3985,
|
|
"step": 20925
|
|
},
|
|
{
|
|
"epoch": 10.2294921875,
|
|
"grad_norm": 3.1418330669403076,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4334,
|
|
"step": 20950
|
|
},
|
|
{
|
|
"epoch": 10.24169921875,
|
|
"grad_norm": 3.525568723678589,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4659,
|
|
"step": 20975
|
|
},
|
|
{
|
|
"epoch": 10.25390625,
|
|
"grad_norm": 3.033616304397583,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4532,
|
|
"step": 21000
|
|
},
|
|
{
|
|
"epoch": 10.25390625,
|
|
"eval_cer": 22.363636363636363,
|
|
"eval_loss": 0.3776665925979614,
|
|
"eval_normalized_cer": 18.372379778051787,
|
|
"eval_runtime": 88.8963,
|
|
"eval_samples_per_second": 1.44,
|
|
"eval_steps_per_second": 0.09,
|
|
"step": 21000
|
|
},
|
|
{
|
|
"epoch": 10.26611328125,
|
|
"grad_norm": 3.0008301734924316,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4229,
|
|
"step": 21025
|
|
},
|
|
{
|
|
"epoch": 10.2783203125,
|
|
"grad_norm": 4.1073102951049805,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.43,
|
|
"step": 21050
|
|
},
|
|
{
|
|
"epoch": 10.29052734375,
|
|
"grad_norm": 2.6684110164642334,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4007,
|
|
"step": 21075
|
|
},
|
|
{
|
|
"epoch": 10.302734375,
|
|
"grad_norm": 2.416282892227173,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4006,
|
|
"step": 21100
|
|
},
|
|
{
|
|
"epoch": 10.31494140625,
|
|
"grad_norm": 3.2998218536376953,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4207,
|
|
"step": 21125
|
|
},
|
|
{
|
|
"epoch": 10.3271484375,
|
|
"grad_norm": 4.176929473876953,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4559,
|
|
"step": 21150
|
|
},
|
|
{
|
|
"epoch": 10.33935546875,
|
|
"grad_norm": 2.522796154022217,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4052,
|
|
"step": 21175
|
|
},
|
|
{
|
|
"epoch": 10.3515625,
|
|
"grad_norm": 2.9196386337280273,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.414,
|
|
"step": 21200
|
|
},
|
|
{
|
|
"epoch": 10.36376953125,
|
|
"grad_norm": 2.877315044403076,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4454,
|
|
"step": 21225
|
|
},
|
|
{
|
|
"epoch": 10.3759765625,
|
|
"grad_norm": 3.0172317028045654,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4294,
|
|
"step": 21250
|
|
},
|
|
{
|
|
"epoch": 10.38818359375,
|
|
"grad_norm": 3.5081894397735596,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4283,
|
|
"step": 21275
|
|
},
|
|
{
|
|
"epoch": 10.400390625,
|
|
"grad_norm": 2.136151075363159,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.429,
|
|
"step": 21300
|
|
},
|
|
{
|
|
"epoch": 10.41259765625,
|
|
"grad_norm": 3.0397658348083496,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4019,
|
|
"step": 21325
|
|
},
|
|
{
|
|
"epoch": 10.4248046875,
|
|
"grad_norm": 3.9129014015197754,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4128,
|
|
"step": 21350
|
|
},
|
|
{
|
|
"epoch": 10.43701171875,
|
|
"grad_norm": 3.1991934776306152,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4,
|
|
"step": 21375
|
|
},
|
|
{
|
|
"epoch": 10.44921875,
|
|
"grad_norm": 2.850768804550171,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4492,
|
|
"step": 21400
|
|
},
|
|
{
|
|
"epoch": 10.46142578125,
|
|
"grad_norm": 3.6122255325317383,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3692,
|
|
"step": 21425
|
|
},
|
|
{
|
|
"epoch": 10.4736328125,
|
|
"grad_norm": 3.7408318519592285,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4239,
|
|
"step": 21450
|
|
},
|
|
{
|
|
"epoch": 10.48583984375,
|
|
"grad_norm": 2.4114573001861572,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3595,
|
|
"step": 21475
|
|
},
|
|
{
|
|
"epoch": 10.498046875,
|
|
"grad_norm": 3.060419797897339,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3665,
|
|
"step": 21500
|
|
},
|
|
{
|
|
"epoch": 10.498046875,
|
|
"eval_cer": 20.945454545454545,
|
|
"eval_loss": 0.3751263916492462,
|
|
"eval_normalized_cer": 17.13933415536375,
|
|
"eval_runtime": 91.0253,
|
|
"eval_samples_per_second": 1.406,
|
|
"eval_steps_per_second": 0.088,
|
|
"step": 21500
|
|
},
|
|
{
|
|
"epoch": 10.51025390625,
|
|
"grad_norm": 2.921161413192749,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4385,
|
|
"step": 21525
|
|
},
|
|
{
|
|
"epoch": 10.5224609375,
|
|
"grad_norm": 3.127946376800537,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4185,
|
|
"step": 21550
|
|
},
|
|
{
|
|
"epoch": 10.53466796875,
|
|
"grad_norm": 2.4071595668792725,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4172,
|
|
"step": 21575
|
|
},
|
|
{
|
|
"epoch": 10.546875,
|
|
"grad_norm": 3.3871352672576904,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3781,
|
|
"step": 21600
|
|
},
|
|
{
|
|
"epoch": 10.55908203125,
|
|
"grad_norm": 2.465686082839966,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4021,
|
|
"step": 21625
|
|
},
|
|
{
|
|
"epoch": 10.5712890625,
|
|
"grad_norm": 3.5105502605438232,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3874,
|
|
"step": 21650
|
|
},
|
|
{
|
|
"epoch": 10.58349609375,
|
|
"grad_norm": 3.2150614261627197,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4203,
|
|
"step": 21675
|
|
},
|
|
{
|
|
"epoch": 10.595703125,
|
|
"grad_norm": 2.755958080291748,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3961,
|
|
"step": 21700
|
|
},
|
|
{
|
|
"epoch": 10.60791015625,
|
|
"grad_norm": 2.6068344116210938,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.425,
|
|
"step": 21725
|
|
},
|
|
{
|
|
"epoch": 10.6201171875,
|
|
"grad_norm": 4.860282897949219,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4188,
|
|
"step": 21750
|
|
},
|
|
{
|
|
"epoch": 10.63232421875,
|
|
"grad_norm": 3.154463529586792,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4108,
|
|
"step": 21775
|
|
},
|
|
{
|
|
"epoch": 10.64453125,
|
|
"grad_norm": 3.5040194988250732,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4295,
|
|
"step": 21800
|
|
},
|
|
{
|
|
"epoch": 10.65673828125,
|
|
"grad_norm": 3.5046403408050537,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4306,
|
|
"step": 21825
|
|
},
|
|
{
|
|
"epoch": 10.6689453125,
|
|
"grad_norm": 3.5825250148773193,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3991,
|
|
"step": 21850
|
|
},
|
|
{
|
|
"epoch": 10.68115234375,
|
|
"grad_norm": 3.1898937225341797,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4365,
|
|
"step": 21875
|
|
},
|
|
{
|
|
"epoch": 10.693359375,
|
|
"grad_norm": 2.6460320949554443,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4053,
|
|
"step": 21900
|
|
},
|
|
{
|
|
"epoch": 10.70556640625,
|
|
"grad_norm": 2.7370963096618652,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3961,
|
|
"step": 21925
|
|
},
|
|
{
|
|
"epoch": 10.7177734375,
|
|
"grad_norm": 2.502699851989746,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3943,
|
|
"step": 21950
|
|
},
|
|
{
|
|
"epoch": 10.72998046875,
|
|
"grad_norm": 3.806084632873535,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.424,
|
|
"step": 21975
|
|
},
|
|
{
|
|
"epoch": 10.7421875,
|
|
"grad_norm": 2.743624687194824,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3897,
|
|
"step": 22000
|
|
},
|
|
{
|
|
"epoch": 10.7421875,
|
|
"eval_cer": 22.763636363636365,
|
|
"eval_loss": 0.3785582482814789,
|
|
"eval_normalized_cer": 18.53678586107686,
|
|
"eval_runtime": 89.903,
|
|
"eval_samples_per_second": 1.424,
|
|
"eval_steps_per_second": 0.089,
|
|
"step": 22000
|
|
},
|
|
{
|
|
"epoch": 10.75439453125,
|
|
"grad_norm": 3.1976799964904785,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.439,
|
|
"step": 22025
|
|
},
|
|
{
|
|
"epoch": 10.7666015625,
|
|
"grad_norm": 2.4848337173461914,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4722,
|
|
"step": 22050
|
|
},
|
|
{
|
|
"epoch": 10.77880859375,
|
|
"grad_norm": 2.554133176803589,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4607,
|
|
"step": 22075
|
|
},
|
|
{
|
|
"epoch": 10.791015625,
|
|
"grad_norm": 3.2724106311798096,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.429,
|
|
"step": 22100
|
|
},
|
|
{
|
|
"epoch": 10.80322265625,
|
|
"grad_norm": 2.920471668243408,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3968,
|
|
"step": 22125
|
|
},
|
|
{
|
|
"epoch": 10.8154296875,
|
|
"grad_norm": 3.323936700820923,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4126,
|
|
"step": 22150
|
|
},
|
|
{
|
|
"epoch": 10.82763671875,
|
|
"grad_norm": 3.5896575450897217,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3986,
|
|
"step": 22175
|
|
},
|
|
{
|
|
"epoch": 10.83984375,
|
|
"grad_norm": 3.3386435508728027,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3979,
|
|
"step": 22200
|
|
},
|
|
{
|
|
"epoch": 10.85205078125,
|
|
"grad_norm": 3.4217638969421387,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3928,
|
|
"step": 22225
|
|
},
|
|
{
|
|
"epoch": 10.8642578125,
|
|
"grad_norm": 2.90092134475708,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3813,
|
|
"step": 22250
|
|
},
|
|
{
|
|
"epoch": 10.87646484375,
|
|
"grad_norm": 2.9017395973205566,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4061,
|
|
"step": 22275
|
|
},
|
|
{
|
|
"epoch": 10.888671875,
|
|
"grad_norm": 3.1124579906463623,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3933,
|
|
"step": 22300
|
|
},
|
|
{
|
|
"epoch": 10.90087890625,
|
|
"grad_norm": 2.763261079788208,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3751,
|
|
"step": 22325
|
|
},
|
|
{
|
|
"epoch": 10.9130859375,
|
|
"grad_norm": 3.8521738052368164,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4537,
|
|
"step": 22350
|
|
},
|
|
{
|
|
"epoch": 10.92529296875,
|
|
"grad_norm": 3.3494768142700195,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3479,
|
|
"step": 22375
|
|
},
|
|
{
|
|
"epoch": 10.9375,
|
|
"grad_norm": 3.752168655395508,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4393,
|
|
"step": 22400
|
|
},
|
|
{
|
|
"epoch": 10.94970703125,
|
|
"grad_norm": 2.980929374694824,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4376,
|
|
"step": 22425
|
|
},
|
|
{
|
|
"epoch": 10.9619140625,
|
|
"grad_norm": 2.931426525115967,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4006,
|
|
"step": 22450
|
|
},
|
|
{
|
|
"epoch": 10.97412109375,
|
|
"grad_norm": 2.798682451248169,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3977,
|
|
"step": 22475
|
|
},
|
|
{
|
|
"epoch": 10.986328125,
|
|
"grad_norm": 3.374363422393799,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3846,
|
|
"step": 22500
|
|
},
|
|
{
|
|
"epoch": 10.986328125,
|
|
"eval_cer": 21.12727272727273,
|
|
"eval_loss": 0.3707554340362549,
|
|
"eval_normalized_cer": 16.93382655158241,
|
|
"eval_runtime": 90.713,
|
|
"eval_samples_per_second": 1.411,
|
|
"eval_steps_per_second": 0.088,
|
|
"step": 22500
|
|
},
|
|
{
|
|
"epoch": 10.99853515625,
|
|
"grad_norm": 2.9657630920410156,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4397,
|
|
"step": 22525
|
|
},
|
|
{
|
|
"epoch": 11.0107421875,
|
|
"grad_norm": 3.373459577560425,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4813,
|
|
"step": 22550
|
|
},
|
|
{
|
|
"epoch": 11.02294921875,
|
|
"grad_norm": 2.293248176574707,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3939,
|
|
"step": 22575
|
|
},
|
|
{
|
|
"epoch": 11.03515625,
|
|
"grad_norm": 3.1755404472351074,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4173,
|
|
"step": 22600
|
|
},
|
|
{
|
|
"epoch": 11.04736328125,
|
|
"grad_norm": 4.223100185394287,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3965,
|
|
"step": 22625
|
|
},
|
|
{
|
|
"epoch": 11.0595703125,
|
|
"grad_norm": 4.098690986633301,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4371,
|
|
"step": 22650
|
|
},
|
|
{
|
|
"epoch": 11.07177734375,
|
|
"grad_norm": 2.937412977218628,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4287,
|
|
"step": 22675
|
|
},
|
|
{
|
|
"epoch": 11.083984375,
|
|
"grad_norm": 2.240269422531128,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4255,
|
|
"step": 22700
|
|
},
|
|
{
|
|
"epoch": 11.09619140625,
|
|
"grad_norm": 3.1162662506103516,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4157,
|
|
"step": 22725
|
|
},
|
|
{
|
|
"epoch": 11.1083984375,
|
|
"grad_norm": 2.875746726989746,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3941,
|
|
"step": 22750
|
|
},
|
|
{
|
|
"epoch": 11.12060546875,
|
|
"grad_norm": 2.4718353748321533,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3975,
|
|
"step": 22775
|
|
},
|
|
{
|
|
"epoch": 11.1328125,
|
|
"grad_norm": 3.1226015090942383,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4033,
|
|
"step": 22800
|
|
},
|
|
{
|
|
"epoch": 11.14501953125,
|
|
"grad_norm": 3.384474515914917,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4701,
|
|
"step": 22825
|
|
},
|
|
{
|
|
"epoch": 11.1572265625,
|
|
"grad_norm": 3.434800148010254,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.404,
|
|
"step": 22850
|
|
},
|
|
{
|
|
"epoch": 11.16943359375,
|
|
"grad_norm": 2.6356847286224365,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.408,
|
|
"step": 22875
|
|
},
|
|
{
|
|
"epoch": 11.181640625,
|
|
"grad_norm": 2.8493409156799316,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3589,
|
|
"step": 22900
|
|
},
|
|
{
|
|
"epoch": 11.19384765625,
|
|
"grad_norm": 3.3821358680725098,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4357,
|
|
"step": 22925
|
|
},
|
|
{
|
|
"epoch": 11.2060546875,
|
|
"grad_norm": 2.6961047649383545,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3989,
|
|
"step": 22950
|
|
},
|
|
{
|
|
"epoch": 11.21826171875,
|
|
"grad_norm": 2.9927561283111572,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4523,
|
|
"step": 22975
|
|
},
|
|
{
|
|
"epoch": 11.23046875,
|
|
"grad_norm": 3.8990068435668945,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4638,
|
|
"step": 23000
|
|
},
|
|
{
|
|
"epoch": 11.23046875,
|
|
"eval_cer": 20.545454545454543,
|
|
"eval_loss": 0.3684229254722595,
|
|
"eval_normalized_cer": 16.974928072338678,
|
|
"eval_runtime": 89.0848,
|
|
"eval_samples_per_second": 1.437,
|
|
"eval_steps_per_second": 0.09,
|
|
"step": 23000
|
|
},
|
|
{
|
|
"epoch": 11.24267578125,
|
|
"grad_norm": 3.1418068408966064,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3857,
|
|
"step": 23025
|
|
},
|
|
{
|
|
"epoch": 11.2548828125,
|
|
"grad_norm": 3.578172445297241,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4817,
|
|
"step": 23050
|
|
},
|
|
{
|
|
"epoch": 11.26708984375,
|
|
"grad_norm": 3.6145339012145996,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3732,
|
|
"step": 23075
|
|
},
|
|
{
|
|
"epoch": 11.279296875,
|
|
"grad_norm": 3.0317516326904297,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.429,
|
|
"step": 23100
|
|
},
|
|
{
|
|
"epoch": 11.29150390625,
|
|
"grad_norm": 3.6998438835144043,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4374,
|
|
"step": 23125
|
|
},
|
|
{
|
|
"epoch": 11.3037109375,
|
|
"grad_norm": 3.1519877910614014,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4285,
|
|
"step": 23150
|
|
},
|
|
{
|
|
"epoch": 11.31591796875,
|
|
"grad_norm": 2.30660343170166,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4427,
|
|
"step": 23175
|
|
},
|
|
{
|
|
"epoch": 11.328125,
|
|
"grad_norm": 2.7679004669189453,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4266,
|
|
"step": 23200
|
|
},
|
|
{
|
|
"epoch": 11.34033203125,
|
|
"grad_norm": 4.154956340789795,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3949,
|
|
"step": 23225
|
|
},
|
|
{
|
|
"epoch": 11.3525390625,
|
|
"grad_norm": 3.537022113800049,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4484,
|
|
"step": 23250
|
|
},
|
|
{
|
|
"epoch": 11.36474609375,
|
|
"grad_norm": 3.042243242263794,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4401,
|
|
"step": 23275
|
|
},
|
|
{
|
|
"epoch": 11.376953125,
|
|
"grad_norm": 3.82000994682312,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3968,
|
|
"step": 23300
|
|
},
|
|
{
|
|
"epoch": 11.38916015625,
|
|
"grad_norm": 2.052567958831787,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3856,
|
|
"step": 23325
|
|
},
|
|
{
|
|
"epoch": 11.4013671875,
|
|
"grad_norm": 3.497328758239746,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3915,
|
|
"step": 23350
|
|
},
|
|
{
|
|
"epoch": 11.41357421875,
|
|
"grad_norm": 3.023749589920044,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4381,
|
|
"step": 23375
|
|
},
|
|
{
|
|
"epoch": 11.42578125,
|
|
"grad_norm": 3.8763065338134766,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3879,
|
|
"step": 23400
|
|
},
|
|
{
|
|
"epoch": 11.43798828125,
|
|
"grad_norm": 2.346403121948242,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4255,
|
|
"step": 23425
|
|
},
|
|
{
|
|
"epoch": 11.4501953125,
|
|
"grad_norm": 2.7521772384643555,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4073,
|
|
"step": 23450
|
|
},
|
|
{
|
|
"epoch": 11.46240234375,
|
|
"grad_norm": 2.2675764560699463,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3837,
|
|
"step": 23475
|
|
},
|
|
{
|
|
"epoch": 11.474609375,
|
|
"grad_norm": 3.2309587001800537,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4423,
|
|
"step": 23500
|
|
},
|
|
{
|
|
"epoch": 11.474609375,
|
|
"eval_cer": 21.454545454545453,
|
|
"eval_loss": 0.3692309558391571,
|
|
"eval_normalized_cer": 16.892725030826142,
|
|
"eval_runtime": 91.354,
|
|
"eval_samples_per_second": 1.401,
|
|
"eval_steps_per_second": 0.088,
|
|
"step": 23500
|
|
},
|
|
{
|
|
"epoch": 11.48681640625,
|
|
"grad_norm": 3.551201581954956,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3808,
|
|
"step": 23525
|
|
},
|
|
{
|
|
"epoch": 11.4990234375,
|
|
"grad_norm": 3.828253984451294,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4081,
|
|
"step": 23550
|
|
},
|
|
{
|
|
"epoch": 11.51123046875,
|
|
"grad_norm": 3.3823349475860596,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3875,
|
|
"step": 23575
|
|
},
|
|
{
|
|
"epoch": 11.5234375,
|
|
"grad_norm": 3.332829713821411,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4167,
|
|
"step": 23600
|
|
},
|
|
{
|
|
"epoch": 11.53564453125,
|
|
"grad_norm": 2.943939685821533,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4397,
|
|
"step": 23625
|
|
},
|
|
{
|
|
"epoch": 11.5478515625,
|
|
"grad_norm": 3.0470168590545654,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4395,
|
|
"step": 23650
|
|
},
|
|
{
|
|
"epoch": 11.56005859375,
|
|
"grad_norm": 3.1324830055236816,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.391,
|
|
"step": 23675
|
|
},
|
|
{
|
|
"epoch": 11.572265625,
|
|
"grad_norm": 2.959723711013794,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4063,
|
|
"step": 23700
|
|
},
|
|
{
|
|
"epoch": 11.58447265625,
|
|
"grad_norm": 3.0424516201019287,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.457,
|
|
"step": 23725
|
|
},
|
|
{
|
|
"epoch": 11.5966796875,
|
|
"grad_norm": 2.7524912357330322,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3916,
|
|
"step": 23750
|
|
},
|
|
{
|
|
"epoch": 11.60888671875,
|
|
"grad_norm": 3.8842151165008545,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4134,
|
|
"step": 23775
|
|
},
|
|
{
|
|
"epoch": 11.62109375,
|
|
"grad_norm": 2.5505778789520264,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4301,
|
|
"step": 23800
|
|
},
|
|
{
|
|
"epoch": 11.63330078125,
|
|
"grad_norm": 2.3430569171905518,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3913,
|
|
"step": 23825
|
|
},
|
|
{
|
|
"epoch": 11.6455078125,
|
|
"grad_norm": 3.5037317276000977,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4213,
|
|
"step": 23850
|
|
},
|
|
{
|
|
"epoch": 11.65771484375,
|
|
"grad_norm": 3.067307710647583,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.395,
|
|
"step": 23875
|
|
},
|
|
{
|
|
"epoch": 11.669921875,
|
|
"grad_norm": 3.3112151622772217,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.429,
|
|
"step": 23900
|
|
},
|
|
{
|
|
"epoch": 11.68212890625,
|
|
"grad_norm": 3.318018913269043,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4593,
|
|
"step": 23925
|
|
},
|
|
{
|
|
"epoch": 11.6943359375,
|
|
"grad_norm": 2.73763370513916,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3565,
|
|
"step": 23950
|
|
},
|
|
{
|
|
"epoch": 11.70654296875,
|
|
"grad_norm": 4.161790370941162,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4181,
|
|
"step": 23975
|
|
},
|
|
{
|
|
"epoch": 11.71875,
|
|
"grad_norm": 2.161367893218994,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3991,
|
|
"step": 24000
|
|
},
|
|
{
|
|
"epoch": 11.71875,
|
|
"eval_cer": 20.763636363636365,
|
|
"eval_loss": 0.37048545479774475,
|
|
"eval_normalized_cer": 16.769420468557335,
|
|
"eval_runtime": 97.2817,
|
|
"eval_samples_per_second": 1.316,
|
|
"eval_steps_per_second": 0.082,
|
|
"step": 24000
|
|
},
|
|
{
|
|
"epoch": 11.73095703125,
|
|
"grad_norm": 2.165503978729248,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3863,
|
|
"step": 24025
|
|
},
|
|
{
|
|
"epoch": 11.7431640625,
|
|
"grad_norm": 3.243612289428711,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4415,
|
|
"step": 24050
|
|
},
|
|
{
|
|
"epoch": 11.75537109375,
|
|
"grad_norm": 3.0755250453948975,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4233,
|
|
"step": 24075
|
|
},
|
|
{
|
|
"epoch": 11.767578125,
|
|
"grad_norm": 3.3624966144561768,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3975,
|
|
"step": 24100
|
|
},
|
|
{
|
|
"epoch": 11.77978515625,
|
|
"grad_norm": 3.3429675102233887,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4349,
|
|
"step": 24125
|
|
},
|
|
{
|
|
"epoch": 11.7919921875,
|
|
"grad_norm": 2.686250925064087,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3938,
|
|
"step": 24150
|
|
},
|
|
{
|
|
"epoch": 11.80419921875,
|
|
"grad_norm": 2.747789144515991,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4164,
|
|
"step": 24175
|
|
},
|
|
{
|
|
"epoch": 11.81640625,
|
|
"grad_norm": 3.79379940032959,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4195,
|
|
"step": 24200
|
|
},
|
|
{
|
|
"epoch": 11.82861328125,
|
|
"grad_norm": 3.6123952865600586,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.399,
|
|
"step": 24225
|
|
},
|
|
{
|
|
"epoch": 11.8408203125,
|
|
"grad_norm": 3.1555302143096924,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4238,
|
|
"step": 24250
|
|
},
|
|
{
|
|
"epoch": 11.85302734375,
|
|
"grad_norm": 2.9324278831481934,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4096,
|
|
"step": 24275
|
|
},
|
|
{
|
|
"epoch": 11.865234375,
|
|
"grad_norm": 2.907283067703247,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4128,
|
|
"step": 24300
|
|
},
|
|
{
|
|
"epoch": 11.87744140625,
|
|
"grad_norm": 3.6261141300201416,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4398,
|
|
"step": 24325
|
|
},
|
|
{
|
|
"epoch": 11.8896484375,
|
|
"grad_norm": 4.799058437347412,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4489,
|
|
"step": 24350
|
|
},
|
|
{
|
|
"epoch": 11.90185546875,
|
|
"grad_norm": 2.681626558303833,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4308,
|
|
"step": 24375
|
|
},
|
|
{
|
|
"epoch": 11.9140625,
|
|
"grad_norm": 2.723325729370117,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3909,
|
|
"step": 24400
|
|
},
|
|
{
|
|
"epoch": 11.92626953125,
|
|
"grad_norm": 4.047337055206299,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3749,
|
|
"step": 24425
|
|
},
|
|
{
|
|
"epoch": 11.9384765625,
|
|
"grad_norm": 3.933332920074463,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4232,
|
|
"step": 24450
|
|
},
|
|
{
|
|
"epoch": 11.95068359375,
|
|
"grad_norm": 2.8435702323913574,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4071,
|
|
"step": 24475
|
|
},
|
|
{
|
|
"epoch": 11.962890625,
|
|
"grad_norm": 3.180521249771118,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4023,
|
|
"step": 24500
|
|
},
|
|
{
|
|
"epoch": 11.962890625,
|
|
"eval_cer": 18.436363636363637,
|
|
"eval_loss": 0.36682504415512085,
|
|
"eval_normalized_cer": 14.26222770242499,
|
|
"eval_runtime": 89.1047,
|
|
"eval_samples_per_second": 1.437,
|
|
"eval_steps_per_second": 0.09,
|
|
"step": 24500
|
|
},
|
|
{
|
|
"epoch": 11.97509765625,
|
|
"grad_norm": 2.902740716934204,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4535,
|
|
"step": 24525
|
|
},
|
|
{
|
|
"epoch": 11.9873046875,
|
|
"grad_norm": 2.6517724990844727,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4093,
|
|
"step": 24550
|
|
},
|
|
{
|
|
"epoch": 11.99951171875,
|
|
"grad_norm": 3.2232935428619385,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4295,
|
|
"step": 24575
|
|
},
|
|
{
|
|
"epoch": 12.01171875,
|
|
"grad_norm": 1.8525205850601196,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4109,
|
|
"step": 24600
|
|
},
|
|
{
|
|
"epoch": 12.02392578125,
|
|
"grad_norm": 2.505067825317383,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3659,
|
|
"step": 24625
|
|
},
|
|
{
|
|
"epoch": 12.0361328125,
|
|
"grad_norm": 2.921861410140991,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.387,
|
|
"step": 24650
|
|
},
|
|
{
|
|
"epoch": 12.04833984375,
|
|
"grad_norm": 3.4830503463745117,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4036,
|
|
"step": 24675
|
|
},
|
|
{
|
|
"epoch": 12.060546875,
|
|
"grad_norm": 3.166236400604248,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4014,
|
|
"step": 24700
|
|
},
|
|
{
|
|
"epoch": 12.07275390625,
|
|
"grad_norm": 2.9270448684692383,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3983,
|
|
"step": 24725
|
|
},
|
|
{
|
|
"epoch": 12.0849609375,
|
|
"grad_norm": 3.857215166091919,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4343,
|
|
"step": 24750
|
|
},
|
|
{
|
|
"epoch": 12.09716796875,
|
|
"grad_norm": 3.039315700531006,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4095,
|
|
"step": 24775
|
|
},
|
|
{
|
|
"epoch": 12.109375,
|
|
"grad_norm": 4.983880043029785,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4108,
|
|
"step": 24800
|
|
},
|
|
{
|
|
"epoch": 12.12158203125,
|
|
"grad_norm": 3.1066694259643555,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.421,
|
|
"step": 24825
|
|
},
|
|
{
|
|
"epoch": 12.1337890625,
|
|
"grad_norm": 3.10778546333313,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4243,
|
|
"step": 24850
|
|
},
|
|
{
|
|
"epoch": 12.14599609375,
|
|
"grad_norm": 2.57086443901062,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4049,
|
|
"step": 24875
|
|
},
|
|
{
|
|
"epoch": 12.158203125,
|
|
"grad_norm": 2.7861294746398926,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3917,
|
|
"step": 24900
|
|
},
|
|
{
|
|
"epoch": 12.17041015625,
|
|
"grad_norm": 3.35518217086792,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4332,
|
|
"step": 24925
|
|
},
|
|
{
|
|
"epoch": 12.1826171875,
|
|
"grad_norm": 2.97426176071167,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4148,
|
|
"step": 24950
|
|
},
|
|
{
|
|
"epoch": 12.19482421875,
|
|
"grad_norm": 2.733003854751587,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3901,
|
|
"step": 24975
|
|
},
|
|
{
|
|
"epoch": 12.20703125,
|
|
"grad_norm": 2.4438769817352295,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4039,
|
|
"step": 25000
|
|
},
|
|
{
|
|
"epoch": 12.20703125,
|
|
"eval_cer": 20.763636363636365,
|
|
"eval_loss": 0.3740121126174927,
|
|
"eval_normalized_cer": 15.454171804356761,
|
|
"eval_runtime": 88.8688,
|
|
"eval_samples_per_second": 1.44,
|
|
"eval_steps_per_second": 0.09,
|
|
"step": 25000
|
|
},
|
|
{
|
|
"epoch": 12.21923828125,
|
|
"grad_norm": 2.897085666656494,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4155,
|
|
"step": 25025
|
|
},
|
|
{
|
|
"epoch": 12.2314453125,
|
|
"grad_norm": 3.21028470993042,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4249,
|
|
"step": 25050
|
|
},
|
|
{
|
|
"epoch": 12.24365234375,
|
|
"grad_norm": 2.8493075370788574,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4279,
|
|
"step": 25075
|
|
},
|
|
{
|
|
"epoch": 12.255859375,
|
|
"grad_norm": 3.8995752334594727,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4169,
|
|
"step": 25100
|
|
},
|
|
{
|
|
"epoch": 12.26806640625,
|
|
"grad_norm": 3.4920387268066406,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4164,
|
|
"step": 25125
|
|
},
|
|
{
|
|
"epoch": 12.2802734375,
|
|
"grad_norm": 3.276472330093384,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.419,
|
|
"step": 25150
|
|
},
|
|
{
|
|
"epoch": 12.29248046875,
|
|
"grad_norm": 2.8615663051605225,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3974,
|
|
"step": 25175
|
|
},
|
|
{
|
|
"epoch": 12.3046875,
|
|
"grad_norm": 4.341454029083252,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4033,
|
|
"step": 25200
|
|
},
|
|
{
|
|
"epoch": 12.31689453125,
|
|
"grad_norm": 4.287600517272949,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4316,
|
|
"step": 25225
|
|
},
|
|
{
|
|
"epoch": 12.3291015625,
|
|
"grad_norm": 2.183636426925659,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4461,
|
|
"step": 25250
|
|
},
|
|
{
|
|
"epoch": 12.34130859375,
|
|
"grad_norm": 2.487642765045166,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.413,
|
|
"step": 25275
|
|
},
|
|
{
|
|
"epoch": 12.353515625,
|
|
"grad_norm": 2.7473361492156982,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.445,
|
|
"step": 25300
|
|
},
|
|
{
|
|
"epoch": 12.36572265625,
|
|
"grad_norm": 3.353499174118042,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4473,
|
|
"step": 25325
|
|
},
|
|
{
|
|
"epoch": 12.3779296875,
|
|
"grad_norm": 2.5522449016571045,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.41,
|
|
"step": 25350
|
|
},
|
|
{
|
|
"epoch": 12.39013671875,
|
|
"grad_norm": 4.818186283111572,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.445,
|
|
"step": 25375
|
|
},
|
|
{
|
|
"epoch": 12.40234375,
|
|
"grad_norm": 4.093599796295166,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4741,
|
|
"step": 25400
|
|
},
|
|
{
|
|
"epoch": 12.41455078125,
|
|
"grad_norm": 3.5378177165985107,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4002,
|
|
"step": 25425
|
|
},
|
|
{
|
|
"epoch": 12.4267578125,
|
|
"grad_norm": 2.7892191410064697,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3966,
|
|
"step": 25450
|
|
},
|
|
{
|
|
"epoch": 12.43896484375,
|
|
"grad_norm": 4.0514044761657715,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4497,
|
|
"step": 25475
|
|
},
|
|
{
|
|
"epoch": 12.451171875,
|
|
"grad_norm": 2.708777904510498,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4542,
|
|
"step": 25500
|
|
},
|
|
{
|
|
"epoch": 12.451171875,
|
|
"eval_cer": 21.78181818181818,
|
|
"eval_loss": 0.3693616986274719,
|
|
"eval_normalized_cer": 17.879161528976574,
|
|
"eval_runtime": 91.1491,
|
|
"eval_samples_per_second": 1.404,
|
|
"eval_steps_per_second": 0.088,
|
|
"step": 25500
|
|
},
|
|
{
|
|
"epoch": 12.46337890625,
|
|
"grad_norm": 2.597757577896118,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.398,
|
|
"step": 25525
|
|
},
|
|
{
|
|
"epoch": 12.4755859375,
|
|
"grad_norm": 4.003383159637451,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3819,
|
|
"step": 25550
|
|
},
|
|
{
|
|
"epoch": 12.48779296875,
|
|
"grad_norm": 2.98598575592041,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3604,
|
|
"step": 25575
|
|
},
|
|
{
|
|
"epoch": 12.5,
|
|
"grad_norm": 2.1123640537261963,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4159,
|
|
"step": 25600
|
|
},
|
|
{
|
|
"epoch": 12.51220703125,
|
|
"grad_norm": 3.579258441925049,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4355,
|
|
"step": 25625
|
|
},
|
|
{
|
|
"epoch": 12.5244140625,
|
|
"grad_norm": 3.140286922454834,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3937,
|
|
"step": 25650
|
|
},
|
|
{
|
|
"epoch": 12.53662109375,
|
|
"grad_norm": 2.7550101280212402,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4045,
|
|
"step": 25675
|
|
},
|
|
{
|
|
"epoch": 12.548828125,
|
|
"grad_norm": 2.977573871612549,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4117,
|
|
"step": 25700
|
|
},
|
|
{
|
|
"epoch": 12.56103515625,
|
|
"grad_norm": 2.6313016414642334,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3912,
|
|
"step": 25725
|
|
},
|
|
{
|
|
"epoch": 12.5732421875,
|
|
"grad_norm": 1.7519272565841675,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4081,
|
|
"step": 25750
|
|
},
|
|
{
|
|
"epoch": 12.58544921875,
|
|
"grad_norm": 2.9577393531799316,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4349,
|
|
"step": 25775
|
|
},
|
|
{
|
|
"epoch": 12.59765625,
|
|
"grad_norm": 3.945467233657837,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.403,
|
|
"step": 25800
|
|
},
|
|
{
|
|
"epoch": 12.60986328125,
|
|
"grad_norm": 3.028316020965576,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4188,
|
|
"step": 25825
|
|
},
|
|
{
|
|
"epoch": 12.6220703125,
|
|
"grad_norm": 2.839320421218872,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4018,
|
|
"step": 25850
|
|
},
|
|
{
|
|
"epoch": 12.63427734375,
|
|
"grad_norm": 2.295088291168213,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.35,
|
|
"step": 25875
|
|
},
|
|
{
|
|
"epoch": 12.646484375,
|
|
"grad_norm": 2.7483038902282715,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4129,
|
|
"step": 25900
|
|
},
|
|
{
|
|
"epoch": 12.65869140625,
|
|
"grad_norm": 3.1008427143096924,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3926,
|
|
"step": 25925
|
|
},
|
|
{
|
|
"epoch": 12.6708984375,
|
|
"grad_norm": 2.92093825340271,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3813,
|
|
"step": 25950
|
|
},
|
|
{
|
|
"epoch": 12.68310546875,
|
|
"grad_norm": 2.930253267288208,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3903,
|
|
"step": 25975
|
|
},
|
|
{
|
|
"epoch": 12.6953125,
|
|
"grad_norm": 2.860041379928589,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4042,
|
|
"step": 26000
|
|
},
|
|
{
|
|
"epoch": 12.6953125,
|
|
"eval_cer": 22.10909090909091,
|
|
"eval_loss": 0.3708876967430115,
|
|
"eval_normalized_cer": 17.79695848746404,
|
|
"eval_runtime": 90.9962,
|
|
"eval_samples_per_second": 1.407,
|
|
"eval_steps_per_second": 0.088,
|
|
"step": 26000
|
|
},
|
|
{
|
|
"epoch": 12.70751953125,
|
|
"grad_norm": 5.0800299644470215,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4279,
|
|
"step": 26025
|
|
},
|
|
{
|
|
"epoch": 12.7197265625,
|
|
"grad_norm": 4.715578556060791,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4126,
|
|
"step": 26050
|
|
},
|
|
{
|
|
"epoch": 12.73193359375,
|
|
"grad_norm": 3.818558692932129,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4176,
|
|
"step": 26075
|
|
},
|
|
{
|
|
"epoch": 12.744140625,
|
|
"grad_norm": 3.0944199562072754,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4155,
|
|
"step": 26100
|
|
},
|
|
{
|
|
"epoch": 12.75634765625,
|
|
"grad_norm": 4.057784080505371,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4359,
|
|
"step": 26125
|
|
},
|
|
{
|
|
"epoch": 12.7685546875,
|
|
"grad_norm": 2.12882661819458,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.427,
|
|
"step": 26150
|
|
},
|
|
{
|
|
"epoch": 12.78076171875,
|
|
"grad_norm": 2.9268953800201416,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4437,
|
|
"step": 26175
|
|
},
|
|
{
|
|
"epoch": 12.79296875,
|
|
"grad_norm": 3.1372625827789307,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3814,
|
|
"step": 26200
|
|
},
|
|
{
|
|
"epoch": 12.80517578125,
|
|
"grad_norm": 3.171339511871338,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.461,
|
|
"step": 26225
|
|
},
|
|
{
|
|
"epoch": 12.8173828125,
|
|
"grad_norm": 2.2062418460845947,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4147,
|
|
"step": 26250
|
|
},
|
|
{
|
|
"epoch": 12.82958984375,
|
|
"grad_norm": 4.490868091583252,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3782,
|
|
"step": 26275
|
|
},
|
|
{
|
|
"epoch": 12.841796875,
|
|
"grad_norm": 4.018364429473877,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4402,
|
|
"step": 26300
|
|
},
|
|
{
|
|
"epoch": 12.85400390625,
|
|
"grad_norm": 3.0183207988739014,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4353,
|
|
"step": 26325
|
|
},
|
|
{
|
|
"epoch": 12.8662109375,
|
|
"grad_norm": 3.209418535232544,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4267,
|
|
"step": 26350
|
|
},
|
|
{
|
|
"epoch": 12.87841796875,
|
|
"grad_norm": 2.75201678276062,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4098,
|
|
"step": 26375
|
|
},
|
|
{
|
|
"epoch": 12.890625,
|
|
"grad_norm": 2.8569531440734863,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4134,
|
|
"step": 26400
|
|
},
|
|
{
|
|
"epoch": 12.90283203125,
|
|
"grad_norm": 3.4323573112487793,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.392,
|
|
"step": 26425
|
|
},
|
|
{
|
|
"epoch": 12.9150390625,
|
|
"grad_norm": 2.4016165733337402,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.407,
|
|
"step": 26450
|
|
},
|
|
{
|
|
"epoch": 12.92724609375,
|
|
"grad_norm": 2.4917731285095215,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.3779,
|
|
"step": 26475
|
|
},
|
|
{
|
|
"epoch": 12.939453125,
|
|
"grad_norm": 2.2973673343658447,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4162,
|
|
"step": 26500
|
|
},
|
|
{
|
|
"epoch": 12.939453125,
|
|
"eval_cer": 20.10909090909091,
|
|
"eval_loss": 0.36156314611434937,
|
|
"eval_normalized_cer": 16.029593094944513,
|
|
"eval_runtime": 88.7052,
|
|
"eval_samples_per_second": 1.443,
|
|
"eval_steps_per_second": 0.09,
|
|
"step": 26500
|
|
},
|
|
{
|
|
"epoch": 12.95166015625,
|
|
"grad_norm": 2.8456501960754395,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4392,
|
|
"step": 26525
|
|
},
|
|
{
|
|
"epoch": 12.9638671875,
|
|
"grad_norm": 2.975572109222412,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4097,
|
|
"step": 26550
|
|
},
|
|
{
|
|
"epoch": 12.97607421875,
|
|
"grad_norm": 4.956911563873291,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4098,
|
|
"step": 26575
|
|
},
|
|
{
|
|
"epoch": 12.98828125,
|
|
"grad_norm": 3.6981606483459473,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.4451,
|
|
"step": 26600
|
|
},
|
|
{
|
|
"epoch": 13.0,
|
|
"step": 26624,
|
|
"total_flos": 6.747472049681203e+20,
|
|
"train_loss": 0.031781960636950456,
|
|
"train_runtime": 2625.017,
|
|
"train_samples_per_second": 162.279,
|
|
"train_steps_per_second": 10.142
|
|
}
|
|
],
|
|
"logging_steps": 25,
|
|
"max_steps": 26624,
|
|
"num_input_tokens_seen": 0,
|
|
"num_train_epochs": 13,
|
|
"save_steps": 500,
|
|
"stateful_callbacks": {
|
|
"TrainerControl": {
|
|
"args": {
|
|
"should_epoch_stop": false,
|
|
"should_evaluate": false,
|
|
"should_log": false,
|
|
"should_save": true,
|
|
"should_training_stop": true
|
|
},
|
|
"attributes": {}
|
|
}
|
|
},
|
|
"total_flos": 6.747472049681203e+20,
|
|
"train_batch_size": 16,
|
|
"trial_name": null,
|
|
"trial_params": null
|
|
}
|
|
|