|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 3653, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.001368738023542294, |
|
"grad_norm": 1.1110930744661907, |
|
"learning_rate": 2.73224043715847e-07, |
|
"loss": 1.71, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.002737476047084588, |
|
"grad_norm": 1.0450218593816867, |
|
"learning_rate": 5.46448087431694e-07, |
|
"loss": 1.7206, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.004106214070626882, |
|
"grad_norm": 1.2104539142499893, |
|
"learning_rate": 8.196721311475409e-07, |
|
"loss": 1.8131, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.005474952094169176, |
|
"grad_norm": 1.1014228621000113, |
|
"learning_rate": 1.092896174863388e-06, |
|
"loss": 1.7571, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.00684369011771147, |
|
"grad_norm": 1.0213348170440788, |
|
"learning_rate": 1.3661202185792352e-06, |
|
"loss": 1.7504, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.008212428141253765, |
|
"grad_norm": 0.907763844914787, |
|
"learning_rate": 1.6393442622950819e-06, |
|
"loss": 1.726, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.009581166164796057, |
|
"grad_norm": 0.9051570945453739, |
|
"learning_rate": 1.912568306010929e-06, |
|
"loss": 1.7158, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.010949904188338352, |
|
"grad_norm": 0.7396508509648158, |
|
"learning_rate": 2.185792349726776e-06, |
|
"loss": 1.7692, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.012318642211880646, |
|
"grad_norm": 0.6316750205627214, |
|
"learning_rate": 2.459016393442623e-06, |
|
"loss": 1.7118, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.01368738023542294, |
|
"grad_norm": 0.6071126407750962, |
|
"learning_rate": 2.7322404371584705e-06, |
|
"loss": 1.6533, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.015056118258965233, |
|
"grad_norm": 0.6129917793535419, |
|
"learning_rate": 3.0054644808743173e-06, |
|
"loss": 1.6471, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.01642485628250753, |
|
"grad_norm": 0.6342687394025227, |
|
"learning_rate": 3.2786885245901638e-06, |
|
"loss": 1.6505, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.017793594306049824, |
|
"grad_norm": 0.5732154266113143, |
|
"learning_rate": 3.551912568306011e-06, |
|
"loss": 1.6261, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.019162332329592115, |
|
"grad_norm": 0.5200300967595174, |
|
"learning_rate": 3.825136612021858e-06, |
|
"loss": 1.5861, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.02053107035313441, |
|
"grad_norm": 0.5356030837166187, |
|
"learning_rate": 4.098360655737705e-06, |
|
"loss": 1.6002, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.021899808376676703, |
|
"grad_norm": 0.4893081141758957, |
|
"learning_rate": 4.371584699453552e-06, |
|
"loss": 1.5388, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.023268546400218998, |
|
"grad_norm": 0.5042585034665032, |
|
"learning_rate": 4.6448087431694e-06, |
|
"loss": 1.5788, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.024637284423761292, |
|
"grad_norm": 0.4756627065397568, |
|
"learning_rate": 4.918032786885246e-06, |
|
"loss": 1.5832, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.026006022447303587, |
|
"grad_norm": 0.4712228645477175, |
|
"learning_rate": 5.191256830601094e-06, |
|
"loss": 1.5221, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.02737476047084588, |
|
"grad_norm": 0.47712827241082845, |
|
"learning_rate": 5.464480874316941e-06, |
|
"loss": 1.5554, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.028743498494388176, |
|
"grad_norm": 0.4747245775681728, |
|
"learning_rate": 5.737704918032787e-06, |
|
"loss": 1.5051, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.030112236517930466, |
|
"grad_norm": 0.4619152125132191, |
|
"learning_rate": 6.010928961748635e-06, |
|
"loss": 1.5731, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.031480974541472764, |
|
"grad_norm": 0.46656513563983254, |
|
"learning_rate": 6.284153005464482e-06, |
|
"loss": 1.5719, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.03284971256501506, |
|
"grad_norm": 0.45441304688271805, |
|
"learning_rate": 6.5573770491803276e-06, |
|
"loss": 1.5359, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.03421845058855735, |
|
"grad_norm": 0.4492251581811985, |
|
"learning_rate": 6.830601092896175e-06, |
|
"loss": 1.5254, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.03558718861209965, |
|
"grad_norm": 0.4901256909260675, |
|
"learning_rate": 7.103825136612022e-06, |
|
"loss": 1.5531, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.036955926635641935, |
|
"grad_norm": 0.4521871797566739, |
|
"learning_rate": 7.3770491803278695e-06, |
|
"loss": 1.5203, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.03832466465918423, |
|
"grad_norm": 0.46301007561132373, |
|
"learning_rate": 7.650273224043716e-06, |
|
"loss": 1.5597, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.039693402682726524, |
|
"grad_norm": 0.4443455697452246, |
|
"learning_rate": 7.923497267759564e-06, |
|
"loss": 1.4751, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.04106214070626882, |
|
"grad_norm": 0.4529255518299648, |
|
"learning_rate": 8.19672131147541e-06, |
|
"loss": 1.5645, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.04243087872981111, |
|
"grad_norm": 0.45575204993786755, |
|
"learning_rate": 8.469945355191259e-06, |
|
"loss": 1.5297, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.04379961675335341, |
|
"grad_norm": 0.4619195990067295, |
|
"learning_rate": 8.743169398907103e-06, |
|
"loss": 1.4945, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.0451683547768957, |
|
"grad_norm": 0.43620516601045234, |
|
"learning_rate": 9.016393442622952e-06, |
|
"loss": 1.5064, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.046537092800437996, |
|
"grad_norm": 0.48507057683395305, |
|
"learning_rate": 9.2896174863388e-06, |
|
"loss": 1.4873, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.04790583082398029, |
|
"grad_norm": 0.4591779126577447, |
|
"learning_rate": 9.562841530054644e-06, |
|
"loss": 1.4989, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.049274568847522585, |
|
"grad_norm": 0.4477837953889451, |
|
"learning_rate": 9.836065573770493e-06, |
|
"loss": 1.4755, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.05064330687106488, |
|
"grad_norm": 0.4580627295319435, |
|
"learning_rate": 1.0109289617486339e-05, |
|
"loss": 1.4545, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.05201204489460717, |
|
"grad_norm": 0.45313033315292356, |
|
"learning_rate": 1.0382513661202187e-05, |
|
"loss": 1.4985, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.05338078291814947, |
|
"grad_norm": 0.42708720087605334, |
|
"learning_rate": 1.0655737704918034e-05, |
|
"loss": 1.4229, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.05474952094169176, |
|
"grad_norm": 0.4211311153329928, |
|
"learning_rate": 1.0928961748633882e-05, |
|
"loss": 1.4669, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.05611825896523406, |
|
"grad_norm": 0.4553582543296074, |
|
"learning_rate": 1.1202185792349727e-05, |
|
"loss": 1.4444, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.05748699698877635, |
|
"grad_norm": 0.43888054055275466, |
|
"learning_rate": 1.1475409836065575e-05, |
|
"loss": 1.4622, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.058855735012318645, |
|
"grad_norm": 0.43839040680573754, |
|
"learning_rate": 1.1748633879781421e-05, |
|
"loss": 1.3921, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.06022447303586093, |
|
"grad_norm": 0.442929256824184, |
|
"learning_rate": 1.202185792349727e-05, |
|
"loss": 1.4695, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.06159321105940323, |
|
"grad_norm": 0.42979267539703714, |
|
"learning_rate": 1.2295081967213116e-05, |
|
"loss": 1.4444, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.06296194908294553, |
|
"grad_norm": 0.43942113048062365, |
|
"learning_rate": 1.2568306010928964e-05, |
|
"loss": 1.4581, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.06433068710648782, |
|
"grad_norm": 0.43105155511527676, |
|
"learning_rate": 1.284153005464481e-05, |
|
"loss": 1.4335, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.06569942513003012, |
|
"grad_norm": 0.4347334195115935, |
|
"learning_rate": 1.3114754098360655e-05, |
|
"loss": 1.5437, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.0670681631535724, |
|
"grad_norm": 0.41558733251657665, |
|
"learning_rate": 1.3387978142076503e-05, |
|
"loss": 1.4246, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.0684369011771147, |
|
"grad_norm": 0.4166574660442, |
|
"learning_rate": 1.366120218579235e-05, |
|
"loss": 1.4672, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.069805639200657, |
|
"grad_norm": 0.40343261677226544, |
|
"learning_rate": 1.3934426229508198e-05, |
|
"loss": 1.4695, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.0711743772241993, |
|
"grad_norm": 0.41945935612278784, |
|
"learning_rate": 1.4207650273224044e-05, |
|
"loss": 1.4173, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.07254311524774158, |
|
"grad_norm": 0.4091410585881606, |
|
"learning_rate": 1.4480874316939892e-05, |
|
"loss": 1.3669, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.07391185327128387, |
|
"grad_norm": 0.40428264801464686, |
|
"learning_rate": 1.4754098360655739e-05, |
|
"loss": 1.4712, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.07528059129482617, |
|
"grad_norm": 0.4074017153449847, |
|
"learning_rate": 1.5027322404371585e-05, |
|
"loss": 1.4477, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.07664932931836846, |
|
"grad_norm": 0.41299926058270664, |
|
"learning_rate": 1.5300546448087432e-05, |
|
"loss": 1.4015, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.07801806734191076, |
|
"grad_norm": 0.39622800283246956, |
|
"learning_rate": 1.5573770491803278e-05, |
|
"loss": 1.443, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.07938680536545305, |
|
"grad_norm": 0.4050370810855124, |
|
"learning_rate": 1.5846994535519128e-05, |
|
"loss": 1.4454, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.08075554338899535, |
|
"grad_norm": 0.405591352468174, |
|
"learning_rate": 1.6120218579234975e-05, |
|
"loss": 1.4077, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.08212428141253764, |
|
"grad_norm": 0.4188674346008248, |
|
"learning_rate": 1.639344262295082e-05, |
|
"loss": 1.4503, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.08349301943607994, |
|
"grad_norm": 0.4015518431467613, |
|
"learning_rate": 1.6666666666666667e-05, |
|
"loss": 1.4027, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.08486175745962223, |
|
"grad_norm": 0.40825550305371383, |
|
"learning_rate": 1.6939890710382517e-05, |
|
"loss": 1.4021, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.08623049548316453, |
|
"grad_norm": 0.4103687720629476, |
|
"learning_rate": 1.721311475409836e-05, |
|
"loss": 1.4304, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.08759923350670681, |
|
"grad_norm": 0.4048205442390437, |
|
"learning_rate": 1.7486338797814207e-05, |
|
"loss": 1.4, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.08896797153024912, |
|
"grad_norm": 0.41661892344943224, |
|
"learning_rate": 1.7759562841530057e-05, |
|
"loss": 1.4035, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.0903367095537914, |
|
"grad_norm": 0.4308332237873975, |
|
"learning_rate": 1.8032786885245903e-05, |
|
"loss": 1.4302, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.0917054475773337, |
|
"grad_norm": 0.41012677210314796, |
|
"learning_rate": 1.830601092896175e-05, |
|
"loss": 1.3969, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.09307418560087599, |
|
"grad_norm": 0.40743633096989823, |
|
"learning_rate": 1.85792349726776e-05, |
|
"loss": 1.4425, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.09444292362441829, |
|
"grad_norm": 0.39899819661715447, |
|
"learning_rate": 1.8852459016393446e-05, |
|
"loss": 1.4303, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.09581166164796058, |
|
"grad_norm": 0.41870768132807307, |
|
"learning_rate": 1.912568306010929e-05, |
|
"loss": 1.4002, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.09718039967150287, |
|
"grad_norm": 0.40412869941796903, |
|
"learning_rate": 1.939890710382514e-05, |
|
"loss": 1.403, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.09854913769504517, |
|
"grad_norm": 0.40478167238457397, |
|
"learning_rate": 1.9672131147540985e-05, |
|
"loss": 1.4227, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.09991787571858746, |
|
"grad_norm": 0.41341003482817107, |
|
"learning_rate": 1.994535519125683e-05, |
|
"loss": 1.4302, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.10128661374212976, |
|
"grad_norm": 0.41658269655258945, |
|
"learning_rate": 1.999992692147127e-05, |
|
"loss": 1.445, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.10265535176567205, |
|
"grad_norm": 0.40746899077621546, |
|
"learning_rate": 1.999963004177886e-05, |
|
"loss": 1.458, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.10402408978921435, |
|
"grad_norm": 0.3960893049170255, |
|
"learning_rate": 1.9999104800289367e-05, |
|
"loss": 1.4382, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.10539282781275663, |
|
"grad_norm": 0.3950983233516754, |
|
"learning_rate": 1.9998351208997734e-05, |
|
"loss": 1.4365, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.10676156583629894, |
|
"grad_norm": 0.38845890924227394, |
|
"learning_rate": 1.9997369285113754e-05, |
|
"loss": 1.3731, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.10813030385984122, |
|
"grad_norm": 0.39526066204949717, |
|
"learning_rate": 1.9996159051061638e-05, |
|
"loss": 1.4111, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.10949904188338352, |
|
"grad_norm": 0.4133763360642962, |
|
"learning_rate": 1.9994720534479543e-05, |
|
"loss": 1.4294, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.11086777990692581, |
|
"grad_norm": 0.39671784657884285, |
|
"learning_rate": 1.999305376821889e-05, |
|
"loss": 1.3708, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.11223651793046811, |
|
"grad_norm": 0.3847825062400942, |
|
"learning_rate": 1.999115879034368e-05, |
|
"loss": 1.4638, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.1136052559540104, |
|
"grad_norm": 0.41459555705408424, |
|
"learning_rate": 1.9989035644129553e-05, |
|
"loss": 1.4105, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.1149739939775527, |
|
"grad_norm": 0.4270268157501032, |
|
"learning_rate": 1.998668437806286e-05, |
|
"loss": 1.4243, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.11634273200109499, |
|
"grad_norm": 0.42083549100480927, |
|
"learning_rate": 1.998410504583952e-05, |
|
"loss": 1.4285, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.11771147002463729, |
|
"grad_norm": 0.3929543512101535, |
|
"learning_rate": 1.998129770636381e-05, |
|
"loss": 1.3692, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.11908020804817958, |
|
"grad_norm": 0.40163841769109054, |
|
"learning_rate": 1.9978262423747003e-05, |
|
"loss": 1.3842, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.12044894607172187, |
|
"grad_norm": 0.40750648380479343, |
|
"learning_rate": 1.997499926730593e-05, |
|
"loss": 1.4412, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.12181768409526417, |
|
"grad_norm": 0.4029670313800167, |
|
"learning_rate": 1.9971508311561373e-05, |
|
"loss": 1.4481, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.12318642211880645, |
|
"grad_norm": 0.3934576991168861, |
|
"learning_rate": 1.996778963623637e-05, |
|
"loss": 1.3706, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.12455516014234876, |
|
"grad_norm": 0.4126991514447529, |
|
"learning_rate": 1.9963843326254406e-05, |
|
"loss": 1.4096, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.12592389816589106, |
|
"grad_norm": 0.400896785065692, |
|
"learning_rate": 1.9959669471737456e-05, |
|
"loss": 1.3284, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.12729263618943334, |
|
"grad_norm": 0.4084884002554967, |
|
"learning_rate": 1.9955268168003938e-05, |
|
"loss": 1.4346, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.12866137421297563, |
|
"grad_norm": 0.4031311283565323, |
|
"learning_rate": 1.9950639515566537e-05, |
|
"loss": 1.3632, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.13003011223651792, |
|
"grad_norm": 0.4017916786924823, |
|
"learning_rate": 1.99457836201299e-05, |
|
"loss": 1.4083, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.13139885026006023, |
|
"grad_norm": 0.38801834482390074, |
|
"learning_rate": 1.9940700592588228e-05, |
|
"loss": 1.3823, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.13276758828360252, |
|
"grad_norm": 0.40725954819087723, |
|
"learning_rate": 1.993539054902275e-05, |
|
"loss": 1.369, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.1341363263071448, |
|
"grad_norm": 0.42191731863948084, |
|
"learning_rate": 1.992985361069906e-05, |
|
"loss": 1.3907, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.1355050643306871, |
|
"grad_norm": 0.3981980013765845, |
|
"learning_rate": 1.9924089904064354e-05, |
|
"loss": 1.4253, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.1368738023542294, |
|
"grad_norm": 0.3968358847514907, |
|
"learning_rate": 1.9918099560744545e-05, |
|
"loss": 1.4478, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.1382425403777717, |
|
"grad_norm": 0.4020407861005804, |
|
"learning_rate": 1.991188271754125e-05, |
|
"loss": 1.4063, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.139611278401314, |
|
"grad_norm": 0.3983008409093381, |
|
"learning_rate": 1.990543951642866e-05, |
|
"loss": 1.3838, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.14098001642485627, |
|
"grad_norm": 0.38772247745474425, |
|
"learning_rate": 1.9898770104550335e-05, |
|
"loss": 1.4193, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.1423487544483986, |
|
"grad_norm": 0.4172192092889838, |
|
"learning_rate": 1.9891874634215784e-05, |
|
"loss": 1.3831, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.14371749247194088, |
|
"grad_norm": 0.39316534732966774, |
|
"learning_rate": 1.9884753262897042e-05, |
|
"loss": 1.3813, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.14508623049548316, |
|
"grad_norm": 0.40825756476362335, |
|
"learning_rate": 1.9877406153225028e-05, |
|
"loss": 1.4233, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.14645496851902545, |
|
"grad_norm": 0.4303738354233706, |
|
"learning_rate": 1.9869833472985882e-05, |
|
"loss": 1.4267, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.14782370654256774, |
|
"grad_norm": 0.39536263390575654, |
|
"learning_rate": 1.9862035395117075e-05, |
|
"loss": 1.3688, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.14919244456611006, |
|
"grad_norm": 0.39303998157704906, |
|
"learning_rate": 1.9854012097703515e-05, |
|
"loss": 1.4259, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.15056118258965234, |
|
"grad_norm": 0.3939411519174778, |
|
"learning_rate": 1.9845763763973433e-05, |
|
"loss": 1.354, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.15192992061319463, |
|
"grad_norm": 0.39277763892363776, |
|
"learning_rate": 1.9837290582294233e-05, |
|
"loss": 1.3468, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.15329865863673692, |
|
"grad_norm": 0.39121990838351217, |
|
"learning_rate": 1.9828592746168172e-05, |
|
"loss": 1.3508, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.15466739666027923, |
|
"grad_norm": 0.4035101254769364, |
|
"learning_rate": 1.981967045422795e-05, |
|
"loss": 1.369, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.15603613468382152, |
|
"grad_norm": 0.38407623227911974, |
|
"learning_rate": 1.9810523910232165e-05, |
|
"loss": 1.4039, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.1574048727073638, |
|
"grad_norm": 0.3915340445821982, |
|
"learning_rate": 1.9801153323060667e-05, |
|
"loss": 1.3634, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.1587736107309061, |
|
"grad_norm": 0.39241210575593644, |
|
"learning_rate": 1.9791558906709787e-05, |
|
"loss": 1.4257, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.1601423487544484, |
|
"grad_norm": 0.3946963973614525, |
|
"learning_rate": 1.9781740880287444e-05, |
|
"loss": 1.3136, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.1615110867779907, |
|
"grad_norm": 0.39148042578465725, |
|
"learning_rate": 1.9771699468008156e-05, |
|
"loss": 1.3654, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.16287982480153299, |
|
"grad_norm": 0.3895106088217061, |
|
"learning_rate": 1.9761434899187893e-05, |
|
"loss": 1.3762, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.16424856282507527, |
|
"grad_norm": 0.3997937267692177, |
|
"learning_rate": 1.9750947408238872e-05, |
|
"loss": 1.4064, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.1656173008486176, |
|
"grad_norm": 0.4002294154735745, |
|
"learning_rate": 1.974023723466418e-05, |
|
"loss": 1.3771, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.16698603887215988, |
|
"grad_norm": 0.40550620529076886, |
|
"learning_rate": 1.9729304623052315e-05, |
|
"loss": 1.4551, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.16835477689570216, |
|
"grad_norm": 0.40410760524836825, |
|
"learning_rate": 1.9718149823071592e-05, |
|
"loss": 1.3394, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.16972351491924445, |
|
"grad_norm": 0.40142195594131513, |
|
"learning_rate": 1.970677308946446e-05, |
|
"loss": 1.3872, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.17109225294278674, |
|
"grad_norm": 0.4041136169142662, |
|
"learning_rate": 1.9695174682041652e-05, |
|
"loss": 1.3644, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.17246099096632905, |
|
"grad_norm": 0.39436310577761385, |
|
"learning_rate": 1.9683354865676298e-05, |
|
"loss": 1.434, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.17382972898987134, |
|
"grad_norm": 0.400517293758074, |
|
"learning_rate": 1.9671313910297826e-05, |
|
"loss": 1.3941, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.17519846701341363, |
|
"grad_norm": 0.4059408812570113, |
|
"learning_rate": 1.9659052090885834e-05, |
|
"loss": 1.3907, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.17656720503695592, |
|
"grad_norm": 0.3929001302562383, |
|
"learning_rate": 1.9646569687463796e-05, |
|
"loss": 1.351, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.17793594306049823, |
|
"grad_norm": 0.4090718300758474, |
|
"learning_rate": 1.9633866985092655e-05, |
|
"loss": 1.4102, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.17930468108404052, |
|
"grad_norm": 0.40995200812054255, |
|
"learning_rate": 1.9620944273864343e-05, |
|
"loss": 1.3575, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.1806734191075828, |
|
"grad_norm": 0.40276642674265933, |
|
"learning_rate": 1.960780184889514e-05, |
|
"loss": 1.3941, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.1820421571311251, |
|
"grad_norm": 0.400136397383414, |
|
"learning_rate": 1.9594440010318924e-05, |
|
"loss": 1.4171, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.1834108951546674, |
|
"grad_norm": 0.40872717576906, |
|
"learning_rate": 1.9580859063280326e-05, |
|
"loss": 1.391, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.1847796331782097, |
|
"grad_norm": 0.39405384932557985, |
|
"learning_rate": 1.956705931792777e-05, |
|
"loss": 1.4042, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.18614837120175198, |
|
"grad_norm": 0.396632571946212, |
|
"learning_rate": 1.9553041089406387e-05, |
|
"loss": 1.3598, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.18751710922529427, |
|
"grad_norm": 0.39535225715229844, |
|
"learning_rate": 1.95388046978508e-05, |
|
"loss": 1.3652, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.18888584724883659, |
|
"grad_norm": 0.39747240968211434, |
|
"learning_rate": 1.9524350468377828e-05, |
|
"loss": 1.3857, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.19025458527237887, |
|
"grad_norm": 0.4045442697980633, |
|
"learning_rate": 1.9509678731079074e-05, |
|
"loss": 1.3724, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.19162332329592116, |
|
"grad_norm": 0.39648112930634666, |
|
"learning_rate": 1.949478982101336e-05, |
|
"loss": 1.3642, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.19299206131946345, |
|
"grad_norm": 0.3966559822195587, |
|
"learning_rate": 1.947968407819909e-05, |
|
"loss": 1.3704, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.19436079934300574, |
|
"grad_norm": 0.3803799928500003, |
|
"learning_rate": 1.9464361847606486e-05, |
|
"loss": 1.3718, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.19572953736654805, |
|
"grad_norm": 0.3920580621341347, |
|
"learning_rate": 1.9448823479149705e-05, |
|
"loss": 1.3994, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.19709827539009034, |
|
"grad_norm": 0.38159551727148144, |
|
"learning_rate": 1.9433069327678847e-05, |
|
"loss": 1.3539, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.19846701341363263, |
|
"grad_norm": 0.3979848267470623, |
|
"learning_rate": 1.9417099752971858e-05, |
|
"loss": 1.3824, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.1998357514371749, |
|
"grad_norm": 0.36228811525683013, |
|
"learning_rate": 1.9400915119726305e-05, |
|
"loss": 1.2942, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.20120448946071723, |
|
"grad_norm": 0.40336839547536973, |
|
"learning_rate": 1.938451579755106e-05, |
|
"loss": 1.3305, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.20257322748425952, |
|
"grad_norm": 0.4093706132368206, |
|
"learning_rate": 1.9367902160957843e-05, |
|
"loss": 1.4047, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.2039419655078018, |
|
"grad_norm": 0.4012678664539993, |
|
"learning_rate": 1.9351074589352684e-05, |
|
"loss": 1.34, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.2053107035313441, |
|
"grad_norm": 0.3982341789573293, |
|
"learning_rate": 1.933403346702725e-05, |
|
"loss": 1.3518, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.2066794415548864, |
|
"grad_norm": 0.3895851509088621, |
|
"learning_rate": 1.931677918315007e-05, |
|
"loss": 1.3698, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.2080481795784287, |
|
"grad_norm": 0.40873783451123424, |
|
"learning_rate": 1.9299312131757645e-05, |
|
"loss": 1.3768, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.20941691760197098, |
|
"grad_norm": 0.38863227885235524, |
|
"learning_rate": 1.928163271174546e-05, |
|
"loss": 1.368, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.21078565562551327, |
|
"grad_norm": 0.40218624042286666, |
|
"learning_rate": 1.9263741326858866e-05, |
|
"loss": 1.4002, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.21215439364905558, |
|
"grad_norm": 0.38750285278794777, |
|
"learning_rate": 1.9245638385683857e-05, |
|
"loss": 1.3808, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.21352313167259787, |
|
"grad_norm": 0.39332958062090156, |
|
"learning_rate": 1.9227324301637747e-05, |
|
"loss": 1.3991, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.21489186969614016, |
|
"grad_norm": 0.39759945858946194, |
|
"learning_rate": 1.9208799492959723e-05, |
|
"loss": 1.3765, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.21626060771968245, |
|
"grad_norm": 0.3779639266645249, |
|
"learning_rate": 1.9190064382701296e-05, |
|
"loss": 1.378, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.21762934574322473, |
|
"grad_norm": 0.3980685187213535, |
|
"learning_rate": 1.917111939871664e-05, |
|
"loss": 1.3559, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.21899808376676705, |
|
"grad_norm": 0.3955990129038042, |
|
"learning_rate": 1.915196497365282e-05, |
|
"loss": 1.3688, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.22036682179030934, |
|
"grad_norm": 0.3814772041353902, |
|
"learning_rate": 1.9132601544939914e-05, |
|
"loss": 1.3106, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.22173555981385162, |
|
"grad_norm": 0.4033095650227253, |
|
"learning_rate": 1.9113029554781014e-05, |
|
"loss": 1.3389, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.2231042978373939, |
|
"grad_norm": 0.3934860150762494, |
|
"learning_rate": 1.9093249450142144e-05, |
|
"loss": 1.3343, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.22447303586093623, |
|
"grad_norm": 0.4008166941226203, |
|
"learning_rate": 1.907326168274204e-05, |
|
"loss": 1.3627, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.2258417738844785, |
|
"grad_norm": 0.38863913805386574, |
|
"learning_rate": 1.905306670904184e-05, |
|
"loss": 1.3978, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.2272105119080208, |
|
"grad_norm": 0.39887884072448215, |
|
"learning_rate": 1.9032664990234648e-05, |
|
"loss": 1.3548, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.2285792499315631, |
|
"grad_norm": 0.3942528539065832, |
|
"learning_rate": 1.9012056992235025e-05, |
|
"loss": 1.3589, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.2299479879551054, |
|
"grad_norm": 0.4003093485695292, |
|
"learning_rate": 1.899124318566832e-05, |
|
"loss": 1.4008, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.2313167259786477, |
|
"grad_norm": 0.38851656838431264, |
|
"learning_rate": 1.897022404585996e-05, |
|
"loss": 1.3396, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.23268546400218998, |
|
"grad_norm": 0.39282411404685663, |
|
"learning_rate": 1.894900005282454e-05, |
|
"loss": 1.3118, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.23405420202573227, |
|
"grad_norm": 0.3995600743628729, |
|
"learning_rate": 1.892757169125492e-05, |
|
"loss": 1.3261, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.23542294004927458, |
|
"grad_norm": 0.40585943733175783, |
|
"learning_rate": 1.8905939450511117e-05, |
|
"loss": 1.2896, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.23679167807281687, |
|
"grad_norm": 0.38676778157652897, |
|
"learning_rate": 1.888410382460915e-05, |
|
"loss": 1.3769, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.23816041609635916, |
|
"grad_norm": 0.4006135132967053, |
|
"learning_rate": 1.8862065312209735e-05, |
|
"loss": 1.3744, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.23952915411990144, |
|
"grad_norm": 0.38183538010624907, |
|
"learning_rate": 1.8839824416606932e-05, |
|
"loss": 1.3994, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.24089789214344373, |
|
"grad_norm": 0.40012172515979694, |
|
"learning_rate": 1.8817381645716613e-05, |
|
"loss": 1.3736, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.24226663016698605, |
|
"grad_norm": 0.4036320450863606, |
|
"learning_rate": 1.879473751206489e-05, |
|
"loss": 1.3679, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.24363536819052833, |
|
"grad_norm": 0.39250548017399456, |
|
"learning_rate": 1.8771892532776406e-05, |
|
"loss": 1.352, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.24500410621407062, |
|
"grad_norm": 0.384618237208868, |
|
"learning_rate": 1.8748847229562504e-05, |
|
"loss": 1.3415, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.2463728442376129, |
|
"grad_norm": 0.3847231065307316, |
|
"learning_rate": 1.8725602128709348e-05, |
|
"loss": 1.436, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.24774158226115522, |
|
"grad_norm": 0.39329176880661687, |
|
"learning_rate": 1.8702157761065877e-05, |
|
"loss": 1.344, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.2491103202846975, |
|
"grad_norm": 0.4202514383981607, |
|
"learning_rate": 1.8678514662031688e-05, |
|
"loss": 1.3384, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.2504790583082398, |
|
"grad_norm": 0.3866724615350219, |
|
"learning_rate": 1.8654673371544815e-05, |
|
"loss": 1.3374, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.2518477963317821, |
|
"grad_norm": 0.4109333525553728, |
|
"learning_rate": 1.8630634434069397e-05, |
|
"loss": 1.3949, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.2532165343553244, |
|
"grad_norm": 0.3782421450027903, |
|
"learning_rate": 1.860639839858324e-05, |
|
"loss": 1.3162, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.2545852723788667, |
|
"grad_norm": 0.3909532902784873, |
|
"learning_rate": 1.8581965818565278e-05, |
|
"loss": 1.3829, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.255954010402409, |
|
"grad_norm": 0.4051170609748319, |
|
"learning_rate": 1.855733725198295e-05, |
|
"loss": 1.3462, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.25732274842595126, |
|
"grad_norm": 0.39349552836368845, |
|
"learning_rate": 1.8532513261279433e-05, |
|
"loss": 1.4015, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.2586914864494936, |
|
"grad_norm": 0.39347934180028793, |
|
"learning_rate": 1.8507494413360808e-05, |
|
"loss": 1.3367, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.26006022447303584, |
|
"grad_norm": 0.4158855876544813, |
|
"learning_rate": 1.848228127958312e-05, |
|
"loss": 1.353, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.26142896249657815, |
|
"grad_norm": 0.3793682443730692, |
|
"learning_rate": 1.8456874435739337e-05, |
|
"loss": 1.3398, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.26279770052012047, |
|
"grad_norm": 0.38047885709566037, |
|
"learning_rate": 1.843127446204616e-05, |
|
"loss": 1.329, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.26416643854366273, |
|
"grad_norm": 0.3991275596683987, |
|
"learning_rate": 1.8405481943130827e-05, |
|
"loss": 1.3115, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.26553517656720504, |
|
"grad_norm": 0.389649193690539, |
|
"learning_rate": 1.8379497468017726e-05, |
|
"loss": 1.3746, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.2669039145907473, |
|
"grad_norm": 0.40470694231468296, |
|
"learning_rate": 1.8353321630114952e-05, |
|
"loss": 1.3354, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.2682726526142896, |
|
"grad_norm": 0.40734933283960667, |
|
"learning_rate": 1.832695502720076e-05, |
|
"loss": 1.3295, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.26964139063783193, |
|
"grad_norm": 0.39323434740751423, |
|
"learning_rate": 1.8300398261409912e-05, |
|
"loss": 1.3069, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.2710101286613742, |
|
"grad_norm": 0.3837841492160018, |
|
"learning_rate": 1.8273651939219914e-05, |
|
"loss": 1.3543, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.2723788666849165, |
|
"grad_norm": 0.38795485386053835, |
|
"learning_rate": 1.8246716671437186e-05, |
|
"loss": 1.3798, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.2737476047084588, |
|
"grad_norm": 0.38509661706339016, |
|
"learning_rate": 1.8219593073183106e-05, |
|
"loss": 1.3604, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.2751163427320011, |
|
"grad_norm": 0.3882331224756399, |
|
"learning_rate": 1.8192281763879946e-05, |
|
"loss": 1.3417, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.2764850807555434, |
|
"grad_norm": 0.3842715356979672, |
|
"learning_rate": 1.816478336723675e-05, |
|
"loss": 1.4061, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.27785381877908566, |
|
"grad_norm": 0.4013303335538485, |
|
"learning_rate": 1.8137098511235084e-05, |
|
"loss": 1.3509, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.279222556802628, |
|
"grad_norm": 0.389250409772092, |
|
"learning_rate": 1.810922782811468e-05, |
|
"loss": 1.3371, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.2805912948261703, |
|
"grad_norm": 0.3891214912082097, |
|
"learning_rate": 1.808117195435901e-05, |
|
"loss": 1.3756, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.28196003284971255, |
|
"grad_norm": 0.39295322317665116, |
|
"learning_rate": 1.805293153068076e-05, |
|
"loss": 1.3525, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.28332877087325486, |
|
"grad_norm": 0.38744159952014173, |
|
"learning_rate": 1.802450720200718e-05, |
|
"loss": 1.3508, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.2846975088967972, |
|
"grad_norm": 0.3971694811346686, |
|
"learning_rate": 1.7995899617465357e-05, |
|
"loss": 1.3242, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.28606624692033944, |
|
"grad_norm": 0.3892979964335795, |
|
"learning_rate": 1.7967109430367406e-05, |
|
"loss": 1.2919, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.28743498494388176, |
|
"grad_norm": 0.40651116778504365, |
|
"learning_rate": 1.793813729819553e-05, |
|
"loss": 1.4047, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.288803722967424, |
|
"grad_norm": 0.39365601735181455, |
|
"learning_rate": 1.7908983882587038e-05, |
|
"loss": 1.3622, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.29017246099096633, |
|
"grad_norm": 0.38324138023204, |
|
"learning_rate": 1.787964984931919e-05, |
|
"loss": 1.3773, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.29154119901450865, |
|
"grad_norm": 0.3887104304366927, |
|
"learning_rate": 1.7850135868294023e-05, |
|
"loss": 1.3973, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.2929099370380509, |
|
"grad_norm": 0.4087532681641359, |
|
"learning_rate": 1.782044261352305e-05, |
|
"loss": 1.3246, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.2942786750615932, |
|
"grad_norm": 0.39381567774906023, |
|
"learning_rate": 1.7790570763111864e-05, |
|
"loss": 1.3683, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.2956474130851355, |
|
"grad_norm": 0.38720936830053604, |
|
"learning_rate": 1.7760520999244638e-05, |
|
"loss": 1.375, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.2970161511086778, |
|
"grad_norm": 0.38699812271583683, |
|
"learning_rate": 1.7730294008168578e-05, |
|
"loss": 1.403, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.2983848891322201, |
|
"grad_norm": 0.3938155567988771, |
|
"learning_rate": 1.7699890480178216e-05, |
|
"loss": 1.3567, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.29975362715576237, |
|
"grad_norm": 0.4051470823072413, |
|
"learning_rate": 1.766931110959967e-05, |
|
"loss": 1.4228, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.3011223651793047, |
|
"grad_norm": 0.4013404711196811, |
|
"learning_rate": 1.763855659477478e-05, |
|
"loss": 1.3689, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.302491103202847, |
|
"grad_norm": 0.4239789885244797, |
|
"learning_rate": 1.7607627638045156e-05, |
|
"loss": 1.3988, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.30385984122638926, |
|
"grad_norm": 0.4005302271519552, |
|
"learning_rate": 1.7576524945736137e-05, |
|
"loss": 1.3368, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.3052285792499316, |
|
"grad_norm": 0.3900827670562639, |
|
"learning_rate": 1.754524922814068e-05, |
|
"loss": 1.3633, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.30659731727347384, |
|
"grad_norm": 0.39004513777273, |
|
"learning_rate": 1.751380119950311e-05, |
|
"loss": 1.4024, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.30796605529701615, |
|
"grad_norm": 0.3955440868581025, |
|
"learning_rate": 1.7482181578002837e-05, |
|
"loss": 1.3667, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.30933479332055847, |
|
"grad_norm": 0.41261816602720663, |
|
"learning_rate": 1.745039108573793e-05, |
|
"loss": 1.357, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.3107035313441007, |
|
"grad_norm": 0.37732770567847945, |
|
"learning_rate": 1.7418430448708644e-05, |
|
"loss": 1.3337, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.31207226936764304, |
|
"grad_norm": 0.3764972884033891, |
|
"learning_rate": 1.738630039680083e-05, |
|
"loss": 1.3556, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.3134410073911853, |
|
"grad_norm": 0.40764805847336166, |
|
"learning_rate": 1.7354001663769278e-05, |
|
"loss": 1.3679, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.3148097454147276, |
|
"grad_norm": 0.3794101029629702, |
|
"learning_rate": 1.7321534987220942e-05, |
|
"loss": 1.3176, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.31617848343826993, |
|
"grad_norm": 0.40248164169870704, |
|
"learning_rate": 1.728890110859812e-05, |
|
"loss": 1.3458, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.3175472214618122, |
|
"grad_norm": 0.39012890957455026, |
|
"learning_rate": 1.7256100773161492e-05, |
|
"loss": 1.296, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.3189159594853545, |
|
"grad_norm": 0.39856381680118136, |
|
"learning_rate": 1.7223134729973134e-05, |
|
"loss": 1.3614, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 0.3202846975088968, |
|
"grad_norm": 0.40493526767721577, |
|
"learning_rate": 1.7190003731879375e-05, |
|
"loss": 1.3533, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.3216534355324391, |
|
"grad_norm": 0.38868285184435786, |
|
"learning_rate": 1.715670853549364e-05, |
|
"loss": 1.3377, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.3230221735559814, |
|
"grad_norm": 0.3840977803535061, |
|
"learning_rate": 1.7123249901179142e-05, |
|
"loss": 1.3753, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.32439091157952366, |
|
"grad_norm": 0.37510550805682613, |
|
"learning_rate": 1.708962859303154e-05, |
|
"loss": 1.3557, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.32575964960306597, |
|
"grad_norm": 0.39070855119805253, |
|
"learning_rate": 1.7055845378861476e-05, |
|
"loss": 1.3584, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.3271283876266083, |
|
"grad_norm": 0.39021160718672115, |
|
"learning_rate": 1.7021901030177036e-05, |
|
"loss": 1.3399, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 0.32849712565015055, |
|
"grad_norm": 0.3983342506003189, |
|
"learning_rate": 1.698779632216615e-05, |
|
"loss": 1.3965, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.32986586367369286, |
|
"grad_norm": 0.38690562360273995, |
|
"learning_rate": 1.6953532033678874e-05, |
|
"loss": 1.4221, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 0.3312346016972352, |
|
"grad_norm": 0.39029540911089444, |
|
"learning_rate": 1.69191089472096e-05, |
|
"loss": 1.3201, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.33260333972077744, |
|
"grad_norm": 0.40512131600219015, |
|
"learning_rate": 1.688452784887921e-05, |
|
"loss": 1.3495, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 0.33397207774431975, |
|
"grad_norm": 0.3804586676740124, |
|
"learning_rate": 1.684978952841709e-05, |
|
"loss": 1.3471, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.335340815767862, |
|
"grad_norm": 0.38428741373037245, |
|
"learning_rate": 1.681489477914312e-05, |
|
"loss": 1.3196, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.3367095537914043, |
|
"grad_norm": 0.384047691017285, |
|
"learning_rate": 1.677984439794954e-05, |
|
"loss": 1.3303, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.33807829181494664, |
|
"grad_norm": 0.3839900646299871, |
|
"learning_rate": 1.6744639185282784e-05, |
|
"loss": 1.2792, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 0.3394470298384889, |
|
"grad_norm": 0.3884042208952324, |
|
"learning_rate": 1.670927994512514e-05, |
|
"loss": 1.3223, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.3408157678620312, |
|
"grad_norm": 0.3911385295369844, |
|
"learning_rate": 1.667376748497646e-05, |
|
"loss": 1.3546, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 0.3421845058855735, |
|
"grad_norm": 0.3807410219808285, |
|
"learning_rate": 1.6638102615835658e-05, |
|
"loss": 1.3148, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.3435532439091158, |
|
"grad_norm": 0.39423887853753, |
|
"learning_rate": 1.6602286152182236e-05, |
|
"loss": 1.361, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 0.3449219819326581, |
|
"grad_norm": 0.39187255483162625, |
|
"learning_rate": 1.6566318911957647e-05, |
|
"loss": 1.339, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.34629071995620037, |
|
"grad_norm": 0.38859511955741677, |
|
"learning_rate": 1.6530201716546647e-05, |
|
"loss": 1.3556, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 0.3476594579797427, |
|
"grad_norm": 0.3993643770992966, |
|
"learning_rate": 1.649393539075851e-05, |
|
"loss": 1.3476, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.349028196003285, |
|
"grad_norm": 0.39035176617866785, |
|
"learning_rate": 1.6457520762808217e-05, |
|
"loss": 1.3228, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.35039693402682726, |
|
"grad_norm": 0.3830707286890397, |
|
"learning_rate": 1.6420958664297514e-05, |
|
"loss": 1.3094, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.35176567205036957, |
|
"grad_norm": 0.36321444690348953, |
|
"learning_rate": 1.638424993019595e-05, |
|
"loss": 1.3853, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 0.35313441007391183, |
|
"grad_norm": 0.371835147945527, |
|
"learning_rate": 1.634739539882178e-05, |
|
"loss": 1.3477, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.35450314809745415, |
|
"grad_norm": 0.3906744323143432, |
|
"learning_rate": 1.6310395911822848e-05, |
|
"loss": 1.3149, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 0.35587188612099646, |
|
"grad_norm": 0.37408867437468385, |
|
"learning_rate": 1.6273252314157352e-05, |
|
"loss": 1.3135, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.3572406241445387, |
|
"grad_norm": 0.37403440281560435, |
|
"learning_rate": 1.6235965454074535e-05, |
|
"loss": 1.2301, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 0.35860936216808104, |
|
"grad_norm": 0.3867708656040568, |
|
"learning_rate": 1.619853618309535e-05, |
|
"loss": 1.3716, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.3599781001916233, |
|
"grad_norm": 0.3885328796144267, |
|
"learning_rate": 1.6160965355992966e-05, |
|
"loss": 1.366, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 0.3613468382151656, |
|
"grad_norm": 0.403719948973195, |
|
"learning_rate": 1.6123253830773293e-05, |
|
"loss": 1.3661, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.3627155762387079, |
|
"grad_norm": 0.3827196192323514, |
|
"learning_rate": 1.6085402468655356e-05, |
|
"loss": 1.3567, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.3640843142622502, |
|
"grad_norm": 0.3968500270407259, |
|
"learning_rate": 1.6047412134051645e-05, |
|
"loss": 1.4044, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.3654530522857925, |
|
"grad_norm": 0.38847711656237954, |
|
"learning_rate": 1.6009283694548365e-05, |
|
"loss": 1.3591, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 0.3668217903093348, |
|
"grad_norm": 0.4036707276972636, |
|
"learning_rate": 1.5971018020885623e-05, |
|
"loss": 1.3916, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.3681905283328771, |
|
"grad_norm": 0.39526777473833846, |
|
"learning_rate": 1.593261598693755e-05, |
|
"loss": 1.3478, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 0.3695592663564194, |
|
"grad_norm": 0.3837685918142903, |
|
"learning_rate": 1.5894078469692343e-05, |
|
"loss": 1.3604, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.37092800437996165, |
|
"grad_norm": 0.4184389090754904, |
|
"learning_rate": 1.585540634923223e-05, |
|
"loss": 1.4042, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 0.37229674240350397, |
|
"grad_norm": 0.4132734637825316, |
|
"learning_rate": 1.5816600508713372e-05, |
|
"loss": 1.3901, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.3736654804270463, |
|
"grad_norm": 0.3876929955451876, |
|
"learning_rate": 1.5777661834345708e-05, |
|
"loss": 1.3296, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 0.37503421845058854, |
|
"grad_norm": 0.39294627819783745, |
|
"learning_rate": 1.57385912153727e-05, |
|
"loss": 1.3581, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.37640295647413086, |
|
"grad_norm": 0.3875865415567979, |
|
"learning_rate": 1.5699389544051028e-05, |
|
"loss": 1.3167, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.37777169449767317, |
|
"grad_norm": 0.3863898154888909, |
|
"learning_rate": 1.566005771563023e-05, |
|
"loss": 1.3694, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.37914043252121543, |
|
"grad_norm": 0.38577432820095847, |
|
"learning_rate": 1.5620596628332242e-05, |
|
"loss": 1.323, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 0.38050917054475775, |
|
"grad_norm": 0.3971392209603939, |
|
"learning_rate": 1.5581007183330877e-05, |
|
"loss": 1.3432, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.3818779085683, |
|
"grad_norm": 0.40311344444707337, |
|
"learning_rate": 1.554129028473127e-05, |
|
"loss": 1.3802, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.3832466465918423, |
|
"grad_norm": 0.40527858159368374, |
|
"learning_rate": 1.5501446839549207e-05, |
|
"loss": 1.3445, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.38461538461538464, |
|
"grad_norm": 0.4025758901452214, |
|
"learning_rate": 1.5461477757690424e-05, |
|
"loss": 1.3321, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 0.3859841226389269, |
|
"grad_norm": 0.3855393398790419, |
|
"learning_rate": 1.542138395192983e-05, |
|
"loss": 1.3405, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.3873528606624692, |
|
"grad_norm": 0.38944762738859545, |
|
"learning_rate": 1.538116633789065e-05, |
|
"loss": 1.3289, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 0.38872159868601147, |
|
"grad_norm": 0.4028684971128393, |
|
"learning_rate": 1.5340825834023526e-05, |
|
"loss": 1.3798, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.3900903367095538, |
|
"grad_norm": 0.37654892707807947, |
|
"learning_rate": 1.530036336158553e-05, |
|
"loss": 1.3432, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.3914590747330961, |
|
"grad_norm": 0.3986727790830818, |
|
"learning_rate": 1.5259779844619152e-05, |
|
"loss": 1.3422, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.39282781275663836, |
|
"grad_norm": 0.4029735444716236, |
|
"learning_rate": 1.5219076209931159e-05, |
|
"loss": 1.3136, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 0.3941965507801807, |
|
"grad_norm": 0.41012078556001347, |
|
"learning_rate": 1.5178253387071458e-05, |
|
"loss": 1.4002, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.395565288803723, |
|
"grad_norm": 0.40771281882367433, |
|
"learning_rate": 1.5137312308311857e-05, |
|
"loss": 1.3684, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 0.39693402682726525, |
|
"grad_norm": 0.38445172076552614, |
|
"learning_rate": 1.5096253908624778e-05, |
|
"loss": 1.3137, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.39830276485080757, |
|
"grad_norm": 0.3869608798791883, |
|
"learning_rate": 1.5055079125661908e-05, |
|
"loss": 1.2812, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 0.3996715028743498, |
|
"grad_norm": 0.3987871366347409, |
|
"learning_rate": 1.5013788899732775e-05, |
|
"loss": 1.3394, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.40104024089789214, |
|
"grad_norm": 0.3970490183809672, |
|
"learning_rate": 1.4972384173783284e-05, |
|
"loss": 1.3544, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 0.40240897892143446, |
|
"grad_norm": 0.3925074770480503, |
|
"learning_rate": 1.493086589337418e-05, |
|
"loss": 1.3294, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.4037777169449767, |
|
"grad_norm": 0.40973551864547075, |
|
"learning_rate": 1.4889235006659448e-05, |
|
"loss": 1.3675, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.40514645496851903, |
|
"grad_norm": 0.3812235133361238, |
|
"learning_rate": 1.484749246436468e-05, |
|
"loss": 1.3288, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.40651519299206135, |
|
"grad_norm": 0.3827389411340829, |
|
"learning_rate": 1.4805639219765337e-05, |
|
"loss": 1.4128, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 0.4078839310156036, |
|
"grad_norm": 0.37343264715636293, |
|
"learning_rate": 1.476367622866499e-05, |
|
"loss": 1.3419, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.4092526690391459, |
|
"grad_norm": 0.4116326320287375, |
|
"learning_rate": 1.4721604449373505e-05, |
|
"loss": 1.297, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 0.4106214070626882, |
|
"grad_norm": 0.38955441577160277, |
|
"learning_rate": 1.4679424842685137e-05, |
|
"loss": 1.3138, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.4119901450862305, |
|
"grad_norm": 0.38444568811838686, |
|
"learning_rate": 1.4637138371856601e-05, |
|
"loss": 1.3284, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 0.4133588831097728, |
|
"grad_norm": 0.3938144538178601, |
|
"learning_rate": 1.4594746002585072e-05, |
|
"loss": 1.3498, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.41472762113331507, |
|
"grad_norm": 0.39696555787686244, |
|
"learning_rate": 1.4552248702986127e-05, |
|
"loss": 1.3524, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 0.4160963591568574, |
|
"grad_norm": 0.3802714110333375, |
|
"learning_rate": 1.4509647443571643e-05, |
|
"loss": 1.288, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.41746509718039965, |
|
"grad_norm": 0.38532238774246486, |
|
"learning_rate": 1.446694319722763e-05, |
|
"loss": 1.3656, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.41883383520394196, |
|
"grad_norm": 0.3866193343216763, |
|
"learning_rate": 1.4424136939192009e-05, |
|
"loss": 1.3696, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.4202025732274843, |
|
"grad_norm": 0.37501111401637777, |
|
"learning_rate": 1.4381229647032346e-05, |
|
"loss": 1.3512, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 0.42157131125102654, |
|
"grad_norm": 0.3700012749063157, |
|
"learning_rate": 1.4338222300623533e-05, |
|
"loss": 1.3092, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.42294004927456885, |
|
"grad_norm": 0.38806617391675347, |
|
"learning_rate": 1.4295115882125393e-05, |
|
"loss": 1.3471, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 0.42430878729811117, |
|
"grad_norm": 0.39116631927262135, |
|
"learning_rate": 1.4251911375960261e-05, |
|
"loss": 1.4043, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.4256775253216534, |
|
"grad_norm": 0.3907489868883888, |
|
"learning_rate": 1.4208609768790513e-05, |
|
"loss": 1.3476, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 0.42704626334519574, |
|
"grad_norm": 0.40188011896979126, |
|
"learning_rate": 1.4165212049496013e-05, |
|
"loss": 1.3398, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.428415001368738, |
|
"grad_norm": 0.39583117671628015, |
|
"learning_rate": 1.4121719209151545e-05, |
|
"loss": 1.3827, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 0.4297837393922803, |
|
"grad_norm": 0.3948036088686505, |
|
"learning_rate": 1.4078132241004174e-05, |
|
"loss": 1.3509, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.43115247741582263, |
|
"grad_norm": 0.3844018510427473, |
|
"learning_rate": 1.4034452140450561e-05, |
|
"loss": 1.3619, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.4325212154393649, |
|
"grad_norm": 0.40654356409937986, |
|
"learning_rate": 1.3990679905014235e-05, |
|
"loss": 1.2983, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.4338899534629072, |
|
"grad_norm": 0.401731234750035, |
|
"learning_rate": 1.3946816534322815e-05, |
|
"loss": 1.3573, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 0.43525869148644947, |
|
"grad_norm": 0.40604771069832035, |
|
"learning_rate": 1.3902863030085176e-05, |
|
"loss": 1.333, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.4366274295099918, |
|
"grad_norm": 0.3821888564061559, |
|
"learning_rate": 1.3858820396068572e-05, |
|
"loss": 1.3062, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 0.4379961675335341, |
|
"grad_norm": 0.3992586017031061, |
|
"learning_rate": 1.3814689638075725e-05, |
|
"loss": 1.3671, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.43936490555707636, |
|
"grad_norm": 0.4018110425948897, |
|
"learning_rate": 1.3770471763921833e-05, |
|
"loss": 1.3709, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 0.4407336435806187, |
|
"grad_norm": 0.387994622633552, |
|
"learning_rate": 1.372616778341158e-05, |
|
"loss": 1.3138, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.442102381604161, |
|
"grad_norm": 0.3974604574879118, |
|
"learning_rate": 1.3681778708316054e-05, |
|
"loss": 1.4087, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 0.44347111962770325, |
|
"grad_norm": 0.39844593969419934, |
|
"learning_rate": 1.3637305552349656e-05, |
|
"loss": 1.3599, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.44483985765124556, |
|
"grad_norm": 0.3857194594528982, |
|
"learning_rate": 1.3592749331146941e-05, |
|
"loss": 1.2897, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.4462085956747878, |
|
"grad_norm": 0.3750305160936592, |
|
"learning_rate": 1.3548111062239432e-05, |
|
"loss": 1.2819, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.44757733369833014, |
|
"grad_norm": 0.4007647993356723, |
|
"learning_rate": 1.350339176503237e-05, |
|
"loss": 1.3194, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 0.44894607172187245, |
|
"grad_norm": 0.4013887994126227, |
|
"learning_rate": 1.3458592460781446e-05, |
|
"loss": 1.3199, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.4503148097454147, |
|
"grad_norm": 0.3981164766269796, |
|
"learning_rate": 1.341371417256947e-05, |
|
"loss": 1.3464, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 0.451683547768957, |
|
"grad_norm": 0.39477092552678616, |
|
"learning_rate": 1.3368757925283015e-05, |
|
"loss": 1.3529, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.45305228579249934, |
|
"grad_norm": 0.3896314211042654, |
|
"learning_rate": 1.3323724745589007e-05, |
|
"loss": 1.3881, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 0.4544210238160416, |
|
"grad_norm": 0.3954835469300208, |
|
"learning_rate": 1.3278615661911274e-05, |
|
"loss": 1.3124, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.4557897618395839, |
|
"grad_norm": 0.3708379387210574, |
|
"learning_rate": 1.3233431704407072e-05, |
|
"loss": 1.2866, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 0.4571584998631262, |
|
"grad_norm": 0.38510327105983166, |
|
"learning_rate": 1.318817390494355e-05, |
|
"loss": 1.2968, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.4585272378866685, |
|
"grad_norm": 0.38362958034859684, |
|
"learning_rate": 1.3142843297074182e-05, |
|
"loss": 1.3155, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.4598959759102108, |
|
"grad_norm": 0.4082500339960413, |
|
"learning_rate": 1.3097440916015179e-05, |
|
"loss": 1.3646, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.46126471393375307, |
|
"grad_norm": 0.3927808793669975, |
|
"learning_rate": 1.3051967798621834e-05, |
|
"loss": 1.3165, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 0.4626334519572954, |
|
"grad_norm": 0.4002328011893471, |
|
"learning_rate": 1.300642498336484e-05, |
|
"loss": 1.3287, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.46400218998083764, |
|
"grad_norm": 0.38096569725907414, |
|
"learning_rate": 1.2960813510306599e-05, |
|
"loss": 1.319, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 0.46537092800437996, |
|
"grad_norm": 0.38869356480860534, |
|
"learning_rate": 1.2915134421077433e-05, |
|
"loss": 1.3763, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.4667396660279223, |
|
"grad_norm": 0.38290589898614796, |
|
"learning_rate": 1.2869388758851828e-05, |
|
"loss": 1.3204, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 0.46810840405146453, |
|
"grad_norm": 0.41880465304512116, |
|
"learning_rate": 1.2823577568324604e-05, |
|
"loss": 1.3858, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.46947714207500685, |
|
"grad_norm": 0.4031243358335159, |
|
"learning_rate": 1.2777701895687034e-05, |
|
"loss": 1.3346, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 0.47084588009854916, |
|
"grad_norm": 0.3919138760842289, |
|
"learning_rate": 1.2731762788602988e-05, |
|
"loss": 1.4046, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.4722146181220914, |
|
"grad_norm": 0.40045145219664335, |
|
"learning_rate": 1.2685761296184987e-05, |
|
"loss": 1.3808, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.47358335614563374, |
|
"grad_norm": 0.38631426280799075, |
|
"learning_rate": 1.2639698468970237e-05, |
|
"loss": 1.3563, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.474952094169176, |
|
"grad_norm": 0.3891224496732836, |
|
"learning_rate": 1.259357535889666e-05, |
|
"loss": 1.368, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 0.4763208321927183, |
|
"grad_norm": 0.40044455029632475, |
|
"learning_rate": 1.2547393019278853e-05, |
|
"loss": 1.3548, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.47768957021626063, |
|
"grad_norm": 0.3899915581292829, |
|
"learning_rate": 1.2501152504784044e-05, |
|
"loss": 1.3291, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 0.4790583082398029, |
|
"grad_norm": 0.3869143385442828, |
|
"learning_rate": 1.2454854871407993e-05, |
|
"loss": 1.3473, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.4804270462633452, |
|
"grad_norm": 0.3824047072867315, |
|
"learning_rate": 1.2408501176450898e-05, |
|
"loss": 1.3196, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 0.48179578428688746, |
|
"grad_norm": 0.4057108906200233, |
|
"learning_rate": 1.2362092478493226e-05, |
|
"loss": 1.3554, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.4831645223104298, |
|
"grad_norm": 0.3776742288321122, |
|
"learning_rate": 1.2315629837371556e-05, |
|
"loss": 1.3014, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 0.4845332603339721, |
|
"grad_norm": 0.3977510874849243, |
|
"learning_rate": 1.2269114314154365e-05, |
|
"loss": 1.3967, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.48590199835751435, |
|
"grad_norm": 0.3697505648941807, |
|
"learning_rate": 1.2222546971117797e-05, |
|
"loss": 1.3171, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.48727073638105667, |
|
"grad_norm": 0.40788370625821474, |
|
"learning_rate": 1.2175928871721411e-05, |
|
"loss": 1.3222, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.488639474404599, |
|
"grad_norm": 0.39090111193092664, |
|
"learning_rate": 1.2129261080583897e-05, |
|
"loss": 1.3676, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 0.49000821242814124, |
|
"grad_norm": 0.3983751803246069, |
|
"learning_rate": 1.2082544663458736e-05, |
|
"loss": 1.3364, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.49137695045168356, |
|
"grad_norm": 0.3848833412781912, |
|
"learning_rate": 1.20357806872099e-05, |
|
"loss": 1.2978, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 0.4927456884752258, |
|
"grad_norm": 0.37897812252926655, |
|
"learning_rate": 1.1988970219787467e-05, |
|
"loss": 1.3333, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.49411442649876813, |
|
"grad_norm": 0.3948100327011319, |
|
"learning_rate": 1.1942114330203227e-05, |
|
"loss": 1.3665, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 0.49548316452231045, |
|
"grad_norm": 0.3844309877664426, |
|
"learning_rate": 1.1895214088506284e-05, |
|
"loss": 1.3383, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.4968519025458527, |
|
"grad_norm": 0.3785571762413702, |
|
"learning_rate": 1.1848270565758616e-05, |
|
"loss": 1.2997, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 0.498220640569395, |
|
"grad_norm": 0.3873700740980167, |
|
"learning_rate": 1.1801284834010596e-05, |
|
"loss": 1.3332, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.49958937859293734, |
|
"grad_norm": 0.3971195600650439, |
|
"learning_rate": 1.1754257966276544e-05, |
|
"loss": 1.3468, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.5009581166164796, |
|
"grad_norm": 0.37924897770225724, |
|
"learning_rate": 1.1707191036510189e-05, |
|
"loss": 1.3152, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.5023268546400219, |
|
"grad_norm": 0.37889309262468557, |
|
"learning_rate": 1.1660085119580165e-05, |
|
"loss": 1.3074, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 0.5036955926635642, |
|
"grad_norm": 0.38526267448234497, |
|
"learning_rate": 1.1612941291245456e-05, |
|
"loss": 1.2863, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.5050643306871064, |
|
"grad_norm": 0.4007327343298376, |
|
"learning_rate": 1.1565760628130824e-05, |
|
"loss": 1.348, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 0.5064330687106487, |
|
"grad_norm": 0.3917232580410847, |
|
"learning_rate": 1.1518544207702238e-05, |
|
"loss": 1.3193, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.5078018067341911, |
|
"grad_norm": 0.41154960813532804, |
|
"learning_rate": 1.1471293108242251e-05, |
|
"loss": 1.3921, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 0.5091705447577334, |
|
"grad_norm": 0.3808027664016428, |
|
"learning_rate": 1.1424008408825383e-05, |
|
"loss": 1.337, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.5105392827812757, |
|
"grad_norm": 0.38767901381709907, |
|
"learning_rate": 1.1376691189293474e-05, |
|
"loss": 1.3142, |
|
"step": 1865 |
|
}, |
|
{ |
|
"epoch": 0.511908020804818, |
|
"grad_norm": 0.39402951617560983, |
|
"learning_rate": 1.1329342530231036e-05, |
|
"loss": 1.358, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.5132767588283602, |
|
"grad_norm": 0.38538496638703307, |
|
"learning_rate": 1.128196351294055e-05, |
|
"loss": 1.2994, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.5146454968519025, |
|
"grad_norm": 0.38082093170114384, |
|
"learning_rate": 1.1234555219417804e-05, |
|
"loss": 1.3208, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.5160142348754448, |
|
"grad_norm": 0.3748307863634792, |
|
"learning_rate": 1.1187118732327167e-05, |
|
"loss": 1.3506, |
|
"step": 1885 |
|
}, |
|
{ |
|
"epoch": 0.5173829728989872, |
|
"grad_norm": 0.38705593276141625, |
|
"learning_rate": 1.1139655134976855e-05, |
|
"loss": 1.2906, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.5187517109225295, |
|
"grad_norm": 0.39888481423655353, |
|
"learning_rate": 1.1092165511294206e-05, |
|
"loss": 1.367, |
|
"step": 1895 |
|
}, |
|
{ |
|
"epoch": 0.5201204489460717, |
|
"grad_norm": 0.3971141048112466, |
|
"learning_rate": 1.104465094580093e-05, |
|
"loss": 1.3405, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.521489186969614, |
|
"grad_norm": 0.3837313790185174, |
|
"learning_rate": 1.0997112523588322e-05, |
|
"loss": 1.3351, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 0.5228579249931563, |
|
"grad_norm": 0.39019433157562, |
|
"learning_rate": 1.0949551330292502e-05, |
|
"loss": 1.2482, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.5242266630166986, |
|
"grad_norm": 0.39062012523465395, |
|
"learning_rate": 1.090196845206961e-05, |
|
"loss": 1.343, |
|
"step": 1915 |
|
}, |
|
{ |
|
"epoch": 0.5255954010402409, |
|
"grad_norm": 0.3948259084399907, |
|
"learning_rate": 1.0854364975571004e-05, |
|
"loss": 1.3001, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.5269641390637831, |
|
"grad_norm": 0.3845620013681871, |
|
"learning_rate": 1.0806741987918448e-05, |
|
"loss": 1.3449, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.5283328770873255, |
|
"grad_norm": 0.3976520791295208, |
|
"learning_rate": 1.075910057667928e-05, |
|
"loss": 1.3007, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.5297016151108678, |
|
"grad_norm": 0.3817709747936168, |
|
"learning_rate": 1.071144182984158e-05, |
|
"loss": 1.3453, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 0.5310703531344101, |
|
"grad_norm": 0.38542609340341144, |
|
"learning_rate": 1.0663766835789327e-05, |
|
"loss": 1.3187, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.5324390911579524, |
|
"grad_norm": 0.3775171570445251, |
|
"learning_rate": 1.0616076683277524e-05, |
|
"loss": 1.3315, |
|
"step": 1945 |
|
}, |
|
{ |
|
"epoch": 0.5338078291814946, |
|
"grad_norm": 0.40024298975169065, |
|
"learning_rate": 1.056837246140736e-05, |
|
"loss": 1.3772, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.5351765672050369, |
|
"grad_norm": 0.3882397104456436, |
|
"learning_rate": 1.0520655259601325e-05, |
|
"loss": 1.3284, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 0.5365453052285792, |
|
"grad_norm": 0.3892143818918897, |
|
"learning_rate": 1.0472926167578323e-05, |
|
"loss": 1.302, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.5379140432521216, |
|
"grad_norm": 0.41282669532669847, |
|
"learning_rate": 1.042518627532881e-05, |
|
"loss": 1.3261, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 0.5392827812756639, |
|
"grad_norm": 0.37154889529358315, |
|
"learning_rate": 1.0377436673089873e-05, |
|
"loss": 1.3142, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.5406515192992062, |
|
"grad_norm": 0.4001478843028637, |
|
"learning_rate": 1.0329678451320352e-05, |
|
"loss": 1.3894, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.5420202573227484, |
|
"grad_norm": 0.4082737912884333, |
|
"learning_rate": 1.0281912700675937e-05, |
|
"loss": 1.38, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.5433889953462907, |
|
"grad_norm": 0.3976694377900842, |
|
"learning_rate": 1.0234140511984246e-05, |
|
"loss": 1.337, |
|
"step": 1985 |
|
}, |
|
{ |
|
"epoch": 0.544757733369833, |
|
"grad_norm": 0.4090490835246688, |
|
"learning_rate": 1.0186362976219926e-05, |
|
"loss": 1.3389, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.5461264713933753, |
|
"grad_norm": 0.39123091240448754, |
|
"learning_rate": 1.0138581184479737e-05, |
|
"loss": 1.2726, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 0.5474952094169176, |
|
"grad_norm": 0.383808685203924, |
|
"learning_rate": 1.0090796227957633e-05, |
|
"loss": 1.3344, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.5488639474404599, |
|
"grad_norm": 0.39282920042398406, |
|
"learning_rate": 1.0043009197919836e-05, |
|
"loss": 1.3188, |
|
"step": 2005 |
|
}, |
|
{ |
|
"epoch": 0.5502326854640022, |
|
"grad_norm": 0.3877735838132022, |
|
"learning_rate": 9.99522118567993e-06, |
|
"loss": 1.3412, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.5516014234875445, |
|
"grad_norm": 0.3855977732261345, |
|
"learning_rate": 9.947433282573926e-06, |
|
"loss": 1.31, |
|
"step": 2015 |
|
}, |
|
{ |
|
"epoch": 0.5529701615110868, |
|
"grad_norm": 0.39182359203627265, |
|
"learning_rate": 9.899646579935336e-06, |
|
"loss": 1.3288, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.5543388995346291, |
|
"grad_norm": 0.37896193597955086, |
|
"learning_rate": 9.85186216907027e-06, |
|
"loss": 1.3222, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 0.5557076375581713, |
|
"grad_norm": 0.39627324276578396, |
|
"learning_rate": 9.80408114123249e-06, |
|
"loss": 1.3659, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.5570763755817136, |
|
"grad_norm": 0.39331176053433153, |
|
"learning_rate": 9.756304587598503e-06, |
|
"loss": 1.4234, |
|
"step": 2035 |
|
}, |
|
{ |
|
"epoch": 0.558445113605256, |
|
"grad_norm": 0.38413569943245757, |
|
"learning_rate": 9.708533599242643e-06, |
|
"loss": 1.2809, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.5598138516287983, |
|
"grad_norm": 0.378050582923272, |
|
"learning_rate": 9.660769267112137e-06, |
|
"loss": 1.3, |
|
"step": 2045 |
|
}, |
|
{ |
|
"epoch": 0.5611825896523406, |
|
"grad_norm": 0.3810287903511996, |
|
"learning_rate": 9.61301268200222e-06, |
|
"loss": 1.328, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.5625513276758828, |
|
"grad_norm": 0.3968883040797777, |
|
"learning_rate": 9.5652649345312e-06, |
|
"loss": 1.3861, |
|
"step": 2055 |
|
}, |
|
{ |
|
"epoch": 0.5639200656994251, |
|
"grad_norm": 0.381754630880566, |
|
"learning_rate": 9.517527115115554e-06, |
|
"loss": 1.2931, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.5652888037229674, |
|
"grad_norm": 0.3934475062551264, |
|
"learning_rate": 9.46980031394504e-06, |
|
"loss": 1.357, |
|
"step": 2065 |
|
}, |
|
{ |
|
"epoch": 0.5666575417465097, |
|
"grad_norm": 0.3764960791780743, |
|
"learning_rate": 9.422085620957795e-06, |
|
"loss": 1.3112, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.568026279770052, |
|
"grad_norm": 0.3819475902039454, |
|
"learning_rate": 9.374384125815427e-06, |
|
"loss": 1.317, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 0.5693950177935944, |
|
"grad_norm": 0.39075935188345634, |
|
"learning_rate": 9.326696917878159e-06, |
|
"loss": 1.329, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.5707637558171366, |
|
"grad_norm": 0.4115410325839856, |
|
"learning_rate": 9.27902508617993e-06, |
|
"loss": 1.3405, |
|
"step": 2085 |
|
}, |
|
{ |
|
"epoch": 0.5721324938406789, |
|
"grad_norm": 0.3929327529237431, |
|
"learning_rate": 9.23136971940353e-06, |
|
"loss": 1.2811, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.5735012318642212, |
|
"grad_norm": 0.38257057379328274, |
|
"learning_rate": 9.183731905855746e-06, |
|
"loss": 1.3234, |
|
"step": 2095 |
|
}, |
|
{ |
|
"epoch": 0.5748699698877635, |
|
"grad_norm": 0.38674820150451145, |
|
"learning_rate": 9.136112733442493e-06, |
|
"loss": 1.3466, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.5762387079113058, |
|
"grad_norm": 0.39368922489348585, |
|
"learning_rate": 9.088513289643982e-06, |
|
"loss": 1.3149, |
|
"step": 2105 |
|
}, |
|
{ |
|
"epoch": 0.577607445934848, |
|
"grad_norm": 0.38199335563674713, |
|
"learning_rate": 9.040934661489885e-06, |
|
"loss": 1.3576, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.5789761839583903, |
|
"grad_norm": 0.39323361965983067, |
|
"learning_rate": 8.9933779355345e-06, |
|
"loss": 1.3159, |
|
"step": 2115 |
|
}, |
|
{ |
|
"epoch": 0.5803449219819327, |
|
"grad_norm": 0.39256654497554794, |
|
"learning_rate": 8.945844197831945e-06, |
|
"loss": 1.3623, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.581713660005475, |
|
"grad_norm": 0.3905410660434604, |
|
"learning_rate": 8.898334533911362e-06, |
|
"loss": 1.3074, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 0.5830823980290173, |
|
"grad_norm": 0.3700729231212675, |
|
"learning_rate": 8.850850028752108e-06, |
|
"loss": 1.3436, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.5844511360525595, |
|
"grad_norm": 0.3701852614302805, |
|
"learning_rate": 8.803391766758998e-06, |
|
"loss": 1.2672, |
|
"step": 2135 |
|
}, |
|
{ |
|
"epoch": 0.5858198740761018, |
|
"grad_norm": 0.3809658344783075, |
|
"learning_rate": 8.755960831737529e-06, |
|
"loss": 1.3638, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.5871886120996441, |
|
"grad_norm": 0.3852501406589481, |
|
"learning_rate": 8.708558306869125e-06, |
|
"loss": 1.3389, |
|
"step": 2145 |
|
}, |
|
{ |
|
"epoch": 0.5885573501231864, |
|
"grad_norm": 0.4003501899309122, |
|
"learning_rate": 8.661185274686418e-06, |
|
"loss": 1.3228, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.5899260881467288, |
|
"grad_norm": 0.38933438340786924, |
|
"learning_rate": 8.613842817048503e-06, |
|
"loss": 1.3627, |
|
"step": 2155 |
|
}, |
|
{ |
|
"epoch": 0.591294826170271, |
|
"grad_norm": 0.39608781513600605, |
|
"learning_rate": 8.566532015116251e-06, |
|
"loss": 1.3626, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.5926635641938133, |
|
"grad_norm": 0.37582004076749886, |
|
"learning_rate": 8.51925394932761e-06, |
|
"loss": 1.2766, |
|
"step": 2165 |
|
}, |
|
{ |
|
"epoch": 0.5940323022173556, |
|
"grad_norm": 0.39046439420533485, |
|
"learning_rate": 8.47200969937293e-06, |
|
"loss": 1.3342, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.5954010402408979, |
|
"grad_norm": 0.3846853640345239, |
|
"learning_rate": 8.424800344170308e-06, |
|
"loss": 1.2973, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 0.5967697782644402, |
|
"grad_norm": 0.3794031254673053, |
|
"learning_rate": 8.37762696184095e-06, |
|
"loss": 1.3288, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.5981385162879824, |
|
"grad_norm": 0.38487683969290326, |
|
"learning_rate": 8.330490629684547e-06, |
|
"loss": 1.3374, |
|
"step": 2185 |
|
}, |
|
{ |
|
"epoch": 0.5995072543115247, |
|
"grad_norm": 0.38666671699682115, |
|
"learning_rate": 8.28339242415468e-06, |
|
"loss": 1.3202, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.600875992335067, |
|
"grad_norm": 0.3788541652969095, |
|
"learning_rate": 8.236333420834216e-06, |
|
"loss": 1.2778, |
|
"step": 2195 |
|
}, |
|
{ |
|
"epoch": 0.6022447303586094, |
|
"grad_norm": 0.4003559394681289, |
|
"learning_rate": 8.189314694410781e-06, |
|
"loss": 1.3349, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.6036134683821517, |
|
"grad_norm": 0.40672129713209243, |
|
"learning_rate": 8.14233731865218e-06, |
|
"loss": 1.3343, |
|
"step": 2205 |
|
}, |
|
{ |
|
"epoch": 0.604982206405694, |
|
"grad_norm": 0.392344465653996, |
|
"learning_rate": 8.0954023663819e-06, |
|
"loss": 1.2804, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.6063509444292362, |
|
"grad_norm": 0.396931681160349, |
|
"learning_rate": 8.048510909454601e-06, |
|
"loss": 1.3712, |
|
"step": 2215 |
|
}, |
|
{ |
|
"epoch": 0.6077196824527785, |
|
"grad_norm": 0.3882127247657791, |
|
"learning_rate": 8.001664018731637e-06, |
|
"loss": 1.3313, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.6090884204763208, |
|
"grad_norm": 0.3727086035483801, |
|
"learning_rate": 7.954862764056602e-06, |
|
"loss": 1.2715, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 0.6104571584998632, |
|
"grad_norm": 0.39519191088539024, |
|
"learning_rate": 7.908108214230902e-06, |
|
"loss": 1.3018, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.6118258965234055, |
|
"grad_norm": 0.39954515244878724, |
|
"learning_rate": 7.861401436989337e-06, |
|
"loss": 1.3697, |
|
"step": 2235 |
|
}, |
|
{ |
|
"epoch": 0.6131946345469477, |
|
"grad_norm": 0.3961862899698687, |
|
"learning_rate": 7.814743498975731e-06, |
|
"loss": 1.3651, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.61456337257049, |
|
"grad_norm": 0.38565731038310297, |
|
"learning_rate": 7.768135465718559e-06, |
|
"loss": 1.3462, |
|
"step": 2245 |
|
}, |
|
{ |
|
"epoch": 0.6159321105940323, |
|
"grad_norm": 0.38796255525719986, |
|
"learning_rate": 7.72157840160662e-06, |
|
"loss": 1.3087, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.6173008486175746, |
|
"grad_norm": 0.3947254708668997, |
|
"learning_rate": 7.67507336986474e-06, |
|
"loss": 1.262, |
|
"step": 2255 |
|
}, |
|
{ |
|
"epoch": 0.6186695866411169, |
|
"grad_norm": 0.39167259958488637, |
|
"learning_rate": 7.628621432529461e-06, |
|
"loss": 1.3117, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.6200383246646591, |
|
"grad_norm": 0.39458361491505556, |
|
"learning_rate": 7.582223650424825e-06, |
|
"loss": 1.3067, |
|
"step": 2265 |
|
}, |
|
{ |
|
"epoch": 0.6214070626882015, |
|
"grad_norm": 0.3948240515553033, |
|
"learning_rate": 7.5358810831381225e-06, |
|
"loss": 1.3177, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.6227758007117438, |
|
"grad_norm": 0.3814626352546649, |
|
"learning_rate": 7.489594788995698e-06, |
|
"loss": 1.3259, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 0.6241445387352861, |
|
"grad_norm": 0.39070667312634827, |
|
"learning_rate": 7.443365825038793e-06, |
|
"loss": 1.3212, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.6255132767588284, |
|
"grad_norm": 0.38913877038236866, |
|
"learning_rate": 7.397195246999391e-06, |
|
"loss": 1.346, |
|
"step": 2285 |
|
}, |
|
{ |
|
"epoch": 0.6268820147823706, |
|
"grad_norm": 0.4070153012541291, |
|
"learning_rate": 7.351084109276119e-06, |
|
"loss": 1.3826, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.6282507528059129, |
|
"grad_norm": 0.36428744429147014, |
|
"learning_rate": 7.305033464910164e-06, |
|
"loss": 1.2659, |
|
"step": 2295 |
|
}, |
|
{ |
|
"epoch": 0.6296194908294552, |
|
"grad_norm": 0.3877290507395246, |
|
"learning_rate": 7.25904436556122e-06, |
|
"loss": 1.304, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.6309882288529975, |
|
"grad_norm": 0.3803920075270922, |
|
"learning_rate": 7.21311786148348e-06, |
|
"loss": 1.3333, |
|
"step": 2305 |
|
}, |
|
{ |
|
"epoch": 0.6323569668765399, |
|
"grad_norm": 0.3822541348646928, |
|
"learning_rate": 7.167255001501651e-06, |
|
"loss": 1.3307, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.6337257049000822, |
|
"grad_norm": 0.3977624877197784, |
|
"learning_rate": 7.121456832986988e-06, |
|
"loss": 1.3329, |
|
"step": 2315 |
|
}, |
|
{ |
|
"epoch": 0.6350944429236244, |
|
"grad_norm": 0.38990040940646387, |
|
"learning_rate": 7.075724401833395e-06, |
|
"loss": 1.3647, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.6364631809471667, |
|
"grad_norm": 0.39550975070062716, |
|
"learning_rate": 7.030058752433526e-06, |
|
"loss": 1.2715, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 0.637831918970709, |
|
"grad_norm": 0.3857148917491234, |
|
"learning_rate": 6.984460927654937e-06, |
|
"loss": 1.3191, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.6392006569942513, |
|
"grad_norm": 0.38855941306596253, |
|
"learning_rate": 6.938931968816275e-06, |
|
"loss": 1.2979, |
|
"step": 2335 |
|
}, |
|
{ |
|
"epoch": 0.6405693950177936, |
|
"grad_norm": 0.3892088605550479, |
|
"learning_rate": 6.893472915663493e-06, |
|
"loss": 1.3826, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.6419381330413358, |
|
"grad_norm": 0.4024108620394257, |
|
"learning_rate": 6.8480848063461035e-06, |
|
"loss": 1.2943, |
|
"step": 2345 |
|
}, |
|
{ |
|
"epoch": 0.6433068710648782, |
|
"grad_norm": 0.3976600247063458, |
|
"learning_rate": 6.8027686773934765e-06, |
|
"loss": 1.3241, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.6446756090884205, |
|
"grad_norm": 0.38538373822690386, |
|
"learning_rate": 6.7575255636911626e-06, |
|
"loss": 1.3316, |
|
"step": 2355 |
|
}, |
|
{ |
|
"epoch": 0.6460443471119628, |
|
"grad_norm": 0.3961847974725629, |
|
"learning_rate": 6.7123564984572596e-06, |
|
"loss": 1.3481, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.6474130851355051, |
|
"grad_norm": 0.3739018408786073, |
|
"learning_rate": 6.667262513218824e-06, |
|
"loss": 1.2857, |
|
"step": 2365 |
|
}, |
|
{ |
|
"epoch": 0.6487818231590473, |
|
"grad_norm": 0.4058027362822848, |
|
"learning_rate": 6.622244637788302e-06, |
|
"loss": 1.3623, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.6501505611825896, |
|
"grad_norm": 0.3887614780397367, |
|
"learning_rate": 6.577303900240023e-06, |
|
"loss": 1.3466, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 0.6515192992061319, |
|
"grad_norm": 0.39583219449776136, |
|
"learning_rate": 6.532441326886716e-06, |
|
"loss": 1.3826, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.6528880372296743, |
|
"grad_norm": 0.38600694323078566, |
|
"learning_rate": 6.487657942256069e-06, |
|
"loss": 1.3124, |
|
"step": 2385 |
|
}, |
|
{ |
|
"epoch": 0.6542567752532166, |
|
"grad_norm": 0.35581732986644143, |
|
"learning_rate": 6.442954769067341e-06, |
|
"loss": 1.3352, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.6556255132767588, |
|
"grad_norm": 0.3977530272775962, |
|
"learning_rate": 6.398332828207996e-06, |
|
"loss": 1.3226, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.6569942513003011, |
|
"grad_norm": 0.3928175408666658, |
|
"learning_rate": 6.3537931387103925e-06, |
|
"loss": 1.3262, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.6583629893238434, |
|
"grad_norm": 0.3795840583774513, |
|
"learning_rate": 6.309336717728516e-06, |
|
"loss": 1.3203, |
|
"step": 2405 |
|
}, |
|
{ |
|
"epoch": 0.6597317273473857, |
|
"grad_norm": 0.396679060826067, |
|
"learning_rate": 6.264964580514745e-06, |
|
"loss": 1.2842, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.661100465370928, |
|
"grad_norm": 0.3699325537006795, |
|
"learning_rate": 6.220677740396668e-06, |
|
"loss": 1.2936, |
|
"step": 2415 |
|
}, |
|
{ |
|
"epoch": 0.6624692033944704, |
|
"grad_norm": 0.38098356916961845, |
|
"learning_rate": 6.176477208753944e-06, |
|
"loss": 1.276, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.6638379414180126, |
|
"grad_norm": 0.38873579028440075, |
|
"learning_rate": 6.132363994995194e-06, |
|
"loss": 1.2348, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 0.6652066794415549, |
|
"grad_norm": 0.39405073870487267, |
|
"learning_rate": 6.088339106534971e-06, |
|
"loss": 1.2659, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.6665754174650972, |
|
"grad_norm": 0.3833620202872981, |
|
"learning_rate": 6.044403548770735e-06, |
|
"loss": 1.3103, |
|
"step": 2435 |
|
}, |
|
{ |
|
"epoch": 0.6679441554886395, |
|
"grad_norm": 0.39527725366864974, |
|
"learning_rate": 6.000558325059894e-06, |
|
"loss": 1.3423, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.6693128935121818, |
|
"grad_norm": 0.38996615233504806, |
|
"learning_rate": 5.956804436696904e-06, |
|
"loss": 1.303, |
|
"step": 2445 |
|
}, |
|
{ |
|
"epoch": 0.670681631535724, |
|
"grad_norm": 0.3943489977366666, |
|
"learning_rate": 5.9131428828903905e-06, |
|
"loss": 1.3547, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.6720503695592663, |
|
"grad_norm": 0.3899208460332646, |
|
"learning_rate": 5.8695746607403285e-06, |
|
"loss": 1.2969, |
|
"step": 2455 |
|
}, |
|
{ |
|
"epoch": 0.6734191075828087, |
|
"grad_norm": 0.3940611248671062, |
|
"learning_rate": 5.826100765215273e-06, |
|
"loss": 1.318, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.674787845606351, |
|
"grad_norm": 0.3930083338553073, |
|
"learning_rate": 5.782722189129655e-06, |
|
"loss": 1.2906, |
|
"step": 2465 |
|
}, |
|
{ |
|
"epoch": 0.6761565836298933, |
|
"grad_norm": 0.39311675749157776, |
|
"learning_rate": 5.739439923121077e-06, |
|
"loss": 1.3569, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.6775253216534355, |
|
"grad_norm": 0.4114954929599597, |
|
"learning_rate": 5.6962549556277134e-06, |
|
"loss": 1.3773, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 0.6788940596769778, |
|
"grad_norm": 0.38652001787654966, |
|
"learning_rate": 5.653168272865724e-06, |
|
"loss": 1.3008, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.6802627977005201, |
|
"grad_norm": 0.3878760723325068, |
|
"learning_rate": 5.6101808588067505e-06, |
|
"loss": 1.2992, |
|
"step": 2485 |
|
}, |
|
{ |
|
"epoch": 0.6816315357240624, |
|
"grad_norm": 0.39884745029373364, |
|
"learning_rate": 5.56729369515542e-06, |
|
"loss": 1.3088, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.6830002737476047, |
|
"grad_norm": 0.38961302860625757, |
|
"learning_rate": 5.52450776132694e-06, |
|
"loss": 1.2904, |
|
"step": 2495 |
|
}, |
|
{ |
|
"epoch": 0.684369011771147, |
|
"grad_norm": 0.3970778873479824, |
|
"learning_rate": 5.481824034424741e-06, |
|
"loss": 1.3464, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.6857377497946893, |
|
"grad_norm": 0.4014161218820373, |
|
"learning_rate": 5.439243489218138e-06, |
|
"loss": 1.3376, |
|
"step": 2505 |
|
}, |
|
{ |
|
"epoch": 0.6871064878182316, |
|
"grad_norm": 0.3874962119825426, |
|
"learning_rate": 5.396767098120087e-06, |
|
"loss": 1.3078, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.6884752258417739, |
|
"grad_norm": 0.37396226223807555, |
|
"learning_rate": 5.354395831164982e-06, |
|
"loss": 1.3097, |
|
"step": 2515 |
|
}, |
|
{ |
|
"epoch": 0.6898439638653162, |
|
"grad_norm": 0.3792883560955228, |
|
"learning_rate": 5.312130655986485e-06, |
|
"loss": 1.3158, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.6912127018888585, |
|
"grad_norm": 0.3798203691318429, |
|
"learning_rate": 5.269972537795434e-06, |
|
"loss": 1.3087, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 0.6925814399124007, |
|
"grad_norm": 0.38253235315284295, |
|
"learning_rate": 5.227922439357823e-06, |
|
"loss": 1.3198, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.693950177935943, |
|
"grad_norm": 0.3817014501814054, |
|
"learning_rate": 5.1859813209727775e-06, |
|
"loss": 1.321, |
|
"step": 2535 |
|
}, |
|
{ |
|
"epoch": 0.6953189159594854, |
|
"grad_norm": 0.3950376035714224, |
|
"learning_rate": 5.144150140450643e-06, |
|
"loss": 1.3024, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.6966876539830277, |
|
"grad_norm": 0.4084629435066235, |
|
"learning_rate": 5.102429853091128e-06, |
|
"loss": 1.3437, |
|
"step": 2545 |
|
}, |
|
{ |
|
"epoch": 0.69805639200657, |
|
"grad_norm": 0.3808185831587325, |
|
"learning_rate": 5.060821411661459e-06, |
|
"loss": 1.3264, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.6994251300301122, |
|
"grad_norm": 0.3937638284493571, |
|
"learning_rate": 5.019325766374625e-06, |
|
"loss": 1.314, |
|
"step": 2555 |
|
}, |
|
{ |
|
"epoch": 0.7007938680536545, |
|
"grad_norm": 0.3920386340433558, |
|
"learning_rate": 4.977943864867712e-06, |
|
"loss": 1.3271, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.7021626060771968, |
|
"grad_norm": 0.3898729624487372, |
|
"learning_rate": 4.936676652180215e-06, |
|
"loss": 1.3467, |
|
"step": 2565 |
|
}, |
|
{ |
|
"epoch": 0.7035313441007391, |
|
"grad_norm": 0.39902316286587364, |
|
"learning_rate": 4.89552507073248e-06, |
|
"loss": 1.3102, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.7049000821242815, |
|
"grad_norm": 0.3673934948431135, |
|
"learning_rate": 4.854490060304192e-06, |
|
"loss": 1.2612, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 0.7062688201478237, |
|
"grad_norm": 0.3922497888421968, |
|
"learning_rate": 4.813572558012892e-06, |
|
"loss": 1.329, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.707637558171366, |
|
"grad_norm": 0.3860080742782407, |
|
"learning_rate": 4.772773498292579e-06, |
|
"loss": 1.3332, |
|
"step": 2585 |
|
}, |
|
{ |
|
"epoch": 0.7090062961949083, |
|
"grad_norm": 0.3695087849805836, |
|
"learning_rate": 4.732093812872391e-06, |
|
"loss": 1.2697, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.7103750342184506, |
|
"grad_norm": 0.4012769440751328, |
|
"learning_rate": 4.691534430755302e-06, |
|
"loss": 1.3213, |
|
"step": 2595 |
|
}, |
|
{ |
|
"epoch": 0.7117437722419929, |
|
"grad_norm": 0.38657730746893026, |
|
"learning_rate": 4.651096278196916e-06, |
|
"loss": 1.3665, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.7131125102655351, |
|
"grad_norm": 0.38560469814966725, |
|
"learning_rate": 4.610780278684315e-06, |
|
"loss": 1.3369, |
|
"step": 2605 |
|
}, |
|
{ |
|
"epoch": 0.7144812482890774, |
|
"grad_norm": 0.39542221048440374, |
|
"learning_rate": 4.570587352914977e-06, |
|
"loss": 1.3048, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.7158499863126198, |
|
"grad_norm": 0.3910002548641959, |
|
"learning_rate": 4.530518418775734e-06, |
|
"loss": 1.3247, |
|
"step": 2615 |
|
}, |
|
{ |
|
"epoch": 0.7172187243361621, |
|
"grad_norm": 0.3825608407076146, |
|
"learning_rate": 4.490574391321814e-06, |
|
"loss": 1.2967, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.7185874623597044, |
|
"grad_norm": 0.4028460389272689, |
|
"learning_rate": 4.450756182755963e-06, |
|
"loss": 1.3671, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 0.7199562003832466, |
|
"grad_norm": 0.40485919090431155, |
|
"learning_rate": 4.411064702407585e-06, |
|
"loss": 1.3556, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.7213249384067889, |
|
"grad_norm": 0.39021669719654245, |
|
"learning_rate": 4.371500856711988e-06, |
|
"loss": 1.3774, |
|
"step": 2635 |
|
}, |
|
{ |
|
"epoch": 0.7226936764303312, |
|
"grad_norm": 0.3958279714364836, |
|
"learning_rate": 4.332065549189697e-06, |
|
"loss": 1.2982, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.7240624144538735, |
|
"grad_norm": 0.4060690524084766, |
|
"learning_rate": 4.292759680425794e-06, |
|
"loss": 1.3767, |
|
"step": 2645 |
|
}, |
|
{ |
|
"epoch": 0.7254311524774159, |
|
"grad_norm": 0.40926421108177974, |
|
"learning_rate": 4.253584148049369e-06, |
|
"loss": 1.339, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.7267998905009582, |
|
"grad_norm": 0.37375698755072967, |
|
"learning_rate": 4.214539846713024e-06, |
|
"loss": 1.2961, |
|
"step": 2655 |
|
}, |
|
{ |
|
"epoch": 0.7281686285245004, |
|
"grad_norm": 0.3968852233360893, |
|
"learning_rate": 4.175627668072425e-06, |
|
"loss": 1.2962, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.7295373665480427, |
|
"grad_norm": 0.4073445114415598, |
|
"learning_rate": 4.136848500765948e-06, |
|
"loss": 1.4056, |
|
"step": 2665 |
|
}, |
|
{ |
|
"epoch": 0.730906104571585, |
|
"grad_norm": 0.3950383171619063, |
|
"learning_rate": 4.098203230394399e-06, |
|
"loss": 1.3192, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.7322748425951273, |
|
"grad_norm": 0.3986322017273241, |
|
"learning_rate": 4.059692739500761e-06, |
|
"loss": 1.3485, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 0.7336435806186696, |
|
"grad_norm": 0.3848930065792978, |
|
"learning_rate": 4.02131790755006e-06, |
|
"loss": 1.378, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.7350123186422118, |
|
"grad_norm": 0.39139666434515563, |
|
"learning_rate": 3.983079610909283e-06, |
|
"loss": 1.3334, |
|
"step": 2685 |
|
}, |
|
{ |
|
"epoch": 0.7363810566657542, |
|
"grad_norm": 0.38858454704969564, |
|
"learning_rate": 3.944978722827347e-06, |
|
"loss": 1.3235, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.7377497946892965, |
|
"grad_norm": 0.390122324459421, |
|
"learning_rate": 3.907016113415166e-06, |
|
"loss": 1.3583, |
|
"step": 2695 |
|
}, |
|
{ |
|
"epoch": 0.7391185327128388, |
|
"grad_norm": 0.3919502066927978, |
|
"learning_rate": 3.869192649625792e-06, |
|
"loss": 1.3092, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.7404872707363811, |
|
"grad_norm": 0.39777490757009926, |
|
"learning_rate": 3.831509195234598e-06, |
|
"loss": 1.3358, |
|
"step": 2705 |
|
}, |
|
{ |
|
"epoch": 0.7418560087599233, |
|
"grad_norm": 0.39353868348346743, |
|
"learning_rate": 3.793966610819545e-06, |
|
"loss": 1.3515, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.7432247467834656, |
|
"grad_norm": 0.39342447998988617, |
|
"learning_rate": 3.756565753741569e-06, |
|
"loss": 1.2885, |
|
"step": 2715 |
|
}, |
|
{ |
|
"epoch": 0.7445934848070079, |
|
"grad_norm": 0.37781062124467246, |
|
"learning_rate": 3.7193074781249585e-06, |
|
"loss": 1.276, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.7459622228305502, |
|
"grad_norm": 0.38397833073923665, |
|
"learning_rate": 3.6821926348378666e-06, |
|
"loss": 1.3031, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 0.7473309608540926, |
|
"grad_norm": 0.3993776914034358, |
|
"learning_rate": 3.6452220714728883e-06, |
|
"loss": 1.3719, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.7486996988776348, |
|
"grad_norm": 0.387591612397714, |
|
"learning_rate": 3.608396632327684e-06, |
|
"loss": 1.3006, |
|
"step": 2735 |
|
}, |
|
{ |
|
"epoch": 0.7500684369011771, |
|
"grad_norm": 0.4059102325008591, |
|
"learning_rate": 3.5717171583857115e-06, |
|
"loss": 1.3689, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.7514371749247194, |
|
"grad_norm": 0.3772375625162727, |
|
"learning_rate": 3.5351844872970233e-06, |
|
"loss": 1.3091, |
|
"step": 2745 |
|
}, |
|
{ |
|
"epoch": 0.7528059129482617, |
|
"grad_norm": 0.3853940006094772, |
|
"learning_rate": 3.498799453359124e-06, |
|
"loss": 1.3454, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.754174650971804, |
|
"grad_norm": 0.3771144405512011, |
|
"learning_rate": 3.462562887497927e-06, |
|
"loss": 1.2998, |
|
"step": 2755 |
|
}, |
|
{ |
|
"epoch": 0.7555433889953463, |
|
"grad_norm": 0.39604150906533137, |
|
"learning_rate": 3.4264756172487813e-06, |
|
"loss": 1.3129, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.7569121270188885, |
|
"grad_norm": 0.3843657920576998, |
|
"learning_rate": 3.390538466737564e-06, |
|
"loss": 1.2981, |
|
"step": 2765 |
|
}, |
|
{ |
|
"epoch": 0.7582808650424309, |
|
"grad_norm": 0.39238834291334906, |
|
"learning_rate": 3.3547522566618593e-06, |
|
"loss": 1.2697, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.7596496030659732, |
|
"grad_norm": 0.4030855832137369, |
|
"learning_rate": 3.319117804272236e-06, |
|
"loss": 1.3152, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 0.7610183410895155, |
|
"grad_norm": 0.38853914068753576, |
|
"learning_rate": 3.283635923353553e-06, |
|
"loss": 1.3125, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.7623870791130578, |
|
"grad_norm": 0.39575307616938343, |
|
"learning_rate": 3.248307424206395e-06, |
|
"loss": 1.4215, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.7637558171366, |
|
"grad_norm": 0.3903008557955737, |
|
"learning_rate": 3.2131331136285717e-06, |
|
"loss": 1.3034, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.7651245551601423, |
|
"grad_norm": 0.38760587663102863, |
|
"learning_rate": 3.1781137948966754e-06, |
|
"loss": 1.291, |
|
"step": 2795 |
|
}, |
|
{ |
|
"epoch": 0.7664932931836846, |
|
"grad_norm": 0.3917779069091076, |
|
"learning_rate": 3.1432502677477494e-06, |
|
"loss": 1.3318, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.767862031207227, |
|
"grad_norm": 0.3978532288818969, |
|
"learning_rate": 3.108543328361017e-06, |
|
"loss": 1.3074, |
|
"step": 2805 |
|
}, |
|
{ |
|
"epoch": 0.7692307692307693, |
|
"grad_norm": 0.3931552217963087, |
|
"learning_rate": 3.0739937693397113e-06, |
|
"loss": 1.3181, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.7705995072543115, |
|
"grad_norm": 0.38132685696899987, |
|
"learning_rate": 3.0396023796929597e-06, |
|
"loss": 1.3202, |
|
"step": 2815 |
|
}, |
|
{ |
|
"epoch": 0.7719682452778538, |
|
"grad_norm": 0.3895674318092009, |
|
"learning_rate": 3.0053699448177687e-06, |
|
"loss": 1.2912, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.7733369833013961, |
|
"grad_norm": 0.3923145835781509, |
|
"learning_rate": 2.971297246481101e-06, |
|
"loss": 1.2874, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 0.7747057213249384, |
|
"grad_norm": 0.3964424700594031, |
|
"learning_rate": 2.937385062802004e-06, |
|
"loss": 1.3723, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.7760744593484807, |
|
"grad_norm": 0.39752354179970156, |
|
"learning_rate": 2.9036341682338466e-06, |
|
"loss": 1.3329, |
|
"step": 2835 |
|
}, |
|
{ |
|
"epoch": 0.7774431973720229, |
|
"grad_norm": 0.39323110376722525, |
|
"learning_rate": 2.870045333546644e-06, |
|
"loss": 1.2905, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.7788119353955653, |
|
"grad_norm": 0.3924911670687984, |
|
"learning_rate": 2.8366193258094355e-06, |
|
"loss": 1.292, |
|
"step": 2845 |
|
}, |
|
{ |
|
"epoch": 0.7801806734191076, |
|
"grad_norm": 0.372337059578975, |
|
"learning_rate": 2.8033569083727797e-06, |
|
"loss": 1.3368, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.7815494114426499, |
|
"grad_norm": 0.386256983183401, |
|
"learning_rate": 2.7702588408513276e-06, |
|
"loss": 1.3236, |
|
"step": 2855 |
|
}, |
|
{ |
|
"epoch": 0.7829181494661922, |
|
"grad_norm": 0.4151085195929508, |
|
"learning_rate": 2.7373258791064572e-06, |
|
"loss": 1.3342, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.7842868874897345, |
|
"grad_norm": 0.39410387314037304, |
|
"learning_rate": 2.7045587752290224e-06, |
|
"loss": 1.28, |
|
"step": 2865 |
|
}, |
|
{ |
|
"epoch": 0.7856556255132767, |
|
"grad_norm": 0.40070665770653285, |
|
"learning_rate": 2.6719582775221862e-06, |
|
"loss": 1.3336, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.787024363536819, |
|
"grad_norm": 0.3842777541641571, |
|
"learning_rate": 2.6395251304843137e-06, |
|
"loss": 1.2757, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 0.7883931015603614, |
|
"grad_norm": 0.40580643578053277, |
|
"learning_rate": 2.6072600747919773e-06, |
|
"loss": 1.3196, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.7897618395839037, |
|
"grad_norm": 0.388179506322075, |
|
"learning_rate": 2.575163847283053e-06, |
|
"loss": 1.293, |
|
"step": 2885 |
|
}, |
|
{ |
|
"epoch": 0.791130577607446, |
|
"grad_norm": 0.3732450173203748, |
|
"learning_rate": 2.543237180939875e-06, |
|
"loss": 1.2751, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.7924993156309882, |
|
"grad_norm": 0.38506878802774175, |
|
"learning_rate": 2.5114808048725035e-06, |
|
"loss": 1.3083, |
|
"step": 2895 |
|
}, |
|
{ |
|
"epoch": 0.7938680536545305, |
|
"grad_norm": 0.4075860384902042, |
|
"learning_rate": 2.479895444302086e-06, |
|
"loss": 1.307, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.7952367916780728, |
|
"grad_norm": 0.39287683770161946, |
|
"learning_rate": 2.4484818205442763e-06, |
|
"loss": 1.3645, |
|
"step": 2905 |
|
}, |
|
{ |
|
"epoch": 0.7966055297016151, |
|
"grad_norm": 0.3882005380231924, |
|
"learning_rate": 2.417240650992767e-06, |
|
"loss": 1.3512, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.7979742677251574, |
|
"grad_norm": 0.3866431713970654, |
|
"learning_rate": 2.3861726491029237e-06, |
|
"loss": 1.2793, |
|
"step": 2915 |
|
}, |
|
{ |
|
"epoch": 0.7993430057486997, |
|
"grad_norm": 0.3798968085022952, |
|
"learning_rate": 2.355278524375465e-06, |
|
"loss": 1.2865, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.800711743772242, |
|
"grad_norm": 0.39453328832547074, |
|
"learning_rate": 2.324558982340275e-06, |
|
"loss": 1.2576, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 0.8020804817957843, |
|
"grad_norm": 0.3764906658011197, |
|
"learning_rate": 2.2940147245402944e-06, |
|
"loss": 1.3167, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.8034492198193266, |
|
"grad_norm": 0.3823354264503374, |
|
"learning_rate": 2.2636464485154875e-06, |
|
"loss": 1.334, |
|
"step": 2935 |
|
}, |
|
{ |
|
"epoch": 0.8048179578428689, |
|
"grad_norm": 0.3911098766218853, |
|
"learning_rate": 2.23345484778692e-06, |
|
"loss": 1.3856, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.8061866958664111, |
|
"grad_norm": 0.4024307217930752, |
|
"learning_rate": 2.2034406118409178e-06, |
|
"loss": 1.3704, |
|
"step": 2945 |
|
}, |
|
{ |
|
"epoch": 0.8075554338899534, |
|
"grad_norm": 0.38312668558103635, |
|
"learning_rate": 2.1736044261133305e-06, |
|
"loss": 1.32, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.8089241719134957, |
|
"grad_norm": 0.39944235808095113, |
|
"learning_rate": 2.1439469719738615e-06, |
|
"loss": 1.3348, |
|
"step": 2955 |
|
}, |
|
{ |
|
"epoch": 0.8102929099370381, |
|
"grad_norm": 0.37362368240209753, |
|
"learning_rate": 2.1144689267105213e-06, |
|
"loss": 1.3058, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.8116616479605804, |
|
"grad_norm": 0.3977426993128578, |
|
"learning_rate": 2.0851709635141526e-06, |
|
"loss": 1.329, |
|
"step": 2965 |
|
}, |
|
{ |
|
"epoch": 0.8130303859841227, |
|
"grad_norm": 0.390250388707541, |
|
"learning_rate": 2.0560537514630595e-06, |
|
"loss": 1.3435, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.8143991240076649, |
|
"grad_norm": 0.41070469591969644, |
|
"learning_rate": 2.0271179555077357e-06, |
|
"loss": 1.3172, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 0.8157678620312072, |
|
"grad_norm": 0.378348117809942, |
|
"learning_rate": 1.998364236455661e-06, |
|
"loss": 1.3027, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.8171366000547495, |
|
"grad_norm": 0.380190609171277, |
|
"learning_rate": 1.969793250956221e-06, |
|
"loss": 1.2577, |
|
"step": 2985 |
|
}, |
|
{ |
|
"epoch": 0.8185053380782918, |
|
"grad_norm": 0.38655656053220244, |
|
"learning_rate": 1.9414056514857205e-06, |
|
"loss": 1.3137, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.8198740761018342, |
|
"grad_norm": 0.38867016212920535, |
|
"learning_rate": 1.913202086332463e-06, |
|
"loss": 1.3597, |
|
"step": 2995 |
|
}, |
|
{ |
|
"epoch": 0.8212428141253764, |
|
"grad_norm": 0.38641157633799067, |
|
"learning_rate": 1.8851831995819569e-06, |
|
"loss": 1.3184, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.8226115521489187, |
|
"grad_norm": 0.39081114681314566, |
|
"learning_rate": 1.8573496311022133e-06, |
|
"loss": 1.3219, |
|
"step": 3005 |
|
}, |
|
{ |
|
"epoch": 0.823980290172461, |
|
"grad_norm": 0.37733544335013924, |
|
"learning_rate": 1.8297020165291158e-06, |
|
"loss": 1.2927, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.8253490281960033, |
|
"grad_norm": 0.38748075262103254, |
|
"learning_rate": 1.8022409872519197e-06, |
|
"loss": 1.3184, |
|
"step": 3015 |
|
}, |
|
{ |
|
"epoch": 0.8267177662195456, |
|
"grad_norm": 0.40138106139790847, |
|
"learning_rate": 1.7749671703988226e-06, |
|
"loss": 1.3312, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.8280865042430878, |
|
"grad_norm": 0.3876362883781664, |
|
"learning_rate": 1.7478811888226555e-06, |
|
"loss": 1.3101, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 0.8294552422666301, |
|
"grad_norm": 0.3936459314036851, |
|
"learning_rate": 1.7209836610866426e-06, |
|
"loss": 1.3431, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.8308239802901725, |
|
"grad_norm": 0.38764243587502206, |
|
"learning_rate": 1.694275201450284e-06, |
|
"loss": 1.3386, |
|
"step": 3035 |
|
}, |
|
{ |
|
"epoch": 0.8321927183137148, |
|
"grad_norm": 0.3908568598595201, |
|
"learning_rate": 1.6677564198553332e-06, |
|
"loss": 1.3342, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.8335614563372571, |
|
"grad_norm": 0.38588972696708185, |
|
"learning_rate": 1.6414279219118568e-06, |
|
"loss": 1.3527, |
|
"step": 3045 |
|
}, |
|
{ |
|
"epoch": 0.8349301943607993, |
|
"grad_norm": 0.3907806499233958, |
|
"learning_rate": 1.6152903088844051e-06, |
|
"loss": 1.3104, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.8362989323843416, |
|
"grad_norm": 0.3825824287507951, |
|
"learning_rate": 1.5893441776782947e-06, |
|
"loss": 1.3062, |
|
"step": 3055 |
|
}, |
|
{ |
|
"epoch": 0.8376676704078839, |
|
"grad_norm": 0.3842728055099776, |
|
"learning_rate": 1.5635901208259608e-06, |
|
"loss": 1.3581, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.8390364084314262, |
|
"grad_norm": 0.3905623616859072, |
|
"learning_rate": 1.5380287264734285e-06, |
|
"loss": 1.3148, |
|
"step": 3065 |
|
}, |
|
{ |
|
"epoch": 0.8404051464549686, |
|
"grad_norm": 0.39844460827488454, |
|
"learning_rate": 1.5126605783668945e-06, |
|
"loss": 1.3074, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.8417738844785108, |
|
"grad_norm": 0.39217288885390583, |
|
"learning_rate": 1.4874862558393787e-06, |
|
"loss": 1.3171, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 0.8431426225020531, |
|
"grad_norm": 0.38303308206260517, |
|
"learning_rate": 1.462506333797501e-06, |
|
"loss": 1.2985, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.8445113605255954, |
|
"grad_norm": 0.37178133903556354, |
|
"learning_rate": 1.4377213827083602e-06, |
|
"loss": 1.278, |
|
"step": 3085 |
|
}, |
|
{ |
|
"epoch": 0.8458800985491377, |
|
"grad_norm": 0.383873936282195, |
|
"learning_rate": 1.413131968586491e-06, |
|
"loss": 1.2989, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.84724883657268, |
|
"grad_norm": 0.3877611566994199, |
|
"learning_rate": 1.3887386529809454e-06, |
|
"loss": 1.3543, |
|
"step": 3095 |
|
}, |
|
{ |
|
"epoch": 0.8486175745962223, |
|
"grad_norm": 0.3844140889489012, |
|
"learning_rate": 1.364541992962476e-06, |
|
"loss": 1.345, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.8499863126197645, |
|
"grad_norm": 0.3791779816963577, |
|
"learning_rate": 1.3405425411108008e-06, |
|
"loss": 1.3202, |
|
"step": 3105 |
|
}, |
|
{ |
|
"epoch": 0.8513550506433069, |
|
"grad_norm": 0.37802806800903216, |
|
"learning_rate": 1.3167408455019903e-06, |
|
"loss": 1.3, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 0.8527237886668492, |
|
"grad_norm": 0.38421255363489804, |
|
"learning_rate": 1.2931374496959548e-06, |
|
"loss": 1.3033, |
|
"step": 3115 |
|
}, |
|
{ |
|
"epoch": 0.8540925266903915, |
|
"grad_norm": 0.3890245703911996, |
|
"learning_rate": 1.2697328927240238e-06, |
|
"loss": 1.3155, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.8554612647139338, |
|
"grad_norm": 0.38965532355017113, |
|
"learning_rate": 1.2465277090766381e-06, |
|
"loss": 1.3408, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 0.856830002737476, |
|
"grad_norm": 0.38847581936098935, |
|
"learning_rate": 1.2235224286911495e-06, |
|
"loss": 1.3619, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 0.8581987407610183, |
|
"grad_norm": 0.3901578768967818, |
|
"learning_rate": 1.2007175769397117e-06, |
|
"loss": 1.3714, |
|
"step": 3135 |
|
}, |
|
{ |
|
"epoch": 0.8595674787845606, |
|
"grad_norm": 0.38170228292380465, |
|
"learning_rate": 1.178113674617285e-06, |
|
"loss": 1.3144, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.860936216808103, |
|
"grad_norm": 0.3852269233134472, |
|
"learning_rate": 1.1557112379297385e-06, |
|
"loss": 1.3542, |
|
"step": 3145 |
|
}, |
|
{ |
|
"epoch": 0.8623049548316453, |
|
"grad_norm": 0.385512160462019, |
|
"learning_rate": 1.1335107784820741e-06, |
|
"loss": 1.3556, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.8636736928551875, |
|
"grad_norm": 0.39758307605071247, |
|
"learning_rate": 1.1115128032667288e-06, |
|
"loss": 1.2992, |
|
"step": 3155 |
|
}, |
|
{ |
|
"epoch": 0.8650424308787298, |
|
"grad_norm": 0.3702170793773961, |
|
"learning_rate": 1.0897178146520014e-06, |
|
"loss": 1.3861, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.8664111689022721, |
|
"grad_norm": 0.3862887179246134, |
|
"learning_rate": 1.0681263103705853e-06, |
|
"loss": 1.3317, |
|
"step": 3165 |
|
}, |
|
{ |
|
"epoch": 0.8677799069258144, |
|
"grad_norm": 0.3979382501364706, |
|
"learning_rate": 1.0467387835081944e-06, |
|
"loss": 1.351, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 0.8691486449493567, |
|
"grad_norm": 0.3853783562962658, |
|
"learning_rate": 1.0255557224923018e-06, |
|
"loss": 1.3474, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 0.8705173829728989, |
|
"grad_norm": 0.3921947417156507, |
|
"learning_rate": 1.004577611080998e-06, |
|
"loss": 1.3162, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.8718861209964412, |
|
"grad_norm": 0.38229199053031476, |
|
"learning_rate": 9.838049283519258e-07, |
|
"loss": 1.3265, |
|
"step": 3185 |
|
}, |
|
{ |
|
"epoch": 0.8732548590199836, |
|
"grad_norm": 0.38057187484097027, |
|
"learning_rate": 9.63238148691351e-07, |
|
"loss": 1.3087, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 0.8746235970435259, |
|
"grad_norm": 0.3942222628186198, |
|
"learning_rate": 9.42877741783328e-07, |
|
"loss": 1.366, |
|
"step": 3195 |
|
}, |
|
{ |
|
"epoch": 0.8759923350670682, |
|
"grad_norm": 0.38514993886291565, |
|
"learning_rate": 9.227241725989699e-07, |
|
"loss": 1.3212, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.8773610730906105, |
|
"grad_norm": 0.3836381492014198, |
|
"learning_rate": 9.027779013858284e-07, |
|
"loss": 1.2787, |
|
"step": 3205 |
|
}, |
|
{ |
|
"epoch": 0.8787298111141527, |
|
"grad_norm": 0.38734970197899504, |
|
"learning_rate": 8.830393836573947e-07, |
|
"loss": 1.3387, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 0.880098549137695, |
|
"grad_norm": 0.3817865137800505, |
|
"learning_rate": 8.635090701826799e-07, |
|
"loss": 1.3753, |
|
"step": 3215 |
|
}, |
|
{ |
|
"epoch": 0.8814672871612373, |
|
"grad_norm": 0.38200788300985267, |
|
"learning_rate": 8.441874069759337e-07, |
|
"loss": 1.2776, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.8828360251847797, |
|
"grad_norm": 0.37300363369723033, |
|
"learning_rate": 8.250748352864546e-07, |
|
"loss": 1.317, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 0.884204763208322, |
|
"grad_norm": 0.39737199946658025, |
|
"learning_rate": 8.061717915885103e-07, |
|
"loss": 1.3048, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 0.8855735012318642, |
|
"grad_norm": 0.3771135467998678, |
|
"learning_rate": 7.874787075713742e-07, |
|
"loss": 1.2507, |
|
"step": 3235 |
|
}, |
|
{ |
|
"epoch": 0.8869422392554065, |
|
"grad_norm": 0.37723109683472905, |
|
"learning_rate": 7.689960101294691e-07, |
|
"loss": 1.3081, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.8883109772789488, |
|
"grad_norm": 0.3871833961992763, |
|
"learning_rate": 7.507241213526073e-07, |
|
"loss": 1.3122, |
|
"step": 3245 |
|
}, |
|
{ |
|
"epoch": 0.8896797153024911, |
|
"grad_norm": 0.3774090621861952, |
|
"learning_rate": 7.326634585163617e-07, |
|
"loss": 1.3243, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.8910484533260334, |
|
"grad_norm": 0.39061075194420053, |
|
"learning_rate": 7.148144340725371e-07, |
|
"loss": 1.3123, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 0.8924171913495756, |
|
"grad_norm": 0.3844913895203542, |
|
"learning_rate": 6.971774556397415e-07, |
|
"loss": 1.3238, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.893785929373118, |
|
"grad_norm": 0.3954658749141559, |
|
"learning_rate": 6.797529259940827e-07, |
|
"loss": 1.3421, |
|
"step": 3265 |
|
}, |
|
{ |
|
"epoch": 0.8951546673966603, |
|
"grad_norm": 0.38848733821104914, |
|
"learning_rate": 6.625412430599765e-07, |
|
"loss": 1.3485, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 0.8965234054202026, |
|
"grad_norm": 0.38275187348582623, |
|
"learning_rate": 6.455427999010466e-07, |
|
"loss": 1.3343, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 0.8978921434437449, |
|
"grad_norm": 0.3946581641011151, |
|
"learning_rate": 6.287579847111569e-07, |
|
"loss": 1.3362, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.8992608814672871, |
|
"grad_norm": 0.3931302904815353, |
|
"learning_rate": 6.121871808055479e-07, |
|
"loss": 1.3095, |
|
"step": 3285 |
|
}, |
|
{ |
|
"epoch": 0.9006296194908294, |
|
"grad_norm": 0.39874206377284055, |
|
"learning_rate": 5.958307666120733e-07, |
|
"loss": 1.2925, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 0.9019983575143717, |
|
"grad_norm": 0.38856468686205425, |
|
"learning_rate": 5.796891156625639e-07, |
|
"loss": 1.2878, |
|
"step": 3295 |
|
}, |
|
{ |
|
"epoch": 0.903367095537914, |
|
"grad_norm": 0.3830311422743617, |
|
"learning_rate": 5.637625965843041e-07, |
|
"loss": 1.3247, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.9047358335614564, |
|
"grad_norm": 0.39286283244773124, |
|
"learning_rate": 5.480515730915992e-07, |
|
"loss": 1.2902, |
|
"step": 3305 |
|
}, |
|
{ |
|
"epoch": 0.9061045715849987, |
|
"grad_norm": 0.37864973007608094, |
|
"learning_rate": 5.325564039774777e-07, |
|
"loss": 1.308, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 0.9074733096085409, |
|
"grad_norm": 0.3932055048918823, |
|
"learning_rate": 5.172774431054995e-07, |
|
"loss": 1.3245, |
|
"step": 3315 |
|
}, |
|
{ |
|
"epoch": 0.9088420476320832, |
|
"grad_norm": 0.3861464572512248, |
|
"learning_rate": 5.022150394016701e-07, |
|
"loss": 1.3345, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.9102107856556255, |
|
"grad_norm": 0.3849211658260205, |
|
"learning_rate": 4.873695368464693e-07, |
|
"loss": 1.3522, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 0.9115795236791678, |
|
"grad_norm": 0.3850036124623499, |
|
"learning_rate": 4.72741274467009e-07, |
|
"loss": 1.3205, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 0.9129482617027102, |
|
"grad_norm": 0.3858389082786252, |
|
"learning_rate": 4.5833058632927417e-07, |
|
"loss": 1.2984, |
|
"step": 3335 |
|
}, |
|
{ |
|
"epoch": 0.9143169997262524, |
|
"grad_norm": 0.3974200110164499, |
|
"learning_rate": 4.441378015305031e-07, |
|
"loss": 1.2907, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.9156857377497947, |
|
"grad_norm": 0.3885026941845055, |
|
"learning_rate": 4.3016324419167365e-07, |
|
"loss": 1.3571, |
|
"step": 3345 |
|
}, |
|
{ |
|
"epoch": 0.917054475773337, |
|
"grad_norm": 0.391701334822345, |
|
"learning_rate": 4.164072334500935e-07, |
|
"loss": 1.2946, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.9184232137968793, |
|
"grad_norm": 0.371298260058105, |
|
"learning_rate": 4.028700834521193e-07, |
|
"loss": 1.2734, |
|
"step": 3355 |
|
}, |
|
{ |
|
"epoch": 0.9197919518204216, |
|
"grad_norm": 0.39257123679148415, |
|
"learning_rate": 3.8955210334597595e-07, |
|
"loss": 1.3792, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.9211606898439638, |
|
"grad_norm": 0.37739382436504915, |
|
"learning_rate": 3.764535972747052e-07, |
|
"loss": 1.3182, |
|
"step": 3365 |
|
}, |
|
{ |
|
"epoch": 0.9225294278675061, |
|
"grad_norm": 0.4040354445639437, |
|
"learning_rate": 3.6357486436921164e-07, |
|
"loss": 1.3149, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 0.9238981658910485, |
|
"grad_norm": 0.3885892463097669, |
|
"learning_rate": 3.5091619874143446e-07, |
|
"loss": 1.3612, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 0.9252669039145908, |
|
"grad_norm": 0.39478973344601664, |
|
"learning_rate": 3.3847788947763194e-07, |
|
"loss": 1.3338, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.9266356419381331, |
|
"grad_norm": 0.4002415191432085, |
|
"learning_rate": 3.2626022063177997e-07, |
|
"loss": 1.3854, |
|
"step": 3385 |
|
}, |
|
{ |
|
"epoch": 0.9280043799616753, |
|
"grad_norm": 0.3871640152525417, |
|
"learning_rate": 3.142634712190795e-07, |
|
"loss": 1.2663, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 0.9293731179852176, |
|
"grad_norm": 0.3844217189319342, |
|
"learning_rate": 3.0248791520959387e-07, |
|
"loss": 1.3304, |
|
"step": 3395 |
|
}, |
|
{ |
|
"epoch": 0.9307418560087599, |
|
"grad_norm": 0.40010784379733844, |
|
"learning_rate": 2.909338215219859e-07, |
|
"loss": 1.3458, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.9321105940323022, |
|
"grad_norm": 0.38274676057072776, |
|
"learning_rate": 2.7960145401737415e-07, |
|
"loss": 1.2606, |
|
"step": 3405 |
|
}, |
|
{ |
|
"epoch": 0.9334793320558445, |
|
"grad_norm": 0.39275920250834023, |
|
"learning_rate": 2.6849107149331756e-07, |
|
"loss": 1.2825, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 0.9348480700793868, |
|
"grad_norm": 0.3947633255379109, |
|
"learning_rate": 2.576029276778924e-07, |
|
"loss": 1.3441, |
|
"step": 3415 |
|
}, |
|
{ |
|
"epoch": 0.9362168081029291, |
|
"grad_norm": 0.41407416506322803, |
|
"learning_rate": 2.4693727122390597e-07, |
|
"loss": 1.371, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.9375855461264714, |
|
"grad_norm": 0.37625661449174036, |
|
"learning_rate": 2.3649434570321984e-07, |
|
"loss": 1.2862, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 0.9389542841500137, |
|
"grad_norm": 0.37104631834710733, |
|
"learning_rate": 2.2627438960117876e-07, |
|
"loss": 1.2833, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 0.940323022173556, |
|
"grad_norm": 0.3868409112001441, |
|
"learning_rate": 2.1627763631117182e-07, |
|
"loss": 1.3551, |
|
"step": 3435 |
|
}, |
|
{ |
|
"epoch": 0.9416917601970983, |
|
"grad_norm": 0.4061569745095073, |
|
"learning_rate": 2.0650431412930104e-07, |
|
"loss": 1.3273, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.9430604982206405, |
|
"grad_norm": 0.3923199404039606, |
|
"learning_rate": 1.969546462491634e-07, |
|
"loss": 1.3093, |
|
"step": 3445 |
|
}, |
|
{ |
|
"epoch": 0.9444292362441828, |
|
"grad_norm": 0.40035298460473323, |
|
"learning_rate": 1.876288507567592e-07, |
|
"loss": 1.2859, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.9457979742677252, |
|
"grad_norm": 0.39270377743419116, |
|
"learning_rate": 1.785271406255107e-07, |
|
"loss": 1.3086, |
|
"step": 3455 |
|
}, |
|
{ |
|
"epoch": 0.9471667122912675, |
|
"grad_norm": 0.391272234134139, |
|
"learning_rate": 1.6964972371139588e-07, |
|
"loss": 1.3324, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.9485354503148098, |
|
"grad_norm": 0.3971408419550116, |
|
"learning_rate": 1.609968027482012e-07, |
|
"loss": 1.3241, |
|
"step": 3465 |
|
}, |
|
{ |
|
"epoch": 0.949904188338352, |
|
"grad_norm": 0.397310674294107, |
|
"learning_rate": 1.5256857534289626e-07, |
|
"loss": 1.344, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 0.9512729263618943, |
|
"grad_norm": 0.37737586107823284, |
|
"learning_rate": 1.443652339711199e-07, |
|
"loss": 1.3227, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 0.9526416643854366, |
|
"grad_norm": 0.3953334287452581, |
|
"learning_rate": 1.3638696597277678e-07, |
|
"loss": 1.3323, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.9540104024089789, |
|
"grad_norm": 0.39727863290285664, |
|
"learning_rate": 1.2863395354777097e-07, |
|
"loss": 1.2965, |
|
"step": 3485 |
|
}, |
|
{ |
|
"epoch": 0.9553791404325213, |
|
"grad_norm": 0.3915026545665889, |
|
"learning_rate": 1.211063737518392e-07, |
|
"loss": 1.2945, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 0.9567478784560635, |
|
"grad_norm": 0.3963237143375255, |
|
"learning_rate": 1.1380439849250414e-07, |
|
"loss": 1.3079, |
|
"step": 3495 |
|
}, |
|
{ |
|
"epoch": 0.9581166164796058, |
|
"grad_norm": 0.38628531323230814, |
|
"learning_rate": 1.0672819452515526e-07, |
|
"loss": 1.347, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.9594853545031481, |
|
"grad_norm": 0.39250735459020125, |
|
"learning_rate": 9.987792344923753e-08, |
|
"loss": 1.3292, |
|
"step": 3505 |
|
}, |
|
{ |
|
"epoch": 0.9608540925266904, |
|
"grad_norm": 0.39214485753203543, |
|
"learning_rate": 9.32537417045576e-08, |
|
"loss": 1.2703, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 0.9622228305502327, |
|
"grad_norm": 0.38156832143297204, |
|
"learning_rate": 8.685580056771781e-08, |
|
"loss": 1.3404, |
|
"step": 3515 |
|
}, |
|
{ |
|
"epoch": 0.9635915685737749, |
|
"grad_norm": 0.3907257773875925, |
|
"learning_rate": 8.0684246148659e-08, |
|
"loss": 1.2681, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.9649603065973172, |
|
"grad_norm": 0.4194193748219591, |
|
"learning_rate": 7.473921938731865e-08, |
|
"loss": 1.382, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 0.9663290446208596, |
|
"grad_norm": 0.3858101635490962, |
|
"learning_rate": 6.902085605042019e-08, |
|
"loss": 1.3671, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 0.9676977826444019, |
|
"grad_norm": 0.3813102160301827, |
|
"learning_rate": 6.352928672836767e-08, |
|
"loss": 1.3013, |
|
"step": 3535 |
|
}, |
|
{ |
|
"epoch": 0.9690665206679442, |
|
"grad_norm": 0.38117598807083153, |
|
"learning_rate": 5.82646368322648e-08, |
|
"loss": 1.3406, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.9704352586914865, |
|
"grad_norm": 0.3869856138402919, |
|
"learning_rate": 5.3227026591049505e-08, |
|
"loss": 1.3311, |
|
"step": 3545 |
|
}, |
|
{ |
|
"epoch": 0.9718039967150287, |
|
"grad_norm": 0.3866858707608126, |
|
"learning_rate": 4.841657104875275e-08, |
|
"loss": 1.3593, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.973172734738571, |
|
"grad_norm": 0.3866568363141188, |
|
"learning_rate": 4.3833380061865104e-08, |
|
"loss": 1.3318, |
|
"step": 3555 |
|
}, |
|
{ |
|
"epoch": 0.9745414727621133, |
|
"grad_norm": 0.39553603732869785, |
|
"learning_rate": 3.947755829683097e-08, |
|
"loss": 1.3403, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.9759102107856557, |
|
"grad_norm": 0.3990993940103786, |
|
"learning_rate": 3.5349205227660496e-08, |
|
"loss": 1.3812, |
|
"step": 3565 |
|
}, |
|
{ |
|
"epoch": 0.977278948809198, |
|
"grad_norm": 0.3925614431869845, |
|
"learning_rate": 3.144841513365249e-08, |
|
"loss": 1.3025, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 0.9786476868327402, |
|
"grad_norm": 0.3883373657835178, |
|
"learning_rate": 2.7775277097247255e-08, |
|
"loss": 1.313, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 0.9800164248562825, |
|
"grad_norm": 0.38572501548529475, |
|
"learning_rate": 2.4329875001989356e-08, |
|
"loss": 1.3058, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.9813851628798248, |
|
"grad_norm": 0.3838975342711645, |
|
"learning_rate": 2.1112287530609122e-08, |
|
"loss": 1.3165, |
|
"step": 3585 |
|
}, |
|
{ |
|
"epoch": 0.9827539009033671, |
|
"grad_norm": 0.3918735272830008, |
|
"learning_rate": 1.812258816323187e-08, |
|
"loss": 1.3388, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 0.9841226389269094, |
|
"grad_norm": 0.38644635827545804, |
|
"learning_rate": 1.5360845175695916e-08, |
|
"loss": 1.3378, |
|
"step": 3595 |
|
}, |
|
{ |
|
"epoch": 0.9854913769504516, |
|
"grad_norm": 0.4046028492421605, |
|
"learning_rate": 1.2827121637992712e-08, |
|
"loss": 1.3104, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.986860114973994, |
|
"grad_norm": 0.37588594630122674, |
|
"learning_rate": 1.0521475412830218e-08, |
|
"loss": 1.3345, |
|
"step": 3605 |
|
}, |
|
{ |
|
"epoch": 0.9882288529975363, |
|
"grad_norm": 0.39556552792879707, |
|
"learning_rate": 8.44395915430729e-09, |
|
"loss": 1.3184, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 0.9895975910210786, |
|
"grad_norm": 0.3745912876489376, |
|
"learning_rate": 6.5946203067135395e-09, |
|
"loss": 1.3036, |
|
"step": 3615 |
|
}, |
|
{ |
|
"epoch": 0.9909663290446209, |
|
"grad_norm": 0.39112472329549625, |
|
"learning_rate": 4.9735011034457434e-09, |
|
"loss": 1.2682, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.9923350670681631, |
|
"grad_norm": 0.380307545731869, |
|
"learning_rate": 3.580638566043071e-09, |
|
"loss": 1.2837, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 0.9937038050917054, |
|
"grad_norm": 0.3955743424810463, |
|
"learning_rate": 2.416064503342197e-09, |
|
"loss": 1.3092, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 0.9950725431152477, |
|
"grad_norm": 0.3853245568354318, |
|
"learning_rate": 1.4798055107489996e-09, |
|
"loss": 1.3025, |
|
"step": 3635 |
|
}, |
|
{ |
|
"epoch": 0.99644128113879, |
|
"grad_norm": 0.39071516823719027, |
|
"learning_rate": 7.718829696334862e-10, |
|
"loss": 1.3332, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.9978100191623324, |
|
"grad_norm": 0.38346889130554324, |
|
"learning_rate": 2.9231304683907667e-10, |
|
"loss": 1.3369, |
|
"step": 3645 |
|
}, |
|
{ |
|
"epoch": 0.9991787571858747, |
|
"grad_norm": 0.4139215786504167, |
|
"learning_rate": 4.1106694317338826e-11, |
|
"loss": 1.3518, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 1.320330262184143, |
|
"eval_runtime": 951.5767, |
|
"eval_samples_per_second": 92.065, |
|
"eval_steps_per_second": 5.755, |
|
"step": 3653 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 3653, |
|
"total_flos": 66190143651840.0, |
|
"train_loss": 1.3573657579367173, |
|
"train_runtime": 8062.8425, |
|
"train_samples_per_second": 7.249, |
|
"train_steps_per_second": 0.453 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 3653, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 66190143651840.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|