|
{ |
|
"best_metric": 0.9291033148765564, |
|
"best_model_checkpoint": "segformer-finetuned-segments-cmp-facade-outputs-transforms/checkpoint-360", |
|
"epoch": 11.25, |
|
"global_step": 360, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000599625, |
|
"loss": 1.4974, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00059925, |
|
"loss": 1.5572, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.000598875, |
|
"loss": 1.5276, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0005985, |
|
"loss": 1.5646, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0005981249999999999, |
|
"loss": 1.3389, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00059775, |
|
"loss": 1.5391, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0005973749999999999, |
|
"loss": 1.3854, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.000597, |
|
"loss": 1.4288, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0005966249999999999, |
|
"loss": 1.3674, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00059625, |
|
"loss": 1.2815, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0005958749999999999, |
|
"loss": 1.3141, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0005955, |
|
"loss": 1.2826, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.000595125, |
|
"loss": 1.2023, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0005947499999999999, |
|
"loss": 1.2487, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.000594375, |
|
"loss": 1.2041, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0005939999999999999, |
|
"loss": 1.3763, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.000593625, |
|
"loss": 1.2913, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.0005932499999999999, |
|
"loss": 1.2048, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 0.000592875, |
|
"loss": 1.2608, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.0005924999999999999, |
|
"loss": 1.2328, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"eval_loss": 1.2321703433990479, |
|
"eval_mean_accuracy": 0.27640981040002394, |
|
"eval_mean_iou": 0.19381676955791272, |
|
"eval_overall_accuracy": 0.5843182279352557, |
|
"eval_per_category_accuracy": [ |
|
0.6484014402727478, |
|
0.7268698138428387, |
|
0.7396621827345261, |
|
0.0, |
|
0.314071721025026, |
|
0.16580774409430243, |
|
0.18421110637031626, |
|
0.0, |
|
0.0, |
|
0.48692030347962645, |
|
0.0, |
|
0.05097341298090315 |
|
], |
|
"eval_per_category_iou": [ |
|
0.46588897486611575, |
|
0.4794379279754499, |
|
0.47516306513696516, |
|
0.0, |
|
0.21479697731322273, |
|
0.1468532929608522, |
|
0.15399766900533918, |
|
0.0, |
|
0.0, |
|
0.34140124873016997, |
|
0.0, |
|
0.04826207870683769 |
|
], |
|
"eval_runtime": 121.98, |
|
"eval_samples_per_second": 0.935, |
|
"eval_steps_per_second": 0.082, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.0005921249999999999, |
|
"loss": 1.2252, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.0005917499999999999, |
|
"loss": 1.1954, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.0005913749999999999, |
|
"loss": 1.2127, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.0005909999999999999, |
|
"loss": 1.1908, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.0005906249999999999, |
|
"loss": 1.1458, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 0.00059025, |
|
"loss": 1.0854, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.000589875, |
|
"loss": 1.1372, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 0.0005895, |
|
"loss": 1.0638, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 0.000589125, |
|
"loss": 1.1055, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 0.00058875, |
|
"loss": 1.1073, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.0005883749999999999, |
|
"loss": 1.0367, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.000588, |
|
"loss": 1.2498, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.0005876249999999999, |
|
"loss": 1.2009, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.00058725, |
|
"loss": 1.2267, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.0005868749999999999, |
|
"loss": 0.9791, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 0.0005865, |
|
"loss": 1.0073, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.000586125, |
|
"loss": 1.0052, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.0005857499999999999, |
|
"loss": 1.0387, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 0.000585375, |
|
"loss": 1.076, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 0.0005849999999999999, |
|
"loss": 1.1158, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"eval_loss": 1.1340055465698242, |
|
"eval_mean_accuracy": 0.3584914050139854, |
|
"eval_mean_iou": 0.25033676612937344, |
|
"eval_overall_accuracy": 0.6050458539996231, |
|
"eval_per_category_accuracy": [ |
|
0.8000948297521079, |
|
0.6337371473364622, |
|
0.6603131057495936, |
|
0.003965968313890952, |
|
0.34629573729708046, |
|
0.3680114248799625, |
|
0.4865078902745655, |
|
0.04561502007369462, |
|
0.0118031463226356, |
|
0.5511836320859765, |
|
0.0, |
|
0.3943689580818554 |
|
], |
|
"eval_per_category_iou": [ |
|
0.5176739126303989, |
|
0.47108775141725884, |
|
0.49329266238923963, |
|
0.003963836494345551, |
|
0.26672845921223864, |
|
0.2701529378697397, |
|
0.31332750965950773, |
|
0.04532147556410988, |
|
0.011623004855842847, |
|
0.3621998398780441, |
|
0.0, |
|
0.24866980358175594 |
|
], |
|
"eval_runtime": 124.8448, |
|
"eval_samples_per_second": 0.913, |
|
"eval_steps_per_second": 0.08, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 0.000584625, |
|
"loss": 1.0057, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.0005842499999999999, |
|
"loss": 0.9904, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 0.000583875, |
|
"loss": 1.1023, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 0.0005834999999999999, |
|
"loss": 1.1263, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 0.000583125, |
|
"loss": 0.9288, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 0.0005827499999999999, |
|
"loss": 0.9942, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 0.0005823749999999999, |
|
"loss": 0.9388, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 0.0005819999999999999, |
|
"loss": 1.0192, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 0.0005816249999999999, |
|
"loss": 1.0228, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 0.00058125, |
|
"loss": 0.9034, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 0.000580875, |
|
"loss": 1.0346, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 0.0005805, |
|
"loss": 0.8394, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 0.000580125, |
|
"loss": 0.9892, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 0.00057975, |
|
"loss": 1.0063, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 0.0005793749999999999, |
|
"loss": 1.0106, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 0.000579, |
|
"loss": 0.9481, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 0.0005786249999999999, |
|
"loss": 1.0124, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 0.00057825, |
|
"loss": 0.9676, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 0.0005778749999999999, |
|
"loss": 0.7901, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 0.0005775, |
|
"loss": 0.9771, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"eval_loss": 1.0994304418563843, |
|
"eval_mean_accuracy": 0.3909737030282887, |
|
"eval_mean_iou": 0.2835007332832486, |
|
"eval_overall_accuracy": 0.6179555591784025, |
|
"eval_per_category_accuracy": [ |
|
0.7168020611741474, |
|
0.7219130768570321, |
|
0.5340207091634352, |
|
0.0431815185182615, |
|
0.45474233550172505, |
|
0.5430138746239189, |
|
0.469268628306114, |
|
0.217241626853214, |
|
0.09605778018331367, |
|
0.5683827928074056, |
|
0.0005080911993094857, |
|
0.3265519411515883 |
|
], |
|
"eval_per_category_iou": [ |
|
0.5113476068813879, |
|
0.49334140108090285, |
|
0.4746839761440892, |
|
0.040214174001835405, |
|
0.31186037529280675, |
|
0.31181019822943934, |
|
0.35578648191496853, |
|
0.20097078621392386, |
|
0.08022661461990363, |
|
0.38374995642386217, |
|
0.0005073985432159898, |
|
0.23750983005264706 |
|
], |
|
"eval_runtime": 125.5081, |
|
"eval_samples_per_second": 0.908, |
|
"eval_steps_per_second": 0.08, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 0.000577125, |
|
"loss": 0.9391, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 0.00057675, |
|
"loss": 1.0473, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 0.000576375, |
|
"loss": 0.918, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 0.0005759999999999999, |
|
"loss": 1.0355, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 0.000575625, |
|
"loss": 1.015, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 0.0005752499999999999, |
|
"loss": 0.9346, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 0.000574875, |
|
"loss": 0.8417, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 0.0005744999999999999, |
|
"loss": 0.8185, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 0.000574125, |
|
"loss": 0.9263, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 0.0005737499999999999, |
|
"loss": 0.8448, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 0.0005733749999999999, |
|
"loss": 1.1335, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 0.0005729999999999999, |
|
"loss": 0.9229, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 0.0005726249999999999, |
|
"loss": 0.8549, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 0.00057225, |
|
"loss": 0.8818, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 0.0005718749999999999, |
|
"loss": 0.831, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 0.0005715, |
|
"loss": 0.8845, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 0.000571125, |
|
"loss": 0.9197, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 0.00057075, |
|
"loss": 0.8579, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 0.000570375, |
|
"loss": 0.8167, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 0.00057, |
|
"loss": 0.8566, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"eval_loss": 1.0861966609954834, |
|
"eval_mean_accuracy": 0.3778088961525232, |
|
"eval_mean_iou": 0.26404526643863496, |
|
"eval_overall_accuracy": 0.6308857432582922, |
|
"eval_per_category_accuracy": [ |
|
0.6129614268456303, |
|
0.7713753471720866, |
|
0.7994049832227069, |
|
0.019437755463182074, |
|
0.5688067632752372, |
|
0.4619781591716599, |
|
0.3827566445038881, |
|
0.08782710157707345, |
|
0.15580996069314926, |
|
0.5327405741537456, |
|
0.0009059939457566733, |
|
0.1397020438061634 |
|
], |
|
"eval_per_category_iou": [ |
|
0.5101947204251175, |
|
0.5344538212406682, |
|
0.49436541874132667, |
|
0.018884843581445522, |
|
0.270377214768106, |
|
0.28449876963719384, |
|
0.32160160861987425, |
|
0.086283029811171, |
|
0.11932598089180631, |
|
0.39498832385101396, |
|
0.0009054175167701066, |
|
0.13266404817912572 |
|
], |
|
"eval_runtime": 123.9, |
|
"eval_samples_per_second": 0.92, |
|
"eval_steps_per_second": 0.081, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 0.0005696249999999999, |
|
"loss": 0.827, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 0.00056925, |
|
"loss": 0.9862, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 0.0005688749999999999, |
|
"loss": 0.8236, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 0.0005685, |
|
"loss": 0.8522, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 0.0005681249999999999, |
|
"loss": 0.9125, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 0.00056775, |
|
"loss": 0.8568, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 0.000567375, |
|
"loss": 0.8791, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 0.0005669999999999999, |
|
"loss": 0.8188, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 0.000566625, |
|
"loss": 0.992, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 0.0005662499999999999, |
|
"loss": 0.8239, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 0.000565875, |
|
"loss": 0.8892, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 0.0005654999999999999, |
|
"loss": 0.8093, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 0.000565125, |
|
"loss": 1.0857, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 0.0005647499999999999, |
|
"loss": 0.8162, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 0.000564375, |
|
"loss": 0.9017, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 0.0005639999999999999, |
|
"loss": 0.8848, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 0.0005636249999999999, |
|
"loss": 0.9075, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 0.00056325, |
|
"loss": 0.8108, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 0.0005628749999999999, |
|
"loss": 0.8238, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 0.0005625, |
|
"loss": 0.7854, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"eval_loss": 1.0004723072052002, |
|
"eval_mean_accuracy": 0.4409761597550485, |
|
"eval_mean_iou": 0.3278147775750712, |
|
"eval_overall_accuracy": 0.6476280145477831, |
|
"eval_per_category_accuracy": [ |
|
0.7394164167070112, |
|
0.7120692919670155, |
|
0.65725755249271, |
|
0.17395090943157923, |
|
0.46355298851110294, |
|
0.31981316619425926, |
|
0.6107480292025965, |
|
0.35138126519456003, |
|
0.1554151176100015, |
|
0.5705962518273815, |
|
0.05702864593695384, |
|
0.4804842819854117 |
|
], |
|
"eval_per_category_iou": [ |
|
0.5451244129938286, |
|
0.5178187584431643, |
|
0.5318888189962618, |
|
0.14778251064896017, |
|
0.3169269253463421, |
|
0.2626134690690619, |
|
0.39125714035001896, |
|
0.3081966042250305, |
|
0.12383697911511432, |
|
0.3977603328560864, |
|
0.04942253722870921, |
|
0.34114884162827647 |
|
], |
|
"eval_runtime": 124.3127, |
|
"eval_samples_per_second": 0.917, |
|
"eval_steps_per_second": 0.08, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 0.0005621249999999999, |
|
"loss": 0.8351, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 0.00056175, |
|
"loss": 0.8611, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 0.000561375, |
|
"loss": 0.9022, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 0.000561, |
|
"loss": 0.8189, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 0.0005606249999999999, |
|
"loss": 0.8431, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 0.00056025, |
|
"loss": 0.9712, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 0.0005598749999999999, |
|
"loss": 0.7789, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 0.0005595, |
|
"loss": 0.7784, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 0.0005591249999999999, |
|
"loss": 0.7459, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 0.00055875, |
|
"loss": 0.7602, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 0.000558375, |
|
"loss": 0.7709, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 0.000558, |
|
"loss": 0.792, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 0.000557625, |
|
"loss": 0.7701, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 0.0005572499999999999, |
|
"loss": 0.8632, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 0.000556875, |
|
"loss": 0.776, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 0.0005564999999999999, |
|
"loss": 0.8366, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 0.000556125, |
|
"loss": 0.731, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 0.0005557499999999999, |
|
"loss": 0.8111, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"learning_rate": 0.000555375, |
|
"loss": 0.7547, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 0.0005549999999999999, |
|
"loss": 0.7275, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"eval_loss": 1.0055418014526367, |
|
"eval_mean_accuracy": 0.46582571564323233, |
|
"eval_mean_iou": 0.34071242663610873, |
|
"eval_overall_accuracy": 0.6500220717045299, |
|
"eval_per_category_accuracy": [ |
|
0.6790541517825218, |
|
0.7248947508922547, |
|
0.702760127940592, |
|
0.22049812284439385, |
|
0.49308863123902213, |
|
0.336202860575091, |
|
0.6458925386011786, |
|
0.40734546148068546, |
|
0.2508473598750699, |
|
0.5798344703996796, |
|
0.12197249573784942, |
|
0.4275176163504493 |
|
], |
|
"eval_per_category_iou": [ |
|
0.5331521735997542, |
|
0.5305308321791462, |
|
0.5588320252288795, |
|
0.16697584226263984, |
|
0.3296551780182928, |
|
0.28048869107094193, |
|
0.35449032046775947, |
|
0.34190939733724723, |
|
0.16527723782325413, |
|
0.39484470653333187, |
|
0.0880505368099854, |
|
0.3443421783020718 |
|
], |
|
"eval_runtime": 124.4108, |
|
"eval_samples_per_second": 0.916, |
|
"eval_steps_per_second": 0.08, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 0.0005546249999999999, |
|
"loss": 0.7266, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 0.00055425, |
|
"loss": 0.8377, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 0.0005538749999999999, |
|
"loss": 0.7775, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 0.0005535, |
|
"loss": 0.8343, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 0.0005531249999999999, |
|
"loss": 0.8143, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 0.00055275, |
|
"loss": 0.8377, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 0.0005523749999999999, |
|
"loss": 0.8807, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 0.000552, |
|
"loss": 0.8941, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 0.000551625, |
|
"loss": 0.9077, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"learning_rate": 0.0005512499999999999, |
|
"loss": 0.8361, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 0.0005508749999999999, |
|
"loss": 0.6058, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"learning_rate": 0.0005505, |
|
"loss": 0.8466, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"learning_rate": 0.0005501249999999999, |
|
"loss": 0.6832, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 0.00054975, |
|
"loss": 0.9442, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 0.000549375, |
|
"loss": 0.7119, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 0.000549, |
|
"loss": 0.8412, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 0.000548625, |
|
"loss": 0.8589, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 0.0005482499999999999, |
|
"loss": 0.7529, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 0.000547875, |
|
"loss": 0.7525, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 4.38, |
|
"learning_rate": 0.0005474999999999999, |
|
"loss": 0.762, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 4.38, |
|
"eval_loss": 1.009921669960022, |
|
"eval_mean_accuracy": 0.4630782466624949, |
|
"eval_mean_iou": 0.3357803602667826, |
|
"eval_overall_accuracy": 0.6476533454761171, |
|
"eval_per_category_accuracy": [ |
|
0.6388404926271997, |
|
0.7725403351834625, |
|
0.6020424381762673, |
|
0.21688954275126474, |
|
0.5520221552773765, |
|
0.4518525508185774, |
|
0.6697900807527053, |
|
0.46810120099302377, |
|
0.1897420654286044, |
|
0.45603045018648025, |
|
0.024152696709344286, |
|
0.5149349510456322 |
|
], |
|
"eval_per_category_iou": [ |
|
0.5210483275809908, |
|
0.5385459417489167, |
|
0.519285461653176, |
|
0.19235183880209375, |
|
0.3408901260438495, |
|
0.3345069902216156, |
|
0.36090171027181106, |
|
0.33539228285452394, |
|
0.1422865859237117, |
|
0.3681340241516971, |
|
0.023384620855608634, |
|
0.35263641309339605 |
|
], |
|
"eval_runtime": 123.6747, |
|
"eval_samples_per_second": 0.922, |
|
"eval_steps_per_second": 0.081, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 0.000547125, |
|
"loss": 0.7219, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 0.0005467499999999999, |
|
"loss": 0.7571, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 0.000546375, |
|
"loss": 0.7736, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 0.0005459999999999999, |
|
"loss": 0.7782, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 0.000545625, |
|
"loss": 0.7814, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 4.56, |
|
"learning_rate": 0.0005452499999999999, |
|
"loss": 0.6899, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"learning_rate": 0.0005448749999999999, |
|
"loss": 0.842, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 0.0005445, |
|
"loss": 0.7052, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 0.0005441249999999999, |
|
"loss": 0.7245, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 4.69, |
|
"learning_rate": 0.00054375, |
|
"loss": 0.8852, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 0.0005433749999999999, |
|
"loss": 0.6868, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 0.000543, |
|
"loss": 0.8091, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"learning_rate": 0.000542625, |
|
"loss": 0.7912, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 0.00054225, |
|
"loss": 0.8428, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 0.0005418749999999999, |
|
"loss": 0.7627, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 4.88, |
|
"learning_rate": 0.0005414999999999999, |
|
"loss": 0.6705, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 0.0005411249999999999, |
|
"loss": 0.7269, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 4.94, |
|
"learning_rate": 0.00054075, |
|
"loss": 0.7688, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"learning_rate": 0.000540375, |
|
"loss": 0.757, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 0.00054, |
|
"loss": 0.7543, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 0.9964900612831116, |
|
"eval_mean_accuracy": 0.47220927209560704, |
|
"eval_mean_iou": 0.35076273333971525, |
|
"eval_overall_accuracy": 0.6553197827255517, |
|
"eval_per_category_accuracy": [ |
|
0.5974941039770494, |
|
0.7656300402304882, |
|
0.7736861734503012, |
|
0.26844019125474494, |
|
0.47782547654307983, |
|
0.5049230780397693, |
|
0.5503440924885342, |
|
0.38577494532904893, |
|
0.29696414469889887, |
|
0.6231179209340338, |
|
0.08607248563724125, |
|
0.33623861256409465 |
|
], |
|
"eval_per_category_iou": [ |
|
0.5029245432853789, |
|
0.5438264868522664, |
|
0.5504928848830769, |
|
0.20921503460100002, |
|
0.3484194245550177, |
|
0.34286066341367244, |
|
0.4044913460614946, |
|
0.3442055933600121, |
|
0.19153018129377833, |
|
0.4114387452368694, |
|
0.07501013614442405, |
|
0.28473776038959164 |
|
], |
|
"eval_runtime": 123.8413, |
|
"eval_samples_per_second": 0.921, |
|
"eval_steps_per_second": 0.081, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 0.000539625, |
|
"loss": 0.639, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 5.06, |
|
"learning_rate": 0.00053925, |
|
"loss": 0.7108, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 5.09, |
|
"learning_rate": 0.000538875, |
|
"loss": 0.6735, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 5.12, |
|
"learning_rate": 0.0005384999999999999, |
|
"loss": 0.7478, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"learning_rate": 0.000538125, |
|
"loss": 0.6023, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 5.19, |
|
"learning_rate": 0.0005377499999999999, |
|
"loss": 0.7976, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 5.22, |
|
"learning_rate": 0.000537375, |
|
"loss": 0.7348, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 5.25, |
|
"learning_rate": 0.0005369999999999999, |
|
"loss": 0.7037, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 5.28, |
|
"learning_rate": 0.000536625, |
|
"loss": 0.694, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 5.31, |
|
"learning_rate": 0.0005362499999999999, |
|
"loss": 0.7427, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 5.34, |
|
"learning_rate": 0.0005358749999999999, |
|
"loss": 0.7319, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 5.38, |
|
"learning_rate": 0.0005355, |
|
"loss": 0.7365, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 5.41, |
|
"learning_rate": 0.0005351249999999999, |
|
"loss": 0.823, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 5.44, |
|
"learning_rate": 0.00053475, |
|
"loss": 0.6932, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 5.47, |
|
"learning_rate": 0.0005343749999999999, |
|
"loss": 0.6911, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"learning_rate": 0.000534, |
|
"loss": 0.7337, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 0.0005336249999999999, |
|
"loss": 0.8752, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 5.56, |
|
"learning_rate": 0.00053325, |
|
"loss": 0.669, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"learning_rate": 0.000532875, |
|
"loss": 0.7305, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 5.62, |
|
"learning_rate": 0.0005324999999999999, |
|
"loss": 0.7142, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 5.62, |
|
"eval_loss": 0.9814519286155701, |
|
"eval_mean_accuracy": 0.497081755026592, |
|
"eval_mean_iou": 0.35939949129333554, |
|
"eval_overall_accuracy": 0.6597169909560889, |
|
"eval_per_category_accuracy": [ |
|
0.6908915714734837, |
|
0.7305051261897875, |
|
0.7036370157962516, |
|
0.2776663590120818, |
|
0.5628060380018922, |
|
0.4295562515651142, |
|
0.5859168478577468, |
|
0.2814967358147499, |
|
0.3033792356902656, |
|
0.6404991026728926, |
|
0.30056349150477635, |
|
0.4580632847400621 |
|
], |
|
"eval_per_category_iou": [ |
|
0.5553982956913027, |
|
0.5378390398854487, |
|
0.5530532106585494, |
|
0.21904999137749334, |
|
0.3437066198831567, |
|
0.32814516520304693, |
|
0.4234087095030862, |
|
0.2666791179990438, |
|
0.18338491161074402, |
|
0.41252681863698343, |
|
0.14896654252535288, |
|
0.34063547254581833 |
|
], |
|
"eval_runtime": 124.8537, |
|
"eval_samples_per_second": 0.913, |
|
"eval_steps_per_second": 0.08, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 5.66, |
|
"learning_rate": 0.0005321249999999999, |
|
"loss": 0.7298, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 0.00053175, |
|
"loss": 0.6999, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 5.72, |
|
"learning_rate": 0.000531375, |
|
"loss": 0.8023, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 5.75, |
|
"learning_rate": 0.000531, |
|
"loss": 0.6922, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 5.78, |
|
"learning_rate": 0.000530625, |
|
"loss": 0.7184, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 5.81, |
|
"learning_rate": 0.00053025, |
|
"loss": 0.7189, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 5.84, |
|
"learning_rate": 0.000529875, |
|
"loss": 0.614, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 5.88, |
|
"learning_rate": 0.0005294999999999999, |
|
"loss": 0.6966, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 5.91, |
|
"learning_rate": 0.000529125, |
|
"loss": 0.7259, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"learning_rate": 0.0005287499999999999, |
|
"loss": 0.7856, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 5.97, |
|
"learning_rate": 0.000528375, |
|
"loss": 0.7678, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 0.0005279999999999999, |
|
"loss": 0.623, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 0.000527625, |
|
"loss": 0.6785, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 6.06, |
|
"learning_rate": 0.0005272499999999999, |
|
"loss": 0.6884, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"learning_rate": 0.000526875, |
|
"loss": 0.6096, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 6.12, |
|
"learning_rate": 0.0005265, |
|
"loss": 0.739, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 6.16, |
|
"learning_rate": 0.0005261249999999999, |
|
"loss": 0.6649, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 6.19, |
|
"learning_rate": 0.00052575, |
|
"loss": 0.6959, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 6.22, |
|
"learning_rate": 0.0005253749999999999, |
|
"loss": 0.6718, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 0.000525, |
|
"loss": 0.7132, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"eval_loss": 0.9577063918113708, |
|
"eval_mean_accuracy": 0.49358942386599547, |
|
"eval_mean_iou": 0.36445855775079217, |
|
"eval_overall_accuracy": 0.6702703174791838, |
|
"eval_per_category_accuracy": [ |
|
0.757485170186219, |
|
0.7039742826269, |
|
0.7879783326232815, |
|
0.3002650918453029, |
|
0.530024898350307, |
|
0.5117678441164408, |
|
0.5200409667652787, |
|
0.282272468662643, |
|
0.18195167830492534, |
|
0.6273017965839506, |
|
0.20833269566867557, |
|
0.511677860658021 |
|
], |
|
"eval_per_category_iou": [ |
|
0.5717339461326115, |
|
0.5411086678260705, |
|
0.5550507730528956, |
|
0.23120184241180683, |
|
0.33923086199004965, |
|
0.35084943122443496, |
|
0.4183980452164044, |
|
0.2680651542162202, |
|
0.15420450059218319, |
|
0.4336026908109024, |
|
0.15032155761093297, |
|
0.3597352219249941 |
|
], |
|
"eval_runtime": 123.8598, |
|
"eval_samples_per_second": 0.92, |
|
"eval_steps_per_second": 0.081, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 6.28, |
|
"learning_rate": 0.0005246249999999999, |
|
"loss": 0.661, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 6.31, |
|
"learning_rate": 0.00052425, |
|
"loss": 0.7055, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 6.34, |
|
"learning_rate": 0.0005238749999999999, |
|
"loss": 0.6852, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"learning_rate": 0.0005235, |
|
"loss": 0.6802, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 6.41, |
|
"learning_rate": 0.0005231249999999999, |
|
"loss": 0.6662, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 6.44, |
|
"learning_rate": 0.0005227499999999999, |
|
"loss": 0.7165, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 6.47, |
|
"learning_rate": 0.0005223749999999999, |
|
"loss": 0.6282, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"learning_rate": 0.000522, |
|
"loss": 0.693, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 0.000521625, |
|
"loss": 0.6498, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"learning_rate": 0.00052125, |
|
"loss": 0.7024, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 6.59, |
|
"learning_rate": 0.000520875, |
|
"loss": 0.573, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 6.62, |
|
"learning_rate": 0.0005205, |
|
"loss": 0.8226, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 6.66, |
|
"learning_rate": 0.000520125, |
|
"loss": 0.5298, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 6.69, |
|
"learning_rate": 0.0005197499999999999, |
|
"loss": 0.6492, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 6.72, |
|
"learning_rate": 0.000519375, |
|
"loss": 0.769, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 6.75, |
|
"learning_rate": 0.0005189999999999999, |
|
"loss": 0.664, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 6.78, |
|
"learning_rate": 0.000518625, |
|
"loss": 0.7115, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"learning_rate": 0.0005182499999999999, |
|
"loss": 0.6526, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 6.84, |
|
"learning_rate": 0.000517875, |
|
"loss": 0.679, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 6.88, |
|
"learning_rate": 0.0005175, |
|
"loss": 0.5572, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 6.88, |
|
"eval_loss": 0.9686026573181152, |
|
"eval_mean_accuracy": 0.49162090929743196, |
|
"eval_mean_iou": 0.3674592922271513, |
|
"eval_overall_accuracy": 0.668428052935684, |
|
"eval_per_category_accuracy": [ |
|
0.8315143436240795, |
|
0.67630998091554, |
|
0.7426204308307264, |
|
0.32498386548323743, |
|
0.49111945333028095, |
|
0.5001669455140652, |
|
0.5374337360209966, |
|
0.43089816792044666, |
|
0.34422109438080617, |
|
0.6128100029154656, |
|
0.11920860204522012, |
|
0.28816428858831905 |
|
], |
|
"eval_per_category_iou": [ |
|
0.5778473653822723, |
|
0.5269478191058978, |
|
0.574435506709114, |
|
0.23668203929496492, |
|
0.3570071601716556, |
|
0.358769802685785, |
|
0.41459541328306615, |
|
0.37143503659747396, |
|
0.2224686465997259, |
|
0.4152571514805167, |
|
0.09885903570114096, |
|
0.25520652971420166 |
|
], |
|
"eval_runtime": 124.6105, |
|
"eval_samples_per_second": 0.915, |
|
"eval_steps_per_second": 0.08, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 6.91, |
|
"learning_rate": 0.0005171249999999999, |
|
"loss": 0.7046, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 6.94, |
|
"learning_rate": 0.00051675, |
|
"loss": 0.6436, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 6.97, |
|
"learning_rate": 0.0005163749999999999, |
|
"loss": 0.6286, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 0.000516, |
|
"loss": 0.621, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 0.0005156249999999999, |
|
"loss": 0.7192, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 7.06, |
|
"learning_rate": 0.00051525, |
|
"loss": 0.5768, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 7.09, |
|
"learning_rate": 0.0005148749999999999, |
|
"loss": 0.6557, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 7.12, |
|
"learning_rate": 0.0005145, |
|
"loss": 0.6194, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 7.16, |
|
"learning_rate": 0.0005141249999999999, |
|
"loss": 0.624, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 7.19, |
|
"learning_rate": 0.0005137499999999999, |
|
"loss": 0.6787, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 7.22, |
|
"learning_rate": 0.0005133749999999999, |
|
"loss": 0.6515, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 7.25, |
|
"learning_rate": 0.0005129999999999999, |
|
"loss": 0.618, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 7.28, |
|
"learning_rate": 0.000512625, |
|
"loss": 0.5743, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 7.31, |
|
"learning_rate": 0.00051225, |
|
"loss": 0.561, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 7.34, |
|
"learning_rate": 0.000511875, |
|
"loss": 0.656, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 7.38, |
|
"learning_rate": 0.0005115, |
|
"loss": 0.5951, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 7.41, |
|
"learning_rate": 0.000511125, |
|
"loss": 0.6536, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 7.44, |
|
"learning_rate": 0.0005107499999999999, |
|
"loss": 0.734, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 7.47, |
|
"learning_rate": 0.000510375, |
|
"loss": 0.6387, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 0.0005099999999999999, |
|
"loss": 0.5676, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"eval_loss": 0.9805742502212524, |
|
"eval_mean_accuracy": 0.5024018526714576, |
|
"eval_mean_iou": 0.371506102641304, |
|
"eval_overall_accuracy": 0.6800810161389803, |
|
"eval_per_category_accuracy": [ |
|
0.7315930355784842, |
|
0.7541007820537842, |
|
0.6881292554956727, |
|
0.30834622938078154, |
|
0.5766004128790718, |
|
0.4276563390300466, |
|
0.6505111399057544, |
|
0.4819251032078268, |
|
0.27926275254429783, |
|
0.5722067988240086, |
|
0.07342223909057798, |
|
0.48506814406718457 |
|
], |
|
"eval_per_category_iou": [ |
|
0.5874312271992882, |
|
0.5604905946583346, |
|
0.5587517085148045, |
|
0.2254583833345596, |
|
0.36177477005679304, |
|
0.33678157470165493, |
|
0.4185254872962371, |
|
0.3728270419920094, |
|
0.20258071367884453, |
|
0.42075932706243124, |
|
0.0692206473059698, |
|
0.34347175589472057 |
|
], |
|
"eval_runtime": 123.7241, |
|
"eval_samples_per_second": 0.921, |
|
"eval_steps_per_second": 0.081, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 7.53, |
|
"learning_rate": 0.000509625, |
|
"loss": 0.6097, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 7.56, |
|
"learning_rate": 0.0005092499999999999, |
|
"loss": 0.6319, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 7.59, |
|
"learning_rate": 0.000508875, |
|
"loss": 0.6219, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 7.62, |
|
"learning_rate": 0.0005085, |
|
"loss": 0.5975, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 7.66, |
|
"learning_rate": 0.000508125, |
|
"loss": 0.651, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 7.69, |
|
"learning_rate": 0.00050775, |
|
"loss": 0.512, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 7.72, |
|
"learning_rate": 0.0005073749999999999, |
|
"loss": 0.5962, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 7.75, |
|
"learning_rate": 0.000507, |
|
"loss": 0.7017, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"learning_rate": 0.0005066249999999999, |
|
"loss": 0.6477, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 7.81, |
|
"learning_rate": 0.00050625, |
|
"loss": 0.6546, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 7.84, |
|
"learning_rate": 0.0005058749999999999, |
|
"loss": 0.6881, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 7.88, |
|
"learning_rate": 0.0005055, |
|
"loss": 0.6078, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 7.91, |
|
"learning_rate": 0.0005051249999999999, |
|
"loss": 0.6891, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 7.94, |
|
"learning_rate": 0.00050475, |
|
"loss": 0.598, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 7.97, |
|
"learning_rate": 0.0005043749999999999, |
|
"loss": 0.638, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 0.0005039999999999999, |
|
"loss": 0.6779, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 0.000503625, |
|
"loss": 0.628, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"learning_rate": 0.0005032499999999999, |
|
"loss": 0.5842, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 8.09, |
|
"learning_rate": 0.000502875, |
|
"loss": 0.5126, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 8.12, |
|
"learning_rate": 0.0005025, |
|
"loss": 0.6533, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 8.12, |
|
"eval_loss": 0.9647707939147949, |
|
"eval_mean_accuracy": 0.5290347832309982, |
|
"eval_mean_iou": 0.3826944916955128, |
|
"eval_overall_accuracy": 0.6695300654361123, |
|
"eval_per_category_accuracy": [ |
|
0.6390227219897978, |
|
0.7622451073333125, |
|
0.6798905616562309, |
|
0.39068084190949404, |
|
0.5202418118057482, |
|
0.5864632592845296, |
|
0.7008623173460086, |
|
0.5284176442793452, |
|
0.26142604899603383, |
|
0.5910629245092603, |
|
0.23104988170045268, |
|
0.45705427796176507 |
|
], |
|
"eval_per_category_iou": [ |
|
0.539363765599418, |
|
0.5541112560719536, |
|
0.564224130990146, |
|
0.28007313069001544, |
|
0.35043846601035505, |
|
0.37497345958257, |
|
0.4145645950481569, |
|
0.40213470319634703, |
|
0.1947903409748277, |
|
0.4211198117656152, |
|
0.15867080332782618, |
|
0.33786943708892114 |
|
], |
|
"eval_runtime": 125.2253, |
|
"eval_samples_per_second": 0.91, |
|
"eval_steps_per_second": 0.08, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 8.16, |
|
"learning_rate": 0.000502125, |
|
"loss": 0.5969, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 8.19, |
|
"learning_rate": 0.00050175, |
|
"loss": 0.5633, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 8.22, |
|
"learning_rate": 0.000501375, |
|
"loss": 0.6202, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 8.25, |
|
"learning_rate": 0.0005009999999999999, |
|
"loss": 0.5568, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 8.28, |
|
"learning_rate": 0.000500625, |
|
"loss": 0.6021, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 8.31, |
|
"learning_rate": 0.0005002499999999999, |
|
"loss": 0.6637, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 8.34, |
|
"learning_rate": 0.000499875, |
|
"loss": 0.6756, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 8.38, |
|
"learning_rate": 0.0004994999999999999, |
|
"loss": 0.6323, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 8.41, |
|
"learning_rate": 0.000499125, |
|
"loss": 0.6591, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 8.44, |
|
"learning_rate": 0.00049875, |
|
"loss": 0.7191, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 8.47, |
|
"learning_rate": 0.0004983749999999999, |
|
"loss": 0.5911, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"learning_rate": 0.000498, |
|
"loss": 0.5841, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 8.53, |
|
"learning_rate": 0.0004976249999999999, |
|
"loss": 0.5837, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 8.56, |
|
"learning_rate": 0.00049725, |
|
"loss": 0.6009, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 8.59, |
|
"learning_rate": 0.0004968749999999999, |
|
"loss": 0.5528, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 8.62, |
|
"learning_rate": 0.0004965, |
|
"loss": 0.5907, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 8.66, |
|
"learning_rate": 0.0004961249999999999, |
|
"loss": 0.5582, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 8.69, |
|
"learning_rate": 0.00049575, |
|
"loss": 0.5714, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 8.72, |
|
"learning_rate": 0.0004953749999999999, |
|
"loss": 0.6407, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"learning_rate": 0.0004949999999999999, |
|
"loss": 0.67, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"eval_loss": 1.0987780094146729, |
|
"eval_mean_accuracy": 0.4580998870390869, |
|
"eval_mean_iou": 0.3479111276469824, |
|
"eval_overall_accuracy": 0.6546705814830044, |
|
"eval_per_category_accuracy": [ |
|
0.48066672745873446, |
|
0.8305757369135304, |
|
0.7676997659254449, |
|
0.21412412127604943, |
|
0.47398841588551216, |
|
0.4325884901955803, |
|
0.5711691272441585, |
|
0.2749094353329169, |
|
0.21508300577624376, |
|
0.6848772870627909, |
|
0.20816741298938213, |
|
0.3433491184086993 |
|
], |
|
"eval_per_category_iou": [ |
|
0.4455015326675329, |
|
0.5573106019694799, |
|
0.5610375857814821, |
|
0.19477330471696624, |
|
0.35352567686407715, |
|
0.32306020230054533, |
|
0.4240965616003975, |
|
0.26065154759824616, |
|
0.18418258317583033, |
|
0.4215337886509578, |
|
0.15863326887955384, |
|
0.2906268775587196 |
|
], |
|
"eval_runtime": 123.9472, |
|
"eval_samples_per_second": 0.92, |
|
"eval_steps_per_second": 0.081, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 8.78, |
|
"learning_rate": 0.000494625, |
|
"loss": 0.6267, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 8.81, |
|
"learning_rate": 0.0004942499999999999, |
|
"loss": 0.7083, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 8.84, |
|
"learning_rate": 0.000493875, |
|
"loss": 0.6574, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 8.88, |
|
"learning_rate": 0.0004935, |
|
"loss": 0.5851, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 8.91, |
|
"learning_rate": 0.000493125, |
|
"loss": 0.5928, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 8.94, |
|
"learning_rate": 0.00049275, |
|
"loss": 0.5435, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 8.97, |
|
"learning_rate": 0.000492375, |
|
"loss": 0.6518, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 0.0004919999999999999, |
|
"loss": 0.654, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 0.000491625, |
|
"loss": 0.5646, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"learning_rate": 0.0004912499999999999, |
|
"loss": 0.5048, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 9.09, |
|
"learning_rate": 0.000490875, |
|
"loss": 0.5574, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 9.12, |
|
"learning_rate": 0.0004904999999999999, |
|
"loss": 0.5995, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 9.16, |
|
"learning_rate": 0.000490125, |
|
"loss": 0.5645, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 9.19, |
|
"learning_rate": 0.00048975, |
|
"loss": 0.5597, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 9.22, |
|
"learning_rate": 0.000489375, |
|
"loss": 0.6295, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 9.25, |
|
"learning_rate": 0.000489, |
|
"loss": 0.5053, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 9.28, |
|
"learning_rate": 0.0004886249999999999, |
|
"loss": 0.4657, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 9.31, |
|
"learning_rate": 0.0004882499999999999, |
|
"loss": 0.6421, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 9.34, |
|
"learning_rate": 0.000487875, |
|
"loss": 0.5838, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 9.38, |
|
"learning_rate": 0.0004875, |
|
"loss": 0.5644, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 9.38, |
|
"eval_loss": 0.983669638633728, |
|
"eval_mean_accuracy": 0.5204264301169051, |
|
"eval_mean_iou": 0.36998021455594227, |
|
"eval_overall_accuracy": 0.6649547710753324, |
|
"eval_per_category_accuracy": [ |
|
0.5940242572097704, |
|
0.8056204971919827, |
|
0.652415122648732, |
|
0.3024337096897315, |
|
0.6671805707779671, |
|
0.7223677954844868, |
|
0.5593419811729217, |
|
0.49315930059755003, |
|
0.35564048871813525, |
|
0.4758290170708594, |
|
0.1260647724455409, |
|
0.491039648395184 |
|
], |
|
"eval_per_category_iou": [ |
|
0.5120521418371993, |
|
0.5615051709697225, |
|
0.5605287190343511, |
|
0.24633585240536524, |
|
0.31312694322687384, |
|
0.3423289627909097, |
|
0.42465141853084787, |
|
0.4075806091690605, |
|
0.2081047227792813, |
|
0.3872001100058186, |
|
0.10902319327019286, |
|
0.36732473065168403 |
|
], |
|
"eval_runtime": 123.8999, |
|
"eval_samples_per_second": 0.92, |
|
"eval_steps_per_second": 0.081, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 9.41, |
|
"learning_rate": 0.000487125, |
|
"loss": 0.6492, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 9.44, |
|
"learning_rate": 0.00048675, |
|
"loss": 0.6658, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 9.47, |
|
"learning_rate": 0.000486375, |
|
"loss": 0.5424, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"learning_rate": 0.000486, |
|
"loss": 0.4951, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 9.53, |
|
"learning_rate": 0.00048562499999999995, |
|
"loss": 0.5663, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 9.56, |
|
"learning_rate": 0.00048524999999999995, |
|
"loss": 0.5105, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 9.59, |
|
"learning_rate": 0.00048487499999999995, |
|
"loss": 0.5594, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 9.62, |
|
"learning_rate": 0.00048449999999999996, |
|
"loss": 0.5925, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 9.66, |
|
"learning_rate": 0.00048412499999999996, |
|
"loss": 0.5676, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 9.69, |
|
"learning_rate": 0.00048374999999999997, |
|
"loss": 0.552, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 9.72, |
|
"learning_rate": 0.00048337499999999997, |
|
"loss": 0.6232, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 9.75, |
|
"learning_rate": 0.000483, |
|
"loss": 0.5877, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 9.78, |
|
"learning_rate": 0.0004826249999999999, |
|
"loss": 0.5313, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 9.81, |
|
"learning_rate": 0.00048224999999999993, |
|
"loss": 0.5456, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 9.84, |
|
"learning_rate": 0.00048187499999999994, |
|
"loss": 0.5813, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 9.88, |
|
"learning_rate": 0.00048149999999999994, |
|
"loss": 0.6503, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 9.91, |
|
"learning_rate": 0.00048112499999999994, |
|
"loss": 0.5662, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 9.94, |
|
"learning_rate": 0.00048074999999999995, |
|
"loss": 0.6718, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 9.97, |
|
"learning_rate": 0.00048037499999999995, |
|
"loss": 0.5636, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 0.00047999999999999996, |
|
"loss": 0.5678, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 0.9780566692352295, |
|
"eval_mean_accuracy": 0.5350595191545178, |
|
"eval_mean_iou": 0.3906962436607775, |
|
"eval_overall_accuracy": 0.6771719413891173, |
|
"eval_per_category_accuracy": [ |
|
0.6554152883827387, |
|
0.7588004293882252, |
|
0.7401698246138683, |
|
0.3209831992838356, |
|
0.5589183297643672, |
|
0.5685601675261941, |
|
0.5751910181705395, |
|
0.45440587328339643, |
|
0.4784721790901751, |
|
0.6453562996222412, |
|
0.21756097859589305, |
|
0.4468806421327391 |
|
], |
|
"eval_per_category_iou": [ |
|
0.5545401246305556, |
|
0.559287046180608, |
|
0.5841953654092598, |
|
0.24684268808453297, |
|
0.3812876262500622, |
|
0.3824948025455624, |
|
0.4319588439462186, |
|
0.384631151257913, |
|
0.2309813395356408, |
|
0.4207958788984324, |
|
0.15715858689932388, |
|
0.3541814702912209 |
|
], |
|
"eval_runtime": 123.6926, |
|
"eval_samples_per_second": 0.922, |
|
"eval_steps_per_second": 0.081, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 0.0004796249999999999, |
|
"loss": 0.5511, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 10.06, |
|
"learning_rate": 0.0004792499999999999, |
|
"loss": 0.611, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 10.09, |
|
"learning_rate": 0.0004788749999999999, |
|
"loss": 0.5803, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 10.12, |
|
"learning_rate": 0.0004785, |
|
"loss": 0.5232, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 10.16, |
|
"learning_rate": 0.000478125, |
|
"loss": 0.5558, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 10.19, |
|
"learning_rate": 0.00047775, |
|
"loss": 0.5829, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 10.22, |
|
"learning_rate": 0.000477375, |
|
"loss": 0.5082, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 10.25, |
|
"learning_rate": 0.000477, |
|
"loss": 0.5775, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 10.28, |
|
"learning_rate": 0.000476625, |
|
"loss": 0.5354, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 10.31, |
|
"learning_rate": 0.00047624999999999995, |
|
"loss": 0.4546, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 10.34, |
|
"learning_rate": 0.00047587499999999995, |
|
"loss": 0.5009, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 10.38, |
|
"learning_rate": 0.00047549999999999996, |
|
"loss": 0.5521, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 10.41, |
|
"learning_rate": 0.00047512499999999996, |
|
"loss": 0.5354, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 10.44, |
|
"learning_rate": 0.00047474999999999997, |
|
"loss": 0.4855, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 10.47, |
|
"learning_rate": 0.00047437499999999997, |
|
"loss": 0.5356, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 10.5, |
|
"learning_rate": 0.000474, |
|
"loss": 0.5868, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 10.53, |
|
"learning_rate": 0.000473625, |
|
"loss": 0.5931, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 10.56, |
|
"learning_rate": 0.00047324999999999993, |
|
"loss": 0.5786, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 10.59, |
|
"learning_rate": 0.00047287499999999993, |
|
"loss": 0.5718, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 10.62, |
|
"learning_rate": 0.00047249999999999994, |
|
"loss": 0.6153, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 10.62, |
|
"eval_loss": 0.9987176060676575, |
|
"eval_mean_accuracy": 0.5497999939840363, |
|
"eval_mean_iou": 0.3857005670859029, |
|
"eval_overall_accuracy": 0.660497130009166, |
|
"eval_per_category_accuracy": [ |
|
0.5231205889958428, |
|
0.7747069218397336, |
|
0.7498118814452185, |
|
0.3063129332898453, |
|
0.546321263677379, |
|
0.6574151027622224, |
|
0.6809347098082498, |
|
0.45994636300225755, |
|
0.4333802117068756, |
|
0.5861572233919434, |
|
0.342909526097829, |
|
0.5365832017910386 |
|
], |
|
"eval_per_category_iou": [ |
|
0.4772630084705572, |
|
0.5541882497823591, |
|
0.5825233033041953, |
|
0.24382429625701532, |
|
0.38442721103951644, |
|
0.38358193468602964, |
|
0.42849690958511605, |
|
0.3960627429416036, |
|
0.21535871992663122, |
|
0.4152440328818222, |
|
0.1834708686588136, |
|
0.36396552749717453 |
|
], |
|
"eval_runtime": 123.7757, |
|
"eval_samples_per_second": 0.921, |
|
"eval_steps_per_second": 0.081, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 10.66, |
|
"learning_rate": 0.00047212499999999994, |
|
"loss": 0.6764, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 10.69, |
|
"learning_rate": 0.00047174999999999995, |
|
"loss": 0.541, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 10.72, |
|
"learning_rate": 0.00047137499999999995, |
|
"loss": 0.5097, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 10.75, |
|
"learning_rate": 0.00047099999999999996, |
|
"loss": 0.4856, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 10.78, |
|
"learning_rate": 0.00047062499999999996, |
|
"loss": 0.5468, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 10.81, |
|
"learning_rate": 0.0004702499999999999, |
|
"loss": 0.5513, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 10.84, |
|
"learning_rate": 0.0004698749999999999, |
|
"loss": 0.6115, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 10.88, |
|
"learning_rate": 0.0004694999999999999, |
|
"loss": 0.5045, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 10.91, |
|
"learning_rate": 0.0004691249999999999, |
|
"loss": 0.5671, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 10.94, |
|
"learning_rate": 0.00046875, |
|
"loss": 0.5062, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 10.97, |
|
"learning_rate": 0.000468375, |
|
"loss": 0.6014, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 0.000468, |
|
"loss": 0.5981, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 0.000467625, |
|
"loss": 0.5369, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 11.06, |
|
"learning_rate": 0.00046725, |
|
"loss": 0.4975, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 11.09, |
|
"learning_rate": 0.00046687499999999995, |
|
"loss": 0.5137, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 11.12, |
|
"learning_rate": 0.00046649999999999996, |
|
"loss": 0.4978, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 11.16, |
|
"learning_rate": 0.00046612499999999996, |
|
"loss": 0.4579, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 11.19, |
|
"learning_rate": 0.00046574999999999996, |
|
"loss": 0.5321, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 11.22, |
|
"learning_rate": 0.00046537499999999997, |
|
"loss": 0.5201, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 11.25, |
|
"learning_rate": 0.00046499999999999997, |
|
"loss": 0.4252, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 11.25, |
|
"eval_loss": 0.9291033148765564, |
|
"eval_mean_accuracy": 0.528009973276733, |
|
"eval_mean_iou": 0.39591471843488896, |
|
"eval_overall_accuracy": 0.6944302341394257, |
|
"eval_per_category_accuracy": [ |
|
0.7758503951264145, |
|
0.7639770669563318, |
|
0.6693553518837109, |
|
0.38660037057341723, |
|
0.5337667415575549, |
|
0.5057342154831077, |
|
0.5763875877943601, |
|
0.6287229015837165, |
|
0.3423444806260703, |
|
0.5546978372878139, |
|
0.15132853605458002, |
|
0.447354194393719 |
|
], |
|
"eval_per_category_iou": [ |
|
0.6138349391900487, |
|
0.5748497846975844, |
|
0.5686854787211605, |
|
0.28537363740676996, |
|
0.3701880790533382, |
|
0.3687162885337687, |
|
0.4387556373721067, |
|
0.4152303639246888, |
|
0.22665730662981398, |
|
0.4070483981278673, |
|
0.12507399555773677, |
|
0.35656271200378264 |
|
], |
|
"eval_runtime": 123.9322, |
|
"eval_samples_per_second": 0.92, |
|
"eval_steps_per_second": 0.081, |
|
"step": 360 |
|
} |
|
], |
|
"max_steps": 1600, |
|
"num_train_epochs": 50, |
|
"total_flos": 7.462066884968448e+16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|