File size: 104,455 Bytes
8cf63df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
2022-05-26 10:46:41,629 INFO [train.py:906] (0/4) Training started
2022-05-26 10:46:41,632 INFO [train.py:916] (0/4) Device: cuda:0
2022-05-26 10:46:41,634 INFO [train.py:934] (0/4) {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Debug', 'k2-with-cuda': True, 'k2-git-sha1': 'ecfe7bd6d9189964bf3ff043038918d889a43185', 'k2-git-date': 'Tue May 10 10:57:55 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': '364bccb-clean', 'icefall-git-date': 'Thu May 26 10:29:08 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_reworked2', 'k2-path': '/ceph-kw/kangwei/code/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-3-0307202051-57dc848959-8tmmp', 'IP address': '10.177.24.138'}, 'world_size': 4, 'master_port': 13498, 'tensorboard': True, 'num_epochs': 50, 'start_epoch': 1, 'start_batch': 0, 'exp_dir': PosixPath('streaming_pruned_transducer_stateless4/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'initial_lr': 0.003, 'lr_batches': 5000, 'lr_epochs': 6, 'context_size': 2, 'prune_range': 5, 'lm_scale': 0.25, 'am_scale': 0.0, 'simple_loss_scale': 0.5, 'seed': 42, 'print_diagnostics': False, 'save_every_n': 8000, 'keep_last_k': 20, 'average_period': 100, 'use_fp16': False, 'dynamic_chunk_training': True, 'causal_convolution': True, 'short_chunk_size': 32, 'num_left_chunks': 4, 'delay_penalty': 0.0, 'return_sym_delay': False, 'full_libri': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 300, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'blank_id': 0, 'vocab_size': 500}
2022-05-26 10:46:41,634 INFO [train.py:936] (0/4) About to create model
2022-05-26 10:46:42,071 INFO [train.py:940] (0/4) Number of model parameters: 78648040
2022-05-26 10:46:47,119 INFO [train.py:955] (0/4) Using DDP
2022-05-26 10:46:47,469 INFO [asr_datamodule.py:391] (0/4) About to get train-clean-100 cuts
2022-05-26 10:46:54,130 INFO [asr_datamodule.py:398] (0/4) About to get train-clean-360 cuts
2022-05-26 10:47:21,021 INFO [asr_datamodule.py:405] (0/4) About to get train-other-500 cuts
2022-05-26 10:48:07,489 INFO [asr_datamodule.py:209] (0/4) Enable MUSAN
2022-05-26 10:48:07,490 INFO [asr_datamodule.py:210] (0/4) About to get Musan cuts
2022-05-26 10:48:08,907 INFO [asr_datamodule.py:238] (0/4) Enable SpecAugment
2022-05-26 10:48:08,907 INFO [asr_datamodule.py:239] (0/4) Time warp factor: 80
2022-05-26 10:48:08,907 INFO [asr_datamodule.py:251] (0/4) Num frame mask: 10
2022-05-26 10:48:08,908 INFO [asr_datamodule.py:264] (0/4) About to create train dataset
2022-05-26 10:48:08,908 INFO [asr_datamodule.py:292] (0/4) Using BucketingSampler.
2022-05-26 10:48:13,977 INFO [asr_datamodule.py:308] (0/4) About to create train dataloader
2022-05-26 10:48:13,978 INFO [asr_datamodule.py:412] (0/4) About to get dev-clean cuts
2022-05-26 10:48:14,266 INFO [asr_datamodule.py:417] (0/4) About to get dev-other cuts
2022-05-26 10:48:14,399 INFO [asr_datamodule.py:339] (0/4) About to create dev dataset
2022-05-26 10:48:14,416 INFO [asr_datamodule.py:358] (0/4) About to create dev dataloader
2022-05-26 10:48:14,417 INFO [train.py:1082] (0/4) Sanity check -- see if any of the batches in epoch 1 would cause OOM.
2022-05-26 10:48:23,779 INFO [distributed.py:874] (0/4) Reducer buckets have been rebuilt in this iteration.
2022-05-26 10:48:40,752 INFO [train.py:842] (0/4) Epoch 1, batch 0, loss[loss=0.784, simple_loss=1.568, pruned_loss=6.534, over 7290.00 frames.], tot_loss[loss=0.784, simple_loss=1.568, pruned_loss=6.534, over 7290.00 frames.], batch size: 17, lr: 3.00e-03
2022-05-26 10:49:19,668 INFO [train.py:842] (0/4) Epoch 1, batch 50, loss[loss=0.5058, simple_loss=1.012, pruned_loss=7.013, over 7164.00 frames.], tot_loss[loss=0.5659, simple_loss=1.132, pruned_loss=7.102, over 323787.58 frames.], batch size: 19, lr: 3.00e-03
2022-05-26 10:49:59,116 INFO [train.py:842] (0/4) Epoch 1, batch 100, loss[loss=0.3827, simple_loss=0.7654, pruned_loss=6.714, over 6988.00 frames.], tot_loss[loss=0.5029, simple_loss=1.006, pruned_loss=6.997, over 566429.21 frames.], batch size: 16, lr: 3.00e-03
2022-05-26 10:50:37,759 INFO [train.py:842] (0/4) Epoch 1, batch 150, loss[loss=0.3718, simple_loss=0.7435, pruned_loss=6.727, over 7447.00 frames.], tot_loss[loss=0.4746, simple_loss=0.9491, pruned_loss=6.936, over 758181.62 frames.], batch size: 17, lr: 3.00e-03
2022-05-26 10:51:16,801 INFO [train.py:842] (0/4) Epoch 1, batch 200, loss[loss=0.4291, simple_loss=0.8581, pruned_loss=6.756, over 7272.00 frames.], tot_loss[loss=0.4535, simple_loss=0.907, pruned_loss=6.887, over 908118.90 frames.], batch size: 25, lr: 3.00e-03
2022-05-26 10:51:55,438 INFO [train.py:842] (0/4) Epoch 1, batch 250, loss[loss=0.4154, simple_loss=0.8308, pruned_loss=6.659, over 7327.00 frames.], tot_loss[loss=0.4399, simple_loss=0.8799, pruned_loss=6.836, over 1017619.37 frames.], batch size: 21, lr: 3.00e-03
2022-05-26 10:52:34,306 INFO [train.py:842] (0/4) Epoch 1, batch 300, loss[loss=0.4177, simple_loss=0.8353, pruned_loss=6.776, over 7271.00 frames.], tot_loss[loss=0.4303, simple_loss=0.8605, pruned_loss=6.797, over 1109133.25 frames.], batch size: 25, lr: 3.00e-03
2022-05-26 10:53:13,223 INFO [train.py:842] (0/4) Epoch 1, batch 350, loss[loss=0.3705, simple_loss=0.7409, pruned_loss=6.61, over 7256.00 frames.], tot_loss[loss=0.4216, simple_loss=0.8432, pruned_loss=6.77, over 1178705.90 frames.], batch size: 19, lr: 3.00e-03
2022-05-26 10:53:52,158 INFO [train.py:842] (0/4) Epoch 1, batch 400, loss[loss=0.3993, simple_loss=0.7985, pruned_loss=6.79, over 7415.00 frames.], tot_loss[loss=0.4141, simple_loss=0.8281, pruned_loss=6.751, over 1230762.81 frames.], batch size: 21, lr: 3.00e-03
2022-05-26 10:54:30,787 INFO [train.py:842] (0/4) Epoch 1, batch 450, loss[loss=0.407, simple_loss=0.8141, pruned_loss=6.668, over 7418.00 frames.], tot_loss[loss=0.4078, simple_loss=0.8156, pruned_loss=6.731, over 1268656.02 frames.], batch size: 21, lr: 2.99e-03
2022-05-26 10:55:09,750 INFO [train.py:842] (0/4) Epoch 1, batch 500, loss[loss=0.394, simple_loss=0.788, pruned_loss=6.678, over 7192.00 frames.], tot_loss[loss=0.401, simple_loss=0.8021, pruned_loss=6.71, over 1304508.80 frames.], batch size: 22, lr: 2.99e-03
2022-05-26 10:55:48,092 INFO [train.py:842] (0/4) Epoch 1, batch 550, loss[loss=0.3586, simple_loss=0.7172, pruned_loss=6.657, over 7344.00 frames.], tot_loss[loss=0.3953, simple_loss=0.7906, pruned_loss=6.705, over 1330878.59 frames.], batch size: 22, lr: 2.99e-03
2022-05-26 10:56:27,036 INFO [train.py:842] (0/4) Epoch 1, batch 600, loss[loss=0.3497, simple_loss=0.6993, pruned_loss=6.711, over 7126.00 frames.], tot_loss[loss=0.3837, simple_loss=0.7675, pruned_loss=6.694, over 1351464.36 frames.], batch size: 21, lr: 2.99e-03
2022-05-26 10:57:05,654 INFO [train.py:842] (0/4) Epoch 1, batch 650, loss[loss=0.2903, simple_loss=0.5806, pruned_loss=6.585, over 7029.00 frames.], tot_loss[loss=0.3717, simple_loss=0.7435, pruned_loss=6.697, over 1369743.70 frames.], batch size: 16, lr: 2.99e-03
2022-05-26 10:57:44,552 INFO [train.py:842] (0/4) Epoch 1, batch 700, loss[loss=0.3273, simple_loss=0.6545, pruned_loss=6.822, over 7208.00 frames.], tot_loss[loss=0.3586, simple_loss=0.7172, pruned_loss=6.695, over 1381299.87 frames.], batch size: 23, lr: 2.99e-03
2022-05-26 10:58:23,525 INFO [train.py:842] (0/4) Epoch 1, batch 750, loss[loss=0.2744, simple_loss=0.5488, pruned_loss=6.648, over 7276.00 frames.], tot_loss[loss=0.346, simple_loss=0.6919, pruned_loss=6.695, over 1392933.57 frames.], batch size: 17, lr: 2.98e-03
2022-05-26 10:59:02,467 INFO [train.py:842] (0/4) Epoch 1, batch 800, loss[loss=0.2985, simple_loss=0.597, pruned_loss=6.75, over 7104.00 frames.], tot_loss[loss=0.3358, simple_loss=0.6716, pruned_loss=6.705, over 1398271.74 frames.], batch size: 21, lr: 2.98e-03
2022-05-26 10:59:41,342 INFO [train.py:842] (0/4) Epoch 1, batch 850, loss[loss=0.3301, simple_loss=0.6602, pruned_loss=6.88, over 7230.00 frames.], tot_loss[loss=0.3258, simple_loss=0.6516, pruned_loss=6.713, over 1403068.37 frames.], batch size: 21, lr: 2.98e-03
2022-05-26 11:00:20,376 INFO [train.py:842] (0/4) Epoch 1, batch 900, loss[loss=0.2737, simple_loss=0.5475, pruned_loss=6.687, over 7337.00 frames.], tot_loss[loss=0.3162, simple_loss=0.6325, pruned_loss=6.717, over 1407943.14 frames.], batch size: 21, lr: 2.98e-03
2022-05-26 11:00:58,853 INFO [train.py:842] (0/4) Epoch 1, batch 950, loss[loss=0.2418, simple_loss=0.4836, pruned_loss=6.59, over 7013.00 frames.], tot_loss[loss=0.3081, simple_loss=0.6162, pruned_loss=6.721, over 1405257.81 frames.], batch size: 16, lr: 2.97e-03
2022-05-26 11:01:37,553 INFO [train.py:842] (0/4) Epoch 1, batch 1000, loss[loss=0.2359, simple_loss=0.4719, pruned_loss=6.638, over 6975.00 frames.], tot_loss[loss=0.301, simple_loss=0.602, pruned_loss=6.725, over 1404991.06 frames.], batch size: 16, lr: 2.97e-03
2022-05-26 11:02:16,197 INFO [train.py:842] (0/4) Epoch 1, batch 1050, loss[loss=0.2498, simple_loss=0.4996, pruned_loss=6.618, over 6990.00 frames.], tot_loss[loss=0.2951, simple_loss=0.5901, pruned_loss=6.729, over 1407362.68 frames.], batch size: 16, lr: 2.97e-03
2022-05-26 11:02:54,893 INFO [train.py:842] (0/4) Epoch 1, batch 1100, loss[loss=0.2648, simple_loss=0.5296, pruned_loss=6.722, over 7217.00 frames.], tot_loss[loss=0.2896, simple_loss=0.5791, pruned_loss=6.73, over 1411554.61 frames.], batch size: 22, lr: 2.96e-03
2022-05-26 11:03:33,468 INFO [train.py:842] (0/4) Epoch 1, batch 1150, loss[loss=0.2694, simple_loss=0.5387, pruned_loss=6.77, over 6785.00 frames.], tot_loss[loss=0.2844, simple_loss=0.5688, pruned_loss=6.737, over 1412326.11 frames.], batch size: 31, lr: 2.96e-03
2022-05-26 11:04:12,434 INFO [train.py:842] (0/4) Epoch 1, batch 1200, loss[loss=0.2591, simple_loss=0.5182, pruned_loss=6.722, over 7209.00 frames.], tot_loss[loss=0.279, simple_loss=0.5581, pruned_loss=6.74, over 1420125.97 frames.], batch size: 26, lr: 2.96e-03
2022-05-26 11:04:50,802 INFO [train.py:842] (0/4) Epoch 1, batch 1250, loss[loss=0.309, simple_loss=0.618, pruned_loss=6.933, over 7386.00 frames.], tot_loss[loss=0.2741, simple_loss=0.5483, pruned_loss=6.742, over 1412927.74 frames.], batch size: 23, lr: 2.95e-03
2022-05-26 11:05:29,859 INFO [train.py:842] (0/4) Epoch 1, batch 1300, loss[loss=0.279, simple_loss=0.558, pruned_loss=6.955, over 7303.00 frames.], tot_loss[loss=0.2694, simple_loss=0.5387, pruned_loss=6.747, over 1421349.61 frames.], batch size: 24, lr: 2.95e-03
2022-05-26 11:06:08,501 INFO [train.py:842] (0/4) Epoch 1, batch 1350, loss[loss=0.2427, simple_loss=0.4855, pruned_loss=6.75, over 7132.00 frames.], tot_loss[loss=0.2652, simple_loss=0.5304, pruned_loss=6.752, over 1422297.40 frames.], batch size: 20, lr: 2.95e-03
2022-05-26 11:06:47,135 INFO [train.py:842] (0/4) Epoch 1, batch 1400, loss[loss=0.2573, simple_loss=0.5146, pruned_loss=6.875, over 7278.00 frames.], tot_loss[loss=0.2635, simple_loss=0.5269, pruned_loss=6.764, over 1418016.83 frames.], batch size: 24, lr: 2.94e-03
2022-05-26 11:07:25,754 INFO [train.py:842] (0/4) Epoch 1, batch 1450, loss[loss=0.2064, simple_loss=0.4128, pruned_loss=6.695, over 7143.00 frames.], tot_loss[loss=0.2599, simple_loss=0.5199, pruned_loss=6.767, over 1417934.65 frames.], batch size: 17, lr: 2.94e-03
2022-05-26 11:08:04,594 INFO [train.py:842] (0/4) Epoch 1, batch 1500, loss[loss=0.2327, simple_loss=0.4654, pruned_loss=6.824, over 7305.00 frames.], tot_loss[loss=0.2564, simple_loss=0.5128, pruned_loss=6.774, over 1421322.57 frames.], batch size: 24, lr: 2.94e-03
2022-05-26 11:08:43,069 INFO [train.py:842] (0/4) Epoch 1, batch 1550, loss[loss=0.2553, simple_loss=0.5106, pruned_loss=6.814, over 7107.00 frames.], tot_loss[loss=0.2532, simple_loss=0.5064, pruned_loss=6.779, over 1421558.94 frames.], batch size: 21, lr: 2.93e-03
2022-05-26 11:09:22,177 INFO [train.py:842] (0/4) Epoch 1, batch 1600, loss[loss=0.2564, simple_loss=0.5127, pruned_loss=6.83, over 7326.00 frames.], tot_loss[loss=0.25, simple_loss=0.5, pruned_loss=6.781, over 1420706.31 frames.], batch size: 20, lr: 2.93e-03
2022-05-26 11:10:01,420 INFO [train.py:842] (0/4) Epoch 1, batch 1650, loss[loss=0.2364, simple_loss=0.4728, pruned_loss=6.815, over 7153.00 frames.], tot_loss[loss=0.2467, simple_loss=0.4935, pruned_loss=6.786, over 1422854.74 frames.], batch size: 18, lr: 2.92e-03
2022-05-26 11:10:40,799 INFO [train.py:842] (0/4) Epoch 1, batch 1700, loss[loss=0.2381, simple_loss=0.4763, pruned_loss=6.822, over 6293.00 frames.], tot_loss[loss=0.2449, simple_loss=0.4898, pruned_loss=6.791, over 1417681.49 frames.], batch size: 37, lr: 2.92e-03
2022-05-26 11:11:19,945 INFO [train.py:842] (0/4) Epoch 1, batch 1750, loss[loss=0.2491, simple_loss=0.4982, pruned_loss=6.739, over 6380.00 frames.], tot_loss[loss=0.2422, simple_loss=0.4843, pruned_loss=6.792, over 1417955.63 frames.], batch size: 37, lr: 2.91e-03
2022-05-26 11:11:59,998 INFO [train.py:842] (0/4) Epoch 1, batch 1800, loss[loss=0.2499, simple_loss=0.4997, pruned_loss=6.874, over 7066.00 frames.], tot_loss[loss=0.2404, simple_loss=0.4807, pruned_loss=6.797, over 1418143.78 frames.], batch size: 28, lr: 2.91e-03
2022-05-26 11:12:39,025 INFO [train.py:842] (0/4) Epoch 1, batch 1850, loss[loss=0.2581, simple_loss=0.5162, pruned_loss=6.869, over 5240.00 frames.], tot_loss[loss=0.2384, simple_loss=0.4768, pruned_loss=6.801, over 1418995.55 frames.], batch size: 55, lr: 2.91e-03
2022-05-26 11:13:18,082 INFO [train.py:842] (0/4) Epoch 1, batch 1900, loss[loss=0.2252, simple_loss=0.4504, pruned_loss=6.893, over 7260.00 frames.], tot_loss[loss=0.2371, simple_loss=0.4742, pruned_loss=6.804, over 1419026.54 frames.], batch size: 19, lr: 2.90e-03
2022-05-26 11:13:56,904 INFO [train.py:842] (0/4) Epoch 1, batch 1950, loss[loss=0.2174, simple_loss=0.4349, pruned_loss=6.805, over 7320.00 frames.], tot_loss[loss=0.235, simple_loss=0.47, pruned_loss=6.797, over 1421939.18 frames.], batch size: 21, lr: 2.90e-03
2022-05-26 11:14:35,965 INFO [train.py:842] (0/4) Epoch 1, batch 2000, loss[loss=0.1976, simple_loss=0.3952, pruned_loss=6.704, over 6831.00 frames.], tot_loss[loss=0.2326, simple_loss=0.4653, pruned_loss=6.792, over 1422015.14 frames.], batch size: 15, lr: 2.89e-03
2022-05-26 11:15:15,095 INFO [train.py:842] (0/4) Epoch 1, batch 2050, loss[loss=0.2332, simple_loss=0.4663, pruned_loss=6.854, over 7170.00 frames.], tot_loss[loss=0.2313, simple_loss=0.4626, pruned_loss=6.792, over 1420442.88 frames.], batch size: 26, lr: 2.89e-03
2022-05-26 11:15:53,879 INFO [train.py:842] (0/4) Epoch 1, batch 2100, loss[loss=0.2075, simple_loss=0.415, pruned_loss=6.655, over 7180.00 frames.], tot_loss[loss=0.2304, simple_loss=0.4609, pruned_loss=6.795, over 1418082.21 frames.], batch size: 18, lr: 2.88e-03
2022-05-26 11:16:32,684 INFO [train.py:842] (0/4) Epoch 1, batch 2150, loss[loss=0.2243, simple_loss=0.4485, pruned_loss=6.843, over 7341.00 frames.], tot_loss[loss=0.2289, simple_loss=0.4579, pruned_loss=6.796, over 1422372.14 frames.], batch size: 22, lr: 2.88e-03
2022-05-26 11:17:11,542 INFO [train.py:842] (0/4) Epoch 1, batch 2200, loss[loss=0.2257, simple_loss=0.4515, pruned_loss=6.891, over 7324.00 frames.], tot_loss[loss=0.2281, simple_loss=0.4563, pruned_loss=6.792, over 1421991.23 frames.], batch size: 25, lr: 2.87e-03
2022-05-26 11:17:50,098 INFO [train.py:842] (0/4) Epoch 1, batch 2250, loss[loss=0.2342, simple_loss=0.4685, pruned_loss=6.834, over 7216.00 frames.], tot_loss[loss=0.227, simple_loss=0.454, pruned_loss=6.796, over 1420357.49 frames.], batch size: 21, lr: 2.86e-03
2022-05-26 11:18:28,792 INFO [train.py:842] (0/4) Epoch 1, batch 2300, loss[loss=0.2194, simple_loss=0.4389, pruned_loss=6.713, over 7270.00 frames.], tot_loss[loss=0.2267, simple_loss=0.4534, pruned_loss=6.801, over 1415292.70 frames.], batch size: 19, lr: 2.86e-03
2022-05-26 11:19:07,806 INFO [train.py:842] (0/4) Epoch 1, batch 2350, loss[loss=0.2265, simple_loss=0.453, pruned_loss=6.892, over 5124.00 frames.], tot_loss[loss=0.2267, simple_loss=0.4534, pruned_loss=6.804, over 1415393.47 frames.], batch size: 52, lr: 2.85e-03
2022-05-26 11:19:47,045 INFO [train.py:842] (0/4) Epoch 1, batch 2400, loss[loss=0.2109, simple_loss=0.4218, pruned_loss=6.796, over 7419.00 frames.], tot_loss[loss=0.2254, simple_loss=0.4508, pruned_loss=6.807, over 1410888.37 frames.], batch size: 20, lr: 2.85e-03
2022-05-26 11:20:25,501 INFO [train.py:842] (0/4) Epoch 1, batch 2450, loss[loss=0.2448, simple_loss=0.4895, pruned_loss=6.841, over 5193.00 frames.], tot_loss[loss=0.2248, simple_loss=0.4496, pruned_loss=6.805, over 1410887.59 frames.], batch size: 53, lr: 2.84e-03
2022-05-26 11:21:04,495 INFO [train.py:842] (0/4) Epoch 1, batch 2500, loss[loss=0.2316, simple_loss=0.4632, pruned_loss=6.867, over 7319.00 frames.], tot_loss[loss=0.222, simple_loss=0.444, pruned_loss=6.803, over 1417263.32 frames.], batch size: 20, lr: 2.84e-03
2022-05-26 11:21:42,892 INFO [train.py:842] (0/4) Epoch 1, batch 2550, loss[loss=0.196, simple_loss=0.392, pruned_loss=6.793, over 7406.00 frames.], tot_loss[loss=0.2227, simple_loss=0.4455, pruned_loss=6.808, over 1418544.68 frames.], batch size: 18, lr: 2.83e-03
2022-05-26 11:22:21,926 INFO [train.py:842] (0/4) Epoch 1, batch 2600, loss[loss=0.2393, simple_loss=0.4785, pruned_loss=6.938, over 7240.00 frames.], tot_loss[loss=0.2206, simple_loss=0.4411, pruned_loss=6.802, over 1421652.87 frames.], batch size: 20, lr: 2.83e-03
2022-05-26 11:23:00,462 INFO [train.py:842] (0/4) Epoch 1, batch 2650, loss[loss=0.2014, simple_loss=0.4029, pruned_loss=6.771, over 7234.00 frames.], tot_loss[loss=0.2186, simple_loss=0.4373, pruned_loss=6.793, over 1422753.88 frames.], batch size: 20, lr: 2.82e-03
2022-05-26 11:23:39,467 INFO [train.py:842] (0/4) Epoch 1, batch 2700, loss[loss=0.2136, simple_loss=0.4273, pruned_loss=6.851, over 7147.00 frames.], tot_loss[loss=0.218, simple_loss=0.436, pruned_loss=6.788, over 1422424.12 frames.], batch size: 20, lr: 2.81e-03
2022-05-26 11:24:17,966 INFO [train.py:842] (0/4) Epoch 1, batch 2750, loss[loss=0.2052, simple_loss=0.4104, pruned_loss=6.834, over 7324.00 frames.], tot_loss[loss=0.2169, simple_loss=0.4339, pruned_loss=6.787, over 1423446.62 frames.], batch size: 20, lr: 2.81e-03
2022-05-26 11:24:56,712 INFO [train.py:842] (0/4) Epoch 1, batch 2800, loss[loss=0.2107, simple_loss=0.4215, pruned_loss=6.795, over 7147.00 frames.], tot_loss[loss=0.2172, simple_loss=0.4344, pruned_loss=6.791, over 1422233.31 frames.], batch size: 20, lr: 2.80e-03
2022-05-26 11:25:35,295 INFO [train.py:842] (0/4) Epoch 1, batch 2850, loss[loss=0.2245, simple_loss=0.4491, pruned_loss=6.74, over 7360.00 frames.], tot_loss[loss=0.2168, simple_loss=0.4337, pruned_loss=6.793, over 1424762.33 frames.], batch size: 19, lr: 2.80e-03
2022-05-26 11:26:13,803 INFO [train.py:842] (0/4) Epoch 1, batch 2900, loss[loss=0.2036, simple_loss=0.4072, pruned_loss=6.778, over 7342.00 frames.], tot_loss[loss=0.2167, simple_loss=0.4334, pruned_loss=6.797, over 1420410.66 frames.], batch size: 20, lr: 2.79e-03
2022-05-26 11:26:52,582 INFO [train.py:842] (0/4) Epoch 1, batch 2950, loss[loss=0.2019, simple_loss=0.4038, pruned_loss=6.794, over 7150.00 frames.], tot_loss[loss=0.2147, simple_loss=0.4294, pruned_loss=6.795, over 1417160.93 frames.], batch size: 26, lr: 2.78e-03
2022-05-26 11:27:31,318 INFO [train.py:842] (0/4) Epoch 1, batch 3000, loss[loss=0.356, simple_loss=0.3797, pruned_loss=1.662, over 7281.00 frames.], tot_loss[loss=0.248, simple_loss=0.4285, pruned_loss=6.765, over 1421118.41 frames.], batch size: 17, lr: 2.78e-03
2022-05-26 11:27:31,319 INFO [train.py:862] (0/4) Computing validation loss
2022-05-26 11:27:40,551 INFO [train.py:871] (0/4) Epoch 1, validation: loss=2.017, simple_loss=0.4861, pruned_loss=1.774, over 868885.00 frames.
2022-05-26 11:28:19,136 INFO [train.py:842] (0/4) Epoch 1, batch 3050, loss[loss=0.3566, simple_loss=0.495, pruned_loss=1.091, over 6463.00 frames.], tot_loss[loss=0.2752, simple_loss=0.4393, pruned_loss=5.562, over 1419999.13 frames.], batch size: 38, lr: 2.77e-03
2022-05-26 11:28:58,769 INFO [train.py:842] (0/4) Epoch 1, batch 3100, loss[loss=0.3004, simple_loss=0.4499, pruned_loss=0.7541, over 7420.00 frames.], tot_loss[loss=0.2798, simple_loss=0.4349, pruned_loss=4.503, over 1426509.22 frames.], batch size: 21, lr: 2.77e-03
2022-05-26 11:29:37,514 INFO [train.py:842] (0/4) Epoch 1, batch 3150, loss[loss=0.2875, simple_loss=0.4663, pruned_loss=0.5437, over 7406.00 frames.], tot_loss[loss=0.2788, simple_loss=0.4336, pruned_loss=3.637, over 1427277.74 frames.], batch size: 21, lr: 2.76e-03
2022-05-26 11:30:16,512 INFO [train.py:842] (0/4) Epoch 1, batch 3200, loss[loss=0.2713, simple_loss=0.4543, pruned_loss=0.4412, over 7291.00 frames.], tot_loss[loss=0.2747, simple_loss=0.4331, pruned_loss=2.936, over 1423862.67 frames.], batch size: 24, lr: 2.75e-03
2022-05-26 11:30:54,979 INFO [train.py:842] (0/4) Epoch 1, batch 3250, loss[loss=0.2604, simple_loss=0.4486, pruned_loss=0.3616, over 7144.00 frames.], tot_loss[loss=0.2696, simple_loss=0.4324, pruned_loss=2.367, over 1423544.71 frames.], batch size: 20, lr: 2.75e-03
2022-05-26 11:31:34,088 INFO [train.py:842] (0/4) Epoch 1, batch 3300, loss[loss=0.2377, simple_loss=0.4174, pruned_loss=0.2897, over 7382.00 frames.], tot_loss[loss=0.265, simple_loss=0.4319, pruned_loss=1.921, over 1419222.53 frames.], batch size: 23, lr: 2.74e-03
2022-05-26 11:32:12,602 INFO [train.py:842] (0/4) Epoch 1, batch 3350, loss[loss=0.241, simple_loss=0.4242, pruned_loss=0.2892, over 7280.00 frames.], tot_loss[loss=0.2586, simple_loss=0.4281, pruned_loss=1.555, over 1423609.77 frames.], batch size: 24, lr: 2.73e-03
2022-05-26 11:32:51,525 INFO [train.py:842] (0/4) Epoch 1, batch 3400, loss[loss=0.2217, simple_loss=0.3933, pruned_loss=0.2504, over 7262.00 frames.], tot_loss[loss=0.2537, simple_loss=0.4258, pruned_loss=1.272, over 1424137.60 frames.], batch size: 19, lr: 2.73e-03
2022-05-26 11:33:30,151 INFO [train.py:842] (0/4) Epoch 1, batch 3450, loss[loss=0.2499, simple_loss=0.4421, pruned_loss=0.2892, over 7343.00 frames.], tot_loss[loss=0.249, simple_loss=0.4227, pruned_loss=1.049, over 1423925.42 frames.], batch size: 25, lr: 2.72e-03
2022-05-26 11:34:09,057 INFO [train.py:842] (0/4) Epoch 1, batch 3500, loss[loss=0.3113, simple_loss=0.5322, pruned_loss=0.4521, over 7183.00 frames.], tot_loss[loss=0.247, simple_loss=0.4232, pruned_loss=0.8774, over 1422489.85 frames.], batch size: 26, lr: 2.72e-03
2022-05-26 11:34:47,674 INFO [train.py:842] (0/4) Epoch 1, batch 3550, loss[loss=0.2196, simple_loss=0.3913, pruned_loss=0.2393, over 7210.00 frames.], tot_loss[loss=0.2431, simple_loss=0.42, pruned_loss=0.7383, over 1423518.80 frames.], batch size: 21, lr: 2.71e-03
2022-05-26 11:35:26,427 INFO [train.py:842] (0/4) Epoch 1, batch 3600, loss[loss=0.1914, simple_loss=0.3459, pruned_loss=0.1845, over 6998.00 frames.], tot_loss[loss=0.2402, simple_loss=0.4181, pruned_loss=0.6294, over 1422338.60 frames.], batch size: 16, lr: 2.70e-03
2022-05-26 11:36:05,092 INFO [train.py:842] (0/4) Epoch 1, batch 3650, loss[loss=0.2059, simple_loss=0.3756, pruned_loss=0.1809, over 7215.00 frames.], tot_loss[loss=0.2375, simple_loss=0.416, pruned_loss=0.5424, over 1422898.08 frames.], batch size: 21, lr: 2.70e-03
2022-05-26 11:36:43,959 INFO [train.py:842] (0/4) Epoch 1, batch 3700, loss[loss=0.2244, simple_loss=0.404, pruned_loss=0.2242, over 6711.00 frames.], tot_loss[loss=0.2355, simple_loss=0.4145, pruned_loss=0.4738, over 1427080.33 frames.], batch size: 31, lr: 2.69e-03
2022-05-26 11:37:22,481 INFO [train.py:842] (0/4) Epoch 1, batch 3750, loss[loss=0.1806, simple_loss=0.3274, pruned_loss=0.1687, over 7293.00 frames.], tot_loss[loss=0.2337, simple_loss=0.413, pruned_loss=0.4223, over 1418626.42 frames.], batch size: 18, lr: 2.68e-03
2022-05-26 11:38:01,292 INFO [train.py:842] (0/4) Epoch 1, batch 3800, loss[loss=0.205, simple_loss=0.3693, pruned_loss=0.2035, over 7133.00 frames.], tot_loss[loss=0.2338, simple_loss=0.4144, pruned_loss=0.3832, over 1417108.20 frames.], batch size: 17, lr: 2.68e-03
2022-05-26 11:38:40,179 INFO [train.py:842] (0/4) Epoch 1, batch 3850, loss[loss=0.2209, simple_loss=0.3919, pruned_loss=0.249, over 7135.00 frames.], tot_loss[loss=0.2324, simple_loss=0.4132, pruned_loss=0.3484, over 1423130.09 frames.], batch size: 17, lr: 2.67e-03
2022-05-26 11:39:18,907 INFO [train.py:842] (0/4) Epoch 1, batch 3900, loss[loss=0.1706, simple_loss=0.3109, pruned_loss=0.1509, over 6825.00 frames.], tot_loss[loss=0.2316, simple_loss=0.4129, pruned_loss=0.3227, over 1419840.09 frames.], batch size: 15, lr: 2.66e-03
2022-05-26 11:39:57,452 INFO [train.py:842] (0/4) Epoch 1, batch 3950, loss[loss=0.2746, simple_loss=0.4878, pruned_loss=0.3067, over 6802.00 frames.], tot_loss[loss=0.2303, simple_loss=0.4114, pruned_loss=0.301, over 1418700.25 frames.], batch size: 31, lr: 2.66e-03
2022-05-26 11:40:36,019 INFO [train.py:842] (0/4) Epoch 1, batch 4000, loss[loss=0.2274, simple_loss=0.4119, pruned_loss=0.2144, over 7188.00 frames.], tot_loss[loss=0.23, simple_loss=0.4118, pruned_loss=0.2841, over 1419833.67 frames.], batch size: 26, lr: 2.65e-03
2022-05-26 11:41:14,456 INFO [train.py:842] (0/4) Epoch 1, batch 4050, loss[loss=0.2715, simple_loss=0.4815, pruned_loss=0.3074, over 4814.00 frames.], tot_loss[loss=0.2285, simple_loss=0.4099, pruned_loss=0.2687, over 1421414.20 frames.], batch size: 52, lr: 2.64e-03
2022-05-26 11:41:53,376 INFO [train.py:842] (0/4) Epoch 1, batch 4100, loss[loss=0.2151, simple_loss=0.3889, pruned_loss=0.2067, over 6294.00 frames.], tot_loss[loss=0.2271, simple_loss=0.4082, pruned_loss=0.2564, over 1419160.90 frames.], batch size: 37, lr: 2.64e-03
2022-05-26 11:42:32,053 INFO [train.py:842] (0/4) Epoch 1, batch 4150, loss[loss=0.1967, simple_loss=0.3607, pruned_loss=0.1633, over 7419.00 frames.], tot_loss[loss=0.2267, simple_loss=0.4078, pruned_loss=0.2477, over 1424525.71 frames.], batch size: 20, lr: 2.63e-03
2022-05-26 11:43:10,837 INFO [train.py:842] (0/4) Epoch 1, batch 4200, loss[loss=0.2246, simple_loss=0.4075, pruned_loss=0.2082, over 7316.00 frames.], tot_loss[loss=0.2253, simple_loss=0.4059, pruned_loss=0.2389, over 1428506.06 frames.], batch size: 21, lr: 2.63e-03
2022-05-26 11:43:49,376 INFO [train.py:842] (0/4) Epoch 1, batch 4250, loss[loss=0.1998, simple_loss=0.365, pruned_loss=0.1727, over 7148.00 frames.], tot_loss[loss=0.2238, simple_loss=0.4038, pruned_loss=0.2317, over 1427654.39 frames.], batch size: 20, lr: 2.62e-03
2022-05-26 11:44:28,192 INFO [train.py:842] (0/4) Epoch 1, batch 4300, loss[loss=0.2763, simple_loss=0.4959, pruned_loss=0.283, over 7190.00 frames.], tot_loss[loss=0.2226, simple_loss=0.4019, pruned_loss=0.2259, over 1424429.29 frames.], batch size: 22, lr: 2.61e-03
2022-05-26 11:45:06,706 INFO [train.py:842] (0/4) Epoch 1, batch 4350, loss[loss=0.2067, simple_loss=0.3765, pruned_loss=0.185, over 7157.00 frames.], tot_loss[loss=0.222, simple_loss=0.4011, pruned_loss=0.2223, over 1426218.39 frames.], batch size: 19, lr: 2.61e-03
2022-05-26 11:45:45,323 INFO [train.py:842] (0/4) Epoch 1, batch 4400, loss[loss=0.2235, simple_loss=0.4071, pruned_loss=0.199, over 7223.00 frames.], tot_loss[loss=0.2234, simple_loss=0.4038, pruned_loss=0.2211, over 1426557.55 frames.], batch size: 21, lr: 2.60e-03
2022-05-26 11:46:23,886 INFO [train.py:842] (0/4) Epoch 1, batch 4450, loss[loss=0.2092, simple_loss=0.3806, pruned_loss=0.1888, over 7165.00 frames.], tot_loss[loss=0.223, simple_loss=0.4031, pruned_loss=0.2183, over 1428932.28 frames.], batch size: 19, lr: 2.59e-03
2022-05-26 11:47:02,775 INFO [train.py:842] (0/4) Epoch 1, batch 4500, loss[loss=0.218, simple_loss=0.3956, pruned_loss=0.2022, over 7255.00 frames.], tot_loss[loss=0.2226, simple_loss=0.4028, pruned_loss=0.2149, over 1431556.74 frames.], batch size: 19, lr: 2.59e-03
2022-05-26 11:47:41,422 INFO [train.py:842] (0/4) Epoch 1, batch 4550, loss[loss=0.1993, simple_loss=0.3643, pruned_loss=0.1713, over 7061.00 frames.], tot_loss[loss=0.2218, simple_loss=0.4016, pruned_loss=0.2134, over 1429692.39 frames.], batch size: 18, lr: 2.58e-03
2022-05-26 11:48:20,132 INFO [train.py:842] (0/4) Epoch 1, batch 4600, loss[loss=0.2043, simple_loss=0.3734, pruned_loss=0.1756, over 7249.00 frames.], tot_loss[loss=0.2204, simple_loss=0.3995, pruned_loss=0.2089, over 1428670.69 frames.], batch size: 19, lr: 2.57e-03
2022-05-26 11:48:58,607 INFO [train.py:842] (0/4) Epoch 1, batch 4650, loss[loss=0.2431, simple_loss=0.4383, pruned_loss=0.2396, over 7090.00 frames.], tot_loss[loss=0.2204, simple_loss=0.3997, pruned_loss=0.2071, over 1429568.51 frames.], batch size: 28, lr: 2.57e-03
2022-05-26 11:49:37,488 INFO [train.py:842] (0/4) Epoch 1, batch 4700, loss[loss=0.2024, simple_loss=0.3686, pruned_loss=0.1814, over 7288.00 frames.], tot_loss[loss=0.2201, simple_loss=0.3991, pruned_loss=0.2064, over 1428974.32 frames.], batch size: 17, lr: 2.56e-03
2022-05-26 11:50:15,957 INFO [train.py:842] (0/4) Epoch 1, batch 4750, loss[loss=0.2578, simple_loss=0.459, pruned_loss=0.283, over 5373.00 frames.], tot_loss[loss=0.2217, simple_loss=0.4019, pruned_loss=0.2081, over 1427664.88 frames.], batch size: 52, lr: 2.55e-03
2022-05-26 11:50:54,731 INFO [train.py:842] (0/4) Epoch 1, batch 4800, loss[loss=0.1903, simple_loss=0.3504, pruned_loss=0.1513, over 7434.00 frames.], tot_loss[loss=0.2203, simple_loss=0.3999, pruned_loss=0.2044, over 1429578.02 frames.], batch size: 20, lr: 2.55e-03
2022-05-26 11:51:33,198 INFO [train.py:842] (0/4) Epoch 1, batch 4850, loss[loss=0.2285, simple_loss=0.4107, pruned_loss=0.2318, over 7256.00 frames.], tot_loss[loss=0.2208, simple_loss=0.4009, pruned_loss=0.2047, over 1426592.52 frames.], batch size: 19, lr: 2.54e-03
2022-05-26 11:52:11,922 INFO [train.py:842] (0/4) Epoch 1, batch 4900, loss[loss=0.21, simple_loss=0.382, pruned_loss=0.1901, over 7321.00 frames.], tot_loss[loss=0.2201, simple_loss=0.3997, pruned_loss=0.2026, over 1428116.24 frames.], batch size: 20, lr: 2.54e-03
2022-05-26 11:52:50,297 INFO [train.py:842] (0/4) Epoch 1, batch 4950, loss[loss=0.2103, simple_loss=0.382, pruned_loss=0.1927, over 7358.00 frames.], tot_loss[loss=0.2207, simple_loss=0.4009, pruned_loss=0.2029, over 1423683.74 frames.], batch size: 19, lr: 2.53e-03
2022-05-26 11:53:29,132 INFO [train.py:842] (0/4) Epoch 1, batch 5000, loss[loss=0.2392, simple_loss=0.4331, pruned_loss=0.2268, over 7341.00 frames.], tot_loss[loss=0.2208, simple_loss=0.4012, pruned_loss=0.2023, over 1423145.91 frames.], batch size: 22, lr: 2.52e-03
2022-05-26 11:54:07,521 INFO [train.py:842] (0/4) Epoch 1, batch 5050, loss[loss=0.2574, simple_loss=0.4624, pruned_loss=0.2623, over 7324.00 frames.], tot_loss[loss=0.2196, simple_loss=0.3993, pruned_loss=0.1997, over 1422617.45 frames.], batch size: 21, lr: 2.52e-03
2022-05-26 11:54:46,104 INFO [train.py:842] (0/4) Epoch 1, batch 5100, loss[loss=0.2179, simple_loss=0.3996, pruned_loss=0.1809, over 7200.00 frames.], tot_loss[loss=0.2181, simple_loss=0.3967, pruned_loss=0.1974, over 1420917.37 frames.], batch size: 22, lr: 2.51e-03
2022-05-26 11:55:24,621 INFO [train.py:842] (0/4) Epoch 1, batch 5150, loss[loss=0.2463, simple_loss=0.4439, pruned_loss=0.2432, over 7429.00 frames.], tot_loss[loss=0.2168, simple_loss=0.3946, pruned_loss=0.1948, over 1423251.49 frames.], batch size: 20, lr: 2.50e-03
2022-05-26 11:56:03,336 INFO [train.py:842] (0/4) Epoch 1, batch 5200, loss[loss=0.2365, simple_loss=0.4296, pruned_loss=0.2166, over 7336.00 frames.], tot_loss[loss=0.2177, simple_loss=0.3962, pruned_loss=0.196, over 1421537.25 frames.], batch size: 25, lr: 2.50e-03
2022-05-26 11:56:41,767 INFO [train.py:842] (0/4) Epoch 1, batch 5250, loss[loss=0.2448, simple_loss=0.4394, pruned_loss=0.2505, over 5025.00 frames.], tot_loss[loss=0.2166, simple_loss=0.3944, pruned_loss=0.1938, over 1419366.50 frames.], batch size: 52, lr: 2.49e-03
2022-05-26 11:57:20,422 INFO [train.py:842] (0/4) Epoch 1, batch 5300, loss[loss=0.2245, simple_loss=0.4025, pruned_loss=0.2319, over 7288.00 frames.], tot_loss[loss=0.2164, simple_loss=0.3941, pruned_loss=0.1933, over 1416829.71 frames.], batch size: 17, lr: 2.49e-03
2022-05-26 11:57:58,747 INFO [train.py:842] (0/4) Epoch 1, batch 5350, loss[loss=0.2178, simple_loss=0.3972, pruned_loss=0.192, over 7385.00 frames.], tot_loss[loss=0.2162, simple_loss=0.3939, pruned_loss=0.1926, over 1414483.02 frames.], batch size: 23, lr: 2.48e-03
2022-05-26 11:58:37,511 INFO [train.py:842] (0/4) Epoch 1, batch 5400, loss[loss=0.2141, simple_loss=0.394, pruned_loss=0.1712, over 7026.00 frames.], tot_loss[loss=0.2166, simple_loss=0.3946, pruned_loss=0.193, over 1419758.10 frames.], batch size: 28, lr: 2.47e-03
2022-05-26 11:59:16,032 INFO [train.py:842] (0/4) Epoch 1, batch 5450, loss[loss=0.2257, simple_loss=0.41, pruned_loss=0.2067, over 7151.00 frames.], tot_loss[loss=0.2158, simple_loss=0.3933, pruned_loss=0.1915, over 1420793.75 frames.], batch size: 20, lr: 2.47e-03
2022-05-26 11:59:54,869 INFO [train.py:842] (0/4) Epoch 1, batch 5500, loss[loss=0.251, simple_loss=0.4516, pruned_loss=0.2516, over 5129.00 frames.], tot_loss[loss=0.2154, simple_loss=0.3927, pruned_loss=0.191, over 1420316.83 frames.], batch size: 52, lr: 2.46e-03
2022-05-26 12:00:33,672 INFO [train.py:842] (0/4) Epoch 1, batch 5550, loss[loss=0.2462, simple_loss=0.4376, pruned_loss=0.2738, over 6824.00 frames.], tot_loss[loss=0.2154, simple_loss=0.3924, pruned_loss=0.1915, over 1422188.07 frames.], batch size: 15, lr: 2.45e-03
2022-05-26 12:01:12,572 INFO [train.py:842] (0/4) Epoch 1, batch 5600, loss[loss=0.2431, simple_loss=0.4391, pruned_loss=0.235, over 6419.00 frames.], tot_loss[loss=0.2153, simple_loss=0.3925, pruned_loss=0.1906, over 1422593.45 frames.], batch size: 38, lr: 2.45e-03
2022-05-26 12:01:51,145 INFO [train.py:842] (0/4) Epoch 1, batch 5650, loss[loss=0.1951, simple_loss=0.356, pruned_loss=0.1713, over 7272.00 frames.], tot_loss[loss=0.215, simple_loss=0.3921, pruned_loss=0.1901, over 1419619.82 frames.], batch size: 17, lr: 2.44e-03
2022-05-26 12:02:29,840 INFO [train.py:842] (0/4) Epoch 1, batch 5700, loss[loss=0.1968, simple_loss=0.3624, pruned_loss=0.156, over 7422.00 frames.], tot_loss[loss=0.215, simple_loss=0.3921, pruned_loss=0.1899, over 1420183.04 frames.], batch size: 20, lr: 2.44e-03
2022-05-26 12:03:08,168 INFO [train.py:842] (0/4) Epoch 1, batch 5750, loss[loss=0.1862, simple_loss=0.3439, pruned_loss=0.143, over 7281.00 frames.], tot_loss[loss=0.2154, simple_loss=0.3928, pruned_loss=0.1902, over 1421928.55 frames.], batch size: 18, lr: 2.43e-03
2022-05-26 12:03:47,051 INFO [train.py:842] (0/4) Epoch 1, batch 5800, loss[loss=0.2041, simple_loss=0.3764, pruned_loss=0.1591, over 7209.00 frames.], tot_loss[loss=0.2149, simple_loss=0.3921, pruned_loss=0.1883, over 1427954.34 frames.], batch size: 22, lr: 2.42e-03
2022-05-26 12:04:25,394 INFO [train.py:842] (0/4) Epoch 1, batch 5850, loss[loss=0.2393, simple_loss=0.431, pruned_loss=0.2385, over 7425.00 frames.], tot_loss[loss=0.2154, simple_loss=0.3929, pruned_loss=0.1893, over 1426989.41 frames.], batch size: 20, lr: 2.42e-03
2022-05-26 12:05:04,399 INFO [train.py:842] (0/4) Epoch 1, batch 5900, loss[loss=0.253, simple_loss=0.4583, pruned_loss=0.2383, over 7322.00 frames.], tot_loss[loss=0.214, simple_loss=0.3906, pruned_loss=0.1873, over 1430197.25 frames.], batch size: 21, lr: 2.41e-03
2022-05-26 12:05:43,139 INFO [train.py:842] (0/4) Epoch 1, batch 5950, loss[loss=0.2184, simple_loss=0.397, pruned_loss=0.1993, over 7168.00 frames.], tot_loss[loss=0.2134, simple_loss=0.3895, pruned_loss=0.1861, over 1430975.40 frames.], batch size: 19, lr: 2.41e-03
2022-05-26 12:06:21,960 INFO [train.py:842] (0/4) Epoch 1, batch 6000, loss[loss=0.4581, simple_loss=0.4171, pruned_loss=0.2495, over 7155.00 frames.], tot_loss[loss=0.2149, simple_loss=0.3888, pruned_loss=0.185, over 1427705.57 frames.], batch size: 26, lr: 2.40e-03
2022-05-26 12:06:21,961 INFO [train.py:862] (0/4) Computing validation loss
2022-05-26 12:06:31,824 INFO [train.py:871] (0/4) Epoch 1, validation: loss=0.2892, simple_loss=0.3436, pruned_loss=0.1174, over 868885.00 frames.
2022-05-26 12:07:10,558 INFO [train.py:842] (0/4) Epoch 1, batch 6050, loss[loss=0.3636, simple_loss=0.3634, pruned_loss=0.1819, over 7266.00 frames.], tot_loss[loss=0.259, simple_loss=0.3925, pruned_loss=0.191, over 1424958.80 frames.], batch size: 16, lr: 2.39e-03
2022-05-26 12:07:49,761 INFO [train.py:842] (0/4) Epoch 1, batch 6100, loss[loss=0.3233, simple_loss=0.3497, pruned_loss=0.1484, over 6806.00 frames.], tot_loss[loss=0.2839, simple_loss=0.3904, pruned_loss=0.1884, over 1427153.91 frames.], batch size: 15, lr: 2.39e-03
2022-05-26 12:08:28,487 INFO [train.py:842] (0/4) Epoch 1, batch 6150, loss[loss=0.3645, simple_loss=0.3782, pruned_loss=0.1754, over 7116.00 frames.], tot_loss[loss=0.3059, simple_loss=0.3913, pruned_loss=0.1878, over 1427397.82 frames.], batch size: 21, lr: 2.38e-03
2022-05-26 12:09:07,159 INFO [train.py:842] (0/4) Epoch 1, batch 6200, loss[loss=0.4063, simple_loss=0.406, pruned_loss=0.2033, over 7342.00 frames.], tot_loss[loss=0.3204, simple_loss=0.391, pruned_loss=0.1852, over 1427919.20 frames.], batch size: 22, lr: 2.38e-03
2022-05-26 12:09:45,799 INFO [train.py:842] (0/4) Epoch 1, batch 6250, loss[loss=0.4259, simple_loss=0.4276, pruned_loss=0.2121, over 7362.00 frames.], tot_loss[loss=0.333, simple_loss=0.3915, pruned_loss=0.1842, over 1429875.68 frames.], batch size: 23, lr: 2.37e-03
2022-05-26 12:10:25,213 INFO [train.py:842] (0/4) Epoch 1, batch 6300, loss[loss=0.4287, simple_loss=0.4311, pruned_loss=0.2132, over 7281.00 frames.], tot_loss[loss=0.3422, simple_loss=0.3912, pruned_loss=0.1831, over 1428116.34 frames.], batch size: 18, lr: 2.37e-03
2022-05-26 12:11:03,817 INFO [train.py:842] (0/4) Epoch 1, batch 6350, loss[loss=0.3284, simple_loss=0.3611, pruned_loss=0.1478, over 7146.00 frames.], tot_loss[loss=0.3468, simple_loss=0.3896, pruned_loss=0.1804, over 1427807.12 frames.], batch size: 20, lr: 2.36e-03
2022-05-26 12:11:42,701 INFO [train.py:842] (0/4) Epoch 1, batch 6400, loss[loss=0.3665, simple_loss=0.3901, pruned_loss=0.1714, over 7363.00 frames.], tot_loss[loss=0.3538, simple_loss=0.3902, pruned_loss=0.1808, over 1427647.66 frames.], batch size: 19, lr: 2.35e-03
2022-05-26 12:12:21,091 INFO [train.py:842] (0/4) Epoch 1, batch 6450, loss[loss=0.4298, simple_loss=0.4262, pruned_loss=0.2167, over 7125.00 frames.], tot_loss[loss=0.3597, simple_loss=0.3918, pruned_loss=0.181, over 1427702.78 frames.], batch size: 21, lr: 2.35e-03
2022-05-26 12:12:59,893 INFO [train.py:842] (0/4) Epoch 1, batch 6500, loss[loss=0.3362, simple_loss=0.3592, pruned_loss=0.1566, over 7138.00 frames.], tot_loss[loss=0.3601, simple_loss=0.3902, pruned_loss=0.1785, over 1423019.08 frames.], batch size: 17, lr: 2.34e-03
2022-05-26 12:13:38,142 INFO [train.py:842] (0/4) Epoch 1, batch 6550, loss[loss=0.4067, simple_loss=0.4215, pruned_loss=0.1959, over 7316.00 frames.], tot_loss[loss=0.3612, simple_loss=0.3896, pruned_loss=0.1769, over 1418883.33 frames.], batch size: 21, lr: 2.34e-03
2022-05-26 12:14:17,053 INFO [train.py:842] (0/4) Epoch 1, batch 6600, loss[loss=0.3335, simple_loss=0.3766, pruned_loss=0.1452, over 7190.00 frames.], tot_loss[loss=0.3639, simple_loss=0.3905, pruned_loss=0.1768, over 1423693.85 frames.], batch size: 26, lr: 2.33e-03
2022-05-26 12:14:55,785 INFO [train.py:842] (0/4) Epoch 1, batch 6650, loss[loss=0.3711, simple_loss=0.3856, pruned_loss=0.1783, over 7063.00 frames.], tot_loss[loss=0.3669, simple_loss=0.3916, pruned_loss=0.1774, over 1421950.43 frames.], batch size: 18, lr: 2.33e-03
2022-05-26 12:15:34,734 INFO [train.py:842] (0/4) Epoch 1, batch 6700, loss[loss=0.584, simple_loss=0.5194, pruned_loss=0.3242, over 4581.00 frames.], tot_loss[loss=0.3671, simple_loss=0.3917, pruned_loss=0.1762, over 1422581.71 frames.], batch size: 52, lr: 2.32e-03
2022-05-26 12:16:13,264 INFO [train.py:842] (0/4) Epoch 1, batch 6750, loss[loss=0.3301, simple_loss=0.3679, pruned_loss=0.1462, over 7283.00 frames.], tot_loss[loss=0.3655, simple_loss=0.3902, pruned_loss=0.1742, over 1426583.31 frames.], batch size: 24, lr: 2.31e-03
2022-05-26 12:16:52,110 INFO [train.py:842] (0/4) Epoch 1, batch 6800, loss[loss=0.3536, simple_loss=0.3706, pruned_loss=0.1683, over 7431.00 frames.], tot_loss[loss=0.3635, simple_loss=0.3889, pruned_loss=0.172, over 1428508.65 frames.], batch size: 20, lr: 2.31e-03
2022-05-26 12:17:30,583 INFO [train.py:842] (0/4) Epoch 1, batch 6850, loss[loss=0.3112, simple_loss=0.3677, pruned_loss=0.1273, over 7196.00 frames.], tot_loss[loss=0.3627, simple_loss=0.3886, pruned_loss=0.1707, over 1426717.85 frames.], batch size: 23, lr: 2.30e-03
2022-05-26 12:18:19,073 INFO [train.py:842] (0/4) Epoch 1, batch 6900, loss[loss=0.3578, simple_loss=0.387, pruned_loss=0.1643, over 7414.00 frames.], tot_loss[loss=0.3603, simple_loss=0.3865, pruned_loss=0.1688, over 1427895.41 frames.], batch size: 21, lr: 2.30e-03
2022-05-26 12:18:57,536 INFO [train.py:842] (0/4) Epoch 1, batch 6950, loss[loss=0.3328, simple_loss=0.3532, pruned_loss=0.1562, over 7271.00 frames.], tot_loss[loss=0.3618, simple_loss=0.3874, pruned_loss=0.1695, over 1422748.48 frames.], batch size: 18, lr: 2.29e-03
2022-05-26 12:19:36,268 INFO [train.py:842] (0/4) Epoch 1, batch 7000, loss[loss=0.3356, simple_loss=0.3666, pruned_loss=0.1523, over 7153.00 frames.], tot_loss[loss=0.3642, simple_loss=0.3892, pruned_loss=0.1707, over 1420545.91 frames.], batch size: 18, lr: 2.29e-03
2022-05-26 12:20:14,735 INFO [train.py:842] (0/4) Epoch 1, batch 7050, loss[loss=0.3505, simple_loss=0.3774, pruned_loss=0.1618, over 7166.00 frames.], tot_loss[loss=0.3618, simple_loss=0.3877, pruned_loss=0.1688, over 1421803.53 frames.], batch size: 19, lr: 2.28e-03
2022-05-26 12:20:53,601 INFO [train.py:842] (0/4) Epoch 1, batch 7100, loss[loss=0.3123, simple_loss=0.3704, pruned_loss=0.1271, over 7342.00 frames.], tot_loss[loss=0.3619, simple_loss=0.3878, pruned_loss=0.1687, over 1423981.86 frames.], batch size: 22, lr: 2.28e-03
2022-05-26 12:21:32,657 INFO [train.py:842] (0/4) Epoch 1, batch 7150, loss[loss=0.3883, simple_loss=0.4022, pruned_loss=0.1872, over 7204.00 frames.], tot_loss[loss=0.3611, simple_loss=0.3867, pruned_loss=0.1683, over 1418171.13 frames.], batch size: 22, lr: 2.27e-03
2022-05-26 12:22:11,492 INFO [train.py:842] (0/4) Epoch 1, batch 7200, loss[loss=0.355, simple_loss=0.3961, pruned_loss=0.1569, over 7346.00 frames.], tot_loss[loss=0.36, simple_loss=0.3869, pruned_loss=0.1669, over 1420770.40 frames.], batch size: 22, lr: 2.27e-03
2022-05-26 12:22:50,077 INFO [train.py:842] (0/4) Epoch 1, batch 7250, loss[loss=0.3941, simple_loss=0.4047, pruned_loss=0.1918, over 7061.00 frames.], tot_loss[loss=0.3622, simple_loss=0.3884, pruned_loss=0.1683, over 1416062.40 frames.], batch size: 18, lr: 2.26e-03
2022-05-26 12:23:28,695 INFO [train.py:842] (0/4) Epoch 1, batch 7300, loss[loss=0.3737, simple_loss=0.4098, pruned_loss=0.1688, over 7067.00 frames.], tot_loss[loss=0.3627, simple_loss=0.3892, pruned_loss=0.1684, over 1416594.02 frames.], batch size: 28, lr: 2.26e-03
2022-05-26 12:24:07,152 INFO [train.py:842] (0/4) Epoch 1, batch 7350, loss[loss=0.31, simple_loss=0.3418, pruned_loss=0.1391, over 7218.00 frames.], tot_loss[loss=0.3606, simple_loss=0.3879, pruned_loss=0.1668, over 1416611.07 frames.], batch size: 16, lr: 2.25e-03
2022-05-26 12:24:45,812 INFO [train.py:842] (0/4) Epoch 1, batch 7400, loss[loss=0.2582, simple_loss=0.3105, pruned_loss=0.1029, over 7402.00 frames.], tot_loss[loss=0.3612, simple_loss=0.3885, pruned_loss=0.1671, over 1417748.21 frames.], batch size: 18, lr: 2.24e-03
2022-05-26 12:25:24,571 INFO [train.py:842] (0/4) Epoch 1, batch 7450, loss[loss=0.3592, simple_loss=0.3845, pruned_loss=0.167, over 7397.00 frames.], tot_loss[loss=0.3629, simple_loss=0.3904, pruned_loss=0.1679, over 1425460.15 frames.], batch size: 18, lr: 2.24e-03
2022-05-26 12:26:03,336 INFO [train.py:842] (0/4) Epoch 1, batch 7500, loss[loss=0.4152, simple_loss=0.4213, pruned_loss=0.2046, over 7432.00 frames.], tot_loss[loss=0.3665, simple_loss=0.3924, pruned_loss=0.1704, over 1422886.18 frames.], batch size: 20, lr: 2.23e-03
2022-05-26 12:26:41,950 INFO [train.py:842] (0/4) Epoch 1, batch 7550, loss[loss=0.3082, simple_loss=0.3492, pruned_loss=0.1337, over 7329.00 frames.], tot_loss[loss=0.3624, simple_loss=0.3894, pruned_loss=0.1677, over 1420722.91 frames.], batch size: 20, lr: 2.23e-03
2022-05-26 12:27:21,041 INFO [train.py:842] (0/4) Epoch 1, batch 7600, loss[loss=0.3289, simple_loss=0.3675, pruned_loss=0.1452, over 7421.00 frames.], tot_loss[loss=0.3578, simple_loss=0.3865, pruned_loss=0.1646, over 1424353.22 frames.], batch size: 21, lr: 2.22e-03
2022-05-26 12:28:28,318 INFO [train.py:842] (0/4) Epoch 1, batch 7650, loss[loss=0.4288, simple_loss=0.4249, pruned_loss=0.2164, over 7333.00 frames.], tot_loss[loss=0.3594, simple_loss=0.3879, pruned_loss=0.1655, over 1427571.25 frames.], batch size: 20, lr: 2.22e-03
2022-05-26 12:29:07,133 INFO [train.py:842] (0/4) Epoch 1, batch 7700, loss[loss=0.365, simple_loss=0.3945, pruned_loss=0.1677, over 7239.00 frames.], tot_loss[loss=0.3602, simple_loss=0.3882, pruned_loss=0.1662, over 1424765.81 frames.], batch size: 20, lr: 2.21e-03
2022-05-26 12:29:46,009 INFO [train.py:842] (0/4) Epoch 1, batch 7750, loss[loss=0.3356, simple_loss=0.3649, pruned_loss=0.1532, over 7360.00 frames.], tot_loss[loss=0.3577, simple_loss=0.3867, pruned_loss=0.1644, over 1425837.59 frames.], batch size: 19, lr: 2.21e-03
2022-05-26 12:30:24,812 INFO [train.py:842] (0/4) Epoch 1, batch 7800, loss[loss=0.4515, simple_loss=0.4411, pruned_loss=0.2309, over 7071.00 frames.], tot_loss[loss=0.3597, simple_loss=0.3878, pruned_loss=0.1658, over 1427722.09 frames.], batch size: 28, lr: 2.20e-03
2022-05-26 12:31:03,146 INFO [train.py:842] (0/4) Epoch 1, batch 7850, loss[loss=0.3404, simple_loss=0.3911, pruned_loss=0.1448, over 7289.00 frames.], tot_loss[loss=0.3585, simple_loss=0.3874, pruned_loss=0.1648, over 1429709.76 frames.], batch size: 24, lr: 2.20e-03
2022-05-26 12:31:41,887 INFO [train.py:842] (0/4) Epoch 1, batch 7900, loss[loss=0.4451, simple_loss=0.4374, pruned_loss=0.2264, over 7433.00 frames.], tot_loss[loss=0.3576, simple_loss=0.3868, pruned_loss=0.1642, over 1427764.73 frames.], batch size: 20, lr: 2.19e-03
2022-05-26 12:32:20,417 INFO [train.py:842] (0/4) Epoch 1, batch 7950, loss[loss=0.3926, simple_loss=0.4128, pruned_loss=0.1863, over 6569.00 frames.], tot_loss[loss=0.3547, simple_loss=0.3848, pruned_loss=0.1623, over 1422853.58 frames.], batch size: 38, lr: 2.19e-03
2022-05-26 12:32:58,422 INFO [checkpoint.py:75] (0/4) Saving checkpoint to streaming_pruned_transducer_stateless4/exp/checkpoint-8000.pt
2022-05-26 12:33:01,920 INFO [train.py:842] (0/4) Epoch 1, batch 8000, loss[loss=0.3569, simple_loss=0.3671, pruned_loss=0.1733, over 7144.00 frames.], tot_loss[loss=0.3537, simple_loss=0.3843, pruned_loss=0.1616, over 1425270.20 frames.], batch size: 17, lr: 2.18e-03
2022-05-26 12:33:40,591 INFO [train.py:842] (0/4) Epoch 1, batch 8050, loss[loss=0.2441, simple_loss=0.2852, pruned_loss=0.1015, over 7146.00 frames.], tot_loss[loss=0.3512, simple_loss=0.3828, pruned_loss=0.1597, over 1428906.56 frames.], batch size: 17, lr: 2.18e-03
2022-05-26 12:34:19,298 INFO [train.py:842] (0/4) Epoch 1, batch 8100, loss[loss=0.4043, simple_loss=0.4158, pruned_loss=0.1964, over 7257.00 frames.], tot_loss[loss=0.3519, simple_loss=0.3835, pruned_loss=0.1601, over 1428116.30 frames.], batch size: 19, lr: 2.17e-03
2022-05-26 12:34:57,730 INFO [train.py:842] (0/4) Epoch 1, batch 8150, loss[loss=0.3794, simple_loss=0.401, pruned_loss=0.1789, over 7212.00 frames.], tot_loss[loss=0.3553, simple_loss=0.3858, pruned_loss=0.1624, over 1423228.70 frames.], batch size: 22, lr: 2.17e-03
2022-05-26 12:35:36,455 INFO [train.py:842] (0/4) Epoch 1, batch 8200, loss[loss=0.3918, simple_loss=0.4046, pruned_loss=0.1895, over 7172.00 frames.], tot_loss[loss=0.3524, simple_loss=0.384, pruned_loss=0.1604, over 1421450.34 frames.], batch size: 18, lr: 2.16e-03
2022-05-26 12:36:15,285 INFO [train.py:842] (0/4) Epoch 1, batch 8250, loss[loss=0.2808, simple_loss=0.3205, pruned_loss=0.1205, over 7247.00 frames.], tot_loss[loss=0.3516, simple_loss=0.3828, pruned_loss=0.1602, over 1421870.53 frames.], batch size: 19, lr: 2.16e-03
2022-05-26 12:36:53,993 INFO [train.py:842] (0/4) Epoch 1, batch 8300, loss[loss=0.3731, simple_loss=0.4047, pruned_loss=0.1707, over 6711.00 frames.], tot_loss[loss=0.3511, simple_loss=0.3829, pruned_loss=0.1596, over 1421967.79 frames.], batch size: 31, lr: 2.15e-03
2022-05-26 12:37:32,673 INFO [train.py:842] (0/4) Epoch 1, batch 8350, loss[loss=0.3089, simple_loss=0.347, pruned_loss=0.1354, over 7276.00 frames.], tot_loss[loss=0.3493, simple_loss=0.3818, pruned_loss=0.1584, over 1424472.66 frames.], batch size: 18, lr: 2.15e-03
2022-05-26 12:38:11,573 INFO [train.py:842] (0/4) Epoch 1, batch 8400, loss[loss=0.4013, simple_loss=0.4311, pruned_loss=0.1858, over 7301.00 frames.], tot_loss[loss=0.3484, simple_loss=0.3819, pruned_loss=0.1574, over 1424227.82 frames.], batch size: 25, lr: 2.15e-03
2022-05-26 12:38:49,920 INFO [train.py:842] (0/4) Epoch 1, batch 8450, loss[loss=0.3543, simple_loss=0.386, pruned_loss=0.1613, over 7116.00 frames.], tot_loss[loss=0.348, simple_loss=0.3819, pruned_loss=0.1571, over 1423616.38 frames.], batch size: 21, lr: 2.14e-03
2022-05-26 12:39:28,659 INFO [train.py:842] (0/4) Epoch 1, batch 8500, loss[loss=0.3258, simple_loss=0.3701, pruned_loss=0.1408, over 7148.00 frames.], tot_loss[loss=0.3477, simple_loss=0.3813, pruned_loss=0.157, over 1423192.48 frames.], batch size: 20, lr: 2.14e-03
2022-05-26 12:40:07,535 INFO [train.py:842] (0/4) Epoch 1, batch 8550, loss[loss=0.3518, simple_loss=0.3732, pruned_loss=0.1652, over 7163.00 frames.], tot_loss[loss=0.3461, simple_loss=0.3801, pruned_loss=0.1561, over 1424251.33 frames.], batch size: 18, lr: 2.13e-03
2022-05-26 12:40:46,238 INFO [train.py:842] (0/4) Epoch 1, batch 8600, loss[loss=0.3705, simple_loss=0.4002, pruned_loss=0.1704, over 7057.00 frames.], tot_loss[loss=0.3484, simple_loss=0.3815, pruned_loss=0.1576, over 1420786.40 frames.], batch size: 18, lr: 2.13e-03
2022-05-26 12:41:24,634 INFO [train.py:842] (0/4) Epoch 1, batch 8650, loss[loss=0.3265, simple_loss=0.3715, pruned_loss=0.1407, over 7312.00 frames.], tot_loss[loss=0.3497, simple_loss=0.383, pruned_loss=0.1582, over 1414157.61 frames.], batch size: 21, lr: 2.12e-03
2022-05-26 12:42:03,383 INFO [train.py:842] (0/4) Epoch 1, batch 8700, loss[loss=0.3132, simple_loss=0.3447, pruned_loss=0.1408, over 7147.00 frames.], tot_loss[loss=0.3506, simple_loss=0.384, pruned_loss=0.1587, over 1411332.89 frames.], batch size: 17, lr: 2.12e-03
2022-05-26 12:42:41,796 INFO [train.py:842] (0/4) Epoch 1, batch 8750, loss[loss=0.3504, simple_loss=0.3974, pruned_loss=0.1517, over 6789.00 frames.], tot_loss[loss=0.3529, simple_loss=0.3855, pruned_loss=0.1601, over 1412986.52 frames.], batch size: 31, lr: 2.11e-03
2022-05-26 12:43:20,372 INFO [train.py:842] (0/4) Epoch 1, batch 8800, loss[loss=0.2974, simple_loss=0.3495, pruned_loss=0.1227, over 6741.00 frames.], tot_loss[loss=0.3494, simple_loss=0.3836, pruned_loss=0.1576, over 1416529.47 frames.], batch size: 31, lr: 2.11e-03
2022-05-26 12:43:58,596 INFO [train.py:842] (0/4) Epoch 1, batch 8850, loss[loss=0.457, simple_loss=0.4651, pruned_loss=0.2244, over 4962.00 frames.], tot_loss[loss=0.3495, simple_loss=0.3841, pruned_loss=0.1575, over 1412717.32 frames.], batch size: 52, lr: 2.10e-03
2022-05-26 12:44:37,288 INFO [train.py:842] (0/4) Epoch 1, batch 8900, loss[loss=0.3112, simple_loss=0.3505, pruned_loss=0.136, over 6969.00 frames.], tot_loss[loss=0.3501, simple_loss=0.384, pruned_loss=0.158, over 1403945.36 frames.], batch size: 16, lr: 2.10e-03
2022-05-26 12:45:15,601 INFO [train.py:842] (0/4) Epoch 1, batch 8950, loss[loss=0.3336, simple_loss=0.3927, pruned_loss=0.1373, over 7313.00 frames.], tot_loss[loss=0.3492, simple_loss=0.3842, pruned_loss=0.1572, over 1406208.79 frames.], batch size: 21, lr: 2.10e-03
2022-05-26 12:45:54,292 INFO [train.py:842] (0/4) Epoch 1, batch 9000, loss[loss=0.4914, simple_loss=0.4665, pruned_loss=0.2581, over 5027.00 frames.], tot_loss[loss=0.3517, simple_loss=0.3867, pruned_loss=0.1584, over 1399371.88 frames.], batch size: 53, lr: 2.09e-03
2022-05-26 12:45:54,294 INFO [train.py:862] (0/4) Computing validation loss
2022-05-26 12:46:03,568 INFO [train.py:871] (0/4) Epoch 1, validation: loss=0.2508, simple_loss=0.3369, pruned_loss=0.08236, over 868885.00 frames.
2022-05-26 12:46:41,376 INFO [train.py:842] (0/4) Epoch 1, batch 9050, loss[loss=0.5309, simple_loss=0.4847, pruned_loss=0.2886, over 5113.00 frames.], tot_loss[loss=0.3535, simple_loss=0.3881, pruned_loss=0.1595, over 1387518.43 frames.], batch size: 55, lr: 2.09e-03
2022-05-26 12:47:18,742 INFO [train.py:842] (0/4) Epoch 1, batch 9100, loss[loss=0.3806, simple_loss=0.4064, pruned_loss=0.1774, over 5177.00 frames.], tot_loss[loss=0.3589, simple_loss=0.3919, pruned_loss=0.163, over 1343375.13 frames.], batch size: 54, lr: 2.08e-03
2022-05-26 12:47:56,211 INFO [train.py:842] (0/4) Epoch 1, batch 9150, loss[loss=0.4318, simple_loss=0.4349, pruned_loss=0.2143, over 5318.00 frames.], tot_loss[loss=0.3643, simple_loss=0.395, pruned_loss=0.1668, over 1286627.94 frames.], batch size: 52, lr: 2.08e-03
2022-05-26 12:48:28,770 INFO [checkpoint.py:75] (0/4) Saving checkpoint to streaming_pruned_transducer_stateless4/exp/epoch-1.pt
2022-05-26 12:48:47,801 INFO [train.py:842] (0/4) Epoch 2, batch 0, loss[loss=0.4658, simple_loss=0.4807, pruned_loss=0.2255, over 7160.00 frames.], tot_loss[loss=0.4658, simple_loss=0.4807, pruned_loss=0.2255, over 7160.00 frames.], batch size: 26, lr: 2.06e-03
2022-05-26 12:49:27,354 INFO [train.py:842] (0/4) Epoch 2, batch 50, loss[loss=0.3564, simple_loss=0.3933, pruned_loss=0.1598, over 7242.00 frames.], tot_loss[loss=0.3464, simple_loss=0.3825, pruned_loss=0.1551, over 310714.08 frames.], batch size: 20, lr: 2.06e-03
2022-05-26 12:50:06,215 INFO [train.py:842] (0/4) Epoch 2, batch 100, loss[loss=0.383, simple_loss=0.4053, pruned_loss=0.1803, over 7425.00 frames.], tot_loss[loss=0.3453, simple_loss=0.3805, pruned_loss=0.1551, over 559080.38 frames.], batch size: 20, lr: 2.05e-03
2022-05-26 12:50:45,170 INFO [train.py:842] (0/4) Epoch 2, batch 150, loss[loss=0.3128, simple_loss=0.3582, pruned_loss=0.1337, over 7320.00 frames.], tot_loss[loss=0.3414, simple_loss=0.3784, pruned_loss=0.1522, over 750064.73 frames.], batch size: 20, lr: 2.05e-03
2022-05-26 12:51:23,759 INFO [train.py:842] (0/4) Epoch 2, batch 200, loss[loss=0.3928, simple_loss=0.405, pruned_loss=0.1903, over 7163.00 frames.], tot_loss[loss=0.3397, simple_loss=0.3769, pruned_loss=0.1513, over 899600.05 frames.], batch size: 19, lr: 2.04e-03
2022-05-26 12:52:03,051 INFO [train.py:842] (0/4) Epoch 2, batch 250, loss[loss=0.4019, simple_loss=0.4243, pruned_loss=0.1898, over 7390.00 frames.], tot_loss[loss=0.3427, simple_loss=0.3787, pruned_loss=0.1534, over 1014665.10 frames.], batch size: 23, lr: 2.04e-03
2022-05-26 12:52:42,055 INFO [train.py:842] (0/4) Epoch 2, batch 300, loss[loss=0.2916, simple_loss=0.3453, pruned_loss=0.1189, over 7259.00 frames.], tot_loss[loss=0.3432, simple_loss=0.3802, pruned_loss=0.1531, over 1103630.59 frames.], batch size: 19, lr: 2.03e-03
2022-05-26 12:53:21,155 INFO [train.py:842] (0/4) Epoch 2, batch 350, loss[loss=0.3706, simple_loss=0.3959, pruned_loss=0.1726, over 7225.00 frames.], tot_loss[loss=0.337, simple_loss=0.3757, pruned_loss=0.1492, over 1172735.56 frames.], batch size: 21, lr: 2.03e-03
2022-05-26 12:53:59,748 INFO [train.py:842] (0/4) Epoch 2, batch 400, loss[loss=0.3729, simple_loss=0.3936, pruned_loss=0.176, over 7135.00 frames.], tot_loss[loss=0.3362, simple_loss=0.3751, pruned_loss=0.1487, over 1229478.66 frames.], batch size: 20, lr: 2.03e-03
2022-05-26 12:54:38,394 INFO [train.py:842] (0/4) Epoch 2, batch 450, loss[loss=0.3196, simple_loss=0.367, pruned_loss=0.1361, over 7158.00 frames.], tot_loss[loss=0.3372, simple_loss=0.376, pruned_loss=0.1492, over 1274709.63 frames.], batch size: 19, lr: 2.02e-03
2022-05-26 12:55:16,799 INFO [train.py:842] (0/4) Epoch 2, batch 500, loss[loss=0.2699, simple_loss=0.325, pruned_loss=0.1073, over 7170.00 frames.], tot_loss[loss=0.3354, simple_loss=0.3746, pruned_loss=0.1481, over 1306170.00 frames.], batch size: 18, lr: 2.02e-03
2022-05-26 12:55:56,046 INFO [train.py:842] (0/4) Epoch 2, batch 550, loss[loss=0.328, simple_loss=0.3634, pruned_loss=0.1463, over 7362.00 frames.], tot_loss[loss=0.3369, simple_loss=0.3756, pruned_loss=0.1491, over 1332006.51 frames.], batch size: 19, lr: 2.01e-03
2022-05-26 12:56:34,277 INFO [train.py:842] (0/4) Epoch 2, batch 600, loss[loss=0.3411, simple_loss=0.3727, pruned_loss=0.1547, over 7370.00 frames.], tot_loss[loss=0.339, simple_loss=0.3775, pruned_loss=0.1503, over 1353927.69 frames.], batch size: 23, lr: 2.01e-03
2022-05-26 12:57:13,120 INFO [train.py:842] (0/4) Epoch 2, batch 650, loss[loss=0.2615, simple_loss=0.3213, pruned_loss=0.1009, over 7272.00 frames.], tot_loss[loss=0.3353, simple_loss=0.375, pruned_loss=0.1477, over 1368032.79 frames.], batch size: 18, lr: 2.01e-03
2022-05-26 12:57:51,836 INFO [train.py:842] (0/4) Epoch 2, batch 700, loss[loss=0.4018, simple_loss=0.4141, pruned_loss=0.1948, over 5187.00 frames.], tot_loss[loss=0.3312, simple_loss=0.3724, pruned_loss=0.145, over 1379724.32 frames.], batch size: 52, lr: 2.00e-03
2022-05-26 12:58:30,909 INFO [train.py:842] (0/4) Epoch 2, batch 750, loss[loss=0.3319, simple_loss=0.3616, pruned_loss=0.1511, over 7256.00 frames.], tot_loss[loss=0.3337, simple_loss=0.3742, pruned_loss=0.1466, over 1390727.13 frames.], batch size: 19, lr: 2.00e-03
2022-05-26 12:59:09,560 INFO [train.py:842] (0/4) Epoch 2, batch 800, loss[loss=0.2663, simple_loss=0.3276, pruned_loss=0.1025, over 7065.00 frames.], tot_loss[loss=0.3325, simple_loss=0.3734, pruned_loss=0.1458, over 1400198.93 frames.], batch size: 18, lr: 1.99e-03
2022-05-26 12:59:48,503 INFO [train.py:842] (0/4) Epoch 2, batch 850, loss[loss=0.3308, simple_loss=0.3606, pruned_loss=0.1505, over 7331.00 frames.], tot_loss[loss=0.3321, simple_loss=0.3728, pruned_loss=0.1457, over 1408694.74 frames.], batch size: 20, lr: 1.99e-03
2022-05-26 13:00:27,172 INFO [train.py:842] (0/4) Epoch 2, batch 900, loss[loss=0.3587, simple_loss=0.3874, pruned_loss=0.165, over 7441.00 frames.], tot_loss[loss=0.3336, simple_loss=0.374, pruned_loss=0.1466, over 1413752.33 frames.], batch size: 20, lr: 1.99e-03
2022-05-26 13:01:06,368 INFO [train.py:842] (0/4) Epoch 2, batch 950, loss[loss=0.3038, simple_loss=0.3387, pruned_loss=0.1345, over 7253.00 frames.], tot_loss[loss=0.3334, simple_loss=0.3739, pruned_loss=0.1464, over 1415166.48 frames.], batch size: 19, lr: 1.98e-03
2022-05-26 13:01:45,059 INFO [train.py:842] (0/4) Epoch 2, batch 1000, loss[loss=0.3836, simple_loss=0.4059, pruned_loss=0.1806, over 6805.00 frames.], tot_loss[loss=0.3306, simple_loss=0.372, pruned_loss=0.1446, over 1416430.60 frames.], batch size: 31, lr: 1.98e-03
2022-05-26 13:02:24,223 INFO [train.py:842] (0/4) Epoch 2, batch 1050, loss[loss=0.302, simple_loss=0.3415, pruned_loss=0.1312, over 7429.00 frames.], tot_loss[loss=0.3325, simple_loss=0.373, pruned_loss=0.146, over 1418576.63 frames.], batch size: 20, lr: 1.97e-03
2022-05-26 13:03:02,483 INFO [train.py:842] (0/4) Epoch 2, batch 1100, loss[loss=0.3411, simple_loss=0.3672, pruned_loss=0.1575, over 7153.00 frames.], tot_loss[loss=0.336, simple_loss=0.3756, pruned_loss=0.1482, over 1419396.12 frames.], batch size: 18, lr: 1.97e-03
2022-05-26 13:03:41,564 INFO [train.py:842] (0/4) Epoch 2, batch 1150, loss[loss=0.3138, simple_loss=0.3706, pruned_loss=0.1285, over 7233.00 frames.], tot_loss[loss=0.3333, simple_loss=0.3737, pruned_loss=0.1464, over 1423282.03 frames.], batch size: 20, lr: 1.97e-03
2022-05-26 13:04:19,996 INFO [train.py:842] (0/4) Epoch 2, batch 1200, loss[loss=0.3399, simple_loss=0.394, pruned_loss=0.1429, over 7054.00 frames.], tot_loss[loss=0.3325, simple_loss=0.3734, pruned_loss=0.1458, over 1422340.50 frames.], batch size: 28, lr: 1.96e-03
2022-05-26 13:04:58,724 INFO [train.py:842] (0/4) Epoch 2, batch 1250, loss[loss=0.2558, simple_loss=0.314, pruned_loss=0.09882, over 7282.00 frames.], tot_loss[loss=0.3326, simple_loss=0.3743, pruned_loss=0.1454, over 1422723.20 frames.], batch size: 18, lr: 1.96e-03
2022-05-26 13:05:37,175 INFO [train.py:842] (0/4) Epoch 2, batch 1300, loss[loss=0.3226, simple_loss=0.3763, pruned_loss=0.1345, over 7218.00 frames.], tot_loss[loss=0.3316, simple_loss=0.3736, pruned_loss=0.1448, over 1416720.90 frames.], batch size: 21, lr: 1.95e-03
2022-05-26 13:06:15,947 INFO [train.py:842] (0/4) Epoch 2, batch 1350, loss[loss=0.2966, simple_loss=0.3213, pruned_loss=0.136, over 7264.00 frames.], tot_loss[loss=0.3329, simple_loss=0.3743, pruned_loss=0.1457, over 1420217.79 frames.], batch size: 17, lr: 1.95e-03
2022-05-26 13:06:54,264 INFO [train.py:842] (0/4) Epoch 2, batch 1400, loss[loss=0.3484, simple_loss=0.3791, pruned_loss=0.1588, over 7224.00 frames.], tot_loss[loss=0.3325, simple_loss=0.3738, pruned_loss=0.1456, over 1419234.68 frames.], batch size: 21, lr: 1.95e-03
2022-05-26 13:07:33,450 INFO [train.py:842] (0/4) Epoch 2, batch 1450, loss[loss=0.3716, simple_loss=0.4086, pruned_loss=0.1672, over 7167.00 frames.], tot_loss[loss=0.3328, simple_loss=0.374, pruned_loss=0.1458, over 1422230.83 frames.], batch size: 26, lr: 1.94e-03
2022-05-26 13:08:12,021 INFO [train.py:842] (0/4) Epoch 2, batch 1500, loss[loss=0.4094, simple_loss=0.426, pruned_loss=0.1964, over 6488.00 frames.], tot_loss[loss=0.3323, simple_loss=0.3736, pruned_loss=0.1455, over 1421195.28 frames.], batch size: 38, lr: 1.94e-03
2022-05-26 13:08:50,744 INFO [train.py:842] (0/4) Epoch 2, batch 1550, loss[loss=0.3139, simple_loss=0.3547, pruned_loss=0.1366, over 7433.00 frames.], tot_loss[loss=0.3304, simple_loss=0.3727, pruned_loss=0.1441, over 1425137.46 frames.], batch size: 20, lr: 1.94e-03
2022-05-26 13:09:29,421 INFO [train.py:842] (0/4) Epoch 2, batch 1600, loss[loss=0.3242, simple_loss=0.3617, pruned_loss=0.1434, over 7156.00 frames.], tot_loss[loss=0.3276, simple_loss=0.37, pruned_loss=0.1426, over 1424882.25 frames.], batch size: 18, lr: 1.93e-03
2022-05-26 13:10:08,283 INFO [train.py:842] (0/4) Epoch 2, batch 1650, loss[loss=0.3435, simple_loss=0.3777, pruned_loss=0.1547, over 7423.00 frames.], tot_loss[loss=0.3243, simple_loss=0.3675, pruned_loss=0.1405, over 1424865.63 frames.], batch size: 20, lr: 1.93e-03
2022-05-26 13:10:46,848 INFO [train.py:842] (0/4) Epoch 2, batch 1700, loss[loss=0.4109, simple_loss=0.4321, pruned_loss=0.1948, over 7406.00 frames.], tot_loss[loss=0.3231, simple_loss=0.3668, pruned_loss=0.1397, over 1423789.21 frames.], batch size: 21, lr: 1.92e-03
2022-05-26 13:11:25,780 INFO [train.py:842] (0/4) Epoch 2, batch 1750, loss[loss=0.2635, simple_loss=0.3118, pruned_loss=0.1076, over 7278.00 frames.], tot_loss[loss=0.3246, simple_loss=0.368, pruned_loss=0.1406, over 1423108.95 frames.], batch size: 18, lr: 1.92e-03
2022-05-26 13:12:04,211 INFO [train.py:842] (0/4) Epoch 2, batch 1800, loss[loss=0.3183, simple_loss=0.3645, pruned_loss=0.1361, over 7362.00 frames.], tot_loss[loss=0.3258, simple_loss=0.3687, pruned_loss=0.1414, over 1424568.56 frames.], batch size: 19, lr: 1.92e-03
2022-05-26 13:12:43,080 INFO [train.py:842] (0/4) Epoch 2, batch 1850, loss[loss=0.2818, simple_loss=0.3339, pruned_loss=0.1148, over 7328.00 frames.], tot_loss[loss=0.322, simple_loss=0.366, pruned_loss=0.139, over 1424261.38 frames.], batch size: 20, lr: 1.91e-03
2022-05-26 13:13:21,382 INFO [train.py:842] (0/4) Epoch 2, batch 1900, loss[loss=0.3056, simple_loss=0.3434, pruned_loss=0.1339, over 6999.00 frames.], tot_loss[loss=0.3226, simple_loss=0.3673, pruned_loss=0.1389, over 1427997.60 frames.], batch size: 16, lr: 1.91e-03
2022-05-26 13:14:00,097 INFO [train.py:842] (0/4) Epoch 2, batch 1950, loss[loss=0.2707, simple_loss=0.3131, pruned_loss=0.1142, over 7258.00 frames.], tot_loss[loss=0.326, simple_loss=0.37, pruned_loss=0.141, over 1428479.21 frames.], batch size: 18, lr: 1.91e-03
2022-05-26 13:14:38,224 INFO [train.py:842] (0/4) Epoch 2, batch 2000, loss[loss=0.3171, simple_loss=0.3763, pruned_loss=0.129, over 7118.00 frames.], tot_loss[loss=0.3269, simple_loss=0.3711, pruned_loss=0.1413, over 1421923.15 frames.], batch size: 21, lr: 1.90e-03
2022-05-26 13:15:17,084 INFO [train.py:842] (0/4) Epoch 2, batch 2050, loss[loss=0.3243, simple_loss=0.3898, pruned_loss=0.1294, over 7130.00 frames.], tot_loss[loss=0.3269, simple_loss=0.3714, pruned_loss=0.1412, over 1423475.31 frames.], batch size: 28, lr: 1.90e-03
2022-05-26 13:15:55,559 INFO [train.py:842] (0/4) Epoch 2, batch 2100, loss[loss=0.3336, simple_loss=0.3584, pruned_loss=0.1544, over 7424.00 frames.], tot_loss[loss=0.3271, simple_loss=0.3712, pruned_loss=0.1415, over 1424556.90 frames.], batch size: 18, lr: 1.90e-03
2022-05-26 13:16:34,411 INFO [train.py:842] (0/4) Epoch 2, batch 2150, loss[loss=0.3572, simple_loss=0.4056, pruned_loss=0.1544, over 7418.00 frames.], tot_loss[loss=0.3269, simple_loss=0.3706, pruned_loss=0.1416, over 1423310.72 frames.], batch size: 21, lr: 1.89e-03
2022-05-26 13:17:12,972 INFO [train.py:842] (0/4) Epoch 2, batch 2200, loss[loss=0.4682, simple_loss=0.456, pruned_loss=0.2402, over 7115.00 frames.], tot_loss[loss=0.3246, simple_loss=0.3691, pruned_loss=0.14, over 1423088.65 frames.], batch size: 21, lr: 1.89e-03
2022-05-26 13:17:52,126 INFO [train.py:842] (0/4) Epoch 2, batch 2250, loss[loss=0.2977, simple_loss=0.3657, pruned_loss=0.1149, over 7224.00 frames.], tot_loss[loss=0.3238, simple_loss=0.3683, pruned_loss=0.1396, over 1424413.58 frames.], batch size: 21, lr: 1.89e-03
2022-05-26 13:18:30,733 INFO [train.py:842] (0/4) Epoch 2, batch 2300, loss[loss=0.3716, simple_loss=0.4032, pruned_loss=0.17, over 7209.00 frames.], tot_loss[loss=0.3237, simple_loss=0.3686, pruned_loss=0.1394, over 1424778.58 frames.], batch size: 22, lr: 1.88e-03
2022-05-26 13:19:09,872 INFO [train.py:842] (0/4) Epoch 2, batch 2350, loss[loss=0.3203, simple_loss=0.363, pruned_loss=0.1388, over 7224.00 frames.], tot_loss[loss=0.3224, simple_loss=0.3675, pruned_loss=0.1386, over 1422216.39 frames.], batch size: 20, lr: 1.88e-03
2022-05-26 13:19:48,354 INFO [train.py:842] (0/4) Epoch 2, batch 2400, loss[loss=0.4103, simple_loss=0.4295, pruned_loss=0.1956, over 7311.00 frames.], tot_loss[loss=0.3238, simple_loss=0.3685, pruned_loss=0.1395, over 1422520.67 frames.], batch size: 21, lr: 1.87e-03
2022-05-26 13:20:27,512 INFO [train.py:842] (0/4) Epoch 2, batch 2450, loss[loss=0.2949, simple_loss=0.3557, pruned_loss=0.1171, over 7308.00 frames.], tot_loss[loss=0.3243, simple_loss=0.3693, pruned_loss=0.1396, over 1426702.09 frames.], batch size: 21, lr: 1.87e-03
2022-05-26 13:21:06,101 INFO [train.py:842] (0/4) Epoch 2, batch 2500, loss[loss=0.3532, simple_loss=0.3894, pruned_loss=0.1585, over 7133.00 frames.], tot_loss[loss=0.3225, simple_loss=0.3682, pruned_loss=0.1385, over 1427509.35 frames.], batch size: 26, lr: 1.87e-03
2022-05-26 13:21:44,924 INFO [train.py:842] (0/4) Epoch 2, batch 2550, loss[loss=0.3168, simple_loss=0.3426, pruned_loss=0.1455, over 7013.00 frames.], tot_loss[loss=0.3225, simple_loss=0.3684, pruned_loss=0.1383, over 1427511.94 frames.], batch size: 16, lr: 1.86e-03
2022-05-26 13:22:23,546 INFO [train.py:842] (0/4) Epoch 2, batch 2600, loss[loss=0.3287, simple_loss=0.3821, pruned_loss=0.1376, over 7180.00 frames.], tot_loss[loss=0.3215, simple_loss=0.3674, pruned_loss=0.1378, over 1429259.73 frames.], batch size: 26, lr: 1.86e-03
2022-05-26 13:23:02,363 INFO [train.py:842] (0/4) Epoch 2, batch 2650, loss[loss=0.3139, simple_loss=0.3634, pruned_loss=0.1322, over 6317.00 frames.], tot_loss[loss=0.3208, simple_loss=0.3669, pruned_loss=0.1374, over 1427139.37 frames.], batch size: 37, lr: 1.86e-03
2022-05-26 13:23:41,053 INFO [train.py:842] (0/4) Epoch 2, batch 2700, loss[loss=0.3396, simple_loss=0.3782, pruned_loss=0.1505, over 6739.00 frames.], tot_loss[loss=0.3187, simple_loss=0.3653, pruned_loss=0.136, over 1426621.94 frames.], batch size: 31, lr: 1.85e-03
2022-05-26 13:24:20,333 INFO [train.py:842] (0/4) Epoch 2, batch 2750, loss[loss=0.3678, simple_loss=0.3976, pruned_loss=0.169, over 7300.00 frames.], tot_loss[loss=0.3189, simple_loss=0.3654, pruned_loss=0.1362, over 1423536.71 frames.], batch size: 24, lr: 1.85e-03
2022-05-26 13:24:58,803 INFO [train.py:842] (0/4) Epoch 2, batch 2800, loss[loss=0.3317, simple_loss=0.3846, pruned_loss=0.1394, over 7206.00 frames.], tot_loss[loss=0.3192, simple_loss=0.3657, pruned_loss=0.1364, over 1426325.48 frames.], batch size: 23, lr: 1.85e-03
2022-05-26 13:25:37,893 INFO [train.py:842] (0/4) Epoch 2, batch 2850, loss[loss=0.3149, simple_loss=0.3628, pruned_loss=0.1335, over 7282.00 frames.], tot_loss[loss=0.3212, simple_loss=0.3667, pruned_loss=0.1378, over 1426213.49 frames.], batch size: 24, lr: 1.84e-03
2022-05-26 13:26:16,459 INFO [train.py:842] (0/4) Epoch 2, batch 2900, loss[loss=0.3949, simple_loss=0.4168, pruned_loss=0.1866, over 7241.00 frames.], tot_loss[loss=0.3234, simple_loss=0.3683, pruned_loss=0.1392, over 1420931.73 frames.], batch size: 20, lr: 1.84e-03
2022-05-26 13:26:55,514 INFO [train.py:842] (0/4) Epoch 2, batch 2950, loss[loss=0.2671, simple_loss=0.3312, pruned_loss=0.1015, over 7228.00 frames.], tot_loss[loss=0.3212, simple_loss=0.3669, pruned_loss=0.1378, over 1422018.71 frames.], batch size: 20, lr: 1.84e-03
2022-05-26 13:27:34,165 INFO [train.py:842] (0/4) Epoch 2, batch 3000, loss[loss=0.2783, simple_loss=0.3305, pruned_loss=0.1131, over 7281.00 frames.], tot_loss[loss=0.3184, simple_loss=0.365, pruned_loss=0.1359, over 1425366.07 frames.], batch size: 17, lr: 1.84e-03
2022-05-26 13:27:34,166 INFO [train.py:862] (0/4) Computing validation loss
2022-05-26 13:27:43,282 INFO [train.py:871] (0/4) Epoch 2, validation: loss=0.238, simple_loss=0.3286, pruned_loss=0.07369, over 868885.00 frames.
2022-05-26 13:28:22,611 INFO [train.py:842] (0/4) Epoch 2, batch 3050, loss[loss=0.2693, simple_loss=0.3169, pruned_loss=0.1108, over 7277.00 frames.], tot_loss[loss=0.3208, simple_loss=0.3665, pruned_loss=0.1375, over 1421208.21 frames.], batch size: 18, lr: 1.83e-03
2022-05-26 13:29:00,967 INFO [train.py:842] (0/4) Epoch 2, batch 3100, loss[loss=0.4686, simple_loss=0.4725, pruned_loss=0.2323, over 5364.00 frames.], tot_loss[loss=0.3206, simple_loss=0.3665, pruned_loss=0.1374, over 1421456.21 frames.], batch size: 52, lr: 1.83e-03
2022-05-26 13:29:39,782 INFO [train.py:842] (0/4) Epoch 2, batch 3150, loss[loss=0.2661, simple_loss=0.3155, pruned_loss=0.1083, over 6788.00 frames.], tot_loss[loss=0.32, simple_loss=0.3659, pruned_loss=0.137, over 1423299.55 frames.], batch size: 15, lr: 1.83e-03
2022-05-26 13:30:18,106 INFO [train.py:842] (0/4) Epoch 2, batch 3200, loss[loss=0.298, simple_loss=0.3361, pruned_loss=0.1299, over 4839.00 frames.], tot_loss[loss=0.3236, simple_loss=0.3687, pruned_loss=0.1393, over 1412342.71 frames.], batch size: 53, lr: 1.82e-03
2022-05-26 13:30:56,890 INFO [train.py:842] (0/4) Epoch 2, batch 3250, loss[loss=0.2998, simple_loss=0.3605, pruned_loss=0.1195, over 7204.00 frames.], tot_loss[loss=0.3251, simple_loss=0.3697, pruned_loss=0.1402, over 1415061.01 frames.], batch size: 23, lr: 1.82e-03
2022-05-26 13:31:35,491 INFO [train.py:842] (0/4) Epoch 2, batch 3300, loss[loss=0.3127, simple_loss=0.3638, pruned_loss=0.1308, over 7209.00 frames.], tot_loss[loss=0.322, simple_loss=0.3675, pruned_loss=0.1383, over 1419727.52 frames.], batch size: 22, lr: 1.82e-03
2022-05-26 13:32:14,257 INFO [train.py:842] (0/4) Epoch 2, batch 3350, loss[loss=0.2973, simple_loss=0.3546, pruned_loss=0.12, over 7191.00 frames.], tot_loss[loss=0.3216, simple_loss=0.3678, pruned_loss=0.1377, over 1422736.40 frames.], batch size: 26, lr: 1.81e-03
2022-05-26 13:32:52,898 INFO [train.py:842] (0/4) Epoch 2, batch 3400, loss[loss=0.2842, simple_loss=0.326, pruned_loss=0.1212, over 7156.00 frames.], tot_loss[loss=0.3206, simple_loss=0.3668, pruned_loss=0.1373, over 1424812.81 frames.], batch size: 17, lr: 1.81e-03
2022-05-26 13:33:31,547 INFO [train.py:842] (0/4) Epoch 2, batch 3450, loss[loss=0.294, simple_loss=0.3609, pruned_loss=0.1135, over 7295.00 frames.], tot_loss[loss=0.3184, simple_loss=0.3656, pruned_loss=0.1356, over 1426655.71 frames.], batch size: 24, lr: 1.81e-03
2022-05-26 13:34:10,026 INFO [train.py:842] (0/4) Epoch 2, batch 3500, loss[loss=0.3436, simple_loss=0.3828, pruned_loss=0.1522, over 6273.00 frames.], tot_loss[loss=0.3177, simple_loss=0.365, pruned_loss=0.1352, over 1423986.62 frames.], batch size: 37, lr: 1.80e-03
2022-05-26 13:34:48,895 INFO [train.py:842] (0/4) Epoch 2, batch 3550, loss[loss=0.3517, simple_loss=0.395, pruned_loss=0.1542, over 7312.00 frames.], tot_loss[loss=0.3194, simple_loss=0.3664, pruned_loss=0.1362, over 1424415.13 frames.], batch size: 25, lr: 1.80e-03
2022-05-26 13:35:27,117 INFO [train.py:842] (0/4) Epoch 2, batch 3600, loss[loss=0.4865, simple_loss=0.4573, pruned_loss=0.2579, over 7230.00 frames.], tot_loss[loss=0.3189, simple_loss=0.3663, pruned_loss=0.1358, over 1425592.74 frames.], batch size: 20, lr: 1.80e-03
2022-05-26 13:36:06,067 INFO [train.py:842] (0/4) Epoch 2, batch 3650, loss[loss=0.2698, simple_loss=0.3231, pruned_loss=0.1082, over 7196.00 frames.], tot_loss[loss=0.317, simple_loss=0.3647, pruned_loss=0.1346, over 1427878.45 frames.], batch size: 16, lr: 1.79e-03
2022-05-26 13:36:44,543 INFO [train.py:842] (0/4) Epoch 2, batch 3700, loss[loss=0.2694, simple_loss=0.3356, pruned_loss=0.1017, over 7156.00 frames.], tot_loss[loss=0.3138, simple_loss=0.3629, pruned_loss=0.1324, over 1429803.89 frames.], batch size: 19, lr: 1.79e-03
2022-05-26 13:37:23,446 INFO [train.py:842] (0/4) Epoch 2, batch 3750, loss[loss=0.2851, simple_loss=0.3527, pruned_loss=0.1087, over 7275.00 frames.], tot_loss[loss=0.3144, simple_loss=0.3634, pruned_loss=0.1327, over 1430441.22 frames.], batch size: 24, lr: 1.79e-03
2022-05-26 13:38:02,063 INFO [train.py:842] (0/4) Epoch 2, batch 3800, loss[loss=0.2788, simple_loss=0.3242, pruned_loss=0.1167, over 6983.00 frames.], tot_loss[loss=0.3152, simple_loss=0.3637, pruned_loss=0.1333, over 1431455.57 frames.], batch size: 16, lr: 1.79e-03
2022-05-26 13:38:40,878 INFO [train.py:842] (0/4) Epoch 2, batch 3850, loss[loss=0.296, simple_loss=0.3562, pruned_loss=0.1179, over 7194.00 frames.], tot_loss[loss=0.313, simple_loss=0.3625, pruned_loss=0.1318, over 1431685.81 frames.], batch size: 22, lr: 1.78e-03
2022-05-26 13:39:19,556 INFO [train.py:842] (0/4) Epoch 2, batch 3900, loss[loss=0.3312, simple_loss=0.3837, pruned_loss=0.1393, over 6668.00 frames.], tot_loss[loss=0.3116, simple_loss=0.3617, pruned_loss=0.1308, over 1434161.73 frames.], batch size: 38, lr: 1.78e-03
2022-05-26 13:39:58,481 INFO [train.py:842] (0/4) Epoch 2, batch 3950, loss[loss=0.3629, simple_loss=0.4066, pruned_loss=0.1596, over 7311.00 frames.], tot_loss[loss=0.311, simple_loss=0.3609, pruned_loss=0.1306, over 1432239.67 frames.], batch size: 21, lr: 1.78e-03
2022-05-26 13:40:36,944 INFO [train.py:842] (0/4) Epoch 2, batch 4000, loss[loss=0.4032, simple_loss=0.4114, pruned_loss=0.1975, over 4998.00 frames.], tot_loss[loss=0.3119, simple_loss=0.362, pruned_loss=0.1309, over 1432240.47 frames.], batch size: 52, lr: 1.77e-03
2022-05-26 13:41:15,590 INFO [train.py:842] (0/4) Epoch 2, batch 4050, loss[loss=0.2992, simple_loss=0.3603, pruned_loss=0.1191, over 6865.00 frames.], tot_loss[loss=0.3137, simple_loss=0.3633, pruned_loss=0.132, over 1427768.06 frames.], batch size: 31, lr: 1.77e-03
2022-05-26 13:41:54,111 INFO [train.py:842] (0/4) Epoch 2, batch 4100, loss[loss=0.3345, simple_loss=0.3903, pruned_loss=0.1393, over 7122.00 frames.], tot_loss[loss=0.3159, simple_loss=0.365, pruned_loss=0.1334, over 1429518.69 frames.], batch size: 28, lr: 1.77e-03
2022-05-26 13:42:32,939 INFO [train.py:842] (0/4) Epoch 2, batch 4150, loss[loss=0.3797, simple_loss=0.4091, pruned_loss=0.1751, over 7134.00 frames.], tot_loss[loss=0.3152, simple_loss=0.3637, pruned_loss=0.1334, over 1427396.59 frames.], batch size: 26, lr: 1.76e-03
2022-05-26 13:43:11,631 INFO [train.py:842] (0/4) Epoch 2, batch 4200, loss[loss=0.2299, simple_loss=0.2872, pruned_loss=0.08627, over 7005.00 frames.], tot_loss[loss=0.3128, simple_loss=0.3622, pruned_loss=0.1317, over 1425642.80 frames.], batch size: 16, lr: 1.76e-03
2022-05-26 13:43:50,287 INFO [train.py:842] (0/4) Epoch 2, batch 4250, loss[loss=0.2626, simple_loss=0.3339, pruned_loss=0.09568, over 7215.00 frames.], tot_loss[loss=0.3122, simple_loss=0.362, pruned_loss=0.1311, over 1424008.27 frames.], batch size: 22, lr: 1.76e-03
2022-05-26 13:44:28,859 INFO [train.py:842] (0/4) Epoch 2, batch 4300, loss[loss=0.3121, simple_loss=0.3646, pruned_loss=0.1298, over 7329.00 frames.], tot_loss[loss=0.309, simple_loss=0.3596, pruned_loss=0.1292, over 1426063.30 frames.], batch size: 22, lr: 1.76e-03
2022-05-26 13:45:07,462 INFO [train.py:842] (0/4) Epoch 2, batch 4350, loss[loss=0.2406, simple_loss=0.3062, pruned_loss=0.08755, over 7163.00 frames.], tot_loss[loss=0.3094, simple_loss=0.3598, pruned_loss=0.1295, over 1422688.45 frames.], batch size: 19, lr: 1.75e-03
2022-05-26 13:45:45,797 INFO [train.py:842] (0/4) Epoch 2, batch 4400, loss[loss=0.2755, simple_loss=0.3373, pruned_loss=0.1069, over 7270.00 frames.], tot_loss[loss=0.3108, simple_loss=0.3611, pruned_loss=0.1302, over 1423356.02 frames.], batch size: 24, lr: 1.75e-03
2022-05-26 13:46:25,232 INFO [train.py:842] (0/4) Epoch 2, batch 4450, loss[loss=0.3168, simple_loss=0.3505, pruned_loss=0.1415, over 7409.00 frames.], tot_loss[loss=0.3112, simple_loss=0.3613, pruned_loss=0.1306, over 1423661.71 frames.], batch size: 18, lr: 1.75e-03
2022-05-26 13:47:03,778 INFO [train.py:842] (0/4) Epoch 2, batch 4500, loss[loss=0.2651, simple_loss=0.3223, pruned_loss=0.104, over 7333.00 frames.], tot_loss[loss=0.3108, simple_loss=0.3608, pruned_loss=0.1304, over 1425320.65 frames.], batch size: 20, lr: 1.74e-03
2022-05-26 13:47:42,573 INFO [train.py:842] (0/4) Epoch 2, batch 4550, loss[loss=0.4127, simple_loss=0.4214, pruned_loss=0.202, over 7275.00 frames.], tot_loss[loss=0.3122, simple_loss=0.3617, pruned_loss=0.1314, over 1426037.82 frames.], batch size: 18, lr: 1.74e-03
2022-05-26 13:48:20,814 INFO [train.py:842] (0/4) Epoch 2, batch 4600, loss[loss=0.2938, simple_loss=0.3535, pruned_loss=0.117, over 7201.00 frames.], tot_loss[loss=0.3121, simple_loss=0.3616, pruned_loss=0.1313, over 1421071.45 frames.], batch size: 22, lr: 1.74e-03
2022-05-26 13:48:59,680 INFO [train.py:842] (0/4) Epoch 2, batch 4650, loss[loss=0.327, simple_loss=0.3816, pruned_loss=0.1362, over 7265.00 frames.], tot_loss[loss=0.3099, simple_loss=0.3602, pruned_loss=0.1298, over 1424222.51 frames.], batch size: 25, lr: 1.74e-03
2022-05-26 13:49:38,508 INFO [train.py:842] (0/4) Epoch 2, batch 4700, loss[loss=0.3141, simple_loss=0.3671, pruned_loss=0.1306, over 7318.00 frames.], tot_loss[loss=0.3102, simple_loss=0.3599, pruned_loss=0.1303, over 1424918.93 frames.], batch size: 21, lr: 1.73e-03
2022-05-26 13:50:16,914 INFO [train.py:842] (0/4) Epoch 2, batch 4750, loss[loss=0.3369, simple_loss=0.3851, pruned_loss=0.1444, over 7410.00 frames.], tot_loss[loss=0.3127, simple_loss=0.3614, pruned_loss=0.132, over 1417559.00 frames.], batch size: 21, lr: 1.73e-03
2022-05-26 13:50:55,337 INFO [train.py:842] (0/4) Epoch 2, batch 4800, loss[loss=0.3368, simple_loss=0.3831, pruned_loss=0.1453, over 7301.00 frames.], tot_loss[loss=0.3143, simple_loss=0.3631, pruned_loss=0.1328, over 1415849.02 frames.], batch size: 24, lr: 1.73e-03
2022-05-26 13:51:34,134 INFO [train.py:842] (0/4) Epoch 2, batch 4850, loss[loss=0.2651, simple_loss=0.3228, pruned_loss=0.1037, over 7150.00 frames.], tot_loss[loss=0.3151, simple_loss=0.3634, pruned_loss=0.1333, over 1415946.88 frames.], batch size: 18, lr: 1.73e-03
2022-05-26 13:52:12,637 INFO [train.py:842] (0/4) Epoch 2, batch 4900, loss[loss=0.2168, simple_loss=0.2759, pruned_loss=0.07888, over 7276.00 frames.], tot_loss[loss=0.3118, simple_loss=0.3611, pruned_loss=0.1313, over 1418653.37 frames.], batch size: 17, lr: 1.72e-03
2022-05-26 13:52:51,828 INFO [train.py:842] (0/4) Epoch 2, batch 4950, loss[loss=0.2968, simple_loss=0.362, pruned_loss=0.1158, over 7240.00 frames.], tot_loss[loss=0.312, simple_loss=0.3609, pruned_loss=0.1316, over 1421182.04 frames.], batch size: 20, lr: 1.72e-03
2022-05-26 13:53:30,406 INFO [train.py:842] (0/4) Epoch 2, batch 5000, loss[loss=0.321, simple_loss=0.3502, pruned_loss=0.1459, over 7285.00 frames.], tot_loss[loss=0.3112, simple_loss=0.3608, pruned_loss=0.1308, over 1423476.22 frames.], batch size: 17, lr: 1.72e-03
2022-05-26 13:54:08,956 INFO [train.py:842] (0/4) Epoch 2, batch 5050, loss[loss=0.3242, simple_loss=0.3755, pruned_loss=0.1364, over 7414.00 frames.], tot_loss[loss=0.3131, simple_loss=0.3625, pruned_loss=0.1319, over 1417393.53 frames.], batch size: 21, lr: 1.71e-03
2022-05-26 13:54:47,653 INFO [train.py:842] (0/4) Epoch 2, batch 5100, loss[loss=0.2672, simple_loss=0.3376, pruned_loss=0.09841, over 7158.00 frames.], tot_loss[loss=0.3108, simple_loss=0.3607, pruned_loss=0.1305, over 1419448.24 frames.], batch size: 19, lr: 1.71e-03
2022-05-26 13:55:26,640 INFO [train.py:842] (0/4) Epoch 2, batch 5150, loss[loss=0.3949, simple_loss=0.4238, pruned_loss=0.183, over 7222.00 frames.], tot_loss[loss=0.3129, simple_loss=0.3621, pruned_loss=0.1318, over 1421226.94 frames.], batch size: 21, lr: 1.71e-03
2022-05-26 13:56:05,013 INFO [train.py:842] (0/4) Epoch 2, batch 5200, loss[loss=0.4194, simple_loss=0.4392, pruned_loss=0.1998, over 7283.00 frames.], tot_loss[loss=0.3116, simple_loss=0.3613, pruned_loss=0.1309, over 1422589.00 frames.], batch size: 25, lr: 1.71e-03
2022-05-26 13:56:43,797 INFO [train.py:842] (0/4) Epoch 2, batch 5250, loss[loss=0.3567, simple_loss=0.3971, pruned_loss=0.1581, over 6952.00 frames.], tot_loss[loss=0.3107, simple_loss=0.3609, pruned_loss=0.1302, over 1424991.62 frames.], batch size: 32, lr: 1.70e-03
2022-05-26 13:57:22,561 INFO [train.py:842] (0/4) Epoch 2, batch 5300, loss[loss=0.2915, simple_loss=0.3475, pruned_loss=0.1177, over 7373.00 frames.], tot_loss[loss=0.3094, simple_loss=0.3598, pruned_loss=0.1294, over 1421943.71 frames.], batch size: 23, lr: 1.70e-03
2022-05-26 13:58:01,638 INFO [train.py:842] (0/4) Epoch 2, batch 5350, loss[loss=0.2939, simple_loss=0.3554, pruned_loss=0.1162, over 7353.00 frames.], tot_loss[loss=0.3076, simple_loss=0.3584, pruned_loss=0.1283, over 1419169.17 frames.], batch size: 19, lr: 1.70e-03
2022-05-26 13:58:40,204 INFO [train.py:842] (0/4) Epoch 2, batch 5400, loss[loss=0.3269, simple_loss=0.3803, pruned_loss=0.1367, over 6296.00 frames.], tot_loss[loss=0.3065, simple_loss=0.3576, pruned_loss=0.1277, over 1419700.28 frames.], batch size: 37, lr: 1.70e-03
2022-05-26 13:59:19,543 INFO [train.py:842] (0/4) Epoch 2, batch 5450, loss[loss=0.3064, simple_loss=0.3457, pruned_loss=0.1336, over 6799.00 frames.], tot_loss[loss=0.3048, simple_loss=0.3561, pruned_loss=0.1267, over 1421266.45 frames.], batch size: 15, lr: 1.69e-03
2022-05-26 13:59:58,074 INFO [train.py:842] (0/4) Epoch 2, batch 5500, loss[loss=0.2624, simple_loss=0.3162, pruned_loss=0.1043, over 7131.00 frames.], tot_loss[loss=0.3058, simple_loss=0.3564, pruned_loss=0.1276, over 1422566.08 frames.], batch size: 17, lr: 1.69e-03
2022-05-26 14:00:37,017 INFO [train.py:842] (0/4) Epoch 2, batch 5550, loss[loss=0.2938, simple_loss=0.3514, pruned_loss=0.1181, over 6992.00 frames.], tot_loss[loss=0.3063, simple_loss=0.3565, pruned_loss=0.128, over 1422978.52 frames.], batch size: 16, lr: 1.69e-03
2022-05-26 14:01:15,380 INFO [train.py:842] (0/4) Epoch 2, batch 5600, loss[loss=0.3209, simple_loss=0.3699, pruned_loss=0.1359, over 7286.00 frames.], tot_loss[loss=0.3091, simple_loss=0.359, pruned_loss=0.1296, over 1423625.70 frames.], batch size: 24, lr: 1.69e-03
2022-05-26 14:01:54,145 INFO [train.py:842] (0/4) Epoch 2, batch 5650, loss[loss=0.3133, simple_loss=0.3796, pruned_loss=0.1235, over 7227.00 frames.], tot_loss[loss=0.3082, simple_loss=0.3587, pruned_loss=0.1289, over 1424675.85 frames.], batch size: 23, lr: 1.68e-03
2022-05-26 14:02:32,792 INFO [train.py:842] (0/4) Epoch 2, batch 5700, loss[loss=0.2217, simple_loss=0.2951, pruned_loss=0.07416, over 7277.00 frames.], tot_loss[loss=0.3059, simple_loss=0.3566, pruned_loss=0.1276, over 1424371.95 frames.], batch size: 18, lr: 1.68e-03
2022-05-26 14:03:11,573 INFO [train.py:842] (0/4) Epoch 2, batch 5750, loss[loss=0.319, simple_loss=0.3799, pruned_loss=0.1291, over 7306.00 frames.], tot_loss[loss=0.3052, simple_loss=0.3564, pruned_loss=0.127, over 1422667.85 frames.], batch size: 21, lr: 1.68e-03
2022-05-26 14:03:50,309 INFO [train.py:842] (0/4) Epoch 2, batch 5800, loss[loss=0.2916, simple_loss=0.3547, pruned_loss=0.1143, over 7145.00 frames.], tot_loss[loss=0.3045, simple_loss=0.356, pruned_loss=0.1265, over 1426674.19 frames.], batch size: 26, lr: 1.68e-03
2022-05-26 14:04:29,348 INFO [train.py:842] (0/4) Epoch 2, batch 5850, loss[loss=0.3512, simple_loss=0.3884, pruned_loss=0.157, over 7405.00 frames.], tot_loss[loss=0.3063, simple_loss=0.3575, pruned_loss=0.1276, over 1422870.48 frames.], batch size: 21, lr: 1.67e-03
2022-05-26 14:05:07,952 INFO [train.py:842] (0/4) Epoch 2, batch 5900, loss[loss=0.3605, simple_loss=0.3763, pruned_loss=0.1724, over 7282.00 frames.], tot_loss[loss=0.306, simple_loss=0.3573, pruned_loss=0.1273, over 1424640.98 frames.], batch size: 17, lr: 1.67e-03
2022-05-26 14:05:46,678 INFO [train.py:842] (0/4) Epoch 2, batch 5950, loss[loss=0.3563, simple_loss=0.3956, pruned_loss=0.1585, over 7200.00 frames.], tot_loss[loss=0.3082, simple_loss=0.3593, pruned_loss=0.1285, over 1423781.91 frames.], batch size: 22, lr: 1.67e-03
2022-05-26 14:06:25,158 INFO [train.py:842] (0/4) Epoch 2, batch 6000, loss[loss=0.2648, simple_loss=0.3325, pruned_loss=0.09852, over 7405.00 frames.], tot_loss[loss=0.3077, simple_loss=0.359, pruned_loss=0.1282, over 1420434.44 frames.], batch size: 21, lr: 1.67e-03
2022-05-26 14:06:25,159 INFO [train.py:862] (0/4) Computing validation loss
2022-05-26 14:06:34,411 INFO [train.py:871] (0/4) Epoch 2, validation: loss=0.2262, simple_loss=0.3196, pruned_loss=0.0664, over 868885.00 frames.
2022-05-26 14:07:13,292 INFO [train.py:842] (0/4) Epoch 2, batch 6050, loss[loss=0.298, simple_loss=0.36, pruned_loss=0.118, over 7200.00 frames.], tot_loss[loss=0.3058, simple_loss=0.3576, pruned_loss=0.1269, over 1424145.28 frames.], batch size: 23, lr: 1.66e-03
2022-05-26 14:07:51,787 INFO [train.py:842] (0/4) Epoch 2, batch 6100, loss[loss=0.2984, simple_loss=0.3614, pruned_loss=0.1177, over 7360.00 frames.], tot_loss[loss=0.3049, simple_loss=0.357, pruned_loss=0.1264, over 1426710.01 frames.], batch size: 23, lr: 1.66e-03
2022-05-26 14:08:30,698 INFO [train.py:842] (0/4) Epoch 2, batch 6150, loss[loss=0.3289, simple_loss=0.3653, pruned_loss=0.1462, over 6988.00 frames.], tot_loss[loss=0.3046, simple_loss=0.3567, pruned_loss=0.1263, over 1426576.95 frames.], batch size: 28, lr: 1.66e-03
2022-05-26 14:09:09,176 INFO [train.py:842] (0/4) Epoch 2, batch 6200, loss[loss=0.4046, simple_loss=0.4329, pruned_loss=0.1882, over 6726.00 frames.], tot_loss[loss=0.3056, simple_loss=0.3576, pruned_loss=0.1268, over 1424669.59 frames.], batch size: 31, lr: 1.66e-03
2022-05-26 14:09:47,987 INFO [train.py:842] (0/4) Epoch 2, batch 6250, loss[loss=0.3359, simple_loss=0.3899, pruned_loss=0.1409, over 7109.00 frames.], tot_loss[loss=0.3051, simple_loss=0.3573, pruned_loss=0.1264, over 1427991.60 frames.], batch size: 21, lr: 1.65e-03
2022-05-26 14:10:27,074 INFO [train.py:842] (0/4) Epoch 2, batch 6300, loss[loss=0.3176, simple_loss=0.3629, pruned_loss=0.1362, over 7149.00 frames.], tot_loss[loss=0.306, simple_loss=0.3581, pruned_loss=0.127, over 1432149.10 frames.], batch size: 26, lr: 1.65e-03
2022-05-26 14:11:05,661 INFO [train.py:842] (0/4) Epoch 2, batch 6350, loss[loss=0.3631, simple_loss=0.4033, pruned_loss=0.1614, over 6481.00 frames.], tot_loss[loss=0.3083, simple_loss=0.3601, pruned_loss=0.1283, over 1430266.23 frames.], batch size: 38, lr: 1.65e-03
2022-05-26 14:11:44,200 INFO [train.py:842] (0/4) Epoch 2, batch 6400, loss[loss=0.2781, simple_loss=0.3499, pruned_loss=0.1031, over 6740.00 frames.], tot_loss[loss=0.3058, simple_loss=0.3577, pruned_loss=0.1269, over 1425780.11 frames.], batch size: 31, lr: 1.65e-03
2022-05-26 14:12:23,047 INFO [train.py:842] (0/4) Epoch 2, batch 6450, loss[loss=0.2486, simple_loss=0.3032, pruned_loss=0.09703, over 7417.00 frames.], tot_loss[loss=0.3049, simple_loss=0.3567, pruned_loss=0.1266, over 1425575.17 frames.], batch size: 18, lr: 1.64e-03
2022-05-26 14:13:01,662 INFO [train.py:842] (0/4) Epoch 2, batch 6500, loss[loss=0.3099, simple_loss=0.3624, pruned_loss=0.1286, over 7201.00 frames.], tot_loss[loss=0.3036, simple_loss=0.3559, pruned_loss=0.1256, over 1424325.44 frames.], batch size: 22, lr: 1.64e-03
2022-05-26 14:13:40,427 INFO [train.py:842] (0/4) Epoch 2, batch 6550, loss[loss=0.2571, simple_loss=0.3183, pruned_loss=0.098, over 7453.00 frames.], tot_loss[loss=0.3044, simple_loss=0.3567, pruned_loss=0.126, over 1422313.58 frames.], batch size: 19, lr: 1.64e-03
2022-05-26 14:14:18,959 INFO [train.py:842] (0/4) Epoch 2, batch 6600, loss[loss=0.2804, simple_loss=0.3318, pruned_loss=0.1145, over 7268.00 frames.], tot_loss[loss=0.302, simple_loss=0.355, pruned_loss=0.1245, over 1421964.36 frames.], batch size: 18, lr: 1.64e-03
2022-05-26 14:14:57,584 INFO [train.py:842] (0/4) Epoch 2, batch 6650, loss[loss=0.3094, simple_loss=0.3675, pruned_loss=0.1256, over 7211.00 frames.], tot_loss[loss=0.3056, simple_loss=0.3575, pruned_loss=0.1268, over 1414573.48 frames.], batch size: 23, lr: 1.63e-03
2022-05-26 14:15:36,108 INFO [train.py:842] (0/4) Epoch 2, batch 6700, loss[loss=0.2928, simple_loss=0.3397, pruned_loss=0.123, over 7276.00 frames.], tot_loss[loss=0.304, simple_loss=0.3565, pruned_loss=0.1257, over 1419701.40 frames.], batch size: 17, lr: 1.63e-03
2022-05-26 14:16:14,706 INFO [train.py:842] (0/4) Epoch 2, batch 6750, loss[loss=0.2973, simple_loss=0.3536, pruned_loss=0.1205, over 7232.00 frames.], tot_loss[loss=0.3016, simple_loss=0.3554, pruned_loss=0.1239, over 1422341.11 frames.], batch size: 20, lr: 1.63e-03
2022-05-26 14:16:53,059 INFO [train.py:842] (0/4) Epoch 2, batch 6800, loss[loss=0.3136, simple_loss=0.371, pruned_loss=0.1281, over 7107.00 frames.], tot_loss[loss=0.3002, simple_loss=0.3545, pruned_loss=0.1229, over 1424525.78 frames.], batch size: 21, lr: 1.63e-03
2022-05-26 14:16:57,935 INFO [checkpoint.py:75] (0/4) Saving checkpoint to streaming_pruned_transducer_stateless4/exp/checkpoint-16000.pt
2022-05-26 14:17:34,587 INFO [train.py:842] (0/4) Epoch 2, batch 6850, loss[loss=0.2625, simple_loss=0.3186, pruned_loss=0.1032, over 7328.00 frames.], tot_loss[loss=0.3007, simple_loss=0.355, pruned_loss=0.1232, over 1421156.74 frames.], batch size: 20, lr: 1.63e-03
2022-05-26 14:18:13,245 INFO [train.py:842] (0/4) Epoch 2, batch 6900, loss[loss=0.2647, simple_loss=0.3242, pruned_loss=0.1026, over 7426.00 frames.], tot_loss[loss=0.3007, simple_loss=0.355, pruned_loss=0.1231, over 1420458.42 frames.], batch size: 20, lr: 1.62e-03
2022-05-26 14:18:52,139 INFO [train.py:842] (0/4) Epoch 2, batch 6950, loss[loss=0.2548, simple_loss=0.3276, pruned_loss=0.09099, over 7288.00 frames.], tot_loss[loss=0.2992, simple_loss=0.3534, pruned_loss=0.1225, over 1420741.03 frames.], batch size: 18, lr: 1.62e-03
2022-05-26 14:19:30,566 INFO [train.py:842] (0/4) Epoch 2, batch 7000, loss[loss=0.2643, simple_loss=0.3223, pruned_loss=0.1032, over 7325.00 frames.], tot_loss[loss=0.2982, simple_loss=0.3533, pruned_loss=0.1216, over 1422796.89 frames.], batch size: 21, lr: 1.62e-03
2022-05-26 14:20:09,754 INFO [train.py:842] (0/4) Epoch 2, batch 7050, loss[loss=0.3176, simple_loss=0.3545, pruned_loss=0.1404, over 5363.00 frames.], tot_loss[loss=0.2966, simple_loss=0.3517, pruned_loss=0.1208, over 1425503.17 frames.], batch size: 52, lr: 1.62e-03
2022-05-26 14:20:48,374 INFO [train.py:842] (0/4) Epoch 2, batch 7100, loss[loss=0.2868, simple_loss=0.347, pruned_loss=0.1133, over 7112.00 frames.], tot_loss[loss=0.3, simple_loss=0.3541, pruned_loss=0.123, over 1425121.55 frames.], batch size: 21, lr: 1.61e-03
2022-05-26 14:21:26,952 INFO [train.py:842] (0/4) Epoch 2, batch 7150, loss[loss=0.2855, simple_loss=0.3479, pruned_loss=0.1115, over 7416.00 frames.], tot_loss[loss=0.3029, simple_loss=0.356, pruned_loss=0.1249, over 1422132.13 frames.], batch size: 21, lr: 1.61e-03
2022-05-26 14:22:05,293 INFO [train.py:842] (0/4) Epoch 2, batch 7200, loss[loss=0.2315, simple_loss=0.2893, pruned_loss=0.08682, over 6991.00 frames.], tot_loss[loss=0.3026, simple_loss=0.3555, pruned_loss=0.1248, over 1419725.14 frames.], batch size: 16, lr: 1.61e-03
2022-05-26 14:22:44,438 INFO [train.py:842] (0/4) Epoch 2, batch 7250, loss[loss=0.2656, simple_loss=0.3353, pruned_loss=0.09794, over 7236.00 frames.], tot_loss[loss=0.3015, simple_loss=0.355, pruned_loss=0.124, over 1425049.89 frames.], batch size: 20, lr: 1.61e-03
2022-05-26 14:23:22,967 INFO [train.py:842] (0/4) Epoch 2, batch 7300, loss[loss=0.3138, simple_loss=0.3659, pruned_loss=0.1309, over 7210.00 frames.], tot_loss[loss=0.3032, simple_loss=0.3558, pruned_loss=0.1254, over 1427524.08 frames.], batch size: 21, lr: 1.60e-03
2022-05-26 14:24:01,873 INFO [train.py:842] (0/4) Epoch 2, batch 7350, loss[loss=0.4031, simple_loss=0.4264, pruned_loss=0.1899, over 4822.00 frames.], tot_loss[loss=0.3004, simple_loss=0.3535, pruned_loss=0.1236, over 1423703.50 frames.], batch size: 52, lr: 1.60e-03
2022-05-26 14:24:40,512 INFO [train.py:842] (0/4) Epoch 2, batch 7400, loss[loss=0.2508, simple_loss=0.313, pruned_loss=0.09431, over 6985.00 frames.], tot_loss[loss=0.2999, simple_loss=0.3532, pruned_loss=0.1233, over 1423095.26 frames.], batch size: 16, lr: 1.60e-03
2022-05-26 14:25:19,330 INFO [train.py:842] (0/4) Epoch 2, batch 7450, loss[loss=0.2896, simple_loss=0.3417, pruned_loss=0.1187, over 7343.00 frames.], tot_loss[loss=0.3006, simple_loss=0.3537, pruned_loss=0.1238, over 1418852.70 frames.], batch size: 19, lr: 1.60e-03
2022-05-26 14:25:57,996 INFO [train.py:842] (0/4) Epoch 2, batch 7500, loss[loss=0.2714, simple_loss=0.3458, pruned_loss=0.0985, over 7213.00 frames.], tot_loss[loss=0.302, simple_loss=0.355, pruned_loss=0.1244, over 1419843.71 frames.], batch size: 21, lr: 1.60e-03
2022-05-26 14:26:36,915 INFO [train.py:842] (0/4) Epoch 2, batch 7550, loss[loss=0.3039, simple_loss=0.3596, pruned_loss=0.1241, over 7404.00 frames.], tot_loss[loss=0.3003, simple_loss=0.354, pruned_loss=0.1233, over 1420963.12 frames.], batch size: 21, lr: 1.59e-03
2022-05-26 14:27:15,663 INFO [train.py:842] (0/4) Epoch 2, batch 7600, loss[loss=0.3808, simple_loss=0.4069, pruned_loss=0.1773, over 5391.00 frames.], tot_loss[loss=0.3001, simple_loss=0.3535, pruned_loss=0.1233, over 1421102.83 frames.], batch size: 52, lr: 1.59e-03
2022-05-26 14:27:54,288 INFO [train.py:842] (0/4) Epoch 2, batch 7650, loss[loss=0.3183, simple_loss=0.3721, pruned_loss=0.1322, over 7420.00 frames.], tot_loss[loss=0.2993, simple_loss=0.3531, pruned_loss=0.1227, over 1421699.63 frames.], batch size: 21, lr: 1.59e-03
2022-05-26 14:28:32,786 INFO [train.py:842] (0/4) Epoch 2, batch 7700, loss[loss=0.3323, simple_loss=0.3852, pruned_loss=0.1397, over 7341.00 frames.], tot_loss[loss=0.2979, simple_loss=0.3521, pruned_loss=0.1218, over 1423316.38 frames.], batch size: 22, lr: 1.59e-03
2022-05-26 14:29:11,466 INFO [train.py:842] (0/4) Epoch 2, batch 7750, loss[loss=0.3157, simple_loss=0.381, pruned_loss=0.1252, over 7078.00 frames.], tot_loss[loss=0.298, simple_loss=0.3526, pruned_loss=0.1217, over 1424497.74 frames.], batch size: 28, lr: 1.59e-03
2022-05-26 14:29:50,000 INFO [train.py:842] (0/4) Epoch 2, batch 7800, loss[loss=0.3816, simple_loss=0.3983, pruned_loss=0.1824, over 7141.00 frames.], tot_loss[loss=0.2995, simple_loss=0.3539, pruned_loss=0.1226, over 1423731.93 frames.], batch size: 20, lr: 1.58e-03
2022-05-26 14:30:28,804 INFO [train.py:842] (0/4) Epoch 2, batch 7850, loss[loss=0.369, simple_loss=0.4027, pruned_loss=0.1677, over 7320.00 frames.], tot_loss[loss=0.2983, simple_loss=0.3529, pruned_loss=0.1218, over 1424272.71 frames.], batch size: 21, lr: 1.58e-03
2022-05-26 14:31:07,199 INFO [train.py:842] (0/4) Epoch 2, batch 7900, loss[loss=0.4059, simple_loss=0.415, pruned_loss=0.1984, over 5202.00 frames.], tot_loss[loss=0.3006, simple_loss=0.3545, pruned_loss=0.1234, over 1426656.01 frames.], batch size: 53, lr: 1.58e-03
2022-05-26 14:31:46,043 INFO [train.py:842] (0/4) Epoch 2, batch 7950, loss[loss=0.288, simple_loss=0.344, pruned_loss=0.1161, over 7159.00 frames.], tot_loss[loss=0.2984, simple_loss=0.3532, pruned_loss=0.1219, over 1428315.63 frames.], batch size: 18, lr: 1.58e-03
2022-05-26 14:32:24,256 INFO [train.py:842] (0/4) Epoch 2, batch 8000, loss[loss=0.2746, simple_loss=0.3423, pruned_loss=0.1034, over 7211.00 frames.], tot_loss[loss=0.2971, simple_loss=0.3524, pruned_loss=0.1209, over 1426745.97 frames.], batch size: 21, lr: 1.57e-03
2022-05-26 14:33:02,919 INFO [train.py:842] (0/4) Epoch 2, batch 8050, loss[loss=0.3373, simple_loss=0.3807, pruned_loss=0.1469, over 6473.00 frames.], tot_loss[loss=0.299, simple_loss=0.3538, pruned_loss=0.122, over 1424708.70 frames.], batch size: 37, lr: 1.57e-03
2022-05-26 14:33:41,415 INFO [train.py:842] (0/4) Epoch 2, batch 8100, loss[loss=0.3343, simple_loss=0.3768, pruned_loss=0.1459, over 7196.00 frames.], tot_loss[loss=0.2982, simple_loss=0.3532, pruned_loss=0.1216, over 1427660.09 frames.], batch size: 26, lr: 1.57e-03
2022-05-26 14:34:20,565 INFO [train.py:842] (0/4) Epoch 2, batch 8150, loss[loss=0.3023, simple_loss=0.3543, pruned_loss=0.1251, over 7076.00 frames.], tot_loss[loss=0.2965, simple_loss=0.3516, pruned_loss=0.1207, over 1429183.01 frames.], batch size: 18, lr: 1.57e-03
2022-05-26 14:34:58,967 INFO [train.py:842] (0/4) Epoch 2, batch 8200, loss[loss=0.2643, simple_loss=0.324, pruned_loss=0.1023, over 7271.00 frames.], tot_loss[loss=0.2971, simple_loss=0.3521, pruned_loss=0.1211, over 1424677.95 frames.], batch size: 18, lr: 1.57e-03
2022-05-26 14:35:38,135 INFO [train.py:842] (0/4) Epoch 2, batch 8250, loss[loss=0.3314, simple_loss=0.3752, pruned_loss=0.1438, over 7070.00 frames.], tot_loss[loss=0.296, simple_loss=0.3505, pruned_loss=0.1208, over 1422250.89 frames.], batch size: 28, lr: 1.56e-03
2022-05-26 14:36:16,475 INFO [train.py:842] (0/4) Epoch 2, batch 8300, loss[loss=0.3009, simple_loss=0.3566, pruned_loss=0.1226, over 7145.00 frames.], tot_loss[loss=0.2958, simple_loss=0.3505, pruned_loss=0.1206, over 1420461.16 frames.], batch size: 20, lr: 1.56e-03
2022-05-26 14:36:55,279 INFO [train.py:842] (0/4) Epoch 2, batch 8350, loss[loss=0.4366, simple_loss=0.4468, pruned_loss=0.2132, over 4922.00 frames.], tot_loss[loss=0.2936, simple_loss=0.3494, pruned_loss=0.1188, over 1420153.56 frames.], batch size: 53, lr: 1.56e-03
2022-05-26 14:37:33,565 INFO [train.py:842] (0/4) Epoch 2, batch 8400, loss[loss=0.2579, simple_loss=0.3177, pruned_loss=0.09903, over 7148.00 frames.], tot_loss[loss=0.2941, simple_loss=0.3499, pruned_loss=0.1191, over 1419376.02 frames.], batch size: 17, lr: 1.56e-03
2022-05-26 14:38:12,064 INFO [train.py:842] (0/4) Epoch 2, batch 8450, loss[loss=0.3288, simple_loss=0.4024, pruned_loss=0.1275, over 7211.00 frames.], tot_loss[loss=0.2953, simple_loss=0.3512, pruned_loss=0.1197, over 1415027.74 frames.], batch size: 22, lr: 1.56e-03
2022-05-26 14:38:50,512 INFO [train.py:842] (0/4) Epoch 2, batch 8500, loss[loss=0.3838, simple_loss=0.3916, pruned_loss=0.188, over 7137.00 frames.], tot_loss[loss=0.2961, simple_loss=0.3512, pruned_loss=0.1205, over 1418660.75 frames.], batch size: 17, lr: 1.55e-03
2022-05-26 14:39:29,175 INFO [train.py:842] (0/4) Epoch 2, batch 8550, loss[loss=0.2275, simple_loss=0.2929, pruned_loss=0.08105, over 7342.00 frames.], tot_loss[loss=0.2967, simple_loss=0.3521, pruned_loss=0.1207, over 1423267.61 frames.], batch size: 19, lr: 1.55e-03
2022-05-26 14:40:07,849 INFO [train.py:842] (0/4) Epoch 2, batch 8600, loss[loss=0.374, simple_loss=0.4082, pruned_loss=0.1699, over 6446.00 frames.], tot_loss[loss=0.2945, simple_loss=0.3498, pruned_loss=0.1196, over 1420207.10 frames.], batch size: 37, lr: 1.55e-03
2022-05-26 14:40:46,956 INFO [train.py:842] (0/4) Epoch 2, batch 8650, loss[loss=0.2873, simple_loss=0.355, pruned_loss=0.1098, over 7149.00 frames.], tot_loss[loss=0.2959, simple_loss=0.3509, pruned_loss=0.1204, over 1422166.54 frames.], batch size: 20, lr: 1.55e-03
2022-05-26 14:41:25,748 INFO [train.py:842] (0/4) Epoch 2, batch 8700, loss[loss=0.269, simple_loss=0.327, pruned_loss=0.1055, over 7075.00 frames.], tot_loss[loss=0.2932, simple_loss=0.3493, pruned_loss=0.1186, over 1421121.19 frames.], batch size: 18, lr: 1.55e-03
2022-05-26 14:42:04,176 INFO [train.py:842] (0/4) Epoch 2, batch 8750, loss[loss=0.2349, simple_loss=0.301, pruned_loss=0.08436, over 7160.00 frames.], tot_loss[loss=0.2922, simple_loss=0.3487, pruned_loss=0.1179, over 1420532.75 frames.], batch size: 18, lr: 1.54e-03
2022-05-26 14:42:42,567 INFO [train.py:842] (0/4) Epoch 2, batch 8800, loss[loss=0.3222, simple_loss=0.3763, pruned_loss=0.134, over 7333.00 frames.], tot_loss[loss=0.2976, simple_loss=0.3518, pruned_loss=0.1217, over 1412882.90 frames.], batch size: 22, lr: 1.54e-03
2022-05-26 14:43:21,145 INFO [train.py:842] (0/4) Epoch 2, batch 8850, loss[loss=0.3963, simple_loss=0.4275, pruned_loss=0.1826, over 7267.00 frames.], tot_loss[loss=0.2979, simple_loss=0.3521, pruned_loss=0.1219, over 1411056.68 frames.], batch size: 24, lr: 1.54e-03
2022-05-26 14:43:59,278 INFO [train.py:842] (0/4) Epoch 2, batch 8900, loss[loss=0.2722, simple_loss=0.3359, pruned_loss=0.1042, over 6667.00 frames.], tot_loss[loss=0.2988, simple_loss=0.3526, pruned_loss=0.1225, over 1401951.45 frames.], batch size: 31, lr: 1.54e-03
2022-05-26 14:44:37,748 INFO [train.py:842] (0/4) Epoch 2, batch 8950, loss[loss=0.3059, simple_loss=0.3656, pruned_loss=0.123, over 7111.00 frames.], tot_loss[loss=0.2978, simple_loss=0.352, pruned_loss=0.1219, over 1402203.66 frames.], batch size: 21, lr: 1.54e-03
2022-05-26 14:45:16,040 INFO [train.py:842] (0/4) Epoch 2, batch 9000, loss[loss=0.3234, simple_loss=0.362, pruned_loss=0.1424, over 7271.00 frames.], tot_loss[loss=0.2993, simple_loss=0.3532, pruned_loss=0.1227, over 1397438.09 frames.], batch size: 18, lr: 1.53e-03
2022-05-26 14:45:16,041 INFO [train.py:862] (0/4) Computing validation loss
2022-05-26 14:45:25,234 INFO [train.py:871] (0/4) Epoch 2, validation: loss=0.2179, simple_loss=0.3144, pruned_loss=0.06069, over 868885.00 frames.
2022-05-26 14:46:03,593 INFO [train.py:842] (0/4) Epoch 2, batch 9050, loss[loss=0.2595, simple_loss=0.3231, pruned_loss=0.09797, over 7274.00 frames.], tot_loss[loss=0.3014, simple_loss=0.3545, pruned_loss=0.1241, over 1382698.79 frames.], batch size: 18, lr: 1.53e-03
2022-05-26 14:46:40,966 INFO [train.py:842] (0/4) Epoch 2, batch 9100, loss[loss=0.3601, simple_loss=0.4026, pruned_loss=0.1588, over 5105.00 frames.], tot_loss[loss=0.3054, simple_loss=0.3572, pruned_loss=0.1268, over 1329781.40 frames.], batch size: 52, lr: 1.53e-03
2022-05-26 14:47:18,817 INFO [train.py:842] (0/4) Epoch 2, batch 9150, loss[loss=0.337, simple_loss=0.3778, pruned_loss=0.1481, over 5021.00 frames.], tot_loss[loss=0.3137, simple_loss=0.3626, pruned_loss=0.1324, over 1258555.56 frames.], batch size: 52, lr: 1.53e-03