Luigi commited on
Commit
3a59c70
1 Parent(s): 0057ce2

Upload INT8-quantized model with calibration

Browse files
yolo_nas_pose_l_int8.onnx.best.engine CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6be26b88c524b32a7fe99eeaa6166ae80bd428e90d7f45b24317115286e1c80
3
- size 57269617
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23bee9f331f0da470fb3b6d1e4244127e7648c848e5cf9265fbeea5a15c2204d
3
+ size 57380810
yolo_nas_pose_l_int8.onnx.best.engine.err CHANGED
@@ -1,7 +1,7 @@
1
- [12/28/2023-19:27:37] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
- [12/28/2023-19:27:37] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
- [12/28/2023-19:27:42] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
- [12/28/2023-20:16:25] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
- [12/28/2023-20:16:25] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
- [12/28/2023-20:16:25] [W] * GPU compute time is unstable, with coefficient of variance = 6.69708%.
7
- [12/28/2023-20:16:25] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
 
1
+ [01/04/2024-16:29:12] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
+ [01/04/2024-16:29:12] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
+ [01/04/2024-16:29:16] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
+ [01/04/2024-17:18:11] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
+ [01/04/2024-17:18:11] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
+ [01/04/2024-17:18:11] [W] * GPU compute time is unstable, with coefficient of variance = 2.14899%.
7
+ [01/04/2024-17:18:11] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
yolo_nas_pose_l_int8.onnx.best.engine.log CHANGED
The diff for this file is too large to render. See raw diff
 
yolo_nas_pose_l_int8.onnx.int8.engine CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0437c2287a6331143c0b7181c0ce945e117b29eae0d9755f226919acbe4a8cff
3
- size 57433967
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0e0b6cf94d6d65e68c969cd873afeffe71477632d733e3a4f8a9d1d9242fa72
3
+ size 57416931
yolo_nas_pose_l_int8.onnx.int8.engine.err CHANGED
@@ -1,7 +1,7 @@
1
- [12/28/2023-20:16:35] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
- [12/28/2023-20:16:35] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
- [12/28/2023-20:16:40] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
- [12/28/2023-20:31:04] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
- [12/28/2023-20:31:04] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
- [12/28/2023-20:31:04] [W] * GPU compute time is unstable, with coefficient of variance = 6.47493%.
7
- [12/28/2023-20:31:04] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
 
1
+ [01/04/2024-17:18:28] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
+ [01/04/2024-17:18:28] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
+ [01/04/2024-17:18:32] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
+ [01/04/2024-17:33:32] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
+ [01/04/2024-17:33:32] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
+ [01/04/2024-17:33:32] [W] * GPU compute time is unstable, with coefficient of variance = 9.56431%.
7
+ [01/04/2024-17:33:32] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
yolo_nas_pose_l_int8.onnx.int8.engine.log CHANGED
The diff for this file is too large to render. See raw diff
 
yolo_nas_pose_m_int8.onnx.best.engine CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:095580091b0635e20541e483c6357dbce9cf2ce19bfab25e1197c375f9ca1807
3
- size 41724504
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d812fb42c4b1a75b40065100ad8e4532a011b7d5dd8954655ba54262cceca67
3
+ size 41711541
yolo_nas_pose_m_int8.onnx.best.engine.err CHANGED
@@ -1,7 +1,7 @@
1
- [12/28/2023-10:38:46] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
- [12/28/2023-10:38:46] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
- [12/28/2023-10:38:50] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
- [12/28/2023-11:17:13] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
- [12/28/2023-11:17:13] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
- [12/28/2023-11:17:13] [W] * GPU compute time is unstable, with coefficient of variance = 6.37849%.
7
- [12/28/2023-11:17:13] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
 
1
+ [01/04/2024-15:38:52] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
+ [01/04/2024-15:38:52] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
+ [01/04/2024-15:38:56] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
+ [01/04/2024-16:17:05] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
+ [01/04/2024-16:17:05] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
+ [01/04/2024-16:17:05] [W] * GPU compute time is unstable, with coefficient of variance = 3.17268%.
7
+ [01/04/2024-16:17:05] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
yolo_nas_pose_m_int8.onnx.best.engine.log CHANGED
The diff for this file is too large to render. See raw diff
 
yolo_nas_pose_m_int8.onnx.int8.engine CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:649a3dfa56cfade0026b0d476fab2983d60458937e7431604527c0fefc3c853c
3
- size 41664431
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bee7381541459b8b03631b7ff27dd33fc54d85b6ef4fdc450591c2062130b3f9
3
+ size 41761906
yolo_nas_pose_m_int8.onnx.int8.engine.err CHANGED
@@ -1,7 +1,7 @@
1
- [12/28/2023-11:17:20] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
- [12/28/2023-11:17:20] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
- [12/28/2023-11:17:24] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
- [12/28/2023-11:29:15] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
- [12/28/2023-11:29:15] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
- [12/28/2023-11:29:15] [W] * GPU compute time is unstable, with coefficient of variance = 6.95793%.
7
- [12/28/2023-11:29:15] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
 
1
+ [01/04/2024-16:17:13] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
+ [01/04/2024-16:17:13] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
+ [01/04/2024-16:17:16] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
+ [01/04/2024-16:29:05] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
+ [01/04/2024-16:29:05] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
+ [01/04/2024-16:29:05] [W] * GPU compute time is unstable, with coefficient of variance = 3.06633%.
7
+ [01/04/2024-16:29:05] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
yolo_nas_pose_m_int8.onnx.int8.engine.log CHANGED
The diff for this file is too large to render. See raw diff
 
yolo_nas_pose_n_int8.onnx.best.engine CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:022d94ab6ef2f6c6deba8f1cb3e09149d10b3df3fe17f24559f62f7277c75177
3
- size 10228093
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:034c3a1e3c8f16b6ba305b1b567429a8e42f2ae6c80cb63d074ca8b794844cf3
3
+ size 10229486
yolo_nas_pose_n_int8.onnx.best.engine.err CHANGED
@@ -1,7 +1,7 @@
1
- [12/27/2023-20:52:51] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
- [12/27/2023-20:52:51] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
- [12/27/2023-20:52:54] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
- [12/27/2023-21:21:24] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
- [12/27/2023-21:21:24] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
- [12/27/2023-21:21:24] [W] * GPU compute time is unstable, with coefficient of variance = 3.55467%.
7
- [12/27/2023-21:21:24] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
 
1
+ [01/04/2024-14:17:33] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
+ [01/04/2024-14:17:33] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
+ [01/04/2024-14:17:36] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
+ [01/04/2024-14:46:17] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
+ [01/04/2024-14:46:17] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
+ [01/04/2024-14:46:17] [W] * GPU compute time is unstable, with coefficient of variance = 2.71486%.
7
+ [01/04/2024-14:46:17] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
yolo_nas_pose_n_int8.onnx.best.engine.log CHANGED
@@ -1,323 +1,325 @@
1
  &&&& RUNNING TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_n_int8.onnx --best --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_n_int8.onnx.best.engine
2
- [12/27/2023-20:52:47] [I] === Model Options ===
3
- [12/27/2023-20:52:47] [I] Format: ONNX
4
- [12/27/2023-20:52:47] [I] Model: yolo_nas_pose_n_int8.onnx
5
- [12/27/2023-20:52:47] [I] Output:
6
- [12/27/2023-20:52:47] [I] === Build Options ===
7
- [12/27/2023-20:52:47] [I] Max batch: explicit batch
8
- [12/27/2023-20:52:47] [I] Memory Pools: workspace: default, dlaSRAM: default, dlaLocalDRAM: default, dlaGlobalDRAM: default
9
- [12/27/2023-20:52:47] [I] minTiming: 1
10
- [12/27/2023-20:52:47] [I] avgTiming: 8
11
- [12/27/2023-20:52:47] [I] Precision: FP32+FP16+INT8
12
- [12/27/2023-20:52:47] [I] LayerPrecisions:
13
- [12/27/2023-20:52:47] [I] Calibration: Dynamic
14
- [12/27/2023-20:52:47] [I] Refit: Disabled
15
- [12/27/2023-20:52:47] [I] Sparsity: Disabled
16
- [12/27/2023-20:52:47] [I] Safe mode: Disabled
17
- [12/27/2023-20:52:47] [I] DirectIO mode: Disabled
18
- [12/27/2023-20:52:47] [I] Restricted mode: Disabled
19
- [12/27/2023-20:52:47] [I] Build only: Disabled
20
- [12/27/2023-20:52:47] [I] Save engine: yolo_nas_pose_n_int8.onnx.best.engine
21
- [12/27/2023-20:52:47] [I] Load engine:
22
- [12/27/2023-20:52:47] [I] Profiling verbosity: 0
23
- [12/27/2023-20:52:47] [I] Tactic sources: Using default tactic sources
24
- [12/27/2023-20:52:47] [I] timingCacheMode: local
25
- [12/27/2023-20:52:47] [I] timingCacheFile:
26
- [12/27/2023-20:52:47] [I] Heuristic: Disabled
27
- [12/27/2023-20:52:47] [I] Preview Features: Use default preview flags.
28
- [12/27/2023-20:52:47] [I] Input(s)s format: fp32:CHW
29
- [12/27/2023-20:52:47] [I] Output(s)s format: fp32:CHW
30
- [12/27/2023-20:52:47] [I] Input build shapes: model
31
- [12/27/2023-20:52:47] [I] Input calibration shapes: model
32
- [12/27/2023-20:52:47] [I] === System Options ===
33
- [12/27/2023-20:52:47] [I] Device: 0
34
- [12/27/2023-20:52:47] [I] DLACore:
35
- [12/27/2023-20:52:47] [I] Plugins:
36
- [12/27/2023-20:52:47] [I] === Inference Options ===
37
- [12/27/2023-20:52:47] [I] Batch: Explicit
38
- [12/27/2023-20:52:47] [I] Input inference shapes: model
39
- [12/27/2023-20:52:47] [I] Iterations: 10
40
- [12/27/2023-20:52:47] [I] Duration: 15s (+ 200ms warm up)
41
- [12/27/2023-20:52:47] [I] Sleep time: 0ms
42
- [12/27/2023-20:52:47] [I] Idle time: 0ms
43
- [12/27/2023-20:52:47] [I] Streams: 1
44
- [12/27/2023-20:52:47] [I] ExposeDMA: Disabled
45
- [12/27/2023-20:52:47] [I] Data transfers: Enabled
46
- [12/27/2023-20:52:47] [I] Spin-wait: Disabled
47
- [12/27/2023-20:52:47] [I] Multithreading: Disabled
48
- [12/27/2023-20:52:47] [I] CUDA Graph: Disabled
49
- [12/27/2023-20:52:47] [I] Separate profiling: Disabled
50
- [12/27/2023-20:52:47] [I] Time Deserialize: Disabled
51
- [12/27/2023-20:52:47] [I] Time Refit: Disabled
52
- [12/27/2023-20:52:47] [I] NVTX verbosity: 0
53
- [12/27/2023-20:52:47] [I] Persistent Cache Ratio: 0
54
- [12/27/2023-20:52:47] [I] Inputs:
55
- [12/27/2023-20:52:47] [I] === Reporting Options ===
56
- [12/27/2023-20:52:47] [I] Verbose: Disabled
57
- [12/27/2023-20:52:47] [I] Averages: 100 inferences
58
- [12/27/2023-20:52:47] [I] Percentiles: 90,95,99
59
- [12/27/2023-20:52:47] [I] Dump refittable layers:Disabled
60
- [12/27/2023-20:52:47] [I] Dump output: Disabled
61
- [12/27/2023-20:52:47] [I] Profile: Disabled
62
- [12/27/2023-20:52:47] [I] Export timing to JSON file:
63
- [12/27/2023-20:52:47] [I] Export output to JSON file:
64
- [12/27/2023-20:52:47] [I] Export profile to JSON file:
65
- [12/27/2023-20:52:47] [I]
66
- [12/27/2023-20:52:47] [I] === Device Information ===
67
- [12/27/2023-20:52:47] [I] Selected Device: Orin
68
- [12/27/2023-20:52:47] [I] Compute Capability: 8.7
69
- [12/27/2023-20:52:47] [I] SMs: 8
70
- [12/27/2023-20:52:47] [I] Compute Clock Rate: 0.624 GHz
71
- [12/27/2023-20:52:47] [I] Device Global Memory: 7471 MiB
72
- [12/27/2023-20:52:47] [I] Shared Memory per SM: 164 KiB
73
- [12/27/2023-20:52:47] [I] Memory Bus Width: 128 bits (ECC disabled)
74
- [12/27/2023-20:52:47] [I] Memory Clock Rate: 0.624 GHz
75
- [12/27/2023-20:52:47] [I]
76
- [12/27/2023-20:52:47] [I] TensorRT version: 8.5.2
77
- [12/27/2023-20:52:48] [I] [TRT] [MemUsageChange] Init CUDA: CPU +220, GPU +0, now: CPU 249, GPU 3041 (MiB)
78
- [12/27/2023-20:52:51] [I] [TRT] [MemUsageChange] Init builder kernel library: CPU +302, GPU +283, now: CPU 574, GPU 3346 (MiB)
79
- [12/27/2023-20:52:51] [I] Start parsing network model
80
- [12/27/2023-20:52:51] [I] [TRT] ----------------------------------------------------------------
81
- [12/27/2023-20:52:51] [I] [TRT] Input filename: yolo_nas_pose_n_int8.onnx
82
- [12/27/2023-20:52:51] [I] [TRT] ONNX IR version: 0.0.8
83
- [12/27/2023-20:52:51] [I] [TRT] Opset version: 17
84
- [12/27/2023-20:52:51] [I] [TRT] Producer name: pytorch
85
- [12/27/2023-20:52:51] [I] [TRT] Producer version: 2.1.2
86
- [12/27/2023-20:52:51] [I] [TRT] Domain:
87
- [12/27/2023-20:52:51] [I] [TRT] Model version: 0
88
- [12/27/2023-20:52:51] [I] [TRT] Doc string:
89
- [12/27/2023-20:52:51] [I] [TRT] ----------------------------------------------------------------
90
- [12/27/2023-20:52:54] [I] Finish parsing network model
91
- [12/27/2023-20:52:58] [I] [TRT] ---------- Layers Running on DLA ----------
92
- [12/27/2023-20:52:58] [I] [TRT] ---------- Layers Running on GPU ----------
93
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation1]
94
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/pre_process/pre_process.0/Cast.../pre_process/pre_process.2/Mul]}
95
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1203) [Constant]
96
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1204) [Constant]
97
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1205) [Constant]
98
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/backbone/stem/conv/rbr_reparam/_input_quantizer/QuantizeLinear
99
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stem.conv.rbr_reparam.weight + /model/backbone/stem/conv/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stem/conv/rbr_reparam/Conv
100
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.downsample.rbr_reparam.weight + /model/backbone/stage1/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/downsample/rbr_reparam/Conv
101
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv2.conv.weight + /model/backbone/stage1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv2/conv/Conv
102
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv1.conv.weight + /model/backbone/stage1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv1/conv/Conv
103
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
104
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
105
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
106
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 485) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Add)
107
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
108
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
109
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
110
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 501) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Add)
111
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv3.conv.weight + /model/backbone/stage1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv3/conv/Conv
112
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_skip2.conv.weight + /model/neck/neck2/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip2/conv/Conv
113
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.downsample.rbr_reparam.weight + /model/backbone/stage2/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/downsample/rbr_reparam/Conv
114
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.downsample.conv.weight + /model/neck/neck2/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/downsample/conv/Conv
115
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv2.conv.weight + /model/backbone/stage2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv2/conv/Conv
116
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv1.conv.weight + /model/backbone/stage2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv1/conv/Conv
117
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
118
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
119
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
120
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 548) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Add)
121
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
122
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
123
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
124
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 564) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Add)
125
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
126
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
127
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
128
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.2.alpha + (Unnamed Layer* 580) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Add)
129
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/Concat_/model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Add_output_0_clone_0 copy
130
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv3.conv.weight + /model/backbone/stage2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv3/conv/Conv
131
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip2.conv.weight + /model/neck/neck1/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip2/conv/Conv || model.neck.neck2.reduce_skip1.conv.weight + /model/neck/neck2/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip1/conv/Conv
132
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.downsample.rbr_reparam.weight + /model/backbone/stage3/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/downsample/rbr_reparam/Conv
133
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.downsample.conv.weight + /model/neck/neck1/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/downsample/conv/Conv
134
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv2.conv.weight + /model/backbone/stage3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv2/conv/Conv
135
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv1.conv.weight + /model/backbone/stage3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv1/conv/Conv
136
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
137
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
138
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
139
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 630) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Add)
140
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
141
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
142
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
143
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 646) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Add)
144
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
145
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
146
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
147
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.2.alpha + (Unnamed Layer* 662) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Add)
148
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
149
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/Conv
150
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/Conv
151
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.3.alpha + (Unnamed Layer* 678) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Add)
152
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv3.conv.weight + /model/backbone/stage3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv3/conv/Conv
153
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip1.conv.weight + /model/neck/neck1/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip1/conv/Conv
154
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.downsample.rbr_reparam.weight + /model/backbone/stage4/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/downsample/rbr_reparam/Conv
155
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv2.conv.weight + /model/backbone/stage4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv2/conv/Conv
156
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv1.conv.weight + /model/backbone/stage4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv1/conv/Conv
157
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
158
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
159
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
160
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 719) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Add)
161
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
162
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
163
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
164
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 735) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Add)
165
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv3.conv.weight + /model/backbone/stage4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv3/conv/Conv
166
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv1.conv.weight + /model/backbone/context_module/cv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv1/conv/Conv
167
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.2/MaxPool
168
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.1/MaxPool
169
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.0/MaxPool
170
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/backbone/context_module/m.2/MaxPool_output_0 copy
171
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv2.conv.weight + /model/backbone/context_module/cv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv2/conv/Conv
172
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.conv.conv.weight + /model/neck/neck1/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/conv/conv/Conv
173
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck1.upsample.weight + /model/neck/neck1/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck1/upsample/ConvTranspose
174
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_after_concat.conv.weight + /model/neck/neck1/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_after_concat/conv/Conv
175
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv2.conv.weight + /model/neck/neck1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv2/conv/Conv
176
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv1.conv.weight + /model/neck/neck1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv1/conv/Conv
177
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
178
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
179
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
180
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 800) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Add)
181
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
182
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
183
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
184
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 816) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Add)
185
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/Concat_/model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
186
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv3.conv.weight + /model/neck/neck1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv3/conv/Conv
187
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.conv.conv.weight + /model/neck/neck2/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/conv/conv/Conv
188
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/upsample/_input_quantizer/QuantizeLinear
189
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck2.upsample.weight + /model/neck/neck2/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck2/upsample/ConvTranspose
190
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/Concat_/model/neck/neck2/reduce_skip1/act/Relu_output_0_clone_1 copy
191
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_after_concat.conv.weight + /model/neck/neck2/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_after_concat/conv/Conv
192
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv2.conv.weight + /model/neck/neck2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv2/conv/Conv
193
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv1.conv.weight + /model/neck/neck2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv1/conv/Conv
194
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
195
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
196
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
197
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 865) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Add)
198
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
199
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
200
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
201
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 881) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Add)
202
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv3.conv.weight + /model/neck/neck2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv3/conv/Conv
203
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.bbox_stem.seq.conv.weight + /model/heads/head1/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/bbox_stem/seq/conv/Conv || model.heads.head1.pose_stem.seq.conv.weight + /model/heads/head1/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_stem/seq/conv/Conv
204
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.conv.conv.weight + /model/neck/neck3/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/conv/conv/Conv
205
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_convs.0.seq.conv.weight + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head1.cls_convs.0.seq.conv.weight + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/Conv
206
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.0.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/Conv
207
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/conv1/conv/_input_quantizer/QuantizeLinear_clone_1
208
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.cls_pred.weight + /model/heads/head1/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_pred/Conv
209
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_pred.weight + /model/heads/head1/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_pred/Conv
210
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.1.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/Conv
211
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv2.conv.weight + /model/neck/neck3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv2/conv/Conv
212
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv1.conv.weight + /model/neck/neck3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv1/conv/Conv
213
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape + /model/heads/Transpose
214
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_pred.weight + /model/heads/head1/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_pred/Conv
215
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
216
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax
217
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
218
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv
219
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
220
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 947) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Add)
221
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
222
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
223
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
224
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 988) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Add)
225
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/Concat_/model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
226
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv3.conv.weight + /model/neck/neck3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv3/conv/Conv
227
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.bbox_stem.seq.conv.weight + /model/heads/head2/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/bbox_stem/seq/conv/Conv || model.heads.head2.pose_stem.seq.conv.weight + /model/heads/head2/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_stem/seq/conv/Conv
228
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.conv.conv.weight + /model/neck/neck4/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/conv/conv/Conv
229
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_convs.0.seq.conv.weight + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head2.cls_convs.0.seq.conv.weight + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/Conv
230
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.0.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/Conv
231
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv2.conv.weight + /model/neck/neck4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv2/conv/Conv
232
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv1.conv.weight + /model/neck/neck4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv1/conv/Conv
233
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.cls_pred.weight + /model/heads/head2/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_pred/Conv
234
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_pred.weight + /model/heads/head2/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_pred/Conv
235
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.1.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/Conv
236
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
237
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_4 + /model/heads/Transpose_3
238
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_pred.weight + /model/heads/head2/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_pred/Conv
239
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
240
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_1
241
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
242
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_1
243
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 1053) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Add)
244
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
245
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
246
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
247
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 1094) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Add)
248
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/Concat_/model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
249
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv3.conv.weight + /model/neck/neck4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv3/conv/Conv
250
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.bbox_stem.seq.conv.weight + /model/heads/head3/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/bbox_stem/seq/conv/Conv || model.heads.head3.pose_stem.seq.conv.weight + /model/heads/head3/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_stem/seq/conv/Conv
251
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_convs.0.seq.conv.weight + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head3.cls_convs.0.seq.conv.weight + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/Conv
252
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.0.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/Conv
253
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.cls_pred.weight + /model/heads/head3/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_pred/Conv
254
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_pred.weight + /model/heads/head3/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_pred/Conv
255
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.1.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/Conv
256
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_8 + /model/heads/Transpose_6
257
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.2.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/Conv
258
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_2
259
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_pred.weight + /model/heads/head3/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_pred/Conv
260
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_2
261
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice_1.../post_process/Reshape_2]}
262
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] NMS: batched_nms_238
263
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] DEVICE_TO_SHAPE_HOST: (Unnamed Layer* 1207) [NMS]_1_output[DevicetoShapeHostCopy]
264
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation2]
265
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice...graph2_/Concat_5]}
266
- [12/27/2023-20:52:58] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation3]
267
- [12/27/2023-20:53:08] [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +534, GPU +373, now: CPU 1168, GPU 3783 (MiB)
268
- [12/27/2023-20:53:10] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +82, GPU +73, now: CPU 1250, GPU 3856 (MiB)
269
- [12/27/2023-20:53:10] [I] [TRT] Local timing cache in use. Profiling results in this builder pass will not be stored.
270
- [12/27/2023-21:20:58] [I] [TRT] Total Activation Memory: 7904134656
271
- [12/27/2023-21:20:58] [I] [TRT] Detected 1 inputs and 1 output network tensors.
272
- [12/27/2023-21:21:05] [I] [TRT] Total Host Persistent Memory: 298176
273
- [12/27/2023-21:21:05] [I] [TRT] Total Device Persistent Memory: 77824
274
- [12/27/2023-21:21:05] [I] [TRT] Total Scratch Memory: 134217728
275
- [12/27/2023-21:21:05] [I] [TRT] [MemUsageStats] Peak memory usage of TRT CPU/GPU memory allocators: CPU 33 MiB, GPU 154 MiB
276
- [12/27/2023-21:21:05] [I] [TRT] [BlockAssignment] Started assigning block shifts. This will take 179 steps to complete.
277
- [12/27/2023-21:21:05] [I] [TRT] [BlockAssignment] Algorithm ShiftNTopDown took 201.747ms to assign 13 blocks to 179 nodes requiring 140788224 bytes.
278
- [12/27/2023-21:21:05] [I] [TRT] Total Activation Memory: 140788224
279
- [12/27/2023-21:21:08] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1593, GPU 4908 (MiB)
280
- [12/27/2023-21:21:08] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in building engine: CPU +6, GPU +8, now: CPU 6, GPU 8 (MiB)
281
- [12/27/2023-21:21:08] [I] Engine built in 1700.74 sec.
282
- [12/27/2023-21:21:09] [I] [TRT] Loaded engine size: 9 MiB
283
- [12/27/2023-21:21:09] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU -1, now: CPU 1244, GPU 4897 (MiB)
284
- [12/27/2023-21:21:09] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in engine deserialization: CPU +0, GPU +7, now: CPU 0, GPU 7 (MiB)
285
- [12/27/2023-21:21:09] [I] Engine deserialized in 0.189318 sec.
286
- [12/27/2023-21:21:09] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1245, GPU 4898 (MiB)
287
- [12/27/2023-21:21:09] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +0, GPU +135, now: CPU 0, GPU 142 (MiB)
288
- [12/27/2023-21:21:09] [I] Setting persistentCacheLimit to 0 bytes.
289
- [12/27/2023-21:21:09] [I] Using random values for input onnx::Cast_0
290
- [12/27/2023-21:21:09] [I] Created input binding for onnx::Cast_0 with dimensions 1x3x640x640
291
- [12/27/2023-21:21:09] [I] Using random values for output graph2_flat_predictions
292
- [12/27/2023-21:21:09] [I] Created output binding for graph2_flat_predictions with dimensions -1x57
293
- [12/27/2023-21:21:09] [I] Starting inference
294
- [12/27/2023-21:21:24] [I] Warmup completed 13 queries over 200 ms
295
- [12/27/2023-21:21:24] [I] Timing trace has 1215 queries over 15.0321 s
296
- [12/27/2023-21:21:24] [I]
297
- [12/27/2023-21:21:24] [I] === Trace details ===
298
- [12/27/2023-21:21:24] [I] Trace averages of 100 runs:
299
- [12/27/2023-21:21:24] [I] Average on 100 runs - GPU latency: 12.6305 ms - Host latency: 12.7431 ms (enqueue 12.7033 ms)
300
- [12/27/2023-21:21:24] [I] Average on 100 runs - GPU latency: 12.2402 ms - Host latency: 12.3526 ms (enqueue 12.3197 ms)
301
- [12/27/2023-21:21:24] [I] Average on 100 runs - GPU latency: 12.256 ms - Host latency: 12.3679 ms (enqueue 12.3361 ms)
302
- [12/27/2023-21:21:24] [I] Average on 100 runs - GPU latency: 12.1815 ms - Host latency: 12.2935 ms (enqueue 12.2622 ms)
303
- [12/27/2023-21:21:24] [I] Average on 100 runs - GPU latency: 12.1642 ms - Host latency: 12.276 ms (enqueue 12.2445 ms)
304
- [12/27/2023-21:21:24] [I] Average on 100 runs - GPU latency: 12.2964 ms - Host latency: 12.4083 ms (enqueue 12.3727 ms)
305
- [12/27/2023-21:21:24] [I] Average on 100 runs - GPU latency: 12.1804 ms - Host latency: 12.2922 ms (enqueue 12.2598 ms)
306
- [12/27/2023-21:21:24] [I] Average on 100 runs - GPU latency: 12.163 ms - Host latency: 12.2745 ms (enqueue 12.2423 ms)
307
- [12/27/2023-21:21:24] [I] Average on 100 runs - GPU latency: 12.173 ms - Host latency: 12.2853 ms (enqueue 12.2536 ms)
308
- [12/27/2023-21:21:24] [I] Average on 100 runs - GPU latency: 12.1462 ms - Host latency: 12.2579 ms (enqueue 12.2245 ms)
309
- [12/27/2023-21:21:24] [I] Average on 100 runs - GPU latency: 12.1697 ms - Host latency: 12.281 ms (enqueue 12.2499 ms)
310
- [12/27/2023-21:21:24] [I] Average on 100 runs - GPU latency: 12.1833 ms - Host latency: 12.2942 ms (enqueue 12.2621 ms)
311
- [12/27/2023-21:21:24] [I]
312
- [12/27/2023-21:21:24] [I] === Performance summary ===
313
- [12/27/2023-21:21:24] [I] Throughput: 80.8271 qps
314
- [12/27/2023-21:21:24] [I] Latency: min = 11.5059 ms, max = 19.759 ms, mean = 12.343 ms, median = 12.2217 ms, percentile(90%) = 12.6807 ms, percentile(95%) = 12.8347 ms, percentile(99%) = 14.0815 ms
315
- [12/27/2023-21:21:24] [I] Enqueue Time: min = 11.4648 ms, max = 21.5256 ms, mean = 12.31 ms, median = 12.1904 ms, percentile(90%) = 12.6436 ms, percentile(95%) = 12.7954 ms, percentile(99%) = 13.7207 ms
316
- [12/27/2023-21:21:24] [I] H2D Latency: min = 0.0810547 ms, max = 0.117188 ms, mean = 0.0989079 ms, median = 0.0996094 ms, percentile(90%) = 0.100586 ms, percentile(95%) = 0.100586 ms, percentile(99%) = 0.101807 ms
317
- [12/27/2023-21:21:24] [I] GPU Compute Time: min = 11.4023 ms, max = 19.6559 ms, mean = 12.2311 ms, median = 12.1094 ms, percentile(90%) = 12.564 ms, percentile(95%) = 12.7192 ms, percentile(99%) = 13.9766 ms
318
- [12/27/2023-21:21:24] [I] D2H Latency: min = 0.00292969 ms, max = 0.0380859 ms, mean = 0.0129526 ms, median = 0.0117188 ms, percentile(90%) = 0.0170898 ms, percentile(95%) = 0.019043 ms, percentile(99%) = 0.027832 ms
319
- [12/27/2023-21:21:24] [I] Total Host Walltime: 15.0321 s
320
- [12/27/2023-21:21:24] [I] Total GPU Compute Time: 14.8608 s
321
- [12/27/2023-21:21:24] [I] Explanations of the performance metrics are printed in the verbose logs.
322
- [12/27/2023-21:21:24] [I]
 
 
323
  &&&& PASSED TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_n_int8.onnx --best --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_n_int8.onnx.best.engine
 
1
  &&&& RUNNING TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_n_int8.onnx --best --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_n_int8.onnx.best.engine
2
+ [01/04/2024-14:17:23] [I] === Model Options ===
3
+ [01/04/2024-14:17:23] [I] Format: ONNX
4
+ [01/04/2024-14:17:23] [I] Model: yolo_nas_pose_n_int8.onnx
5
+ [01/04/2024-14:17:23] [I] Output:
6
+ [01/04/2024-14:17:23] [I] === Build Options ===
7
+ [01/04/2024-14:17:23] [I] Max batch: explicit batch
8
+ [01/04/2024-14:17:23] [I] Memory Pools: workspace: default, dlaSRAM: default, dlaLocalDRAM: default, dlaGlobalDRAM: default
9
+ [01/04/2024-14:17:23] [I] minTiming: 1
10
+ [01/04/2024-14:17:23] [I] avgTiming: 8
11
+ [01/04/2024-14:17:23] [I] Precision: FP32+FP16+INT8
12
+ [01/04/2024-14:17:23] [I] LayerPrecisions:
13
+ [01/04/2024-14:17:23] [I] Calibration: Dynamic
14
+ [01/04/2024-14:17:23] [I] Refit: Disabled
15
+ [01/04/2024-14:17:23] [I] Sparsity: Disabled
16
+ [01/04/2024-14:17:23] [I] Safe mode: Disabled
17
+ [01/04/2024-14:17:23] [I] DirectIO mode: Disabled
18
+ [01/04/2024-14:17:23] [I] Restricted mode: Disabled
19
+ [01/04/2024-14:17:23] [I] Build only: Disabled
20
+ [01/04/2024-14:17:23] [I] Save engine: yolo_nas_pose_n_int8.onnx.best.engine
21
+ [01/04/2024-14:17:23] [I] Load engine:
22
+ [01/04/2024-14:17:23] [I] Profiling verbosity: 0
23
+ [01/04/2024-14:17:23] [I] Tactic sources: Using default tactic sources
24
+ [01/04/2024-14:17:23] [I] timingCacheMode: local
25
+ [01/04/2024-14:17:23] [I] timingCacheFile:
26
+ [01/04/2024-14:17:23] [I] Heuristic: Disabled
27
+ [01/04/2024-14:17:23] [I] Preview Features: Use default preview flags.
28
+ [01/04/2024-14:17:23] [I] Input(s)s format: fp32:CHW
29
+ [01/04/2024-14:17:23] [I] Output(s)s format: fp32:CHW
30
+ [01/04/2024-14:17:23] [I] Input build shapes: model
31
+ [01/04/2024-14:17:23] [I] Input calibration shapes: model
32
+ [01/04/2024-14:17:23] [I] === System Options ===
33
+ [01/04/2024-14:17:23] [I] Device: 0
34
+ [01/04/2024-14:17:23] [I] DLACore:
35
+ [01/04/2024-14:17:23] [I] Plugins:
36
+ [01/04/2024-14:17:23] [I] === Inference Options ===
37
+ [01/04/2024-14:17:23] [I] Batch: Explicit
38
+ [01/04/2024-14:17:23] [I] Input inference shapes: model
39
+ [01/04/2024-14:17:23] [I] Iterations: 10
40
+ [01/04/2024-14:17:23] [I] Duration: 15s (+ 200ms warm up)
41
+ [01/04/2024-14:17:23] [I] Sleep time: 0ms
42
+ [01/04/2024-14:17:23] [I] Idle time: 0ms
43
+ [01/04/2024-14:17:23] [I] Streams: 1
44
+ [01/04/2024-14:17:23] [I] ExposeDMA: Disabled
45
+ [01/04/2024-14:17:23] [I] Data transfers: Enabled
46
+ [01/04/2024-14:17:23] [I] Spin-wait: Disabled
47
+ [01/04/2024-14:17:23] [I] Multithreading: Disabled
48
+ [01/04/2024-14:17:23] [I] CUDA Graph: Disabled
49
+ [01/04/2024-14:17:23] [I] Separate profiling: Disabled
50
+ [01/04/2024-14:17:23] [I] Time Deserialize: Disabled
51
+ [01/04/2024-14:17:23] [I] Time Refit: Disabled
52
+ [01/04/2024-14:17:23] [I] NVTX verbosity: 0
53
+ [01/04/2024-14:17:23] [I] Persistent Cache Ratio: 0
54
+ [01/04/2024-14:17:23] [I] Inputs:
55
+ [01/04/2024-14:17:23] [I] === Reporting Options ===
56
+ [01/04/2024-14:17:23] [I] Verbose: Disabled
57
+ [01/04/2024-14:17:23] [I] Averages: 100 inferences
58
+ [01/04/2024-14:17:23] [I] Percentiles: 90,95,99
59
+ [01/04/2024-14:17:23] [I] Dump refittable layers:Disabled
60
+ [01/04/2024-14:17:23] [I] Dump output: Disabled
61
+ [01/04/2024-14:17:23] [I] Profile: Disabled
62
+ [01/04/2024-14:17:23] [I] Export timing to JSON file:
63
+ [01/04/2024-14:17:23] [I] Export output to JSON file:
64
+ [01/04/2024-14:17:23] [I] Export profile to JSON file:
65
+ [01/04/2024-14:17:23] [I]
66
+ [01/04/2024-14:17:24] [I] === Device Information ===
67
+ [01/04/2024-14:17:24] [I] Selected Device: Orin
68
+ [01/04/2024-14:17:24] [I] Compute Capability: 8.7
69
+ [01/04/2024-14:17:24] [I] SMs: 8
70
+ [01/04/2024-14:17:24] [I] Compute Clock Rate: 0.624 GHz
71
+ [01/04/2024-14:17:24] [I] Device Global Memory: 7471 MiB
72
+ [01/04/2024-14:17:24] [I] Shared Memory per SM: 164 KiB
73
+ [01/04/2024-14:17:24] [I] Memory Bus Width: 128 bits (ECC disabled)
74
+ [01/04/2024-14:17:24] [I] Memory Clock Rate: 0.624 GHz
75
+ [01/04/2024-14:17:24] [I]
76
+ [01/04/2024-14:17:24] [I] TensorRT version: 8.5.2
77
+ [01/04/2024-14:17:29] [I] [TRT] [MemUsageChange] Init CUDA: CPU +220, GPU +0, now: CPU 249, GPU 2718 (MiB)
78
+ [01/04/2024-14:17:33] [I] [TRT] [MemUsageChange] Init builder kernel library: CPU +302, GPU +435, now: CPU 574, GPU 3215 (MiB)
79
+ [01/04/2024-14:17:33] [I] Start parsing network model
80
+ [01/04/2024-14:17:33] [I] [TRT] ----------------------------------------------------------------
81
+ [01/04/2024-14:17:33] [I] [TRT] Input filename: yolo_nas_pose_n_int8.onnx
82
+ [01/04/2024-14:17:33] [I] [TRT] ONNX IR version: 0.0.8
83
+ [01/04/2024-14:17:33] [I] [TRT] Opset version: 17
84
+ [01/04/2024-14:17:33] [I] [TRT] Producer name: pytorch
85
+ [01/04/2024-14:17:33] [I] [TRT] Producer version: 2.1.2
86
+ [01/04/2024-14:17:33] [I] [TRT] Domain:
87
+ [01/04/2024-14:17:33] [I] [TRT] Model version: 0
88
+ [01/04/2024-14:17:33] [I] [TRT] Doc string:
89
+ [01/04/2024-14:17:33] [I] [TRT] ----------------------------------------------------------------
90
+ [01/04/2024-14:17:36] [I] Finish parsing network model
91
+ [01/04/2024-14:17:40] [I] [TRT] ---------- Layers Running on DLA ----------
92
+ [01/04/2024-14:17:40] [I] [TRT] ---------- Layers Running on GPU ----------
93
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation1]
94
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/pre_process/pre_process.0/Cast.../pre_process/pre_process.2/Mul]}
95
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1204) [Constant]
96
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1205) [Constant]
97
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1206) [Constant]
98
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/backbone/stem/conv/rbr_reparam/_input_quantizer/QuantizeLinear
99
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stem.conv.rbr_reparam.weight + /model/backbone/stem/conv/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stem/conv/rbr_reparam/Conv
100
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.downsample.rbr_reparam.weight + /model/backbone/stage1/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/downsample/rbr_reparam/Conv
101
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv2.conv.weight + /model/backbone/stage1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv2/conv/Conv
102
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv1.conv.weight + /model/backbone/stage1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv1/conv/Conv
103
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
104
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
105
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
106
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 485) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Add)
107
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
108
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
109
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
110
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 501) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Add)
111
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv3.conv.weight + /model/backbone/stage1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv3/conv/Conv
112
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_skip2.conv.weight + /model/neck/neck2/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip2/conv/Conv
113
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.downsample.rbr_reparam.weight + /model/backbone/stage2/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/downsample/rbr_reparam/Conv
114
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.downsample.conv.weight + /model/neck/neck2/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/downsample/conv/Conv
115
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv2.conv.weight + /model/backbone/stage2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv2/conv/Conv
116
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv1.conv.weight + /model/backbone/stage2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv1/conv/Conv
117
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
118
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
119
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
120
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 548) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Add)
121
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
122
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
123
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
124
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 564) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Add)
125
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
126
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
127
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
128
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.2.alpha + (Unnamed Layer* 580) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Add)
129
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/Concat_/model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Add_output_0_clone_0 copy
130
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv3.conv.weight + /model/backbone/stage2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv3/conv/Conv
131
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip2.conv.weight + /model/neck/neck1/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip2/conv/Conv || model.neck.neck2.reduce_skip1.conv.weight + /model/neck/neck2/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip1/conv/Conv
132
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.downsample.rbr_reparam.weight + /model/backbone/stage3/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/downsample/rbr_reparam/Conv
133
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.downsample.conv.weight + /model/neck/neck1/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/downsample/conv/Conv
134
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv2.conv.weight + /model/backbone/stage3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv2/conv/Conv
135
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv1.conv.weight + /model/backbone/stage3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv1/conv/Conv
136
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
137
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
138
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
139
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 630) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Add)
140
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
141
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
142
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
143
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 646) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Add)
144
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
145
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
146
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
147
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.2.alpha + (Unnamed Layer* 662) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Add)
148
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
149
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/Conv
150
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/Conv
151
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.3.alpha + (Unnamed Layer* 678) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Add)
152
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv3.conv.weight + /model/backbone/stage3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv3/conv/Conv
153
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip1.conv.weight + /model/neck/neck1/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip1/conv/Conv
154
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.downsample.rbr_reparam.weight + /model/backbone/stage4/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/downsample/rbr_reparam/Conv
155
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv2.conv.weight + /model/backbone/stage4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv2/conv/Conv
156
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv1.conv.weight + /model/backbone/stage4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv1/conv/Conv
157
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
158
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
159
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
160
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 719) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Add)
161
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
162
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
163
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
164
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 735) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Add)
165
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv3.conv.weight + /model/backbone/stage4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv3/conv/Conv
166
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv1.conv.weight + /model/backbone/context_module/cv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv1/conv/Conv
167
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.2/MaxPool
168
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.1/MaxPool
169
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.0/MaxPool
170
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/backbone/context_module/m.2/MaxPool_output_0 copy
171
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv2.conv.weight + /model/backbone/context_module/cv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv2/conv/Conv
172
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.conv.conv.weight + /model/neck/neck1/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/conv/conv/Conv
173
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/upsample/_input_quantizer/QuantizeLinear
174
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck1.upsample.weight + /model/neck/neck1/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck1/upsample/ConvTranspose
175
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_after_concat.conv.weight + /model/neck/neck1/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_after_concat/conv/Conv
176
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv2.conv.weight + /model/neck/neck1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv2/conv/Conv
177
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv1.conv.weight + /model/neck/neck1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv1/conv/Conv
178
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
179
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
180
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
181
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 800) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Add)
182
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
183
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
184
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
185
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 816) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Add)
186
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/Concat_/model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
187
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv3.conv.weight + /model/neck/neck1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv3/conv/Conv
188
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.conv.conv.weight + /model/neck/neck2/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/conv/conv/Conv
189
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/upsample/_input_quantizer/QuantizeLinear
190
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck2.upsample.weight + /model/neck/neck2/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck2/upsample/ConvTranspose
191
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/Concat_/model/neck/neck2/reduce_skip1/act/Relu_output_0_clone_1 copy
192
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_after_concat.conv.weight + /model/neck/neck2/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_after_concat/conv/Conv
193
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv2.conv.weight + /model/neck/neck2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv2/conv/Conv
194
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv1.conv.weight + /model/neck/neck2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv1/conv/Conv
195
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
196
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
197
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
198
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 865) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Add)
199
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
200
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
201
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
202
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 881) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Add)
203
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv3.conv.weight + /model/neck/neck2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv3/conv/Conv
204
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.bbox_stem.seq.conv.weight + /model/heads/head1/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/bbox_stem/seq/conv/Conv || model.heads.head1.pose_stem.seq.conv.weight + /model/heads/head1/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_stem/seq/conv/Conv
205
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.conv.conv.weight + /model/neck/neck3/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/conv/conv/Conv
206
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_convs.0.seq.conv.weight + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head1.cls_convs.0.seq.conv.weight + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/Conv
207
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.0.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/Conv
208
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/conv1/conv/_input_quantizer/QuantizeLinear_clone_1
209
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.cls_pred.weight + /model/heads/head1/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_pred/Conv
210
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_pred.weight + /model/heads/head1/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_pred/Conv
211
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.1.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/Conv
212
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv2.conv.weight + /model/neck/neck3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv2/conv/Conv
213
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv1.conv.weight + /model/neck/neck3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv1/conv/Conv
214
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape + /model/heads/Transpose
215
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_pred.weight + /model/heads/head1/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_pred/Conv
216
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
217
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax
218
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
219
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv
220
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
221
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 947) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Add)
222
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
223
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
224
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
225
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 988) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Add)
226
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/Concat_/model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
227
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv3.conv.weight + /model/neck/neck3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv3/conv/Conv
228
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.bbox_stem.seq.conv.weight + /model/heads/head2/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/bbox_stem/seq/conv/Conv || model.heads.head2.pose_stem.seq.conv.weight + /model/heads/head2/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_stem/seq/conv/Conv
229
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.conv.conv.weight + /model/neck/neck4/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/conv/conv/Conv
230
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_convs.0.seq.conv.weight + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head2.cls_convs.0.seq.conv.weight + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/Conv
231
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.0.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/Conv
232
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/conv1/conv/_input_quantizer/QuantizeLinear_clone_1
233
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.cls_pred.weight + /model/heads/head2/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_pred/Conv
234
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_pred.weight + /model/heads/head2/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_pred/Conv
235
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.1.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/Conv
236
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv2.conv.weight + /model/neck/neck4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv2/conv/Conv
237
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv1.conv.weight + /model/neck/neck4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv1/conv/Conv
238
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_4 + /model/heads/Transpose_3
239
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_pred.weight + /model/heads/head2/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_pred/Conv
240
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
241
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_1
242
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
243
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_1
244
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
245
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 1054) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Add)
246
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
247
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
248
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
249
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 1095) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Add)
250
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/Concat_/model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
251
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv3.conv.weight + /model/neck/neck4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv3/conv/Conv
252
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.bbox_stem.seq.conv.weight + /model/heads/head3/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/bbox_stem/seq/conv/Conv || model.heads.head3.pose_stem.seq.conv.weight + /model/heads/head3/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_stem/seq/conv/Conv
253
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_convs.0.seq.conv.weight + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head3.cls_convs.0.seq.conv.weight + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/Conv
254
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.0.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/Conv
255
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.cls_pred.weight + /model/heads/head3/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_pred/Conv
256
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_pred.weight + /model/heads/head3/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_pred/Conv
257
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.1.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/Conv
258
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_8 + /model/heads/Transpose_6
259
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.2.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/Conv
260
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_2
261
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_pred.weight + /model/heads/head3/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_pred/Conv
262
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_2
263
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice_1.../post_process/Reshape_2]}
264
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] NMS: batched_nms_238
265
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] DEVICE_TO_SHAPE_HOST: (Unnamed Layer* 1208) [NMS]_1_output[DevicetoShapeHostCopy]
266
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation2]
267
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice...graph2_/Concat_5]}
268
+ [01/04/2024-14:17:40] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation3]
269
+ [01/04/2024-14:17:55] [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +534, GPU +815, now: CPU 1168, GPU 4076 (MiB)
270
+ [01/04/2024-14:17:58] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +82, GPU +138, now: CPU 1250, GPU 4214 (MiB)
271
+ [01/04/2024-14:17:58] [I] [TRT] Local timing cache in use. Profiling results in this builder pass will not be stored.
272
+ [01/04/2024-14:45:51] [I] [TRT] Total Activation Memory: 7900558848
273
+ [01/04/2024-14:45:51] [I] [TRT] Detected 1 inputs and 1 output network tensors.
274
+ [01/04/2024-14:45:58] [I] [TRT] Total Host Persistent Memory: 300704
275
+ [01/04/2024-14:45:58] [I] [TRT] Total Device Persistent Memory: 116736
276
+ [01/04/2024-14:45:58] [I] [TRT] Total Scratch Memory: 134217728
277
+ [01/04/2024-14:45:58] [I] [TRT] [MemUsageStats] Peak memory usage of TRT CPU/GPU memory allocators: CPU 33 MiB, GPU 154 MiB
278
+ [01/04/2024-14:45:58] [I] [TRT] [BlockAssignment] Started assigning block shifts. This will take 176 steps to complete.
279
+ [01/04/2024-14:45:58] [I] [TRT] [BlockAssignment] Algorithm ShiftNTopDown took 60.5874ms to assign 13 blocks to 176 nodes requiring 140788224 bytes.
280
+ [01/04/2024-14:45:58] [I] [TRT] Total Activation Memory: 140788224
281
+ [01/04/2024-14:46:01] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1593, GPU 5386 (MiB)
282
+ [01/04/2024-14:46:01] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in building engine: CPU +6, GPU +8, now: CPU 6, GPU 8 (MiB)
283
+ [01/04/2024-14:46:01] [I] Engine built in 1717.31 sec.
284
+ [01/04/2024-14:46:01] [I] [TRT] Loaded engine size: 9 MiB
285
+ [01/04/2024-14:46:02] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1244, GPU 5378 (MiB)
286
+ [01/04/2024-14:46:02] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in engine deserialization: CPU +0, GPU +7, now: CPU 0, GPU 7 (MiB)
287
+ [01/04/2024-14:46:02] [I] Engine deserialized in 0.18586 sec.
288
+ [01/04/2024-14:46:02] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU -1, now: CPU 1245, GPU 5378 (MiB)
289
+ [01/04/2024-14:46:02] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +0, GPU +135, now: CPU 0, GPU 142 (MiB)
290
+ [01/04/2024-14:46:02] [I] Setting persistentCacheLimit to 0 bytes.
291
+ [01/04/2024-14:46:02] [I] Using random values for input onnx::Cast_0
292
+ [01/04/2024-14:46:02] [I] Created input binding for onnx::Cast_0 with dimensions 1x3x640x640
293
+ [01/04/2024-14:46:02] [I] Using random values for output graph2_flat_predictions
294
+ [01/04/2024-14:46:02] [I] Created output binding for graph2_flat_predictions with dimensions -1x57
295
+ [01/04/2024-14:46:02] [I] Starting inference
296
+ [01/04/2024-14:46:17] [I] Warmup completed 3 queries over 200 ms
297
+ [01/04/2024-14:46:17] [I] Timing trace has 1232 queries over 15.0315 s
298
+ [01/04/2024-14:46:17] [I]
299
+ [01/04/2024-14:46:17] [I] === Trace details ===
300
+ [01/04/2024-14:46:17] [I] Trace averages of 100 runs:
301
+ [01/04/2024-14:46:17] [I] Average on 100 runs - GPU latency: 12.29 ms - Host latency: 12.4041 ms (enqueue 12.3732 ms)
302
+ [01/04/2024-14:46:17] [I] Average on 100 runs - GPU latency: 12.0468 ms - Host latency: 12.1594 ms (enqueue 12.1269 ms)
303
+ [01/04/2024-14:46:17] [I] Average on 100 runs - GPU latency: 12.0958 ms - Host latency: 12.2081 ms (enqueue 12.1741 ms)
304
+ [01/04/2024-14:46:17] [I] Average on 100 runs - GPU latency: 12.0328 ms - Host latency: 12.1444 ms (enqueue 12.1128 ms)
305
+ [01/04/2024-14:46:17] [I] Average on 100 runs - GPU latency: 12.0594 ms - Host latency: 12.1718 ms (enqueue 12.1396 ms)
306
+ [01/04/2024-14:46:17] [I] Average on 100 runs - GPU latency: 12.065 ms - Host latency: 12.1767 ms (enqueue 12.145 ms)
307
+ [01/04/2024-14:46:17] [I] Average on 100 runs - GPU latency: 12.0382 ms - Host latency: 12.1502 ms (enqueue 12.1181 ms)
308
+ [01/04/2024-14:46:17] [I] Average on 100 runs - GPU latency: 12.0485 ms - Host latency: 12.1607 ms (enqueue 12.1285 ms)
309
+ [01/04/2024-14:46:17] [I] Average on 100 runs - GPU latency: 12.0218 ms - Host latency: 12.1333 ms (enqueue 12.1027 ms)
310
+ [01/04/2024-14:46:17] [I] Average on 100 runs - GPU latency: 11.9903 ms - Host latency: 12.1026 ms (enqueue 12.0704 ms)
311
+ [01/04/2024-14:46:17] [I] Average on 100 runs - GPU latency: 11.9893 ms - Host latency: 12.1013 ms (enqueue 12.0701 ms)
312
+ [01/04/2024-14:46:17] [I] Average on 100 runs - GPU latency: 12.0489 ms - Host latency: 12.1609 ms (enqueue 12.1343 ms)
313
+ [01/04/2024-14:46:17] [I]
314
+ [01/04/2024-14:46:17] [I] === Performance summary ===
315
+ [01/04/2024-14:46:17] [I] Throughput: 81.9611 qps
316
+ [01/04/2024-14:46:17] [I] Latency: min = 11.582 ms, max = 16.703 ms, mean = 12.1721 ms, median = 12.1212 ms, percentile(90%) = 12.4434 ms, percentile(95%) = 12.5928 ms, percentile(99%) = 13.4438 ms
317
+ [01/04/2024-14:46:17] [I] Enqueue Time: min = 11.5547 ms, max = 16.675 ms, mean = 12.1407 ms, median = 12.0908 ms, percentile(90%) = 12.4111 ms, percentile(95%) = 12.5623 ms, percentile(99%) = 13.3818 ms
318
+ [01/04/2024-14:46:17] [I] H2D Latency: min = 0.0830078 ms, max = 0.146088 ms, mean = 0.1 ms, median = 0.0996094 ms, percentile(90%) = 0.10083 ms, percentile(95%) = 0.101562 ms, percentile(99%) = 0.115234 ms
319
+ [01/04/2024-14:46:17] [I] GPU Compute Time: min = 11.4717 ms, max = 16.5476 ms, mean = 12.0599 ms, median = 12.0098 ms, percentile(90%) = 12.332 ms, percentile(95%) = 12.4795 ms, percentile(99%) = 13.3123 ms
320
+ [01/04/2024-14:46:17] [I] D2H Latency: min = 0.00415039 ms, max = 0.0344238 ms, mean = 0.0122015 ms, median = 0.0117188 ms, percentile(90%) = 0.0146484 ms, percentile(95%) = 0.0159912 ms, percentile(99%) = 0.0258789 ms
321
+ [01/04/2024-14:46:17] [I] Total Host Walltime: 15.0315 s
322
+ [01/04/2024-14:46:17] [I] Total GPU Compute Time: 14.8578 s
323
+ [01/04/2024-14:46:17] [I] Explanations of the performance metrics are printed in the verbose logs.
324
+ [01/04/2024-14:46:17] [I]
325
  &&&& PASSED TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_n_int8.onnx --best --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_n_int8.onnx.best.engine
yolo_nas_pose_n_int8.onnx.int8.engine CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1235190611d52c7c4bb3e4edb7652062a0e7c08bf3d721d882d72cef0b6eb7cc
3
- size 10088004
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b19fa98d82cf9039bb036f74f7f3f994127f88c46b7bfab06353cf82e5f8dc09
3
+ size 10115566
yolo_nas_pose_n_int8.onnx.int8.engine.err CHANGED
@@ -1,7 +1,7 @@
1
- [12/27/2023-21:21:30] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
- [12/27/2023-21:21:30] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
- [12/27/2023-21:21:33] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
- [12/27/2023-21:30:30] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
- [12/27/2023-21:30:30] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
- [12/27/2023-21:30:30] [W] * GPU compute time is unstable, with coefficient of variance = 5.79476%.
7
- [12/27/2023-21:30:30] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
 
1
+ [01/04/2024-14:46:24] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
+ [01/04/2024-14:46:24] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
+ [01/04/2024-14:46:27] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
+ [01/04/2024-14:55:32] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
+ [01/04/2024-14:55:32] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
+ [01/04/2024-14:55:32] [W] * GPU compute time is unstable, with coefficient of variance = 4.3012%.
7
+ [01/04/2024-14:55:32] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
yolo_nas_pose_n_int8.onnx.int8.engine.log CHANGED
@@ -1,323 +1,325 @@
1
  &&&& RUNNING TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_n_int8.onnx --int8 --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_n_int8.onnx.int8.engine
2
- [12/27/2023-21:21:26] [I] === Model Options ===
3
- [12/27/2023-21:21:26] [I] Format: ONNX
4
- [12/27/2023-21:21:26] [I] Model: yolo_nas_pose_n_int8.onnx
5
- [12/27/2023-21:21:26] [I] Output:
6
- [12/27/2023-21:21:26] [I] === Build Options ===
7
- [12/27/2023-21:21:26] [I] Max batch: explicit batch
8
- [12/27/2023-21:21:26] [I] Memory Pools: workspace: default, dlaSRAM: default, dlaLocalDRAM: default, dlaGlobalDRAM: default
9
- [12/27/2023-21:21:26] [I] minTiming: 1
10
- [12/27/2023-21:21:26] [I] avgTiming: 8
11
- [12/27/2023-21:21:26] [I] Precision: FP32+INT8
12
- [12/27/2023-21:21:26] [I] LayerPrecisions:
13
- [12/27/2023-21:21:26] [I] Calibration: Dynamic
14
- [12/27/2023-21:21:26] [I] Refit: Disabled
15
- [12/27/2023-21:21:26] [I] Sparsity: Disabled
16
- [12/27/2023-21:21:26] [I] Safe mode: Disabled
17
- [12/27/2023-21:21:26] [I] DirectIO mode: Disabled
18
- [12/27/2023-21:21:26] [I] Restricted mode: Disabled
19
- [12/27/2023-21:21:26] [I] Build only: Disabled
20
- [12/27/2023-21:21:26] [I] Save engine: yolo_nas_pose_n_int8.onnx.int8.engine
21
- [12/27/2023-21:21:26] [I] Load engine:
22
- [12/27/2023-21:21:26] [I] Profiling verbosity: 0
23
- [12/27/2023-21:21:26] [I] Tactic sources: Using default tactic sources
24
- [12/27/2023-21:21:26] [I] timingCacheMode: local
25
- [12/27/2023-21:21:26] [I] timingCacheFile:
26
- [12/27/2023-21:21:26] [I] Heuristic: Disabled
27
- [12/27/2023-21:21:26] [I] Preview Features: Use default preview flags.
28
- [12/27/2023-21:21:26] [I] Input(s)s format: fp32:CHW
29
- [12/27/2023-21:21:26] [I] Output(s)s format: fp32:CHW
30
- [12/27/2023-21:21:26] [I] Input build shapes: model
31
- [12/27/2023-21:21:26] [I] Input calibration shapes: model
32
- [12/27/2023-21:21:26] [I] === System Options ===
33
- [12/27/2023-21:21:26] [I] Device: 0
34
- [12/27/2023-21:21:26] [I] DLACore:
35
- [12/27/2023-21:21:26] [I] Plugins:
36
- [12/27/2023-21:21:26] [I] === Inference Options ===
37
- [12/27/2023-21:21:26] [I] Batch: Explicit
38
- [12/27/2023-21:21:26] [I] Input inference shapes: model
39
- [12/27/2023-21:21:26] [I] Iterations: 10
40
- [12/27/2023-21:21:26] [I] Duration: 15s (+ 200ms warm up)
41
- [12/27/2023-21:21:26] [I] Sleep time: 0ms
42
- [12/27/2023-21:21:26] [I] Idle time: 0ms
43
- [12/27/2023-21:21:26] [I] Streams: 1
44
- [12/27/2023-21:21:26] [I] ExposeDMA: Disabled
45
- [12/27/2023-21:21:26] [I] Data transfers: Enabled
46
- [12/27/2023-21:21:26] [I] Spin-wait: Disabled
47
- [12/27/2023-21:21:26] [I] Multithreading: Disabled
48
- [12/27/2023-21:21:26] [I] CUDA Graph: Disabled
49
- [12/27/2023-21:21:26] [I] Separate profiling: Disabled
50
- [12/27/2023-21:21:26] [I] Time Deserialize: Disabled
51
- [12/27/2023-21:21:26] [I] Time Refit: Disabled
52
- [12/27/2023-21:21:26] [I] NVTX verbosity: 0
53
- [12/27/2023-21:21:26] [I] Persistent Cache Ratio: 0
54
- [12/27/2023-21:21:26] [I] Inputs:
55
- [12/27/2023-21:21:26] [I] === Reporting Options ===
56
- [12/27/2023-21:21:26] [I] Verbose: Disabled
57
- [12/27/2023-21:21:26] [I] Averages: 100 inferences
58
- [12/27/2023-21:21:26] [I] Percentiles: 90,95,99
59
- [12/27/2023-21:21:26] [I] Dump refittable layers:Disabled
60
- [12/27/2023-21:21:26] [I] Dump output: Disabled
61
- [12/27/2023-21:21:26] [I] Profile: Disabled
62
- [12/27/2023-21:21:26] [I] Export timing to JSON file:
63
- [12/27/2023-21:21:26] [I] Export output to JSON file:
64
- [12/27/2023-21:21:26] [I] Export profile to JSON file:
65
- [12/27/2023-21:21:26] [I]
66
- [12/27/2023-21:21:26] [I] === Device Information ===
67
- [12/27/2023-21:21:26] [I] Selected Device: Orin
68
- [12/27/2023-21:21:26] [I] Compute Capability: 8.7
69
- [12/27/2023-21:21:26] [I] SMs: 8
70
- [12/27/2023-21:21:26] [I] Compute Clock Rate: 0.624 GHz
71
- [12/27/2023-21:21:26] [I] Device Global Memory: 7471 MiB
72
- [12/27/2023-21:21:26] [I] Shared Memory per SM: 164 KiB
73
- [12/27/2023-21:21:26] [I] Memory Bus Width: 128 bits (ECC disabled)
74
- [12/27/2023-21:21:26] [I] Memory Clock Rate: 0.624 GHz
75
- [12/27/2023-21:21:26] [I]
76
- [12/27/2023-21:21:26] [I] TensorRT version: 8.5.2
77
- [12/27/2023-21:21:27] [I] [TRT] [MemUsageChange] Init CUDA: CPU +220, GPU +0, now: CPU 249, GPU 3026 (MiB)
78
- [12/27/2023-21:21:30] [I] [TRT] [MemUsageChange] Init builder kernel library: CPU +302, GPU +281, now: CPU 574, GPU 3329 (MiB)
79
- [12/27/2023-21:21:30] [I] Start parsing network model
80
- [12/27/2023-21:21:30] [I] [TRT] ----------------------------------------------------------------
81
- [12/27/2023-21:21:30] [I] [TRT] Input filename: yolo_nas_pose_n_int8.onnx
82
- [12/27/2023-21:21:30] [I] [TRT] ONNX IR version: 0.0.8
83
- [12/27/2023-21:21:30] [I] [TRT] Opset version: 17
84
- [12/27/2023-21:21:30] [I] [TRT] Producer name: pytorch
85
- [12/27/2023-21:21:30] [I] [TRT] Producer version: 2.1.2
86
- [12/27/2023-21:21:30] [I] [TRT] Domain:
87
- [12/27/2023-21:21:30] [I] [TRT] Model version: 0
88
- [12/27/2023-21:21:30] [I] [TRT] Doc string:
89
- [12/27/2023-21:21:30] [I] [TRT] ----------------------------------------------------------------
90
- [12/27/2023-21:21:32] [I] Finish parsing network model
91
- [12/27/2023-21:21:32] [I] FP32 and INT8 precisions have been specified - more performance might be enabled by additionally specifying --fp16 or --best
92
- [12/27/2023-21:21:36] [I] [TRT] ---------- Layers Running on DLA ----------
93
- [12/27/2023-21:21:36] [I] [TRT] ---------- Layers Running on GPU ----------
94
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation1]
95
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/pre_process/pre_process.0/Cast.../pre_process/pre_process.2/Mul]}
96
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1203) [Constant]
97
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1204) [Constant]
98
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1205) [Constant]
99
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/backbone/stem/conv/rbr_reparam/_input_quantizer/QuantizeLinear
100
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stem.conv.rbr_reparam.weight + /model/backbone/stem/conv/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stem/conv/rbr_reparam/Conv
101
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.downsample.rbr_reparam.weight + /model/backbone/stage1/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/downsample/rbr_reparam/Conv
102
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv2.conv.weight + /model/backbone/stage1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv2/conv/Conv
103
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv1.conv.weight + /model/backbone/stage1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv1/conv/Conv
104
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
105
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
106
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
107
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 485) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Add)
108
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
109
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
110
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
111
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 501) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Add)
112
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv3.conv.weight + /model/backbone/stage1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv3/conv/Conv
113
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_skip2.conv.weight + /model/neck/neck2/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip2/conv/Conv
114
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.downsample.rbr_reparam.weight + /model/backbone/stage2/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/downsample/rbr_reparam/Conv
115
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.downsample.conv.weight + /model/neck/neck2/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/downsample/conv/Conv
116
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv2.conv.weight + /model/backbone/stage2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv2/conv/Conv
117
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv1.conv.weight + /model/backbone/stage2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv1/conv/Conv
118
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
119
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
120
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
121
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 548) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Add)
122
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
123
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
124
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
125
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 564) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Add)
126
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
127
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
128
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
129
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.2.alpha + (Unnamed Layer* 580) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Add)
130
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/Concat_/model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Add_output_0_clone_0 copy
131
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv3.conv.weight + /model/backbone/stage2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv3/conv/Conv
132
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip2.conv.weight + /model/neck/neck1/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip2/conv/Conv || model.neck.neck2.reduce_skip1.conv.weight + /model/neck/neck2/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip1/conv/Conv
133
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.downsample.rbr_reparam.weight + /model/backbone/stage3/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/downsample/rbr_reparam/Conv
134
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.downsample.conv.weight + /model/neck/neck1/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/downsample/conv/Conv
135
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv2.conv.weight + /model/backbone/stage3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv2/conv/Conv
136
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv1.conv.weight + /model/backbone/stage3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv1/conv/Conv
137
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
138
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
139
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
140
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 630) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Add)
141
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
142
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
143
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
144
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 646) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Add)
145
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
146
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
147
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
148
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.2.alpha + (Unnamed Layer* 662) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Add)
149
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
150
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/Conv
151
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/Conv
152
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.3.alpha + (Unnamed Layer* 678) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Add)
153
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv3.conv.weight + /model/backbone/stage3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv3/conv/Conv
154
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip1.conv.weight + /model/neck/neck1/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip1/conv/Conv
155
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.downsample.rbr_reparam.weight + /model/backbone/stage4/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/downsample/rbr_reparam/Conv
156
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv2.conv.weight + /model/backbone/stage4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv2/conv/Conv
157
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv1.conv.weight + /model/backbone/stage4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv1/conv/Conv
158
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
159
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
160
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
161
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 719) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Add)
162
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
163
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
164
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
165
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 735) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Add)
166
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv3.conv.weight + /model/backbone/stage4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv3/conv/Conv
167
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv1.conv.weight + /model/backbone/context_module/cv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv1/conv/Conv
168
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.2/MaxPool
169
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.1/MaxPool
170
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.0/MaxPool
171
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/backbone/context_module/m.2/MaxPool_output_0 copy
172
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv2.conv.weight + /model/backbone/context_module/cv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv2/conv/Conv
173
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.conv.conv.weight + /model/neck/neck1/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/conv/conv/Conv
174
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck1.upsample.weight + /model/neck/neck1/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck1/upsample/ConvTranspose
175
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_after_concat.conv.weight + /model/neck/neck1/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_after_concat/conv/Conv
176
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv2.conv.weight + /model/neck/neck1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv2/conv/Conv
177
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv1.conv.weight + /model/neck/neck1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv1/conv/Conv
178
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
179
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
180
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
181
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 800) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Add)
182
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
183
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
184
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
185
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 816) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Add)
186
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/Concat_/model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
187
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv3.conv.weight + /model/neck/neck1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv3/conv/Conv
188
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.conv.conv.weight + /model/neck/neck2/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/conv/conv/Conv
189
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/upsample/_input_quantizer/QuantizeLinear
190
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck2.upsample.weight + /model/neck/neck2/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck2/upsample/ConvTranspose
191
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/Concat_/model/neck/neck2/reduce_skip1/act/Relu_output_0_clone_1 copy
192
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_after_concat.conv.weight + /model/neck/neck2/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_after_concat/conv/Conv
193
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv2.conv.weight + /model/neck/neck2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv2/conv/Conv
194
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv1.conv.weight + /model/neck/neck2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv1/conv/Conv
195
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
196
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
197
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
198
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 865) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Add)
199
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
200
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
201
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
202
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 881) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Add)
203
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv3.conv.weight + /model/neck/neck2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv3/conv/Conv
204
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.bbox_stem.seq.conv.weight + /model/heads/head1/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/bbox_stem/seq/conv/Conv || model.heads.head1.pose_stem.seq.conv.weight + /model/heads/head1/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_stem/seq/conv/Conv
205
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.conv.conv.weight + /model/neck/neck3/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/conv/conv/Conv
206
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_convs.0.seq.conv.weight + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head1.cls_convs.0.seq.conv.weight + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/Conv
207
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.0.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/Conv
208
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/conv1/conv/_input_quantizer/QuantizeLinear_clone_1
209
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.cls_pred.weight + /model/heads/head1/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_pred/Conv
210
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_pred.weight + /model/heads/head1/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_pred/Conv
211
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.1.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/Conv
212
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv2.conv.weight + /model/neck/neck3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv2/conv/Conv
213
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv1.conv.weight + /model/neck/neck3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv1/conv/Conv
214
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape + /model/heads/Transpose
215
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_pred.weight + /model/heads/head1/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_pred/Conv
216
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
217
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax
218
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
219
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv
220
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
221
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 947) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Add)
222
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
223
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
224
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
225
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 988) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Add)
226
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/Concat_/model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
227
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv3.conv.weight + /model/neck/neck3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv3/conv/Conv
228
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.bbox_stem.seq.conv.weight + /model/heads/head2/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/bbox_stem/seq/conv/Conv || model.heads.head2.pose_stem.seq.conv.weight + /model/heads/head2/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_stem/seq/conv/Conv
229
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.conv.conv.weight + /model/neck/neck4/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/conv/conv/Conv
230
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_convs.0.seq.conv.weight + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head2.cls_convs.0.seq.conv.weight + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/Conv
231
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.0.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/Conv
232
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv2.conv.weight + /model/neck/neck4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv2/conv/Conv
233
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv1.conv.weight + /model/neck/neck4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv1/conv/Conv
234
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.cls_pred.weight + /model/heads/head2/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_pred/Conv
235
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_pred.weight + /model/heads/head2/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_pred/Conv
236
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.1.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/Conv
237
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
238
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_4 + /model/heads/Transpose_3
239
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_pred.weight + /model/heads/head2/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_pred/Conv
240
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
241
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_1
242
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
243
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_1
244
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 1053) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Add)
245
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
246
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
247
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
248
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 1094) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Add)
249
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/Concat_/model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
250
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv3.conv.weight + /model/neck/neck4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv3/conv/Conv
251
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.bbox_stem.seq.conv.weight + /model/heads/head3/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/bbox_stem/seq/conv/Conv || model.heads.head3.pose_stem.seq.conv.weight + /model/heads/head3/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_stem/seq/conv/Conv
252
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_convs.0.seq.conv.weight + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head3.cls_convs.0.seq.conv.weight + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/Conv
253
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.0.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/Conv
254
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.cls_pred.weight + /model/heads/head3/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_pred/Conv
255
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_pred.weight + /model/heads/head3/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_pred/Conv
256
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.1.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/Conv
257
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_8 + /model/heads/Transpose_6
258
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.2.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/Conv
259
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_2
260
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_pred.weight + /model/heads/head3/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_pred/Conv
261
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_2
262
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice_1.../post_process/Reshape_2]}
263
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] NMS: batched_nms_238
264
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] DEVICE_TO_SHAPE_HOST: (Unnamed Layer* 1207) [NMS]_1_output[DevicetoShapeHostCopy]
265
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation2]
266
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice...graph2_/Concat_5]}
267
- [12/27/2023-21:21:36] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation3]
268
- [12/27/2023-21:21:37] [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +534, GPU +498, now: CPU 1168, GPU 3892 (MiB)
269
- [12/27/2023-21:21:38] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +82, GPU +61, now: CPU 1250, GPU 3953 (MiB)
270
- [12/27/2023-21:21:38] [I] [TRT] Local timing cache in use. Profiling results in this builder pass will not be stored.
271
- [12/27/2023-21:30:13] [I] [TRT] Total Activation Memory: 7918594560
272
- [12/27/2023-21:30:13] [I] [TRT] Detected 1 inputs and 1 output network tensors.
273
- [12/27/2023-21:30:14] [I] [TRT] Total Host Persistent Memory: 299584
274
- [12/27/2023-21:30:14] [I] [TRT] Total Device Persistent Memory: 116736
275
- [12/27/2023-21:30:14] [I] [TRT] Total Scratch Memory: 134217728
276
- [12/27/2023-21:30:14] [I] [TRT] [MemUsageStats] Peak memory usage of TRT CPU/GPU memory allocators: CPU 33 MiB, GPU 132 MiB
277
- [12/27/2023-21:30:14] [I] [TRT] [BlockAssignment] Started assigning block shifts. This will take 180 steps to complete.
278
- [12/27/2023-21:30:14] [I] [TRT] [BlockAssignment] Algorithm ShiftNTopDown took 39.0902ms to assign 13 blocks to 180 nodes requiring 144065024 bytes.
279
- [12/27/2023-21:30:14] [I] [TRT] Total Activation Memory: 144065024
280
- [12/27/2023-21:30:14] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1590, GPU 4967 (MiB)
281
- [12/27/2023-21:30:14] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in building engine: CPU +6, GPU +8, now: CPU 6, GPU 8 (MiB)
282
- [12/27/2023-21:30:15] [I] Engine built in 528.25 sec.
283
- [12/27/2023-21:30:15] [I] [TRT] Loaded engine size: 9 MiB
284
- [12/27/2023-21:30:15] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1240, GPU 4970 (MiB)
285
- [12/27/2023-21:30:15] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in engine deserialization: CPU +0, GPU +7, now: CPU 0, GPU 7 (MiB)
286
- [12/27/2023-21:30:15] [I] Engine deserialized in 0.110555 sec.
287
- [12/27/2023-21:30:15] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1241, GPU 4970 (MiB)
288
- [12/27/2023-21:30:15] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +0, GPU +138, now: CPU 0, GPU 145 (MiB)
289
- [12/27/2023-21:30:15] [I] Setting persistentCacheLimit to 0 bytes.
290
- [12/27/2023-21:30:15] [I] Using random values for input onnx::Cast_0
291
- [12/27/2023-21:30:15] [I] Created input binding for onnx::Cast_0 with dimensions 1x3x640x640
292
- [12/27/2023-21:30:15] [I] Using random values for output graph2_flat_predictions
293
- [12/27/2023-21:30:15] [I] Created output binding for graph2_flat_predictions with dimensions -1x57
294
- [12/27/2023-21:30:15] [I] Starting inference
295
- [12/27/2023-21:30:30] [I] Warmup completed 11 queries over 200 ms
296
- [12/27/2023-21:30:30] [I] Timing trace has 1116 queries over 15.0271 s
297
- [12/27/2023-21:30:30] [I]
298
- [12/27/2023-21:30:30] [I] === Trace details ===
299
- [12/27/2023-21:30:30] [I] Trace averages of 100 runs:
300
- [12/27/2023-21:30:30] [I] Average on 100 runs - GPU latency: 13.303 ms - Host latency: 13.4154 ms (enqueue 13.3824 ms)
301
- [12/27/2023-21:30:30] [I] Average on 100 runs - GPU latency: 13.1087 ms - Host latency: 13.2205 ms (enqueue 13.189 ms)
302
- [12/27/2023-21:30:30] [I] Average on 100 runs - GPU latency: 13.1353 ms - Host latency: 13.2476 ms (enqueue 13.2151 ms)
303
- [12/27/2023-21:30:30] [I] Average on 100 runs - GPU latency: 13.1584 ms - Host latency: 13.27 ms (enqueue 13.2379 ms)
304
- [12/27/2023-21:30:30] [I] Average on 100 runs - GPU latency: 13.3766 ms - Host latency: 13.4898 ms (enqueue 13.4566 ms)
305
- [12/27/2023-21:30:30] [I] Average on 100 runs - GPU latency: 13.1268 ms - Host latency: 13.2384 ms (enqueue 13.2038 ms)
306
- [12/27/2023-21:30:30] [I] Average on 100 runs - GPU latency: 13.4482 ms - Host latency: 13.5607 ms (enqueue 13.5189 ms)
307
- [12/27/2023-21:30:30] [I] Average on 100 runs - GPU latency: 13.5223 ms - Host latency: 13.637 ms (enqueue 13.5947 ms)
308
- [12/27/2023-21:30:30] [I] Average on 100 runs - GPU latency: 13.7048 ms - Host latency: 13.8213 ms (enqueue 13.7771 ms)
309
- [12/27/2023-21:30:30] [I] Average on 100 runs - GPU latency: 13.5633 ms - Host latency: 13.6782 ms (enqueue 13.6358 ms)
310
- [12/27/2023-21:30:30] [I] Average on 100 runs - GPU latency: 13.1192 ms - Host latency: 13.2312 ms (enqueue 13.1992 ms)
311
- [12/27/2023-21:30:30] [I]
312
- [12/27/2023-21:30:30] [I] === Performance summary ===
313
- [12/27/2023-21:30:30] [I] Throughput: 74.2658 qps
314
- [12/27/2023-21:30:30] [I] Latency: min = 12.3477 ms, max = 28.8486 ms, mean = 13.4348 ms, median = 13.165 ms, percentile(90%) = 13.8916 ms, percentile(95%) = 13.9805 ms, percentile(99%) = 16.4219 ms
315
- [12/27/2023-21:30:30] [I] Enqueue Time: min = 12.3198 ms, max = 28.8143 ms, mean = 13.3985 ms, median = 13.1329 ms, percentile(90%) = 13.8477 ms, percentile(95%) = 13.9395 ms, percentile(99%) = 16.3662 ms
316
- [12/27/2023-21:30:30] [I] H2D Latency: min = 0.0810547 ms, max = 0.185547 ms, mean = 0.0987979 ms, median = 0.0986328 ms, percentile(90%) = 0.100586 ms, percentile(95%) = 0.100586 ms, percentile(99%) = 0.111328 ms
317
- [12/27/2023-21:30:30] [I] GPU Compute Time: min = 12.2375 ms, max = 28.7328 ms, mean = 13.3217 ms, median = 13.0537 ms, percentile(90%) = 13.7734 ms, percentile(95%) = 13.8643 ms, percentile(99%) = 16.2988 ms
318
- [12/27/2023-21:30:30] [I] D2H Latency: min = 0.00292969 ms, max = 0.0380859 ms, mean = 0.0142309 ms, median = 0.0126953 ms, percentile(90%) = 0.0195312 ms, percentile(95%) = 0.0214844 ms, percentile(99%) = 0.0302734 ms
319
- [12/27/2023-21:30:30] [I] Total Host Walltime: 15.0271 s
320
- [12/27/2023-21:30:30] [I] Total GPU Compute Time: 14.8671 s
321
- [12/27/2023-21:30:30] [I] Explanations of the performance metrics are printed in the verbose logs.
322
- [12/27/2023-21:30:30] [I]
 
 
323
  &&&& PASSED TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_n_int8.onnx --int8 --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_n_int8.onnx.int8.engine
 
1
  &&&& RUNNING TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_n_int8.onnx --int8 --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_n_int8.onnx.int8.engine
2
+ [01/04/2024-14:46:20] [I] === Model Options ===
3
+ [01/04/2024-14:46:20] [I] Format: ONNX
4
+ [01/04/2024-14:46:20] [I] Model: yolo_nas_pose_n_int8.onnx
5
+ [01/04/2024-14:46:20] [I] Output:
6
+ [01/04/2024-14:46:20] [I] === Build Options ===
7
+ [01/04/2024-14:46:20] [I] Max batch: explicit batch
8
+ [01/04/2024-14:46:20] [I] Memory Pools: workspace: default, dlaSRAM: default, dlaLocalDRAM: default, dlaGlobalDRAM: default
9
+ [01/04/2024-14:46:20] [I] minTiming: 1
10
+ [01/04/2024-14:46:20] [I] avgTiming: 8
11
+ [01/04/2024-14:46:20] [I] Precision: FP32+INT8
12
+ [01/04/2024-14:46:20] [I] LayerPrecisions:
13
+ [01/04/2024-14:46:20] [I] Calibration: Dynamic
14
+ [01/04/2024-14:46:20] [I] Refit: Disabled
15
+ [01/04/2024-14:46:20] [I] Sparsity: Disabled
16
+ [01/04/2024-14:46:20] [I] Safe mode: Disabled
17
+ [01/04/2024-14:46:20] [I] DirectIO mode: Disabled
18
+ [01/04/2024-14:46:20] [I] Restricted mode: Disabled
19
+ [01/04/2024-14:46:20] [I] Build only: Disabled
20
+ [01/04/2024-14:46:20] [I] Save engine: yolo_nas_pose_n_int8.onnx.int8.engine
21
+ [01/04/2024-14:46:20] [I] Load engine:
22
+ [01/04/2024-14:46:20] [I] Profiling verbosity: 0
23
+ [01/04/2024-14:46:20] [I] Tactic sources: Using default tactic sources
24
+ [01/04/2024-14:46:20] [I] timingCacheMode: local
25
+ [01/04/2024-14:46:20] [I] timingCacheFile:
26
+ [01/04/2024-14:46:20] [I] Heuristic: Disabled
27
+ [01/04/2024-14:46:20] [I] Preview Features: Use default preview flags.
28
+ [01/04/2024-14:46:20] [I] Input(s)s format: fp32:CHW
29
+ [01/04/2024-14:46:20] [I] Output(s)s format: fp32:CHW
30
+ [01/04/2024-14:46:20] [I] Input build shapes: model
31
+ [01/04/2024-14:46:20] [I] Input calibration shapes: model
32
+ [01/04/2024-14:46:20] [I] === System Options ===
33
+ [01/04/2024-14:46:20] [I] Device: 0
34
+ [01/04/2024-14:46:20] [I] DLACore:
35
+ [01/04/2024-14:46:20] [I] Plugins:
36
+ [01/04/2024-14:46:20] [I] === Inference Options ===
37
+ [01/04/2024-14:46:20] [I] Batch: Explicit
38
+ [01/04/2024-14:46:20] [I] Input inference shapes: model
39
+ [01/04/2024-14:46:20] [I] Iterations: 10
40
+ [01/04/2024-14:46:20] [I] Duration: 15s (+ 200ms warm up)
41
+ [01/04/2024-14:46:20] [I] Sleep time: 0ms
42
+ [01/04/2024-14:46:20] [I] Idle time: 0ms
43
+ [01/04/2024-14:46:20] [I] Streams: 1
44
+ [01/04/2024-14:46:20] [I] ExposeDMA: Disabled
45
+ [01/04/2024-14:46:20] [I] Data transfers: Enabled
46
+ [01/04/2024-14:46:20] [I] Spin-wait: Disabled
47
+ [01/04/2024-14:46:20] [I] Multithreading: Disabled
48
+ [01/04/2024-14:46:20] [I] CUDA Graph: Disabled
49
+ [01/04/2024-14:46:20] [I] Separate profiling: Disabled
50
+ [01/04/2024-14:46:20] [I] Time Deserialize: Disabled
51
+ [01/04/2024-14:46:20] [I] Time Refit: Disabled
52
+ [01/04/2024-14:46:20] [I] NVTX verbosity: 0
53
+ [01/04/2024-14:46:20] [I] Persistent Cache Ratio: 0
54
+ [01/04/2024-14:46:20] [I] Inputs:
55
+ [01/04/2024-14:46:20] [I] === Reporting Options ===
56
+ [01/04/2024-14:46:20] [I] Verbose: Disabled
57
+ [01/04/2024-14:46:20] [I] Averages: 100 inferences
58
+ [01/04/2024-14:46:20] [I] Percentiles: 90,95,99
59
+ [01/04/2024-14:46:20] [I] Dump refittable layers:Disabled
60
+ [01/04/2024-14:46:20] [I] Dump output: Disabled
61
+ [01/04/2024-14:46:20] [I] Profile: Disabled
62
+ [01/04/2024-14:46:20] [I] Export timing to JSON file:
63
+ [01/04/2024-14:46:20] [I] Export output to JSON file:
64
+ [01/04/2024-14:46:20] [I] Export profile to JSON file:
65
+ [01/04/2024-14:46:20] [I]
66
+ [01/04/2024-14:46:20] [I] === Device Information ===
67
+ [01/04/2024-14:46:20] [I] Selected Device: Orin
68
+ [01/04/2024-14:46:20] [I] Compute Capability: 8.7
69
+ [01/04/2024-14:46:20] [I] SMs: 8
70
+ [01/04/2024-14:46:20] [I] Compute Clock Rate: 0.624 GHz
71
+ [01/04/2024-14:46:20] [I] Device Global Memory: 7471 MiB
72
+ [01/04/2024-14:46:20] [I] Shared Memory per SM: 164 KiB
73
+ [01/04/2024-14:46:20] [I] Memory Bus Width: 128 bits (ECC disabled)
74
+ [01/04/2024-14:46:20] [I] Memory Clock Rate: 0.624 GHz
75
+ [01/04/2024-14:46:20] [I]
76
+ [01/04/2024-14:46:20] [I] TensorRT version: 8.5.2
77
+ [01/04/2024-14:46:20] [I] [TRT] [MemUsageChange] Init CUDA: CPU +220, GPU +0, now: CPU 249, GPU 3636 (MiB)
78
+ [01/04/2024-14:46:24] [I] [TRT] [MemUsageChange] Init builder kernel library: CPU +302, GPU +284, now: CPU 574, GPU 3941 (MiB)
79
+ [01/04/2024-14:46:24] [I] Start parsing network model
80
+ [01/04/2024-14:46:24] [I] [TRT] ----------------------------------------------------------------
81
+ [01/04/2024-14:46:24] [I] [TRT] Input filename: yolo_nas_pose_n_int8.onnx
82
+ [01/04/2024-14:46:24] [I] [TRT] ONNX IR version: 0.0.8
83
+ [01/04/2024-14:46:24] [I] [TRT] Opset version: 17
84
+ [01/04/2024-14:46:24] [I] [TRT] Producer name: pytorch
85
+ [01/04/2024-14:46:24] [I] [TRT] Producer version: 2.1.2
86
+ [01/04/2024-14:46:24] [I] [TRT] Domain:
87
+ [01/04/2024-14:46:24] [I] [TRT] Model version: 0
88
+ [01/04/2024-14:46:24] [I] [TRT] Doc string:
89
+ [01/04/2024-14:46:24] [I] [TRT] ----------------------------------------------------------------
90
+ [01/04/2024-14:46:27] [I] Finish parsing network model
91
+ [01/04/2024-14:46:27] [I] FP32 and INT8 precisions have been specified - more performance might be enabled by additionally specifying --fp16 or --best
92
+ [01/04/2024-14:46:31] [I] [TRT] ---------- Layers Running on DLA ----------
93
+ [01/04/2024-14:46:31] [I] [TRT] ---------- Layers Running on GPU ----------
94
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation1]
95
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/pre_process/pre_process.0/Cast.../pre_process/pre_process.2/Mul]}
96
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1204) [Constant]
97
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1205) [Constant]
98
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1206) [Constant]
99
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/backbone/stem/conv/rbr_reparam/_input_quantizer/QuantizeLinear
100
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stem.conv.rbr_reparam.weight + /model/backbone/stem/conv/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stem/conv/rbr_reparam/Conv
101
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.downsample.rbr_reparam.weight + /model/backbone/stage1/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/downsample/rbr_reparam/Conv
102
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv2.conv.weight + /model/backbone/stage1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv2/conv/Conv
103
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv1.conv.weight + /model/backbone/stage1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv1/conv/Conv
104
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
105
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
106
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
107
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 485) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Add)
108
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
109
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
110
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
111
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 501) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Add)
112
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv3.conv.weight + /model/backbone/stage1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv3/conv/Conv
113
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_skip2.conv.weight + /model/neck/neck2/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip2/conv/Conv
114
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.downsample.rbr_reparam.weight + /model/backbone/stage2/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/downsample/rbr_reparam/Conv
115
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.downsample.conv.weight + /model/neck/neck2/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/downsample/conv/Conv
116
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv2.conv.weight + /model/backbone/stage2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv2/conv/Conv
117
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv1.conv.weight + /model/backbone/stage2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv1/conv/Conv
118
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
119
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
120
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
121
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 548) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Add)
122
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
123
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
124
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
125
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 564) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Add)
126
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
127
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
128
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
129
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.2.alpha + (Unnamed Layer* 580) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Add)
130
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/Concat_/model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Add_output_0_clone_0 copy
131
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv3.conv.weight + /model/backbone/stage2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv3/conv/Conv
132
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip2.conv.weight + /model/neck/neck1/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip2/conv/Conv || model.neck.neck2.reduce_skip1.conv.weight + /model/neck/neck2/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip1/conv/Conv
133
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.downsample.rbr_reparam.weight + /model/backbone/stage3/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/downsample/rbr_reparam/Conv
134
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.downsample.conv.weight + /model/neck/neck1/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/downsample/conv/Conv
135
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv2.conv.weight + /model/backbone/stage3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv2/conv/Conv
136
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv1.conv.weight + /model/backbone/stage3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv1/conv/Conv
137
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
138
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
139
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
140
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 630) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Add)
141
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
142
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
143
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
144
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 646) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Add)
145
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
146
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
147
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
148
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.2.alpha + (Unnamed Layer* 662) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Add)
149
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
150
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/Conv
151
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/Conv
152
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.3.alpha + (Unnamed Layer* 678) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Add)
153
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv3.conv.weight + /model/backbone/stage3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv3/conv/Conv
154
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip1.conv.weight + /model/neck/neck1/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip1/conv/Conv
155
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.downsample.rbr_reparam.weight + /model/backbone/stage4/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/downsample/rbr_reparam/Conv
156
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv2.conv.weight + /model/backbone/stage4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv2/conv/Conv
157
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv1.conv.weight + /model/backbone/stage4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv1/conv/Conv
158
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
159
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
160
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
161
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 719) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Add)
162
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
163
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
164
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
165
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 735) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Add)
166
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv3.conv.weight + /model/backbone/stage4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv3/conv/Conv
167
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv1.conv.weight + /model/backbone/context_module/cv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv1/conv/Conv
168
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.2/MaxPool
169
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.1/MaxPool
170
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.0/MaxPool
171
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/backbone/context_module/m.2/MaxPool_output_0 copy
172
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv2.conv.weight + /model/backbone/context_module/cv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv2/conv/Conv
173
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.conv.conv.weight + /model/neck/neck1/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/conv/conv/Conv
174
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/upsample/_input_quantizer/QuantizeLinear
175
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck1.upsample.weight + /model/neck/neck1/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck1/upsample/ConvTranspose
176
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_after_concat.conv.weight + /model/neck/neck1/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_after_concat/conv/Conv
177
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv2.conv.weight + /model/neck/neck1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv2/conv/Conv
178
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv1.conv.weight + /model/neck/neck1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv1/conv/Conv
179
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
180
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
181
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
182
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 800) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Add)
183
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
184
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
185
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
186
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 816) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Add)
187
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/Concat_/model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
188
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv3.conv.weight + /model/neck/neck1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv3/conv/Conv
189
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.conv.conv.weight + /model/neck/neck2/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/conv/conv/Conv
190
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/upsample/_input_quantizer/QuantizeLinear
191
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck2.upsample.weight + /model/neck/neck2/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck2/upsample/ConvTranspose
192
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/Concat_/model/neck/neck2/reduce_skip1/act/Relu_output_0_clone_1 copy
193
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_after_concat.conv.weight + /model/neck/neck2/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_after_concat/conv/Conv
194
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv2.conv.weight + /model/neck/neck2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv2/conv/Conv
195
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv1.conv.weight + /model/neck/neck2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv1/conv/Conv
196
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
197
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
198
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
199
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 865) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Add)
200
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
201
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
202
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
203
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 881) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Add)
204
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv3.conv.weight + /model/neck/neck2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv3/conv/Conv
205
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.bbox_stem.seq.conv.weight + /model/heads/head1/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/bbox_stem/seq/conv/Conv || model.heads.head1.pose_stem.seq.conv.weight + /model/heads/head1/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_stem/seq/conv/Conv
206
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.conv.conv.weight + /model/neck/neck3/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/conv/conv/Conv
207
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_convs.0.seq.conv.weight + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head1.cls_convs.0.seq.conv.weight + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/Conv
208
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.0.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/Conv
209
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/conv1/conv/_input_quantizer/QuantizeLinear_clone_1
210
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.cls_pred.weight + /model/heads/head1/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_pred/Conv
211
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_pred.weight + /model/heads/head1/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_pred/Conv
212
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.1.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/Conv
213
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv2.conv.weight + /model/neck/neck3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv2/conv/Conv
214
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv1.conv.weight + /model/neck/neck3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv1/conv/Conv
215
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape + /model/heads/Transpose
216
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_pred.weight + /model/heads/head1/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_pred/Conv
217
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
218
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax
219
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
220
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv
221
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
222
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 947) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Add)
223
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
224
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
225
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
226
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 988) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Add)
227
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/Concat_/model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
228
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv3.conv.weight + /model/neck/neck3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv3/conv/Conv
229
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.bbox_stem.seq.conv.weight + /model/heads/head2/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/bbox_stem/seq/conv/Conv || model.heads.head2.pose_stem.seq.conv.weight + /model/heads/head2/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_stem/seq/conv/Conv
230
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.conv.conv.weight + /model/neck/neck4/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/conv/conv/Conv
231
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_convs.0.seq.conv.weight + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head2.cls_convs.0.seq.conv.weight + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/Conv
232
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.0.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/Conv
233
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/conv1/conv/_input_quantizer/QuantizeLinear_clone_1
234
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.cls_pred.weight + /model/heads/head2/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_pred/Conv
235
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_pred.weight + /model/heads/head2/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_pred/Conv
236
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.1.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/Conv
237
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv2.conv.weight + /model/neck/neck4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv2/conv/Conv
238
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv1.conv.weight + /model/neck/neck4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv1/conv/Conv
239
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_4 + /model/heads/Transpose_3
240
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_pred.weight + /model/heads/head2/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_pred/Conv
241
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
242
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_1
243
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
244
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_1
245
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
246
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 1054) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Add)
247
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
248
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
249
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
250
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 1095) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Add)
251
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/Concat_/model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
252
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv3.conv.weight + /model/neck/neck4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv3/conv/Conv
253
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.bbox_stem.seq.conv.weight + /model/heads/head3/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/bbox_stem/seq/conv/Conv || model.heads.head3.pose_stem.seq.conv.weight + /model/heads/head3/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_stem/seq/conv/Conv
254
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_convs.0.seq.conv.weight + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head3.cls_convs.0.seq.conv.weight + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/Conv
255
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.0.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/Conv
256
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.cls_pred.weight + /model/heads/head3/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_pred/Conv
257
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_pred.weight + /model/heads/head3/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_pred/Conv
258
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.1.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/Conv
259
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_8 + /model/heads/Transpose_6
260
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.2.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/Conv
261
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_2
262
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_pred.weight + /model/heads/head3/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_pred/Conv
263
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_2
264
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice_1.../post_process/Reshape_2]}
265
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] NMS: batched_nms_238
266
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] DEVICE_TO_SHAPE_HOST: (Unnamed Layer* 1208) [NMS]_1_output[DevicetoShapeHostCopy]
267
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation2]
268
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice...graph2_/Concat_5]}
269
+ [01/04/2024-14:46:31] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation3]
270
+ [01/04/2024-14:46:36] [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +534, GPU +436, now: CPU 1168, GPU 4440 (MiB)
271
+ [01/04/2024-14:46:37] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +82, GPU +43, now: CPU 1250, GPU 4483 (MiB)
272
+ [01/04/2024-14:46:37] [I] [TRT] Local timing cache in use. Profiling results in this builder pass will not be stored.
273
+ [01/04/2024-14:55:15] [I] [TRT] Total Activation Memory: 7920254464
274
+ [01/04/2024-14:55:15] [I] [TRT] Detected 1 inputs and 1 output network tensors.
275
+ [01/04/2024-14:55:15] [I] [TRT] Total Host Persistent Memory: 300608
276
+ [01/04/2024-14:55:15] [I] [TRT] Total Device Persistent Memory: 119296
277
+ [01/04/2024-14:55:15] [I] [TRT] Total Scratch Memory: 134217728
278
+ [01/04/2024-14:55:15] [I] [TRT] [MemUsageStats] Peak memory usage of TRT CPU/GPU memory allocators: CPU 33 MiB, GPU 132 MiB
279
+ [01/04/2024-14:55:15] [I] [TRT] [BlockAssignment] Started assigning block shifts. This will take 185 steps to complete.
280
+ [01/04/2024-14:55:15] [I] [TRT] [BlockAssignment] Algorithm ShiftNTopDown took 40.3888ms to assign 13 blocks to 185 nodes requiring 144141824 bytes.
281
+ [01/04/2024-14:55:15] [I] [TRT] Total Activation Memory: 144141824
282
+ [01/04/2024-14:55:16] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1590, GPU 5342 (MiB)
283
+ [01/04/2024-14:55:16] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in building engine: CPU +6, GPU +8, now: CPU 6, GPU 8 (MiB)
284
+ [01/04/2024-14:55:16] [I] Engine built in 536.559 sec.
285
+ [01/04/2024-14:55:17] [I] [TRT] Loaded engine size: 9 MiB
286
+ [01/04/2024-14:55:17] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1240, GPU 5345 (MiB)
287
+ [01/04/2024-14:55:17] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in engine deserialization: CPU +0, GPU +7, now: CPU 0, GPU 7 (MiB)
288
+ [01/04/2024-14:55:17] [I] Engine deserialized in 0.120887 sec.
289
+ [01/04/2024-14:55:17] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1241, GPU 5345 (MiB)
290
+ [01/04/2024-14:55:17] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +0, GPU +138, now: CPU 0, GPU 145 (MiB)
291
+ [01/04/2024-14:55:17] [I] Setting persistentCacheLimit to 0 bytes.
292
+ [01/04/2024-14:55:17] [I] Using random values for input onnx::Cast_0
293
+ [01/04/2024-14:55:17] [I] Created input binding for onnx::Cast_0 with dimensions 1x3x640x640
294
+ [01/04/2024-14:55:17] [I] Using random values for output graph2_flat_predictions
295
+ [01/04/2024-14:55:17] [I] Created output binding for graph2_flat_predictions with dimensions -1x57
296
+ [01/04/2024-14:55:17] [I] Starting inference
297
+ [01/04/2024-14:55:32] [I] Warmup completed 11 queries over 200 ms
298
+ [01/04/2024-14:55:32] [I] Timing trace has 1108 queries over 15.0289 s
299
+ [01/04/2024-14:55:32] [I]
300
+ [01/04/2024-14:55:32] [I] === Trace details ===
301
+ [01/04/2024-14:55:32] [I] Trace averages of 100 runs:
302
+ [01/04/2024-14:55:32] [I] Average on 100 runs - GPU latency: 13.2196 ms - Host latency: 13.3319 ms (enqueue 13.2998 ms)
303
+ [01/04/2024-14:55:32] [I] Average on 100 runs - GPU latency: 13.3938 ms - Host latency: 13.5071 ms (enqueue 13.4684 ms)
304
+ [01/04/2024-14:55:32] [I] Average on 100 runs - GPU latency: 13.2194 ms - Host latency: 13.3318 ms (enqueue 13.2975 ms)
305
+ [01/04/2024-14:55:32] [I] Average on 100 runs - GPU latency: 13.2789 ms - Host latency: 13.3906 ms (enqueue 13.3589 ms)
306
+ [01/04/2024-14:55:32] [I] Average on 100 runs - GPU latency: 13.6773 ms - Host latency: 13.7933 ms (enqueue 13.7451 ms)
307
+ [01/04/2024-14:55:32] [I] Average on 100 runs - GPU latency: 13.7471 ms - Host latency: 13.8643 ms (enqueue 13.8164 ms)
308
+ [01/04/2024-14:55:32] [I] Average on 100 runs - GPU latency: 13.5782 ms - Host latency: 13.6935 ms (enqueue 13.6539 ms)
309
+ [01/04/2024-14:55:32] [I] Average on 100 runs - GPU latency: 13.3669 ms - Host latency: 13.4813 ms (enqueue 13.4435 ms)
310
+ [01/04/2024-14:55:32] [I] Average on 100 runs - GPU latency: 13.2753 ms - Host latency: 13.3888 ms (enqueue 13.3504 ms)
311
+ [01/04/2024-14:55:32] [I] Average on 100 runs - GPU latency: 13.3138 ms - Host latency: 13.4258 ms (enqueue 13.3842 ms)
312
+ [01/04/2024-14:55:32] [I] Average on 100 runs - GPU latency: 13.511 ms - Host latency: 13.6262 ms (enqueue 13.5945 ms)
313
+ [01/04/2024-14:55:32] [I]
314
+ [01/04/2024-14:55:32] [I] === Performance summary ===
315
+ [01/04/2024-14:55:32] [I] Throughput: 73.7247 qps
316
+ [01/04/2024-14:55:32] [I] Latency: min = 12.3433 ms, max = 18.3281 ms, mean = 13.5316 ms, median = 13.4639 ms, percentile(90%) = 14.1807 ms, percentile(95%) = 14.3574 ms, percentile(99%) = 15.7993 ms
317
+ [01/04/2024-14:55:32] [I] Enqueue Time: min = 12.313 ms, max = 18.2793 ms, mean = 13.4932 ms, median = 13.4253 ms, percentile(90%) = 14.1348 ms, percentile(95%) = 14.3091 ms, percentile(99%) = 15.7402 ms
318
+ [01/04/2024-14:55:32] [I] H2D Latency: min = 0.0810547 ms, max = 0.114258 ms, mean = 0.0982483 ms, median = 0.0986328 ms, percentile(90%) = 0.0998535 ms, percentile(95%) = 0.100586 ms, percentile(99%) = 0.102539 ms
319
+ [01/04/2024-14:55:32] [I] GPU Compute Time: min = 12.2305 ms, max = 18.21 ms, mean = 13.4177 ms, median = 13.3489 ms, percentile(90%) = 14.0645 ms, percentile(95%) = 14.2349 ms, percentile(99%) = 15.7002 ms
320
+ [01/04/2024-14:55:32] [I] D2H Latency: min = 0.00292969 ms, max = 0.0498047 ms, mean = 0.0156769 ms, median = 0.0146484 ms, percentile(90%) = 0.0205078 ms, percentile(95%) = 0.0230713 ms, percentile(99%) = 0.03125 ms
321
+ [01/04/2024-14:55:32] [I] Total Host Walltime: 15.0289 s
322
+ [01/04/2024-14:55:32] [I] Total GPU Compute Time: 14.8668 s
323
+ [01/04/2024-14:55:32] [I] Explanations of the performance metrics are printed in the verbose logs.
324
+ [01/04/2024-14:55:32] [I]
325
  &&&& PASSED TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_n_int8.onnx --int8 --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_n_int8.onnx.int8.engine
yolo_nas_pose_s_int8.onnx.best.engine CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19f0e5f7536fe7f1f802536c4f9de7a56072c2f442043fe76c02813262b89a90
3
- size 18008336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24cb45f2e6bbc0f6183888eae7fa29fbe79918082902ae47814c721c3f57bb68
3
+ size 18052506
yolo_nas_pose_s_int8.onnx.best.engine.err CHANGED
@@ -1,7 +1,7 @@
1
- [12/28/2023-03:03:31] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
- [12/28/2023-03:03:31] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
- [12/28/2023-03:03:34] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
- [12/28/2023-03:36:12] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
- [12/28/2023-03:36:12] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
- [12/28/2023-03:36:12] [W] * GPU compute time is unstable, with coefficient of variance = 3.47567%.
7
- [12/28/2023-03:36:12] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
 
1
+ [01/04/2024-14:55:38] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
+ [01/04/2024-14:55:38] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
+ [01/04/2024-14:55:41] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
+ [01/04/2024-15:28:09] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
+ [01/04/2024-15:28:10] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
+ [01/04/2024-15:28:10] [W] * GPU compute time is unstable, with coefficient of variance = 4.19166%.
7
+ [01/04/2024-15:28:10] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
yolo_nas_pose_s_int8.onnx.best.engine.log CHANGED
@@ -1,323 +1,324 @@
1
  &&&& RUNNING TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_s_int8.onnx --best --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_s_int8.onnx.best.engine
2
- [12/28/2023-03:03:27] [I] === Model Options ===
3
- [12/28/2023-03:03:27] [I] Format: ONNX
4
- [12/28/2023-03:03:27] [I] Model: yolo_nas_pose_s_int8.onnx
5
- [12/28/2023-03:03:27] [I] Output:
6
- [12/28/2023-03:03:27] [I] === Build Options ===
7
- [12/28/2023-03:03:27] [I] Max batch: explicit batch
8
- [12/28/2023-03:03:27] [I] Memory Pools: workspace: default, dlaSRAM: default, dlaLocalDRAM: default, dlaGlobalDRAM: default
9
- [12/28/2023-03:03:27] [I] minTiming: 1
10
- [12/28/2023-03:03:27] [I] avgTiming: 8
11
- [12/28/2023-03:03:27] [I] Precision: FP32+FP16+INT8
12
- [12/28/2023-03:03:27] [I] LayerPrecisions:
13
- [12/28/2023-03:03:27] [I] Calibration: Dynamic
14
- [12/28/2023-03:03:27] [I] Refit: Disabled
15
- [12/28/2023-03:03:27] [I] Sparsity: Disabled
16
- [12/28/2023-03:03:27] [I] Safe mode: Disabled
17
- [12/28/2023-03:03:27] [I] DirectIO mode: Disabled
18
- [12/28/2023-03:03:27] [I] Restricted mode: Disabled
19
- [12/28/2023-03:03:27] [I] Build only: Disabled
20
- [12/28/2023-03:03:27] [I] Save engine: yolo_nas_pose_s_int8.onnx.best.engine
21
- [12/28/2023-03:03:27] [I] Load engine:
22
- [12/28/2023-03:03:27] [I] Profiling verbosity: 0
23
- [12/28/2023-03:03:27] [I] Tactic sources: Using default tactic sources
24
- [12/28/2023-03:03:27] [I] timingCacheMode: local
25
- [12/28/2023-03:03:27] [I] timingCacheFile:
26
- [12/28/2023-03:03:27] [I] Heuristic: Disabled
27
- [12/28/2023-03:03:27] [I] Preview Features: Use default preview flags.
28
- [12/28/2023-03:03:27] [I] Input(s)s format: fp32:CHW
29
- [12/28/2023-03:03:27] [I] Output(s)s format: fp32:CHW
30
- [12/28/2023-03:03:27] [I] Input build shapes: model
31
- [12/28/2023-03:03:27] [I] Input calibration shapes: model
32
- [12/28/2023-03:03:27] [I] === System Options ===
33
- [12/28/2023-03:03:27] [I] Device: 0
34
- [12/28/2023-03:03:27] [I] DLACore:
35
- [12/28/2023-03:03:27] [I] Plugins:
36
- [12/28/2023-03:03:27] [I] === Inference Options ===
37
- [12/28/2023-03:03:27] [I] Batch: Explicit
38
- [12/28/2023-03:03:27] [I] Input inference shapes: model
39
- [12/28/2023-03:03:27] [I] Iterations: 10
40
- [12/28/2023-03:03:27] [I] Duration: 15s (+ 200ms warm up)
41
- [12/28/2023-03:03:27] [I] Sleep time: 0ms
42
- [12/28/2023-03:03:27] [I] Idle time: 0ms
43
- [12/28/2023-03:03:27] [I] Streams: 1
44
- [12/28/2023-03:03:27] [I] ExposeDMA: Disabled
45
- [12/28/2023-03:03:27] [I] Data transfers: Enabled
46
- [12/28/2023-03:03:27] [I] Spin-wait: Disabled
47
- [12/28/2023-03:03:27] [I] Multithreading: Disabled
48
- [12/28/2023-03:03:27] [I] CUDA Graph: Disabled
49
- [12/28/2023-03:03:27] [I] Separate profiling: Disabled
50
- [12/28/2023-03:03:27] [I] Time Deserialize: Disabled
51
- [12/28/2023-03:03:27] [I] Time Refit: Disabled
52
- [12/28/2023-03:03:27] [I] NVTX verbosity: 0
53
- [12/28/2023-03:03:27] [I] Persistent Cache Ratio: 0
54
- [12/28/2023-03:03:27] [I] Inputs:
55
- [12/28/2023-03:03:27] [I] === Reporting Options ===
56
- [12/28/2023-03:03:27] [I] Verbose: Disabled
57
- [12/28/2023-03:03:27] [I] Averages: 100 inferences
58
- [12/28/2023-03:03:27] [I] Percentiles: 90,95,99
59
- [12/28/2023-03:03:27] [I] Dump refittable layers:Disabled
60
- [12/28/2023-03:03:27] [I] Dump output: Disabled
61
- [12/28/2023-03:03:27] [I] Profile: Disabled
62
- [12/28/2023-03:03:27] [I] Export timing to JSON file:
63
- [12/28/2023-03:03:27] [I] Export output to JSON file:
64
- [12/28/2023-03:03:27] [I] Export profile to JSON file:
65
- [12/28/2023-03:03:27] [I]
66
- [12/28/2023-03:03:27] [I] === Device Information ===
67
- [12/28/2023-03:03:27] [I] Selected Device: Orin
68
- [12/28/2023-03:03:27] [I] Compute Capability: 8.7
69
- [12/28/2023-03:03:27] [I] SMs: 8
70
- [12/28/2023-03:03:27] [I] Compute Clock Rate: 0.624 GHz
71
- [12/28/2023-03:03:27] [I] Device Global Memory: 7471 MiB
72
- [12/28/2023-03:03:27] [I] Shared Memory per SM: 164 KiB
73
- [12/28/2023-03:03:27] [I] Memory Bus Width: 128 bits (ECC disabled)
74
- [12/28/2023-03:03:27] [I] Memory Clock Rate: 0.624 GHz
75
- [12/28/2023-03:03:27] [I]
76
- [12/28/2023-03:03:27] [I] TensorRT version: 8.5.2
77
- [12/28/2023-03:03:28] [I] [TRT] [MemUsageChange] Init CUDA: CPU +220, GPU +0, now: CPU 249, GPU 3018 (MiB)
78
- [12/28/2023-03:03:31] [I] [TRT] [MemUsageChange] Init builder kernel library: CPU +302, GPU +284, now: CPU 574, GPU 3324 (MiB)
79
- [12/28/2023-03:03:31] [I] Start parsing network model
80
- [12/28/2023-03:03:31] [I] [TRT] ----------------------------------------------------------------
81
- [12/28/2023-03:03:31] [I] [TRT] Input filename: yolo_nas_pose_s_int8.onnx
82
- [12/28/2023-03:03:31] [I] [TRT] ONNX IR version: 0.0.8
83
- [12/28/2023-03:03:31] [I] [TRT] Opset version: 17
84
- [12/28/2023-03:03:31] [I] [TRT] Producer name: pytorch
85
- [12/28/2023-03:03:31] [I] [TRT] Producer version: 2.1.2
86
- [12/28/2023-03:03:31] [I] [TRT] Domain:
87
- [12/28/2023-03:03:31] [I] [TRT] Model version: 0
88
- [12/28/2023-03:03:31] [I] [TRT] Doc string:
89
- [12/28/2023-03:03:31] [I] [TRT] ----------------------------------------------------------------
90
- [12/28/2023-03:03:34] [I] Finish parsing network model
91
- [12/28/2023-03:03:38] [I] [TRT] ---------- Layers Running on DLA ----------
92
- [12/28/2023-03:03:38] [I] [TRT] ---------- Layers Running on GPU ----------
93
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation1]
94
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/pre_process/pre_process.0/Cast.../pre_process/pre_process.2/Mul]}
95
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1228) [Constant]
96
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1229) [Constant]
97
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1230) [Constant]
98
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/backbone/stem/conv/rbr_reparam/_input_quantizer/QuantizeLinear
99
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stem.conv.rbr_reparam.weight + /model/backbone/stem/conv/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stem/conv/rbr_reparam/Conv
100
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.downsample.rbr_reparam.weight + /model/backbone/stage1/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/downsample/rbr_reparam/Conv
101
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv2.conv.weight + /model/backbone/stage1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv2/conv/Conv
102
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv1.conv.weight + /model/backbone/stage1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv1/conv/Conv
103
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
104
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
105
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
106
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 494) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Add)
107
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
108
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
109
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
110
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 510) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Add)
111
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv3.conv.weight + /model/backbone/stage1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv3/conv/Conv
112
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_skip2.conv.weight + /model/neck/neck2/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip2/conv/Conv
113
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.downsample.rbr_reparam.weight + /model/backbone/stage2/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/downsample/rbr_reparam/Conv
114
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.downsample.conv.weight + /model/neck/neck2/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/downsample/conv/Conv
115
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv2.conv.weight + /model/backbone/stage2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv2/conv/Conv
116
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv1.conv.weight + /model/backbone/stage2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv1/conv/Conv
117
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
118
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
119
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
120
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 557) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Add)
121
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
122
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
123
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
124
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 573) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Add)
125
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
126
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
127
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
128
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.2.alpha + (Unnamed Layer* 589) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Add)
129
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv3.conv.weight + /model/backbone/stage2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv3/conv/Conv
130
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip2.conv.weight + /model/neck/neck1/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip2/conv/Conv || model.neck.neck2.reduce_skip1.conv.weight + /model/neck/neck2/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip1/conv/Conv
131
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.downsample.rbr_reparam.weight + /model/backbone/stage3/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/downsample/rbr_reparam/Conv
132
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.downsample.conv.weight + /model/neck/neck1/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/downsample/conv/Conv
133
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv2.conv.weight + /model/backbone/stage3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv2/conv/Conv
134
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv1.conv.weight + /model/backbone/stage3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv1/conv/Conv
135
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
136
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
137
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
138
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 639) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Add)
139
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
140
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
141
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
142
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 655) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Add)
143
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
144
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
145
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
146
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.2.alpha + (Unnamed Layer* 671) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Add)
147
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
148
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/Conv
149
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/Conv
150
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.3.alpha + (Unnamed Layer* 687) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Add)
151
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
152
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.4.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv1/rbr_reparam/Conv
153
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.4.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv2/rbr_reparam/Conv
154
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.4.alpha + (Unnamed Layer* 703) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/Add)
155
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv3.conv.weight + /model/backbone/stage3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv3/conv/Conv
156
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip1.conv.weight + /model/neck/neck1/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip1/conv/Conv
157
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.downsample.rbr_reparam.weight + /model/backbone/stage4/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/downsample/rbr_reparam/Conv
158
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv2.conv.weight + /model/backbone/stage4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv2/conv/Conv
159
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv1.conv.weight + /model/backbone/stage4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv1/conv/Conv
160
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
161
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
162
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
163
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 744) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Add)
164
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
165
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
166
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
167
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 760) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Add)
168
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv3.conv.weight + /model/backbone/stage4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv3/conv/Conv
169
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv1.conv.weight + /model/backbone/context_module/cv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv1/conv/Conv
170
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.2/MaxPool
171
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.1/MaxPool
172
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.0/MaxPool
173
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/backbone/context_module/m.2/MaxPool_output_0 copy
174
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv2.conv.weight + /model/backbone/context_module/cv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv2/conv/Conv
175
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.conv.conv.weight + /model/neck/neck1/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/conv/conv/Conv
176
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck1.upsample.weight + /model/neck/neck1/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck1/upsample/ConvTranspose
177
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_after_concat.conv.weight + /model/neck/neck1/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_after_concat/conv/Conv
178
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv2.conv.weight + /model/neck/neck1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv2/conv/Conv
179
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv1.conv.weight + /model/neck/neck1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv1/conv/Conv
180
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
181
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
182
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
183
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 825) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Add)
184
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
185
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
186
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
187
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 841) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Add)
188
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv3.conv.weight + /model/neck/neck1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv3/conv/Conv
189
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.conv.conv.weight + /model/neck/neck2/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/conv/conv/Conv
190
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/upsample/_input_quantizer/QuantizeLinear
191
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck2.upsample.weight + /model/neck/neck2/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck2/upsample/ConvTranspose
192
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/Concat_/model/neck/neck2/reduce_skip1/act/Relu_output_0_clone_1 copy
193
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_after_concat.conv.weight + /model/neck/neck2/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_after_concat/conv/Conv
194
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv2.conv.weight + /model/neck/neck2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv2/conv/Conv
195
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv1.conv.weight + /model/neck/neck2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv1/conv/Conv
196
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
197
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
198
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
199
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 890) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Add)
200
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
201
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
202
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
203
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 906) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Add)
204
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/Concat_/model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
205
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv3.conv.weight + /model/neck/neck2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv3/conv/Conv
206
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.bbox_stem.seq.conv.weight + /model/heads/head1/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/bbox_stem/seq/conv/Conv || model.heads.head1.pose_stem.seq.conv.weight + /model/heads/head1/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_stem/seq/conv/Conv
207
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.conv.conv.weight + /model/neck/neck3/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/conv/conv/Conv
208
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_convs.0.seq.conv.weight + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head1.cls_convs.0.seq.conv.weight + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/Conv
209
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.0.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/Conv
210
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/conv1/conv/_input_quantizer/QuantizeLinear_clone_1
211
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.cls_pred.weight + /model/heads/head1/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_pred/Conv
212
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_pred.weight + /model/heads/head1/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_pred/Conv
213
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.1.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/Conv
214
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv2.conv.weight + /model/neck/neck3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv2/conv/Conv
215
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv1.conv.weight + /model/neck/neck3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv1/conv/Conv
216
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape + /model/heads/Transpose
217
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_pred.weight + /model/heads/head1/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_pred/Conv
218
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
219
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax
220
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
221
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv
222
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
223
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 972) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Add)
224
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
225
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
226
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
227
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 1013) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Add)
228
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv3.conv.weight + /model/neck/neck3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv3/conv/Conv
229
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_stem.seq.conv.weight + /model/heads/head2/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_stem/seq/conv/Conv || model.heads.head2.bbox_stem.seq.conv.weight + /model/heads/head2/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/bbox_stem/seq/conv/Conv
230
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.conv.conv.weight + /model/neck/neck4/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/conv/conv/Conv
231
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_convs.0.seq.conv.weight + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head2.cls_convs.0.seq.conv.weight + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/Conv
232
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.0.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/Conv
233
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv2.conv.weight + /model/neck/neck4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv2/conv/Conv
234
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv1.conv.weight + /model/neck/neck4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv1/conv/Conv
235
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.cls_pred.weight + /model/heads/head2/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_pred/Conv
236
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_pred.weight + /model/heads/head2/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_pred/Conv
237
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.1.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/Conv
238
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
239
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_4 + /model/heads/Transpose_3
240
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_pred.weight + /model/heads/head2/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_pred/Conv
241
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
242
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_1
243
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
244
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_1
245
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 1078) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Add)
246
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
247
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
248
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
249
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 1119) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Add)
250
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv3.conv.weight + /model/neck/neck4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv3/conv/Conv
251
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.bbox_stem.seq.conv.weight + /model/heads/head3/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/bbox_stem/seq/conv/Conv || model.heads.head3.pose_stem.seq.conv.weight + /model/heads/head3/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_stem/seq/conv/Conv
252
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_convs.0.seq.conv.weight + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head3.cls_convs.0.seq.conv.weight + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/Conv
253
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.0.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/Conv
254
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.cls_pred.weight + /model/heads/head3/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_pred/Conv
255
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_pred.weight + /model/heads/head3/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_pred/Conv
256
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.1.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/Conv
257
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_8 + /model/heads/Transpose_6
258
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.2.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/Conv
259
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_2
260
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_pred.weight + /model/heads/head3/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_pred/Conv
261
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_2
262
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice_1.../post_process/Reshape_2]}
263
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] NMS: batched_nms_243
264
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] DEVICE_TO_SHAPE_HOST: (Unnamed Layer* 1232) [NMS]_1_output[DevicetoShapeHostCopy]
265
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation2]
266
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice...graph2_/Concat_5]}
267
- [12/28/2023-03:03:38] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation3]
268
- [12/28/2023-03:03:51] [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +534, GPU +341, now: CPU 1233, GPU 3794 (MiB)
269
- [12/28/2023-03:03:53] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +82, GPU +76, now: CPU 1315, GPU 3870 (MiB)
270
- [12/28/2023-03:03:53] [I] [TRT] Local timing cache in use. Profiling results in this builder pass will not be stored.
271
- [12/28/2023-03:35:46] [I] [TRT] Total Activation Memory: 7917157888
272
- [12/28/2023-03:35:46] [I] [TRT] Detected 1 inputs and 1 output network tensors.
273
- [12/28/2023-03:35:53] [I] [TRT] Total Host Persistent Memory: 308256
274
- [12/28/2023-03:35:53] [I] [TRT] Total Device Persistent Memory: 38912
275
- [12/28/2023-03:35:53] [I] [TRT] Total Scratch Memory: 134217728
276
- [12/28/2023-03:35:53] [I] [TRT] [MemUsageStats] Peak memory usage of TRT CPU/GPU memory allocators: CPU 74 MiB, GPU 154 MiB
277
- [12/28/2023-03:35:53] [I] [TRT] [BlockAssignment] Started assigning block shifts. This will take 164 steps to complete.
278
- [12/28/2023-03:35:53] [I] [TRT] [BlockAssignment] Algorithm ShiftNTopDown took 59.5318ms to assign 13 blocks to 164 nodes requiring 142029824 bytes.
279
- [12/28/2023-03:35:53] [I] [TRT] Total Activation Memory: 142029824
280
- [12/28/2023-03:35:56] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1667, GPU 5515 (MiB)
281
- [12/28/2023-03:35:56] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in building engine: CPU +15, GPU +16, now: CPU 15, GPU 16 (MiB)
282
- [12/28/2023-03:35:56] [I] Engine built in 1949.31 sec.
283
- [12/28/2023-03:35:57] [I] [TRT] Loaded engine size: 17 MiB
284
- [12/28/2023-03:35:57] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1252, GPU 5486 (MiB)
285
- [12/28/2023-03:35:57] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in engine deserialization: CPU +0, GPU +15, now: CPU 0, GPU 15 (MiB)
286
- [12/28/2023-03:35:57] [I] Engine deserialized in 0.208393 sec.
287
- [12/28/2023-03:35:57] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1252, GPU 5486 (MiB)
288
- [12/28/2023-03:35:57] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +0, GPU +135, now: CPU 0, GPU 150 (MiB)
289
- [12/28/2023-03:35:57] [I] Setting persistentCacheLimit to 0 bytes.
290
- [12/28/2023-03:35:57] [I] Using random values for input onnx::Cast_0
291
- [12/28/2023-03:35:57] [I] Created input binding for onnx::Cast_0 with dimensions 1x3x640x640
292
- [12/28/2023-03:35:57] [I] Using random values for output graph2_flat_predictions
293
- [12/28/2023-03:35:57] [I] Created output binding for graph2_flat_predictions with dimensions -1x57
294
- [12/28/2023-03:35:57] [I] Starting inference
295
- [12/28/2023-03:36:12] [I] Warmup completed 11 queries over 200 ms
296
- [12/28/2023-03:36:12] [I] Timing trace has 1116 queries over 15.0304 s
297
- [12/28/2023-03:36:12] [I]
298
- [12/28/2023-03:36:12] [I] === Trace details ===
299
- [12/28/2023-03:36:12] [I] Trace averages of 100 runs:
300
- [12/28/2023-03:36:12] [I] Average on 100 runs - GPU latency: 13.3509 ms - Host latency: 13.4657 ms (enqueue 13.4291 ms)
301
- [12/28/2023-03:36:12] [I] Average on 100 runs - GPU latency: 13.6114 ms - Host latency: 13.7267 ms (enqueue 13.6814 ms)
302
- [12/28/2023-03:36:12] [I] Average on 100 runs - GPU latency: 13.2966 ms - Host latency: 13.4103 ms (enqueue 13.3777 ms)
303
- [12/28/2023-03:36:12] [I] Average on 100 runs - GPU latency: 13.3229 ms - Host latency: 13.4369 ms (enqueue 13.4032 ms)
304
- [12/28/2023-03:36:12] [I] Average on 100 runs - GPU latency: 13.3265 ms - Host latency: 13.4397 ms (enqueue 13.4066 ms)
305
- [12/28/2023-03:36:12] [I] Average on 100 runs - GPU latency: 13.3048 ms - Host latency: 13.4183 ms (enqueue 13.3859 ms)
306
- [12/28/2023-03:36:12] [I] Average on 100 runs - GPU latency: 13.2897 ms - Host latency: 13.4033 ms (enqueue 13.375 ms)
307
- [12/28/2023-03:36:12] [I] Average on 100 runs - GPU latency: 13.2914 ms - Host latency: 13.4051 ms (enqueue 13.3713 ms)
308
- [12/28/2023-03:36:12] [I] Average on 100 runs - GPU latency: 13.2786 ms - Host latency: 13.3924 ms (enqueue 13.3595 ms)
309
- [12/28/2023-03:36:12] [I] Average on 100 runs - GPU latency: 13.2361 ms - Host latency: 13.3498 ms (enqueue 13.3173 ms)
310
- [12/28/2023-03:36:12] [I] Average on 100 runs - GPU latency: 13.2792 ms - Host latency: 13.3929 ms (enqueue 13.3575 ms)
311
- [12/28/2023-03:36:12] [I]
312
- [12/28/2023-03:36:12] [I] === Performance summary ===
313
- [12/28/2023-03:36:12] [I] Throughput: 74.2494 qps
314
- [12/28/2023-03:36:12] [I] Latency: min = 12.5967 ms, max = 18.459 ms, mean = 13.4387 ms, median = 13.2627 ms, percentile(90%) = 13.9451 ms, percentile(95%) = 14.0898 ms, percentile(99%) = 15.3137 ms
315
- [12/28/2023-03:36:12] [I] Enqueue Time: min = 12.5688 ms, max = 18.406 ms, mean = 13.4045 ms, median = 13.2305 ms, percentile(90%) = 13.9082 ms, percentile(95%) = 14.0646 ms, percentile(99%) = 14.8955 ms
316
- [12/28/2023-03:36:12] [I] H2D Latency: min = 0.0820312 ms, max = 0.120605 ms, mean = 0.100692 ms, median = 0.101074 ms, percentile(90%) = 0.102295 ms, percentile(95%) = 0.102539 ms, percentile(99%) = 0.113281 ms
317
- [12/28/2023-03:36:12] [I] GPU Compute Time: min = 12.4849 ms, max = 18.3318 ms, mean = 13.3247 ms, median = 13.1485 ms, percentile(90%) = 13.8306 ms, percentile(95%) = 13.973 ms, percentile(99%) = 15.1929 ms
318
- [12/28/2023-03:36:12] [I] D2H Latency: min = 0.00292969 ms, max = 0.0505371 ms, mean = 0.0132362 ms, median = 0.0117188 ms, percentile(90%) = 0.0186768 ms, percentile(95%) = 0.0222168 ms, percentile(99%) = 0.0280762 ms
319
- [12/28/2023-03:36:12] [I] Total Host Walltime: 15.0304 s
320
- [12/28/2023-03:36:12] [I] Total GPU Compute Time: 14.8704 s
321
- [12/28/2023-03:36:12] [I] Explanations of the performance metrics are printed in the verbose logs.
322
- [12/28/2023-03:36:12] [I]
 
323
  &&&& PASSED TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_s_int8.onnx --best --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_s_int8.onnx.best.engine
 
1
  &&&& RUNNING TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_s_int8.onnx --best --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_s_int8.onnx.best.engine
2
+ [01/04/2024-14:55:33] [I] === Model Options ===
3
+ [01/04/2024-14:55:33] [I] Format: ONNX
4
+ [01/04/2024-14:55:33] [I] Model: yolo_nas_pose_s_int8.onnx
5
+ [01/04/2024-14:55:33] [I] Output:
6
+ [01/04/2024-14:55:33] [I] === Build Options ===
7
+ [01/04/2024-14:55:33] [I] Max batch: explicit batch
8
+ [01/04/2024-14:55:33] [I] Memory Pools: workspace: default, dlaSRAM: default, dlaLocalDRAM: default, dlaGlobalDRAM: default
9
+ [01/04/2024-14:55:33] [I] minTiming: 1
10
+ [01/04/2024-14:55:33] [I] avgTiming: 8
11
+ [01/04/2024-14:55:33] [I] Precision: FP32+FP16+INT8
12
+ [01/04/2024-14:55:33] [I] LayerPrecisions:
13
+ [01/04/2024-14:55:33] [I] Calibration: Dynamic
14
+ [01/04/2024-14:55:33] [I] Refit: Disabled
15
+ [01/04/2024-14:55:33] [I] Sparsity: Disabled
16
+ [01/04/2024-14:55:33] [I] Safe mode: Disabled
17
+ [01/04/2024-14:55:33] [I] DirectIO mode: Disabled
18
+ [01/04/2024-14:55:33] [I] Restricted mode: Disabled
19
+ [01/04/2024-14:55:33] [I] Build only: Disabled
20
+ [01/04/2024-14:55:33] [I] Save engine: yolo_nas_pose_s_int8.onnx.best.engine
21
+ [01/04/2024-14:55:33] [I] Load engine:
22
+ [01/04/2024-14:55:33] [I] Profiling verbosity: 0
23
+ [01/04/2024-14:55:33] [I] Tactic sources: Using default tactic sources
24
+ [01/04/2024-14:55:33] [I] timingCacheMode: local
25
+ [01/04/2024-14:55:33] [I] timingCacheFile:
26
+ [01/04/2024-14:55:33] [I] Heuristic: Disabled
27
+ [01/04/2024-14:55:33] [I] Preview Features: Use default preview flags.
28
+ [01/04/2024-14:55:33] [I] Input(s)s format: fp32:CHW
29
+ [01/04/2024-14:55:33] [I] Output(s)s format: fp32:CHW
30
+ [01/04/2024-14:55:33] [I] Input build shapes: model
31
+ [01/04/2024-14:55:33] [I] Input calibration shapes: model
32
+ [01/04/2024-14:55:33] [I] === System Options ===
33
+ [01/04/2024-14:55:33] [I] Device: 0
34
+ [01/04/2024-14:55:33] [I] DLACore:
35
+ [01/04/2024-14:55:33] [I] Plugins:
36
+ [01/04/2024-14:55:33] [I] === Inference Options ===
37
+ [01/04/2024-14:55:33] [I] Batch: Explicit
38
+ [01/04/2024-14:55:33] [I] Input inference shapes: model
39
+ [01/04/2024-14:55:33] [I] Iterations: 10
40
+ [01/04/2024-14:55:33] [I] Duration: 15s (+ 200ms warm up)
41
+ [01/04/2024-14:55:33] [I] Sleep time: 0ms
42
+ [01/04/2024-14:55:33] [I] Idle time: 0ms
43
+ [01/04/2024-14:55:33] [I] Streams: 1
44
+ [01/04/2024-14:55:33] [I] ExposeDMA: Disabled
45
+ [01/04/2024-14:55:33] [I] Data transfers: Enabled
46
+ [01/04/2024-14:55:33] [I] Spin-wait: Disabled
47
+ [01/04/2024-14:55:33] [I] Multithreading: Disabled
48
+ [01/04/2024-14:55:33] [I] CUDA Graph: Disabled
49
+ [01/04/2024-14:55:33] [I] Separate profiling: Disabled
50
+ [01/04/2024-14:55:33] [I] Time Deserialize: Disabled
51
+ [01/04/2024-14:55:33] [I] Time Refit: Disabled
52
+ [01/04/2024-14:55:33] [I] NVTX verbosity: 0
53
+ [01/04/2024-14:55:33] [I] Persistent Cache Ratio: 0
54
+ [01/04/2024-14:55:33] [I] Inputs:
55
+ [01/04/2024-14:55:33] [I] === Reporting Options ===
56
+ [01/04/2024-14:55:33] [I] Verbose: Disabled
57
+ [01/04/2024-14:55:33] [I] Averages: 100 inferences
58
+ [01/04/2024-14:55:33] [I] Percentiles: 90,95,99
59
+ [01/04/2024-14:55:33] [I] Dump refittable layers:Disabled
60
+ [01/04/2024-14:55:33] [I] Dump output: Disabled
61
+ [01/04/2024-14:55:33] [I] Profile: Disabled
62
+ [01/04/2024-14:55:33] [I] Export timing to JSON file:
63
+ [01/04/2024-14:55:33] [I] Export output to JSON file:
64
+ [01/04/2024-14:55:33] [I] Export profile to JSON file:
65
+ [01/04/2024-14:55:33] [I]
66
+ [01/04/2024-14:55:33] [I] === Device Information ===
67
+ [01/04/2024-14:55:33] [I] Selected Device: Orin
68
+ [01/04/2024-14:55:33] [I] Compute Capability: 8.7
69
+ [01/04/2024-14:55:33] [I] SMs: 8
70
+ [01/04/2024-14:55:33] [I] Compute Clock Rate: 0.624 GHz
71
+ [01/04/2024-14:55:33] [I] Device Global Memory: 7471 MiB
72
+ [01/04/2024-14:55:33] [I] Shared Memory per SM: 164 KiB
73
+ [01/04/2024-14:55:33] [I] Memory Bus Width: 128 bits (ECC disabled)
74
+ [01/04/2024-14:55:33] [I] Memory Clock Rate: 0.624 GHz
75
+ [01/04/2024-14:55:33] [I]
76
+ [01/04/2024-14:55:33] [I] TensorRT version: 8.5.2
77
+ [01/04/2024-14:55:34] [I] [TRT] [MemUsageChange] Init CUDA: CPU +220, GPU +0, now: CPU 249, GPU 3779 (MiB)
78
+ [01/04/2024-14:55:37] [I] [TRT] [MemUsageChange] Init builder kernel library: CPU +302, GPU +284, now: CPU 574, GPU 4083 (MiB)
79
+ [01/04/2024-14:55:37] [I] Start parsing network model
80
+ [01/04/2024-14:55:38] [I] [TRT] ----------------------------------------------------------------
81
+ [01/04/2024-14:55:38] [I] [TRT] Input filename: yolo_nas_pose_s_int8.onnx
82
+ [01/04/2024-14:55:38] [I] [TRT] ONNX IR version: 0.0.8
83
+ [01/04/2024-14:55:38] [I] [TRT] Opset version: 17
84
+ [01/04/2024-14:55:38] [I] [TRT] Producer name: pytorch
85
+ [01/04/2024-14:55:38] [I] [TRT] Producer version: 2.1.2
86
+ [01/04/2024-14:55:38] [I] [TRT] Domain:
87
+ [01/04/2024-14:55:38] [I] [TRT] Model version: 0
88
+ [01/04/2024-14:55:38] [I] [TRT] Doc string:
89
+ [01/04/2024-14:55:38] [I] [TRT] ----------------------------------------------------------------
90
+ [01/04/2024-14:55:41] [I] Finish parsing network model
91
+ [01/04/2024-14:55:45] [I] [TRT] ---------- Layers Running on DLA ----------
92
+ [01/04/2024-14:55:45] [I] [TRT] ---------- Layers Running on GPU ----------
93
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation1]
94
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/pre_process/pre_process.0/Cast.../pre_process/pre_process.2/Mul]}
95
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1229) [Constant]
96
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1230) [Constant]
97
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1231) [Constant]
98
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stem/conv/rbr_reparam/_input_quantizer/QuantizeLinear
99
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stem.conv.rbr_reparam.weight + /model/backbone/stem/conv/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stem/conv/rbr_reparam/Conv
100
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.downsample.rbr_reparam.weight + /model/backbone/stage1/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/downsample/rbr_reparam/Conv
101
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv2.conv.weight + /model/backbone/stage1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv2/conv/Conv
102
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv1.conv.weight + /model/backbone/stage1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv1/conv/Conv
103
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
104
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
105
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
106
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 494) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Add)
107
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
108
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
109
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
110
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 510) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Add)
111
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv3.conv.weight + /model/backbone/stage1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv3/conv/Conv
112
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_skip2.conv.weight + /model/neck/neck2/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip2/conv/Conv
113
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.downsample.rbr_reparam.weight + /model/backbone/stage2/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/downsample/rbr_reparam/Conv
114
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.downsample.conv.weight + /model/neck/neck2/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/downsample/conv/Conv
115
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv2.conv.weight + /model/backbone/stage2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv2/conv/Conv
116
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv1.conv.weight + /model/backbone/stage2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv1/conv/Conv
117
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
118
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
119
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
120
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 557) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Add)
121
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
122
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
123
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
124
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 573) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Add)
125
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
126
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
127
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
128
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.2.alpha + (Unnamed Layer* 589) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Add)
129
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv3.conv.weight + /model/backbone/stage2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv3/conv/Conv
130
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip2.conv.weight + /model/neck/neck1/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip2/conv/Conv || model.neck.neck2.reduce_skip1.conv.weight + /model/neck/neck2/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip1/conv/Conv
131
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.downsample.rbr_reparam.weight + /model/backbone/stage3/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/downsample/rbr_reparam/Conv
132
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.downsample.conv.weight + /model/neck/neck1/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/downsample/conv/Conv
133
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv2.conv.weight + /model/backbone/stage3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv2/conv/Conv
134
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv1.conv.weight + /model/backbone/stage3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv1/conv/Conv
135
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
136
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
137
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
138
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 639) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Add)
139
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
140
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
141
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
142
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 655) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Add)
143
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
144
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
145
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
146
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.2.alpha + (Unnamed Layer* 671) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Add)
147
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
148
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/Conv
149
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/Conv
150
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.3.alpha + (Unnamed Layer* 687) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Add)
151
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
152
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.4.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv1/rbr_reparam/Conv
153
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.4.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv2/rbr_reparam/Conv
154
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.4.alpha + (Unnamed Layer* 703) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/Add)
155
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv3.conv.weight + /model/backbone/stage3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv3/conv/Conv
156
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip1.conv.weight + /model/neck/neck1/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip1/conv/Conv
157
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.downsample.rbr_reparam.weight + /model/backbone/stage4/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/downsample/rbr_reparam/Conv
158
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv2.conv.weight + /model/backbone/stage4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv2/conv/Conv
159
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv1.conv.weight + /model/backbone/stage4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv1/conv/Conv
160
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
161
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
162
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
163
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 744) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Add)
164
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
165
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
166
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
167
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 760) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Add)
168
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv3.conv.weight + /model/backbone/stage4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv3/conv/Conv
169
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv1.conv.weight + /model/backbone/context_module/cv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv1/conv/Conv
170
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.2/MaxPool
171
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.1/MaxPool
172
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.0/MaxPool
173
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/context_module/m.2/MaxPool_output_0 copy
174
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv2.conv.weight + /model/backbone/context_module/cv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv2/conv/Conv
175
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.conv.conv.weight + /model/neck/neck1/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/conv/conv/Conv
176
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/upsample/_input_quantizer/QuantizeLinear
177
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck1.upsample.weight + /model/neck/neck1/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck1/upsample/ConvTranspose
178
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_after_concat.conv.weight + /model/neck/neck1/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_after_concat/conv/Conv
179
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv2.conv.weight + /model/neck/neck1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv2/conv/Conv
180
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv1.conv.weight + /model/neck/neck1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv1/conv/Conv
181
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
182
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
183
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
184
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 825) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Add)
185
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
186
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
187
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
188
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 841) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Add)
189
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv3.conv.weight + /model/neck/neck1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv3/conv/Conv
190
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.conv.conv.weight + /model/neck/neck2/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/conv/conv/Conv
191
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/upsample/_input_quantizer/QuantizeLinear
192
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck2.upsample.weight + /model/neck/neck2/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck2/upsample/ConvTranspose
193
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/Concat_/model/neck/neck2/reduce_skip1/act/Relu_output_0_clone_1 copy
194
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_after_concat.conv.weight + /model/neck/neck2/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_after_concat/conv/Conv
195
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv2.conv.weight + /model/neck/neck2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv2/conv/Conv
196
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv1.conv.weight + /model/neck/neck2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv1/conv/Conv
197
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
198
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
199
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
200
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 890) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Add)
201
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
202
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
203
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
204
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 906) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Add)
205
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/Concat_/model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
206
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv3.conv.weight + /model/neck/neck2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv3/conv/Conv
207
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.bbox_stem.seq.conv.weight + /model/heads/head1/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/bbox_stem/seq/conv/Conv || model.heads.head1.pose_stem.seq.conv.weight + /model/heads/head1/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_stem/seq/conv/Conv
208
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.conv.conv.weight + /model/neck/neck3/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/conv/conv/Conv
209
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_convs.0.seq.conv.weight + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head1.cls_convs.0.seq.conv.weight + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/Conv
210
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.0.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/Conv
211
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/conv1/conv/_input_quantizer/QuantizeLinear_clone_1
212
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.cls_pred.weight + /model/heads/head1/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_pred/Conv
213
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_pred.weight + /model/heads/head1/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_pred/Conv
214
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.1.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/Conv
215
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv2.conv.weight + /model/neck/neck3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv2/conv/Conv
216
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv1.conv.weight + /model/neck/neck3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv1/conv/Conv
217
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape + /model/heads/Transpose
218
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_pred.weight + /model/heads/head1/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_pred/Conv
219
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
220
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax
221
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
222
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv
223
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
224
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 972) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Add)
225
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
226
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
227
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
228
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 1013) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Add)
229
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv3.conv.weight + /model/neck/neck3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv3/conv/Conv
230
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_stem.seq.conv.weight + /model/heads/head2/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_stem/seq/conv/Conv || model.heads.head2.bbox_stem.seq.conv.weight + /model/heads/head2/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/bbox_stem/seq/conv/Conv
231
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.conv.conv.weight + /model/neck/neck4/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/conv/conv/Conv
232
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_convs.0.seq.conv.weight + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head2.cls_convs.0.seq.conv.weight + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/Conv
233
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.0.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/Conv
234
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/conv1/conv/_input_quantizer/QuantizeLinear_clone_1
235
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.cls_pred.weight + /model/heads/head2/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_pred/Conv
236
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_pred.weight + /model/heads/head2/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_pred/Conv
237
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.1.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/Conv
238
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv2.conv.weight + /model/neck/neck4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv2/conv/Conv
239
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv1.conv.weight + /model/neck/neck4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv1/conv/Conv
240
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_4 + /model/heads/Transpose_3
241
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_pred.weight + /model/heads/head2/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_pred/Conv
242
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
243
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_1
244
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
245
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_1
246
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
247
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 1079) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Add)
248
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
249
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
250
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
251
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 1120) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Add)
252
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv3.conv.weight + /model/neck/neck4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv3/conv/Conv
253
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.bbox_stem.seq.conv.weight + /model/heads/head3/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/bbox_stem/seq/conv/Conv || model.heads.head3.pose_stem.seq.conv.weight + /model/heads/head3/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_stem/seq/conv/Conv
254
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_convs.0.seq.conv.weight + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head3.cls_convs.0.seq.conv.weight + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/Conv
255
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.0.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/Conv
256
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.cls_pred.weight + /model/heads/head3/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_pred/Conv
257
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_pred.weight + /model/heads/head3/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_pred/Conv
258
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.1.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/Conv
259
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_8 + /model/heads/Transpose_6
260
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.2.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/Conv
261
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_2
262
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_pred.weight + /model/heads/head3/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_pred/Conv
263
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_2
264
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice_1.../post_process/Reshape_2]}
265
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] NMS: batched_nms_243
266
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] DEVICE_TO_SHAPE_HOST: (Unnamed Layer* 1233) [NMS]_1_output[DevicetoShapeHostCopy]
267
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation2]
268
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice...graph2_/Concat_5]}
269
+ [01/04/2024-14:55:45] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation3]
270
+ [01/04/2024-14:55:46] [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +534, GPU +459, now: CPU 1233, GPU 4661 (MiB)
271
+ [01/04/2024-14:55:46] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +82, GPU +38, now: CPU 1315, GPU 4699 (MiB)
272
+ [01/04/2024-14:55:46] [I] [TRT] Local timing cache in use. Profiling results in this builder pass will not be stored.
273
+ [01/04/2024-15:27:42] [I] [TRT] Total Activation Memory: 7917384192
274
+ [01/04/2024-15:27:42] [I] [TRT] Detected 1 inputs and 1 output network tensors.
275
+ [01/04/2024-15:27:50] [I] [TRT] Total Host Persistent Memory: 308448
276
+ [01/04/2024-15:27:50] [I] [TRT] Total Device Persistent Memory: 653824
277
+ [01/04/2024-15:27:50] [I] [TRT] Total Scratch Memory: 134217728
278
+ [01/04/2024-15:27:50] [I] [TRT] [MemUsageStats] Peak memory usage of TRT CPU/GPU memory allocators: CPU 74 MiB, GPU 154 MiB
279
+ [01/04/2024-15:27:50] [I] [TRT] [BlockAssignment] Started assigning block shifts. This will take 166 steps to complete.
280
+ [01/04/2024-15:27:50] [I] [TRT] [BlockAssignment] Algorithm ShiftNTopDown took 59.522ms to assign 13 blocks to 166 nodes requiring 141982720 bytes.
281
+ [01/04/2024-15:27:50] [I] [TRT] Total Activation Memory: 141982720
282
+ [01/04/2024-15:27:53] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1667, GPU 5744 (MiB)
283
+ [01/04/2024-15:27:53] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in building engine: CPU +15, GPU +16, now: CPU 15, GPU 16 (MiB)
284
+ [01/04/2024-15:27:53] [I] Engine built in 1940.22 sec.
285
+ [01/04/2024-15:27:54] [I] [TRT] Loaded engine size: 17 MiB
286
+ [01/04/2024-15:27:54] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1252, GPU 5595 (MiB)
287
+ [01/04/2024-15:27:54] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in engine deserialization: CPU +0, GPU +15, now: CPU 0, GPU 15 (MiB)
288
+ [01/04/2024-15:27:54] [I] Engine deserialized in 0.210553 sec.
289
+ [01/04/2024-15:27:54] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU -1, now: CPU 1252, GPU 5594 (MiB)
290
+ [01/04/2024-15:27:54] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +0, GPU +136, now: CPU 0, GPU 151 (MiB)
291
+ [01/04/2024-15:27:54] [I] Setting persistentCacheLimit to 0 bytes.
292
+ [01/04/2024-15:27:54] [I] Using random values for input onnx::Cast_0
293
+ [01/04/2024-15:27:54] [I] Created input binding for onnx::Cast_0 with dimensions 1x3x640x640
294
+ [01/04/2024-15:27:54] [I] Using random values for output graph2_flat_predictions
295
+ [01/04/2024-15:27:54] [I] Created output binding for graph2_flat_predictions with dimensions -1x57
296
+ [01/04/2024-15:27:54] [I] Starting inference
297
+ [01/04/2024-15:28:09] [I] Warmup completed 12 queries over 200 ms
298
+ [01/04/2024-15:28:09] [I] Timing trace has 1074 queries over 15.0266 s
299
+ [01/04/2024-15:28:09] [I]
300
+ [01/04/2024-15:28:09] [I] === Trace details ===
301
+ [01/04/2024-15:28:09] [I] Trace averages of 100 runs:
302
+ [01/04/2024-15:28:09] [I] Average on 100 runs - GPU latency: 13.6253 ms - Host latency: 13.7361 ms (enqueue 13.703 ms)
303
+ [01/04/2024-15:28:09] [I] Average on 100 runs - GPU latency: 13.9431 ms - Host latency: 14.0566 ms (enqueue 14.0098 ms)
304
+ [01/04/2024-15:28:09] [I] Average on 100 runs - GPU latency: 13.8369 ms - Host latency: 13.9494 ms (enqueue 13.9083 ms)
305
+ [01/04/2024-15:28:09] [I] Average on 100 runs - GPU latency: 13.8257 ms - Host latency: 13.9381 ms (enqueue 13.8989 ms)
306
+ [01/04/2024-15:28:09] [I] Average on 100 runs - GPU latency: 13.6064 ms - Host latency: 13.7172 ms (enqueue 13.6832 ms)
307
+ [01/04/2024-15:28:09] [I] Average on 100 runs - GPU latency: 14.264 ms - Host latency: 14.3781 ms (enqueue 14.3258 ms)
308
+ [01/04/2024-15:28:09] [I] Average on 100 runs - GPU latency: 13.6034 ms - Host latency: 13.7146 ms (enqueue 13.682 ms)
309
+ [01/04/2024-15:28:09] [I] Average on 100 runs - GPU latency: 14.1877 ms - Host latency: 14.3027 ms (enqueue 14.2525 ms)
310
+ [01/04/2024-15:28:09] [I] Average on 100 runs - GPU latency: 13.7484 ms - Host latency: 13.8601 ms (enqueue 13.8257 ms)
311
+ [01/04/2024-15:28:09] [I] Average on 100 runs - GPU latency: 13.7575 ms - Host latency: 13.8697 ms (enqueue 13.8349 ms)
312
+ [01/04/2024-15:28:09] [I]
313
+ [01/04/2024-15:28:09] [I] === Performance summary ===
314
+ [01/04/2024-15:28:09] [I] Throughput: 71.4732 qps
315
+ [01/04/2024-15:28:09] [I] Latency: min = 13.0068 ms, max = 17.7432 ms, mean = 13.9607 ms, median = 13.9542 ms, percentile(90%) = 14.7441 ms, percentile(95%) = 14.9595 ms, percentile(99%) = 15.5879 ms
316
+ [01/04/2024-15:28:09] [I] Enqueue Time: min = 12.9634 ms, max = 18.0693 ms, mean = 13.9208 ms, median = 13.9097 ms, percentile(90%) = 14.6982 ms, percentile(95%) = 14.8809 ms, percentile(99%) = 15.5361 ms
317
+ [01/04/2024-15:28:09] [I] H2D Latency: min = 0.0809937 ms, max = 0.114258 ms, mean = 0.0973303 ms, median = 0.0976562 ms, percentile(90%) = 0.0991211 ms, percentile(95%) = 0.0996094 ms, percentile(99%) = 0.101562 ms
318
+ [01/04/2024-15:28:09] [I] GPU Compute Time: min = 12.8984 ms, max = 17.6377 ms, mean = 13.8482 ms, median = 13.8396 ms, percentile(90%) = 14.6279 ms, percentile(95%) = 14.8496 ms, percentile(99%) = 15.4727 ms
319
+ [01/04/2024-15:28:09] [I] D2H Latency: min = 0.00390625 ms, max = 0.0466309 ms, mean = 0.0151338 ms, median = 0.0130615 ms, percentile(90%) = 0.0224609 ms, percentile(95%) = 0.0244141 ms, percentile(99%) = 0.03125 ms
320
+ [01/04/2024-15:28:09] [I] Total Host Walltime: 15.0266 s
321
+ [01/04/2024-15:28:09] [I] Total GPU Compute Time: 14.873 s
322
+ [01/04/2024-15:28:10] [I] Explanations of the performance metrics are printed in the verbose logs.
323
+ [01/04/2024-15:28:10] [I]
324
  &&&& PASSED TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_s_int8.onnx --best --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_s_int8.onnx.best.engine
yolo_nas_pose_s_int8.onnx.int8.engine CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b03544a46558d98432ec484cef728f0be750bcbb79762856b78125143998b2d
3
- size 17958627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8795eb441b2790005ae4c651d93cda57424a970bde3e5a5bceff132b34cf7c78
3
+ size 17990796
yolo_nas_pose_s_int8.onnx.int8.engine.err CHANGED
@@ -1,7 +1,7 @@
1
- [12/28/2023-03:36:18] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
- [12/28/2023-03:36:18] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
- [12/28/2023-03:36:21] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
- [12/28/2023-03:46:46] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
- [12/28/2023-03:46:46] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
- [12/28/2023-03:46:46] [W] * GPU compute time is unstable, with coefficient of variance = 3.57305%.
7
- [12/28/2023-03:46:46] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
 
1
+ [01/04/2024-15:28:15] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2
+ [01/04/2024-15:28:15] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
3
+ [01/04/2024-15:28:18] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
4
+ [01/04/2024-15:38:46] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
5
+ [01/04/2024-15:38:46] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
6
+ [01/04/2024-15:38:46] [W] * GPU compute time is unstable, with coefficient of variance = 4.08535%.
7
+ [01/04/2024-15:38:46] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.
yolo_nas_pose_s_int8.onnx.int8.engine.log CHANGED
@@ -1,322 +1,324 @@
1
  &&&& RUNNING TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_s_int8.onnx --int8 --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_s_int8.onnx.int8.engine
2
- [12/28/2023-03:36:14] [I] === Model Options ===
3
- [12/28/2023-03:36:14] [I] Format: ONNX
4
- [12/28/2023-03:36:14] [I] Model: yolo_nas_pose_s_int8.onnx
5
- [12/28/2023-03:36:14] [I] Output:
6
- [12/28/2023-03:36:14] [I] === Build Options ===
7
- [12/28/2023-03:36:14] [I] Max batch: explicit batch
8
- [12/28/2023-03:36:14] [I] Memory Pools: workspace: default, dlaSRAM: default, dlaLocalDRAM: default, dlaGlobalDRAM: default
9
- [12/28/2023-03:36:14] [I] minTiming: 1
10
- [12/28/2023-03:36:14] [I] avgTiming: 8
11
- [12/28/2023-03:36:14] [I] Precision: FP32+INT8
12
- [12/28/2023-03:36:14] [I] LayerPrecisions:
13
- [12/28/2023-03:36:14] [I] Calibration: Dynamic
14
- [12/28/2023-03:36:14] [I] Refit: Disabled
15
- [12/28/2023-03:36:14] [I] Sparsity: Disabled
16
- [12/28/2023-03:36:14] [I] Safe mode: Disabled
17
- [12/28/2023-03:36:14] [I] DirectIO mode: Disabled
18
- [12/28/2023-03:36:14] [I] Restricted mode: Disabled
19
- [12/28/2023-03:36:14] [I] Build only: Disabled
20
- [12/28/2023-03:36:14] [I] Save engine: yolo_nas_pose_s_int8.onnx.int8.engine
21
- [12/28/2023-03:36:14] [I] Load engine:
22
- [12/28/2023-03:36:14] [I] Profiling verbosity: 0
23
- [12/28/2023-03:36:14] [I] Tactic sources: Using default tactic sources
24
- [12/28/2023-03:36:14] [I] timingCacheMode: local
25
- [12/28/2023-03:36:14] [I] timingCacheFile:
26
- [12/28/2023-03:36:14] [I] Heuristic: Disabled
27
- [12/28/2023-03:36:14] [I] Preview Features: Use default preview flags.
28
- [12/28/2023-03:36:14] [I] Input(s)s format: fp32:CHW
29
- [12/28/2023-03:36:14] [I] Output(s)s format: fp32:CHW
30
- [12/28/2023-03:36:14] [I] Input build shapes: model
31
- [12/28/2023-03:36:14] [I] Input calibration shapes: model
32
- [12/28/2023-03:36:14] [I] === System Options ===
33
- [12/28/2023-03:36:14] [I] Device: 0
34
- [12/28/2023-03:36:14] [I] DLACore:
35
- [12/28/2023-03:36:14] [I] Plugins:
36
- [12/28/2023-03:36:14] [I] === Inference Options ===
37
- [12/28/2023-03:36:14] [I] Batch: Explicit
38
- [12/28/2023-03:36:14] [I] Input inference shapes: model
39
- [12/28/2023-03:36:14] [I] Iterations: 10
40
- [12/28/2023-03:36:14] [I] Duration: 15s (+ 200ms warm up)
41
- [12/28/2023-03:36:14] [I] Sleep time: 0ms
42
- [12/28/2023-03:36:14] [I] Idle time: 0ms
43
- [12/28/2023-03:36:14] [I] Streams: 1
44
- [12/28/2023-03:36:14] [I] ExposeDMA: Disabled
45
- [12/28/2023-03:36:14] [I] Data transfers: Enabled
46
- [12/28/2023-03:36:14] [I] Spin-wait: Disabled
47
- [12/28/2023-03:36:14] [I] Multithreading: Disabled
48
- [12/28/2023-03:36:14] [I] CUDA Graph: Disabled
49
- [12/28/2023-03:36:14] [I] Separate profiling: Disabled
50
- [12/28/2023-03:36:14] [I] Time Deserialize: Disabled
51
- [12/28/2023-03:36:14] [I] Time Refit: Disabled
52
- [12/28/2023-03:36:14] [I] NVTX verbosity: 0
53
- [12/28/2023-03:36:14] [I] Persistent Cache Ratio: 0
54
- [12/28/2023-03:36:14] [I] Inputs:
55
- [12/28/2023-03:36:14] [I] === Reporting Options ===
56
- [12/28/2023-03:36:14] [I] Verbose: Disabled
57
- [12/28/2023-03:36:14] [I] Averages: 100 inferences
58
- [12/28/2023-03:36:14] [I] Percentiles: 90,95,99
59
- [12/28/2023-03:36:14] [I] Dump refittable layers:Disabled
60
- [12/28/2023-03:36:14] [I] Dump output: Disabled
61
- [12/28/2023-03:36:14] [I] Profile: Disabled
62
- [12/28/2023-03:36:14] [I] Export timing to JSON file:
63
- [12/28/2023-03:36:14] [I] Export output to JSON file:
64
- [12/28/2023-03:36:14] [I] Export profile to JSON file:
65
- [12/28/2023-03:36:14] [I]
66
- [12/28/2023-03:36:14] [I] === Device Information ===
67
- [12/28/2023-03:36:14] [I] Selected Device: Orin
68
- [12/28/2023-03:36:14] [I] Compute Capability: 8.7
69
- [12/28/2023-03:36:14] [I] SMs: 8
70
- [12/28/2023-03:36:14] [I] Compute Clock Rate: 0.624 GHz
71
- [12/28/2023-03:36:14] [I] Device Global Memory: 7471 MiB
72
- [12/28/2023-03:36:14] [I] Shared Memory per SM: 164 KiB
73
- [12/28/2023-03:36:14] [I] Memory Bus Width: 128 bits (ECC disabled)
74
- [12/28/2023-03:36:14] [I] Memory Clock Rate: 0.624 GHz
75
- [12/28/2023-03:36:14] [I]
76
- [12/28/2023-03:36:14] [I] TensorRT version: 8.5.2
77
- [12/28/2023-03:36:15] [I] [TRT] [MemUsageChange] Init CUDA: CPU +220, GPU +0, now: CPU 249, GPU 2931 (MiB)
78
- [12/28/2023-03:36:17] [I] [TRT] [MemUsageChange] Init builder kernel library: CPU +302, GPU +285, now: CPU 574, GPU 3236 (MiB)
79
- [12/28/2023-03:36:18] [I] Start parsing network model
80
- [12/28/2023-03:36:18] [I] [TRT] ----------------------------------------------------------------
81
- [12/28/2023-03:36:18] [I] [TRT] Input filename: yolo_nas_pose_s_int8.onnx
82
- [12/28/2023-03:36:18] [I] [TRT] ONNX IR version: 0.0.8
83
- [12/28/2023-03:36:18] [I] [TRT] Opset version: 17
84
- [12/28/2023-03:36:18] [I] [TRT] Producer name: pytorch
85
- [12/28/2023-03:36:18] [I] [TRT] Producer version: 2.1.2
86
- [12/28/2023-03:36:18] [I] [TRT] Domain:
87
- [12/28/2023-03:36:18] [I] [TRT] Model version: 0
88
- [12/28/2023-03:36:18] [I] [TRT] Doc string:
89
- [12/28/2023-03:36:18] [I] [TRT] ----------------------------------------------------------------
90
- [12/28/2023-03:36:21] [I] Finish parsing network model
91
- [12/28/2023-03:36:21] [I] FP32 and INT8 precisions have been specified - more performance might be enabled by additionally specifying --fp16 or --best
92
- [12/28/2023-03:36:24] [I] [TRT] ---------- Layers Running on DLA ----------
93
- [12/28/2023-03:36:24] [I] [TRT] ---------- Layers Running on GPU ----------
94
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation1]
95
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/pre_process/pre_process.0/Cast.../pre_process/pre_process.2/Mul]}
96
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1228) [Constant]
97
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1229) [Constant]
98
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1230) [Constant]
99
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/backbone/stem/conv/rbr_reparam/_input_quantizer/QuantizeLinear
100
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stem.conv.rbr_reparam.weight + /model/backbone/stem/conv/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stem/conv/rbr_reparam/Conv
101
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.downsample.rbr_reparam.weight + /model/backbone/stage1/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/downsample/rbr_reparam/Conv
102
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv2.conv.weight + /model/backbone/stage1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv2/conv/Conv
103
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv1.conv.weight + /model/backbone/stage1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv1/conv/Conv
104
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
105
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
106
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
107
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 494) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Add)
108
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
109
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
110
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
111
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 510) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Add)
112
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv3.conv.weight + /model/backbone/stage1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv3/conv/Conv
113
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_skip2.conv.weight + /model/neck/neck2/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip2/conv/Conv
114
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.downsample.rbr_reparam.weight + /model/backbone/stage2/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/downsample/rbr_reparam/Conv
115
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.downsample.conv.weight + /model/neck/neck2/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/downsample/conv/Conv
116
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv2.conv.weight + /model/backbone/stage2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv2/conv/Conv
117
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv1.conv.weight + /model/backbone/stage2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv1/conv/Conv
118
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
119
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
120
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
121
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 557) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Add)
122
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
123
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
124
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
125
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 573) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Add)
126
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
127
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
128
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
129
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.2.alpha + (Unnamed Layer* 589) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Add)
130
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv3.conv.weight + /model/backbone/stage2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv3/conv/Conv
131
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip2.conv.weight + /model/neck/neck1/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip2/conv/Conv || model.neck.neck2.reduce_skip1.conv.weight + /model/neck/neck2/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip1/conv/Conv
132
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.downsample.rbr_reparam.weight + /model/backbone/stage3/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/downsample/rbr_reparam/Conv
133
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.downsample.conv.weight + /model/neck/neck1/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/downsample/conv/Conv
134
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv2.conv.weight + /model/backbone/stage3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv2/conv/Conv
135
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv1.conv.weight + /model/backbone/stage3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv1/conv/Conv
136
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
137
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
138
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
139
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 639) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Add)
140
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
141
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
142
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
143
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 655) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Add)
144
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
145
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
146
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
147
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.2.alpha + (Unnamed Layer* 671) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Add)
148
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
149
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/Conv
150
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/Conv
151
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.3.alpha + (Unnamed Layer* 687) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Add)
152
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
153
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.4.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv1/rbr_reparam/Conv
154
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.4.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv2/rbr_reparam/Conv
155
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.4.alpha + (Unnamed Layer* 703) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/Add)
156
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv3.conv.weight + /model/backbone/stage3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv3/conv/Conv
157
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip1.conv.weight + /model/neck/neck1/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip1/conv/Conv
158
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.downsample.rbr_reparam.weight + /model/backbone/stage4/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/downsample/rbr_reparam/Conv
159
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv2.conv.weight + /model/backbone/stage4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv2/conv/Conv
160
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv1.conv.weight + /model/backbone/stage4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv1/conv/Conv
161
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
162
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
163
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
164
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 744) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Add)
165
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
166
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
167
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
168
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 760) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Add)
169
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv3.conv.weight + /model/backbone/stage4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv3/conv/Conv
170
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv1.conv.weight + /model/backbone/context_module/cv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv1/conv/Conv
171
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.2/MaxPool
172
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.1/MaxPool
173
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.0/MaxPool
174
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/backbone/context_module/m.2/MaxPool_output_0 copy
175
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv2.conv.weight + /model/backbone/context_module/cv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv2/conv/Conv
176
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.conv.conv.weight + /model/neck/neck1/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/conv/conv/Conv
177
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck1.upsample.weight + /model/neck/neck1/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck1/upsample/ConvTranspose
178
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_after_concat.conv.weight + /model/neck/neck1/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_after_concat/conv/Conv
179
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv2.conv.weight + /model/neck/neck1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv2/conv/Conv
180
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv1.conv.weight + /model/neck/neck1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv1/conv/Conv
181
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
182
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
183
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
184
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 825) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Add)
185
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
186
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
187
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
188
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 841) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Add)
189
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv3.conv.weight + /model/neck/neck1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv3/conv/Conv
190
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.conv.conv.weight + /model/neck/neck2/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/conv/conv/Conv
191
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/upsample/_input_quantizer/QuantizeLinear
192
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck2.upsample.weight + /model/neck/neck2/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck2/upsample/ConvTranspose
193
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/Concat_/model/neck/neck2/reduce_skip1/act/Relu_output_0_clone_1 copy
194
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_after_concat.conv.weight + /model/neck/neck2/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_after_concat/conv/Conv
195
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv2.conv.weight + /model/neck/neck2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv2/conv/Conv
196
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv1.conv.weight + /model/neck/neck2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv1/conv/Conv
197
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
198
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
199
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
200
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 890) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Add)
201
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
202
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
203
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
204
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 906) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Add)
205
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/Concat_/model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
206
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv3.conv.weight + /model/neck/neck2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv3/conv/Conv
207
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.bbox_stem.seq.conv.weight + /model/heads/head1/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/bbox_stem/seq/conv/Conv || model.heads.head1.pose_stem.seq.conv.weight + /model/heads/head1/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_stem/seq/conv/Conv
208
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.conv.conv.weight + /model/neck/neck3/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/conv/conv/Conv
209
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_convs.0.seq.conv.weight + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head1.cls_convs.0.seq.conv.weight + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/Conv
210
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.0.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/Conv
211
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/conv1/conv/_input_quantizer/QuantizeLinear_clone_1
212
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.cls_pred.weight + /model/heads/head1/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_pred/Conv
213
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_pred.weight + /model/heads/head1/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_pred/Conv
214
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.1.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/Conv
215
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv2.conv.weight + /model/neck/neck3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv2/conv/Conv
216
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv1.conv.weight + /model/neck/neck3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv1/conv/Conv
217
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape + /model/heads/Transpose
218
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_pred.weight + /model/heads/head1/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_pred/Conv
219
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
220
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax
221
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
222
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv
223
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
224
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 972) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Add)
225
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
226
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
227
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
228
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 1013) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Add)
229
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv3.conv.weight + /model/neck/neck3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv3/conv/Conv
230
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_stem.seq.conv.weight + /model/heads/head2/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_stem/seq/conv/Conv || model.heads.head2.bbox_stem.seq.conv.weight + /model/heads/head2/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/bbox_stem/seq/conv/Conv
231
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.conv.conv.weight + /model/neck/neck4/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/conv/conv/Conv
232
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_convs.0.seq.conv.weight + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head2.cls_convs.0.seq.conv.weight + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/Conv
233
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.0.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/Conv
234
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv2.conv.weight + /model/neck/neck4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv2/conv/Conv
235
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv1.conv.weight + /model/neck/neck4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv1/conv/Conv
236
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.cls_pred.weight + /model/heads/head2/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_pred/Conv
237
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_pred.weight + /model/heads/head2/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_pred/Conv
238
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.1.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/Conv
239
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
240
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_4 + /model/heads/Transpose_3
241
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_pred.weight + /model/heads/head2/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_pred/Conv
242
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
243
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_1
244
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
245
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_1
246
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 1078) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Add)
247
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
248
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
249
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
250
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 1119) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Add)
251
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv3.conv.weight + /model/neck/neck4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv3/conv/Conv
252
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.bbox_stem.seq.conv.weight + /model/heads/head3/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/bbox_stem/seq/conv/Conv || model.heads.head3.pose_stem.seq.conv.weight + /model/heads/head3/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_stem/seq/conv/Conv
253
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_convs.0.seq.conv.weight + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head3.cls_convs.0.seq.conv.weight + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/Conv
254
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.0.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/Conv
255
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.cls_pred.weight + /model/heads/head3/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_pred/Conv
256
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_pred.weight + /model/heads/head3/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_pred/Conv
257
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.1.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/Conv
258
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_8 + /model/heads/Transpose_6
259
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.2.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/Conv
260
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_2
261
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_pred.weight + /model/heads/head3/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_pred/Conv
262
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_2
263
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice_1.../post_process/Reshape_2]}
264
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] NMS: batched_nms_243
265
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] DEVICE_TO_SHAPE_HOST: (Unnamed Layer* 1232) [NMS]_1_output[DevicetoShapeHostCopy]
266
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation2]
267
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice...graph2_/Concat_5]}
268
- [12/28/2023-03:36:24] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation3]
269
- [12/28/2023-03:36:26] [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +534, GPU +773, now: CPU 1233, GPU 4134 (MiB)
270
- [12/28/2023-03:36:26] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +82, GPU +118, now: CPU 1315, GPU 4252 (MiB)
271
- [12/28/2023-03:36:26] [I] [TRT] Local timing cache in use. Profiling results in this builder pass will not be stored.
272
- [12/28/2023-03:46:28] [I] [TRT] Total Activation Memory: 7940674048
273
- [12/28/2023-03:46:28] [I] [TRT] Detected 1 inputs and 1 output network tensors.
274
- [12/28/2023-03:46:29] [I] [TRT] Total Host Persistent Memory: 307232
275
- [12/28/2023-03:46:29] [I] [TRT] Total Device Persistent Memory: 38912
276
- [12/28/2023-03:46:29] [I] [TRT] Total Scratch Memory: 134217728
277
- [12/28/2023-03:46:29] [I] [TRT] [MemUsageStats] Peak memory usage of TRT CPU/GPU memory allocators: CPU 74 MiB, GPU 132 MiB
278
- [12/28/2023-03:46:29] [I] [TRT] [BlockAssignment] Started assigning block shifts. This will take 173 steps to complete.
279
- [12/28/2023-03:46:29] [I] [TRT] [BlockAssignment] Algorithm ShiftNTopDown took 42.4134ms to assign 13 blocks to 173 nodes requiring 144692224 bytes.
280
- [12/28/2023-03:46:29] [I] [TRT] Total Activation Memory: 144692224
281
- [12/28/2023-03:46:30] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +6, now: CPU 1665, GPU 5446 (MiB)
282
- [12/28/2023-03:46:30] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in building engine: CPU +15, GPU +16, now: CPU 15, GPU 16 (MiB)
283
- [12/28/2023-03:46:31] [I] Engine built in 616.071 sec.
284
- [12/28/2023-03:46:31] [I] [TRT] Loaded engine size: 17 MiB
285
- [12/28/2023-03:46:31] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +12, now: CPU 1250, GPU 5268 (MiB)
286
- [12/28/2023-03:46:31] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in engine deserialization: CPU +0, GPU +15, now: CPU 0, GPU 15 (MiB)
287
- [12/28/2023-03:46:31] [I] Engine deserialized in 0.135656 sec.
288
- [12/28/2023-03:46:31] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +7, now: CPU 1251, GPU 5268 (MiB)
289
- [12/28/2023-03:46:31] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +0, GPU +138, now: CPU 0, GPU 153 (MiB)
290
- [12/28/2023-03:46:31] [I] Setting persistentCacheLimit to 0 bytes.
291
- [12/28/2023-03:46:31] [I] Using random values for input onnx::Cast_0
292
- [12/28/2023-03:46:31] [I] Created input binding for onnx::Cast_0 with dimensions 1x3x640x640
293
- [12/28/2023-03:46:31] [I] Using random values for output graph2_flat_predictions
294
- [12/28/2023-03:46:31] [I] Created output binding for graph2_flat_predictions with dimensions -1x57
295
- [12/28/2023-03:46:31] [I] Starting inference
296
- [12/28/2023-03:46:46] [I] Warmup completed 8 queries over 200 ms
297
- [12/28/2023-03:46:46] [I] Timing trace has 959 queries over 15.0421 s
298
- [12/28/2023-03:46:46] [I]
299
- [12/28/2023-03:46:46] [I] === Trace details ===
300
- [12/28/2023-03:46:46] [I] Trace averages of 100 runs:
301
- [12/28/2023-03:46:46] [I] Average on 100 runs - GPU latency: 15.6862 ms - Host latency: 15.799 ms (enqueue 15.7577 ms)
302
- [12/28/2023-03:46:46] [I] Average on 100 runs - GPU latency: 15.8638 ms - Host latency: 15.9795 ms (enqueue 15.9265 ms)
303
- [12/28/2023-03:46:46] [I] Average on 100 runs - GPU latency: 16.1783 ms - Host latency: 16.2971 ms (enqueue 16.24 ms)
304
- [12/28/2023-03:46:46] [I] Average on 100 runs - GPU latency: 15.4097 ms - Host latency: 15.5201 ms (enqueue 15.487 ms)
305
- [12/28/2023-03:46:46] [I] Average on 100 runs - GPU latency: 15.3654 ms - Host latency: 15.4758 ms (enqueue 15.4444 ms)
306
- [12/28/2023-03:46:46] [I] Average on 100 runs - GPU latency: 15.4303 ms - Host latency: 15.5405 ms (enqueue 15.5048 ms)
307
- [12/28/2023-03:46:46] [I] Average on 100 runs - GPU latency: 15.388 ms - Host latency: 15.5003 ms (enqueue 15.463 ms)
308
- [12/28/2023-03:46:46] [I] Average on 100 runs - GPU latency: 15.3653 ms - Host latency: 15.4758 ms (enqueue 15.4443 ms)
309
- [12/28/2023-03:46:46] [I] Average on 100 runs - GPU latency: 15.3726 ms - Host latency: 15.4828 ms (enqueue 15.4505 ms)
310
- [12/28/2023-03:46:46] [I]
311
- [12/28/2023-03:46:46] [I] === Performance summary ===
312
- [12/28/2023-03:46:46] [I] Throughput: 63.7546 qps
313
- [12/28/2023-03:46:46] [I] Latency: min = 14.4111 ms, max = 20.8093 ms, mean = 15.6545 ms, median = 15.4062 ms, percentile(90%) = 16.3127 ms, percentile(95%) = 16.4324 ms, percentile(99%) = 17.4062 ms
314
- [12/28/2023-03:46:46] [I] Enqueue Time: min = 14.3828 ms, max = 20.7617 ms, mean = 15.6158 ms, median = 15.3755 ms, percentile(90%) = 16.2577 ms, percentile(95%) = 16.3701 ms, percentile(99%) = 17.3357 ms
315
- [12/28/2023-03:46:46] [I] H2D Latency: min = 0.0805664 ms, max = 0.117188 ms, mean = 0.0972066 ms, median = 0.0976562 ms, percentile(90%) = 0.0996094 ms, percentile(95%) = 0.0998535 ms, percentile(99%) = 0.111328 ms
316
- [12/28/2023-03:46:46] [I] GPU Compute Time: min = 14.3018 ms, max = 20.6875 ms, mean = 15.5423 ms, median = 15.2939 ms, percentile(90%) = 16.1956 ms, percentile(95%) = 16.3105 ms, percentile(99%) = 17.293 ms
317
- [12/28/2023-03:46:46] [I] D2H Latency: min = 0.00292969 ms, max = 0.0727539 ms, mean = 0.014973 ms, median = 0.0120239 ms, percentile(90%) = 0.0234375 ms, percentile(95%) = 0.0253906 ms, percentile(99%) = 0.0373535 ms
318
- [12/28/2023-03:46:46] [I] Total Host Walltime: 15.0421 s
319
- [12/28/2023-03:46:46] [I] Total GPU Compute Time: 14.9051 s
320
- [12/28/2023-03:46:46] [I] Explanations of the performance metrics are printed in the verbose logs.
321
- [12/28/2023-03:46:46] [I]
 
 
322
  &&&& PASSED TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_s_int8.onnx --int8 --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_s_int8.onnx.int8.engine
 
1
  &&&& RUNNING TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_s_int8.onnx --int8 --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_s_int8.onnx.int8.engine
2
+ [01/04/2024-15:28:11] [I] === Model Options ===
3
+ [01/04/2024-15:28:11] [I] Format: ONNX
4
+ [01/04/2024-15:28:11] [I] Model: yolo_nas_pose_s_int8.onnx
5
+ [01/04/2024-15:28:11] [I] Output:
6
+ [01/04/2024-15:28:11] [I] === Build Options ===
7
+ [01/04/2024-15:28:11] [I] Max batch: explicit batch
8
+ [01/04/2024-15:28:11] [I] Memory Pools: workspace: default, dlaSRAM: default, dlaLocalDRAM: default, dlaGlobalDRAM: default
9
+ [01/04/2024-15:28:11] [I] minTiming: 1
10
+ [01/04/2024-15:28:11] [I] avgTiming: 8
11
+ [01/04/2024-15:28:11] [I] Precision: FP32+INT8
12
+ [01/04/2024-15:28:11] [I] LayerPrecisions:
13
+ [01/04/2024-15:28:11] [I] Calibration: Dynamic
14
+ [01/04/2024-15:28:11] [I] Refit: Disabled
15
+ [01/04/2024-15:28:11] [I] Sparsity: Disabled
16
+ [01/04/2024-15:28:11] [I] Safe mode: Disabled
17
+ [01/04/2024-15:28:11] [I] DirectIO mode: Disabled
18
+ [01/04/2024-15:28:11] [I] Restricted mode: Disabled
19
+ [01/04/2024-15:28:11] [I] Build only: Disabled
20
+ [01/04/2024-15:28:11] [I] Save engine: yolo_nas_pose_s_int8.onnx.int8.engine
21
+ [01/04/2024-15:28:11] [I] Load engine:
22
+ [01/04/2024-15:28:11] [I] Profiling verbosity: 0
23
+ [01/04/2024-15:28:11] [I] Tactic sources: Using default tactic sources
24
+ [01/04/2024-15:28:11] [I] timingCacheMode: local
25
+ [01/04/2024-15:28:11] [I] timingCacheFile:
26
+ [01/04/2024-15:28:11] [I] Heuristic: Disabled
27
+ [01/04/2024-15:28:11] [I] Preview Features: Use default preview flags.
28
+ [01/04/2024-15:28:11] [I] Input(s)s format: fp32:CHW
29
+ [01/04/2024-15:28:11] [I] Output(s)s format: fp32:CHW
30
+ [01/04/2024-15:28:11] [I] Input build shapes: model
31
+ [01/04/2024-15:28:11] [I] Input calibration shapes: model
32
+ [01/04/2024-15:28:11] [I] === System Options ===
33
+ [01/04/2024-15:28:11] [I] Device: 0
34
+ [01/04/2024-15:28:11] [I] DLACore:
35
+ [01/04/2024-15:28:11] [I] Plugins:
36
+ [01/04/2024-15:28:11] [I] === Inference Options ===
37
+ [01/04/2024-15:28:11] [I] Batch: Explicit
38
+ [01/04/2024-15:28:11] [I] Input inference shapes: model
39
+ [01/04/2024-15:28:11] [I] Iterations: 10
40
+ [01/04/2024-15:28:11] [I] Duration: 15s (+ 200ms warm up)
41
+ [01/04/2024-15:28:11] [I] Sleep time: 0ms
42
+ [01/04/2024-15:28:11] [I] Idle time: 0ms
43
+ [01/04/2024-15:28:11] [I] Streams: 1
44
+ [01/04/2024-15:28:11] [I] ExposeDMA: Disabled
45
+ [01/04/2024-15:28:11] [I] Data transfers: Enabled
46
+ [01/04/2024-15:28:11] [I] Spin-wait: Disabled
47
+ [01/04/2024-15:28:11] [I] Multithreading: Disabled
48
+ [01/04/2024-15:28:11] [I] CUDA Graph: Disabled
49
+ [01/04/2024-15:28:11] [I] Separate profiling: Disabled
50
+ [01/04/2024-15:28:11] [I] Time Deserialize: Disabled
51
+ [01/04/2024-15:28:11] [I] Time Refit: Disabled
52
+ [01/04/2024-15:28:11] [I] NVTX verbosity: 0
53
+ [01/04/2024-15:28:11] [I] Persistent Cache Ratio: 0
54
+ [01/04/2024-15:28:11] [I] Inputs:
55
+ [01/04/2024-15:28:11] [I] === Reporting Options ===
56
+ [01/04/2024-15:28:11] [I] Verbose: Disabled
57
+ [01/04/2024-15:28:11] [I] Averages: 100 inferences
58
+ [01/04/2024-15:28:11] [I] Percentiles: 90,95,99
59
+ [01/04/2024-15:28:11] [I] Dump refittable layers:Disabled
60
+ [01/04/2024-15:28:11] [I] Dump output: Disabled
61
+ [01/04/2024-15:28:11] [I] Profile: Disabled
62
+ [01/04/2024-15:28:11] [I] Export timing to JSON file:
63
+ [01/04/2024-15:28:11] [I] Export output to JSON file:
64
+ [01/04/2024-15:28:11] [I] Export profile to JSON file:
65
+ [01/04/2024-15:28:11] [I]
66
+ [01/04/2024-15:28:11] [I] === Device Information ===
67
+ [01/04/2024-15:28:11] [I] Selected Device: Orin
68
+ [01/04/2024-15:28:11] [I] Compute Capability: 8.7
69
+ [01/04/2024-15:28:11] [I] SMs: 8
70
+ [01/04/2024-15:28:11] [I] Compute Clock Rate: 0.624 GHz
71
+ [01/04/2024-15:28:11] [I] Device Global Memory: 7471 MiB
72
+ [01/04/2024-15:28:11] [I] Shared Memory per SM: 164 KiB
73
+ [01/04/2024-15:28:11] [I] Memory Bus Width: 128 bits (ECC disabled)
74
+ [01/04/2024-15:28:11] [I] Memory Clock Rate: 0.624 GHz
75
+ [01/04/2024-15:28:11] [I]
76
+ [01/04/2024-15:28:11] [I] TensorRT version: 8.5.2
77
+ [01/04/2024-15:28:11] [I] [TRT] [MemUsageChange] Init CUDA: CPU +220, GPU +0, now: CPU 249, GPU 3760 (MiB)
78
+ [01/04/2024-15:28:15] [I] [TRT] [MemUsageChange] Init builder kernel library: CPU +302, GPU +284, now: CPU 574, GPU 4064 (MiB)
79
+ [01/04/2024-15:28:15] [I] Start parsing network model
80
+ [01/04/2024-15:28:15] [I] [TRT] ----------------------------------------------------------------
81
+ [01/04/2024-15:28:15] [I] [TRT] Input filename: yolo_nas_pose_s_int8.onnx
82
+ [01/04/2024-15:28:15] [I] [TRT] ONNX IR version: 0.0.8
83
+ [01/04/2024-15:28:15] [I] [TRT] Opset version: 17
84
+ [01/04/2024-15:28:15] [I] [TRT] Producer name: pytorch
85
+ [01/04/2024-15:28:15] [I] [TRT] Producer version: 2.1.2
86
+ [01/04/2024-15:28:15] [I] [TRT] Domain:
87
+ [01/04/2024-15:28:15] [I] [TRT] Model version: 0
88
+ [01/04/2024-15:28:15] [I] [TRT] Doc string:
89
+ [01/04/2024-15:28:15] [I] [TRT] ----------------------------------------------------------------
90
+ [01/04/2024-15:28:18] [I] Finish parsing network model
91
+ [01/04/2024-15:28:18] [I] FP32 and INT8 precisions have been specified - more performance might be enabled by additionally specifying --fp16 or --best
92
+ [01/04/2024-15:28:22] [I] [TRT] ---------- Layers Running on DLA ----------
93
+ [01/04/2024-15:28:22] [I] [TRT] ---------- Layers Running on GPU ----------
94
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation1]
95
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/pre_process/pre_process.0/Cast.../pre_process/pre_process.2/Mul]}
96
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1229) [Constant]
97
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1230) [Constant]
98
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1231) [Constant]
99
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/backbone/stem/conv/rbr_reparam/_input_quantizer/QuantizeLinear
100
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stem.conv.rbr_reparam.weight + /model/backbone/stem/conv/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stem/conv/rbr_reparam/Conv
101
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.downsample.rbr_reparam.weight + /model/backbone/stage1/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/downsample/rbr_reparam/Conv
102
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv2.conv.weight + /model/backbone/stage1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv2/conv/Conv
103
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv1.conv.weight + /model/backbone/stage1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv1/conv/Conv
104
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
105
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
106
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
107
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 494) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Add)
108
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
109
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
110
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
111
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 510) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Add)
112
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv3.conv.weight + /model/backbone/stage1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv3/conv/Conv
113
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_skip2.conv.weight + /model/neck/neck2/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip2/conv/Conv
114
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.downsample.rbr_reparam.weight + /model/backbone/stage2/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/downsample/rbr_reparam/Conv
115
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.downsample.conv.weight + /model/neck/neck2/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/downsample/conv/Conv
116
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv2.conv.weight + /model/backbone/stage2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv2/conv/Conv
117
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv1.conv.weight + /model/backbone/stage2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv1/conv/Conv
118
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
119
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
120
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
121
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 557) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Add)
122
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
123
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
124
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
125
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 573) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Add)
126
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
127
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
128
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
129
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.2.alpha + (Unnamed Layer* 589) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Add)
130
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv3.conv.weight + /model/backbone/stage2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv3/conv/Conv
131
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip2.conv.weight + /model/neck/neck1/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip2/conv/Conv || model.neck.neck2.reduce_skip1.conv.weight + /model/neck/neck2/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip1/conv/Conv
132
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.downsample.rbr_reparam.weight + /model/backbone/stage3/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/downsample/rbr_reparam/Conv
133
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.downsample.conv.weight + /model/neck/neck1/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/downsample/conv/Conv
134
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv2.conv.weight + /model/backbone/stage3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv2/conv/Conv
135
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv1.conv.weight + /model/backbone/stage3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv1/conv/Conv
136
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
137
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
138
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
139
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 639) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Add)
140
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
141
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
142
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
143
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 655) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Add)
144
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
145
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
146
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
147
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.2.alpha + (Unnamed Layer* 671) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Add)
148
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
149
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/Conv
150
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/Conv
151
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.3.alpha + (Unnamed Layer* 687) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Add)
152
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
153
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.4.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv1/rbr_reparam/Conv
154
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.4.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv2/rbr_reparam/Conv
155
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.4.alpha + (Unnamed Layer* 703) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/Add)
156
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv3.conv.weight + /model/backbone/stage3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv3/conv/Conv
157
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip1.conv.weight + /model/neck/neck1/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip1/conv/Conv
158
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.downsample.rbr_reparam.weight + /model/backbone/stage4/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/downsample/rbr_reparam/Conv
159
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv2.conv.weight + /model/backbone/stage4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv2/conv/Conv
160
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv1.conv.weight + /model/backbone/stage4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv1/conv/Conv
161
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
162
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
163
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
164
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 744) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Add)
165
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
166
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
167
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
168
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 760) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Add)
169
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv3.conv.weight + /model/backbone/stage4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv3/conv/Conv
170
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv1.conv.weight + /model/backbone/context_module/cv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv1/conv/Conv
171
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.2/MaxPool
172
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.1/MaxPool
173
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.0/MaxPool
174
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/backbone/context_module/m.2/MaxPool_output_0 copy
175
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv2.conv.weight + /model/backbone/context_module/cv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv2/conv/Conv
176
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.conv.conv.weight + /model/neck/neck1/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/conv/conv/Conv
177
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/upsample/_input_quantizer/QuantizeLinear
178
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck1.upsample.weight + /model/neck/neck1/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck1/upsample/ConvTranspose
179
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_after_concat.conv.weight + /model/neck/neck1/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_after_concat/conv/Conv
180
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv2.conv.weight + /model/neck/neck1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv2/conv/Conv
181
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv1.conv.weight + /model/neck/neck1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv1/conv/Conv
182
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
183
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
184
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
185
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 825) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Add)
186
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
187
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
188
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
189
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 841) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Add)
190
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv3.conv.weight + /model/neck/neck1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv3/conv/Conv
191
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.conv.conv.weight + /model/neck/neck2/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/conv/conv/Conv
192
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/upsample/_input_quantizer/QuantizeLinear
193
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck2.upsample.weight + /model/neck/neck2/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck2/upsample/ConvTranspose
194
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/Concat_/model/neck/neck2/reduce_skip1/act/Relu_output_0_clone_1 copy
195
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_after_concat.conv.weight + /model/neck/neck2/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_after_concat/conv/Conv
196
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv2.conv.weight + /model/neck/neck2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv2/conv/Conv
197
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv1.conv.weight + /model/neck/neck2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv1/conv/Conv
198
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
199
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
200
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
201
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 890) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Add)
202
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
203
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
204
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
205
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 906) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Add)
206
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/Concat_/model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Add_output_0_clone_0 copy
207
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv3.conv.weight + /model/neck/neck2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv3/conv/Conv
208
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.bbox_stem.seq.conv.weight + /model/heads/head1/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/bbox_stem/seq/conv/Conv || model.heads.head1.pose_stem.seq.conv.weight + /model/heads/head1/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_stem/seq/conv/Conv
209
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.conv.conv.weight + /model/neck/neck3/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/conv/conv/Conv
210
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_convs.0.seq.conv.weight + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head1.cls_convs.0.seq.conv.weight + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/Conv
211
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.0.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/Conv
212
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/conv1/conv/_input_quantizer/QuantizeLinear_clone_1
213
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.cls_pred.weight + /model/heads/head1/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_pred/Conv
214
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_pred.weight + /model/heads/head1/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_pred/Conv
215
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.1.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/Conv
216
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv2.conv.weight + /model/neck/neck3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv2/conv/Conv
217
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv1.conv.weight + /model/neck/neck3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv1/conv/Conv
218
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape + /model/heads/Transpose
219
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_pred.weight + /model/heads/head1/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_pred/Conv
220
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
221
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax
222
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
223
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv
224
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
225
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 972) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Add)
226
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
227
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
228
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
229
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 1013) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Add)
230
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv3.conv.weight + /model/neck/neck3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv3/conv/Conv
231
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_stem.seq.conv.weight + /model/heads/head2/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_stem/seq/conv/Conv || model.heads.head2.bbox_stem.seq.conv.weight + /model/heads/head2/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/bbox_stem/seq/conv/Conv
232
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.conv.conv.weight + /model/neck/neck4/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/conv/conv/Conv
233
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_convs.0.seq.conv.weight + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head2.cls_convs.0.seq.conv.weight + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/Conv
234
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.0.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/Conv
235
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/conv1/conv/_input_quantizer/QuantizeLinear_clone_1
236
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.cls_pred.weight + /model/heads/head2/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_pred/Conv
237
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_pred.weight + /model/heads/head2/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_pred/Conv
238
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.1.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/Conv
239
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv2.conv.weight + /model/neck/neck4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv2/conv/Conv
240
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv1.conv.weight + /model/neck/neck4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv1/conv/Conv
241
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_4 + /model/heads/Transpose_3
242
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_pred.weight + /model/heads/head2/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_pred/Conv
243
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
244
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_1
245
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
246
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_1
247
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
248
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 1079) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Add)
249
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
250
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
251
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
252
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 1120) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Add)
253
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv3.conv.weight + /model/neck/neck4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv3/conv/Conv
254
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.bbox_stem.seq.conv.weight + /model/heads/head3/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/bbox_stem/seq/conv/Conv || model.heads.head3.pose_stem.seq.conv.weight + /model/heads/head3/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_stem/seq/conv/Conv
255
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_convs.0.seq.conv.weight + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head3.cls_convs.0.seq.conv.weight + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/Conv
256
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.0.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/Conv
257
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.cls_pred.weight + /model/heads/head3/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_pred/Conv
258
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_pred.weight + /model/heads/head3/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_pred/Conv
259
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.1.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/Conv
260
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_8 + /model/heads/Transpose_6
261
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.2.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/Conv
262
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_2
263
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_pred.weight + /model/heads/head3/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_pred/Conv
264
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_2
265
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice_1.../post_process/Reshape_2]}
266
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] NMS: batched_nms_243
267
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] DEVICE_TO_SHAPE_HOST: (Unnamed Layer* 1233) [NMS]_1_output[DevicetoShapeHostCopy]
268
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation2]
269
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice...graph2_/Concat_5]}
270
+ [01/04/2024-15:28:22] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation3]
271
+ [01/04/2024-15:28:23] [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +534, GPU +499, now: CPU 1233, GPU 4691 (MiB)
272
+ [01/04/2024-15:28:23] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +82, GPU +74, now: CPU 1315, GPU 4765 (MiB)
273
+ [01/04/2024-15:28:23] [I] [TRT] Local timing cache in use. Profiling results in this builder pass will not be stored.
274
+ [01/04/2024-15:38:28] [I] [TRT] Total Activation Memory: 7939057152
275
+ [01/04/2024-15:38:28] [I] [TRT] Detected 1 inputs and 1 output network tensors.
276
+ [01/04/2024-15:38:29] [I] [TRT] Total Host Persistent Memory: 309280
277
+ [01/04/2024-15:38:29] [I] [TRT] Total Device Persistent Memory: 38912
278
+ [01/04/2024-15:38:29] [I] [TRT] Total Scratch Memory: 134217728
279
+ [01/04/2024-15:38:29] [I] [TRT] [MemUsageStats] Peak memory usage of TRT CPU/GPU memory allocators: CPU 74 MiB, GPU 132 MiB
280
+ [01/04/2024-15:38:29] [I] [TRT] [BlockAssignment] Started assigning block shifts. This will take 171 steps to complete.
281
+ [01/04/2024-15:38:29] [I] [TRT] [BlockAssignment] Algorithm ShiftNTopDown took 40.1195ms to assign 13 blocks to 171 nodes requiring 144747520 bytes.
282
+ [01/04/2024-15:38:29] [I] [TRT] Total Activation Memory: 144747520
283
+ [01/04/2024-15:38:30] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1665, GPU 5901 (MiB)
284
+ [01/04/2024-15:38:30] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in building engine: CPU +15, GPU +16, now: CPU 15, GPU 16 (MiB)
285
+ [01/04/2024-15:38:30] [I] Engine built in 619.407 sec.
286
+ [01/04/2024-15:38:31] [I] [TRT] Loaded engine size: 17 MiB
287
+ [01/04/2024-15:38:31] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +1, GPU +0, now: CPU 1251, GPU 5877 (MiB)
288
+ [01/04/2024-15:38:31] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in engine deserialization: CPU +0, GPU +15, now: CPU 0, GPU 15 (MiB)
289
+ [01/04/2024-15:38:31] [I] Engine deserialized in 0.128553 sec.
290
+ [01/04/2024-15:38:31] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1251, GPU 5877 (MiB)
291
+ [01/04/2024-15:38:31] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +0, GPU +138, now: CPU 0, GPU 153 (MiB)
292
+ [01/04/2024-15:38:31] [I] Setting persistentCacheLimit to 0 bytes.
293
+ [01/04/2024-15:38:31] [I] Using random values for input onnx::Cast_0
294
+ [01/04/2024-15:38:31] [I] Created input binding for onnx::Cast_0 with dimensions 1x3x640x640
295
+ [01/04/2024-15:38:31] [I] Using random values for output graph2_flat_predictions
296
+ [01/04/2024-15:38:31] [I] Created output binding for graph2_flat_predictions with dimensions -1x57
297
+ [01/04/2024-15:38:31] [I] Starting inference
298
+ [01/04/2024-15:38:46] [I] Warmup completed 10 queries over 200 ms
299
+ [01/04/2024-15:38:46] [I] Timing trace has 924 queries over 15.0277 s
300
+ [01/04/2024-15:38:46] [I]
301
+ [01/04/2024-15:38:46] [I] === Trace details ===
302
+ [01/04/2024-15:38:46] [I] Trace averages of 100 runs:
303
+ [01/04/2024-15:38:46] [I] Average on 100 runs - GPU latency: 16.5012 ms - Host latency: 16.6184 ms (enqueue 16.5675 ms)
304
+ [01/04/2024-15:38:46] [I] Average on 100 runs - GPU latency: 16.52 ms - Host latency: 16.6389 ms (enqueue 16.5841 ms)
305
+ [01/04/2024-15:38:46] [I] Average on 100 runs - GPU latency: 15.9968 ms - Host latency: 16.11 ms (enqueue 16.0689 ms)
306
+ [01/04/2024-15:38:46] [I] Average on 100 runs - GPU latency: 15.7859 ms - Host latency: 15.8982 ms (enqueue 15.8629 ms)
307
+ [01/04/2024-15:38:46] [I] Average on 100 runs - GPU latency: 15.6698 ms - Host latency: 15.7823 ms (enqueue 15.7438 ms)
308
+ [01/04/2024-15:38:46] [I] Average on 100 runs - GPU latency: 16.1195 ms - Host latency: 16.2344 ms (enqueue 16.191 ms)
309
+ [01/04/2024-15:38:46] [I] Average on 100 runs - GPU latency: 16.3087 ms - Host latency: 16.4258 ms (enqueue 16.3729 ms)
310
+ [01/04/2024-15:38:46] [I] Average on 100 runs - GPU latency: 15.5609 ms - Host latency: 15.6727 ms (enqueue 15.6373 ms)
311
+ [01/04/2024-15:38:46] [I] Average on 100 runs - GPU latency: 16.4593 ms - Host latency: 16.577 ms (enqueue 16.5201 ms)
312
+ [01/04/2024-15:38:46] [I]
313
+ [01/04/2024-15:38:46] [I] === Performance summary ===
314
+ [01/04/2024-15:38:46] [I] Throughput: 61.4865 qps
315
+ [01/04/2024-15:38:46] [I] Latency: min = 14.6812 ms, max = 18.0088 ms, mean = 16.2285 ms, median = 16.2705 ms, percentile(90%) = 17.0212 ms, percentile(95%) = 17.188 ms, percentile(99%) = 17.4453 ms
316
+ [01/04/2024-15:38:46] [I] Enqueue Time: min = 14.6509 ms, max = 17.9592 ms, mean = 16.1828 ms, median = 16.2139 ms, percentile(90%) = 16.9602 ms, percentile(95%) = 17.1279 ms, percentile(99%) = 17.3662 ms
317
+ [01/04/2024-15:38:46] [I] H2D Latency: min = 0.0830078 ms, max = 0.121094 ms, mean = 0.0968298 ms, median = 0.0967102 ms, percentile(90%) = 0.0991211 ms, percentile(95%) = 0.0996094 ms, percentile(99%) = 0.112915 ms
318
+ [01/04/2024-15:38:46] [I] GPU Compute Time: min = 14.5688 ms, max = 17.8901 ms, mean = 16.1133 ms, median = 16.1515 ms, percentile(90%) = 16.9043 ms, percentile(95%) = 17.0674 ms, percentile(99%) = 17.3271 ms
319
+ [01/04/2024-15:38:46] [I] D2H Latency: min = 0.00341797 ms, max = 0.0561523 ms, mean = 0.0183549 ms, median = 0.0194397 ms, percentile(90%) = 0.0244141 ms, percentile(95%) = 0.0263672 ms, percentile(99%) = 0.0390625 ms
320
+ [01/04/2024-15:38:46] [I] Total Host Walltime: 15.0277 s
321
+ [01/04/2024-15:38:46] [I] Total GPU Compute Time: 14.8887 s
322
+ [01/04/2024-15:38:46] [I] Explanations of the performance metrics are printed in the verbose logs.
323
+ [01/04/2024-15:38:46] [I]
324
  &&&& PASSED TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_s_int8.onnx --int8 --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_s_int8.onnx.int8.engine