File size: 57,324 Bytes
127e10d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 |
&&&& RUNNING TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_l_int8.onnx --int8 --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_l_int8.onnx.int8.engine
[12/28/2023-20:16:29] [I] === Model Options ===
[12/28/2023-20:16:29] [I] Format: ONNX
[12/28/2023-20:16:29] [I] Model: yolo_nas_pose_l_int8.onnx
[12/28/2023-20:16:29] [I] Output:
[12/28/2023-20:16:29] [I] === Build Options ===
[12/28/2023-20:16:29] [I] Max batch: explicit batch
[12/28/2023-20:16:29] [I] Memory Pools: workspace: default, dlaSRAM: default, dlaLocalDRAM: default, dlaGlobalDRAM: default
[12/28/2023-20:16:29] [I] minTiming: 1
[12/28/2023-20:16:29] [I] avgTiming: 8
[12/28/2023-20:16:29] [I] Precision: FP32+INT8
[12/28/2023-20:16:29] [I] LayerPrecisions:
[12/28/2023-20:16:29] [I] Calibration: Dynamic
[12/28/2023-20:16:29] [I] Refit: Disabled
[12/28/2023-20:16:29] [I] Sparsity: Disabled
[12/28/2023-20:16:29] [I] Safe mode: Disabled
[12/28/2023-20:16:29] [I] DirectIO mode: Disabled
[12/28/2023-20:16:29] [I] Restricted mode: Disabled
[12/28/2023-20:16:29] [I] Build only: Disabled
[12/28/2023-20:16:29] [I] Save engine: yolo_nas_pose_l_int8.onnx.int8.engine
[12/28/2023-20:16:29] [I] Load engine:
[12/28/2023-20:16:29] [I] Profiling verbosity: 0
[12/28/2023-20:16:29] [I] Tactic sources: Using default tactic sources
[12/28/2023-20:16:29] [I] timingCacheMode: local
[12/28/2023-20:16:29] [I] timingCacheFile:
[12/28/2023-20:16:29] [I] Heuristic: Disabled
[12/28/2023-20:16:29] [I] Preview Features: Use default preview flags.
[12/28/2023-20:16:29] [I] Input(s)s format: fp32:CHW
[12/28/2023-20:16:29] [I] Output(s)s format: fp32:CHW
[12/28/2023-20:16:29] [I] Input build shapes: model
[12/28/2023-20:16:29] [I] Input calibration shapes: model
[12/28/2023-20:16:29] [I] === System Options ===
[12/28/2023-20:16:29] [I] Device: 0
[12/28/2023-20:16:29] [I] DLACore:
[12/28/2023-20:16:29] [I] Plugins:
[12/28/2023-20:16:29] [I] === Inference Options ===
[12/28/2023-20:16:29] [I] Batch: Explicit
[12/28/2023-20:16:29] [I] Input inference shapes: model
[12/28/2023-20:16:29] [I] Iterations: 10
[12/28/2023-20:16:29] [I] Duration: 15s (+ 200ms warm up)
[12/28/2023-20:16:29] [I] Sleep time: 0ms
[12/28/2023-20:16:29] [I] Idle time: 0ms
[12/28/2023-20:16:29] [I] Streams: 1
[12/28/2023-20:16:29] [I] ExposeDMA: Disabled
[12/28/2023-20:16:29] [I] Data transfers: Enabled
[12/28/2023-20:16:29] [I] Spin-wait: Disabled
[12/28/2023-20:16:29] [I] Multithreading: Disabled
[12/28/2023-20:16:29] [I] CUDA Graph: Disabled
[12/28/2023-20:16:29] [I] Separate profiling: Disabled
[12/28/2023-20:16:29] [I] Time Deserialize: Disabled
[12/28/2023-20:16:29] [I] Time Refit: Disabled
[12/28/2023-20:16:29] [I] NVTX verbosity: 0
[12/28/2023-20:16:29] [I] Persistent Cache Ratio: 0
[12/28/2023-20:16:29] [I] Inputs:
[12/28/2023-20:16:29] [I] === Reporting Options ===
[12/28/2023-20:16:29] [I] Verbose: Disabled
[12/28/2023-20:16:29] [I] Averages: 100 inferences
[12/28/2023-20:16:29] [I] Percentiles: 90,95,99
[12/28/2023-20:16:29] [I] Dump refittable layers:Disabled
[12/28/2023-20:16:29] [I] Dump output: Disabled
[12/28/2023-20:16:29] [I] Profile: Disabled
[12/28/2023-20:16:29] [I] Export timing to JSON file:
[12/28/2023-20:16:29] [I] Export output to JSON file:
[12/28/2023-20:16:29] [I] Export profile to JSON file:
[12/28/2023-20:16:29] [I]
[12/28/2023-20:16:29] [I] === Device Information ===
[12/28/2023-20:16:29] [I] Selected Device: Orin
[12/28/2023-20:16:29] [I] Compute Capability: 8.7
[12/28/2023-20:16:29] [I] SMs: 8
[12/28/2023-20:16:29] [I] Compute Clock Rate: 0.624 GHz
[12/28/2023-20:16:29] [I] Device Global Memory: 7471 MiB
[12/28/2023-20:16:29] [I] Shared Memory per SM: 164 KiB
[12/28/2023-20:16:29] [I] Memory Bus Width: 128 bits (ECC disabled)
[12/28/2023-20:16:29] [I] Memory Clock Rate: 0.624 GHz
[12/28/2023-20:16:29] [I]
[12/28/2023-20:16:29] [I] TensorRT version: 8.5.2
[12/28/2023-20:16:29] [I] [TRT] [MemUsageChange] Init CUDA: CPU +220, GPU +0, now: CPU 249, GPU 2967 (MiB)
[12/28/2023-20:16:33] [I] [TRT] [MemUsageChange] Init builder kernel library: CPU +302, GPU +283, now: CPU 574, GPU 3271 (MiB)
[12/28/2023-20:16:33] [I] Start parsing network model
[12/28/2023-20:16:35] [I] [TRT] ----------------------------------------------------------------
[12/28/2023-20:16:35] [I] [TRT] Input filename: yolo_nas_pose_l_int8.onnx
[12/28/2023-20:16:35] [I] [TRT] ONNX IR version: 0.0.8
[12/28/2023-20:16:35] [I] [TRT] Opset version: 17
[12/28/2023-20:16:35] [I] [TRT] Producer name: pytorch
[12/28/2023-20:16:35] [I] [TRT] Producer version: 2.1.2
[12/28/2023-20:16:35] [I] [TRT] Domain:
[12/28/2023-20:16:35] [I] [TRT] Model version: 0
[12/28/2023-20:16:35] [I] [TRT] Doc string:
[12/28/2023-20:16:35] [I] [TRT] ----------------------------------------------------------------
[12/28/2023-20:16:40] [I] Finish parsing network model
[12/28/2023-20:16:40] [I] FP32 and INT8 precisions have been specified - more performance might be enabled by additionally specifying --fp16 or --best
[12/28/2023-20:16:45] [I] [TRT] ---------- Layers Running on DLA ----------
[12/28/2023-20:16:45] [I] [TRT] ---------- Layers Running on GPU ----------
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation1]
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/pre_process/pre_process.0/Cast.../pre_process/pre_process.2/Mul]}
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1417) [Constant]
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1418) [Constant]
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONSTANT: (Unnamed Layer* 1419) [Constant]
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stem/conv/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stem.conv.rbr_reparam.weight + /model/backbone/stem/conv/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stem/conv/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.downsample.rbr_reparam.weight + /model/backbone/stage1/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/downsample/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv2.conv.weight + /model/backbone/stage1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv1.conv.weight + /model/backbone/stage1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 557) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.0/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 573) [Shuffle] + /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage1/blocks/bottlenecks/bottlenecks.1/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/conv3/conv/_input_quantizer/QuantizeLinear_clone_1
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage1/blocks/conv3/conv/_input_quantizer/QuantizeLinear_clone_0
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage1.blocks.conv3.conv.weight + /model/backbone/stage1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage1/blocks/conv3/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_skip2.conv.weight + /model/neck/neck2/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.downsample.rbr_reparam.weight + /model/backbone/stage2/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/downsample/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.downsample.conv.weight + /model/neck/neck2/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/downsample/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv2.conv.weight + /model/backbone/stage2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv1.conv.weight + /model/backbone/stage2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 620) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.0/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 636) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.1/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage2.blocks.bottlenecks.2.alpha + (Unnamed Layer* 652) [Shuffle] + /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage2/blocks/bottlenecks/bottlenecks.2/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/conv3/conv/_input_quantizer/QuantizeLinear_clone_2
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/conv3/conv/_input_quantizer/QuantizeLinear_clone_1
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage2/blocks/conv3/conv/_input_quantizer/QuantizeLinear_clone_0
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage2.blocks.conv3.conv.weight + /model/backbone/stage2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage2/blocks/conv3/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip2.conv.weight + /model/neck/neck1/reduce_skip2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip2/conv/Conv || model.neck.neck2.reduce_skip1.conv.weight + /model/neck/neck2/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_skip1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.downsample.rbr_reparam.weight + /model/backbone/stage3/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/downsample/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.downsample.conv.weight + /model/neck/neck1/downsample/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/downsample/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv2.conv.weight + /model/backbone/stage3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv1.conv.weight + /model/backbone/stage3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 702) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.0/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 718) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.1/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.2.alpha + (Unnamed Layer* 734) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.2/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.3.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.3.alpha + (Unnamed Layer* 750) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.3/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.4.cv1.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.bottlenecks.4.cv2.rbr_reparam.weight + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage3.blocks.bottlenecks.4.alpha + (Unnamed Layer* 766) [Shuffle] + /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/Mul, /model/backbone/stage3/blocks/bottlenecks/bottlenecks.4/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/conv3/conv/_input_quantizer/QuantizeLinear_clone_4
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/conv3/conv/_input_quantizer/QuantizeLinear_clone_2
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/conv3/conv/_input_quantizer/QuantizeLinear_clone_1
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage3/blocks/conv3/conv/_input_quantizer/QuantizeLinear_clone_0
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage3.blocks.conv3.conv.weight + /model/backbone/stage3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage3/blocks/conv3/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_skip1.conv.weight + /model/neck/neck1/reduce_skip1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_skip1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.downsample.rbr_reparam.weight + /model/backbone/stage4/downsample/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/downsample/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv2.conv.weight + /model/backbone/stage4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv1.conv.weight + /model/backbone/stage4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 806) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.0/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.backbone.stage4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 822) [Shuffle] + /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Mul, /model/backbone/stage4/blocks/bottlenecks/bottlenecks.1/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/conv3/conv/_input_quantizer/QuantizeLinear_clone_1
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/stage4/blocks/conv3/conv/_input_quantizer/QuantizeLinear_clone_0
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.stage4.blocks.conv3.conv.weight + /model/backbone/stage4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/backbone/stage4/blocks/conv3/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv1.conv.weight + /model/backbone/context_module/cv1/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.2/MaxPool
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.1/MaxPool
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POOLING: /model/backbone/context_module/m.0/MaxPool
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/backbone/context_module/m.2/MaxPool_output_0 copy
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.backbone.context_module.cv2.conv.weight + /model/backbone/context_module/cv2/conv/_weight_quantizer/QuantizeLinear + /model/backbone/context_module/cv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.conv.conv.weight + /model/neck/neck1/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/conv/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck1.upsample.weight + /model/neck/neck1/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck1/upsample/ConvTranspose
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.reduce_after_concat.conv.weight + /model/neck/neck1/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/reduce_after_concat/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv2.conv.weight + /model/neck/neck1/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv1.conv.weight + /model/neck/neck1/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.0.alpha + (Unnamed Layer* 887) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.0/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.1.alpha + (Unnamed Layer* 903) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.1/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.2.alpha + (Unnamed Layer* 919) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.2/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.2/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck1/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.3.cv1.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.bottlenecks.3.cv2.rbr_reparam.weight + /model/neck/neck1/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck1.blocks.bottlenecks.3.alpha + (Unnamed Layer* 935) [Shuffle] + /model/neck/neck1/blocks/bottlenecks/bottlenecks.3/Mul, /model/neck/neck1/blocks/bottlenecks/bottlenecks.3/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck1.blocks.conv3.conv.weight + /model/neck/neck1/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck1/blocks/conv3/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.conv.conv.weight + /model/neck/neck2/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/conv/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] DECONVOLUTION: model.neck.neck2.upsample.weight + /model/neck/neck2/upsample/_weight_quantizer/QuantizeLinear + /model/neck/neck2/upsample/ConvTranspose
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/Concat_/model/neck/neck2/reduce_skip1/act/Relu_output_0_clone_1 copy
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.reduce_after_concat.conv.weight + /model/neck/neck2/reduce_after_concat/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/reduce_after_concat/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv2.conv.weight + /model/neck/neck2/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv1.conv.weight + /model/neck/neck2/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.0.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.0.alpha + (Unnamed Layer* 984) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.0/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.1.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.1.alpha + (Unnamed Layer* 1000) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.1/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.2.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.2/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.2.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.2/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.2.alpha + (Unnamed Layer* 1016) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.2/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.2/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck2/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.3.cv1.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.3/cv1/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.bottlenecks.3.cv2.rbr_reparam.weight + /model/neck/neck2/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/bottlenecks/bottlenecks.3/cv2/rbr_reparam/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck2.blocks.bottlenecks.3.alpha + (Unnamed Layer* 1032) [Shuffle] + /model/neck/neck2/blocks/bottlenecks/bottlenecks.3/Mul, /model/neck/neck2/blocks/bottlenecks/bottlenecks.3/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck2.blocks.conv3.conv.weight + /model/neck/neck2/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck2/blocks/conv3/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.bbox_stem.seq.conv.weight + /model/heads/head1/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/bbox_stem/seq/conv/Conv || model.heads.head1.pose_stem.seq.conv.weight + /model/heads/head1/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_stem/seq/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.conv.conv.weight + /model/neck/neck3/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/conv/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_convs.0.seq.conv.weight + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head1.cls_convs.0.seq.conv.weight + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_convs/cls_convs.0/seq/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.0.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.0/seq/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv2.conv.weight + /model/neck/neck3/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv1.conv.weight + /model/neck/neck3/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.cls_pred.weight + /model/heads/head1/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/cls_pred/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.reg_pred.weight + /model/heads/head1/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/reg_pred/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_convs.1.seq.conv.weight + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_convs/pose_convs.1/seq/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape + /model/heads/Transpose
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head1.pose_pred.weight + /model/heads/head1/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head1/pose_pred/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.0.alpha + (Unnamed Layer* 1097) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.0/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.1.alpha + (Unnamed Layer* 1138) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.1/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.2/cv1/conv/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.2.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.2/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.2/cv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.2.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.2/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.2/cv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.2.alpha + (Unnamed Layer* 1154) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.2/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.2/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck3/blocks/bottlenecks/bottlenecks.3/cv1/conv/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.3.cv1.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.3/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.3/cv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.bottlenecks.3.cv2.conv.weight + /model/neck/neck3/blocks/bottlenecks/bottlenecks.3/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/bottlenecks/bottlenecks.3/cv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck3.blocks.bottlenecks.3.alpha + (Unnamed Layer* 1170) [Shuffle] + /model/neck/neck3/blocks/bottlenecks/bottlenecks.3/Mul, /model/neck/neck3/blocks/bottlenecks/bottlenecks.3/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck3.blocks.conv3.conv.weight + /model/neck/neck3/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck3/blocks/conv3/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_stem.seq.conv.weight + /model/heads/head2/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_stem/seq/conv/Conv || model.heads.head2.bbox_stem.seq.conv.weight + /model/heads/head2/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/bbox_stem/seq/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.conv.conv.weight + /model/neck/neck4/conv/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/conv/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_convs.0.seq.conv.weight + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head2.cls_convs.0.seq.conv.weight + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_convs/cls_convs.0/seq/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.0.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.0/seq/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv2.conv.weight + /model/neck/neck4/blocks/conv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv1.conv.weight + /model/neck/neck4/blocks/conv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.cls_pred.weight + /model/heads/head2/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/cls_pred/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.reg_pred.weight + /model/heads/head2/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/reg_pred/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_convs.1.seq.conv.weight + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_convs/pose_convs.1/seq/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_4 + /model/heads/Transpose_3
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head2.pose_pred.weight + /model/heads/head2/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head2/pose_pred/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_1
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.0.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/cv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_1
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.0.alpha + (Unnamed Layer* 1235) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.0/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.1.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/cv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.1.alpha + (Unnamed Layer* 1276) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.1/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.2/cv1/conv/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.2.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.2/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.2/cv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.2.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.2/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.2/cv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.2.alpha + (Unnamed Layer* 1292) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.2/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.2/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] COPY: /model/neck/neck4/blocks/bottlenecks/bottlenecks.3/cv1/conv/_input_quantizer/QuantizeLinear
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.3.cv1.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.3/cv1/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.3/cv1/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.bottlenecks.3.cv2.conv.weight + /model/neck/neck4/blocks/bottlenecks/bottlenecks.3/cv2/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/bottlenecks/bottlenecks.3/cv2/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] POINTWISE: PWN(model.neck.neck4.blocks.bottlenecks.3.alpha + (Unnamed Layer* 1308) [Shuffle] + /model/neck/neck4/blocks/bottlenecks/bottlenecks.3/Mul, /model/neck/neck4/blocks/bottlenecks/bottlenecks.3/Add)
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.neck.neck4.blocks.conv3.conv.weight + /model/neck/neck4/blocks/conv3/conv/_weight_quantizer/QuantizeLinear + /model/neck/neck4/blocks/conv3/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.bbox_stem.seq.conv.weight + /model/heads/head3/bbox_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/bbox_stem/seq/conv/Conv || model.heads.head3.pose_stem.seq.conv.weight + /model/heads/head3/pose_stem/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_stem/seq/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_convs.0.seq.conv.weight + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_convs/reg_convs.0/seq/conv/Conv || model.heads.head3.cls_convs.0.seq.conv.weight + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_convs/cls_convs.0/seq/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.0.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.0/seq/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.cls_pred.weight + /model/heads/head3/cls_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/cls_pred/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.reg_pred.weight + /model/heads/head3/reg_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/reg_pred/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.1.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.1/seq/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] SHUFFLE: /model/heads/Reshape_8 + /model/heads/Transpose_6
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_convs.2.seq.conv.weight + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_convs/pose_convs.2/seq/conv/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] SOFTMAX: /model/heads/Softmax_2
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: model.heads.head3.pose_pred.weight + /model/heads/head3/pose_pred/_weight_quantizer/QuantizeLinear + /model/heads/head3/pose_pred/Conv
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] CONVOLUTION: /model/heads/Conv_2
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice_1.../post_process/Reshape_2]}
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] NMS: batched_nms_274
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] DEVICE_TO_SHAPE_HOST: (Unnamed Layer* 1421) [NMS]_1_output[DevicetoShapeHostCopy]
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation2]
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] MYELIN: {ForeignNode[/model/heads/head1/Slice...graph2_/Concat_5]}
[12/28/2023-20:16:45] [I] [TRT] [GpuLayer] TRAIN_STATION: [trainStation3]
[12/28/2023-20:16:46] [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +534, GPU +481, now: CPU 1532, GPU 4180 (MiB)
[12/28/2023-20:16:47] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +83, GPU +76, now: CPU 1615, GPU 4256 (MiB)
[12/28/2023-20:16:47] [I] [TRT] Local timing cache in use. Profiling results in this builder pass will not be stored.
[12/28/2023-20:30:40] [I] [TRT] Total Activation Memory: 8076738048
[12/28/2023-20:30:40] [I] [TRT] Detected 1 inputs and 1 output network tensors.
[12/28/2023-20:30:46] [I] [TRT] Total Host Persistent Memory: 355104
[12/28/2023-20:30:46] [I] [TRT] Total Device Persistent Memory: 38912
[12/28/2023-20:30:46] [I] [TRT] Total Scratch Memory: 134217728
[12/28/2023-20:30:46] [I] [TRT] [MemUsageStats] Peak memory usage of TRT CPU/GPU memory allocators: CPU 264 MiB, GPU 132 MiB
[12/28/2023-20:30:46] [I] [TRT] [BlockAssignment] Started assigning block shifts. This will take 194 steps to complete.
[12/28/2023-20:30:46] [I] [TRT] [BlockAssignment] Algorithm ShiftNTopDown took 71.4503ms to assign 13 blocks to 194 nodes requiring 183278080 bytes.
[12/28/2023-20:30:46] [I] [TRT] Total Activation Memory: 183278080
[12/28/2023-20:30:48] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +1, GPU +7, now: CPU 2003, GPU 5811 (MiB)
[12/28/2023-20:30:48] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in building engine: CPU +52, GPU +64, now: CPU 52, GPU 64 (MiB)
[12/28/2023-20:30:48] [I] Engine built in 859.207 sec.
[12/28/2023-20:30:48] [I] [TRT] Loaded engine size: 54 MiB
[12/28/2023-20:30:48] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +11, now: CPU 1289, GPU 5551 (MiB)
[12/28/2023-20:30:48] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in engine deserialization: CPU +0, GPU +52, now: CPU 0, GPU 52 (MiB)
[12/28/2023-20:30:48] [I] Engine deserialized in 0.183216 sec.
[12/28/2023-20:30:49] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +6, now: CPU 1289, GPU 5551 (MiB)
[12/28/2023-20:30:49] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +0, GPU +175, now: CPU 0, GPU 227 (MiB)
[12/28/2023-20:30:49] [I] Setting persistentCacheLimit to 0 bytes.
[12/28/2023-20:30:49] [I] Using random values for input onnx::Cast_0
[12/28/2023-20:30:49] [I] Created input binding for onnx::Cast_0 with dimensions 1x3x640x640
[12/28/2023-20:30:49] [I] Using random values for output graph2_flat_predictions
[12/28/2023-20:30:49] [I] Created output binding for graph2_flat_predictions with dimensions -1x57
[12/28/2023-20:30:49] [I] Starting inference
[12/28/2023-20:31:04] [I] Warmup completed 5 queries over 200 ms
[12/28/2023-20:31:04] [I] Timing trace has 467 queries over 15.0803 s
[12/28/2023-20:31:04] [I]
[12/28/2023-20:31:04] [I] === Trace details ===
[12/28/2023-20:31:04] [I] Trace averages of 100 runs:
[12/28/2023-20:31:04] [I] Average on 100 runs - GPU latency: 32.4935 ms - Host latency: 32.6124 ms (enqueue 32.5531 ms)
[12/28/2023-20:31:04] [I] Average on 100 runs - GPU latency: 32.0289 ms - Host latency: 32.1442 ms (enqueue 32.092 ms)
[12/28/2023-20:31:04] [I] Average on 100 runs - GPU latency: 31.9715 ms - Host latency: 32.0855 ms (enqueue 32.0299 ms)
[12/28/2023-20:31:04] [I] Average on 100 runs - GPU latency: 32.1741 ms - Host latency: 32.2872 ms (enqueue 32.2448 ms)
[12/28/2023-20:31:04] [I]
[12/28/2023-20:31:04] [I] === Performance summary ===
[12/28/2023-20:31:04] [I] Throughput: 30.9676 qps
[12/28/2023-20:31:04] [I] Latency: min = 29.5527 ms, max = 48.669 ms, mean = 32.255 ms, median = 31.8447 ms, percentile(90%) = 33.2435 ms, percentile(95%) = 35.3904 ms, percentile(99%) = 42.4307 ms
[12/28/2023-20:31:04] [I] Enqueue Time: min = 29.5156 ms, max = 48.6032 ms, mean = 32.204 ms, median = 31.8164 ms, percentile(90%) = 32.9639 ms, percentile(95%) = 35.3633 ms, percentile(99%) = 42.3721 ms
[12/28/2023-20:31:04] [I] H2D Latency: min = 0.0800781 ms, max = 0.116211 ms, mean = 0.0943801 ms, median = 0.0952148 ms, percentile(90%) = 0.0976562 ms, percentile(95%) = 0.0981445 ms, percentile(99%) = 0.107422 ms
[12/28/2023-20:31:04] [I] GPU Compute Time: min = 29.4414 ms, max = 48.5475 ms, mean = 32.1403 ms, median = 31.7275 ms, percentile(90%) = 33.1393 ms, percentile(95%) = 35.2798 ms, percentile(99%) = 42.3066 ms
[12/28/2023-20:31:04] [I] D2H Latency: min = 0.00292969 ms, max = 0.0655518 ms, mean = 0.0203297 ms, median = 0.0205078 ms, percentile(90%) = 0.0288086 ms, percentile(95%) = 0.03125 ms, percentile(99%) = 0.0375977 ms
[12/28/2023-20:31:04] [I] Total Host Walltime: 15.0803 s
[12/28/2023-20:31:04] [I] Total GPU Compute Time: 15.0095 s
[12/28/2023-20:31:04] [I] Explanations of the performance metrics are printed in the verbose logs.
[12/28/2023-20:31:04] [I]
&&&& PASSED TensorRT.trtexec [TensorRT v8502] # /usr/src/tensorrt/bin/trtexec --onnx=yolo_nas_pose_l_int8.onnx --int8 --avgRuns=100 --duration=15 --saveEngine=yolo_nas_pose_l_int8.onnx.int8.engine
|