pesi
/

Luigi commited on
Commit
0a36dd1
1 Parent(s): 1f0f5d8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -4
README.md CHANGED
@@ -4,11 +4,25 @@ pipeline_tag: object-detection
4
  tags:
5
  - Pose Estimation
6
  ---
7
- RTMO Running CUDA DEMO Package
8
 
9
  - `demo.sh`: DEMO main program, which will first install rtmlib, and then use rtmo-s to analyze the .mp4 files in the video folder.
10
- - `rtmo_gpu.py`: Defines an RTMO_GPU class, making fine adjustments to CUDA settings.
11
- - `rtmo_demo.py`: Python main program, which has two arguments:
 
12
  - `path`: The folder location that contains the .mp4 files to be analyzed.
13
  - `model_path`: The local path to the ONNX model or a URL pointing to the RTMO model published on mmpose.
14
- - `video`: Contains two test videos provided by Sam's family.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  tags:
5
  - Pose Estimation
6
  ---
7
+ RTMO / YOLO-NAS-Pose Inference with CUDAExecutionProvider / TensorrtExecutionProvider DEMO
8
 
9
  - `demo.sh`: DEMO main program, which will first install rtmlib, and then use rtmo-s to analyze the .mp4 files in the video folder.
10
+ - `demo_batch.sh`: Multi-batch version of demo.sh
11
+ - `rtmo_gpu.py`: Defines an RTMO_GPU (& RTMO_GPU_BATCH) class, making fine adjustments to CUDA & TensorRT settings.
12
+ - `rtmo_demo.py`: Python main program, which has three arguments:
13
  - `path`: The folder location that contains the .mp4 files to be analyzed.
14
  - `model_path`: The local path to the ONNX model or a URL pointing to the RTMO model published on mmpose.
15
+ - `--yolo_nas_pose`: If you run inference with YOLO NAS Pose Model instead of RTMO model.
16
+ - `rtmo_demo_batch.py`: Multi-batch version of demo_batch.sh
17
+ - `video`: Contains one test video.
18
+
19
+ Original ONNX models come from [](https://github.com/open-mmlab/mmpose/tree/main/projects/rtmo) trained on body7. We did only
20
+
21
+ We did the following to make them work with TensorRTExecutionProvdier
22
+
23
+ 1. Shape inference
24
+ 2. batch size 1,2,4 fixation
25
+
26
+ Note: TensorrtExecutionProvider only supports Models with fixed batch size (*_batchN.onnx) while CUDAExecutionProvider can run with dynamic batch size.
27
+
28
+ FP16 ONNX model is also provided.