Skip to content

Files

Latest commit

2f69b14 · Feb 11, 2025

History

History

keypoint_detection

Keypoint detection demo

This application demonstrates the human body key point detection model.

The model is yolov8n-pose from ultralytics. It is exported to ONNX using ultralytics cli yolo export model=yolov8n-pose.pt format=onnx dynamic simplify.

Preview:

Tested on platforms:

  • Nvidia Turing, Ampere
  • Nvidia Jetson Orin family

Demonstrated adapters:

  • Video loop source adapter;
  • Always-ON RTSP sink adapter.

Prerequisites

git clone https://github.com/insight-platform/Savant.git
cd Savant
git lfs pull
./utils/check-environment-compatible

Note: Ubuntu 22.04 runtime configuration guide helps to configure the runtime to run Savant pipelines.

Build Engines

The demo uses models that are compiled into TensorRT engines the first time the demo is run. This takes time. Optionally, you can prepare the engines before running the demo by using the command:

# you are expected to be in Savant/ directory

./scripts/run_module.py --build-engines samples/keypoint_detection/module.yml

Run Demo

# you are expected to be in Savant/ directory

# if x86
docker compose -f samples/keypoint_detection/docker-compose.x86.yml up

# if Jetson
docker compose -f samples/keypoint_detection/docker-compose.l4t.yml up

# open 'rtsp://127.0.0.1:554/stream/video' in your player
# or visit 'http://127.0.0.1:888/stream/video/' (LL-HLS)

# Ctrl+C to stop running the compose bundle

Performance Measurement

Download the video file to the data folder. For example:

# you are expected to be in Savant/ directory

mkdir -p data && curl -o data/shuffle_dance.mp4 \
https://eu-central-1.linodeobjects.com/savant-data/demo/shuffle_dance.mp4

Now you are ready to run the performance benchmark with the following command:

./samples/keypoint_detection/run_perf.sh 

Note: Change the value of the DATA_LOCATION variable in the run_perf.sh script if you changed the video.