Skip to content

Single-view real-time motion capture built up upon Google Mediapipe.

License

Notifications You must be signed in to change notification settings

EasternJournalist/live_mocap

Repository files navigation

Live MoCap

Ganyu-ji

Requirements

  • python>=3.8
    • mediapipe
    • pytorch (cpu version is ok)
  • blender >= 3.0 (for reading assets and binding animation)

How to use

  1. Prepare your character model

    Currently this script uses Blender to load model skeleton and bind animation. Your model should be saved as .blend file.

    You may edit your model to assure that

    • Model must be in rest pose (clear all bone rotation/translation/scale in pose mode). And the rest pose should be close to T pose.

    • Clear previous bone animation data and constraints.

    • Name related bones as below (in lower case). You may refer the mixamo example assets/mixamo.blend to see to name the bones so that they can be recogonzed and binded.

    • Save the model as .blend file somewhere.

  2. Run the script mocap.py.

    python mocap.py --blend your_character_model.blend --video your_video.mp4 [other options] 
    

    The program will read and capture motion from the video, save the animation data, and then open Blender and bind the animation to your charactor model. After everything is done, you should be able to see the Blender window with your character already animated.

Future work

  • Now working on face capture.

About

Single-view real-time motion capture built up upon Google Mediapipe.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published