Skip to content
Switch branches/tags


Failed to load latest commit information.
Latest commit message
Commit time

AMASS: Archive of Motion Capture as Surface Shapes

alt text

AMASS is a large database of human motion unifying different optical marker-based motion capture datasets by representing them within a common framework and parameterization. AMASS is readily useful for animation, visualization, and generating training data for deep learning.

Here we provide tools and tutorials to use AMASS in your research projects. More specifically:

  • Following the recommended splits of data by AMASS, we provide three non-overlapping train/validation/test splits.
  • AMASS uses an extended version of SMPL+H with DMPLs. Here we show how to load different components and visualize a body model with AMASS data.
  • AMASS is also compatible with SMPL and SMPL-X body models. We show how to use the body data from AMASS to animate these models.

Table of Contents



Clone this repo and run the following from the root folder:

python install -r requirements.txt
python develop

Body Models

AMASS uses MoSh++ pipeline to fit SMPL+H body model to human optical marker based motion capture (mocap) data. In the paper we use SMPL+H with extended shape space, i.e. 16 betas, and 8 DMPLs. Please download models and place them them in body_models folder of this repository after you obtained the code from GitHub.


We release tools and Jupyter notebooks to demonstrate how to use AMASS to animate SMPL+H body model.

Furthermore, as promised in the supplementary material of the paper, we release code to produce synthetic mocap using DFaust registrations.

Please refer to tutorials for further details.


Please cite the following paper if you use this code directly or indirectly in your research/projects:

  title={AMASS: Archive of Motion Capture as Surface Shapes},
  author={Mahmood, Naureen and Ghorbani, Nima and F. Troje, Nikolaus and Pons-Moll, Gerard and Black, Michael J.},
  booktitle = {The IEEE International Conference on Computer Vision (ICCV)},
  month = {Oct},
  url = {},
  month_numeric = {10}


Software Copyright License for non-commercial scientific research purposes. Please read carefully the terms and conditions and any accompanying documentation before you download and/or use the AMASS dataset, and software, (the "Model & Software"). By downloading and/or using the Model & Software (including downloading, cloning, installing, and any other use of this GitHub repository), you acknowledge that you have read these terms and conditions, understand them, and agree to be bound by them. If you do not agree with these terms and conditions, you must not download and/or use the Model & Software. Any infringement of the terms of this agreement will automatically terminate your rights under this License.


The code in this repository is developed by Nima Ghorbani.

If you have any questions you can contact us at

For commercial licensing, please contact

To find out about the latest developments stay tuned to AMASS twitter.

Contribute to AMASS

The research community needs more human motion data. If you have interesting marker based motion capture data, and you are willing to share it for research purposes, then we will label and clean your mocap and MoSh it for you and add it to the AMASS dataset, naturally citing you as the original owner of the marker data. For this purposes feel free to contact