A minimal Python package for the Momentum Human Rig - a parametric 3D human body model with identity, pose, and facial expression parameterization.
MHR (Momentum Human Rig) is a high-fidelity 3D human body model that provides:
- Identity Parameterization: 45 shape parameters controlling body identity
- Pose Parameterization: 204 model parameters for full-body articulation
- Facial Expression: 72 expression parameters for detailed face animation
- Multiple LOD Levels: 7 levels of detail (LOD 0-6) for different performance requirements
- Non-linear Pose Correctives: Neural network-based pose-dependent deformations
- PyTorch Integration: GPU-accelerated inference for real-time applications
- PyMomentum Integration: Compatible to fast CPU solver
# Download the torchscript model
curl -OL https://github.com/facebookresearch/MHR/releases/download/v.1.0.0-beta/assets.zip
# Unzip torchscript
unzip -p assets.zip assets/mhr_model.pt > mhr_model.pt
# Start using the torchscript modelNew to TorchScript model? In short it's a Graph mode of pytorch models. More details here. You can take ./demo.py as a reference to start using th torchscript model.
- Advantage: no codebase or model assets are required.
- Disadvantage: Currently only support for LOD 1; limited access to model properties.
# Clone the repository
git clone git@github.com:facebookresearch/MHR.git
cd MHR
# Download the and unzip model assets
curl -OL https://github.com/facebookresearch/MHR/releases/download/v.1.0.0-beta/assets.zip
unzip assets.zip
# Install dependencies with Pixi
pixi install
# Activate the environment
pixi shell# Pip install
pip install mhr .
# Download and unzip the model assets
curl -OL https://github.com/facebookresearch/MHR/releases/download/v.1.0.0-beta/assets.zip
unzip assets.zip- Python >= 3.11
- PyTorch
- pymomentum >= 0.1.90
- trimesh >= 4.8.3 (Only for demo.py)
python demo.pyThis will generate a test MHR mesh and compare outputs with the TorchScript model.
import torch
from mhr.mhr import MHR
# Load MHR model (LOD 1, on CPU)
mhr_model = MHR.from_files(device=torch.device("cpu"), lod=1)
# Define parameters
batch_size = 2
identity_coeffs = 0.8 * torch.randn(batch_size, 45) # Identity
model_parameters = 0.2 * (torch.rand(batch_size, 204) - 0.5) # Pose
face_expr_coeffs = 0.3 * torch.randn(batch_size, 72) # Facial expression
# Generate mesh vertices and skeleton information (joint orientation and positions).
vertices, skeleton_state = mhr_model(identity_coeffs, model_parameters, face_expr_coeffs)- Shape:
[batch_size, 45] - Description: The first 20 control body shape identity, second 20 control head, and the last 5 for hands.
- Typical Range: -3 to +3 (zero-mean, unit variance)
- Shape:
[batch_size, 204] - Description: Joint angles and scalings
- Shape:
[batch_size, 72] - Description: Facial expression blendshape weights
- Typical Range: -1 to +1
Interactive Jupyter notebook for MHR visualization. See tools/mhr_visualization/README.md.
MHR/
├── assets/ # Model assets
│ ├── rig_lod*.fbx # Rig files for each LOD
│ ├── corrective_blendshapes_lod*.npz # Blendshapes
│ ├── corrective_activation.npz # None-linear pose correctives
│ └── model_definition.model # Model parameterization
├── mhr/ # Main package
│ ├── __init__.py
│ ├── mhr.py # MHR model implementation
│ ├── io.py # Asset loading utilities
│ └── utils.py # Helper functions
├── tools/ # Additional tools
│ ├── mhr_visualization/ # Jupyter visualization
├── tests/ # Unit tests
├── demo.py # Basic demo script
└── pyproject.toml # Project configuration
Run the test suite:
# Run all tests
pixi run pytest tests/
# Run specific test
pixi run pytest tests/test_mhr.pyIf you want to do Human Motion Recovery with MHR, head to Sam3D.
We welcome contributions! Please see CONTRIBUTING.md for guidelines.
Please read our CODE_OF_CONDUCT.md before contributing.
If you use MHR in your research, please cite:
@inproceedings{MHR:2025,
author = {Ferguson, Aaron and Osman, Ahmed A. A. and Bescos, Berta and Stoll, Carsten and Twigg, Chris and Lassner, Christoph and Otte, David and Vignola, Eric and Bogo, Federica and Santesteban, Igor and Romero, Javier and Zarate, Jenna and Lee, Jeongseok and Park, Jinhyung and Yang, Jinlong and Doublestein, John and Venkateshan, Kishore and Kitani, Kris and Kavan, Ladislav and Dal Farra, Marco and Hu, Matthew and Cioffi, Matthew and Fabris, Michael and Ranieri, Michael and Modarres, Mohammad and Kadlecek, Petr and Khirodkar, Rawal and Abdrashitov, Rinat and Prévost, Romain and Rajbhandari, Roman and Mallet, Ronald and Pearsall, Russel and Kao, Sandy and Kumar, Sanjeev and Parrish, Scott and Saito, Shunsuke and Wang, Te-Li and Tung, Tony and Dong, Yuan and Chen, Yuhua and Xu, Yuanlu and Ye, Yuting and Jiang, Zhongshi},
title = {MHR: Momentum Human Rig},
booktitle = {Tech Report},
year = {2025},
url = {https://arxiv.org/abs/your-arxiv-id}
}MHR is licensed under the Apache Software License 2.0, as found in the LICENSE file.