Skip to content
/ MoMa Public

Official repository for MoMa: Skinned Motion Retargeting Using Masked Pose Modeling

Notifications You must be signed in to change notification settings

mmlab-cv/MoMa

Repository files navigation

MoMa

Code repository for the paper:
MoMa: Skinned Motion Retargeting Using Masked Pose Modeling
Giulia Martinelli, Nicola Garau, Niccolò Bisagno, Nicola Conci
CVIU 2024
[paper] [project page]

image

🚀 How to Run

1️⃣ Set Up the Virtual Environment

Run the following command to configure the virtual environment:
pipenv install

2️⃣ Install Dependencies

pip install -r requirements.txt

3️⃣ Install Additional Dependencies

📂 Datasets

Download the following datasets and organize them in the specified folders:

🏃‍♂️ Mixamo

  • Download animations from Mixamo.
  • Organize them as:
MIXAMO/
├── Character_1/
│   ├── animation_1.bvh
│   ├── animation_2.bvh
│   └── ...
├── Character_2/
├── Character_3/
└── ...

🐶 Dog Dataset

  • Download the dog animations from AI4Animation.
  • Place all .bvh files inside:
HumanDog/
├── Dog/
│   ├── animation_1.bvh
│   ├── animation_2.bvh
│   └── ...

🧍‍♂️ Human Dataset

HumanDog/
├── Human/
│   ├── animation_1.bvh
│   ├── animation_2.bvh
│   └── ...

🛠 Setup

Before starting training, configure the dataset settings in data_flags.py.

🔧 Dataset Configuration

Modify the following variables:

  • dataset_path: Path to the dataset directory.

    • Mixamo Dataset: The main folder should contain multiple character subfolders, each with .bvh animation files.
    • HumanDog Dataset: The main folder should contain two subfolders:
      • Human/ → Contains .bvh animation files for human motions.
      • Dog/ → Contains .bvh animation files for dog motions.
  • dataset: Choose the dataset name:

    • "MIXAMO" → for the Mixamo dataset.
    • "HumanDog" → for the HumanDog dataset.
  • n_joints: Number of joints in the dataset’s superskeleton:

    • 25 → for Mixamo.
    • 26 → for HumanDog.
  • mode: Set the mode to:

    • "train" → to enable training.

🚆 Training

Once the dataset configuration is set, start training by running:

python main.py

Once it is trained, the collision in the resulting animation can be solved by running:

python shape_optimization.py

Remember to set the correct character name and bvh_path in the shape_optimization.py file

Checkpoints

💾 Checkpoints

The checkpoints can be downloaded here:

🔗 Download Checkpoints

📜 Citing

If you find this code useful for your research or use the data generated by our method, please consider citing the following paper:

@article{martinelli2024moma,
  title={MoMa: Skinned motion retargeting using masked pose modeling},
  author={Martinelli, Giulia and Garau, Nicola and Bisagno, Niccol{\'o} and Conci, Nicola},
  journal={Computer Vision and Image Understanding},
  volume={249},
  pages={104141},
  year={2024},
  publisher={Elsevier}
}

About

Official repository for MoMa: Skinned Motion Retargeting Using Masked Pose Modeling

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages