| --- |
| pretty_name: MotionMillion |
| size_categories: |
| - n<1T |
| task_categories: |
| - other |
| - robotics |
| language: |
| - en |
| tags: |
| - Large Human Motion |
| - Humanoid |
| - Humanoid Locomotion |
| extra_gated_prompt: >- |
| ### MotionMillion COMMUNITY LICENSE AGREEMENT |
| |
| MotionMillion Release Date: July 30, 2025 All the data and code within this |
| repo are under [CC BY-NC-SA |
| 4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/). |
| extra_gated_fields: |
| First Name: text |
| Last Name: text |
| Email: text |
| Country: country |
| Affiliation: text |
| Phone: text |
| Job title: |
| type: select |
| options: |
| - Student |
| - Research Graduate |
| - AI researcher |
| - AI developer/engineer |
| - Reporter |
| - Other |
| Research interest: text |
| geo: ip_location |
| By clicking Submit below I accept the terms of the license and acknowledge that the information I provide will be collected stored processed and shared in accordance with the InternData Privacy Policy: checkbox |
| extra_gated_description: >- |
| The information you provide will be collected, stored, processed and shared in |
| accordance with the InternData Privacy Policy. |
| extra_gated_button_content: Submit |
| --- |
| |
| # 🔑 Key Features |
| - **Over 2000 hours of high-quality human motion** captured from web-scale human video data, covering: |
| - Martial Arts (23.7%) |
| - Fitness (26.4%) |
| - Performance (17.5%) |
| - Dance (14.9%) |
| - Non-Human (2.9%) |
| - Sports (2.4%) |
| - **Over 20 detailed annotations per motion**, including: |
| - Age |
| - Body Characteristics |
| - Movement Styles |
| - Emotions |
| - Environments |
|
|
| <img src="assets/motionmillion_teaser.png" alt="Image Alt Text" width="70%" style="display: block; margin-left: auto; margin-right: auto;" /> |
|
|
| </br> |
|
|
| # Table of Contents |
| - [Key Features](#key-features-) |
| - [Get Started](#get-started-) |
| - [Download the Dataset](#download-the-dataset) |
| - [Dataset Structure](#dataset-structure) |
| - [License and Citation](#license-and-citation) |
|
|
| # 👨🏫 Get Started |
| ## Download the Dataset |
| To download the full dataset, use the following code. If you encounter any issues, refer to the official Hugging Face documentation. |
|
|
| ```bash |
| # Ensure git-lfs is installed (https://git-lfs.com) |
| git lfs install |
| # When prompted for a password, use an access token with write permissions. |
| # Generate one in your settings: https://huggingface.co/settings/tokens |
| git clone https://huggingface.co/datasets/InternRobotics/MotionMillion |
| # To clone without large files (only their pointers) |
| GIT_LFS_SKIP_SMUDGE=1 git clone https://huggingface.co/datasets/InternRobotics/MotionMillion |
| ``` |
|
|
| If you only need to download a specific dataset (e.g., `MotionGV/folder1.tar.gz`), use the following code: |
|
|
| ```bash |
| # Ensure git-lfs is installed (https://git-lfs.com) |
| git lfs install |
| # Initialize an empty Git repository |
| git init MotionMillion |
| cd MotionMillion |
| # Set the remote repository |
| git remote add origin https://huggingface.co/datasets/InternRobotics/MotionMillion |
| # Enable sparse-checkout |
| git sparse-checkout init |
| # Specify the target folders and files |
| git sparse-checkout set MotionGV/folder1.tar.gz |
| # Pull the data |
| git pull origin main |
| ``` |
|
|
| ## Dataset Processing |
| ### Folder Hierarchy |
| ``` |
| MotionMillion |
| |-- motion_272rpr |
| | |-- Mirror_MotionGV |
| | | |-- folder0.tar.gz |
| | | |-- folder1.tar.gz |
| | | |-- folder2.tar.gz |
| | | |-- folder3.tar.gz |
| | | |-- folder4.tar.gz |
| | | |-- folder5.tar.gz |
| | | |-- folder6.tar.gz |
| | | |-- folder7.tar.gz |
| | | |-- folder8.tar.gz |
| | | `-- folder9.tar.gz |
| | |-- Mirror_MotionLLAMA |
| | | |-- finedance.tar.gz |
| | | |-- fit3d.tar.gz |
| | | |-- hi4d.tar.gz |
| | | |-- humansc3d.tar.gz |
| | | |-- interhuman.tar.gz |
| | | |-- interx.tar.gz |
| | | `-- trumans.tar.gz |
| | |-- Mirror_MotionUnion |
| | | |-- 100STYLE_smpl.tar.gz |
| | | |-- CombatMotion_seperate.tar.gz |
| | | |-- EgoBody.tar.gz |
| | | |-- animation.tar.gz |
| | | |-- fitness.tar.gz |
| | | |-- game_motion.tar.gz |
| | | |-- haa500.tar.gz |
| | | |-- humman.tar.gz |
| | | |-- idea400.tar.gz |
| | | |-- kungfu.tar.gz |
| | | |-- music.tar.gz |
| | | `-- perform.tar.gz |
| | |-- MotionGV |
| | | |-- folder0.tar.gz |
| | | |-- folder1.tar.gz |
| | | |-- folder2.tar.gz |
| | | |-- folder3.tar.gz |
| | | |-- folder4.tar.gz |
| | | |-- folder5.tar.gz |
| | | |-- folder6.tar.gz |
| | | |-- folder7.tar.gz |
| | | |-- folder8.tar.gz |
| | | `-- folder9.tar.gz |
| | |-- MotionLLAMA |
| | | |-- finedance.tar.gz |
| | | |-- fit3d.tar.gz |
| | | |-- hi4d.tar.gz |
| | | |-- humansc3d.tar.gz |
| | | |-- interhuman.tar.gz |
| | | |-- interx.tar.gz |
| | | `-- trumans.tar.gz |
| | |-- MotionUnion |
| | | |-- 100STYLE_smpl.tar.gz |
| | | |-- CombatMotion_seperate.tar.gz |
| | | |-- EgoBody.tar.gz |
| | | |-- animation.tar.gz |
| | | |-- fitness.tar.gz |
| | | |-- game_motion.tar.gz |
| | | |-- haa500.tar.gz |
| | | |-- humman.tar.gz |
| | | |-- idea400.tar.gz |
| | | |-- kungfu.tar.gz |
| | | |-- music.tar.gz |
| | | `-- perform.tar.gz |
| |-- mean_std |
| │ |-- Mean.npy |
| │ `-- Std.npy |
| |-- texts.tar.gz |
| |-- splits.tar.gz |
| ``` |
|
|
| Due to data licensing restrictions, we only provide parts of the processed motion data with 272-representation. Among these, MotionGV contains motions captured by our motion capture algorithm; the remaining data is merged from other datasets. |
|
|
| Due to copyright constraints, BABEL, AIST and HumanML3D cannot be released directly. We will provide detailed data processing workflows. |
|
|
| ### Data Processing Steps |
|
|
| 1. For all tar.gz files, use `tar -xzvf x.tar.gz` to extract them. |
| 2. For HumanML3D, please refer to [data_process/HumanML3D](data_process/HumanML3D/README.md). |
| 3. For BABEL, please refer to [data_process/BABEL](data_process/BABEL/README.md). |
| 4. For AIST, please refer to [data_process/AIST](data_process/AIST/README.md). |
|
|
| ### Processed Data Hierarchy |
| ``` |
| MotionMillion |
| |-- motion_272rpr |
| | |-- BABEL |
| | |-- Mirror_BABEL |
| | |-- Mirror_MotionGV |
| | | |-- folder0 |
| | | |-- folder1 |
| | | |-- folder2 |
| | | |-- folder3 |
| | | |-- folder4 |
| | | |-- folder5 |
| | | |-- folder6 |
| | | |-- folder7 |
| | | |-- folder8 |
| | | `-- folder9 |
| | |-- Mirror_MotionLLAMA |
| | | |-- aist |
| | | |-- finedance |
| | | |-- fit3d |
| | | |-- hi4d |
| | | |-- humansc3d |
| | | |-- interhuman |
| | | |-- interx |
| | | `-- trumans |
| | |-- Mirror_MotionUnion |
| | | |-- 100STYLE_smpl |
| | | |-- CombatMotion_seperate |
| | | |-- EgoBody |
| | | |-- animation |
| | | |-- fitness |
| | | |-- game_motion |
| | | |-- haa500 |
| | | |-- humanml |
| | | |-- humman |
| | | |-- idea400 |
| | | |-- kungfu |
| | | |-- music |
| | | `-- perform |
| | |-- Mirror_PhantomDanceDatav1.1 |
| | |-- MotionGV |
| | | |-- folder0 |
| | | |-- folder1 |
| | | |-- folder2 |
| | | |-- folder3 |
| | | |-- folder4 |
| | | |-- folder5 |
| | | |-- folder6 |
| | | |-- folder7 |
| | | |-- folder8 |
| | | `-- folder9 |
| | |-- MotionLLAMA |
| | | |-- aist |
| | | |-- finedance |
| | | |-- fit3d |
| | | |-- hi4d |
| | | |-- humansc3d |
| | | |-- interhuman |
| | | |-- interx |
| | | `-- trumans |
| | |-- MotionUnion |
| | | |-- 100STYLE_smpl |
| | | |-- CombatMotion_seperate |
| | | |-- EgoBody |
| | | |-- animation |
| | | |-- fitness |
| | | |-- game_motion |
| | | |-- haa500 |
| | | |-- humanml |
| | | |-- humman |
| | | |-- idea400 |
| | | |-- kungfu |
| | | |-- music |
| | | `-- perform |
| | `-- PhantomDanceDatav1.1 |
| |-- texts |
| | |-- Mirror_MotionGV |
| | |-- Mirror_MotionLLAMA |
| | |-- Mirror_MotionUnion |
| | |-- MotionGV |
| | |-- MotionLLAMA |
| | `-- MotionUnion |
| |-- mean_std |
| │ |-- Mean.npy |
| │ `-- Std.npy |
| |-- split |
| | `-- version1 |
| | |-- t2m_60_300 |
| | | |-- all.txt |
| | | |-- test.txt |
| | | |-- train.txt |
| | | `-- val.txt |
| | `-- tokenizer_96 |
| | |-- test.txt |
| | |-- train.txt |
| | `-- val.txt |
| ``` |
|
|
|
|
| # License and Citation |
| All the data and code within this repo are under [CC BY-NC-SA 4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/). Please consider citing our project if it helps your research. |
|
|
| ```BibTeX |
| @article{fan2025go, |
| title={Go to Zero: Towards Zero-shot Motion Generation with Million-scale Data}, |
| author={Fan, Ke and Lu, Shunlin and Dai, Minyue and Yu, Runyi and Xiao, Lixing and Dou, Zhiyang and Dong, Junting and Ma, Lizhuang and Wang, Jingbo}, |
| journal={arXiv preprint arXiv:2507.07095}, |
| year={2025} |
| } |
| ``` |
|
|
| In addition, please cite the following literature: |
| ```BibTeX |
| @article{xiao2025motionstreamer, |
| title={MotionStreamer: Streaming Motion Generation via Diffusion-based Autoregressive Model in Causal Latent Space}, |
| author={Xiao, Lixing and Lu, Shunlin and Pi, Huaijin and Fan, Ke and Pan, Liang and Zhou, Yueer and Feng, Ziyong and Zhou, Xiaowei and Peng, Sida and Wang, Jingbo}, |
| journal={arXiv preprint arXiv:2503.15451}, |
| year={2025} |
| } |
| |
| @inproceedings{amass, |
| title={AMASS: Archive of motion capture as surface shapes}, |
| author={Mahmood, Naureen and Ghorbani, Nima and Troje, Nikolaus F and Pons-Moll, Gerard and Black, Michael J}, |
| booktitle={ICCV}, |
| pages={5442--5451}, |
| year={2019} |
| } |
| |
| @InProceedings{Guo_2022_CVPR, |
| author = {Guo, Chuan and Zou, Shihao and Zuo, Xinxin and Wang, Sen and Ji, Wei and Li, Xingyu and Cheng, Li}, |
| title = {Generating Diverse and Natural 3D Human Motions From Text}, |
| booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)}, |
| month = {June}, |
| year = {2022}, |
| pages = {5152-5161} |
| } |
| |
| @inproceedings{babel, |
| title={BABEL: Bodies, action and behavior with english labels}, |
| author={Punnakkal, Abhinanda R and Chandrasekaran, Arjun and Athanasiou, Nikos and Quiros-Ramirez, Alejandra and Black, Michael J}, |
| booktitle={CVPR}, |
| pages={722--731}, |
| year={2021} |
| } |
| |
| @inproceedings{flag3d, |
| title={Flag3d: A 3d fitness activity dataset with language instruction}, |
| author={Tang, Yansong and Liu, Jinpeng and Liu, Aoyang and Yang, Bin and Dai, Wenxun and Rao, Yongming and Lu, Jiwen and Zhou, Jie and Li, Xiu}, |
| booktitle={CVPR}, |
| pages={22106--22117}, |
| year={2023} |
| } |
| |
| @inproceedings{li2023finedance, |
| title={FineDance: A Fine-grained Choreography Dataset for 3D Full Body Dance Generation}, |
| author={Li, Ronghui and Zhao, Junfan and Zhang, Yachao and Su, Mingyang and Ren, Zeping and Zhang, Han and Tang, Yansong and Li, Xiu}, |
| booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision}, |
| pages={10234--10243}, |
| year={2023} |
| } |
| |
| @article{motionx, |
| title={Motion-x: A large-scale 3d expressive whole-body human motion dataset}, |
| author={Lin, Jing and Zeng, Ailing and Lu, Shunlin and Cai, Yuanhao and Zhang, Ruimao and Wang, Haoqian and Zhang, Lei}, |
| journal={NeurIPS}, |
| year={2024} |
| } |
| |
| @article{liang2024intergen, |
| title={Intergen: Diffusion-based multi-human motion generation under complex interactions}, |
| author={Liang, Han and Zhang, Wenqian and Li, Wenxuan and Yu, Jingyi and Xu, Lan}, |
| journal={International Journal of Computer Vision}, |
| volume={132}, |
| number={9}, |
| pages={3463--3483}, |
| year={2024}, |
| publisher={Springer} |
| } |
| |
| @inproceedings{interx, |
| title={Inter-x: Towards versatile human-human interaction analysis}, |
| author={Xu, Liang and Lv, Xintao and Yan, Yichao and Jin, Xin and Wu, Shuwen and Xu, Congsheng and Liu, Yifan and Zhou, Yizhou and Rao, Fengyun and Sheng, Xingdong and others}, |
| booktitle={CVPR}, |
| pages={22260--22271}, |
| year={2024} |
| } |
| |
| @inproceedings{aist-dance-db, |
| author = {Shuhei Tsuchida and Satoru Fukayama and Masahiro Hamasaki and Masataka Goto}, |
| title = {AIST Dance Video Database: Multi-genre, Multi-dancer, and Multi-camera Database for Dance Information Processing}, |
| booktitle = {Proceedings of the 20th International Society for Music Information Retrieval Conference, {ISMIR} 2019}, |
| address = {Delft, Netherlands}, |
| year = 2019, |
| month = nov} |
| |
| @inproceedings{jiang2024scaling, |
| title={Scaling up dynamic human-scene interaction modeling}, |
| author={Jiang, Nan and Zhang, Zhiyuan and Li, Hongjie and Ma, Xiaoxuan and Wang, Zan and Chen, Yixin and Liu, Tengyu and Zhu, Yixin and Huang, Siyuan}, |
| booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition}, |
| pages={1737--1747}, |
| year={2024} |
| } |
| |
| @inproceedings{yin2023hi4d, |
| title={Hi4d: 4d instance segmentation of close human interaction}, |
| author={Yin, Yifei and Guo, Chen and Kaufmann, Manuel and Zarate, Juan Jose and Song, Jie and Hilliges, Otmar}, |
| booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition}, |
| pages={17016--17027}, |
| year={2023} |
| } |
| |
| @inproceedings{fieraru2021learning, |
| title={Learning complex 3D human self-contact}, |
| author={Fieraru, Mihai and Zanfir, Mihai and Oneata, Elisabeta and Popa, Alin-Ionut and Olaru, Vlad and Sminchisescu, Cristian}, |
| booktitle={Proceedings of the AAAI Conference on Artificial Intelligence}, |
| volume={35}, |
| number={2}, |
| pages={1343--1351}, |
| year={2021} |
| } |
| |
| @inproceedings{danceformer, |
| title={Danceformer: Music conditioned 3d dance generation with parametric motion transformer}, |
| author={Li, Buyu and Zhao, Yongchi and Zhelun, Shi and Sheng, Lu}, |
| booktitle={AAAI}, |
| pages={1272--1279}, |
| year={2022} |
| } |
| |
| ``` |
|
|
| **Special Notes** |
| - We would like to express our gratitude to the authors of [FineDance](https://github.com/li-ronghui/FineDance) for granting permission to directly open-source the preprocessed motion data. It is important to note that when generating the 272-dimensional motion representation, we utilized the SMPL-X data provided in [MotionLLAMA](https://github.com/ZeyuLing/MotionLLaMA) with all beta values set to 0, which may differ from the original FineDance data. |
| - If you intend to use the merged data (excluding MotionGV), please strictly adhere to their respective licenses. |