We study how tempo-changing mechanisms shape expressive music-to-motion generation. JoruriPuppet is a dataset of Japanese Ningyo Joruri puppet performances paired with music, enabling models to learn tempo-aware behavior. We introduce two tempo features (instantaneous BPM and its change) and three metrics for Jo–Ha–Kyu correlation, S-curve aesthetics, and head–hand theatrical contrast, revealing limitations of beat-only evaluations and improving state-of-the-art methods.
| Subject | Motion/BVH | Motion/SMPL-BVH | Motion/SMPL‑NPZ | Motion/SMPL-TRC | Audio/WAV |
|---|---|---|---|---|---|
| Shamisen | BVH | BVH | NPZ | TRC | WAV |
| Without Shamisen | BVH | BVH | NPZ | TRC | WAV |
Details and file structure are described in the Appendix. See Appendix.
@inproceedings{dong2025joruripuppet,
title={JoruriPuppet: Learning Tempo-Changing Mechanisms Beyond the Beat for Music-to-Motion Generation with Expressive Metrics},
author={Dong, Ran and Ni, Shaowen and Yang, Xi},
booktitle={Proceedings of the SIGGRAPH Asia 2025 Conference Papers},
pages={1--11},
year={2025}
}