We study how tempo-changing mechanisms shape expressive music-to-motion generation. JoruriPuppet is a dataset of Japanese Ningyo Joruri puppet performances paired with music, enabling models to learn tempo-aware behavior. We introduce two tempo features (instantaneous BPM and its change) and three metrics for Jo–Ha–Kyu correlation, S-curve aesthetics, and head–hand theatrical contrast, revealing limitations of beat-only evaluations and improving state-of-the-art methods.
Subject | Motion/BVH | Motion/BVH‑SMPL | Motion/SMPL‑NPZ | Motion/TRC | Audio/WAV |
---|---|---|---|---|---|
Shamisen | BVH | BVH | NPZ | TRC | WAV |
Without Shamisen | BVH | BVH | NPZ | TRC | WAV |
Details and file structure are described in the Appendix. See Appendix (PDF).
@inproceedings{dong2025joruripuppet, title = {JoruriPuppet: Learning Tempo-Changing Mechanisms Beyond the Beat for Music-to-Motion Generation}, author = {Dong, Ran and Ni, Shaowen and Yang, Xi}, title = {Proceedings of the Special Interest Group on Computer Graphics and Interactive Techniques Conference Conference Papers}, pages = {1--11}, year = {2025}, doi = {10.1145/3757377.3764006} }