Qijun Gan1, Song Wang1, Shengtao Wu2 ,Jianke Zhu1 📧
1 Zhejiang University, 2 Hangzhou Dianzi University
(📧) corresponding author.
ArXiv Preprint Project Page Dataset[Google Drive]/[Hyper.ai]/[zenodo]
Jun. 14th, 2024
: Paper is available at arxiv. ☕️Jun. 1st, 2024
: We released our code and datasets! Paper is coming soon. Please stay tuned! ☕️
Recently, artificial intelligence techniques for education have been received increasing attentions, while it still remains an open problem to design the effective music instrument instructing systems. Although key presses can be directly derived from sheet music, the transitional movements among key presses require more extensive guidance in piano performance. In this work, we construct a piano-hand motion generation benchmark to guide hand movements and fingerings for piano playing. To this end, we collect an annotated dataset, PianoMotion10M, consisting of 116 hours of piano playing videos from a bird's-eye view with 10 million annotated hand poses. We also introduce a powerful baseline model that generates hand motions from piano audios through a position predictor and a position-guided gesture generator. Furthermore, a series of evaluation metrics are designed to assess the performance of the baseline model, including motion similarity, smoothness, positional accuracy of left and right hands, and overall fidelity of movement distribution. Despite that piano key presses with respect to music scores or audios are already accessible, PianoMotion10M aims to provide guidance on piano fingering for instruction purposes.
Overview of our framework. We collect videos of expert piano performances from the internet and annotated and processed them to obtain a large-scale dataset, PianoMotion10M, comprising piano music and hand motions. Building upon this dataset, we establish a benchmark aimed at generating hand movements from piano music.
Sample results of our generation model.
sample_out_1.mp4
sample_out_2.mp4
sample_out_3.mp4
Results from the PianoMotion10M paper
Method | Backbone | Decoder | FID | PARAMs |
---|---|---|---|---|
EmoTalk | HuBert | Transformer | 4.645 | 308 |
LivelySpeaker | HuBert | Transformer | 4.157 | 321 |
Our-Base | Wav2Vec2.0 | SSM | 3.587 | 320 |
Our-Base | Wav2Vec2.0 | Transformer | 3.608 | 323 |
Our-Base | HuBert | SSM | 3.412 | 320 |
Our-Base | HuBert | Transformer | 3.529 | 323 |
Our-Large | Wav2Vec2.0 | SSM | 3.453 | 539 |
Our-Large | Wav2Vec2.0 | Transformer | 3.376 | 557 |
Our-Large | HuBert | SSM | 3.395 | 539 |
Our-Large | HuBert | Transformer | 3.281 | 557 |
Notes:
- All the experiments are performed on 1 NVIDIA GeForce RTX 3090Ti GPU.
If you find PianoMotion10M is useful in your research or applications, please consider giving us a star 🌟 and citing it by the following BibTeX entry.
@inproceedings{gan2024pianomotion,
title={PianoMotion10M: Dataset and Benchmark for Hand Motion Generation in Piano Performance},
author={Gan, Qijun and Wang, Song and Wu, Shengtao and Zhu, Jianke},
year={2024},
}