Visualising Pianists' Touch: Transcribing Expressive Piano Performance from Audio to Piano Key Motion

要旨

Detailed measurements of piano key motion capture touch, timing, and dynamic control, providing crucial performance insights. Such expressive gestures are overlooked in MIDI, which only records pitch onset, duration, and velocity. Here, we introduce a novel transcription technique that directly maps audio from expressive piano performance to continuous piano key motion. User studies reveal a preference to the transcribed key motion trajectories over MIDI in representing sound, and over 80% accuracy in matching transcribed trajectories to audio from contrasting piano expressions. Follow-up interviews further indicate that the visualised trajectories can reveal subtle performance nuances and provide actionable guidance for both teaching and practice. An interface example for pedagogy and performance analysis utilising our technique is also illustrated. By providing a physically grounded performance representation that musicians can interpret and act upon, this work establishes a foundation for future interactive tools in music pedagogy, performance feedback, and embodied musical learning.

著者
Jingjing Tang
Queen Mary University of London, London, United Kingdom
Shinichi Furuya
Sony Computer Science Laboratories Inc., Shinagawa, Tokyo, Japan
Hayato Nishioka
Sony Computer Science Laboratories Inc., Tokyo, Japan
Momoko Shioki
Sony Computer Science Laboratories Inc., Shinagawa, Tokyo, Japan
Geraint A.. Wiggins
Queen Mary University of London, London, United Kingdom
George Fazekas
Queen Mary University of London, London, United Kingdom
Vincent K.M.. Cheung
Sony Computer Science Laboratories Inc., Shinagawa, Tokyo, Japan

会議: CHI 2026

ACM CHI Conference on Human Factors in Computing Systems

セッション: Immersive and Spatial Visualization

P1 - Room 128
7 件の発表
2026-04-15 20:15:00
2026-04-15 21:45:00