Rendering Music Performance With Interpretation Variations Using Conditional Variational RNN
Akira Maezawa, Kazuhiko Yamamoto, Takuya Fujishima
Abstract
Capturing and generating a wide variety of musical expression is important in music performance rendering, but current methods fail to model such a variation. This paper presents a music performance rendering method that could explicitly model differences in interpretations for a given piece of music. Conditional variational auto-encoder is used to jointly train, conditioned on the music score, an encoder from performance to a latent code and a decoder from the latent code to music performance. Evaluation demonstrates the method is capable of generating a wide variety of human-like expressive music performances as the latent code is varied.
Related Papers
RenderBox: Expressive Performance Rendering with Text Control2025-02-11DExter: Learning and Controlling Performance Expression with Diffusion Models2024-06-21ScorePerformer: Expressive Piano Performance Rendering With Fine-Grained Control2023-11-04Reconstructing Human Expressiveness in Piano Performances with a Transformer Network2023-06-09VirtuosoNet: A Hierarchical RNN-based System for Modeling Expressive Piano Performance2019-11-04Graph Neural Network for Music Score Data and Modeling Expressive Piano Performance2019-06-11