🤖 AI Summary
Estimating dynamic levels in piano performance remains a fundamental challenge in computational music analysis. This paper introduces a novel multi-task, multi-scale neural network that achieves, for the first time, end-to-end joint modeling of four interrelated musical attributes directly from audio: dynamic level (e.g., *p*, *f*), dynamic change points, beat positions, and downbeats. Leveraging Bark-scale loudness features and a shared latent representation architecture, the model significantly reduces parameter count while preserving temporal modeling capacity; it supports 60-second audio sequences, balancing expressiveness and computational efficiency. Evaluated on the MazurkaBL dataset, our approach achieves state-of-the-art performance across all four tasks, compressing model parameters from 14.7M to just 0.5M. This work establishes the first compact, efficient, and structurally coherent benchmark model for piano expressivity analysis.
📝 Abstract
Estimating piano dynamic from audio recordings is a fundamental challenge in computational music analysis. In this paper, we propose an efficient multi-task network that jointly predicts dynamic levels, change points, beats, and downbeats from a shared latent representation. These four targets form the metrical structure of dynamics in the music score. Inspired by recent vocal dynamic research, we use a multi-scale network as the backbone, which takes Bark-scale specific loudness as the input feature. Compared to log-Mel as input, this reduces model size from 14.7 M to 0.5 M, enabling long sequential input. We use a 60-second audio length in audio segmentation, which doubled the length of beat tracking commonly used. Evaluated on the public MazurkaBL dataset, our model achieves state-of-the-art results across all tasks. This work sets a new benchmark for piano dynamic estimation and delivers a powerful and compact tool, paving the way for large-scale, resource-efficient analysis of musical expression.