🤖 AI Summary
This work addresses the challenge of jointly achieving fine-grained representation learning and perceptual consistency in instrument-level music similarity modeling—without requiring clean, isolated instrument signals during inference. Methodologically, we propose a novel learning framework integrating (i) cascade-style end-to-end fine-tuning (Cascade+E2E-FT), (ii) perceptual alignment fine-tuning (PAFT), and (iii) disentangled-feature-driven multi-task direct learning—built upon music source separation (MSS). This enables joint optimization of representation disentanglement and auditory preference alignment. Experiments demonstrate that our approach significantly outperforms the Direct multi-task baseline on instrument-level similarity tasks, while concurrently improving separation robustness, feature disentanglement quality, and perceptual consistency. The framework establishes a new paradigm for unsupervised and weakly supervised instrument perception modeling.
📝 Abstract
This paper proposes music similarity representation learning (MSRL) based on individual instrument sounds (InMSRL) utilizing music source separation (MSS) and human preference without requiring clean instrument sounds during inference. We propose three methods that effectively improve performance. First, we introduce end-to-end fine-tuning (E2E-FT) for the Cascade approach that sequentially performs MSS and music similarity feature extraction. E2E-FT allows the model to minimize the adverse effects of a separation error on the feature extraction. Second, we propose multi-task learning for the Direct approach that directly extracts disentangled music similarity features using a single music similarity feature extractor. Multi-task learning, which is based on the disentangled music similarity feature extraction and MSS based on reconstruction with disentangled music similarity features, further enhances instrument feature disentanglement. Third, we employ perception-aware fine-tuning (PAFT). PAFT utilizes human preference, allowing the model to perform InMSRL aligned with human perceptual similarity. We conduct experimental evaluations and demonstrate that 1) E2E-FT for Cascade significantly improves InMSRL performance, 2) the multi-task learning for Direct is also helpful to improve disentanglement performance in the feature extraction, 3) PAFT significantly enhances the perceptual InMSRL performance, and 4) Cascade with E2E-FT and PAFT outperforms Direct with the multi-task learning and PAFT.