🤖 AI Summary
Traditional harmonic tools require users to manually specify key signatures or chords, imposing high demands on musical literacy. This paper introduces AI Harmonizer—the first end-to-end neuro-symbolic generation system capable of automatically producing musically coherent four-part harmonizations from monophonic melodies, without key annotation, chord input, or human intervention. Methodologically, it integrates high-accuracy pitch detection, voice-leading modeling, and a customized symbolic music generation model to enable real-time melody-to-harmony mapping. Its core contribution lies in the first deep coupling of generative AI with symbolic music rules, ensuring both voice-leading correctness and expressive musicality. The system yields harmonizations with high consistency and stylistic richness, and supports offline deployment. Source code is publicly released, establishing a novel paradigm for real-time human–machine choral collaboration.
📝 Abstract
Vocals harmonizers are powerful tools to help solo vocalists enrich their melodies with harmonically supportive voices. These tools exist in various forms, from commercially available pedals and software to custom-built systems, each employing different methods to generate harmonies. Traditional harmonizers often require users to manually specify a key or tonal center, while others allow pitch selection via an external keyboard-both approaches demanding some degree of musical expertise. The AI Harmonizer introduces a novel approach by autonomously generating musically coherent four-part harmonies without requiring prior harmonic input from the user. By integrating state-of-the-art generative AI techniques for pitch detection and voice modeling with custom-trained symbolic music models, our system arranges any vocal melody into rich choral textures. In this paper, we present our methods, explore potential applications in performance and composition, and discuss future directions for real-time implementations. While our system currently operates offline, we believe it represents a significant step toward AI-assisted vocal performance and expressive musical augmentation. We release our implementation on GitHub.