π€ AI Summary
In real-world scenarios, audio-visual speech recognition (AVSR) suffers severe performance degradation under acoustic noise and visual occlusion, hindering practical deployment. To address this, we propose a robust and scalable AVSR framework tailored for open environments, jointly optimizing representation, architecture, and system design. First, we introduce hierarchical noise- and occlusion-robust audio-visual unified representation learningβa novel paradigm for joint multimodal representation. Second, we design an input-adaptive multimodal computation allocation mechanism to dynamically distribute computational resources across modalities. Third, we establish a modular, plug-and-play collaborative system architecture supporting seamless integration of large language models (LLMs) and audio-visual foundation models. Our method incorporates cross-modal feature alignment and fusion, dynamic resource scheduling, and extensible functional interfaces. Evaluated under diverse noise and occlusion conditions, our approach achieves a 32% relative reduction in word error rate (WER) and a 2.1Γ speedup in inference latency, significantly enhancing robustness and scalability in realistic settings.
π Abstract
The practical deployment of Audio-Visual Speech Recognition (AVSR) systems is fundamentally challenged by significant performance degradation in real-world environments, characterized by unpredictable acoustic noise and visual interference. This dissertation posits that a systematic, hierarchical approach is essential to overcome these challenges, achieving the robust scalability at the representation, architecture, and system levels. At the representation level, we investigate methods for building a unified model that learns audio-visual features inherently robust to diverse real-world corruptions, thereby enabling generalization to new environments without specialized modules. To address architectural scalability, we explore how to efficiently expand model capacity while ensuring the adaptive and reliable use of multimodal inputs, developing a framework that intelligently allocates computational resources based on the input characteristics. Finally, at the system level, we present methods to expand the system's functionality through modular integration with large-scale foundation models, leveraging their powerful cognitive and generative capabilities to maximize final recognition accuracy. By systematically providing solutions at each of these three levels, this dissertation aims to build a next-generation, robust, and scalable AVSR system with high reliability in real-world applications.