🤖 AI Summary
Facial behavior analysis—encompassing facial landmark localization, action unit recognition, gaze estimation, and emotion recognition—faces significant challenges in cross-scenario generalization and computational efficiency. To address these, this work introduces the first lightweight unified multi-task model for facial behavior analysis. Our approach employs a shared backbone network coupled with task-specific lightweight heads, end-to-end joint optimization, and a cross-domain robust training paradigm to enhance generalization across diverse demographics, head poses, illumination conditions, and image resolutions. The resulting model achieves state-of-the-art or competitive accuracy on multiple benchmarks while reducing inference latency by 40% and memory footprint by 55%. All code and pre-trained models are fully open-sourced, enabling plug-and-play deployment and community-driven extension.
📝 Abstract
In recent years, there has been increasing interest in automatic facial behavior analysis systems from computing communities such as vision, multimodal interaction, robotics, and affective computing. Building upon the widespread utility of prior open-source facial analysis systems, we introduce OpenFace 3.0, an open-source toolkit capable of facial landmark detection, facial action unit detection, eye-gaze estimation, and facial emotion recognition. OpenFace 3.0 contributes a lightweight unified model for facial analysis, trained with a multi-task architecture across diverse populations, head poses, lighting conditions, video resolutions, and facial analysis tasks. By leveraging the benefits of parameter sharing through a unified model and training paradigm, OpenFace 3.0 exhibits improvements in prediction performance, inference speed, and memory efficiency over similar toolkits and rivals state-of-the-art models. OpenFace 3.0 can be installed and run with a single line of code and operate in real-time without specialized hardware. OpenFace 3.0 code for training models and running the system is freely available for research purposes and supports contributions from the community.