π€ AI Summary
Addressing the challenge of modeling subtle facial muscle movements in facial expression quality assessment (FEQA) for neurological disorder diagnosis, this paper proposes TraMP-Formerβa trajectory-guided motion-aware Transformer architecture. To capture fine-grained dynamic facial behavior, TraMP-Former uniquely fuses structured facial landmark trajectories with RGB semantic features via trajectory encoding, multimodal alignment, spatiotemporal attention, and a regression-oriented Transformer decoder, enabling end-to-end prediction of FEQA scores. Evaluated on PFED5 and an enhanced Toronto NeuroFace dataset, it achieves new state-of-the-art performance, improving accuracy by 6.51% and 7.62%, respectively. Ablation studies quantitatively validate that explicit modeling of landmark trajectories yields substantial gains for FEQA, confirming their critical role in characterizing neuromotor deficits.
π Abstract
Automated facial expression quality assessment (FEQA) in neurological disorders is critical for enhancing diagnostic accuracy and improving patient care, yet effectively capturing the subtle motions and nuances of facial muscle movements remains a challenge. We propose to analyse facial landmark trajectories, a compact yet informative representation, that encodes these subtle motions from a high-level structural perspective. Hence, we introduce Trajectory-guided Motion Perception Transformer (TraMP-Former), a novel FEQA framework that fuses landmark trajectory features for fine-grained motion capture with visual semantic cues from RGB frames, ultimately regressing the combined features into a quality score. Extensive experiments demonstrate that TraMP-Former achieves new state-of-the-art performance on benchmark datasets with neurological disorders, including PFED5 (up by 6.51%) and an augmented Toronto NeuroFace (up by 7.62%). Our ablation studies further validate the efficiency and effectiveness of landmark trajectories in FEQA. Our code is available at https://github.com/shuchaoduan/TraMP-Former.