REFA: Real-time Egocentric Facial Animations for Virtual Reality

πŸ“… 2024-06-17
πŸ›οΈ 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)
πŸ“ˆ Citations: 4
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the lack of non-invasive, calibration-free solutions for real-time facial expression capture in current virtual reality systems. The authors propose a distillation-based heterogeneous data training framework that leverages egocentric facial images captured by built-in infrared cameras in VR headsets. By fusing synthetic and real multi-source data and integrating a lightweight capture system with a differentiable rendering pipeline, the method achieves high-fidelity, calibration-free facial animation. The study introduces a large-scale facial dataset encompassing 18,000 diverse users and demonstrates low-latency real-time performance under a mobile–VR collaborative architecture, making it suitable for applications such as video conferencing, gaming, and remote collaboration.

Technology Category

Application Category

πŸ“ Abstract
We present a novel system for real-time tracking of facial expressions using egocentric views captured from a set of infrared cameras embedded in a virtual reality (VR) headset. Our technology facilitates any user to accurately drive the facial expressions of virtual characters in a non-intrusive manner and without the need of a lengthy calibration step. At the core of our system is a distillation based approach to train a machine learning model on heterogeneous data and labels coming form multiple sources, e.g. synthetic and real images. As part of our dataset, we collected 18k diverse subjects using a lightweight capture setup consisting of a mobile phone and a custom VR headset with extra cameras. To process this data, we developed a robust differentiable rendering pipeline enabling us to automatically extract facial expression labels. Our system opens up new avenues for communication and expression in virtual environments, with applications in video conferencing, gaming, entertainment, and remote collaboration.
Problem

Research questions and friction points this paper is trying to address.

facial animation
virtual reality
real-time tracking
egocentric vision
expression transfer
Innovation

Methods, ideas, or system contributions that make the work stand out.

egocentric facial animation
real-time expression tracking
distillation-based learning
differentiable rendering
VR headset
πŸ”Ž Similar Papers
No similar papers found.
Q
Qiang Zhang
Reality Labs at Meta, Menlo Park, USA
Tong Xiao
Tong Xiao
Facebook Inc
Computer VisionDeep Learning
H
Haroun Habeeb
Reality Labs at Meta, Menlo Park, USA
Larissa Laich
Larissa Laich
Unknown affiliation
Sofien Bouaziz
Sofien Bouaziz
Meta
Computer GraphicsComputer VisionMachine Learning
Patrick Snape
Patrick Snape
Meta
Computer VisionPattern Recognition
W
Wenjing Zhang
Reality Labs at Meta, Menlo Park, USA
M
Matthew Cioffi
Reality Labs at Meta, Menlo Park, USA
Peizhao Zhang
Peizhao Zhang
Research Scientist, Meta MSL
Computer VisionComputer Graphics
P
Pavel Pidlypenskyi
Reality Labs at Meta, Menlo Park, USA
W
Winnie Lin
Reality Labs at Meta, Menlo Park, USA
L
Luming Ma
Reality Labs at Meta, Menlo Park, USA
M
Mengjiao Wang
Reality Labs at Meta, Menlo Park, USA
Kunpeng Li
Kunpeng Li
Research Scientist, Meta Superintelligence Labs
Computer VisionDeep Learning
Chengjiang Long
Chengjiang Long
Research Engineer/Tech Leader at ByteDance Inc.
Computer VisionComputer GraphicsMultimediaMachine LearningArtificial Intelligence
Steven Song
Steven Song
University of Chicago
machine learning for healthcare
M
Martin Prazak
Reality Labs at Meta, Menlo Park, USA
A
Alexander Sjoholm
Reality Labs at Meta, Menlo Park, USA
A
Ajinkya Deogade
Reality Labs at Meta, Menlo Park, USA
J
Jaebong Lee
Reality Labs at Meta, Menlo Park, USA
Julio Delgado Mangas
Julio Delgado Mangas
Principal Engineer, Machine Learning
Machine Learningroboticsauto-coding
A
A. Aubel
Reality Labs at Meta, Menlo Park, USA