Mingfei Chen
Scholar

Mingfei Chen

Google Scholar ID: uK7MW8QAAAAJ
University of Washington
Computer Vision
Citations & Impact
All-time
Citations
364
 
H-index
6
 
i10-index
5
 
Publications
10
 
Co-authors
16
list available
Resume (English only)
Academic Achievements
  • Awarded the Google PhD Fellowship 2025 in Machine Perception (North America). One first-author paper on spatial audio-visual LLMs accepted by NeurIPS as Oral (acceptance rate < 0.4%). One first-author paper on spatial audio-visual reconstruction accepted by CVPR2025 as Highlight (acceptance rate < 2.9%). One first-author paper on spatial audio-visual reconstruction accepted by NeurIPS2024. One first-author paper on audio-visual learning accepted by ICCV2023. One co-first author paper on implicit neural acoustic fields accepted by NeurIPS2022. One first-author paper on 3D photo-realistic digital human rendering accepted by ECCV2022. One first-author paper on human-object interaction accepted by CVPR2021. Undergraduate thesis on Language-guided Video Retrieval awarded Outstanding undergraduate graduation thesis at Huazhong University of Science and Technology.
Research Experience
  • Worked on human detection and segmentation at Bytedance AI Lab; collaborated with Prof. Si Liu on human-object interaction (HOI); worked with Prof. Jenq-Neng Hwang on multiple-object tracking (MOT) in IPL lab; conducted 3D photo-realistic digital human rendering with Prof. Shuicheng Yan and Prof. Jiashi Feng in Sea AI Lab; and engaged in wonderful multi-modality 3D research at Meta Reality Labs.
Education
  • Received a Bachelor’s degree in Computer Science and Technology from the Experimental Program for Exemplary Engineers at Huazhong University of Science and Technology (HUST), China, in 2020. Currently pursuing a Ph.D. at the University of Washington, Seattle, advised by Prof. Eli Shlizerman.
Background
  • Currently a third-year Ph.D. student in the Electrical and Computer Engineering department at the University of Washington, Seattle. Research interests include multi-modal research for spatial reasoning in 3D scenes, as well as related applications on multi-modal LLMs, XR devices, and robotics. Supported by the Google PhD Fellowship.
Miscellany
  • Open to research collaboration. Please email (lasiafly [at] uw.edu) if interested in exploring more on multi-modal for spatial reasoning in 3D together!