Browse publications on Google Scholar (top-right) ↗
Resume (English only)
Academic Achievements
Received the PAMI Young Researcher Award. Published multiple papers including NExF: Learning Neural Exposure Fields for View Synthesis, Bolt3D: Generating 3D Scenes in Seconds, EVER: Exact Volumetric Ellipsoid Rendering for Real-time View Synthesis, CAT4D: Create Anything in 4D with Multi-View Video Diffusion Models, Generative Multiview Relighting for 3D Reconstruction under Extreme Illumination Variation, SimVS: Simulating World Inconsistencies for Robust View Synthesis, A Power Transform, CAT3D: Create Anything in 3D with Multi-View Diffusion Models.
Research Experience
Principal research scientist at Google DeepMind in San Francisco, leading a small team that mostly works on NeRF. At Google, he has worked on projects such as Glass, Lens Blur, HDR+, VR, Portrait Mode, Portrait Light, Maps, and Shopping.
Education
Ph.D. from UC Berkeley, advised by Jitendra Malik.
Background
Research interests include computer vision, deep learning, generative AI, and image processing. Most of his research is about inferring the physical world (shape, motion, color, light, etc) from images, usually with radiance fields.