Published multiple papers including 'Flattery, Fluff, and Fog: Diagnosing and Mitigating Idiosyncratic Biases in Preference Models' and 'EvalAgent: Discovering Implicit Evaluation Criteria from the Web'. Involved in projects like DOLIMITES and AssistantBench. Received the Outstanding Paper Award at ACL 2023.
Research Experience
Currently a Senior Research Scientist at Google DeepMind, working on evaluation and post-training of language models. During his Ph.D., he worked with the Semantic Scholar team at Ai2 and the Gemini team at Google DeepMind. Previously, he was a predoctoral young investigator at the Allen Institute for Artificial Intelligence.
Education
Ph.D. in Computer Science from the University of Pennsylvania, advised by Mark Yatskar and Dan Roth; Master's degree from the Language Technologies Institute at Carnegie Mellon University; Bachelor's degree from Nanyang Technological University, Singapore.
Background
Research interests: evaluation and benchmarking of language models in realistic scenarios, aligning models to diverse users (especially domain experts), learning from human feedback. Broadly, interested in how to use human knowledge and expertise to improve, align, and evaluate language models. Also keen on cognitive science and linguistics.
Miscellany
Personal interests include cognitive science and linguistics.