Published numerous papers covering topics such as multimodal reward benchmarks, robustness of reward models, content-adaptive image tokenization, large language models as analogical reasoners, and more. Some of these papers were presented at top conferences like ICLR 2024, NeurIPS 2024, etc.
Research Experience
Worked as a researcher at Google DeepMind and Meta.
Education
Received a PhD in Computer Science from Stanford, advised by Percy Liang, Jure Leskovec, and Chris Manning.
Background
Interested in building LLMs and agents that assist humans in diverse tasks. His work spans: Post-training (RL, reward models, and evaluation), reasoning, retrieval and tool use for LLMs, and multimodality (vision-language models).