Reasoning-Enhanced Self-Training for Long-Form Personalized Text Generation (arXiv preprint, 2025)
Bridging the Preference Gap between Retrievers and LLMs (to appear in ACL '24)
PRewrite: Prompt Rewriting with Reinforcement Learning (to appear in ACL '24)
Gemini: A Family of Highly Capable Multimodal Models (in arXiv, 2023)
SparseEmbed: Learning Sparse Lexical Representations with Contextual Embeddings for Retrieval (SIGIR 2023, Best Short Paper Award)
Multi-Aspect Dense Retrieval (KDD 2022)
Natural Language Understanding with Privacy-Preserving BERT (CIKM 2021)
Research Experience
Formerly a Staff Research Scientist at Google DeepMind, focusing on knowledge augmentation and personalization for LLMs; now a researcher at OpenAI.
Education
Ph.D. student, advised by Prof. James Allan at the Center for Intelligent Information Retrieval (CIIR), University of Massachusetts Amherst; previously worked with Prof. Yiqun Liu in the Information Retrieval Group, Tsinghua University.
Background
Research interests include knowledge augmentation (retrieval-augmented generation and long-context models) and personalization for LLMs. Currently a researcher at OpenAI, focusing on agents.