Published multiple papers in top conferences such as ICML, ACL, NeurIPS, ICLR. Examples include: 'SWE-smith: Scaling Data for Software Engineering Agent', 'Organize the Web: Constructing Domains Enhances Pre-Training Data Curation', 'Metadata Conditioning Accelerates Language Model Pre-training', etc.
Research Experience
Interned at Ai2 during PhD. Currently working on training coding agents at Cursor.
Education
PhD in Computer Science at Princeton University, advised by Danqi Chen; Undergraduate at the University of Cambridge, advised by Adrian Weller.
Background
Research interests: building and understanding large language models, with a particular focus on their training data. Worked on projects like QuRating, WebOrganizer, ProLong, Masking Rates. Also studied why LMs are easy to adapt (via Kernel Behavior) and how to interpret their internal workings (Transformer Programs, Edge Pruning). Part of the team that built SWE-bench and SWE-agent.
Miscellany
Personal interests and other information not provided.