Two open-source projects have been downloaded by 20,000 people and liked by 5,000 since their release; led or co-led research projects that received distinctions such as Best Paper Award (ACL'25), Best Paper Award (NeurIPS'24 Pluralistic Alignment Workshop), and Spotlight (NeurIPS'24).
Research Experience
Currently an Anthropic AI Safety Fellow; previously worked with the UC Berkeley Center for Human-Compatible AI, and has also been a member of the PKU Alignment Team; part-time mentor for the Supervised Program for Alignment Research and the Algoverse AI Safety Fellowship.
Background
Research Interests: AI alignment, human truth-seeking and moral progress. Professional Fields: Computer Scientist, Machine Learning Researcher/Engineer, Statistician.
Miscellany
Personal interests include experimental research, loves to dig deep into findings in search for a generalizeable & falsifiable theory.