Published multiple papers including 'DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning', 'DeepSeek-V3 Technical Report', 'DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence', 'DeepSeekMath: Pushing the limits of mathematical reasoning in open language models', 'DeepSeek-Coder: When the Large Language Model Meets Programming--The Rise of Code Intelligence', 'GraphCodeBERT: Pre-training Code Representations with Data Flow', and 'CodeBERT: A Pre-Trained Model for Programming and Natural Languages'.
Research Experience
AI Researcher at DeepSeek (July 2023 to Present), working on code intelligence and LLM reasoning. Core contributor to projects such as DeepSeek-Coder, DeepSeekMath, etc. Research Intern at Microsoft Research Asia (May 2020 to May 2023), mentored by Dr. Nan Duan; Research Intern at Microsoft Research Asia (July 2017 to May 2020), mentored by Dr. Duyu Tang.
Education
PhD in Computer Science and Technology from Sun Yat-sen University (August 2018 to June 2023), supervised by Prof. Jian Yin and Dr. Ming Zhou. B.S. in Computer Science and Technology from Sun Yat-sen University (August 2014 to June 2018).
Background
Research interests: natural language processing and code intelligence. Long-term goal is to develop artificial general intelligence to revolutionize the way computers interact with humans and handle complex tasks. Research areas include: Large Language Model; Code Intelligence.