- 'MC-BERT: Efficient Language Pre-Training via a Meta Controller', 2020
- 'Microsoft Research Asia's Systems for WMT19', WMT19 (ACL 2019 Workshop)
- 'Efficient training of BERT by progressively stacking', ICML 2019
Research Experience
- Research projects involve the application of large language models in code generation, infilling, transpilation, and understanding.
- Participated in multiple research projects related to code generation and understanding, such as AST-T5, SAFIM, etc.
Education
- Ph.D. in Computer Science, 2020 - Present
- University: University of California, Berkeley
- Advisors: Prof. Alvin Cheung and Prof. Dawn Song
- B.S. in Computer Science, 2016 - 2020
- University: Peking University, Beijing, China
- Advisors: Prof. Liwei Wang and Prof. Di He
Background
Research Interests: Artificial Intelligence, Natural Language Processing, Large Language Models. Specializes in pretraining, prompting, and evaluation methodologies for a variety of language models, including BERT, T5, and GPT-like LLMs. Recent research focuses on leveraging LLMs for code generation, infilling, transpilation, and understanding.