Published multiple papers in top international conferences such as NeurIPS, EMNLP, ACL, ICML, ICLR, COLING, ICASSP, AAAI between 2020 and 2025.
Research Experience
Worked at Natural Language Processing Lab; Internship at Microsoft Research Asia (MSRA) Natural Language Computing (NLC) from May 2022 to May 2023; Started a new internship at Machine Learning Group (ML) in December 2023.
Education
Bachelor's degree in 2017 from Northeastern University, majoring in Computer Science and Technology; Master's degree in 2020 from Northeastern University, majoring in Computer Software and Theory; Ph.D. from the Department of Computer Science and Technology at Northeastern University, supervised by Prof. Tong Xiao and Prof. Jingbo Zhu.
Background
Research interests include complex architecture modeling, deep transformers, multimodal modeling, and machine learning. Currently focusing on large language models such as prompt engineering via deliberation (DTG), evolutionary algorithms-based prompt search (EvoPrompt), foundation models (PCformer and its subsequent work), and DPO improvements (temporal-decay based DPO). Primary research domain is sequence generation tasks, including machine translation, abstractive summarization, etc.