Alp-kd: Attention-based layer projection for knowledge distillation (AAAI 2021)
JABER and SABER: Junior and Senior Arabic BERT (ArXiv 2021)
Why skip if you can combine: A simple knowledge distillation technique for intermediate layers (EMNLP 2020)
Research Experience
Serves as an NLP researcher at Huawei Noah’s Ark Lab, Hong Kong, focusing on building pre-trained models and conducting research related to model compression. Worked at Huawei Canada Research Centre for 3 years prior to this role.
Education
Completed Master’s degree in Electrical and Computer Engineering from McGill University in 2019, under the supervision of Dr. Ioannis Psaromiligkos. Received Bachelor’s degree in Biomedical Engineering from Tianjin University in 2017.
Background
Currently an NLP researcher at Huawei Noah’s Ark Lab in Hong Kong SAR. Previously worked at Huawei Canada Research Centre for 3 years. Research interests include LLM pretraining, knowledge distillation, and machine translation.
Miscellany
Looking for research interns interested in LLMs, especially MoE pretraining.