Zhisheng Zheng
Scholar

Zhisheng Zheng

Google Scholar ID: WYwBrzAAAAAJ
The University of Texas at Austin
Speech and Language ProcessingNatural Language ProcessingMultimodal Learning
Citations & Impact
All-time
Citations
514
 
H-index
10
 
i10-index
10
 
Publications
16
 
Co-authors
5
list available
Resume (English only)
Academic Achievements
  • 1. 2 papers accepted by EMNLP 2025; 2. BAT accepted by ICML 2024; 3. EAT: Self-Supervised Pre-Training with Efficient Audio Transformer accepted by IJCAI 2024; 4. Released emotion2vec, the first universal speech emotion model that excels across diverse emotional tasks, languages; 5. 1 paper accepted by ICASSP 2024; 6. Released Fast-HuBERT, accelerating HuBERT pre-training in 5.2X speedup without performance drop; 7. 2 papers accepted by IEEE ASRU 2023; 8. MT4SSL nominated in ISCA Interspeech Best Student Paper Shortlist; 9. 3 papers accepted by ISCA INTERSPEECH 2023; 10. 1 paper accepted by ICASSP 2023.
Research Experience
  • 1. Research Intern at Microsoft Research Asia, mentored by Lei He and Xu Tan, focusing on Multilingual Text-to-Speech; 2. Research Intern at SALT Lab at UT-Austin during the summer of 2023, collaborating with Prof. David Harwath and Prof. Eunsol Choi; 3. Research Intern at X-Lance Lab at SJTU since 2021, supervised by Prof. Xie Chen.
Education
  • Ph.D. in Computer Science, 2024 - 2028 (expected), The University of Texas at Austin; BSc in Electrical Engineering & Zhiyuan Honors Program of Engineering, 2020 - 2024, Shanghai Jiao Tong University.
Background
  • Research Interests: Multimodal Large Language Model, Self-Supervised Learning, Speech and Audio Understanding. Background: He is a second-year Ph.D. student in Computer Science at the University of Texas at Austin.