NoteIt: A System Converting Instructional Videos to Interactable Notes Through Multimodal Video Understanding (UIST 2025)
DietGlance: Dietary Monitoring and Personalized Analysis at a Glance with Knowledge-Empowered AI Assistant (arXiv:2502.01317)
USpeech: Ultrasound-Enhanced Speech with Minimal Human Effort via Cross-Modal Synthesis (IMWUT/UbiComp 2025)
Internal Cross-layer Gradients for Extending Homogeneity to Heterogeneity in Federated Learning (ICLR 2024)
Radio2Text: Streaming Speech Recognition Using mmWave Radio Signals (IMWUT/UbiComp 2023)
SPACE: Speaker Adaptation for Acoustic Eavesdropping using mmWave Radio Signals (IEEE Transactions on Mobile Computing)
Radio2Speech: High Quality Speech Recovery from Radio Frequency Signals (INTERSPEECH 2022)
Continual Learning with Strategic Selection and Forgetting for Network Intrusion Detection (IEEE INFOCOM 2025)
Explainable framework to detect Parkinson's disease related depression from EEG (IEEE EMBC 2024)
Research Experience
Previously worked with Dr. Xun Qian from Google and Prof. Hang Zhao from IIIS, Tsinghua University.
Education
PhD student in the Department of Electrical and Electronic Engineering at The University of Hong Kong, advised by Prof. Edith C.H. Ngai.
Background
Research interests include human-AI interaction, mobile and IoT computing, and multimodal learning, with a particular emphasis on using different modalities to enhance human creativity and perception of the physical world.