Scholar
Yingfa Chen
Google Scholar ID: IgPWvEQAAAAJ
PhD at Tsinghua University
machine learning
long-context modeling
language modeling
Follow
Homepage
↗
Google Scholar
↗
Citations & Impact
All-time
Citations
434
H-index
8
i10-index
7
Publications
14
Co-authors
0
Contact
Twitter
Open ↗
GitHub
Open ↗
Publications
8 items
Student-in-the-Loop Chain-of-Thought Distillation via Generation-Time Selection
2026
Cited
0
MiniCPM-SALA: Hybridizing Sparse and Linear Attention for Efficient Long-Context Modeling
2026
Cited
0
Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts
2026
Cited
1
StateX: Enhancing RNN Recall via Post-training State Expansion
2025
Cited
0
BlockFFN: Towards End-Side Acceleration-Friendly Mixture-of-Experts with Chunk-Level Activation Sparsity
2025
Cited
0
Cost-Optimal Grouped-Query Attention for Long-Context LLMs
2025
Cited
0
Sparsing Law: Towards Large Language Models with Greater Activation Sparsity
arXiv.org · 2024
Cited
1
Stuffed Mamba: State Collapse and State Capacity of RNN-Based Long-Context Modeling
arXiv.org · 2024
Cited
2
Resume (English only)
Academic Achievements
Contributed to the COLING 2024 paper 'Robust and Scalable Model Editing for Large Language Models'.
Research Experience
Currently working on large language models, long-context modeling, and continual learning.
Education
PhD student at Tsinghua University, advised by Prof. Zhiyuan Liu.
Background
NLP PhD student at Tsinghua University, interested in LLM architectures, long-context modeling, and continual learning.
Miscellany
Sports: Badminton, jogging, etc.
Co-authors
0 total
Co-authors: 0 (list not available)
×
Welcome back
Sign in to Agora
Welcome back! Please sign in to continue.
Email address
Password
Forgot password?
Continue
Do not have an account?
Sign up