🤖 AI Summary
Systematic trend analysis of research addressing large language model (LLM) limitations—such as reasoning failures, hallucination, and weak multilingual capabilities—remains lacking. Method: Leveraging 250,000 papers from ACL and arXiv (2022–2024), this work proposes the first conceptual framework and reproducible quantitative paradigm for “LLM Limitations Research” (LLLMS). It integrates keyword filtering, LLM-assisted classification, dual clustering via HDBSCAN and BERTopic, LlooM-based topic modeling, and expert validation to construct the first publicly annotated LLLMS corpus (14,648 papers). Contribution/Results: LLLMS constitutes over 30% of all LLM-related publications. Temporal analysis reveals divergent disciplinary emphases: arXiv exhibits a marked shift toward safety, controllability, and multimodality, whereas ACL consistently prioritizes foundational capability deficits. This framework enables scalable, interpretable, and domain-grounded analysis of LLM limitations research.
📝 Abstract
Large language model (LLM) research has grown rapidly, along with increasing concern about their limitations such as failures in reasoning, hallucinations, and limited multilingual capability. In this survey, we conduct a data-driven, semi-automated review of research on limitations of LLM (LLLMs) from 2022 to 2024 using a bottom-up approach. From a corpus of 250,000 ACL and arXiv papers, we identify 14,648 relevant papers using keyword filtering, LLM-based classification, validated against expert labels, and topic clustering (via two approaches, HDBSCAN+BERTopic and LlooM). We find that LLM-related research increases over fivefold in ACL and fourfold in arXiv. Since 2022, LLLMs research grows even faster, reaching over 30% of LLM papers by late 2024. Reasoning remains the most studied limitation, followed by generalization, hallucination, bias, and security. The distribution of topics in the ACL dataset stays relatively stable over time, while arXiv shifts toward safety and controllability (with topics like security risks, alignment, hallucinations, knowledge editing), and multimodality between 2022 and 2024. We release a dataset of annotated abstracts and a validated methodology, and offer a quantitative view of trends in LLM limitations research.