🤖 AI Summary
Large language models are vulnerable to resource-consuming attacks—such as those maliciously inducing excessive generation—which jeopardize service availability and economic sustainability. This work presents the first systematic survey of this emerging threat class, introducing a unified analytical framework that encompasses threat elicitation, operational mechanisms, and mitigation strategies. Through a comprehensive literature review integrated with threat modeling, efficiency analysis, and categorization of defense mechanisms, the study clearly delineates the problem boundaries and maps the research landscape. The resulting synthesis establishes a foundational theoretical basis and outlines a coherent research agenda for future efforts in threat characterization, detection, and defense.
📝 Abstract
Given limited and costly computational infrastructure, resource efficiency is a key requirement for large language models (LLMs). Efficient LLMs increase service capacity for providers and reduce latency and API costs for users. Recent resource consumption threats induce excessive generation, degrading model efficiency and harming both service availability and economic sustainability. This survey presents a systematic review of threats to resource consumption in LLMs. We further establish a unified view of this emerging area by clarifying its scope and examining the problem along the full pipeline from threat induction to mechanism understanding and mitigation. Our goal is to clarify the problem landscape for this emerging area, thereby providing a clearer foundation for characterization and mitigation.