DiSC-AMC: Token- and Parameter-Efficient Discretized Statistics In-Context Automatic Modulation Classification

📅 2025-09-30
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large language models (LLMs) face significant deployment challenges in open-set automatic modulation classification (AMC) due to excessive context length and parameter count. Method: We propose a lightweight in-context learning framework that (i) discretizes higher-order statistics and cumulants into symbolic tokens; (ii) integrates a k-top neural pre-filtering mechanism coupled with LLM feedback-driven exemplar selection; and (iii) employs a calibration-aware compact prompt template enforcing structured label output. Contribution/Results: Our approach drastically reduces contextual and model-scale requirements: on a ten-class AMC benchmark, a ~5B-parameter LLM achieves 45.5% accuracy—over 8× higher than a 7B baseline—while halving both input/output token count and model parameters. Inference cost is reduced by more than 2×, enabling the first practical, efficient deployment of small-scale LLMs for AMC.

Technology Category

Application Category

📝 Abstract
Large Language Models (LLMs) can perform Automatic Modulation Classification (AMC) in an open-set manner without LLM fine-tuning when equipped with carefully designed in-context prompts~cite{rostami2025plug}. Building on this prior work, we target the practical bottlenecks of long prompt contexts and large model sizes that impede in-the-loop deployment. We present Discretized Statistics in-Context Automatic Modulation Classification (DiSC-AMC), a token- and parameter-efficient variant that: (i) discretizes higher-order statistics and cumulants into compact symbolic tokens, (ii) prunes the exemplar list via a lightweight k-top neural prefilter and filters misleading/low-impact features using rationales extracted from prior LLM responses, and (iii) enforces label-only predictions through a calibrated prompt template. Together, these changes reduce both input/output tokens and the model parameter footprint by more than half while maintaining competitive accuracy. On synthetic AMC with ten modulation types under noise, a 7B extit{DeepSeek-R1-Distill-Qwen} baseline achieves 5.2% accuracy, whereas our system, using an approximately 5B-parameter extit{Gemini-2.5-Flash}~cite{comanici2025gemini} model, attains 45.5% accuracy. These results demonstrate that careful discretization and context selection can cut inference cost by over 2x while preserving the advantages of prompt-based AMC and enabling practical in-the-loop use.
Problem

Research questions and friction points this paper is trying to address.

Reduces long prompt contexts and large model sizes for practical deployment
Improves token and parameter efficiency in automatic modulation classification
Maintains competitive accuracy while cutting inference costs significantly
Innovation

Methods, ideas, or system contributions that make the work stand out.

Discretizes statistics into compact symbolic tokens
Prunes exemplars via lightweight neural prefilter
Enforces label-only predictions with calibrated prompt template
🔎 Similar Papers
No similar papers found.
M
Mohammad Rostami
Department of Electrical and Computer Engineering, Rowan University, Glassboro, NJ, USA
Atik Faysal
Atik Faysal
Graduate Research Assistant
Machine learningartificial intelligence
R
Reihaneh Gh. Roshan
Department of Computer Science, Stevens Institute of Technology, Hoboken, NJ, USA
Huaxia Wang
Huaxia Wang
Assistant Professor, Rowan University, Glassboro, NJ
Adversarial NetworksDeep Reinforcement LearningCognitive RadioWireless Communications
Nikhil Muralidhar
Nikhil Muralidhar
Stevens Institute of Technology
Data MiningMachine LearningPhysics Informed Machine LearningTime Series ForecastingAnomaly Detection
Y
Yu-Dong Yao
Department of Electrical and Computer Engineering, Stevens Institute of Technology, Hoboken, NJ, USA