Quantum Large Language Model Fine-Tuning

📅 2025-04-11
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the limited accuracy of large language models (LLMs) on complex semantic tasks—such as sentiment prediction—during fine-tuning. We propose the first scalable hybrid quantum-classical architecture, co-designing a sentence-level Transformer with a parameterized quantum circuit (PQC) featuring long-range entanglement, augmented by a quantum state reloading mechanism and a noise-robust training strategy. Our method pioneers systematic integration of quantum modules into LLM fine-tuning. We rigorously validate component contributions via finite-sampling noise simulation, hyperparameter sensitivity analysis, and ablation studies. Experiments demonstrate that, at comparable parameter counts, our architecture achieves up to a 3.14% absolute accuracy gain over classical baselines; moreover, increasing qubit count yields consistent performance improvements, confirming both the efficacy and scalability of quantum enhancement.

Technology Category

Application Category

📝 Abstract
We introduce a hybrid quantum-classical deep learning architecture for large language model fine-tuning. The classical portion of the architecture is a sentence transformer that is powerful enough to display significant accuracy for complex tasks such as sentiment prediction. The quantum portion of the architecture consists of parameterized quantum circuits that utilize long-range connections between qubits. We analyze the performance of the hybrid models for various settings of hyperparameters, including the number of qubits, the depth of the quantum circuits, learning rate, number of re-uploading steps, etc. Based on a screening study of main effects, we show an overall improvement in prediction accuracy over a comparable classical baseline, with a trend of increasing accuracy with number of qubits. We observe up to $3.14%$ improvements in accuracy over classical architectures of comparable model size, within the set of hyperparameters probed in this study. We demonstrate the contribution of each module in our architecture through ablation studies. Our studies are based on finite shot-counts and include simulations based on noisy quantum gates.
Problem

Research questions and friction points this paper is trying to address.

Hybrid quantum-classical architecture for fine-tuning large language models
Improving prediction accuracy using parameterized quantum circuits
Analyzing performance under various hyperparameters and noise conditions
Innovation

Methods, ideas, or system contributions that make the work stand out.

Hybrid quantum-classical deep learning architecture
Parameterized quantum circuits with long-range qubit connections
Improved accuracy over classical baselines via quantum components
🔎 Similar Papers
No similar papers found.
S
Sang Hyub Kim
IonQ Inc., 4505 Campus Dr, College Park, MD 20740, USA
Jonathan Mei
Jonathan Mei
Staff Research Scientist at IonQ
Machine LearningComputer VisionQuantum ComputingSpectral Graph TheoryStatistical Signal/Image Processing
C
Claudio Girotto
IonQ Inc., 4505 Campus Dr, College Park, MD 20740, USA
M
Masako Yamada
IonQ Inc., 4505 Campus Dr, College Park, MD 20740, USA
Martin Roetteler
Martin Roetteler
IonQ
Quantum computingquantum applicationsquantum programmingquantum solutions