Quantum Machine Learning via Contrastive Training

📅 2025-11-17
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the poor generalization of quantum machine learning (QML) under scarce labeled data, this paper introduces the first end-to-end contrastive pretraining framework implemented natively on trapped-ion quantum hardware. Methodologically, images are encoded as quantum states, and a self-supervised contrastive loss—computed directly on hardware via quantum state overlap—is employed for representation learning, followed by downstream classification fine-tuning. The key contributions are: (i) the first hardware-native, label-free contrastive pretraining paradigm for QML; and (ii) learned quantum representations that exhibit strong robustness in few-shot settings, significantly improving mean classification accuracy while reducing runtime variance. Experimental results demonstrate enhanced learning efficiency and stability of QML models, establishing a novel pathway for efficient, quantum-native representation learning.

Technology Category

Application Category

📝 Abstract
Quantum machine learning (QML) has attracted growing interest with the rapid parallel advances in large-scale classical machine learning and quantum technologies. Similar to classical machine learning, QML models also face challenges arising from the scarcity of labeled data, particularly as their scale and complexity increase. Here, we introduce self-supervised pretraining of quantum representations that reduces reliance on labeled data by learning invariances from unlabeled examples. We implement this paradigm on a programmable trapped-ion quantum computer, encoding images as quantum states. In situ contrastive pretraining on hardware yields a representation that, when fine-tuned, classifies image families with higher mean test accuracy and lower run-to-run variability than models trained from random initialization. Performance improvement is especially significant in regimes with limited labeled training data. We show that the learned invariances generalize beyond the pretraining image samples. Unlike prior work, our pipeline derives similarity from measured quantum overlaps and executes all training and classification stages on hardware. These results establish a label-efficient route to quantum representation learning, with direct relevance to quantum-native datasets and a clear path to larger classical inputs.
Problem

Research questions and friction points this paper is trying to address.

Reducing labeled data dependency in quantum machine learning models
Implementing self-supervised contrastive training on quantum hardware
Improving classification accuracy for quantum-encoded image datasets
Innovation

Methods, ideas, or system contributions that make the work stand out.

Self-supervised pretraining of quantum representations
Contrastive training using measured quantum overlaps
Full training pipeline executed on quantum hardware
🔎 Similar Papers
No similar papers found.
L
Liudmila A. Zhukas
Duke Quantum Center, Duke University, Durham, NC 27701.
V
Vivian Ni Zhang
Duke Quantum Center, Duke University, Durham, NC 27701.
Q
Qiang Miao
Duke Quantum Center, Duke University, Durham, NC 27701.
Q
Qingfeng Wang
Department of Physics and Astronomy, Tufts University, Medford, MA 02115.
M
Marko Cetina
Duke Quantum Center, Duke University, Durham, NC 27701.
Jungsang Kim
Jungsang Kim
Duke Quantum Center, Duke University, Durham, NC 27701.
Lawrence Carin
Lawrence Carin
Duke University
machine learningartificial intelligencestatistics
Christopher Monroe
Christopher Monroe
Duke Quantum Center, Duke University, Durham, NC 27701.