Bridging Classical and Quantum Machine Learning: Knowledge Transfer From Classical to Quantum Neural Networks Using Knowledge Distillation

📅 2023-11-23
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the high resource consumption, noise sensitivity, and hardware scalability limitations of quantum neural networks (QNNs), this paper proposes the first cross-paradigm knowledge distillation framework—classical CNN to QNN. Methodologically, it integrates amplitude encoding with dimensionality-reduction techniques (PCA, pooling, or center cropping) in fully connected layers to construct parameter-efficient and error-robust parametrized quantum circuits (PQCs). A systematic comparison of amplitude, angle, and qubit encodings reveals amplitude encoding achieves superior performance on image classification tasks. Experiments demonstrate that 4- and 8-qubit QNNs trained via this framework attain significantly improved accuracy on MNIST, Fashion-MNIST, and CIFAR-10—matching classical CNN performance while using orders-of-magnitude fewer parameters. These results validate the framework’s effectiveness and generalizability under stringent resource constraints.
📝 Abstract
Quantum neural networks (QNNs), harnessing superposition and entanglement, have shown potential to surpass classical methods in complex learning tasks but remain limited by hardware constraints and noisy conditions. In this work, we present a novel framework for transferring knowledge from classical convolutional neural networks (CNNs) to QNNs via knowledge distillation, thereby reducing the need for resource intensive quantum training and error mitigation. We conduct extensive experiments using two parameterized quantum circuits (PQCs) with 4 and 8 qubits on MNIST, Fashion MNIST, and CIFAR10 datasets. The approach demonstrates consistent accuracy improvements attributed to distilled knowledge from larger classical networks. Through ablation studies, we systematically compare the effect of state of the art dimensionality reduction techniques fully connected layers, center cropping, principal component analysis, and pooling to compress high-dimensional image data prior to quantum encoding. Our findings reveal that fully connected layers retain the most salient features for QNN inference, thereby surpassing other down sampling approaches. Additionally, we examine state of the art data encoding methods (amplitude, angle, and qubit encoding) and identify amplitude encoding as the optimal strategy, yielding superior accuracy across all tested datasets and qubit configurations. Through computational analyses, we show that our distilled 4-qubit and 8-qubit QNNs achieve competitive performance while utilizing significantly fewer parameters than their classical counterparts. Our results establish a promising paradigm for bridging classical deep learning and emerging quantum computing, paving the way for more powerful, resource conscious models in quantum machine intelligence.
Problem

Research questions and friction points this paper is trying to address.

Transfer knowledge from classical to quantum neural networks.
Reduce resource-intensive quantum training and error mitigation.
Improve QNN accuracy using distilled knowledge from classical networks.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Knowledge distillation from classical to quantum networks
Optimal data encoding using amplitude encoding
Fully connected layers for salient feature retention
Mohammad Junayed Hasan
Mohammad Junayed Hasan
MSE in CS at Johns Hopkins University
Natural Language ProcessingComputer VisionMachine LearningQuantum Machine Learning
M
M.R.C. Mahdy
Department of Electrical and Computer Engineering, North South University, Bashundhara, Dhaka