QuPAINT: Physics-Aware Instruction Tuning Approach to Quantum Material Discovery

📅 2026-02-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenges of identifying two-dimensional quantum materials from optical microscopy images, which include low interlayer contrast, scarcity of annotated data, and poor generalization across diverse experimental conditions. To overcome these limitations, the authors propose a physics-aware multimodal framework that integrates a thin-film interference–based synthetic data generator (Synthia), a physics-informed attention mechanism, and instruction tuning of multimodal large language models. They also introduce QMat-Instruct, the first large-scale instruction dataset for quantum materials, along with QF-Bench, a comprehensive evaluation benchmark. The proposed approach substantially reduces reliance on manual annotations and achieves more robust and generalizable recognition performance across various materials, substrates, and imaging conditions, thereby advancing the integration of physical knowledge with multimodal foundation models.

Technology Category

Application Category

📝 Abstract
Characterizing two-dimensional quantum materials from optical microscopy images is challenging due to the subtle layer-dependent contrast, limited labeled data, and significant variation across laboratories and imaging setups. Existing vision models struggle in this domain since they lack physical priors and cannot generalize to new materials or hardware conditions. This work presents a new physics-aware multimodal framework that addresses these limitations from both the data and model perspectives. We first present Synthia, a physics-based synthetic data generator that simulates realistic optical responses of quantum material flakes under thin-film interference. Synthia produces diverse and high-quality samples, helping reduce the dependence on expert manual annotation. We introduce QMat-Instruct, the first large-scale instruction dataset for quantum materials, comprising multimodal, physics-informed question-answer pairs designed to teach Multimodal Large Language Models (MLLMs) to understand the appearance and thickness of flakes. Then, we propose Physics-Aware Instruction Tuning (QuPAINT), a multimodal architecture that incorporates a Physics-Informed Attention module to fuse visual embeddings with optical priors, enabling more robust and discriminative flake representations. Finally, we establish QF-Bench, a comprehensive benchmark spanning multiple materials, substrates, and imaging settings, offering standardized protocols for fair and reproducible evaluation.
Problem

Research questions and friction points this paper is trying to address.

quantum materials
optical microscopy
layer-dependent contrast
limited labeled data
generalization across imaging setups
Innovation

Methods, ideas, or system contributions that make the work stand out.

Physics-Aware Instruction Tuning
Synthetic Data Generation
Multimodal Large Language Models
Thin-Film Interference
Quantum Material Discovery
🔎 Similar Papers
No similar papers found.
X
Xuan-Bac Nguyen
CVIU Lab, University of Arkansas, USA
H
Hoang-Quan Nguyen
CVIU Lab, University of Arkansas, USA
S
Sankalp Pandey
CVIU Lab, University of Arkansas, USA
T
Tim Faltermeier
University of Utah, USA
N
Nicholas Borys
University of Utah, USA
H
Hugh Churchill
Department of Physics, University of Arkansas, USA
Khoa Luu
Khoa Luu
EECS Department, University of Arkansas
Smart HealthBiometricsAutonomous DrivingQuantum Machine LearningPrecision Agriculture