🤖 AI Summary
Raman spectroscopy on autonomous exploration platforms (e.g., Mars rovers, deep-sea landers) suffers from fluorescence-induced baseline distortion, peak shift, and severe label scarcity. Method: We propose an end-to-end 1D convolutional neural network framework that directly classifies raw spectra without baseline correction or handcrafted peak extraction. Contribution/Results: Key innovations include a baseline-agnostic classification mechanism, tunable pooling for robustness control, a semi-supervised efficient learning strategy, and a constant-time adaptive transfer method. Technically, the framework integrates contrastive pretraining, semi-supervised GANs, and frozen-backbone fine-tuning. Evaluated on the RRUFF subset, it achieves an 11% accuracy gain using only 10% labeled data, enables O(1)-time adaptation to novel minerals, and significantly outperforms Siamese networks and traditional feature-engineering approaches—while maintaining high robustness and suitability for edge deployment.
📝 Abstract
Autonomous Raman instruments on Mars rovers, deep-sea landers, and field robots must interpret raw spectra distorted by fluorescence baselines, peak shifts, and limited ground-truth labels. Using curated subsets of the RRUFF database, we evaluate one-dimensional convolutional neural networks (CNNs) and report four advances: (i) Baseline-independent classification: compact CNNs surpass $k$-nearest-neighbors and support-vector machines on handcrafted features, removing background-correction and peak-picking stages while ensuring reproducibility through released data splits and scripts. (ii) Pooling-controlled robustness: tuning a single pooling parameter accommodates Raman shifts up to $30 ,mathrm{cm}^{-1}$, balancing translational invariance with spectral resolution. (iii) Label-efficient learning: semi-supervised generative adversarial networks and contrastive pretraining raise accuracy by up to $11%$ with only $10%$ labels, valuable for autonomous deployments with scarce annotation. (iv) Constant-time adaptation: freezing the CNN backbone and retraining only the softmax layer transfers models to unseen minerals at $mathcal{O}(1)$ cost, outperforming Siamese networks on resource-limited processors. This workflow, which involves training on raw spectra, tuning pooling, adding semi-supervision when labels are scarce, and fine-tuning lightly for new targets, provides a practical path toward robust, low-footprint Raman classification in autonomous exploration.