๐ค AI Summary
This work addresses the challenges of cross-modal 3D reconstruction from heterogeneous RGB (appearance details) and X-ray (internal structure) imagesโnamely, severe modality discrepancy, difficult image registration, and scarcity of paired training data. We propose a radiography-aware Gaussian splatting method tailored for cross-modal reconstruction. Its core innovation lies in a hierarchical geometry-appearance joint optimization strategy, incorporating an X-ray projection consistency loss to enable, for the first time, end-to-end co-modeling of RGB and X-ray signals within the 3D Gaussian splatting framework. The method requires no strict pixel-level correspondence and supports non-invasive, unified 3D representation integrating both surface appearance and internal anatomy. Evaluated on a newly constructed paired dataset, our approach significantly improves surface visual fidelity and internal geometric consistency of reconstructions. It establishes a novel, interpretable, and high-fidelity multi-modal 3D modeling paradigm with direct applicability to medical diagnosis and non-destructive cultural heritage inspection.
๐ Abstract
We introduce InsideOut, an extension of 3D Gaussian splatting (3DGS) that bridges the gap between high-fidelity RGB surface details and subsurface X-ray structures. The fusion of RGB and X-ray imaging is invaluable in fields such as medical diagnostics, cultural heritage restoration, and manufacturing. We collect new paired RGB and X-ray data, perform hierarchical fitting to align RGB and X-ray radiative Gaussian splats, and propose an X-ray reference loss to ensure consistent internal structures. InsideOut effectively addresses the challenges posed by disparate data representations between the two modalities and limited paired datasets. This approach significantly extends the applicability of 3DGS, enhancing visualization, simulation, and non-destructive testing capabilities across various domains.