🤖 AI Summary
Current neural rendering methods (e.g., NeRF, Plenoxel) output view-dependent RGB images, which are incompatible with material jetting 3D printing—causing color inaccuracies and geometric distortions. To address this, we propose the first end-to-end printable voxel reconstruction framework tailored for multi-material jetting printing. Our approach: (1) introduces *Printable-Voxel*, a view-invariant representation that decouples density and color; (2) designs a differentiable neural network mapping RGB to the physical CMYKWCl color space, respecting ink absorption and spectral mixing; and (3) incorporates voxel-wise printability constraints during optimization. Built upon Plenoxel’s voxelized neural field, our framework directly outputs high-fidelity, full-color, printer-ready voxel grids. Experiments demonstrate seamless 4K-resolution printing on standard industrial printers without post-processing, achieving significant improvements in both color fidelity and geometric accuracy.
📝 Abstract
Recent advancements in 3D reconstruction, especially through neural rendering approaches like Neural Radiance Fields (NeRF) and Plenoxel, have led to high-quality 3D visualizations. However, these methods are optimized for digital environments and employ view-dependent color models (RGB) and 2D splatting techniques, which do not translate well to physical 3D printing. This paper introduces"Poxel", which stands for Printable-Voxel, a voxel-based 3D reconstruction framework optimized for photopolymer jetting 3D printing, which allows for high-resolution, full-color 3D models using a CMYKWCl color model. Our framework directly outputs printable voxel grids by removing view-dependency and converting the digital RGB color space to a physical CMYKWCl color space suitable for multi-material jetting. The proposed system achieves better fidelity and quality in printed models, aligning with the requirements of physical 3D objects.