🤖 AI Summary
This work addresses the challenge of balancing platform mobility and deposition accuracy in mobile 3D printing over unstructured terrain. The authors propose a perception–learning–execution closed-loop system that embeds AI-driven disturbance prediction within a three-layer hierarchical control architecture, shifting the paradigm from passive correction to proactive compensation. By fusing data from IMU, visual, and depth sensors, the system enables co-optimization of path planning, predictive control of the chassis–manipulator coordination, and high-precision material deposition. Experimental results demonstrate that the approach achieves sub-centimeter printing accuracy on outdoor terrains featuring slopes and uneven surfaces, while preserving full all-terrain mobility of the robotic platform.
📝 Abstract
Mobile 3D printing on unstructured terrain remains challenging due to the conflict between platform mobility and deposition precision. Existing gantry-based systems achieve high accuracy but lack mobility, while mobile platforms struggle to maintain print quality on uneven ground. We present a framework that tightly integrates AI-driven disturbance prediction with multi-modal sensor fusion and hierarchical hardware control, forming a closed-loop perception-learning-actuation system. The AI module learns terrain-to-perturbation mappings from IMU, vision, and depth sensors, enabling proactive compensation rather than reactive correction. This intelligence is embedded into a three-layer control architecture: path planning, predictive chassis-manipulator coordination, and precision hardware execution. Through outdoor experiments on terrain with slopes and surface irregularities, we demonstrate sub-centimeter printing accuracy while maintaining full platform mobility. This AI-hardware integration establishes a practical foundation for autonomous construction in unstructured environments.