🤖 AI Summary
This work addresses the challenge that existing automatic code generation methods often produce structurally invalid or physically inconsistent models, which are unsuitable for engineering simulation. To ensure physical consistency and simulatability, the authors propose a procedural modeling framework that integrates domain knowledge injection, constraint-guided fine-tuning, and closed-loop simulation validation. Key contributions include CivilInstruct—the first instruction-following dataset tailored for structural engineering—along with a two-stage fine-tuning strategy and MBEval, a validation-driven evaluation benchmark. Experimental results demonstrate that the proposed approach significantly outperforms baseline methods across multiple rigorous metrics, effectively suppressing hallucinations and constraint violations, and enabling the direct use of generated models in structural dynamics simulations.
📝 Abstract
Structural modeling is a fundamental component of computational engineering science, in which even minor physical inconsistencies or specification violations may invalidate downstream simulations. The potential of large language models (LLMs) for automatic generation of modeling code has been demonstrated. However, non-executable or physically inconsistent outputs remain prevalent under stringent engineering constraints. A framework for physics-consistent automatic building modeling is therefore proposed, integrating domain knowledge construction, constraint-oriented model alignment, and verification-driven evaluation. CivilInstruct is introduced as a domain-specific dataset that formalizes structural engineering knowledge and constraint reasoning to enable simulation-ready model generation. A two-stage fine-tuning strategy is further employed to enforce constraint satisfaction and application programming interface compliance, substantially reducing hallucinated and non-conforming outputs. MBEval is presented as a verification-driven benchmark that evaluates executability and structural dynamics consistency through closed-loop validation. Experimental results show consistent improvements over baselines across rigorous verification metrics. Our code is available at https://github.com/Jovanqing/AutoBM.