🤖 AI Summary
Existing neural-symbolic approaches face limitations in modeling uncertainty, handling real-world scenarios, and reasoning over complex relational structures—such as causal, hierarchical, or multi-hop dependencies—especially when scaling to large-scale problems. To address these challenges, we propose an end-to-end differentiable relational inductive bias module that jointly integrates relation-aware attention, differentiable graph learning, and symbolic logic embedding. This enables dynamic graph construction and cross-relation path reasoning, overcoming the implicit relational discovery bottlenecks inherent in standard GNNs and Transformers. We further introduce contrastive relational representation learning and a hierarchical relational aggregation architecture. Our method achieves state-of-the-art performance on CLEVR, Sort-of-CLEVR, and Raven benchmarks, improving generalization ability by 23% and sample efficiency by 31% over prior approaches.