🤖 AI Summary
Existing evaluation methods conflate multiple reasoning types, making it difficult to isolate and study analogical reasoning mechanisms. This work addresses this gap through theoretical analysis and large-scale experiments (up to 1.5B parameters), providing the first evidence that Transformers perform analogical reasoning via representational alignment. The study further reveals the critical influence of training order and data structure on reasoning capabilities. Key contributions include the proposal of a representational alignment mechanism, the demonstration of the necessity of curriculum learning for effective analogical reasoning, and the unification of two-hop reasoning as a form of analogical reasoning—thereby elucidating the underlying principles of multi-hop reasoning in Transformer models.
📝 Abstract
Understanding reasoning in large language models is complicated by evaluations that conflate multiple reasoning types. We isolate analogical reasoning (inferring shared properties between entities based on known similarities) and analyze its emergence in transformers. We theoretically prove three key results: (1) Joint training on similarity and attribution premises enables analogical reasoning through aligned representations; (2) Sequential training succeeds only when similarity structure is learned before specific attributes, revealing a necessary curriculum; (3) Two-hop reasoning ($a \to b, b \to c \implies a \to c$) reduces to analogical reasoning with identity bridges ($b = b$), which must appear explicitly in training data. These results reveal a unified mechanism: transformers encode entities with similar properties into similar representations, enabling property transfer through feature alignment. Experiments with architectures up to 1.5B parameters validate our theory and demonstrate how representational geometry shapes inductive reasoning capabilities.