Discrete Neural Algorithmic Reasoning

📅 2024-02-18
🏛️ arXiv.org
📈 Citations: 3
Influential: 0
📄 PDF
🤖 AI Summary
Neural algorithmic reasoning models exhibit poor out-of-distribution generalization, whereas classical algorithms guarantee distribution-agnostic determinism. To bridge this gap, we propose a discrete-continuous dual-stream architecture that enforces strict alignment between neural execution traces and the target algorithm’s finite-state machine: discrete state encoding and transition supervision model state evolution; a dual-stream interaction mechanism enables tight coordination between neural computation and formal semantics; and a verifiable framework supports formal behavioral proofs for arbitrary inputs. Our approach achieves 100% test accuracy on both single-task and multi-task algorithmic reasoning benchmarks—matching oracle performance—and, for the first time, enables end-to-end provable correctness. This significantly enhances neural models’ generalization to classical algorithms, while improving interpretability and reliability through formal guarantees.

Technology Category

Application Category

📝 Abstract
Neural algorithmic reasoning aims to capture computations with neural networks via learning the models to imitate the execution of classic algorithms. While common architectures are expressive enough to contain the correct model in the weights space, current neural reasoners are struggling to generalize well on out-of-distribution data. On the other hand, classic computations are not affected by distributional shifts as they can be described as transitions between discrete computational states. In this work, we propose to force neural reasoners to maintain the execution trajectory as a combination of finite predefined states. To achieve that, we separate discrete and continuous data flows and describe the interaction between them. Trained with supervision on the algorithm's state transitions, such models are able to perfectly align with the original algorithm. To show this, we evaluate our approach on multiple algorithmic problems and get perfect test scores both in single-task and multitask setups. Moreover, the proposed architectural choice allows us to prove the correctness of the learned algorithms for any test~data.
Problem

Research questions and friction points this paper is trying to address.

Improve neural algorithmic reasoning generalization on out-of-distribution data
Maintain algorithm execution trajectory through discrete computational states
Separate discrete and continuous data flows for correct algorithm alignment
Innovation

Methods, ideas, or system contributions that make the work stand out.

Combining discrete and continuous data flows
Using finite predefined states for execution
Supervising state transitions for algorithm alignment
🔎 Similar Papers
No similar papers found.
G
Gleb Rodionov
Yandex Research, Moscow, Russia
L
Liudmila Prokhorenkova
Yandex Research, Amsterdam, The Netherlands