PICASO: Permutation-Invariant Context Composition with State Space Models

📅 2025-02-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large language models (LLMs) benefit from multi-segment context in reasoning, yet long-context processing incurs substantial computational overhead; existing state space models (SSMs) struggle to effectively integrate states from multiple independent contexts. This paper proposes a novel SSM-based framework for permutation-invariant multi-context state mapping and aggregation: it introduces the first commutative, order-agnostic multi-state fusion mechanism—mathematically approximating text concatenation—and enforces permutation invariance by averaging over all context permutations. Furthermore, we design a dynamic weighting algorithm to adaptively calibrate contributions based on contextual importance. Evaluated on WikiText and MSMARCO, our method matches or exceeds the strongest baselines under both zero-shot and fine-tuned settings, while achieving an average 5.4× speedup in inference latency—significantly alleviating the computational bottleneck of multi-context integration.

Technology Category

Application Category

📝 Abstract
Providing Large Language Models with relevant contextual knowledge at inference time has been shown to greatly improve the quality of their generations. This is often achieved by prepending informative passages of text, or 'contexts', retrieved from external knowledge bases to their input. However, processing additional contexts online incurs significant computation costs that scale with their length. State Space Models (SSMs) offer a promising solution by allowing a database of contexts to be mapped onto fixed-dimensional states from which to start the generation. A key challenge arises when attempting to leverage information present across multiple contexts, since there is no straightforward way to condition generation on multiple independent states in existing SSMs. To address this, we leverage a simple mathematical relation derived from SSM dynamics to compose multiple states into one that efficiently approximates the effect of concatenating textual contexts. Since the temporal ordering of contexts can often be uninformative, we enforce permutation-invariance by efficiently averaging states obtained via our composition algorithm across all possible context orderings. We evaluate our resulting method on WikiText and MSMARCO in both zero-shot and fine-tuned settings, and show that we can match the strongest performing baseline while enjoying on average 5.4x speedup.
Problem

Research questions and friction points this paper is trying to address.

Enhancing LLMs with contextual knowledge efficiently
Composing multiple states in SSMs for better generation
Achieving permutation-invariance in context composition
Innovation

Methods, ideas, or system contributions that make the work stand out.

State Space Models composition
Permutation-invariant context averaging
Efficient online context processing
🔎 Similar Papers
No similar papers found.
Tian Yu Liu
Tian Yu Liu
University of California Los Angeles
Machine LearningComputer Vision
A
A. Achille
AWS AI Labs
Matthew Trager
Matthew Trager
AWS AI Labs
Computer VisionMachine LearningApplied Algebraic Geometry
A
Aditya Golatkar
AWS AI Labs
L
L. Zancato
AWS AI Labs
S
S. Soatto
AWS AI Labs