LLM Flow Processes for Text-Conditioned Regression

๐Ÿ“… 2026-01-05
๐Ÿ›๏ธ arXiv.org
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This work proposes a general framework that addresses the limitations of existing meta-learning regression methods, which struggle to effectively integrate expert priors with textual metadata, and the constrained regression performance of large language models (LLMs) despite their rich semantic capabilities. The approach uniquely combines the text-conditional probability outputs of LLMs with neural diffusion or flow-matching processes through a product-of-experts mechanism, enabling joint sampling between a binning-based probabilistic density โ€œexpertโ€ model and a diffusion generative model. By doing so, the method preserves the semantic understanding of LLMs while significantly enhancing regression accuracy. Empirical evaluations across multiple benchmark tasks demonstrate consistent superiority over both standalone LLMs and neural processโ€“based approaches, validating the effective contribution of text-conditioned knowledge to regression performance.

Technology Category

Application Category

๐Ÿ“ Abstract
Meta-learning methods for regression like Neural (Diffusion) Processes achieve impressive results, but with these models it can be difficult to incorporate expert prior knowledge and information contained in metadata. Large Language Models (LLMs) are trained on giant corpora including varied real-world regression datasets alongside their descriptions and metadata, leading to impressive performance on a range of downstream tasks. Recent work has extended this to regression tasks and is able to leverage such prior knowledge and metadata, achieving surprisingly good performance, but this still rarely matches dedicated meta-learning methods. Here we introduce a general method for sampling from a product-of-experts of a diffusion or flow matching model and an `expert'with binned probability density; we apply this to combine neural diffusion processes with LLM token probabilities for regression (which may incorporate textual knowledge), exceeding the empirical performance of either alone.
Problem

Research questions and friction points this paper is trying to address.

text-conditioned regression
large language models
meta-learning
prior knowledge
neural processes
Innovation

Methods, ideas, or system contributions that make the work stand out.

product-of-experts
diffusion processes
flow matching
large language models
text-conditioned regression
๐Ÿ”Ž Similar Papers
No similar papers found.