Neural Operators Can Discover Functional Clusters

๐Ÿ“… 2026-02-26
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This work addresses the challenge of unsupervised clustering in infinite-dimensional function spaces under non-convex and disconnected cluster structures. We propose a general neural operatorโ€“based clustering framework that maps discrete trajectories into continuous features in a reproducing kernel Hilbert space via a pretrained encoder, followed by a lightweight trainable head to produce soft cluster assignments. Theoretically, we establish, for the first time, that neural operators can approximate any finite collection of closed sets in this space to arbitrary precision, thereby formulating a universal clustering theory in the upper Kuratowski topology. Experimentally, our method successfully recovers latent dynamical structures across multiple synthetic ordinary differential equation trajectory benchmarks, significantly outperforming classical approaches and demonstrating both effectiveness and practical utility.

Technology Category

Application Category

๐Ÿ“ Abstract
Operator learning is reshaping scientific computing by amortizing inference across infinite families of problems. While neural operators (NOs) are increasingly well understood for regression, far less is known for classification and its unsupervised analogue: clustering. We prove that sample-based neural operators can learn any finite collection of classes in an infinite-dimensional reproducing kernel Hilbert space, even when the classes are neither convex nor connected, under mild kernel sampling assumptions. Our universal clustering theorem shows that any $K$ closed classes can be approximated to arbitrary precision by NO-parameterized classes in the upper Kuratowski topology on closed sets, a notion that can be interpreted as disallowing false-positive misclassifications. Building on this, we develop an NO-powered clustering pipeline for functional data and apply it to unlabeled families of ordinary differential equation (ODE) trajectories. Discretized trajectories are lifted by a fixed pre-trained encoder into a continuous feature map and mapped to soft assignments by a lightweight trainable head. Experiments on diverse synthetic ODE benchmarks show that the resulting practical SNO recovers latent dynamical structure in regimes where classical methods fail, providing evidence consistent with our universal clustering theory.
Problem

Research questions and friction points this paper is trying to address.

neural operators
functional clustering
unsupervised learning
infinite-dimensional data
ODE trajectories
Innovation

Methods, ideas, or system contributions that make the work stand out.

neural operators
functional clustering
universal approximation
Kuratowski topology
ODE trajectories
๐Ÿ”Ž Similar Papers
No similar papers found.
Yicen Li
Yicen Li
PhD Student, McMaster University
deep clustering
Jose Antonio Lara Benitez
Jose Antonio Lara Benitez
Applied Mathematics, Rice University
Applied MathematicsComputational PhysicsMachine Learning
R
Ruiyang Hong
McMaster University and the Vector Institute, Main St., Hamilton, L8S 4L8, Ontario, Canada
Anastasis Kratsios
Anastasis Kratsios
McMaster University and Vector Institute
Mathematics of AIGeometric Deep LearningApproximation TheoryLearning TheoryFinance
P
Paul David McNicholas
McMaster University and the Vector Institute, Main St., Hamilton, L8S 4L8, Ontario, Canada
M
Maarten Valentijn de Hoop
Rice University, Main St., Houston, 77005, Texas, United States