On Transferring Transferability: Towards a Theory for Size Generalization

📅 2025-05-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Non-Euclidean structures—such as graphs, sets, and point clouds—exhibit poor generalization across varying input sizes, posing a fundamental challenge for learning invariant representations. Method: This work establishes the first rigorous theoretical framework for cross-dimensional transferability, formalizing size transferability as continuity in a limiting space and constructing dimension-agnostic generalization models via data- and task-driven equivalence identification. We propose design principles for transferable architectures and apply them to adapt mainstream graph neural networks as well as set- and point-cloud models. Contribution/Results: We derive necessary and sufficient conditions for transferability and prove their theoretical validity. Extensive experiments on multiple benchmarks demonstrate that the adapted models achieve significantly improved generalization performance on out-of-distribution size tasks, strongly corroborating the theoretical analysis.

Technology Category

Application Category

📝 Abstract
Many modern learning tasks require models that can take inputs of varying sizes. Consequently, dimension-independent architectures have been proposed for domains where the inputs are graphs, sets, and point clouds. Recent work on graph neural networks has explored whether a model trained on low-dimensional data can transfer its performance to higher-dimensional inputs. We extend this body of work by introducing a general framework for transferability across dimensions. We show that transferability corresponds precisely to continuity in a limit space formed by identifying small problem instances with equivalent large ones. This identification is driven by the data and the learning task. We instantiate our framework on existing architectures, and implement the necessary changes to ensure their transferability. Finally, we provide design principles for designing new transferable models. Numerical experiments support our findings.
Problem

Research questions and friction points this paper is trying to address.

Develops theory for size generalization in learning models
Explores transferability across varying input dimensions
Provides design principles for transferable model architectures
Innovation

Methods, ideas, or system contributions that make the work stand out.

General framework for dimension transferability
Continuity in limit space ensures transferability
Design principles for transferable models
Eitan Levin
Eitan Levin
California Institute of Technology
OptimizationGeometrySymmetryData science
Y
Yuxin Ma
Department of Applied Mathematics and Statistics and the Mathematical Institute for Data Science, Johns Hopkins University, Baltimore, MD 21218, USA
M
Mateo D'iaz
Department of Applied Mathematics and Statistics and the Mathematical Institute for Data Science, Johns Hopkins University, Baltimore, MD 21218, USA
Soledad Villar
Soledad Villar
Johns Hopkins University
mathematics of datageometric deep learningcomputational harmonic analysis