🤖 AI Summary
This paper addresses the efficient approximation of high-dimensional irregular functions. We propose P1-KAN, a novel learnable network grounded in the Kolmogorov–Arnold representation theorem. Methodologically, it parameterizes univariate basis functions using piecewise polynomials, enabling structure-aware and theory-driven architecture design. Its key contributions are threefold: (i) the first smoothness-dependent theoretical error bound for KAN-type models; (ii) universal approximation capability without assuming function differentiability; and (iii) superior empirical performance—P1-KAN achieves state-of-the-art accuracy and convergence speed on irregular functions, outperforming MLPs and existing KAN variants; attains comparable accuracy to spline-based KAN on smooth functions while accelerating convergence by up to 40%; and delivers optimal performance on the French hydraulic valley optimization task.
📝 Abstract
A new Kolmogorov-Arnold network (KAN) is proposed to approximate potentially irregular functions in high dimensions. We provide error bounds for this approximation, assuming that the Kolmogorov-Arnold expansion functions are sufficiently smooth. When the function is only continuous, we also provide universal approximation theorems. We show that it outperforms multilayer perceptrons in terms of accuracy and convergence speed. We also compare it with several proposed KAN networks: it outperforms all networks for irregular functions and achieves similar accuracy to the original spline-based KAN network for smooth functions. Finally, we compare some of the KAN networks in optimizing a French hydraulic valley.