A Dual Optimization View to Empirical Risk Minimization with f-Divergence Regularization

📅 2025-08-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In joint optimization of empirical risk minimization (ERM) and $f$-divergence regularization ($f$-DR), the normalization function lacks a closed-form solution, hindering scalability and efficiency. Method: We propose a dual-theoretic framework grounded in Legendre–Fenchel duality and the implicit function theorem, rigorously linking the primal and dual problems. This analysis yields a nonlinear ordinary differential equation (ODE) characterizing the normalization function, which admits efficient numerical integration under mild regularity conditions. Contribution/Results: The method enables rapid, accurate computation of the normalization function, significantly enhancing the scalability and computational efficiency of ERM-$f$-DR models. It establishes a novel theoretical pathway for $f$-divergence-regularized learning and provides a practical numerical tool. Empirical and theoretical validation confirms the effectiveness and broad applicability of dual analysis in statistical learning.

Technology Category

Application Category

📝 Abstract
The dual formulation of empirical risk minimization with f-divergence regularization (ERM-fDR) is introduced. The solution of the dual optimization problem to the ERM-fDR is connected to the notion of normalization function introduced as an implicit function. This dual approach leverages the Legendre-Fenchel transform and the implicit function theorem to provide a nonlinear ODE expression to the normalization function. Furthermore, the nonlinear ODE expression and its properties provide a computationally efficient method to calculate the normalization function of the ERM-fDR solution under a mild condition.
Problem

Research questions and friction points this paper is trying to address.

Dual optimization for empirical risk minimization with f-divergence regularization
Connecting dual solution to implicit normalization function via ODE
Efficient computation of normalization function under mild conditions
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dual optimization with f-divergence regularization
Legendre-Fenchel transform for nonlinear ODE
Implicit function theorem for efficient computation