Federated Sinkhorn

📅 2025-02-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the data silo problem among clients in federated learning by adapting entropy-regularized optimal transport (Sinkhorn) to a distributed setting—its first such application—enabling collaborative discrete OT across multiple clients. We propose several federated Sinkhorn variants (synchronous/asynchronous, fully connected/star topologies), unified under a distributed marginal-constrained optimization framework. Each variant achieves privacy preservation and resource efficiency via local Sinkhorn iterations coupled with periodic potential function exchanges. We establish theoretical convergence guarantees for all variants and, for the first time, quantify the computational–communication trade-off in relation to problem scale and marginal sparsity. Extensive experiments on synthetic data and a real-world financial risk assessment task validate efficacy; empirical analysis characterizes the Pareto frontier between communication cost and accuracy across network topologies.

Technology Category

Application Category

📝 Abstract
In this work we investigate the potential of solving the discrete Optimal Transport (OT) problem with entropy regularization in a federated learning setting. Recall that the celebrated Sinkhorn algorithm transforms the classical OT linear program into strongly convex constrained optimization, facilitating first order methods for otherwise intractably large problems. A common contemporary setting that remains an open problem as far as the application of Sinkhorn is the presence of data spread across clients with distributed inter-communication, either due to clients whose privacy is a concern, or simply by necessity of processing and memory hardware limitations. In this work we investigate various natural procedures, which we refer to as Federated Sinkhorn, that handle distributed environments where data is partitioned across multiple clients. We formulate the problem as minimizing the transport cost with an entropy regularization term, subject to marginal constraints, where block components of the source and target distribution vectors are locally known to clients corresponding to each block. We consider both synchronous and asynchronous variants as well as all-to-all and server-client communication topology protocols. Each procedure allows clients to compute local operations on their data partition while periodically exchanging information with others. We provide theoretical guarantees on convergence for the different variants under different possible conditions. We empirically demonstrate the algorithms performance on synthetic datasets and a real-world financial risk assessment application. The investigation highlights the subtle tradeoffs associated with computation and communication time in different settings and how they depend on problem size and sparsity.
Problem

Research questions and friction points this paper is trying to address.

Solving discrete Optimal Transport with entropy regularization
Handling data partitioned across multiple federated clients
Minimizing transport cost under marginal constraints
Innovation

Methods, ideas, or system contributions that make the work stand out.

Federated Sinkhorn for distributed OT
Handles data privacy and hardware limits
Convergence guarantees in varied conditions
🔎 Similar Papers
No similar papers found.
J
Jeremy Kulcsar
HSBC Quantum Technologies Group, Innovation & Ventures, HSBC, Hong Kong
Vyacheslav Kungurtsev
Vyacheslav Kungurtsev
Czech Technical University in Prague
Georgios Korpas
Georgios Korpas
HSBC and Czech Technical University in Prague
Applied MathematicsOptimizationArtificial IntelligenceQuantum Computing
G
Giulio Giaconi
HSBC Quantum Technologies Group, Innovation & Ventures, HSBC, United Kingdom
W
William Shoosmith
HSBC Quantum Technologies Group, Innovation & Ventures, HSBC, United Kingdom