Decentralized Quantile Regression for Feature-Distributed Massive Datasets with Privacy Guarantees

πŸ“… 2025-04-23
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This paper addresses privacy-preserving quantile regression in large-scale, feature-distributed settings over decentralized networks. Method: We propose DSG-cqrβ€”the first decentralized surrogate gradient algorithm for conditional quantile regression that avoids conjugate gradient computation. It integrates convolution-type smoothing approximation with the Gaussian mechanism to achieve (Ξ΅,Ξ΄)-differential privacy (with Ξ΅ ≀ 1) under feature partitioning, without requiring raw data sharing or global coordination. Auxiliary variables are introduced for residual estimation, and Wald statistics are employed to construct confidence intervals. Contribution/Results: We establish linear convergence of DSG-cqr to statistical accuracy. Empirical evaluations demonstrate its superior performance in estimation precision, differential privacy guarantees, and communication efficiency compared to existing approaches.

Technology Category

Application Category

πŸ“ Abstract
In this paper, we introduce a novel decentralized surrogate gradient-based algorithm for quantile regression in a feature-distributed setting, where global features are dispersed across multiple machines within a decentralized network. The proposed algorithm, exttt{DSG-cqr}, utilizes a convolution-type smoothing approach to address the non-smooth nature of the quantile loss function. exttt{DSG-cqr} is fully decentralized, conjugate-free, easy to implement, and achieves linear convergence up to statistical precision. To ensure privacy, we adopt the Gaussian mechanism to provide $(epsilon,delta)$-differential privacy. To overcome the exact residual calculation problem, we estimate residuals using auxiliary variables and develop a confidence interval construction method based on Wald statistics. Theoretical properties are established, and the practical utility of the methods is also demonstrated through extensive simulations and a real-world data application.
Problem

Research questions and friction points this paper is trying to address.

Decentralized quantile regression for feature-distributed massive datasets
Privacy-preserving algorithm with Gaussian differential privacy guarantees
Convolution smoothing for non-smooth quantile loss optimization
Innovation

Methods, ideas, or system contributions that make the work stand out.

Decentralized surrogate gradient algorithm for quantile regression
Convolution-type smoothing for non-smooth quantile loss
Gaussian mechanism for differential privacy guarantees
πŸ”Ž Similar Papers
No similar papers found.
P
Peiwen Xiao
Key Laboratory of Data Science in Finance and Economics, and School of Statistics and Data Science, Jiangxi University of Finance and Economics, Nanchang, Jiangxi, China
X
Xiaohui Liu
Key Laboratory of Data Science in Finance and Economics, and School of Statistics and Data Science, Jiangxi University of Finance and Economics, Nanchang, Jiangxi, China
Guangming Pan
Guangming Pan
Nanyang Technological University
random matrixinformation theorystatistics
W
Wei Long
Department of Economics, Tulane University, New Orleans, Louisiana, United States