Private and Federated Stochastic Convex Optimization: Efficient Strategies for Centralized Systems
0citations
PDF0
Citations
#10
in ICML 2024
of 2635 papers
2
Authors
1
Data Points
Authors
Topics
Abstract
This paper addresses the challenge of preserving privacy in Federated Learning (FL) within centralized systems, focusing on both trusted and untrusted server scenarios. We analyze this setting within the Stochastic Convex Optimization (SCO) framework, and devise methods that ensure Differential Privacy (DP) while maintaining optimal convergence rates for homogeneous and heterogeneous data distributions. Our approach, based on a recent stochastic optimization technique, offers linear computational complexity, comparable to non-private FL methods, and reduced gradient obfuscation. This work enhances the practicality of DP in FL, balancing privacy, efficiency, and robustness in a variety of server trust environments.
Citation History
Jan 28, 2026
0