29
45

SwiftAgg+: Achieving Asymptotically Optimal Communication Loads in Secure Aggregation for Federated Learning

Abstract

We propose SwiftAgg+, a novel secure aggregation protocol for federated learning systems, where a central server aggregates local models of NNN \in \mathbb{N} distributed users, each of size LNL \in \mathbb{N}, trained on their local data, in a privacy-preserving manner. SwiftAgg+ can significantly reduce the communication overheads without any compromise on security, and achieve optimal communication loads within diminishing gaps. Specifically, in presence of at most D=o(N)D=o(N) dropout users, SwiftAgg+ achieves a per-user communication load of (1+O(1N))L(1+\mathcal{O}(\frac{1}{N}))L symbols and a server communication load of (1+O(1N))L(1+\mathcal{O}(\frac{1}{N}))L symbols, with a worst-case information-theoretic security guarantee, against any subset of up to T=o(N)T=o(N) semi-honest users who may also collude with the curious server. Moreover, the proposed SwiftAgg+ allows for a flexible trade-off between communication loads and the number of active communication links. In particular, for T<NDT<N-D and for any KNK\in\mathbb{N}, SwiftAgg+ can achieve the server communication load of (1+TK)L(1+\frac{T}{K})L symbols, and per-user communication load of up to (1+T+DK)L(1+\frac{T+D}{K})L symbols, where the number of pair-wise active connections in the network is N2(K+T+D+1)\frac{N}{2}(K+T+D+1).

View on arXiv
Comments on this paper