Federated learning allows each client to keep its data locally when training machine learning models in a distributed setting. Significant recent research established the requirements that the input must satisfy in order to guarantee convergence of the training loop. This line of work uses averaging as the aggregation rule for the training models. In particular, we are interested in whether federated learning is robust to Byzantine behavior, and observe and investigate a tradeoff between the average/centroid and the validity conditions from distributed computing. We show that the various validity conditions alone do not guarantee a good approximation of the average. Furthermore, we show that reaching good approximation does not give good results in experimental settings due to possible Byzantine outliers. Our main contribution is the first lower bound of on the centroid approximation under box validity that is often considered in the literature, where is the number of clients, the upper bound on the number of Byzantine faults, and is the dimension of the machine learning model. We complement this lower bound by an upper bound of , by providing a new analysis for the case . In addition, we present a new algorithm that achieves a -approximation under convex validity, which also proves that the existing lower bound in the literature is tight. We show that all presented bounds can also be achieved in the distributed peer-to-peer setting. We complement our analytical results with empirical evaluations in federated stochastic gradient descent and federated averaging settings.
View on arXiv@article{cambus2025_2506.15264, title={ Centroid Approximation for Byzantine-Tolerant Federated Learning }, author={ Mélanie Cambus and Darya Melnyk and Tijana Milentijević and Stefan Schmid }, journal={arXiv preprint arXiv:2506.15264}, year={ 2025 } }