Byzantine-Robust and Communication-Efficient Distributed Training: Compressive and Cyclic Gradient Coding
arXiv:2603.28780v1 Announce Type: new Abstract: In this paper, we study the problem of distributed training (DT) under Byzantine attacks with communication constraints. While prior work has developed various robust aggregation rules at the server to enhance robustness to Byzantine attacks, the existing methods suffer from a critical limitation in that the solution error does not diminish when the local gradients sent by different devices vary considerably, as a result of data heterogeneity among the subsets held by different […]