TY - JOUR
T1 - ResFed
T2 - Communication0Efficient Federated Learning with Deep Compressed Residuals
AU - Song, Rui
AU - Zhou, Liguo
AU - Lyu, Lingjuan
AU - Festag, Andreas
AU - Knoll, Alois
N1 - Publisher Copyright:
© 2014 IEEE.
PY - 2024/3/15
Y1 - 2024/3/15
N2 - Federated learning allows for cooperative training among distributed clients by sharing their locally learned model parameters, such as weights or gradients. However, as model size increases, the communication bandwidth required for deployment in wireless networks becomes a bottleneck. To address this, we propose a residual-based federated learning framework (ResFed) that transmits residuals instead of gradients or weights in networks. By predicting model updates at both clients and the server, residuals are calculated as the difference between updated and predicted models and contain more dense information than weights or gradients. We find that the residuals are less sensitive to an increasing compression ratio than other parameters, and hence use lossy compression techniques on residuals to improve communication efficiency for training in federated settings. With the same compression ratio, ResFed outperforms current methods (weight- or gradient-based federated learning) by over 1.4× on federated data sets, including MNIST, FashionMNIST, SVHN, CIFAR-10, CIFAR-100, and FEMNIST, in client-to-server communication, and can also be applied to reduce communication costs for server-to-client communication.
AB - Federated learning allows for cooperative training among distributed clients by sharing their locally learned model parameters, such as weights or gradients. However, as model size increases, the communication bandwidth required for deployment in wireless networks becomes a bottleneck. To address this, we propose a residual-based federated learning framework (ResFed) that transmits residuals instead of gradients or weights in networks. By predicting model updates at both clients and the server, residuals are calculated as the difference between updated and predicted models and contain more dense information than weights or gradients. We find that the residuals are less sensitive to an increasing compression ratio than other parameters, and hence use lossy compression techniques on residuals to improve communication efficiency for training in federated settings. With the same compression ratio, ResFed outperforms current methods (weight- or gradient-based federated learning) by over 1.4× on federated data sets, including MNIST, FashionMNIST, SVHN, CIFAR-10, CIFAR-100, and FEMNIST, in client-to-server communication, and can also be applied to reduce communication costs for server-to-client communication.
KW - Communication efficiency
KW - deep compression
KW - federated learning
KW - protocol design
UR - http://www.scopus.com/inward/record.url?scp=85174854756&partnerID=8YFLogxK
U2 - 10.1109/JIOT.2023.3324079
DO - 10.1109/JIOT.2023.3324079
M3 - Article
AN - SCOPUS:85174854756
SN - 2327-4662
VL - 11
SP - 9458
EP - 9472
JO - IEEE Internet of Things Journal
JF - IEEE Internet of Things Journal
IS - 6
ER -