Federated Learning (FL) is gaining momentum as a promising solution to enable the efficient and privacy-preserving distributed training of Machine Learning (ML) models. Unlike centralized ML solutions, only the ML model and its updates are transferred between the clients and the aggregator server, eliminating the need to share large datasets. Notwithstanding, poor connectivity conditions experienced over the path that interconnects the FL clients and the aggregator server, either due to (wireless) channel losses or congestion, may deteriorate the training convergence. Several methods have been devised to reduce the training duration, primarily by minimizing data transfer through the design of ML algorithms at the application level. However, these solutions still exhibit unsettled issues, as they may only reduce the communication footprint but do not improve the communication process as a whole. Differently, in this work, our aim is to improve FL data exchange from a networking perspective by promoting Information Centric Networking (ICN) approaches rather than host-centric TCP/IP-based solutions. To this aim, we analyze the impact that host-centric transport protocols as well as ICN approaches have on the FL performance, in terms of duration of the model training and exchanged data (model and updates) load, under different channel loss settings. We show that ICN-based FL solutions significantly reduce the network data load and decrease the duration of the training round by up to an order of magnitude for high channel loss rates.
Improving communication performance of Federated Learning: A networking perspective
Amadeo, Marica;Molinaro, Antonella
2025-01-01
Abstract
Federated Learning (FL) is gaining momentum as a promising solution to enable the efficient and privacy-preserving distributed training of Machine Learning (ML) models. Unlike centralized ML solutions, only the ML model and its updates are transferred between the clients and the aggregator server, eliminating the need to share large datasets. Notwithstanding, poor connectivity conditions experienced over the path that interconnects the FL clients and the aggregator server, either due to (wireless) channel losses or congestion, may deteriorate the training convergence. Several methods have been devised to reduce the training duration, primarily by minimizing data transfer through the design of ML algorithms at the application level. However, these solutions still exhibit unsettled issues, as they may only reduce the communication footprint but do not improve the communication process as a whole. Differently, in this work, our aim is to improve FL data exchange from a networking perspective by promoting Information Centric Networking (ICN) approaches rather than host-centric TCP/IP-based solutions. To this aim, we analyze the impact that host-centric transport protocols as well as ICN approaches have on the FL performance, in terms of duration of the model training and exchanged data (model and updates) load, under different channel loss settings. We show that ICN-based FL solutions significantly reduce the network data load and decrease the duration of the training round by up to an order of magnitude for high channel loss rates.Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


