In Cross-device Federated Learning, communication efficiency is of paramount importance. Sparse Ternary Compression (STC) is one of the most effective techniques for considerably reducing the per-round communication cost of Federated Learning (FL) without significantly degrading the accuracy of the global model, by using ternary quantization in series to topk sparsification. In this paper, we propose an original variant of STC that is specifically designed and implemented for convolutional layers. Our variant is originally based on the experimental evidence that a pattern exists in the distribution of client updates, namely, the difference between the received global model and the locally trained model. In particular, we have experimentally found that the largest (in absolute value) updates for convolutional layers tend to form clusters in a kernel-wise fashion. Therefore, our primary novel idea is to a-priori restrict the elements of STC updates to lay on such a structured pattern, thus allowing us to further reduce the STC communication cost. We have designed, implemented, and evaluated our novel technique, called Structured Sparse Ternary Compression (SSTC). Reported experimental results show that SSTC shrinks compressed updates by a factor of x3 with respect to traditional STC and with a reduction up to x104 with respect to uncompressed FedAvg, at the expense of negligible degradation of the global model accuracy.

Structured Sparse Ternary Compression for Convolutional Layers in Federated Learning / Mora, A; Foschini, L; Bellavista, P. - ELETTRONICO. - 2022:(2022), pp. 1-5. (Intervento presentato al convegno 95th IEEE Vehicular Technology Conference - Spring, VTC 2022 - tenutosi a Helsinki, Finland nel 19-22 June 2022) [10.1109/VTC2022-Spring54318.2022.9860833].

Structured Sparse Ternary Compression for Convolutional Layers in Federated Learning

Mora, A;Foschini, L;Bellavista, P
2022

Abstract

In Cross-device Federated Learning, communication efficiency is of paramount importance. Sparse Ternary Compression (STC) is one of the most effective techniques for considerably reducing the per-round communication cost of Federated Learning (FL) without significantly degrading the accuracy of the global model, by using ternary quantization in series to topk sparsification. In this paper, we propose an original variant of STC that is specifically designed and implemented for convolutional layers. Our variant is originally based on the experimental evidence that a pattern exists in the distribution of client updates, namely, the difference between the received global model and the locally trained model. In particular, we have experimentally found that the largest (in absolute value) updates for convolutional layers tend to form clusters in a kernel-wise fashion. Therefore, our primary novel idea is to a-priori restrict the elements of STC updates to lay on such a structured pattern, thus allowing us to further reduce the STC communication cost. We have designed, implemented, and evaluated our novel technique, called Structured Sparse Ternary Compression (SSTC). Reported experimental results show that SSTC shrinks compressed updates by a factor of x3 with respect to traditional STC and with a reduction up to x104 with respect to uncompressed FedAvg, at the expense of negligible degradation of the global model accuracy.
2022
2022 IEEE 95th Vehicular Technology Conference: (VTC2022-Spring)
1
5
Structured Sparse Ternary Compression for Convolutional Layers in Federated Learning / Mora, A; Foschini, L; Bellavista, P. - ELETTRONICO. - 2022:(2022), pp. 1-5. (Intervento presentato al convegno 95th IEEE Vehicular Technology Conference - Spring, VTC 2022 - tenutosi a Helsinki, Finland nel 19-22 June 2022) [10.1109/VTC2022-Spring54318.2022.9860833].
Mora, A; Foschini, L; Bellavista, P
File in questo prodotto:
File Dimensione Formato  
SSTC_IEEE_VTC_Edge_AI_am.pdf

accesso aperto

Tipo: Postprint
Licenza: Licenza per accesso libero gratuito
Dimensione 896.14 kB
Formato Adobe PDF
896.14 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/899393
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? 0
social impact