Binary Neural Networks (BNNs) use 1-bit weights and acti- vations to e ciently execute deep convolutional neural networks on edge devices. Nevertheless, the binarization of the rst layer is conventionally excluded, as it leads to a large accuracy loss. The few works address- ing the rst layer binarization, typically increase the number of input channels to enhance data representation; such data expansion raises the amount of operations needed and it is feasible only on systems with enough computational resources. In this work, we present a new method to binarize the rst layer using directly the 8-bit representation of input data; we exploit the standard bit-planes encoding to extract features bit-wise (using depth-wise convolutions); after a re-weighting stage, fea- tures are fused again. The resulting model is fully binarized and our rst layer binarization approach is model independent. The concept is evalu- ated on three classi cation datasets (CIFAR10, SVHN and CIFAR100) for di erent model architectures (VGG and ResNet) and, the proposed technique outperforms state of the art methods both in accuracy and BMACs reduction.

Vorabbi L., Maltoni D., Santi S. (2023). Input Layer Binarization with Bit-Plane Encoding [10.1007/978-3-031-44198-1_33].

Input Layer Binarization with Bit-Plane Encoding

Vorabbi L.
Primo
;
Maltoni D.
Secondo
;
2023

Abstract

Binary Neural Networks (BNNs) use 1-bit weights and acti- vations to e ciently execute deep convolutional neural networks on edge devices. Nevertheless, the binarization of the rst layer is conventionally excluded, as it leads to a large accuracy loss. The few works address- ing the rst layer binarization, typically increase the number of input channels to enhance data representation; such data expansion raises the amount of operations needed and it is feasible only on systems with enough computational resources. In this work, we present a new method to binarize the rst layer using directly the 8-bit representation of input data; we exploit the standard bit-planes encoding to extract features bit-wise (using depth-wise convolutions); after a re-weighting stage, fea- tures are fused again. The resulting model is fully binarized and our rst layer binarization approach is model independent. The concept is evalu- ated on three classi cation datasets (CIFAR10, SVHN and CIFAR100) for di erent model architectures (VGG and ResNet) and, the proposed technique outperforms state of the art methods both in accuracy and BMACs reduction.
2023
Artificial Neural Networks and Machine Learning – ICANN 2023
395
406
Vorabbi L., Maltoni D., Santi S. (2023). Input Layer Binarization with Bit-Plane Encoding [10.1007/978-3-031-44198-1_33].
Vorabbi L.; Maltoni D.; Santi S.
File in questo prodotto:
File Dimensione Formato  
Input Layer Binarization with Bit-Plane Encoding.pdf

Open Access dal 23/09/2024

Tipo: Postprint
Licenza: Licenza per accesso libero gratuito
Dimensione 1.23 MB
Formato Adobe PDF
1.23 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/952173
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? ND
social impact