Activation functions provide the non-linearity to deep neural networks, which are crucial for the optimization and performance improvement. In this paper, we propose a learnable continuous piece-wise nonlinear activation function (or CPN in short), which improves the widely used ReLU from three directions, i.e., finer pieces, non-linear terms and learnable parameterization. CPN is a continuous activation function with multiple pieces and incorporates non-linear terms in every interval. We give a general formulation of CPN and provide different implementations according to three key factors: whether the activation space is divided uniformly or not, whether the non-linear terms exist or not, and whether the activation function is continuous or not. We demonstrate the effectiveness of our method on image classification and single image super-resolution tasks by simply changing the activation function. For example, CPN improves 4.78% / 4.52% top-1 accuracy over ReLU on MobileNetV2_0.25 / MobileNetV2_0.35 for ImageNet classification and achieves better PSNR on several benchmarks for super-resolution. Our implementation is available at https: //github.com/xc-G/CPN.

Gao, X., Li, Y., Li, W., Duan, L., Van Gool, L., Benini, L., et al. (2023). Learning continuous piecewise non-linear activation functions for deep neural networks. 345 E 47TH ST, NEW YORK, NY 10017 USA : IEEE [10.1109/ICME55011.2023.00315].

Learning continuous piecewise non-linear activation functions for deep neural networks

Benini, Luca;Magno, Michele
2023

Abstract

Activation functions provide the non-linearity to deep neural networks, which are crucial for the optimization and performance improvement. In this paper, we propose a learnable continuous piece-wise nonlinear activation function (or CPN in short), which improves the widely used ReLU from three directions, i.e., finer pieces, non-linear terms and learnable parameterization. CPN is a continuous activation function with multiple pieces and incorporates non-linear terms in every interval. We give a general formulation of CPN and provide different implementations according to three key factors: whether the activation space is divided uniformly or not, whether the non-linear terms exist or not, and whether the activation function is continuous or not. We demonstrate the effectiveness of our method on image classification and single image super-resolution tasks by simply changing the activation function. For example, CPN improves 4.78% / 4.52% top-1 accuracy over ReLU on MobileNetV2_0.25 / MobileNetV2_0.35 for ImageNet classification and achieves better PSNR on several benchmarks for super-resolution. Our implementation is available at https: //github.com/xc-G/CPN.
2023
2023 IEEE International Conference on Multimedia and Expo (ICME)
1835
1840
Gao, X., Li, Y., Li, W., Duan, L., Van Gool, L., Benini, L., et al. (2023). Learning continuous piecewise non-linear activation functions for deep neural networks. 345 E 47TH ST, NEW YORK, NY 10017 USA : IEEE [10.1109/ICME55011.2023.00315].
Gao, Xinchen; Li, Yawei; Li, Wen; Duan, Lixin; Van Gool, Luc; Benini, Luca; Magno, Michele
File in questo prodotto:
Eventuali allegati, non sono esposti

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/958755
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? 0
social impact