Dynamic Vision Sensors (DVS) offer the opportunity to scale the energy consumption in image acquisition proportionally to the activity in the captured scene by only transmitting data when the captured image changes. Their potential for energy-proportional sensing makes them highly attractive for severely energy-constrained sensing nodes at the edge. Most approaches to the processing of DVS data employ Spiking Neural Networks to classify the input from the sensor. In this paper, we propose an alternative, event frame-based approach to the classification of DVS video data. We assemble ternary video frames from the event stream and process them with a fully ternarized Temporal Convolutional Network which can be mapped to CUTIE, a highly energy-efficient Ternary Neural Network accelerator. The network mapped to the accelerator achieves a classification accuracy of 94.5%, matching the state of the art for embedded implementations. We implement the processing pipeline in a modern 22nm FDX technology and perform post-synthesis power simulation of the network running on the system, achieving an inference energy of 1.7 mu J, which is 647x lower than previously reported results based on Spiking Neural Networks.

Rutishauser, G., Scherer, M., Fischer, T., Benini, L. (2022). Ternarized TCN for mu J/Inference Gesture Recognition from DVS Event Frames. NEW YORK : IEEE [10.23919/DATE54114.2022.9774592].

Ternarized TCN for mu J/Inference Gesture Recognition from DVS Event Frames

Scherer, M;Benini, L
2022

Abstract

Dynamic Vision Sensors (DVS) offer the opportunity to scale the energy consumption in image acquisition proportionally to the activity in the captured scene by only transmitting data when the captured image changes. Their potential for energy-proportional sensing makes them highly attractive for severely energy-constrained sensing nodes at the edge. Most approaches to the processing of DVS data employ Spiking Neural Networks to classify the input from the sensor. In this paper, we propose an alternative, event frame-based approach to the classification of DVS video data. We assemble ternary video frames from the event stream and process them with a fully ternarized Temporal Convolutional Network which can be mapped to CUTIE, a highly energy-efficient Ternary Neural Network accelerator. The network mapped to the accelerator achieves a classification accuracy of 94.5%, matching the state of the art for embedded implementations. We implement the processing pipeline in a modern 22nm FDX technology and perform post-synthesis power simulation of the network running on the system, achieving an inference energy of 1.7 mu J, which is 647x lower than previously reported results based on Spiking Neural Networks.
2022
2022 Design, Automation & Test in Europe Conference & Exhibition (DATE)
736
741
Rutishauser, G., Scherer, M., Fischer, T., Benini, L. (2022). Ternarized TCN for mu J/Inference Gesture Recognition from DVS Event Frames. NEW YORK : IEEE [10.23919/DATE54114.2022.9774592].
Rutishauser, G; Scherer, M; Fischer, T; Benini, L
File in questo prodotto:
File Dimensione Formato  
857_file_Paper.pdf

accesso aperto

Tipo: Postprint
Licenza: Licenza per Accesso Aperto. Creative Commons Attribuzione (CCBY)
Dimensione 338 kB
Formato Adobe PDF
338 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/905406
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 3
  • ???jsp.display-item.citation.isi??? 3
social impact