A Low Power, Fully Event-Based Gesture Recognition System

Published: 22 Jul 2017, Last Modified: 01 Apr 2024CVPR 2017EveryoneRevisionsCC BY 4.0
Abstract: We present the first gesture recognition system im- plemented end-to-end on event-based hardware, using a TrueNorth neurosynaptic processor to recognize hand ges- tures in real-time at low power from events streamed live by a Dynamic Vision Sensor (DVS). The biologically inspired DVS transmits data only when a pixel detects a change, un- like traditional frame-based cameras which sample every pixel at a fixed frame rate. This sparse, asynchronous data representation lets event-based cameras operate at much lower power than frame-based cameras. However, much of the energy efficiency is lost if, as in previous work, the event stream is interpreted by conventional synchronous proces- sors. Here, for the first time, we process a live DVS event stream using TrueNorth, a natively event-based processor with 1 million spiking neurons. Configured here as a con- volutional neural network (CNN), the TrueNorth chip iden- tifies the onset of a gesture with a latency of 105 ms while consuming less than 200mW. The CNN achieves 96.5% out-of-sample accuracy on a newly collected DVS dataset (DvsGesture) comprising 11 hand gesture categories from 29 subjects under 3 illumination conditions.
Loading