A Low Power, Fully Event-Based Gesture Recognition System
Abstract: We present the first gesture recognition system im-
plemented end-to-end on event-based hardware, using a
TrueNorth neurosynaptic processor to recognize hand ges-
tures in real-time at low power from events streamed live by
a Dynamic Vision Sensor (DVS). The biologically inspired
DVS transmits data only when a pixel detects a change, un-
like traditional frame-based cameras which sample every
pixel at a fixed frame rate. This sparse, asynchronous data
representation lets event-based cameras operate at much
lower power than frame-based cameras. However, much of
the energy efficiency is lost if, as in previous work, the event
stream is interpreted by conventional synchronous proces-
sors. Here, for the first time, we process a live DVS event
stream using TrueNorth, a natively event-based processor
with 1 million spiking neurons. Configured here as a con-
volutional neural network (CNN), the TrueNorth chip iden-
tifies the onset of a gesture with a latency of 105 ms while
consuming less than 200mW. The CNN achieves 96.5%
out-of-sample accuracy on a newly collected DVS dataset
(DvsGesture) comprising 11 hand gesture categories from
29 subjects under 3 illumination conditions.
Loading