Convolutional networks for fast, energy-efficient neuromorphic computing
Top Cited Papers
- 20 September 2016
- journal article
- research article
- Published by Proceedings of the National Academy of Sciences in Proceedings of the National Academy of Sciences
- Vol. 113 (41) , 11441-11446
- https://doi.org/10.1073/pnas.1604850113
Abstract
Deep networks are now able to achieve human-level performance on a broad spectrum of recognition tasks. Independently, neuromorphic computing has now demonstrated unprecedented energy-efficiency through a new chip architecture based on spiking neurons, low precision synapses, and a scalable communication network. Here, we demonstrate that neuromorphic computing, despite its novel architectural primitives, can implement deep convolution networks that (i) approach state-of-the-art classification accuracy across eight standard datasets encompassing vision and speech, (ii) perform inference while preserving the hardware’s underlying energy-efficiency and high throughput, running on the aforementioned datasets at between 1,200 and 2,600 frames/s and using between 25 and 275 mW (effectively >6,000 frames/s per Watt), and (iii) can be specified and trained using backpropagation with the same ease-of-use as contemporary deep learning. This approach allows the algorithmic power of deep learning to be merged with the efficiency of neuromorphic processors, bringing the promise of embedded, intelligent, brain-inspired computing one step closer. Significance Brain-inspired computing seeks to develop new technologies that solve real-world problems while remaining grounded in the physical requirements of energy, speed, and size. Meeting these challenges requires high-performing algorithms that are capable of running on efficient hardware. Here, we adapt deep convolutional neural networks, which are today’s state-of-the-art approach for machine perception in many domains, to perform classification tasks on neuromorphic hardware, which is today’s most efficient platform for running neural networks. Using our approach, we demonstrate near state-of-the-art accuracy on eight datasets, while running at between 1,200 and 2,600 frames/s and using between 25 and 275 mW.Keywords
All Related Versions
Funding Information
- DOD | Defense Advanced Research Projects Agency (FA9453-15-C-0055)
This publication has 30 references indexed in Scilit:
- Deep learningNature, 2015
- A Ultra-Low-Energy Convolution Engine for Fast Brain-Inspired Vision in Multicore ClustersPublished by EDAA ,2015
- A million spiking-neuron integrated circuit with a scalable communication network and interfaceScience, 2014
- Six Networks on a Universal Neuromorphic Computing SubstrateFrontiers in Neuroscience, 2013
- Multi-column deep neural network for traffic sign classificationNeural Networks, 2012
- Man vs. computer: Benchmarking machine learning algorithms for traffic sign recognitionNeural Networks, 2012
- Assessment for automatic speech recognition: II. NOISEX-92: A database and an experiment to study the effect of additive noise on speech recognition systemsSpeech Communication, 1993
- Neuromorphic electronic systemsProceedings of the IEEE, 1990
- Learning representations by back-propagating errorsNature, 1986
- Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in positionBiological Cybernetics, 1980