ImageNet classification with deep convolutional neural networks
Top Cited Papers
- 24 May 2017
- journal article
- research article
- Published by Association for Computing Machinery (ACM) in Communications of the ACM
- Vol. 60 (6) , 84-90
- https://doi.org/10.1145/3065386
Abstract
We trained a large, deep convolutional neural network to classify the 1.2 million high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 different classes. On the test data, we achieved top-1 and top-5 error rates of 37.5% and 17.0%, respectively, which is considerably better than the previous state-of-the-art. The neural network, which has 60 million parameters and 650,000 neurons, consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully connected layers with a final 1000-way softmax. To make training faster, we used non-saturating neurons and a very efficient GPU implementation of the convolution operation. To reduce overfitting in the fully connected layers we employed a recently developed regularization method called "dropout" that proved to be very effective. We also entered a variant of this model in the ILSVRC-2012 competition and achieved a winning top-5 test error rate of 15.3%, compared to 26.2% achieved by the second-best entry.This publication has 16 references indexed in Scilit:
- Going deeper with convolutionsPublished by Institute of Electrical and Electronics Engineers (IEEE) ,2015
- High-dimensional signature compression for large-scale image classificationPublished by Institute of Electrical and Electronics Engineers (IEEE) ,2011
- Convolutional Networks Can Learn to Generate Affinity Graphs for Image SegmentationNeural Computation, 2010
- A High-Throughput Screening Approach to Discovering Good Forms of Biologically Inspired Visual RepresentationPLoS Computational Biology, 2009
- Why is Real-World Visual Object Recognition Hard?PLoS Computational Biology, 2008
- Lessons from the Netflix prize challengeACM SIGKDD Explorations Newsletter, 2007
- LabelMe: A Database and Web-Based Tool for Image AnnotationInternational Journal of Computer Vision, 2007
- Learning generative visual models from few training examples: An incremental Bayesian approach tested on 101 object categoriesComputer Vision and Image Understanding, 2007
- Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in positionBiological Cybernetics, 1980
- Taylor expansion of the accumulated rounding errorBIT Numerical Mathematics, 1976