• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

DMCA

Imagenet classification with deep convolutional neural networks. (2012)

Cached

  • Download as a PDF

Download Links

  • [vision.stanford.edu]
  • [www.cs.toronto.edu]
  • [learning.cs.toronto.edu]
  • [www.cs.toronto.edu]
  • [learning.cs.toronto.edu]
  • [www.cs.utoronto.ca]
  • [www.cs.utoronto.ca]
  • [www.cs.utoronto.ca]
  • [books.nips.cc]
  • [www.cs.toronto.edu]
  • [www.cs.utoronto.ca]
  • [www.nvidia.co.uk]
  • [machinelearning.wustl.edu]
  • [papers.nips.cc]

  • Save to List
  • Add to Collection
  • Correct Errors
  • Monitor Changes
by Alex Krizhevsky , Ilya Sutskever , Geoffrey E Hinton
Venue:In Advances in the Neural Information Processing System,
Citations:1008 - 11 self
  • Summary
  • Citations
  • Active Bibliography
  • Co-citation
  • Clustered Documents
  • Version History

BibTeX

@INPROCEEDINGS{Krizhevsky12imagenetclassification,
    author = {Alex Krizhevsky and Ilya Sutskever and Geoffrey E Hinton},
    title = {Imagenet classification with deep convolutional neural networks.},
    booktitle = {In Advances in the Neural Information Processing System,},
    year = {2012}
}

Share

Facebook Twitter Reddit Bibsonomy

OpenURL

 

Abstract

Abstract We trained a large, deep convolutional neural network to classify the 1.2 million high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 different classes. On the test data, we achieved top-1 and top-5 error rates of 37.5% and 17.0% which is considerably better than the previous state-of-the-art. The neural network, which has 60 million parameters and 650,000 neurons, consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax. To make training faster, we used non-saturating neurons and a very efficient GPU implementation of the convolution operation. To reduce overfitting in the fully-connected layers we employed a recently-developed regularization method called "dropout" that proved to be very effective. We also entered a variant of this model in the ILSVRC-2012 competition and achieved a winning top-5 test error rate of 15.3%, compared to 26.2% achieved by the second-best entry.

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University