Deep Big Simple Neural Nets Excel on Handwritten Digit Recognition

Dan Claudiu Ciresan, Ueli Meier, Luca Maria Gambardella, Juergen Schmidhuber
IDSIA, Galleria 2, 6928 Manno-Lugano, Switzerland
arXiv:1003.0358v1 [cs.NE] (1 Mar 2010)


   author={Claudiu Ciresan}, D. and {Meier}, U. and {Gambardella}, L.~M. and {Schmidhuber}, J.},

   title={“{Deep Big Simple Neural Nets Excel on Handwritten Digit Recognition}”},

   journal={ArXiv e-prints},



   keywords={Computer Science – Neural and Evolutionary Computing, Computer Science – Artificial Intelligence},




   adsnote={Provided by the SAO/NASA Astrophysics Data System}


Download Download (PDF)   View View   Source Source   



Good old on-line back-propagation for plain multi-layer perceptrons yields a very low 0.35% error rate on the famous MNIST handwritten digits benchmark. All we need to achieve this best result so far are many hidden layers, many neurons per layer, numerous deformed training images, and graphics cards to greatly speed up learning.
No votes yet.
Please wait...

* * *

* * *

HGPU group © 2010-2021 hgpu.org

All rights belong to the respective authors

Contact us: