Abstract
Good old online backpropagation for plain multilayer perceptrons yields a very low 0.35% error rate on the MNIST handwritten digits benchmark. All we need to achieve this best result so far are many hidden layers, many neurons per layer, numerous deformed training images to avoid overfitting, and graphics cards to greatly speed up learning. © 2010 Massachusetts Institute of Technology.
Original language | English (US) |
---|---|
Pages (from-to) | 3207-3220 |
Number of pages | 14 |
Journal | Neural Computation |
Volume | 22 |
Issue number | 12 |
DOIs | |
State | Published - Dec 1 2010 |
Externally published | Yes |
ASJC Scopus subject areas
- Cognitive Neuroscience