TY - GEN
T1 - Modular deep belief networks that do not forget
AU - Pape, Leo
AU - Gomez, Faustino
AU - Ring, Mark
AU - Schmidhuber, Jurgen
N1 - Generated from Scopus record by KAUST IRTS on 2022-09-14
PY - 2011/10/24
Y1 - 2011/10/24
N2 - Deep belief networks (DBNs) are popular for learning compact representations of high-dimensional data. However, most approaches so far rely on having a single, complete training set. If the distribution of relevant features changes during subsequent training stages, the features learned in earlier stages are gradually forgotten. Often it is desirable for learning algorithms to retain what they have previously learned, even if the input distribution temporarily changes. This paper introduces the M-DBN, an unsupervised modular DBN that addresses the forgetting problem. M-DBNs are composed of a number of modules that are trained only on samples they best reconstruct. While modularization by itself does not prevent forgetting, the M-DBN additionally uses a learning method that adjusts each module's learning rate proportionally to the fraction of best reconstructed samples. On the MNIST handwritten digit dataset module specialization largely corresponds to the digits discerned by humans. Furthermore, in several learning tasks with changing MNIST digits, M-DBNs retain learned features even after those features are removed from the training data, while monolithic DBNs of comparable size forget feature mappings learned before. © 2011 IEEE.
AB - Deep belief networks (DBNs) are popular for learning compact representations of high-dimensional data. However, most approaches so far rely on having a single, complete training set. If the distribution of relevant features changes during subsequent training stages, the features learned in earlier stages are gradually forgotten. Often it is desirable for learning algorithms to retain what they have previously learned, even if the input distribution temporarily changes. This paper introduces the M-DBN, an unsupervised modular DBN that addresses the forgetting problem. M-DBNs are composed of a number of modules that are trained only on samples they best reconstruct. While modularization by itself does not prevent forgetting, the M-DBN additionally uses a learning method that adjusts each module's learning rate proportionally to the fraction of best reconstructed samples. On the MNIST handwritten digit dataset module specialization largely corresponds to the digits discerned by humans. Furthermore, in several learning tasks with changing MNIST digits, M-DBNs retain learned features even after those features are removed from the training data, while monolithic DBNs of comparable size forget feature mappings learned before. © 2011 IEEE.
UR - http://ieeexplore.ieee.org/document/6033359/
UR - http://www.scopus.com/inward/record.url?scp=80054728044&partnerID=8YFLogxK
U2 - 10.1109/IJCNN.2011.6033359
DO - 10.1109/IJCNN.2011.6033359
M3 - Conference contribution
SN - 9781457710865
SP - 1191
EP - 1198
BT - Proceedings of the International Joint Conference on Neural Networks
ER -