A Fast Learning Algorithm for Deep Belief Nets

Neural Computation - Tập 18 Số 7 - Trang 1527-1554 - 2006
Geoffrey E. Hinton, Simon Osindero1, Yee-Whye Teh2
1Department of Computer Science, University of Toronto, Toronto, Canada M5S 3G4
2Department of Computer Science, National University of Singapore, Singapore 117543

Tóm tắt

We show how to use “complementary priors” to eliminate the explaining-away effects that make inference difficult in densely connected belief nets that have many hidden layers. Using complementary priors, we derive a fast, greedy algorithm that can learn deep, directed belief networks one layer at a time, provided the top two layers form an undirected associative memory. The fast, greedy algorithm is used to initialize a slower learning procedure that fine-tunes the weights using a contrastive version of the wake-sleep algorithm. After fine-tuning, a network with three hidden layers forms a very good generative model of the joint distribution of handwritten digit images and their labels. This generative model gives better digit classification than the best discriminative learning algorithms. The low-dimensional manifolds on which the digits lie are modeled by long ravines in the free-energy landscape of the top-level associative memory, and it is easy to explore these ravines by using the directed connections to display what the associative memory has in mind.

Từ khóa


Tài liệu tham khảo

10.1109/34.993558

10.1023/A:1012454411458

10.1006/inco.1995.1136

10.2307/2287576

10.1162/089976602760128018

10.1126/science.7761831

10.1109/5.726791

10.1364/JOSAA.20.001434

10.1109/34.982899

10.1016/0004-3702(92)90065-6

10.1109/TIP.2005.852470

10.1016/0893-6080(89)90044-0

10.1162/jmlr.2003.4.7-8.1235