培训(气象学)
图层(电子)
计算机科学
人工智能
地理
材料科学
复合材料
气象学
作者
Yoshua Bengio,Pascal Lamblin,Dan Popovici,Hugo Larochelle
出处
期刊:The MIT Press eBooks
[The MIT Press]
日期:2007-09-07
卷期号:: 153-160
被引量:1590
标识
DOI:10.7551/mitpress/7503.003.0024
摘要
Complexity theory of circuits strongly suggests that deep architectures can be much more efficient (sometimes exponentially) than shallow architectures, in terms of computational elements required to represent some functions. Deep multi-layer neural networks have many levels of non-linearities allowing them to compactly represent highly non-linear and highly-varying functions. However, until recently it was not clear how to train such deep networks, since gradient-based optimization starting from random initialization appears to often get stuck in poor solutions. Hinton et al. recently introduced a greedy layer-wise unsupervised learning algorithm for Deep Belief Networks (DBN), a generative model with many layers of hidden causal variables. In the context of the above optimization problem, we study this algorithm empirically and explore variants to better understand its success and extend it to cases where the inputs are continuous or where the structure of the input distribution is not revealing enough about the variable to be predicted in a supervised task. Our experiments also confirm the hypothesis that the greedy layer-wise unsupervised training strategy mostly helps the optimization, by initializing weights in a region near a good local minimum, giving rise to internal distributed representations that are high-level abstractions of the input, bringing better generalization.
科研通智能强力驱动
Strongly Powered by AbleSci AI