Publication: Avoiding pathologies in very deep networks
Open/View Files
Date
2014
Published Version
Published Version
Journal Title
Journal ISSN
Volume Title
Publisher
Journal of Machine Learning Research
The Harvard community has made this article openly available. Please share how this access benefits you.
Citation
Duvenaud, David, Oren Rippel, Ryan Adams, and Zoubin Ghahramani. 2014. "Avoiding pathologies in very deep networks." In Proceedings of the 17th International Conference on Artificial Intelligence and Statistics (AISTATS) 2014, Reykjavik, Iceland. JMLR: W&CP volume 33: 202-210.
Research Data
Abstract
Choosing appropriate architectures and regularization strategies of deep networks is crucial to good predictive performance. To shed light on this problem, we analyze the analogous problem of constructing useful priors on compositions of functions. Specifically, we study the deep Gaussian process, a type of infinitely-wide, deep neural network. We show that in standard architectures, the representational capacity of the network tends to capture fewer degrees of freedom as the number of layers increases, retaining only a single degree of freedom in the limit. We propose an alternate network architecture which does not suffer from this pathology. We also examine deep covariance functions, obtained by composing infinitely many feature transforms. Lastly, we characterize the class of models obtained by performing dropout on Gaussian processes.
Description
Other Available Sources
Keywords
Terms of Use
This article is made available under the terms and conditions applicable to Other Posted Material (LAA), as set forth at Terms of Service