Deep Learning: Computational Aspects
2018-08-26Unverified0· sign in to hype
Nicholas Polson, Vadim Sokolov
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
In this article we review computational aspects of Deep Learning (DL). Deep learning uses network architectures consisting of hierarchical layers of latent variables to construct predictors for high-dimensional input-output models. Training a deep learning architecture is computationally intensive, and efficient linear algebra libraries is the key for training and inference. Stochastic gradient descent (SGD) optimization and batch sampling are used to learn from massive data sets.