2000 character limit reached
Deep Learning: Computational Aspects (1808.08618v2)
Published 26 Aug 2018 in cs.LG, stat.CO, and stat.ML
Abstract: In this article we review computational aspects of Deep Learning (DL). Deep learning uses network architectures consisting of hierarchical layers of latent variables to construct predictors for high-dimensional input-output models. Training a deep learning architecture is computationally intensive, and efficient linear algebra libraries is the key for training and inference. Stochastic gradient descent (SGD) optimization and batch sampling are used to learn from massive data sets.
- Nicholas Polson (18 papers)
- Vadim Sokolov (38 papers)