TY - JOUR
T1 - Deep learning from a statistical perspective
AU - Yuan, Yubai
AU - Deng, Yujia
AU - Zhang, Yanqing
AU - Qu, Annie
N1 - Publisher Copyright:
© 2020 John Wiley & Sons, Ltd.
PY - 2020
Y1 - 2020
N2 - As one of the most rapidly developing artificial intelligence techniques, deep learning has been applied in various machine learning tasks and has received great attention in data science and statistics. Regardless of the complex model structure, deep neural networks can be viewed as a nonlinear and nonparametric generalization of existing statistical models. In this review, we introduce several popular deep learning models including convolutional neural networks, generative adversarial networks, recurrent neural networks, and autoencoders, with their applications in image data, sequential data and recommender systems. We review the architecture of each model and highlight their connections and differences compared with conventional statistical models. In particular, we provide a brief survey of the recent works on the unique overparameterization phenomenon, which explains the strengths and advantages of using an extremely large number of parameters in deep learning. In addition, we provide a practical guidance on optimization algorithms, hyperparameter tuning, and computing resources.
AB - As one of the most rapidly developing artificial intelligence techniques, deep learning has been applied in various machine learning tasks and has received great attention in data science and statistics. Regardless of the complex model structure, deep neural networks can be viewed as a nonlinear and nonparametric generalization of existing statistical models. In this review, we introduce several popular deep learning models including convolutional neural networks, generative adversarial networks, recurrent neural networks, and autoencoders, with their applications in image data, sequential data and recommender systems. We review the architecture of each model and highlight their connections and differences compared with conventional statistical models. In particular, we provide a brief survey of the recent works on the unique overparameterization phenomenon, which explains the strengths and advantages of using an extremely large number of parameters in deep learning. In addition, we provide a practical guidance on optimization algorithms, hyperparameter tuning, and computing resources.
UR - http://www.scopus.com/inward/record.url?scp=85094636136&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85094636136&partnerID=8YFLogxK
U2 - 10.1002/sta4.294
DO - 10.1002/sta4.294
M3 - Article
AN - SCOPUS:85094636136
SN - 2049-1573
VL - 9
JO - Stat
JF - Stat
IS - 1
M1 - e294
ER -