Deep learning from a statistical perspective

Yubai Yuan, Yujia Deng, Yanqing Zhang, Annie Qu

Research output: Contribution to journalArticlepeer-review

7 Scopus citations

Abstract

As one of the most rapidly developing artificial intelligence techniques, deep learning has been applied in various machine learning tasks and has received great attention in data science and statistics. Regardless of the complex model structure, deep neural networks can be viewed as a nonlinear and nonparametric generalization of existing statistical models. In this review, we introduce several popular deep learning models including convolutional neural networks, generative adversarial networks, recurrent neural networks, and autoencoders, with their applications in image data, sequential data and recommender systems. We review the architecture of each model and highlight their connections and differences compared with conventional statistical models. In particular, we provide a brief survey of the recent works on the unique overparameterization phenomenon, which explains the strengths and advantages of using an extremely large number of parameters in deep learning. In addition, we provide a practical guidance on optimization algorithms, hyperparameter tuning, and computing resources.

Original languageEnglish (US)
Article numbere294
JournalStat
Volume9
Issue number1
DOIs
StatePublished - 2020

All Science Journal Classification (ASJC) codes

  • Statistics and Probability
  • Statistics, Probability and Uncertainty

Fingerprint

Dive into the research topics of 'Deep learning from a statistical perspective'. Together they form a unique fingerprint.

Cite this