Learning sparse deep neural networks with a spike-and-slab prior

Research output: Contribution to journalArticlepeer-review

13 Scopus citations

Abstract

Deep learning has achieved great successes in many machine learning tasks. However, the deep neural networks (DNNs) are often severely over-parameterized, making them computationally expensive, memory intensive, less interpretable and mis-calibrated. We study sparse DNNs under the Bayesian framework: we establish posterior consistency and structure selection consistency for Bayesian DNNs with a spike-and-slab prior, and illustrate their performance using examples on high-dimensional nonlinear variable selection, large network compression and model calibration. Our numerical results indicate that sparsity is essential for improving the prediction accuracy and calibration of the DNN.

Original languageEnglish (US)
Article number109246
JournalStatistics and Probability Letters
Volume180
DOIs
StatePublished - Jan 2022
Externally publishedYes

All Science Journal Classification (ASJC) codes

  • Statistics and Probability
  • Statistics, Probability and Uncertainty

Keywords

  • Bayesian neural network
  • High-dimensional nonlinear variable selection
  • Model calibration
  • Posterior consistency
  • Sparse deep learning

Fingerprint

Dive into the research topics of 'Learning sparse deep neural networks with a spike-and-slab prior'. Together they form a unique fingerprint.

Cite this