Abstract
Deep learning has achieved great successes in many machine learning tasks. However, the deep neural networks (DNNs) are often severely over-parameterized, making them computationally expensive, memory intensive, less interpretable and mis-calibrated. We study sparse DNNs under the Bayesian framework: we establish posterior consistency and structure selection consistency for Bayesian DNNs with a spike-and-slab prior, and illustrate their performance using examples on high-dimensional nonlinear variable selection, large network compression and model calibration. Our numerical results indicate that sparsity is essential for improving the prediction accuracy and calibration of the DNN.
| Original language | English (US) |
|---|---|
| Article number | 109246 |
| Journal | Statistics and Probability Letters |
| Volume | 180 |
| DOIs | |
| State | Published - Jan 2022 |
| Externally published | Yes |
All Science Journal Classification (ASJC) codes
- Statistics and Probability
- Statistics, Probability and Uncertainty
Keywords
- Bayesian neural network
- High-dimensional nonlinear variable selection
- Model calibration
- Posterior consistency
- Sparse deep learning