Abstract
To perform well, deep learning (DL) models have to be trained well. Which optimizer should be adopted? We answer this question by discussing how optimizers have evolved from traditional methods like gradient descent to more advanced techniques to address challenges posed by high-dimensional and non-convex problem space. Ongoing challenges include their hyperparameter sensitivity, balancing between convergence and generalization performance, and improving interpretability of optimization processes. Researchers continue to seek robust, efficient, and universally applicable optimizers to advance the field of DL across various domains.
Original language | English (US) |
---|---|
Pages (from-to) | 2039-2042 |
Number of pages | 4 |
Journal | IEEE/CAA Journal of Automatica Sinica |
Volume | 11 |
Issue number | 10 |
DOIs | |
State | Published - 2024 |
All Science Journal Classification (ASJC) codes
- Control and Systems Engineering
- Information Systems
- Control and Optimization
- Artificial Intelligence