Evolution and Role of Optimizers in Training Deep Learning Models

Xiao Hao Wen, Meng Chu Zhou

Research output: Contribution to journalArticlepeer-review

6 Scopus citations

Abstract

To perform well, deep learning (DL) models have to be trained well. Which optimizer should be adopted? We answer this question by discussing how optimizers have evolved from traditional methods like gradient descent to more advanced techniques to address challenges posed by high-dimensional and non-convex problem space. Ongoing challenges include their hyperparameter sensitivity, balancing between convergence and generalization performance, and improving interpretability of optimization processes. Researchers continue to seek robust, efficient, and universally applicable optimizers to advance the field of DL across various domains.

Original languageEnglish (US)
Pages (from-to)2039-2042
Number of pages4
JournalIEEE/CAA Journal of Automatica Sinica
Volume11
Issue number10
DOIs
StatePublished - 2024

All Science Journal Classification (ASJC) codes

  • Control and Systems Engineering
  • Information Systems
  • Control and Optimization
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Evolution and Role of Optimizers in Training Deep Learning Models'. Together they form a unique fingerprint.

Cite this