@inproceedings{01ba5eed0ff64261bbce709d7ac1630b,
title = "Guaranteed Convergence of Training Convolutional Neural Networks via Accelerated Gradient Descent",
abstract = "In this paper, we study the linear regression problem of training an one-hidden-layer non-overlapping convolutional neural networks (ConvNNs) with the rectified linear unit (ReLU) activation functions. Given a set of training data that contains the inputs (feature vectors) and outputs (labels), the outputs are assumed to be generated from a ConvNN with unknown weights, and our goal is to recover the ground-truth weights by minimizing a non-convex optimization problem whose object function is the empirical loss function. We have proved that if the inputs belong to Gaussian distribution, then the optimization problem can be solved by accelerated gradient descent (AGD) algorithm with a well-designed initial point and enough samples, and the iterates via AGD algorithm converge linearly to the ground-truth weights.",
keywords = "accelerated gradient descent, convolutional neural networks, global optimality, linear convergence",
author = "Shuai Zhang and Meng Wang and Sijia Liu and Chen, {Pin Yu} and Jinjun Xiong",
note = "Publisher Copyright: {\textcopyright} 2020 IEEE.; 54th Annual Conference on Information Sciences and Systems, CISS 2020 ; Conference date: 18-03-2020 Through 20-03-2020",
year = "2020",
month = mar,
doi = "10.1109/CISS48834.2020.1570627111",
language = "English (US)",
series = "2020 54th Annual Conference on Information Sciences and Systems, CISS 2020",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
booktitle = "2020 54th Annual Conference on Information Sciences and Systems, CISS 2020",
address = "United States",
}