Backtracking Gradient Descent Method and some Applications in Large Scale Optimisation Part 1: Theory

Authors

  • Tuyen Trung Truong,Hang-Tuan Nguyen Author

Keywords:

Backtracking, deep learning, global convergence, gradient descent, line search method, optimisation, random dynamical systems.

Abstract

Deep Neural Networks (DNN) are essential in many realistic applications, including Data Science. At the core of DNN is numerical optimisation, in particular gradient descent methods (GD). The purpose of this paper is twofold. First, we prove some new results on the backtracking variant of GD under very general situations. Second, we present a comprehensive comparison of our new results to the previously known results in the literature, providing pros and cons of these methods. To illustrate the efficiency of Backtracking line search, we will present some experimental results (on validation accuracy, training time and so on) on CIFAR10, based on implemetations developed in another paper by the authors. Source codes for the experiments are available on GitHub. 

Downloads

Download data is not yet available.

Downloads

Published

2022-01-05

How to Cite

Backtracking Gradient Descent Method and some Applications in Large Scale Optimisation Part 1: Theory. (2022). Minimax Theory and Its Applications, 7(1), 79–108. https://journalmta.com/index.php/jmta/article/view/175