Accelerated gradient descent methods with line search (Q5961879): Difference between revisions
From MaRDI portal
Latest revision as of 04:52, 3 July 2024
scientific article; zbMATH DE number 5786243
Language | Label | Description | Also known as |
---|---|---|---|
English | Accelerated gradient descent methods with line search |
scientific article; zbMATH DE number 5786243 |
Statements
Accelerated gradient descent methods with line search (English)
0 references
16 September 2010
0 references
The authors consider the unconstrained minimization problem on \(\mathbb{R}^n\) with a twice differentiable objective function \(f: \mathbb{R}^n\to\mathbb{R}\). A class of gradient descent methods is based on the multiplication of the iteration step size by an appropriate accleration parameter. The step-size is computed by a line search procedure. The methods of this class are called accelerated gradient descent algorithms with the line search. In the further part of the paper, a special accelerated gradient descent method arising from the Newton method with the line search is proposed. The acceleration parameter of this method is obtained by replacing the Hessian by an appropriately generated diagonal matrix. Linear convergence of the proposed algorithm is proved for uniformly convex objective functions satisfying some additional conditions. Reported numerical results show that the proposed method produces better results than alternative methods known from the literature.
0 references
line search
0 references
gradient descent methods
0 references
Newton method
0 references
convergence rate
0 references
0 references
0 references
0 references