Two descent hybrid conjugate gradient methods for optimization (Q2483351)
From MaRDI portal
scientific article
Language | Label | Description | Also known as |
---|---|---|---|
English | Two descent hybrid conjugate gradient methods for optimization |
scientific article |
Statements
Two descent hybrid conjugate gradient methods for optimization (English)
0 references
28 April 2008
0 references
The aim of the paper is to study convergence and computational properties of two new descent hybrid conjugate gradient methods for nonlinear optimization problems consisting in the global minimization of a continuously differentiable function of \(n\) variables over \(\mathbb{R}^n\). The methods require no restarts and produce a sufficient descent search direction in each iteration. No convexity assumptions are required. The obtained results hold for functions with bounded level sets and bounded Lipschitz continuous gradients. The numerical results presented at the end of the paper show a good efficiency of the proposed methods.
0 references
conjugate gradient method
0 references
descent direction
0 references
global convergence
0 references
global optimization
0 references
nonlinear optimization
0 references
numerical results
0 references
0 references
0 references
0 references