A Conjugate Gradient Method Based on a Modified Secant Relation for Unconstrained Optimization
From MaRDI portal
Publication:4959904
DOI10.1080/01630563.2019.1669641zbMath1441.90156OpenAlexW2985720973WikidataQ126825221 ScholiaQ126825221MaRDI QIDQ4959904
Hassan Fahs, Mohammad-Mehdi Hosseini, Razieh Dehghani, Narges Bidabadi
Publication date: 7 April 2020
Published in: Numerical Functional Analysis and Optimization (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1080/01630563.2019.1669641
Numerical mathematical programming methods (65K05) Nonlinear programming (90C30) Methods of quasi-Newton type (90C53) Numerical methods based on nonlinear programming (49M37)
Related Items (2)
An online conjugate gradient algorithm for large-scale data analysis in machine learning ⋮ An Efficient Mixed Conjugate Gradient Method for Solving Unconstrained Optimisation Problems
Uses Software
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- On the sufficient descent condition of the Hager-Zhang conjugate gradient methods
- Convergence analysis of a modified BFGS method on convex minimizations
- New conjugacy condition and related new conjugate gradient methods for unconstrained optimization
- Efficient generalized conjugate gradient algorithms. I: Theory
- Accelerated hybrid conjugate gradient algorithm with modified secant condition for unconstrained optimization
- Two new conjugate gradient methods based on modified secant equations
- Hybrid conjugate gradient algorithm for unconstrained optimization
- New quasi-Newton equation and related methods for unconstrained optimization
- Global convergence result for conjugate gradient methods
- Global convergence properties of nonlinear conjugate gradient methods with modified secant condition
- New quasi-Newton methods for unconstrained optimization problems
- A note on global convergence result for conjugate gradient methods
- A nonlinear conjugate gradient method based on the MBFGS secant condition
- A hybridization of the Hestenes–Stiefel and Dai–Yuan conjugate gradient methods based on a least-squares approach
- Global Convergence Properties of Conjugate Gradient Methods for Optimization
- A Nonlinear Conjugate Gradient Method with a Strong Global Convergence Property
- A New Conjugate Gradient Method with Guaranteed Descent and an Efficient Line Search
- Function minimization by conjugate gradients
- CUTEr and SifDec
- The conjugate gradient method in extremal problems
- Methods of conjugate gradients for solving linear systems
- New conjugacy conditions and related nonlinear conjugate gradient methods
- A modified BFGS method and its global convergence in nonconvex minimization
- Properties and numerical performance of quasi-Newton methods with modified quasi-Newton equations
- Benchmarking optimization software with performance profiles.
This page was built for publication: A Conjugate Gradient Method Based on a Modified Secant Relation for Unconstrained Optimization