Citation
Abstract
This paper proposes a nonmonotone spectral gradient method for solving large-scale unconstrained optimization problems. The spectral parameter is derived from the eigenvalues of an optimally sized memoryless symmetric rank-one matrix obtained under the measure defined as a ratio of the determinant of updating matrix over its largest eigenvalue. Coupled with a nonmonotone line search strategy where backtracking-type line search is applied selectively, the spectral parameter acts as a stepsize during iterations when no line search is performed and as a milder form of quasi-Newton update when backtracking line search is employed. Convergence properties of the proposed method are established for uniformly convex functions. Extensive numerical experiments are conducted and the results indicate that our proposed spectral gradient method outperforms some standard conjugate-gradient methods.
Download File
Full text not available from this repository.
Official URL or Download Paper: https://www.aimsciences.org/article/doi/10.3934/ji...
|
Additional Metadata
Item Type: | Article |
---|---|
Divisions: | Faculty of Science Institute for Mathematical Research |
DOI Number: | https://doi.org/10.3934/jimo.2021143 |
Publisher: | American Institute of Mathematical Sciences |
Keywords: | Large-scale unconstrained optimization; Spectral gradient method; Nonmonotone line search; Memoryless symmetric rank-one update; Quasi-Newton update |
Depositing User: | Ms. Nur Faseha Mohd Kadim |
Date Deposited: | 04 Apr 2023 04:25 |
Last Modified: | 04 Apr 2023 04:25 |
Altmetrics: | http://www.altmetric.com/details.php?domain=psasir.upm.edu.my&doi=10.3934/jimo.2021143 |
URI: | http://psasir.upm.edu.my/id/eprint/94373 |
Statistic Details: | View Download Statistic |
Actions (login required)
View Item |