Citation
Abstract
We will tackle the l0-norm sparse optimization problem using an underdetermined system as a constraint in this research. This problem is turned into an unconstrained optimization problem using the Lagrangian method and solved using the proximal variable metric method. This approach combines the proximal and variable metric methods by substituting a diagonal matrix for the approximation of the full rank Hessian matrix. Hence, the memory requirement is reduced to O(n) storage instead of O(n2 ) storage. The diagonal updating matrix is derived from the same variational technique used in the derivation of variable metric or quasi-Newton updates but incorporated with some weaker form of quasi-Newton relation. Convergence analysis of this method is established. The efficiency of the proposed method is compared against existing versions of proximal gradient methods on simulated datasets and large real-world MNIST datasets using Python software. Numerical results show that our proposed method is more robust and stable for finding sparse solutions to the linear system.
Download File
Full text not available from this repository.
Official URL or Download Paper: https://linkinghub.elsevier.com/retrieve/pii/S0016...
|
Additional Metadata
Item Type: | Article |
---|---|
Divisions: | Institute for Mathematical Research |
DOI Number: | https://doi.org/10.1016/j.jfranklin.2023.02.035 |
Publisher: | Elsevier |
Keywords: | Sparse optimization; Lagrangian method; Proximal variable metric; Diagonal updating matrix; Memory reduction; Quasi-Newton relation; Convergence analysis; Numerical efficiency; MNIST dataset |
Depositing User: | Ms. Nur Faseha Mohd Kadim |
Date Deposited: | 14 Oct 2024 06:59 |
Last Modified: | 14 Oct 2024 06:59 |
Altmetrics: | http://www.altmetric.com/details.php?domain=psasir.upm.edu.my&doi=10.1016/j.jfranklin.2023.02.035 |
URI: | http://psasir.upm.edu.my/id/eprint/109005 |
Statistic Details: | View Download Statistic |
Actions (login required)
View Item |