A recent proximal gradient algorithm for convex minimization problem using double inertial extrapolations
In this study, we suggest a new class of forward-backward (FB) algorithms designed to solve convex minimization problems. Our method incorporates a linesearch technique, eliminating the need to choose Lipschitz assumptions explicitly. Additionally, we apply double inertial extrapolations to enhance...
        Saved in:
      
    
          | Published in | AIMS mathematics Vol. 9; no. 7; pp. 18841 - 18859 | 
|---|---|
| Main Authors | , , , | 
| Format | Journal Article | 
| Language | English | 
| Published | 
            AIMS Press
    
        01.01.2024
     | 
| Subjects | |
| Online Access | Get full text | 
| ISSN | 2473-6988 2473-6988  | 
| DOI | 10.3934/math.2024917 | 
Cover
| Summary: | In this study, we suggest a new class of forward-backward (FB) algorithms designed to solve convex minimization problems. Our method incorporates a linesearch technique, eliminating the need to choose Lipschitz assumptions explicitly. Additionally, we apply double inertial extrapolations to enhance the algorithm's convergence rate. We establish a weak convergence theorem under some mild conditions. Furthermore, we perform numerical tests, and apply the algorithm to image restoration and data classification as a practical application. The experimental results show our approach's superior performance and effectiveness, surpassing some existing methods in the literature. | 
|---|---|
| ISSN: | 2473-6988 2473-6988  | 
| DOI: | 10.3934/math.2024917 |