Distributed Gradient Tracking for Unbalanced Optimization With Different Constraint Sets
Gradient tracking methods have become popular for distributed optimization in recent years, partially because they achieve linear convergence using only a constant step-size for strongly convex optimization. In this article, we construct a counterexample on constrained optimization to show that dire...
        Saved in:
      
    
          | Published in | IEEE transactions on automatic control Vol. 68; no. 6; pp. 3633 - 3640 | 
|---|---|
| Main Authors | , , , | 
| Format | Journal Article | 
| Language | English | 
| Published | 
        New York
          IEEE
    
        01.06.2023
     The Institute of Electrical and Electronics Engineers, Inc. (IEEE)  | 
| Subjects | |
| Online Access | Get full text | 
| ISSN | 0018-9286 1558-2523  | 
| DOI | 10.1109/TAC.2022.3192316 | 
Cover
| Summary: | Gradient tracking methods have become popular for distributed optimization in recent years, partially because they achieve linear convergence using only a constant step-size for strongly convex optimization. In this article, we construct a counterexample on constrained optimization to show that direct extension of gradient tracking by using projections cannot guarantee the correctness. Then, we propose projected gradient tracking algorithms with diminishing step-sizes rather than a constant one for distributed strongly convex optimization with different constraint sets and unbalanced graphs. Our basic algorithm can achieve <inline-formula><tex-math notation="LaTeX">O(\ln T/{T})</tex-math></inline-formula> convergence rate. Moreover, we design an epoch iteration scheme and improve the convergence rate as <inline-formula><tex-math notation="LaTeX">O(1/{T})</tex-math></inline-formula>. | 
|---|---|
| Bibliography: | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 14  | 
| ISSN: | 0018-9286 1558-2523  | 
| DOI: | 10.1109/TAC.2022.3192316 |