Abstract
In this paper, we provide a simple convergence analysis of proximal gradient algorithm with Bregman distance, which provides a tighter bound than existing result. In particular, for the problem of minimizing a class of convex objective functions, we show that proximal gradient algorithm with Bregman distance can be viewed as proximal point algorithm that incorporates another Bregman distance. Consequently, the convergence result of the proximal gradient algorithm with Bregman distance follows directly from that of the proximal point algorithm with Bregman distance, and this leads to a simpler convergence analysis with a tighter convergence bound than existing ones. We further propose and analyze the backtracking line-search variant of the proximal gradient algorithm with Bregman distance.
Original language | English (US) |
---|---|
Pages (from-to) | 903-912 |
Number of pages | 10 |
Journal | Computational Optimization and Applications |
Volume | 73 |
Issue number | 3 |
DOIs | |
State | Published - Jul 1 2019 |
Keywords
- Bregman distance
- Convergence analysis
- Line-search
- Proximal algorithms
ASJC Scopus subject areas
- Control and Optimization
- Computational Mathematics
- Applied Mathematics