A note on the convergence of the monotone inclusion version of the primal-dual hybrid gradient algorithm
Abstract
The note contains a direct extension of the convergence proof of the primal-dual hybrid gradient (PDHG) algorithm in [3] to the case of monotone inclusions.
1 Introduction
Assume that are Hilbert spaces, and , are maximally monotone maps. Furthermore, assume that is a non-zero bounded linear operator, and consider the following pair of primal-dual monotone inclusions
(1) |
When are subdifferential maps of proper convex lower semicontinuous functions, this previous problem reduces to a pair of primal-dual convex programs or a convex-concave saddle point problem. More specifically, if , for , then (1) is equivalent to
(2) |
In [2, 3], the authors introduced a first-order primal-dual splitting scheme for solving (2), which in its simplest form reads as
(3) |
where . The main results in [2, 3] provide convergence of ergodic sequences
(4) |
under the assumption
(5) |
In [6], the author considers a more general version of (1) and introduces a splitting scheme, which in its simplest form reads as
(6) |
Using techniques different from the ones in [2, 3], the author in [6] proves the convergence of the iterates in (6) to the solution of (1) under the same assumption (5). The key idea is to rewrite (6) in the form of a forward-backward splitting algorithm analyzed in [4].
2 Notation and hypotheses
Throughout the note, we assume that are Hilbert spaces, are maximally monotone, and is a non-zero bounded linear operator. Furthermore, assume that and are continuously Fréchet differentiable convex functions, and denote by
(7) |
their Bregman divergences. We assume that there exists such that
(8) |
Taking we obtain
(9) |
which means that is -strongly convex. Similarly, we have that
(10) |
and so is also -strongly convex.
Lemma 1.
Assume that is a Hilbert space, is a continuously Fréchet differentiable strongly convex function, and is a maximally monotone operator. Furthermore, denote by
Then the map
is surjective for all .
Proof.
Fix an arbitrary . Since is convex and smooth [1, Theorem 20.25] yields that is maximally monotone with a domain . Hence, by [5, Theorem 1] we have that is maximally monotone.
Next, let . Then for every we have that
Furthermore, the strong convexity of yields that
for some , and from Cauchy-Schwarz inequality we obtain that
Hence
which implies
and [1, Corollary 21.24] concludes the proof. ∎
3 The algorithm and its convergence
Considering the following primal-dual splitting algorithm
(11) |
This previous algorithm is an extension of [3, Algorithm 1], where the subdifferential maps are replaced by general maximally monotone maps. When
we obtain
and (11) reduces to (6). Moreover the existence of an such that (8) holds is equivalent to (5).
Furthermore, Lemma 1 guarantees that all steps in (11) are well defined, and the algorithm will not halt.
Theorem 1.
Proof.
We introduce the following function
(12) |
where we set the supremum of an empty set to be . As pointed out in [3], the basic building block of (11) is the iteration
(13) |
for suitable choices of and . In an expanded form, (13) can be written as
(14) |
where . Thus, we first obtain estimates for the general iteration (14) and then apply them to (11).
Let (14) hold, and , be arbitrary. Then by the monotonicity of and (14) we have that
(15) |
where we also used the identity
Similarly, using the monotonicity of we obtain
(16) |
Combining (15), (16), we obtain
Since are arbitrary, we obtain that
(17) |
As in [3], this previous inequality is the key inequality in the proof. Indeed, (11) corresponds to choosing
Hence, by the convexity of , we obtain
(18) |
for all , , and . Note that (8) guarantees that the expressions in the curly brackets are nonnegative.
Recall that is a solution of (1), and so
(19) |
But then by the definition of we have that
In particular, we have that
(20) |
and (18) yields that
and (8) implies that
Therefore, is a bounded sequence, and the convexity of the norm yields the boundedness of the ergodic sequence with the same bounds; that is,
Assume that is a weak (subsequential) limit of . Invoking (18) again, we obtain
(21) |
for all , , and . Let be arbitrary. Then we have that
and so the weak convergence and (21) yield
Therefore we have that
(22) |
Taking in (22) we obtain
and so maximal monotonicity of yields that
(23) |
Similarly, plugging in in (22) we find that
and the maximal monotonicity of yields that
(24) |
Combining (23) and (24) we obtain that is a solution of (1). ∎
References
- [1] H. H. Bauschke and P. L. Combettes. Convex analysis and monotone operator theory in Hilbert spaces. CMS Books in Mathematics/Ouvrages de Mathématiques de la SMC. Springer, Cham, second edition, 2017. With a foreword by Hédy Attouch.
- [2] A. Chambolle and T. Pock. A first-order primal-dual algorithm for convex problems with applications to imaging. J. Math. Imaging Vision, 40(1):120–145, 2011.
- [3] A. Chambolle and T. Pock. On the ergodic convergence rates of a first-order primal-dual algorithm. Math. Program., 159(1-2):253–287, 2016.
- [4] P. L. Combettes. Solving monotone inclusions via compositions of nonexpansive averaged operators. Optimization, 53(5-6):475–504, 2004.
- [5] R. T. Rockafellar. On the maximality of sums of nonlinear monotone operators. Trans. Amer. Math. Soc., 149:75–88, 1970.
- [6] B. C. Vũ. A splitting algorithm for dual monotone inclusions involving cocoercive operators. Adv. Comput. Math., 38(3):667–681, 2013.