∎ ∎
[email protected]
L.P. Tang 33institutetext: National Center for Applied Mathematics of Chongqing, Chongqing 401331, China
[email protected]
🖂X.M. Yang 44institutetext: National Center for Applied Mathematics of Chongqing, and School of Mathematical Sciences, Chongqing Normal University, Chongqing 401331, China
[email protected]
Convergence rates analysis of Interior Bregman Gradient Method for Vector Optimization Problems
Abstract
In recent years, by using Bregman distance, the Lipschitz gradient continuity and strong convexity were lifted and replaced by relative smoothness and relative strong convexity. Under the mild assumptions, it was proved that gradient methods with Bregman regularity converge linearly for single-objective optimization problems (SOPs). In this paper, we extend the relative smoothness and relative strong convexity to vector-valued functions and analyze the convergence of an interior Bregman gradient method for vector optimization problems (VOPs). Specifically, the global convergence rates are and for convex and relative strongly convex VOPs, respectively. Moreover, the proposed method converges linearly for VOPs that satisfy a vector Bregman-PL inequality.
Keywords:
Vector optimization Bregman distance Relative smoothness Relative strong convexity Linear convergenceMSC:
65K05 90C26 90C29 90C301 Introduction
Let be a closed, convex, and pointed cone with a non-empty interior. The partial order induced by :
In this paper, we consider the the following vector optimization problem
(VOP) |
where every component is differentiable, and is closed and convex with a non-empty interior. For the optimal solutions of (VOP), none of the objectives can be improved without sacrificing the others. The concept of optimality is thus defined as efficiency. It is worth noting that (VOP) corresponds to a multiobjective optimization problem when , where is the non-negative orthant of . Some applications of multiobjective optimization problems can be found in engineering 1a , economics 1b ; eco , and machine learning mac1 ; m2 ; mac2 , etc.
Scalarization approaches mnp ; luc ; joh are widely explored for VOPs, which convert a VOP into a single-objective optimization problem so that classical numerical optimization methods can be applied to obtain the solutions to the original problem. Unfortunately, the approaches burden the decision-maker with some parameters which are unknown in advance. To overcome the drawback, Fliege and Svaiter 6 invented the parameter-free method, called the steepest descent method for multiobjective optimization. The corresponding method for VOPs is described in vsd . Motivated by the work of Fliege and Svaiter, some standard first-order method methods are extended to solve MOPs, and VOPs 8 ; 9a ; 9b ; 9c ; 9d ; chen . The main idea of these methods is to compute a descent direction; then, a line search technique is performed along the direction to ensure sufficient decrease for all objective functions. By the strategy, these descent methods produce a sequence of points that enjoy global convergence. Very recently, it was proved that the convergence rates of the multiobjective gradient descent method com and proximal gradient method pcom are the same as the ones for SOPs.
Similar to the first-order method for SOPs, a widely used assumption for MOPs is that the objective function gradients have to be globally Lipschitz continuous, which guarantees a sufficient descent in each iteration. By using Bregman distance, the restrictive assumption was replaced by the Lipschitz-like condition lc , or relative smoothness rs for SOPs. On the other hand, the relative strong convexity was proposed to replace strong convexity in rs . Under the mild assumptions, the linear convergence results were established for gradient rs and proximal gradient method as with Bregman regularization. Moreover, the linear convergence results of the gradient method were obtained with Bregman error bound condition nlns . Auslender and Teboulle ip presented another advantage of the Bregman method: the produced sequence can be restricted in the interior of the feasible set by choosing a suitable Bregman distance and thus eliminate the constraints.
The Bregman gradient methods were also explored in VOPs. Villacorta and Oliveira mip developed the interior Bregman gradient method with a modified convergence sensing condition for convex VOPs with a non-polyhedral set. In cz . Chen et al. introduced the vector-valued Bregman function for MOPs, and the corresponding vector-valued Bregman distance replaced the Euclidean distance in the proximal point algorithm (PPA). Very recently, the PPA with vector-valued Bregman distance was applied to solve multiobjective DC programming in 2022 . Moreover, the proximal gradient method with Bregman functions for MOPs has been investigated by Ansary and Dutta 2022a .
It should be noted that convergence analysis with the appealing potential of Bregman distance was not discussed in the methods mentioned above. Naturally, a question arises can we obtain linear convergence results of the Bregman gradient method for VOPs without Lipschitz gradient continuity and strong convexity. The paper’s main contribution is to answer the question positively. More specifically:
-
We extend the relative smoothness and relative strong convexity to vector-valued functions in the context of VOPs;
-
We present two merit functions and analyze their properties, such as optimality, continuity, and interrelation. We also proposed a generalized vector Bregman-PL inequality, which will be used to prove the linear convergence of the proposed method.
-
We propose an interior Bregman gradient method for VOPs that restricts the produced sequence inside the feasible set. It is proved that every accumulation point of the produced sequence is -stationary point for VOP, and the whole sequence converges to a weakly efficient set with -convex objective functions.
-
With relative smooth and strongly convex objective functions, we prove the produced sequence converges linearly to a weakly efficient point in the sense of Bregman distance. We also prove that a merit function converges linearly with the vector Bregman-PL inequality.
The organization of the paper is as follows. Some notations and definitions are given in Sect. 2 for our later use. Sect. 3 discusses the merit functions for VOPs. Sect. 4 is devoted to introducing the interior Bregman gradient method and proving the convergence results for the proposed method. At the end of the paper, some conclusions are drawn.
2 Preliminaries
2.1 Notations
-
.
-
the relative interior of -dimensional unit simplex.
-
the Euclidean norm, the inner product.
-
and the interior and the closure of a set, respectively.
-
the convex hull of a set.
-
the closed ball centred at with radius .
-
and the Jacobian matrix and the gradient of at , respectively.
-
the positive polar cone of .
-
.
-
the conjugate function of .
2.2 Vector optimization
Recall some definitions of solutions to (VOP) as follows.
Definition 1.
Definition 2.
For a non-stationary point , there exists a descent direction that is defined as:
Definition 4.
Remark 1.
If is a non-stationary point, then there exists a vector such that . Note that is an open set, the relation is also valid for some . It follows from Definition 4 that is a -descent direction.
Definition 5.
Since is differentiable, -convexity of on is equivalent to
By using the positive polar cone , an equivalent characterization of -convexity of on is presented as follows.
Lemma 1.
luc A function is -convex on if and only if is convex on for all .
Remark 2.
Note that for all , then is -convex on if and only if is convex on for all .
In -convex case, we derive the equivalence between -stationary point and weakly efficient solution.
2.3 Relative smoothness and relative strong convexity
Definition 6 (Legendre function).
roc Let be a proper closed convex function, where . Then
-
is essentially smooth if is differentiable on and for all converges to some boundary point of ;
-
is a Legendre function if is essentially smooth and strictly convex on .
Lemma 3.
roc For a Legendre function , we have the following useful properties.
-
is Legendre if and only if its conjugate is Legendre.
-
.
-
.
-
with .
Using the Legendre function . the Bregman distance w.r.t. is defined as
Since is strictly convex on , we have for all , and the equality holds if and only if . In what follows, we recall a basic property for that will be used in the sequel:
(1) |
In general, is not symmetric. The measure of symmetry for was introduced in lc . The symmetric coefficient is given by
Remark 3.
In the context of vector optimization, we propose the relative smoothness of relative to Legendre function .
Definition 7.
For a vector , we call is -smooth relative to on if for all , there exists such that
The following proposition presents some properties for relative smoothness.
Proposition 1.
Assume is -smooth relative to on , then
-
it is equivalent to is -convex on ;
-
for any , is -smooth relative to on , where .
Proof.
(i) From Remark 2, -convexity of on is equivalent to is convex on for all . Then, we have
This is equivalent to , assertion (i) is proved.
(ii) From assertion (i), we obtain is convex on for all . Since is convex, for any , it follows that is convex on for all when holds for all . The inequality holds by setting , and is well-defined due to and the compactness of .
We also propose the relative strong convex of relative to Legendre function .
Definition 8.
For a vector , we call is -strongly convex relative to on if for all , there exists such that
Similarly, we present some properties for relative strong convexity as follows.
Proposition 2.
Assume is -strongly convex relative to on , then
-
it is equivalent to is -convex on ;
-
for any , is -strongly convex relative to on , where .
Proof.
The proof is similar to the arguments used in Proposition 1, we omit it here.
Remark 4.
From the above two propositions, the constants of relative smoothness and relative strong convexity depend on , and . Recall that the quotient plays a key role in the geometric convergence of first-order methods in the presence of strong convexity. If is large, the function descreases slowly when first-order methods are applied. We call the function is ill-conditioned in the situation. To alleviate this dilemma, for the general and , the vector is selected as:
(2) |
When , the vector defined as (2) corresponds to . For simplicity, we select a suitable such that
(3) |
On the other hand, since is a closed, convex, and pointed cone, we have . Then we defined
(4) |
3 Merit Functions for VOPs
A function is called merit function for (VOP) if it returns at the solutions of (VOP) and strictly positive values otherwise. In this section, we first present the following two merit functions.
(5) |
(6) |
where is a constant and .
3.1 Properties of merit functions
We can show that and are merit functions in the sense of weak efficiency and stationarity, respectively.
Theorem 1.
Proof.
(i) Since is a weakly efficient solution of (VOP), we have
Then
which is equivalent to
On the other hand, the definition of implies that
which is equivalent to .
(ii) Since is -stationary point of (VOP), we have
which, along the fact that , yields
On the other hand, the definition of implies that
Hence, .
Conversely, assume that is not -stationary when . Then, there exists such that
For all , . Hence,
On the other hand, the differentiability of implies . This together with the above inequality and yield . This contradicts . The proof is completed.
We denote by the optimal solution of . Hence,
(7) |
In the sequel, we systematically assume that is well-defined on , i.e., is a single valued mapping from to . We give a sufficient condition for the assumption.
Lemma 4.
If is supercoercive, i.e.,
then is a single valued mapping from to .
Proof.
For any fixed point , and . We define the function for as follows
(8) |
Then . Substituting the equality of Bregman distance into equality (8), we have
Taking , then is bounded due to the compactness of . This combines with the boundness of and supercoercivity of implies . Hence, is level bounded on . It follows from continuity of and Weierstrass’s theorem (rocv, , Theorem 1.9) that is nonempty. The uniqueness of is given by the strict convexity of . From the optimality condition of , it follows that is nonempty. Then Lemma 3(iv) implies that .
Remark 5.
Recall that , we have . From Lemma 4, if is supercoercive, then is a single valued mapping from to .
Proposition 3.
Proof.
(i) Note that is convex, is compact and convex, and is convex for and concave for . Therefore, it follows by Sion’s minimax theorem 32 that there exists such that
and
and
Hence, we obtain (10), and (9) follows by the strict convexity of .
(ii) From the optimality condition for (9) and the fact that , we have
(11) |
Then, assertion (i) follows from the fact that .
(iv) Using (9) and the fact that is the unique solution of , we obtain
where the second equality is given by assertion (ii). The proof is completed.
We can also show the continuity of and .
Proposition 4.
For all , the and are continuous on .
Proof.
From Proposition 3(iv), it is sufficient to prove the continuity of . For all , the optimality condition (11) gives
For and a sequence satisfying . Substituting and into the above equality, respectively, and sum them up, we have
where the inequality is given by (10). For simplicity, we denote the right hand side of the above inequality as . Then, we have
Notice that is strictly convex on , we have
On the other hand, since and are continuous, and , we obtain and tend to . Then tends to . It follows by the strict convexity of that . Hence, the continuity of follows from the arbitrary of and .
3.2 Vector Bregman-PL inequality
In this subsection, a noval vector Bregman-PL inequality is derived with merit functions. At first, we present the connection between and for some , which will be used to establish the connection between and .
Lemma 5.
Let , and . If , we have
(12) |
Proof.
The proof is similar to the arguments in the proof of (nlns, , Lemma 3.5), we omit it here.
Note that it is not clear whether the inequality holds for in Lemma 5. To overcome this difficulty we present the following assumption.
Assumption 1.
Let , and . Then there exists such that
(13) |
The Assumption 1 seems strict in general. Indeed, we have the following sufficient condition for Assumption 1, which is first presented in nlns .
Remark 6.
We are now in a position to present the relation between and .
Proposition 5.
Proof.
The left hand side of inequalities (14) follows directly from the definition of . Next, we proof the right hand side of inequalities (14). From (9) and the definition of , we have
Using the similar argument in the proof of Proposition 3(iv), we obtain
We further use inequality (13) to get
where the equality is given by Proposition 3(iv). Therefore, we obtain the right hand side of inequalities (14).
Moreover, we present the relations between and .
Proposition 6.
For the merit functions and , the following statements hold.
-
If is -smooth relative to for some , then
(15) -
If is -strongly convex relative to for some , then
(16)
Proof.
(i) From the -smoothness of , we have
Hence,
where the last inequality is given by (3). It follows that .
From the -strong comvexity of , we have
Therefore,
where the last inequality is given by (4). Hence, .
We propose the following error bound property, which will be used in convergence rate analysis.
Definition 9.
We say that vector Bregman-PL inequality holds for on when
(17) |
where .
Remark 7.
The vector Bregman-PL inequality (17) corresponds to the general PL inequality when and . It is also a generalization for gradient dominated inequality (nlns, , Definition 3.2) in the context of vector optimization problems. Recently, a multiobjective proximal-PL inequality was established in (pcom, , Definition 4.1). The vector Bregman-PL inequality coincides with multiobjective proximal-PL inequality when and for all .
By virtue of Propositions 5 and 6, we give the sufficient conditions for vector Bregman-PL inequality.
Proposition 7.
4 Interior Bregman Gradient Method for VOPs
In this section, we first introduce interior Bregman gradient method for VOPs.
The proposed algorithm generates a sequence via the following iterates:
From Lemma 4 and the strict convexity of , the algorithm is well-defined, and the generated sequence lies in . Note that is the unique solution of
then
Define mapping as
From the iterates in Algorithm 1, we deduce the following sufficient descent property.
Lemma 6 (Sufficient descent property).
Assume that is -smooth relative to , and . Then the sequence produced by Algorithm 1 satisfies
(18) |
Furthermore, the sufficient descent property holds when
Proof.
We can see that Algorithm 1 terminates with an -stationary point in a finite number of iterations or generates an infinite sequence. In the sequel, we will suppose that Algorithm 1 generates an infinite sequence of nonstationary points. Firstly, we present the global convergence of Algorithm 1 in nonconvex case.
Theorem 2.
Assume that is bounded, is -smooth relative to for some , and . Let be the sequence generated by Algorithm 1. Then, we have
-
there exists for all such that .
-
.
-
has at least one accumulation point, and every accumulation point is a -stationary point. Moreover, if is -Lipschitz continuous on , every accumulation point is a -stationary point.
-
Proof.
(i) From Lemma 6, it follows that is descreasing under the partial order induced by . This combined with boundness of implies is bounded and there exists for all such that .
(ii) Summing the inequality (18) over , we obtain
where the second inequality follows from the subadditiveness of , and the last inequality is given by the definition of and . Hence
(iii) From Proposition 3(iv), we have
(19) |
this together with assertion (ii) imply
On the other hand, the boundness of yields that there exists a subsequence tends to , where is an accumulation point. Next, we prove is a -stationary point. We distinguish two cases: or . If , it follows from the continuity of that
Therefore, is a -stationary point. If , suppose by contrary that is a non-stationary point. From Remark 1, there exists such that
Then, we denote by
From the continuity of and the compactness of , we have
for is sufficient large. Then, we have
for is sufficient large and all , where the first inequality follows from , the third inequality is given by -Lipschitz continuity of , and the last inequality is due to that is sufficient large so that, without loss of generality, we have . Recall that the above inequalities holds for all , then there exists such that . This contradicts . Therefore, is a -stationary point.
(iv) In the proof of assertion (ii), we obtained
this combines with (19) yield
Then, assertion (iv) holds due to
Remark 8.
In Theorem 2(iii), if the accumulation lies in , the continuity of and can not be derived due to . Recall that when , the continuity of , corresponds to , implied the stationarity of accumulation points 6 ; vsd . Alternatively, since the gradient of is -Lipschitz continuous, the stationarity of accumulation points follows directly from a subsequence . The result is crucial when the continuity of and are unknown. For example, the authors used but didn’t prove the continuity of in (dsd, , Theorem 3.2). However, the result is valid, which can be proved by .
4.1 Strong convergence
In this subsection, we discuss the strong convergence of Algorithm 1 in convex case. Firstly, we present the following identity, known as three points lemma.
Lemma 7 (Three points lemma).
Applying the three points lemma, we can now establish a fundamental inequality for Algorithm 1 with -convex objective function.
Lemma 8.
Assume that is -smooth and -strongly convex relative to for some and . Let for some . Then, for all , we have
(20) |
Proof.
Before presenting the convergence result in convex case, we recall the following result on nonnegative sequences.
Lemma 9.
We also recall the following assumption.
Assumption 2.
(lc, , Assumption H) Assume the Legendre function satisfies
-
If converges to some point , then .
-
Reciprocally, if and if is such that , then .
Remark 9.
In what follows, we give the convergence result for convex case.
Theorem 3.
Proof.
Since all assumptions of Theorem 2 hold, we obtain assertions (i) and (ii).
(iii) Denote by =. From the proof of Theorem 2(iii), there exists is an accumulation point of . Next, we prove is a weakly efficient point. Suppose by the contrary that there exists such that . Substituting and into inequality (20), we have
(21) |
Summing the above inequality over to , we obtain
where the second inequality follows from . Multiplying by , we obtain
(22) |
Since , this together with imply
(23) |
where . However, as , the right hand side of inequality (22) tends to due to . This contradicts ineuqality (23). Therefore, is a weakly efficient point. In what follows, we prove the whole sequence converges to . Substituting and into inequality (20), we have
(24) |
The left hand side of the above inequality is nonnegative due to , then
This combines with and Lemma 9 imply converges. On the other hand, is an accumulation point of . It follows from Assumption 2(i) that converges to . Therefore, by Assumption 2(ii), we obtain sequence converges to .
(iv) Summing inequality (24) over to , we obtain
where the second inequality follows from . Multiplying by , we obtain assertion (iv).
(v) Assertion (v) follows directly from assertion (iv).
4.2 Linear convergence
This subsection is devoted to discuss the linear convergence of Algorithm 1. Fristly, we present the following results under the assumption that is -strongly convex relative to for .
Theorem 4.
Proof.
Since -strong convexity is sharper than -convexity, the assertions (i)-(iii) hold directly from Theorem 3.
Next, we show the Q-linear convergence result of with vector Bregman-PL inequality.
Theorem 5.
5 Conclusions
We extended the relative smoothness and relative strong convexity to vector-valued functions in the context of VOPs. Then we presented convergence rates for the interior Bregman gradient method using the generalized assumptions. In addition, we also devised a vector Bregman-PL inequality, which was used to prove linear convergence of the proposed method. It is worth noting that the convergence results extended the ones in com to non-Lipschitz gradient continuous and non-strongly convex case.
References
- (1) Marler, R. T., Arora, J. S.: Survey of multi-objective optimization methods for engineering. Struct. Multidisc. Optim. 26(6), 369-395 (2004)
- (2) Tapia, M., Coello, C.: Applications of multi-objective evolutionary algorithms in economics and finance: A survey. IEEE Congress on Evolutionary Computation. 532-539 (2007)
- (3) Fliege, J., Werner, R.: Robust multiobjective optimization & applications in portfolio optimization. Eur. J. Oper. Res. 234(2), 422-433 (2014)
- (4) Sener, O., Koltun, V.: Multi-task learning as multiobjective optimization. Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), 527-538 (2018)
- (5) Mahapatra, D., Rajan, V.: Multi-task learning with user preferences: gradient descent with controlled ascent in Pareto optimization. Proc. Int. Conf. Mach. Learn. (ICML), 119, 6597-6601 (2020)
- (6) Ye, F., Lin, B., Yue, Z., Guo, P., Xiao, Q., Zhang, Y.: Multi-objective meta learning. Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), 2021
- (7) Miettinen, K.: Nonlinear Multiobjective Optimization. Kluwer Academic Publishers, Boston, Massachusetts (1999)
- (8) Luc, T. D.: Theory of Vector Optimization, Lecture Notes in Economics and Mathematical Systems. Springer-Verlag, Berlin (1989)
- (9) John, J.: Vector Optimization: Theory, Applications and Extensions. 2nd ed. Springer, Berlin (2011)
- (10) Fliege, J., Svaiter, B. F.: Steepest descent methods for multicriteria optimization. Math. Method. Oper. Res. 51(3), 479-494 (2000)
- (11) Drummond, L. M., Svaiter, B. F.: A steepest descent method for vector optimization problems. J. Comput. Appl. Math. 175, 395–414 (2005)
- (12) Drummong, L. M. G., Iusem, A. N.: A projected gradient method for vector optimization problems. Comput. Optim. Appl. 28(1), 5-29 (2004)
- (13) Bonnel, H., Iusem, A. N., Svaiter, B. F.: Proximal methods in vector optimization. SIAM J. Optim. 15(4), 953-970 (2005)
- (14) Lucambio Pérez, L. R., Prudente, L. F.: Nonlinear conjugate gradient methods for vector optimization. SIAM J. Optim. 20(3), 2690-2720 (2018)
- (15) Bento, G. C., Cruz Neto, J. X., López, G., Soubeyran, A., Souza, J. C. O.: The proximal point method for locally Lipschitz functions in multiobjective optimization with application to the compromise problem. SIAM J. Optim. 28(2), 1104-1120 (2018)
- (16) Tanabe, H., Fukuda, E. H., Yamashita, N.: Proximal gradient methods for multiobjective optimization and their applications. Comput. Optim. Appl. 72, 339-361 (2019)
- (17) Chen, J., Tang, L. P., Yang, X. M.: A Barzilai-Borwein descent method for multiobjective optimization problems. arXiv:2204.08616 (2022)
- (18) Fliege, J., Vaz, A. I. F., Vicente, L. N.: Complexity of gradient descent for multiobjective optimization. Optim. Methods Softw. 34(5), 949-959 (2019)
- (19) Tanabe, H., Fukuda, E. H., Yamashita, N.: Convergence rates analysis of a multiobjective proximal gradient method. Optim. Lett. (2022)
- (20) Bauschke, H. H., Bolte, J., Teboulle, M.: A descent lemma beyond Lipschitz gradient continuity: first-order methods revisited and applications. Math. Oper. Res. 42(2), 330-348 (2016)
- (21) Lu, H., Freund, Robert M., Nesterov, Y.: Relatively smooth convex optimization by first-order methods, and applications. SIAM J. Optim. 28(1), 333-354 (2018)
- (22) Teboulle, M.: A simplified view of first order methods for optimization. Math. Program. 170, 67-96 (2018)
- (23) Bauschke, H. H., Bolte, J., Chen, J., Teboulle, M., Wang, X.: On linear convergence of non-Euclidean gradient methods without strong convexity and Lipschitz gradient continuity. J. Optim. Theory Appl. 182, 1068-1087 (2019)
- (24) Auslender, A., Teboulle, M.: Interior gradient and proximal methods for convex and conic optimization. SIAM J. Optim. 16(3), 697-725 (2006)
- (25) Villacorta, K. D. V., Oliveira, P. R.: An interior proximal method in vector optimization. Eur. J. Oper. Res. 214(3), 485-492 (2011)
- (26) Chen, Z., Huang, X. X., Yang, X. Q.: Generalized proximal point algorithms for multiobjective optimization problems. Appl. Anal. 90(6), 935-949 (2011)
- (27) Souza, Carlos de O., et al.: The proximal point method with a vectorial Bregman regularization in multiobjective DC programming. Optim. 71(1), 263-284 (2022)
- (28) Ansary, M. A. T., Dutta, J.: A proximal gradient method for multi-objective optimization problems using Bregman functions. optimization-online (2022)
- (29) Rockafellar, R. T.: Convex Analysis. Princeton University Press, Princeton (1970)
- (30) Bauschke, H. H., Borwein, J. M.: Joint and separate convexity of the Bregman distance. In: Butnariu, D., Censor, Y., Reich, S. (eds.) Inherently Parallel Algorithms in Feasibility and Optimization and their Applications (Haifa 2000), pp. 23-36. Elsevier, Amsterdam (2001)
- (31) Chen, W., Yang, X. M., Zhao, Y.: Conditional gradient method for vector optimization. arXiv:2109.11296 (2022)
- (32) Rockafellar, R. T., Wets, R. J. B.: Variational Analysis, Grundlehren Math. Wiss. 317, Springer, Berlin (1998)
- (33) Sion, M.: On general minimax theorems. Pac. J. Math., 8(1), 171-176 (1958)
- (34) Moudden, M. E., Mouatasim, A. E.: Accelerated diagonal steepest descent method for unconstrained multiobjective optimization. J. Optim. Theory Appl. 188(1), 220-242 (2021)
- (35) Chen, G., Teboulle, M.: Convergence analysis of a proximal-like minimization algorithm using Bregman functions. SIAM J. Optim. 3, 538-543 (1993)
- (36) Polyak, B. T.: Introduction to Optimization. Optimization Software, Inc., New York (1987)