Local saddles of relaxed averaged alternating reflections algorithms on phase retrieval
Abstract
Phase retrieval can be expressed as a non-convex constrained optimization problem to identify one phase minimizer one a torus. Many iterative transform techniques have been proposed to identify the minimizer, e.g., relaxed averaged alternating reflections(RAAR) algorithms. In this paper, we present one optimization viewpoint on the RAAR algorithm. RAAR algorithm is one alternating direction method of multipliers(ADMM) with one penalty parameter. Pairing with multipliers (dual vectors), phase vectors on the primal space are lifted to higher dimensional vectors, RAAR algorithm is one continuation algorithm, which searches for local saddles in the primal-dual space. The dual iteration approximates one gradient ascent flow, which drives the corresponding local minimizers in a positive-definite Hessian region. Altering penalty parameters, the RAAR avoids the stagnation of these corresponding local minimizers in the primal space and thus screens out many stationary points corresponding to non-local minimizers.
Keywords: Phase retrieval, relaxed averaged alternating reflections, alternating direction method of multipliers, Nash equilibrium, local saddles
1 Introduction
Phase retrieval has recently attracted attentions in the mathematics community (see one review [Shechtman2015] and references therein). The problem of phase retrieval is motivated by the inability of photo detectors to directly measure the phase of an electromagnetic wave at frequencies of THz (terahertz) and higher. The problem of phase retrieval aims to reconstruct an unknown object from its magnitude measurement data , where represents some isometric matrix and represents the Hermitian adjoint of . Introduce one non-convex -dimensional torus associated with its normalized torus The whole problem is equivalent to reconstructing the missing phase information and the unknown object via solving the constrained least squares problem \beqq min_x∈\IC^n, —u—=1 { ∥b⊙u-A^* x∥^2: u∈\IC^N}= min_z∈\cZ ∥A_⊥z∥^2, \eeqqwhere is an isometric matrix with unitary matrix ,
Here, represents the component-wise multiplication between two vectors , respectively. The isometric condition is not very restrictive in applications, since Fourier transforms are commonly applied in phase retrieval. Even for non-Fourier transforms, we can still obtain equivalent problems via a QR-factorization, see [Chen2017]. Let denote the set in consisting of all the local minimizers of (1). A vector minimizes (1) is called a global solution. In the noiseless measurement case, or for some and some . Numerically, it is a nontrivial task to obtain a global minimizer on the non-convex torus. The error reduction method is one traditional method [Gerchberg1972], which could produce a local solution of poor quality for (1), if no proper initialization is taken. During last decades, researchers propose various spectral initialization algorithms to overcome this challenge[NetrapalliEtAl2013Phase, ChenCandes2015Solving, CandesEtAl2015Phase, NIPS2016_83adc922, Chen2017, Chen2017a, LuoEtAl2018Optimal, LuLi2017Phase, Mondelli2019, duchi2019solving]. On the other hand, phase retrieval can be also tackled by another class of algorithms, including the well-known hybrid input-output algorithm(HIO)[Fienup1982, Fienup2013], the hybrid projection–reflection method[Bauschke2003], Fourier Douglas-Rachford algorithm (FDR)[CHEN2018665], alternating direction methods[Wen_2012] and relaxed averaged alternating reflections(RAAR) algorithms[Luke_2004]. An important feature of these algorithms is the empirical ability to avoid local minima and converge to a global mini-mum for noise-free oversampled diffraction patterns. For instance, the empirical study of FDR indicates the disappearance of the stagnation at poor local solutions under sufficiently many random masks. A limit point of FDR is a global solution in (1) and the limit point with appropriate spectral gap conditions reconstructs the phase retrieval solution [CHEN2018665]. Traditional convergence study on Douglas-Rachford splitting algorithm [Eckstein1992, He2015] heavily relies on the convexity assumption. Noise-free measurement is a strict requirement for HIO and FDR, which motivates the proposal of relaxed averaged alternating reflections algorithm [Luke_2004, Li]. Let denote the sets and , respectively. Let and denote the projector on and , respectively. Let denote the reflectors corresponding to . With one parameter relaxing the original feasibility problem (the intersection of and ), the -RAAR algorithm [Luke_2004] is defined as the iterations for some initialization ,
(1) | |||||
(2) | |||||
(3) |
Fourier Douglas-Rachford algorithm can be deemed as an extreme case of -RAAR family with . As RAAR converges to a fixed point , we could retrieve the phase information for (1). Any in yields a fixed point . Empirically, RAAR fixed points can produce local solutions of high quality, if a large value is properly chosen for , as reported in [Wen_2012, Li].
In this work, we disclose the relation between RAAR and the local minimizers in (1). As HIO can be reformulated as one alternating direction method of multipliers in [Wen_2012], we identify RAAR as one ADMM with penalty parameter applied to the constrained optimization problem in (1), e.g., Theorem. 2.3. This perspective links -RAAR with a small parameter to multiplier methods with large penalty . It is known in optimization that convergence of a multiplier method relies on a sufficiently large penalty (e.g., see Prop. 2.7 in [Bertsekas1996]). From this perspective, it is not surprising that the convergence of RAAR to its fixed point also requires a large penalty parameter. Actually, large penalty has been employed to ensure various ADMM iterations converging to stationary points [Hong2016, li2015global, wang2019global]. For instance, ADMM [wang2019global] is applied to solve the minimization of nonconvex nonsmooth functions. Global convergence to a stationary point can be established, when sufficiently large penalty parameters are used.
Saddle plays a fundamental role in the theory and the application of convex optimization[Bert], in particular, the convergence of ADMM, e.g., [Boyd]. For the application on phase retrieval, Sun et al[Sun] conduct saddle analysis on a quatradic objective function of Gaussian measurements. The geometric analysis shows that with high probability the global solution is the one local minimizer, when is sufficiently large. Most of critical points are saddles at actually. We believe that saddle analysis is also one key ingredient in explaining the avoidance of undesired critical points for the Lagrangian of RAAR. To some extent, promising empirical performance of non-convex ADMM conveys the impression that saddles exist in the Lagrangian function, which is not evident in the context of phase retrieval. This is a motivation of the current study. Recently, researchers have been cognizant of the importance of saddle structure in non-convex optimization research. Analysis of critical points in non-concave-convex problems leads to many interesting results in various applications. For instance, Lee et al. used a dynamical system approach to show that many gradient-descent algorithms almost surely converge to local minimizers with random initialization, even though they can get stuck at critical points theoretically [Lee2016, NIPS2017_f79921bb, jin2017escape]. The terminology “local saddle” is a crucial concept in understanding the min-max algorithm employed in modern machine learning research, e.g., gradient descent-ascent algorithms in generative adversarial networks (GANs)[goodfellow2020generative] and multi-agent reinforcement learning[omidshafiei2017deep]. With proper Hessian adjustment, [Adolphs2018] and [Daskalakis2018] proposed novel saddle algorithms to escape undesired critical points and to reach local saddles of min-max problems almost surely with random initialization. Jin et al. [Jin2020] proposed one non-symmetric definition of local saddles to address one basic question, “ what is a proper definition of local optima for the local saddle?” Later, Dai and Zhang gave saddle analysis on the constraint minimization problems [Dai2020].
Our study starts with one characterization of all the fixed point of RAAR algorithms in Theorem 2.2. These fixed points are critical points of (1). By varying , some of the fixed points become “local saddles” of a concave-nonconvex function , \beqq max_λmin_z {F(z,λ; β):=(β2 ∥A_⊥(z-λ)∥^2-12 ∥λ∥^2 ), z∈\cZ, λ∈\IC^N}. \eeqqTo characterize RAAR iterates, we investigate saddles in (1) lying in a high dimensional primal-dual space. Our study aims to answer a few intuitive questions, whether these local dimensional critical points on in the primal space can be lift to local saddles of (1) in a primal-dual space under some spectral gap condition, and how the ADMM iterates avoid or converge to these local saddles under a proper penalty parameter? The line of thought motivates the current study on local saddles of (1). Unfortunately, the definition of local saddles in [Jin2020] can not be employed to analyze the RAAR convergence, since the objective function in phase retrieval shares phase invariance, see Remark LABEL:alpha-79.
The main goal of the present work is to establish an optimization view to illustrate the convergence of RAAR, and show by analysis and numerics, under the framework for phase retrieval with coded diffraction patterns in [Fannjiang_2012], RAAR has a basin of attraction at a local saddle . For noiseless measurement, is a strictly local minimizer of (1). In practice, numerical stagnation of RAAR on noiseless measurements disappears under sufficient large values. Specifically, Theorem 2.3 shows that RAAR is actually one ADMM to solve the constrained problem in (1). Based on this identification, Theorem LABEL:Main1 show that each limit of RAAR iterates can be viewed as a “local saddle” of in (1).
The rest of the paper is organized as follows. In section 2, we examine the fixed point condition of RAAR algorithm. By identifying RAAR as ADMM, we disclose the concave-non-convex function for the dynamics of RAAR, which provides a continuation viewpoint on RAAR iteration. In section LABEL:sec:3, we present one proper definition for local saddles and show the existence of local saddles for oversampled coded diffraction patterns. In section LABEL:sec:4, we show the convergence of RAAR to a local saddle under a sufficiently large parameter. Last, we provide experiments to illustrate the behavior of RAAR, (i)comparison experiments between RAAR and Douglas Rachford splitting proposed in [fannjiang2020fixed]; (ii) applications of RAAR on coded diffraction patterns.
2 RAAR algorithms
2.1 Critical points
The following gives the first order optimality of the problem in (1). This is a special case of Prop. LABEL:convex1 with . We skip the proof. {prop} Let be a local minimizer of the problem in (1). Let . Then the first-order optimality condition is \beqq q_0:=z_0^-1⊙(A_⊥^* A_⊥z_0)∈\IR^N, \eeqqand the second-order necessary condition is that for all , \beqq ξ^⊤(K_z_0^⊥-\diag(q_0))ξ≥0. \eeqq
Once a local solution is obtained, the unknown object of phase retrieval in (1) can be estimated by . On the other hand, using , we can express the first order condition as \beqqu^-1⊙(A^*A(b⊙u))=b⊙(1-q_0)∈\IR^N.\eeqqUsing canonical vectors of , we have a componentwise lower bound on from (2.1): . In general, there exists many local minimizers on , satisfying (2.1) and (2.1).
2.2 Fixed point conditions of RAAR
We begin with fixed point conditions of -RAAR iterations in (3). For each , introduce one auxiliary parameter defined by \beqq β=β’ 1+β’, i.e., β’=β1-β. \eeqqWe shall show the reduction in the cardinality of fixed points under with a small penalty parameter. {theorem} Consider the application of the -RAAR algorithm on the constrained problem in (1). Write in polar form . For each , let and \beqq c:= (1-1-ββ) b+1-ββ —w—∈\IR^N. \eeqqThen is a fixed point of -RAAR, if and only if satisfies the phase condition \beqq A^*A(b⊙u)=c⊙u, \eeqqand the magnitude condition, \beqq —w—=β’ c+ (1-β’)b≥0, i.e., c≥(1-β’^-1) b. \eeqqIn particular, for , we have from (2.2). Observe that the inequality in (2.2) ensures the well-defined magnitude vector . Hence, the fixed points are critical points of (1).
Rearranging (3), we obtain the fixed point condition of RAAR, \beqq ((1-β)—w—-(1-2β)b)⊙w—w—=β A^*A { (2b-—w—)⊙w—w—}. \eeqqEquivalently, taking the projections and on (2.2) yield
(4) | |||
(5) |
For the only-if part, let be a fixed point of RAAR with (4,5). With the definition of , (4) gives \beqq A^*A ((c-b)⊙u)=β’A^*A((—w—-b)⊙u)=0, \eeqqand (5) gives \beqqA_⊥^* A_⊥(c⊙w—w—)=A_⊥^* A_⊥({ b-β’^-1 (b-—w—)}⊙w—w—)=0,\eeqqwhich implies in the range of . Together with (2.2), we have (2.2). Also, (2.2) is the result of the non-negativeness of in (2.2).
To verify the if-part, we need to show that constructed from a phase vector satisfying (2.2) and a magnitude vector satisfying (2.2) meets (4,5). From (2.2) and (2.2), we have (4), i.e., \beqqA^*A((b-—w—)⊙u)=A^*A{ β’(b-c)⊙u}=β’{c⊙u-c⊙u}=0. \eeqqWith the aid of (4, 2.2), the fixed point condition in (5) is ensured by the computation: \beqq(I-A^*A){(b-β’^-1(b-—w—))⊙u} =(I-A^*A){c⊙u}=0. \eeqq
Finally, the condition in (2.2) is identical to the first optimality condition in (2.1). Hence, the fixed points must be critical points of (1). Theorem 2.2 indicates that each fixed point can be re-parameterized by satisfying (2.2) and (2.2). The condition in (2.2) always hold for sufficiently small.
The first order optimality in (2.1) yields that the phase condition in (2.2) is actually the critical point condition of in (1). Fix one critical point and let be the corresponding vector given from (2.2). From Theorem 2.2, given from the polar form with (2.2) is a fixed point of -RAAR, if satisfies the condition in (2.2). To further examine (2.2), we parameterize the fixed point by . Let denote the threshold vector, where
The fixed point condition in (2.2) indicates that gives a fixed point of -RAAR with any . That is, the corresponding parameter must exceed the threshold vector, \beqq (β’)^-1=1-ββ≥b^-1⊙(K^⊥b). \eeqqSince can be viewed as one penalty parameter in the associated Lagrangian in (2.3), we call (2.2) the penalty-threshold condition of RAAR fixed points. In general, the cardinality of RAAR fixed points decreases under a large parameter . See Fig. 1.
For , RAAR reduces to FDR, whose fixed point satisfies and thus . When phase retrieval has uniqueness property, gives the reconstruction. On the other hand, for , (3) gives \beqq S(w)=A^*A (b⊙w—w—)+12(I-A^*A) w. \eeqqSuppose a RAAR initialization is chosen from the range of . The second term in (2.2) always vanishes and thus RAAR iterations reduce to alternating projection iterations(AP) in [Chen2017]. From this perspective, one can regard -RAAR as one family of algorithms interpolating AP and FDR, varying from to .
[width=0.4]beta_fig.eps
2.3 Alternative directions method of multipliers
Next, we present one relation between RAAR and the alternating direction method of multipliers (ADMM). The alternating direction method of multipliers was originally introduced in the 1970s[Glowinski1975, Gabay1976] and can be regarded as an approximation of the augmented Lagrangian method, whose primal update step is replaced by one iteration of the alternating minimization. Although ADMM is classified as one first-order method, practically ADMM could produce a solution with modest accuracy within a reasonable amount of time. Due to the algorithm simplicity, nowadays this approach is popular in many applications, in particular, applications of nonsmooth optimalication. See [Boyd, yang2009, Deka2019] and the references therein.
Use the standard inner product
To solve the problem in (1), introduce one auxiliary variable with one constraint and one associated multiplier , and form the Lagrangian function with some parameter in (2.2), \beqq β’2 ∥A_⊥y∥^2+¡ λ, y-z¿+12∥y-z∥^2, y∈\IC^N, z∈\cZ. \eeqqEquivalently, we have the augmented Lagrangian function, \beqq L(y,z, λ):= 12 ∥A_⊥y∥^2+ β’^-1¡ λ, y-z¿+12β’∥y-z∥^2, \eeqqwhen we identify as a penalty parameter. To solve , ADMM starts with some initialization and and generates the sequence with stepsize , according to rules,
Introducing one projection operator on , for . Algebraic computation yields
(6) | |||
(7) | |||
(8) |
From the -update in (6), one reconstruction for the unknown object in (1) can be computed by \beqq x=Ay=A(I-βA_⊥^* A_⊥)(z-λ)=A(z-λ). \eeqq
Theorem 2.3 indicates that RAAR is actually an ADMM with proper initialization applied to the problem in (1). In general, the step size of the dual vector should be chosen properly to ensure the convergence. The following shows the relation between RAAR and the ADMM with . Hence, we shall focus in this paper.
Consider one -RAAR iteration with nonzero initialization . Let Generate one ADMM sequence with dual step size , according to (6, 7, 8) with initialization . Construct a sequence from , \beqq w’_k:=y_k+1+λ_k=(I-βA_⊥^* A_⊥) z_k+βA_⊥^* A_⊥λ_k. \eeqqThen the sequence is exactly the -RAAR sequence, i.e., for all .