Fair Principal Component Analysis and Filter Design
Abstract
We consider Fair Principal Component Analysis (FPCA) and search for a low dimensional subspace that spans multiple target vectors in a fair manner. FPCA is defined as a non-concave maximization of the worst projected target norm within a given set. The problem arises in filter design in signal processing, and when incorporating fairness into dimensionality reduction schemes. The state of the art approach to FPCA is via semidefinite programming followed by rank reduction methods. Instead, we propose to address FPCA using simple sub-gradient descent. We analyze the landscape of the underlying optimization in the case of orthogonal targets. We prove that the landscape is benign and that all local minima are globally optimal. Interestingly, the SDR approach leads to sub-optimal solutions in this orthogonal case. Finally, we discuss the equivalence between orthogonal FPCA and the design of normalized tight frames.
Index Terms:
Dimensionality Reduction, SDP, Fairness, Normalized Tight Frame, PCAI Introduction
Dimensionality reduction is a fundamental problem in signal processing and machine learning. In particular, Principal Component Analysis (PCA) is among the most popular data science tool. It involves a non-concave maximization but has a tight semidefinite relaxation (SDR). Its optimization landscape, saddle points and extreme points are all well understood and it is routinely solved using scalable first order methods [1]. PCA maximizes the average performance across a given set of vector targets. In many settings, worst case metrics are preferred in order to ensure fairness and equal performance across all targets. This gives rise to Fair PCA (FPCA) [2] which will be defined formally in the next section. Unfortunately, changing the average PCA objective to a worst case FPCA objective results in an NP-hard problem [2] which is poorly understood. There is a growing body of works on convex relaxations via SDR for FPCA [2, 3], but these methods do not scale well and are inapplicable to many realistic settings. Therefore, the goal of this paper to consider scalable first order solutions to FPCA and shed more light on the landscape of this important optimization problem.
Due to the significance of PCA it is non-trivial to track the origins of FPCA. In the context of filter design, FPCA with rank one constraints is known as multicast beamforming and there is a huge body of literature on this topic, e.g., [4, 5, 6]. In the modern context of fairness in machine learning, FPCA was considered in [7, 8, 2, 3]. It was shown that SDR with an iterative rounding technique provides near optimal performance when the rank is much larger than the squared root of the number of targets. More generally, by interpreting the worst case operator as an norm, FPCA is a special case of norm optimizations. Classical PCA corresponds to . Robust PCA algorithms as [9] rely on , and FPCA is the other extreme using . Most of these works capitalize on the use of SDR that leads to conic optimizations with provable performance guarantees. Finally, [10] proposed a different definition to fairness in PCA via multi-objective optimization. They developed a first order method for attaining random solutions on the PCA Pareto frontier. FPCA defined above may be considered as one of these points.
SDR and nuclear norm relaxations are currently state of the art in a wide range of subspace recovery problems. Unfortunately, SDR is known to scale poorly to high dimensions. Therefore, there is a growing body of works on first order solutions to semidefinite programs. The main trick is to factorize the low rank matrix and show that the landscape of the resulting non-convex objective is benign [11, 12, 13, 14, 15, 16, 17]. The SDR of FPCA involves two types of linear matrix inequalities and still poses a challenge. Therefore, we first reformulate the problem and then apply sub-gradient descent on the factorized formulation.
The main contribution of this paper is the observation that the landscape of the factorized FPCA optimization is benign when the targets are orthogonal. This is the case in which dimensionality reduction is most lossy. Yet, we show that it is easy from an optimization perspective. The maximization is non-concave but every (non-connected) local minima is globally optimal. Surprisingly, we show that this case is challenging for SDR. Its objective is tight but it is not trivial to project its solution onto the feasible set. Numerical experiments with synthetic data suggest that these properties also hold in more realistic near-orthogonal settings. Finally, a direct corollary of our analysis is an equivalence between orthogonal FPCA and the design of finite normalized tight frames [18]. This characterization may be useful in future works on data-driven normalized tight frame design.
Notations:
We used bold uppercase letters (e.g. P) for matrices, bold lowercase letters (e.g. ) for vectors and non-bold letters (e.g. ) for scalars. We used pythonic notation for indices of matrices: for the ’th row, for the ’th column and for the ’th entry of matrix. The set of () semi-orthogonal matrices (matrices with orthonormal columns) is denoted by , the set of positive semidefinite matrices by , and the set of projection matrices of rank at most by (and ). Given a function , we define the set of indices . Finally we define a projection operator onto the set of projection matrices of rank at most : as follows: Let (EVD decomposition), where: then: .
II Problem formulation
The goal of this paper is to identify a low dimensional subspace that maximizes the smallest norm of a given set of projected targets. More specifically, let be the set of targets, we consider the problem:
(1) |
Our motivation to FPCA arises in the context of filter design for detection. We are interested in the design of a linear sampling device from to that will allow detection of known targets denoted by . The motivation for using a small rank is that the cost of power, space and/or time resources typically scales with . Detection accuracy in additive white Gaussian noise decreases exponentially with the received signal to noise ratio (SNR), and it is therefore natural to maximize the worst SNR across all the targets. Hopefully, this will lead to a fair solution with equal norms for all the targets.
FPCA with is concerned with the design of a single beaamforming filter, and is equivalent to multicast downlink transmit beamforming [4, 5]
(2) |
Practical systems typically satisfy , e.g., the design of a few expensive sensors that downsample a high resolution digital signal (or even an infinite dimension analog signal). Without loss of optimality, we assume a first stage of dimensionality reduction via PCA that results in effective dimensions such that .
As detailed in the introduction, FPCA was also recently introduced in the context of fair machine learning. There, it is more natural to assume a block structure. The targets are divided into blocks, denoted by matrices , and fairness needs to be respected with respect to properties as gender or race [7, 2, 3]:
(3) |
Throughout this paper, we will consider the simpler non-block FPCA formulation corresponding to filter design. Preliminary experiments suggest that most of the results also hold in the block case.
FPCA is known to be NP-hard [4, 2]. The state of the art approach to FPCA is SDR. Specifically, we relax the rank constraint by its convex hull, the nuclear norm, and the projection constraint by linear matrix inequalities [5, 2]. This yields the SDP:
(4) |
The computational complexity of solving an SDR using an Interior Point method is intractable for most applications, but [2] propose a practical and efficient multiplicative weight update. Unfortunately, the optimal solution to SDR might not be a feasible projection, and may have any rank. Due to the relaxation, SDR always results in an upper bound on FPCA. To obtain a feasible approximation, it is customary to define
(5) |
PrSDR is a feasible projection matrix of rank , and is therefore a lower bound on FPCA. Better approximations may be obtained via randomized procedures [5]. Recently, an iterative rounding technique was proven to provide a approximation [2]. This result is near optimal in the block case where it is reasonable to assume . It is less applicable to filter design where is large and smaller ranks are required.
The goal of this paper is to provide a scalable, yet accurate solution to FPCA, without the need for additional rank reduction schemes. Motivated by the growing success of simple gradient based methods in complex optimization problems, e.g., deep learning, we consider the application of sub-gradient descent to FPCA and analyze its optimization landscape.
III Algorithm
In this section, we propose an alternative and more scalable approach for solving SDR. The two optimization challenges in FPCA are the projection and rank constraints. We confront the first challenge by reformulating the problem using a quadratic objective, and the second by decomposing the projection matrix using its low rank factors. Together, we define factorized FPCA:
(6) |
where
(7) |
Proposition 1.
Before proving the proposition, we note that the projection is only needed in order to handle a degenerate case in which the dimension of the subspace spanned by the targets is smaller than . Typically, this projection is not needed and is feasible.
Proof.
We rely on the observation that F-FPCA has an optimal solution with orthogonal matrix, and for orthogonal matrix we have:
In addition the function is a surjective function from to , so the optimization over both sets is equivalent. More details are available in the Appendix. ∎
The advantage of solving F-FPCA rather than FPCA is that it forces a low rank solution via an unconstrained optimization. A member of the sub-gradient of F-FPCA objective can be computed in . In particular, Algorithm 1 describes a promising sub-gradient descent method for its minimization.
The obvious downside of using F-FPCA is its non-convexity that may cause descent algorithms to converge to bad stationary points. Its convergence analysis is more difficult due to the non-smooth maximum function. Nonetheless, in the next section, we prove that there are no bad local minima when the targets are orthogonal. This is also demonstrated in the experimental section where we show the advantages of F-FPCA in terms of accuracy.
Relation to other low rank optimization papers: We note in passing that there is a large body of literature on global optimality properties of low rank optimizations [16, 17]. These provide sufficient conditions for convergence to global optimum in factorized formulations, e.g., Restricted Strong Convexity and Smoothness. Observe that these guarantees require the existence of a low rank optimal solution in the original problem. These conditions do not hold in FPCA, and therefore our analysis below takes a different approach.
IV Analysis - the orthogonal case
In this section, we analyze the FPCA in the special case of orthogonal targets. As explained, FPCA is NP-hard and we do not expect a scalable and accurate solution for arbitray targets. Interestingly, our analysis shows that the problem becomes significantly easier when the targets are orthogonal. This is the case for example when the targets are randomly generated and the number of targets is much smaller than their dimension.
We will use the following assumptions:
-
A1:
The targets are orthogonal vectors.
-
A2:
The problem is not degenerate in the sense that
where
(the harmonic mean of the squared norms of ).
Assumption A1 is the main property that simplifies the landscape and allows a tractable solution and analysis. On the other hand, assumption A2 is a technical condition that prevents a trivial degenerate solution based on the norms of the targets.
Using these assumptions, we have the following results.
Proposition 2.
Under assumptions A1-A2, any local minimizer of F-FPCA is a global maximizer of FPCA and FPCA.
Proof.
The proof consists of the following lemmas (proofs in the appendix):
Lemma 1.
Under assumptions A1-A2, let be a local minimizer of F-FPCA, then .
Lemma 2.
Under assumptions A1-A2, let a local minimizer of F-FPCA, then: for all .
Intuitively, if the property in Lemma 2 is violated, then can be infinitesimally changed in a manner that decreases the correlation of with some direction such that for all . We can decrease the value of for some without harming the objective function using a sequence of Givens rotations with respect to the pairs for each . After decreasing for all the objective will also be decreased.
Finally, in order to prove global optimality we define:
(8) |
If :
We have:
Rearranging yields . Together with the equivalence in Proposition 1 we conclude that all local minima yield an identical objective of which is globally optimal.∎
Proposition 2 justifies the use of Algorithm 1 when the targets are orthogonal. Numerical results in the next section suggest that bad local minima are rare even in more realistic near-orthogonal scenarios.
Given the favourable properties of F-FPCA in the orthogonal case, it is interesting to analyze the performance of SDR in this case.
Proposition 3.
Under assumptions A1-A2, SDR is tight and its optimal objective value is
However, the optimal solution may be full rank and infeasible for FPCA.
Proof.
See Appendix. ∎
Observe that the rank constraint is hard, and a rank reduction procedure such as PrSDR is necessary for finding a feasible solution. The iterative rounding algorithm of [2] relies on finding an extreme point solution, and guarantees an upper bound on its rank. The bound is not always tight. For example, their algorithm fails to find an optimal low rank solution in the orthogonal case. On the other hand, Algorithm 1 easily finds the global solution.
Finally, we conclude this section by noting an interesting relation between FPCA with orthogonal targets and the design of Finite Tight Frames [18]. Recall the following definition:
Definition 1.
.
-
•
Let . If then is frame for .
-
•
A frame is tight with frame bound A if :
-
•
A frame is a ’Normalized Tight Frame’ if is tight frame and for all .
A straight forward consequence is the following result.
Corollary 1.
Under assumptions A1-A2, if is an optimal solution for F-FPCA, then is a tight frame. In particular, if the targets are the standard basis, then is a normalized tight frame.
Sketch of proof (the proof in the appendix): As proved before, the solution of F-FPCA is in and the transposition of any is a tight frame. The second part is true since the optimal solution of F-FPCA is satisfied for all k: . For the standard basis we get for all : i.e. the norm of all rows of are equals.
It is well known that normalized tight frames can be derived as minimizers of frame potential functions [18]. The corollary provides an alternative derivation via FPCA with different targets . Depending on the properties of the targets, this allows a flexible data-driven design that will be pursued in future work.
V Experimental results
In this section, we illustrate the efficacy of the different algorithms using numerical experiments. We compare the following competitors:
- •
-
•
PIRSDR - the projection of SDR onto the feasible set using eigenvalue decomposition. Before project the solution, the iterative rounding rank reduction from [2] was performed.
-
•
F-FPCA - the solution to (6) via Algorithm 1 with a random initialization.
-
•
F-FPCAi - the solution to (6) via Algorithm 1 with PIRSDR initialization.
To allow easy comparison, we normalize the results by the value of SDR, so that a ratio of corresponds to a tight solution.
V-A Synthetic simulations
We begin with experiments on synthetic targets with independent, zero mean, unit variance, Gaussian random variables. This is clearly a simplistic setting but it allows control over the different parameters , and . Each of the experiments was performed times and we report the average performance.
Rank effect: The first experiment is presented in Fig. 1 and illustrates the dependency on the rank . It is easy to see that even with very small rank, the gap between the upper and lower bounds vanishes. We conclude that in this non-orthogonal setting, the landscape of FPCA is benign as long as the rank is not very small.
Orthogonality effect: The second experiment is presented in Fig. 2 and addresses the effect of orthogonality. As explained, the targets are drawn randomly and they tend to orthogonality as increases. Our analysis proved that the gap should vanish when the targets are exactly orthogonal. The numerical results suggest that this is also true for more realistic and near-orthogonal targets. The optimality gap clearly decreases as increases.
V-B Minerals dataset
In order to illustrate the performance in a more realistic setting we also considered a real world dataset. We consider the design of hyperspectral matched filters for detection of known minerals. We downloaded spectral signatures of minerals from the Spectral Library of United States Geological Survey (USGS). We experimented with different minerals, each with bands in the range . Some of the measurements were missing and their corresponding bands were omitted. We then performed PCA and reduced the dimension from to . These vectors were normalized and then centered. Fig. 3 provides the signatures of the first minerals before and after the pre-processing. Finally, we performed fair dimension reduction to . Fig. 4 summarizes the quality of the approximation of the different algorithms. As before, it is easy to see that F-FPCA is near optimal at very small ranks. Interestingly, PIRSDR is beneficial as an initialization but shows inferior and non-monotonic performance on its own. As expected, all the algorithms easily attain the optimal performance at higher values of .
V-C Credit dataset
Next, we continue to the credit dataset [21] that was considered for block-FPCA in [7, 2, 3]. Following these works, we consider the functions
(9) |
where and is the objective of the standard PCA function of which is independent of . The results in Fig 6 are identical to those in [2] with our additional F-FPCA algorithm. PIRSDR achieves the SDR lower bound at all ranks excepts . Remarkably, F-FPCA is optimal in this specific setting and attains the bound for all without exception. Apparently, the landscape of the credit dataset is benign. We emphasize that this is pure luck and we can easily find other non-orthogonal examples with spurious local minima. In real applications, we recommend running F-FPCA using multiple initializations and choosing the best solution.
VI Conclusion
In this paper, we suggested to tackle the problem of fairness in linear dimension reduction by simply using first order methods over a non-convex optimization problem. We provided an analysis of the landscape of this problem in the special case where the targets are orthogonal to each other. We also provided experimental results which support our approach by showing that sub gradient descent is successful also in the near orthogonal case and real world data.
There are many interesting extensions to this paper that are worth pursuing. Analysis of the near-orthogonal case is still an open question. In addition, a drawback of our approach is the non smoothness of the landscape which might prevent the use of standard convergence bounds for first order methods. This can be treated by approximating the in our formulation by log-sum-exp or norm for functions. Experimental results show that our results can be extended to the block case that is more relevant to machine learning. Finally, we only considered the case of classical linear dimension reduction. Future work may focus on extensions to non-linear methods and tensor decompositions.
Appendix A Proof of Proposition 1
Let a truncated EVD decomposition of , then:
Observe that this function is minimized when for all , so:
So F-FPCA is equivalent to the following problem (over the orthogonal matrices):
Now for any orthogonal matrix we get:
Finally, observe that:
-
•
is a feasible solution for the problem above iff is a feasible solution for FPCA.
-
•
The objective function of FPCA in is equal to the objective function of the problem above in (multiplied by ).
So we conclude that the problems are equivalent.
Appendix B Proof of Lemma 1
We begin with the following lemma:
Lemma 3.
Let . If A2 holds then .
Proof.
Assume in contradiction that there exists () such that: , and let . We get for all :
Now recall the definition of in (8) and observe that:
This means that A2 does not hold. ∎
We will now show that if is not orthogonal, then we can decrease either the size of or the value of by choosing an arbitrarily close .
Lemma 4.
Let , then for any there exists a such that:
-
1.
.
-
2.
Either , or .
Proof.
Let an EVD decomposition of , then:
Due to , there is an such that , and an such that . Observe that: has a local minimum only in . Therefore, define where:
and we get for all such that .
If there exists an such that and then we are done.
Otherwise, pick some with , and after the procedure above take and define the projection of onto (by Lemma 3 ). Define by adding to the singular vector of and define . Now we get that for all :
Similarly, for we get:
as required. ∎
Appendix C Proof of Lemma 2
We begin with the following lemma that states that we can utilize the orthogonality of the targets in order to infinitesimally change in a manner that increases the value of for some , decreases the value of for some and does not change the value of for .
Lemma 5.
Let such that there exist with . Then, there exists an such that: for any there exist such that:
-
1.
.
-
2.
-
3.
.
-
4.
Proof.
Define , a Given Rotation (for some ) over the axes, i.e.:
along with two orthogonal vectors
(11) |
and an orthonormal basis for their orthogonal complement in : . Now define: and we get:
1+2 is true, since:
are continuous functions.
3 is true, since:
For all thus and:
In order to show 4 we use the equality in (12) (In the next page, proof is in the appendix, since it is quite technical):
(12) |
Now, if :
-
•
If: then for any : .
-
•
If: then for any : .
and since , for small enough we get: .
On the other hand, By Lemma 3 , so if then:
∎
Appendix D Proof of Proposition 3
Proof.
Given , and recall the definition of in (8). In the SDR problem we solve:
where and we have used the orthogonality of .
Now, define as a diagonal matrix with .
It is easy to verify that is feasible and yields an objective of . Now let such that , then:
so is not feasible and we conclude that is optimal for SDR. By Proposition 2 the optimal value for F-FPCA is so by Proposition 1 the value of FPCA is so SDR=FPCA (but this solution might not be low rank, as in our positive definite construction). ∎
Appendix E Proof of Corollary 1
We start the proof of the proposition by the observation that tight frame is characterized by the standard basis:
Lemma 6.
A frame is tight with frame bound A if and only if in the standard basis of :
Proof.
Observe that if we have: than we get for all :
∎
Now we use the observation above to claim that tight frame is actual the transposition of semi orthogonal matrix:
Lemma 7.
Let . is a tight frame with frame bound iff has orthogonal columns with norm .
Proof.
Consider equality 1 below:
Observe that iff has orthogonal columns with norm . Equality 1 also holds iff which holds iff is a tight frame with frame bound (by Lemma 6), so we conclude that the conditions are equivalent.
∎
Lemma 8.
If for all k: , then: .
Proof.
Now, let an optimal solution for F-FPCA, by Lemma 1 the columns of are orthonormal, so by Lemma 7 is tight frame and by Proposition 2 we have for all :
On the other hand, let a tight frame as above, by Lemma 7 the columns of are orthogonal and have the same norm. By Lemma 8 so the columns has unit norms, i.e. the columns are orthonormal and for all :
i.e. which is the optimal target. Finally, if , then F-FPCA is reduced to the problem of finding ’normalized tight frame’.
Acknowledgment
The authors would like to thank Uri Okon who initiated this research and defined the problem, as well as Gal Elidan. This work was partially supported by ISF grant 1339/15.
References
- [1] J. Sun, Q. Qu, and J. Wright, “When are nonconvex problems not scary?” arXiv preprint arXiv:1510.06096, 2015.
- [2] J. Morgenstern, S. Samadi, M. Singh, U. Tantipongpipat, and S. Vempala, “Fair dimensionality reduction and iterative rounding for sdps,” arXiv preprint arXiv:1902.11281, 2019.
- [3] S. Samadi, U. Tantipongpipat, J. H. Morgenstern, M. Singh, and S. Vempala, “The price of fair pca: One extra dimension,” in Advances in Neural Information Processing Systems, 2018, pp. 10 976–10 987.
- [4] N. D. Sidiropoulos, T. N. Davidson, and Z.-Q. Luo, “Transmit beamforming for physical-layer multicasting,” IEEE Trans. Signal Processing, vol. 54, no. 6-1, pp. 2239–2251, 2006.
- [5] W.-K. K. Ma, “Semidefinite relaxation of quadratic optimization problems and applications,” IEEE Signal Processing Magazine, vol. 1053, no. 5888/10, 2010.
- [6] A. Cheriyadat and L. M. Bruce, “Why principal component analysis is not an appropriate feature extraction method for hyperspectral data,” in IGARSS 2003. 2003 IEEE International Geoscience and Remote Sensing Symposium. Proceedings (IEEE Cat. No. 03CH37477), vol. 6. IEEE, 2003, pp. 3420–3422.
- [7] M. Olfat and A. Aswani, “Convex formulations for fair principal component analysis,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, 2019, pp. 663–670.
- [8] W. Bian and D. Tao, “Max-min distance analysis by using sequential sdp relaxation for dimension reduction,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 33, no. 5, pp. 1037–1050, 2010.
- [9] G. Lerman, M. B. McCoy, J. A. Tropp, and T. Zhang, “Robust computation of linear models by convex relaxation,” Foundations of Computational Mathematics, vol. 15, no. 2, pp. 363–410, 2015.
- [10] M. M. Kamani, F. Haddadpour, R. Forsati, and M. Mahdavi, “Efficient fair principal component analysis,” arXiv preprint arXiv:1911.04931, 2019.
- [11] S. Burer and R. D. Monteiro, “A nonlinear programming algorithm for solving semidefinite programs via low-rank factorization,” Mathematical Programming, vol. 95, no. 2, pp. 329–357, 2003.
- [12] ——, “Local minima and convergence in low-rank semidefinite programming,” Mathematical Programming, vol. 103, no. 3, pp. 427–444, 2005.
- [13] N. Boumal, V. Voroninski, and A. S. Bandeira, “Deterministic guarantees for burer-monteiro factorizations of smooth semidefinite programs,” Communications on Pure and Applied Mathematics, 2018.
- [14] N. Boumal, V. Voroninski, and A. Bandeira, “The non-convex burer-monteiro approach works on smooth semidefinite programs,” in Advances in Neural Information Processing Systems, 2016, pp. 2757–2765.
- [15] D. Cifuentes, “Burer-monteiro guarantees for general semidefinite programs,” arXiv preprint arXiv:1904.07147, 2019.
- [16] Z. Zhu, Q. Li, G. Tang, and M. B. Wakin, “Global optimality in low-rank matrix optimization,” IEEE Transactions on Signal Processing, vol. 66, no. 13, pp. 3614–3628, 2018.
- [17] Q. Li, Z. Zhu, and G. Tang, “The non-convex geometry of low-rank matrix optimization,” Information and Inference: A Journal of the IMA, vol. 8, no. 1, pp. 51–96, 2018.
- [18] J. J. Benedetto and M. Fickus, “Finite normalized tight frames,” Advances in Computational Mathematics, vol. 18, no. 2-4, pp. 357–385, 2003.
- [19] S. Diamond and S. Boyd, “CVXPY: A Python-embedded modeling language for convex optimization,” Journal of Machine Learning Research, vol. 17, no. 83, pp. 1–5, 2016.
- [20] A. Agrawal, R. Verschueren, S. Diamond, and S. Boyd, “A rewriting system for convex optimization problems,” Journal of Control and Decision, vol. 5, no. 1, pp. 42–60, 2018.
- [21] I.-C. Yeh and C.-h. Lien, “The comparisons of data mining techniques for the predictive accuracy of probability of default of credit card clients,” Expert Systems with Applications, vol. 36, no. 2, pp. 2473–2480, 2009.
Appendix F Proof of equation (11)
Lemma 9.
Let and assume . Define:
for some and complete these vectors to orthonormal basis: , and define: , then:
Proof.
So finally:
. ∎
Lemma 10.
Let , (a Given rotation over the axes i,j), .
Proof.
Observe that:
Since , we also have:
So:
∎