On the Robustness of Cross-Concentrated Sampling for Matrix Completion
††thanks: The authors contributed equally. This work was partially supported by NSF DMS 2304489.
(Corresponding authors: HanQin Cai and Longxiu Huang.)
Abstract
Matrix completion is one of the crucial tools in modern data science research. Recently, a novel sampling model for matrix completion coined cross-concentrated sampling (CCS) has caught much attention. However, the robustness of the CCS model against sparse outliers remains unclear in the existing studies. In this paper, we aim to answer this question by exploring a novel Robust CCS Completion problem. A highly efficient non-convex iterative algorithm, dubbed Robust CUR Completion (RCURC), is proposed. The empirical performance of the proposed algorithm, in terms of both efficiency and robustness, is verified in synthetic and real datasets.
Index Terms:
Robust matrix completion, cross-concentrated sampling, CUR decomposition, outlier detection.I Introduction
Matrix completion problem [1, 2] aims to recover the underlying low-rank matrix from some entry-wise partial observation. It has received tremendous attention in the past decades and has been widely applied in real-world applications such as collaborative filtering [3, 4], image processing [5, 6], and signal processing [7, 8]. While the vanilla matrix completion studies are often based on entry-wise uniform or Bernoulli sampling models, the recent development of a novel sampling model called Cross-Concentrated Sampling (CCS) [9] has caught the attention of the matrix completion community. The CCS model is summarized in Procedure 1.
As illustrated in Figure 1, the CCS model bridges two popular sampling models for matrix completion: Uniform sampling and CUR sampling which recovers the underlying low-rank matrix from observed rows and columns. The CCS model provides extra flexibility, with a theoretical guideline, for samples concentrated on certain rows and columns. This model has the potential for more cost- and time-efficient sampling strategies in applications such as recommendation systems. However, data in real-world applications are often corrupted by sparse outliers, and uniformly sampled data must be processed by some robust completion algorithms for reliable recovery results. That being said, a crucial question has to be answered before the CCS model can be widely used: “Is CCS-based matrix completion robust to sparse outliers under some robust algorithms, like the uniform sampling model?”
Mathematically, provided the partial observation of a corrupted data matrix , we study the Robust CCS Completion problem:
(1) |
where is the sampling operator defined as:
and the observation set is generated according to the CCS model. Similar to the standard robust matrix completion, a sparse variable is introduced for better outlier tolerance, thus one can recover the underlying low-rank matrix accurately.




I-A Related Work
The problem of low-rank matrix recovery in the presence of sparse outliers has been well-studied under the settings of uniform sampling and Bernoulli sampling. This problem is known as robust principal component analysis (RPCA) when the corrupted data matrix is fully observed, and it is called robust matrix completion (RMC) when data is partially observed. The seminal work [10] considers both RPCA and RMC problems via convex relaxed formulations and provides recovery guarantees. In particular, under the -incoherence assumption for the low-rank , [10] requires the positions of outliers to be placed uniformly at random, and at least entries are observed uniformly at random. Later, a series of non-convex algorithms [11, 12, 13, 14, 15, 16] tackle RPCA and/or RMC problems with an improved, non-random -sparsity assumptions for the outlier matrix . The typical recovery guarantee shows a linear convergence of a non-convex algorithm, provided ; moreover, random samples are typically required for the RMC cases. Another line of work [17, 7, 18, 19, 8] focuses on the robust recovery of structured low-rank matrices, e.g., Hankel matrices, and they typically require merely samples by utilizing the structure, even in the presence of structured outliers. More recently, [20, 21, 22] study the robust CUR decomposition problem, that is, recovering the low-rank matrix from row- and column-wise observations with entry-wise corruptions. The studies show that robust CUR decomposition is as robust as RPCA, with better computational efficiency and data interoperability.
On the other hand, [9] shows that CCS-based matrix completion requires samples which is only a factor of worse than the state-of-the-art result; however, its outlier tolerance has not been studied.
I-B Notation
For a matrix , , , , and denote its -th entry, its row submatrix with row indices , its column submatrix with column indices , and its submatrix with row indices and column indices , respectively. denotes the Frobenius norm, denotes the largest row-wise -norm, denote the largest entry-wise magnitude, represents the Moore–Penrose inverse of , and is the transpose of . denotes the Frobenius inner product. We denote as the -th standard basis of the real vector space. The symbol denotes the set for all . Throughout the paper, uniform sampling is referred to as uniform sampling with replacement.
II Preliminaries
II-A Assumptions
As discussed in the related work, -incoherence and -sparsity are commonly used assumptions in RMC problems. We will use the same assumptions in this study and give the formal definition here:
Assumption 1 (-incoherence).
Let be a rank- matrix. It is said -incoherent if
for some positive , where is the compact SVD of .
Assumption 2 (-sparsity).
is said -sparse if there are at most fraction of non-zero entries in each of its row and column. That is, for all and , it holds
Note that no randomness is required for the sparsity pattern.
II-B CUR Approximation
CUR approximation, also known as skeleton decomposition, is the backbone of our algorithm design. We shall go over some background of CUR approximation for the reader’s interest. CUR approximation is a self-expressive technique that fills the interpretability gap in matrix dimensionality reduction. Given a rank- matrix , if we select rows and columns from that span its row and column spaces respectively, then can be reconstructed from these submatrices. This has been affirmed in previous studies:
Theorem 3 ([23]).
Consider row and column index sets . Denote submatrices , and . If , then .
For further details of CUR approximation, including the historical context and formal proof, one can refer to [23] and reference therein. Note that through various row- and column-wise sampling methodologies, meeting this rank equivalence condition in Theorem 3 is highly probable, especially when a sufficient number of rows and columns are sampled. An example of this is detailed in Theorem 4.
Theorem 4 ([24, Theorem 1.1]).
Let satisfy 1, and suppose we sample rows and columns uniformly at random. Then with probability at least .
There also exist many other sampling methodologies that ensure the recoverability of CUR approximation; for instance, the Discrete Empirical Interpolation Method (DEIM) [25, 26] and leverage scores sampling method [27]. The interested reader can find more details in [28, Table 1]. Nonetheless, they all sample completed rows and columns, thus are substantially different than the CCS model.
III Proposed Algorithm
In this section, we will present a novel non-convex algorithm for solving the proposed Robust CCS Completion problem (1). This iterative algorithm is based on the idea of projected gradient descent where CUR approximation is used for fast low-rank approximation in each iteration. The algorithm is dubbed Robust CUR Completion (RCURC) and summarized in Algorithm 2.
We will go over the algorithm step by step in the following paragraphs. For the ease of the presentation, we start with the low-rank component.
Updating low-rank component. To utilize the structure of cross-concentrated samples, it is efficient to enforce the low-rank constraint of with the CUR approximation technique. Let , , and . Applying gradient descent directly on and gives:
where and are the stepsizes. However, When it comes to the intersection submatrix , it is more complicated as and can have overlaps. We abuse the notation and define an operator called union sum here:
Basically, we take whatever value we have for the non-overlapped entries and take a weighted average for the overlaps where the weights are determined by the stepsizes used in the updates of and . To ensure the rank- constraint, at least one of , or should be rank-. For computational efficiency, we choose to put it on the smallest one. Thus,
where is the best rank- approximation operator via truncated SVD. After replacing the intersection part in the previously updated and , we have the new estimation of low-rank component:
(2) |
However, (2) is just a conceptual step and one should never compute it. In fact, the full matrix is never needed and should not be formed in the algorithm as updating the corresponding CUR components is sufficient.
Updating sparse component. We detect the outliers and put them into the sparse matrix via hard-thresholding operator:
The hard-thresholding on residue , paired with iterative decayed thresholding values:
has shown promising performance in outlier detection in prior art [14, 16, 29, 30]. Notice that we only need to remove outliers located on the selected rows and columns, i.e., and , since they are the only components needed to update the low-rank component later. Therefore, for computational efficiency, we should only compute on the selected rows and columns to update correspondingly—as said, one should never compute the full in this algorithm. In particular,
Stopping criteria and stepsizes. We finish the algorithm with the last few pieces. The stopping criteria is set to be where is the targeted accuracy and the computational error is
(3) |
The recommended stepsizes are and where and are the observation rates of and respectively. Smaller stepsizes should be used with larger , i.e., more outliers.
IV Numerical Experiments
In this section, we will verify the empirical performance of RCURC with both synthetic and real datasets. All experiments are implemented on Matlab R2022b and executed on a laptop equipped with Intel i7-11800H CPU (2.3GHz @ 8 cores) and 16GB DDR4 RAM.
IV-A Synthetic Datasets
In this simulation, we assess the computational efficiency of our algorithm, RCURC, in addressing the robust CCS completion problem. We construct , a matrix with , where is a randomly generated rank- matrix. To create the sparse outlier tensor , we randomly select percent entries to form the support of . The values of the non-zero entries are then uniformly sampled from the range . To generate the robust CCS completion problems, we set , , and . The results are obtained by averaging over runs and reported in Figure 2. Both figures in Figure 2 depict the relationship between the relative error and computational time for our RCURC method with varying rank and outlier amplification factor . It is noteworthy that RCURC consistently achieves nearly linear convergence rates across different scenarios.


IV-B Video Background Subtraction
We have applied our robust CCS model and RCURC to the problem of background separation. Two popular videos, shoppingmall and restaurant, are used as benchmarks. The dimensions of the datasets are documented in Table I.
To evaluate the performance of our method, we apply IRCUR [31] to the entire video dataset. We use the background derived from IRCUR as the benchmark ground truth. Subsequently, we randomly select of the rows and of the columns to create subrow and subcolumn matrices. At the next stage, we generate observations on these submatrices. The reconstructed backgrounds obtained using RCURC are illustrated in Figure 3. It is evident that the background recovered by RCURC closely resembles that obtained via IRCUR. Furthermore, we utilize the Peak Signal-to-Noise Ratio (PSNR) as a quantitative measure to evaluate the reconstruction quality of RCURC in comparison to the results achieved with IRCUR. These findings are detailed in Table I.
Frame Size | Frame Number | Runtime | PSNR | |
---|---|---|---|---|
s | ||||
s |








V Conclusion Remarks
This paper introduces a novel mathematical model for robust matrix completion problems with cross-concentrated samples. A highly efficient non-convex algorithm, dubbed RCURC, has been developed for the proposed model. The key techniques are projected gradient descent and CUR approximation. The numerical experiments, with both synthetic and real datasets, show high potential. In particular, we consistently observe linear convergence on RCURC.
As for future work, we will study the statistical properties of the proposed robust CCS completion model, such as theoretical sample complexities and outlier tolerance. The recovery guarantee with a linear convergence rate will also be established for RCURC. We will also explore other real-world applications that suit the proposed model.
References
- [1] E. J. Candès and B. Recht, “Exact matrix completion via convex optimization,” Found. Comput. Math., vol. 9, no. 6, pp. 717–772, 2009.
- [2] B. Recht, “A simpler approach to matrix completion.” J. Mach. Learn. Res., vol. 12, no. 12, 2011.
- [3] J. Bennett, C. Elkan, B. Liu, P. Smyth, and D. Tikk, “KDD cup and workshop 2007,” SIGKDD Explor. Newsl., vol. 9, no. 2, p. 51–52, 2007.
- [4] D. Goldberg, D. Nichols, B. M. Oki, and D. Terry, “Using collaborative filtering to weave an information tapestry,” Commun. ACM, vol. 35, no. 12, pp. 61–70, 1992.
- [5] P. Chen and D. Suter, “Recovering the missing components in a large noisy low-rank matrix: Application to sfm,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 26, no. 8, pp. 1051–1063, 2004.
- [6] Y. Hu, D. Zhang, J. Ye, X. Li, and X. He, “Fast and accurate matrix completion via truncated nuclear norm regularization,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 35, no. 9, pp. 2117–2130, 2012.
- [7] J.-F. Cai, T. Wang, and K. Wei, “Fast and provable algorithms for spectrally sparse signal reconstruction via low-rank Hankel matrix completion,” Appl. Comput. Harmon. Anal., vol. 46, no. 1, pp. 94–121, 2019.
- [8] H. Cai, J.-F. Cai, and J. You, “Structured gradient descent for fast robust low-rank Hankel matrix completion,” SIAM J. Sci. Comput., vol. 45, no. 3, pp. A1172–A1198, 2023.
- [9] H. Cai, L. Huang, P. Li, and D. Needell, “Matrix completion with cross-concentrated sampling: Bridging uniform sampling and CUR sampling,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 45, no. 8, pp. 10 100–10 113, 2023.
- [10] E. J. Candès, X. Li, Y. Ma, and J. Wright, “Robust principal component analysis?” Journal of the ACM, vol. 58, no. 3, pp. 1–37, 2011.
- [11] P. Netrapalli, U. Niranjan, S. Sanghavi, A. Anandkumar, and P. Jain, “Non-convex robust PCA,” in Adv. Neural Inf. Process Syst., 2014, pp. 1107–1115.
- [12] X. Yi, D. Park, Y. Chen, and C. Caramanis, “Fast algorithms for robust PCA via gradient descent,” in Adv. Neural Inf. Process Syst., 2016, pp. 4152–4160.
- [13] T. Zhang and Y. Yang, “Robust pca by manifold optimization,” J. Mach. Learn. Res., vol. 19, no. 1, pp. 3101–3139, 2018.
- [14] H. Cai, J.-F. Cai, and K. Wei, “Accelerated alternating projections for robust principal component analysis,” J. Mach. Learn. Res., vol. 20, no. 1, pp. 685–717, 2019.
- [15] T. Tong, C. Ma, and Y. Chi, “Accelerating ill-conditioned low-rank matrix estimation via scaled gradient descent,” J. Mach. Learn. Res., vol. 22, no. 150, pp. 1–63, 2021.
- [16] H. Cai, J. Liu, and W. Yin, “Learned robust PCA: A scalable deep unfolding approach for high-dimensional outlier detection,” in Adv. Neural Inf. Process Syst., vol. 34, 2021, pp. 16 977–16 989.
- [17] Y. Chen and Y. Chi, “Robust spectral compressed sensing via structured matrix completion,” IEEE Trans. Inf. Theory, vol. 60, no. 10, pp. 6576–6601, 2014.
- [18] S. Zhang and M. Wang, “Correction of corrupted columns through fast robust hankel matrix completion,” IEEE Trans. Signal Process., vol. 67, no. 10, pp. 2580–2594, 2019.
- [19] H. Cai, J.-F. Cai, T. Wang, and G. Yin, “Accelerated structured alternating projections for robust spectrally sparse signal recovery,” IEEE Trans. Signal Process., vol. 69, pp. 809–821, 2021.
- [20] H. Cai, K. Hamm, L. Huang, J. Li, and T. Wang, “Rapid robust principal component analysis: CUR accelerated inexact low rank estimation,” IEEE Signal Process. Lett., vol. 28, pp. 116–120, 2021.
- [21] H. Cai, K. Hamm, L. Huang, and D. Needell, “Robust CUR decomposition: Theory and imaging applications,” SIAM J. Imaging Sci., vol. 14, no. 4, pp. 1472–1503, 2021.
- [22] K. Hamm, M. Meskini, and H. Cai, “Riemannian CUR decompositions for robust principal component analysis,” in Topological, Algebraic and Geometric Learning Workshops, 2022, pp. 152–160.
- [23] K. Hamm and L. Huang, “Perspectives on CUR decompositions,” Appl. Comput. Harmon. Anal., vol. 48, no. 3, pp. 1088–1099, 2020.
- [24] J. Chiu and L. Demanet, “Sublinear randomized algorithms for skeleton decompositions,” SIAM J. Matrix Anal. Appl., vol. 34, no. 3, pp. 1361–1383, 2013.
- [25] S. Chaturantabut and D. C. Sorensen, “Nonlinear model reduction via discrete empirical interpolation,” SIAM J. Sci. Comput., vol. 32, no. 5, pp. 2737–2764, 2010.
- [26] D. C. Sorensen and M. Embree, “A DEIM induced CUR factorization,” SIAM J. Sci. Comput., vol. 38, no. 3, pp. A1454–A1482, 2016.
- [27] P. Drineas, M. W. Mahoney, and S. Muthukrishnan, “Relative-error CUR matrix decompositions,” SIAM J. Matrix Anal. Appl., vol. 30, no. 2, pp. 844–881, 2008.
- [28] K. Hamm and L. Huang, “Stability of sampling for CUR decompositions,” Found. Data Sci., vol. 2, no. 2, pp. 83–99, 2020.
- [29] H. Cai, Z. Chao, L. Huang, and D. Needell, “Fast robust tensor principal component analysis via fiber CUR decomposition,” in Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, 2021, pp. 189–197.
- [30] ——, “Robust tensor CUR decompositions: Rapid low-tucker-rank tensor recovery with sparse corruption,” SIAM J. Imaging Sci., 2024.
- [31] H. Cai, K. Hamm, L. Huang, J. Li, and T. Wang, “Rapid robust principal component analysis: CUR accelerated inexact low rank estimation,” IEEE Signal Process. Lett., vol. 28, pp. 116–120, 2020.