Third-Order Statistics Reconstruction from Compressive Measurements
Abstract
Estimation of third-order statistics relies on the availability of a huge amount of data records, which can pose severe challenges on the data collecting hardware in terms of considerable storage costs, overwhelming energy consumption, and unaffordably high sampling rate especially when dealing with high-dimensional data such as wideband signals. To overcome these challenges, this paper focuses on the reconstruction of the third-order cumulants under the compressive sensing framework. Specifically, this paper derives a transformed linear system that directly connects the cross-cumulants of compressive measurements to the desired third-order statistics. We provide sufficient conditions for lossless third-order cumulant reconstruction via solving simple least-squares, and quantify the strongest achievable compression ratio. To reduce the computational burden, we also propose an approach to recover diagonal cumulant slices directly from compressive measurements, which is useful when the cumulant slices are sufficient for the inference task at hand. As testified by extensive simulations, the developed joint sampling and reconstruction approaches to third-order statistics estimation are able to reduce the required sampling rates significantly by exploiting the cumulant structure resulting from signal stationarity, even in the absence of any sparsity constraints on the signal or cumulants.
Index Terms:
Higher-order statistics, cumulants, moments, compressive sampling, dense sampler, sparse samplerI Introduction
High-dimensional signals are ubiquitously featured in contemporary systems, which provide rich information for enhancing statistical inference performance, but at the same time pose unprecedented challenges to traditional signal processing theories and methods. As the signal bandwidth increases significantly, the sampling rate for digital conversion ruled by the Shannon-Nyquist sampling theorem can be exceedingly high in the wideband regime and hence easily reach the limit of analog-to-digital converter (ADC) hardware which is a main bottleneck of many current signal processing systems. In order to alleviate the burden on ADCs, compressed sensing (CS) is shown to be able to accurately recover signals from sub-Nyquist-rate samples by capitalizing on signal sparsity in the spectrum or the edge spectrum [3] [4] [5]. Closely related works can also be found in the so-called spectrum blind sampling (SBS) framework [6] [7] where analog-to-digital sub-Nyquist samplers are designed for spectrum reconstruction under the sparse spectrum assumption.
Another line of compressive sampling research advocates direct second-order statistics (SOS) reconstruction from sub-Nyquist samples even in the absence of signal sparsity. In [8] [9] [10], a compressive covariance sensing (CCS) framework is proposed to recover sparse cyclic spectrum from compressive measurements for cyclostationary processes. Taking stationary signals as a special case therein, it shows that lossless power spectrum recovery is possible even for non-sparse signals by capitalizing on the Toeplitz structure of covariances. For improved reconstruction accuracy, generalizations are considered in [11] [12], where all significant lags of cross-correlation functions among the compressive outputs are exploited to induce an overdetermined system for non-sparse cyclic spectrum. In [12] [13] [14], sampler design for CCS is formulated into a minimal sparse ruler problem with the goal of achieving the strongest compression. Other sparse sampler designs with closed-form expressions can be found in [15] [16], in which two uniform sub-Nyquist-rate samplers are exploited for sinusoids frequency estimation based on the co-array concept and properties of coprime numbers.
For non-Gaussian signals appeared in many real-world applications, higher-order statistics (HOS) preserve non-trivial information that are absent in SOS [17] [18] [19]. Such information can be used in many tasks that cannot be accomplished using SOS only, such as suppressing additive colored Gaussian noise, blindly identifying non-minimum phase systems, to name just a few. In practice, accurate HOS estimation hinges on the availability of a huge amount of data records, which poses challenges for data acquisition. Besides the exceedingly high cost in data storage, the sampler consumes substantial energy in order to collect a huge amount of data over a long sampling time, which is unaffordable for instance in power-hungry sensor networks. These challenges become even pronounced when dealing with wideband signals because high sampling rate requirements not only overburden ADCs but also consume a large amount of power. In view of these challenges, it is well motivated to develop compressive sampling techniques for HOS estimation.
Despite of the well-documented merits of HOS in various signal processing tasks, the literature on HOS estimation under the compressive sampling framework is scant. In [20], third-order cumulants of compressive measurements are calculated and directly used for spectrum sensing tasks, but it does not concern the reconstruction of uncompressed third-order cumulants. In [21], estimating higher-order cyclostationary features from compressive samples is studied by taking into account of sparsity in the frequency domain. However, [21] recovers only a part of cumulant values instead of the entire cumulant tensor and does not generalize to cases in the absence of sparsity assumption. To reduce the computational complexity, the generalization of co-prime sampling for HOS estimation is studied in [22]. In [23], sparse representation of HOS is explored. However, the sampler design, the HOS recovery guarantees and the achievable sampling rate reduction are not considered in [22] and [23].
In this paper, we put forward a systematic and simple approach to the estimation of third-order statistics under the compressive sampling framework with recovery guarantees. Different from existing compressed sensing techniques for recovering one-dimensional deterministic signals, the collected compressive samples do not have a direct linear relationship with the third-order statistics to be recovered. To circumvent this difficulty, this paper derives a transformed linear system that connects the cross-cumulants of compressive outputs to the desired third-order statistics. Theoretically, the proposed approach can losslessly recover the unknown third-order statistics of a wide-sense stationary signal via solving simple least-squares (LS). We analyze sufficient conditions for lossless third-order statistics reconstruction by using the proposed approach, and derive the strongest achievable compression ratio. Realizing that the diagonal cumulant slice is sufficient for many signal processing tasks, we also propose an approach to reconstruct arbitrary-order cumulant slices directly from compressive measurements without having to recover the entire cumulant tensor, which interestingly subsumes CCS as a special case. Extensive simulations demonstrate that the proposed approaches are able to reduce the required sampling rate significantly without imposing sparsity constraints on the signal or its cumulants.
The rest of the paper is organized as follows. Section II reviews some key definitions on third-order statistics. Section III gives the signal model and problem statement. Section IV describes the proposed compressive cumulant estimation approaches. Section V considers the symmetry property of third-order cumulants to further enhance the estimation performance. Section VI derives a compressive cumulant slice sensing approach to recover the diagonal cumulant slice directly. Corroborating simulation results are provided in Section VII, followed by a concluding summary in Section VIII.
II Preliminaries
This section introduces relevant notation and reviews some basic definitions of third-order statistics.
II-A Notation
We use lower case boldface letters, upper case boldface letters, and calligraphic boldface letters, such as , , , to denote vectors, matrices and tensors respectively. The operator stacks all columns (slices) of a matrix (tensor) into a large column vector. Calligraphic letters such as are used to denote sets, and represents the corresponding cardinality. We use to denote the convolution operator, the outer product, and the Kronecker product. Three-fold Kronecker product of vector is denoted as . We use to represent transposition and the Moore-Penrose pseudo inverse. We write the -norm of a vector as . We use to denote the identity matrix of size .
II-B Third-Order Statistics
This subsection reviews some basic definitions that are used in the ensuing sections [18].
Definition 1: For zero-mean real random variables , the third-order cumulant is given by:
(1) |
Definition 2: Let be a wide-sense stationary random process. The third-order cumulant of this process, denoted as , is defined as the joint third-order cumulant of random variables , i.e.,
(2) |
Because is stationary, its cumulant is dependent only on time-lags and while independent on the time origin .
Definition 3: Let be a non-stationary random process. Its third-order cumulant is defined as
(3) |
which is dependent on both time-lags and and the time origin because of the nonstationarity.
Definition 4: Let be a vector random process of dimension , i.e., . The third-order cross-cumulant of elements in is defined as
(4) |
Definition 5: Let be a zero-mean th-order stationary process. The th-order cumulant of this process, denoted by , is defined as the joint th-order cumulant of random variables ,
(5) | ||||
III System Model and Problem Statement
The third- and higher-order cumulants of real-valued stationary processes can be useful in many real-world applications such as real harmonics retrieval in the additive colored noise setting, quadratic phase coupling detection in nonlinear processes, and time delay estimation for source bearing and range calculation [17] [18]. Consider w.l.o.g. a real-valued wide-sense stationary analog signal which is bandlimited with bandwidth . For digital processing at a sub-Nyquist sampling rate, an analog to information converter (AIC) can be adopted [12], as shown in Fig. 1. Equipped with branches, the adopted sampler first modulates the signal with a periodic waveform of period in the th branch, and then convert the modulated signal in each branch to the digital version via an integrate-and-dump device operating with period . Through the AIC, a total of samples are collected for every seconds with , resulting in an effective sampling rate of that is times the Nyquist rate. The compression ratio determines the reduced sampling rate. Evidently, the output of the th branch at the th sampling instant is given by
(6) | ||||
where for and otherwise. By assuming that is a piecewise constant function having constant values in every interval of length , i.e., for for , where , the output (6) can be rewritten as
(7) | ||||
where is the Nyquist digital version of , which is not acquired in practical implementation for sake of reducing the sampling rate. The choice of depends on the sampler design. Random Gaussian sampler, similar to [24], is adopted in this paper, which will be detailed in the ensuing sections.
The sampling strategy in (7) can be alternatively interpreted as a two-step operation. It first convolves with a length- digital filter and then feeds the outcome to an -fold decimation operator to obtain the compressive outputs, i.e., , where
(8) |
It is worth noting that our system model can adopt any linear compressive sampler besides the AIC, for which (7) and (8) hold for some . The relationship is useful, which is illustrated via the AIC.
With the above model in mind, the overarching goal of this paper is to accurately reconstruct the third-order cumulant of based on the obtained sub-Nyquist-rate samples . The main contribution of this work is that all the different cross-cumulants for are leveraged to recover unknown variables for , which makes the reconstruction feasible from sub-Nyquist-rate samples even in the absence of sparsity constraints on or its cumulants. Attractively, such a reconstruction strategy can be formulated as an overdetermined linear system as analyzed next, leading to a unique solution via solving simple LS.
IV Third-Order Cumulants Reconstruction
In this section, we develop two compressive third-order cumulants sensing (C3CS) approaches.
IV-A The direct C3CS Approach
This subsection presents a direct C3CS approach for reconstruction from for all . Considering that , the third-order cross-cumulant of sub-Nyquist-rate samples and that of Nyquist-rate samples are related through:
(9) | ||||
Viewing as the convolution of and , can be obtained from a two-dimensional linear convolution [17]
(10) | ||||
where is the “deterministic” third-order cross-cumulant of , and in the form
(11) |
(12) | ||||
where the concise vector representation in the last equality is based on the following vector definitions:
(13) |
(14) |
(15) |
(16) |
We next stack all the cross-cumulant values for into a vector which according to (12) can be expressed as
(17) |
where matrices (for and ) are of size .
Although the bandlimitedness assumption on in the frequency domain enforces and thus to have unlimited support in the time domain, it is reasonable to assume that the supports of and are limited within a certain range because those and outside such a range can be negligible in many real-world applications. By stacking values and for respectively, we obtain vectors and as follows:
(18) | ||||
(19) | ||||
The reconstruction of can be achieved by taking advantage of all significant lags of cross-cumulants among the compressive outputs in (18) as long as the linear relationship between (18) and (19) is figured out. To that end, several key observations are in order. First, given the definition of in (13) and the fact that has finite support within , the support of should be limited within . As a result, the last entries in the vector are zero; for , the elements in the vector are all zero except those indexed by ; for , the elements in the vector are all zero except the first elements. Similarly, given the definition in (14) along with the fact that has limited support within , the columns of matrix indexed by are all zero; the first columns of matrix are all zero; the first columns and other columns of matrix that indexed by are all zero. By capitalizing on these crucial observations, we can rewrite the two-dimensional linear convolution in (17) into a one-dimensional circular convolution, which further allows us to figure out the relation between in (18) and in (19) as
(20) |
where is given by (21).
(21) |
Evidently, recovering boils down to solving the linear system (20), which is invertible if has full column rank.
By noting that in (21) is a block circulant matrix with blocks of size , the computational load and memory requirements for solving (20) can be greatly reduced. Specifically, it is known that block circulant matrices can be diagonalized by discrete Fourier transform (DFT) matrices [25]. By using the -point (inverse) discrete Fourier transform ((I)DFT), the block circulant matrix can be represented as
(22) |
where is the DFT matrix, and is a block diagonal matrix with each block being of size . Such a fact allows us to rewrite (20) as
(23) |
where the vectors and are given by
(24) |
(25) |
Plugging (18) and (19) into (24) and (25) yields
(26) |
(27) |
with and given by:
(28) |
(29) |
where and denotes a submatrix. The block structure in (26) and (27) enables us to decompose (23) into matrix equations:
(30) |
Note that given , the linear systems (30) for can be solved via LS in a parallel fashion, as long as has full column rank. Once is obtained, can be recovered based on (25).
To summarize the implementation of the proposed direct C3CS approach, firstly, is calculated based on (9) and (18) and is constructed based on (11) and (21); secondly, is obtained according to (24) and (26); and finally can be recovered by solving the linear system (23) or (30), which returns based on the relation (25).
IV-B The Alternative C3CS Approach
Different from the direct C3CS approach derived above, an alternative C3CS approach is presented in this subsection. First, by defining the measurement vectors and the vector sequence as
(31) |
(32) |
we rewrite (7) in the matrix-vector form
(33) |
where compressive sampling matrix is given by
(34) |
with .
Next, we compute the three-way cumulant tensor of in (31) by using tensor outer product . It is obvious that contains all third-order cumulants that can be calculated from . Similarly, we can also construct the cumulant tensor of in (32) as . By using multilinear algebra, the relationship between and can be expressed as
(35) |
where represents the th mode product between a tensor and a matrix [26]. By vectorizing , it is evident that (35) can be rewritten as
(36) | ||||
Thus far, the linear relationship between cross-cumulants of compressive samples and the cumulants of the uncompressed signal is explicitly figured out.
Alternatively, the third-order cumulants of and can also be calculated directly in the vector form as of dimension and of dimension . By using the linear transformation property of Kronecker product or equivalently [27], and are related by
(37) |
Actually, and , where and are defined in (18) and (19) respectively. Therefore, the linear systems (36) and (37) are equivalent.
Since of size is rank-deficient, the linear system in (37) is under-determined, and hence one cannot recover accurately from if not given any further constraints. Fortunately, it turns out that , or equivalently , can actually be represented in a subspace with much smaller dimension, which consequently transforms the seemingly under-determined system (37) to be an over-determined one. To be specific, by noticing that the entry amounts to according to (2), we can form an index set with elements for by removing repeated indices, and the cardinalty equals to the degrees of freedom contained in . It can be easily verified that . Since tensor with elements contains only degrees of freedom, can be condensed into a vector whose elements are and we can write
(38) |
where is a special mapping matrix. We will show how such deterministic mapping matrix can be constructed in the next section after accounting for further information.
Plugging (38) into (37) yields
(39) |
where is of dimension . Evidently, obtaining is equivalent to recovering because both of them contain the same set of values. The unique recovery of is guaranteed by the following theorem.
Theorem 1 (Sufficient Conditions): can be losslessly recovered from in the linear system (39) via solving simple LS, if one of the following two conditions holds:
1) The matrix is deterministic and full column rank.
2) The matrix is random (Guassian random matrix for example) and , such that 1) is satisfied with high probability.
Proof.
: It is straightforward to see that condition 1) is sufficient for lossless recovery of . We directly go to examine condition 2), which is more practical for the compressive sampling matrix design. To figure out if and when the linear system (39) is over-determined, we need to determined the degrees of freedom contained in . Observing the compressive sampling strategy (33), each element of is generated by filtering with an individual filter indicating by the corresponding row of , where are mutually incoherent with high probability if a random Guassian matrix is used. As a result, is not stationary, although is, which implies that the third-order cumulant of is not only dependent on time lags and , but also dependent on time origin . Therefore, the redundant values in come only from the inherent symmetry of the Kronecker product, e.g., [28]. In that case, the number of unique elements contained in is actually . In addition, the degrees of freedom in is as analyzed in (38). Consequently, we conclude that as long as , the linear system (39) is over-determined with high probability and a unique solution is permitted via solving simple LS. ∎
Theorem 1 provides sufficient conditions on matrices and to guarantee lossless recovery, and it should be noted that the elements of are determined by as shown in (34). That is to say, , or equivalently , should be carefully chosen according to Theorem 1 so that the mutual incoherence of matrix can be guaranteed with high probability. Besides, the above sufficient condition 2) fundamentally reveals the strongest achievable compression when using the proposed alternative C3CS approach. Specifically, the ratio reflects the saving of the sampling cost and the strongest compression is given when the equality in condition 2) holds, that is, for a given value of . A more straightforward illustration of compression efficiency will be provided in the next section. In general, we choose to be slightly greater than to satisfy condition 2), which results in the computational complexity of our method to be comparable to calculating cumulants directly from uncompressive measurements.
Remark 1 (comparison to the direct C3CS approach): Generally speaking, the direct C3CS approach is able to estimate more cumulant lags than the alternative C3CS approach. The reason is that the direct C3CS approach takes advantage of all inter-data-blocks cross-cumulants , of compressive outputs, whereas the alternative C3CS approach considers only cross-cumulants within one data block, i.e., . Once is chosen, all the cumulant values can be estimated. Obviously, the larger is chosen, the more cumulant values can be estimated, which, however, also increases computational complexity because more cumulant values need to be calculated.
Remark 2 (relation to CCS): The multilinear system (35) is closely related to CCS where cumulant tensors and reduce to covariance matrices and and they are related by . In CCS, it recovers by capitalizing on its Toeplitz structure. The similarity between CCS and our approach is that it is the stationarity that results in the special structures of and , and hence allows the reduction in the degrees of freedom to be estimated to achieve over-determined systems.
Remark 3 (generalization to higher-order statistics): Our proposed approach can be easily generalized to higher than third-order statistics. Specifically, the th-order moments of and can be arranged into -way tensors as and , and they are related through a multilinear system like (35). By capitalizing on the stationarity of , it turns out that the degrees of freedom contained in can be condensed into a vector of much smaller dimension like in (38). Then an over-determined system can be obtained similar to (39).
V Additional Symmetry Information
This section explores the use of symmetry information of third-order cumulants in order to improve both compression efficiency and estimation performance.
Symmetry Property [17]: The definition of the third-order cumulant of stationary random process in (2) implies an important symmetry property: .
By using these symmetry equations, the plane can be divided into six triangular regions. Giving the cumulants in any region, the cumulants in the other five regions can be determined. The principal region is as shown in the shadowed sector in Fig. 2, which is defined by . The degrees of freedom to be estimated are completely determined by the elements located in the principal region. Hence, for stationary random process , the number of unique values contained in is exactly . We stack these unique values in a vector denoted as follows
(40) | ||||
Evidently, obtaining the values of is equivalent to recovering , as long as the relationship between them can be figured out explicitly.
Relating to : Both to contain the same set of nonzero elements, and hence can be linearly related to each other as:
(41) |
where is a deterministic mapping matrix constructed as follows:
(42) |
where denotes the element in matrix , and and are given by
(43) | ||||
The derivation of is given in Appendix A.
Plugging (41) into (37) yields:
(44) |
where is of dimension . It is worth noting that contains less degree of freedom to be estimated than , and hence stronger compression and better reconstruction performance can be expected. The unique recovery of is guaranteed by the following theorem.
Theorem 2 (Sufficient Conditions): can be losslessly recovered from in the linear system (44) via solving simple LS, if one of the following two conditions holds:
1) The matrix is deterministic and full column rank.
2) The matrix is random (Guassian random matrix for example) and , such that 1) is satisfied with high probability.
The proof for Theorem 2 is similar to that for Theorem 1. Theoretically, the strongest compression is achieved when the equality holds in 2), say . Evidently, Theorem 2 provides stronger compression than Theorem 1 due to the proper use of symmetry information. To gain intuition on the compression ratio of our proposed approach, Fig. 3 shows how strongest achievable compression ratio varies as increases. We can see that the relation can be approximated by as increases, which suggests the strongest compression ratio
(45) |
The approximation (45) makes it evident that our proposed cumulant reconstruction approach can achieve significant compression even without any sparsity assumption. As is shown in Fig. 3, the strongest compression ratio of our proposed cumulant reconstruction approach can be even lower than when is greater than . We provide the achievable strongest compression ratio corresponding to a variety of values of and in Table LABEL:table:M/N.
N | The smallest M | Compression ratio M/N |
---|---|---|
20 | 11 | 0.550 |
40 | 17 | 0.425 |
80 | 27 | 0.338 |
160 | 43 | 0.269 |
320 | 68 | 0.212 |
Remark 4 (compression efficiency comparison): The achievable strongest compression (45) for the third-order cumulant reconstruction in C3CS is weaker than that for covariance reconstruction in CCS which is as given in [9]. This is intuitively understandable because the third-order statistics contain richer information than second-order statistics and thus weaker compression is permitted in order to preserve such useful information.
To summarize, the alternative C3CS approach accounting for cumulant symmetry information is composed of the compressive sampling matrix design according to Theorem 2 and a cumulant recovery method. Given , the mapping matrix and the predefined sampling matrix , the cumulant recovery problem can be formulated as a simple LS problem without any constraints for reconstructing as follows:
(46) |
The over-determined least-squares formulation in (46) leads to a closed-form solution. Although sparsity is not entailed in C3CS, the sparsity prior, if exists, can also be properly incorporated to enhance the recovery performance. In addition, a key observation is that the matrix is extremely sparse with each row having only one non-zero element, which facilitates the solving of (46) in large scale settings.
VI Diagonal Cumulant Slice Reconstruction
In many signal processing applications, the 1-D slice of the higher-order cumulant, defined as , is sufficient for inference tasks therein, such as harmonics retrieval [29], channel estimation [30], system identification [31], to name just a few. In that case, it would be desirable to recover the 1-D cumulant slice only instead of all the cumulant values for the purpose of reducing computational complexity. To that end, this section proposes an approach termed compressive cumulant slice sensing (CCSS) to reconstruct the 1-D diagonal cumulant slice directly from compressive measurements. In CCSS, a sparse sampling strategy is adopted, which links the compressive outputs and the 1-D diagonal slice of cumulants explicitly demonstrating the possibility of lossless recovery of the 1-D diagonal slice straightforwardly. In general, CCSS is able to achieve significant compression for arbitrary-order cumulant diagonal slice recovery.
To motivate the adopted sparse sampling strategy in CCSS, we first introduce the structure embedded in the 1-D diagonal cumulant slice of the th-order cumulant. Recall the sampling system in (33). An intuitive way of calculating the 1-D diagonal cumulant slice of is given by for any , where denote the indices of the elements in the vector . Because of the stationarity of , it always holds that as long as , for and where , similar to , also denote the indices of the elements in the vector . By stacking the non-zero diagonal cumulant slice values of into a vector as , , the above analysis on the stationary property implies that not every value in is required to calculate , but merely those values indexed by the elements of a set defined by such that . This suggests that can be accurately estimated from a subset of entries of , which enables appealing signal compression for coping with wideband signals. Evidently, the choice of , if exists, is crucial for compressive efficiency. It turns out that several solutions to are available, including the minimal sparse ruler, [13], nested sampling and co-prime sampling [16] [32]. This observation sheds light on the sub-Nyquist sampler design and the reconstruction of from compressive measurements.
VI-A Sampling Strategy for CCSS
We adopt the nonuniform sub-sampling to reduce the average sampling rate [13]. Specifically, the sampling matrix in (33) is designed as a sparse matrix with its elements indexed by () being ones and otherwise zeros, where . As a result, only a small number of samples are collected into the measurements vector in (31), the elements of which are indexed by a subset of the Nyquist grid:
(47) |
where .
It should be noted that is wide-sense stationary but is generally not because the nature of the compressive sampling matrix , and hence the third-order cumulant of is dependent on both the time origin and time-lags. This fact makes it possible that and contain the same degree of freedom even though the dimension of is smaller than that of , where stacks all time-variant cumulant slice values of .
To relate and , it can be easily found that
(48) | ||||
This simple relationship in (48) fundamentally reveals that can be completely recovered from , as long as for every , there exists at least one pair , in satisfying .
To achieve the strongest compression ratio, the compressive sampler design basically boils down to determining the minimal set such that can be estimated accurately from compressively collected samples indexed by . Mathematically, this is a minimal sparse ruler design problem:
(49) |
where and denotes the cardinality of the set . The minimal sparse ruler problem is well studied in [33] and has many precomputed solutions at hand for our design of the sparse sampling matrix . In practice, is composed of the rows of identity matrix indexed by the elements of .
VI-B Cumulant Slice Reconstruction
Assume that multiple blocks of data records ( blocks) are available and each record is denoted by . Adopting the same sampler () for all blocks, the compressive sample vector is obtained as , . Based on (48), the finite-sample estimate of the cumulant slice is obtained by averaging over all records:
(50) |
where is the estimate based on the th subrecord . For example, the second-, third-and fourth-order () unbiased estimate are:
(51) | ||||
(52) | ||||
(53) | ||||
where denotes the number of averaged terms including all satisfying for all .
Evidently, CCSS is able to accurately recover arbitrary th-order cumulant slice () from compressive measurements, which subsumes CCS as a special case with . The compression ratio of sparse ruler based CCSS should be the same as that of CCS because both of them depends only on the cardinalty of , the analysis of which can be found in detail in [34]. The computation complexity of cumulant slice estimation is comparable to that of covariance estimation as is analyzed in [29], which is much lighter than the entire cumulant estimation.
VII SIMULATIONS
This section presents simulation results to corroborate the effectiveness of the proposed approaches on the sampling rate reduction as well as colored Gaussian noise suppression. Specific applications including correlation function estimation and line spectrum estimation are considered.
VII-A C3CS: Robustness to Sampling Rate Reduction
The discrete-time data used for simulation is generated by passing a zero-mean, independent and exponentially distributed driven process through an MA linear system with coefficients [1, 0.9, 0.385,-0.771], say , which has a finite number of cumulant and correlation values. We process block by block with each data block of equal length . The compressive measurements of size is obtained from (33), and the value of that decides the compression ratio will be specified in each simulation case. As analyzed in the Remark 1, given fixed , the proposed direct C3CS approach is able to estimate cumulants within lags , while the alternative C3CS approach within . In this simulation, since cumulants at lags outside is negligible, we test the alternative C3CS approach directly.
First, it is of great interest to study the impact of the compression ratio on the error performance, we define the performance metric as the normalized mean-square error (MSE) of the estimated with respect to the uncompressed one, for a given compression ratio . That is, . In the noise-free setting, Fig. 4 depicts the MSE of C3CS versus the compression ratio , for a varying number of measurement vectors . Fig. 4 shows that the estimation performance of C3CS is satisfactory even at a compression ratio without posing any sparsity constraint on the signal, and the performance improves with increasing . This is understandable because weaker compression always leads to better performance, and stands for no compression. In this simulation, the sharp drop of MSE happens when , so we say is the strongest achievable compression ratio. It is worth pointing out that the achievable strongest compression ratio can be much smaller than and even below when is large as illustrated in Fig. 3. In this simulation, is set to be small for a simple illustration. Furthermore, the error performance also improves with increasing but the improvement becomes slight when is large enough (observe the curves for and in Fig. 4). This is because a large number of data blocks can effectively reduce the finite sample effect and when is large enough the finite sample effect has been adequately alleviated.
Suppressing colored Gaussian noise is one of the most important merits of third-order cumulants. The error performance of C3CS in the colored Gaussian noise setting is evaluated in Fig. 5, where the added dB colored Gaussian noise is generated by passing white Gaussian noise through the MA(5) model with coefficients [1, -2.33, 0.75, 0.5, -1.3, -1.4]. In this setting, the estimation error comes from not only the compression but also the strong noise. Theoretically, the third-order cumulants of colored Gaussian noise are identical to zero. However, practical estimation only accesses to a finite number of data records, which can not average the third-order cumulants of colored Gaussian noise to be exactly zero. As a result, the performance of C3CS in Fig. 5 is not as good as that in Fig. 4. Even so, we can see that the third-order cumulants can still be recovered satisfactorily from compressive measurements as long as given a large number of data records ( or larger) such that the Gaussian noise is averaged to be nearly zero. The error performance also improves with increasing compression ratio , which is similar to the trends in the noise-free setting. Fig. 5 demonstrates the effectiveness of C3CS on the robustness to colored Gaussian noise, which is very useful in practical application as illustrated next.
VII-B C3CS: Application to Correlation Function Estimation
The correlation function is useful in many signal processing tasks, such as spectral density estimation, DoA esitmation, and etc. However, under the strong colored Gaussian noise setting, the correlation estimate is highly biased if the spectrum of noise is unknown for prewhitening. In this section, we consider estimating the unbiased correlation function from compressive measurements by using C3CS even when the uncompressed signal is contaminated by strong colored Gaussian noise. Specifically, We estimate the unbiased correlation function based on the output of C3CS followed by the method proposed in [19]. As a benchmark, we also evaluate the correlation function estimation based on CCS [9]. The data used for the simulation here is the same as in Part A.
In the noise-free setting, the error performance of correlation estimation based on C3CS and CCS is depicted in Fig. 6. It signifies that CCS-based approach achieves better MSE performance than the C3CS-based one. This is because given the same number of data blocks, the variance of the covariance estimate is smaller than that of the cumulant estimate. Moreover, the strongest achievable compression ratio of C3CS is about in this simulation, while that of CCS is much smaller at least below . In other words, CCS achieves stronger compression than C3CS as analyzed in Remark 4. For correlation function estimation application, although CCS based approach shows appealing advantages in the noise-free setting, we will show next how C3CS-based approach outperforms the CCS-based one in the colored Gaussian noise setting.
In the colored Gaussian noise setting, the error performance of correlation estimation based on C3CS and CCS is shown in Fig. 7. It shows that the curves corresponding to CCS approach with various are almost overlapped and the corresponding error is always greater than , which means that the CCS-based approach fails to estimate the correlation function (no matter how large is) due to the presence of strong colored Gaussian noise. In contrast, the C3CS-based approach still works as long as given enough data blocks to reduce the finite sampling effect and average the colored Gaussian noise to be nearly zero. We emphasize again that the achievable strongest compression ratio of C3CS can be much smaller when is large in practical applications.
To gain a more straightforward comparison between the C3CS-based and CCS-based approaches for correlation estimation, we also plot the estimated correlation values at several lags under different compression ratios in Fig. 8, while the true correlation values are also plotted in the solid line as a benchmark. The estimated correlation values obtained from both the C3CS-based and the CCS-based approaches are calculated by averaging over multiple Monte Carlo runs. We can see that the CCS-based approach is highly biased even without any compression () while the C3CS-based approach achieves quite good estimation even with , which validates the superiority of the proposed C3CS approach in terms of sampling rate reduction and noise suppression.
VII-C CCSS: Robustness to Sampling Rate Reduction
In this subsection, we consider a signal containing harmonics with analog frequencies and and phases are i.i.d. uniformly distributed over [-, ]. Assume a conservative sampling rate to get a digital counter part with digital frequencies and . It has been shown in [29] that both cumulant slice and correlation can be used to retrieve harmonics contained in , while the former is more robust to colored Gaussian noise. In this simulation, and are estimated from independent realizations, each consisting of samples. We sample at the sub-Nyquist rate via (33) to test CCSS.
We first examine the robustness to the sampling rate reduction of our proposed CCSS approach. Since the length-16 () minimal sparse ruler has distance marks, the sampler is designed by choosing the 7 rows corresponding to the distance marks from the identity matrix . In this case, the sampler achieves the strongest compression rate . Then the larger cases are generated by randomly adding additional rows of to the already chosen 7 rows. The normalized MSE between the estimated cumulant slice and the true one is calculated for sparse ruler sampling (). In the noise-free setting, Fig. 9 shows the MSE under various values of . It indicates that the estimation performance of CCSS is satisfactory even when , and can be improved with increasing and . In addition, Fig. 9 also shows that CCS achieves better MSE performance than CCSS for fixed and , which is due to the faster convergence rate of sample second-order statistics. However, the superiority of CCSS arises when strong colored Gaussian noise ( dB) is present as shown in Fig. 10 where the added colored Gaussian noise is generated by passing white Gaussian noise through an ARMA filter with AR parameters [1, 1.4563, 0.81] and MA parameters [1, 2, 1]. It turns out that CCSS still works as long as given large enough to reduce finite sample effect and hence suppressing Gaussian noise, while, in contrast, CCS does not work at all because its estimation error is almost constant greater than . We next show the advantage of C3CS in a practical task.
VII-D CCSS: Application to Line Spectrum Estimation
In this subsection, we test the performance of CCSS in the line spectrum estimation application and the data used here is the same as in Part C. Specifically, we first estimate the cumulant slice and the correlation based on CCSS and CCS respectively, and then use the estimates to retrieve harmonics as in [29].
Fig. 11 depicts the MUSIC spectrum based on CCS and CCSS. Here, the cumulant slice estimate and the covariance estimate are calculated from data blocks under the same noise setting as above. Note this colored noise spectrum has a strong pole at frequency . We can see that the frequency estimation performance is already quite good when for both the cumulant slice and covariance based methods, and can be further improved by increasing the compression rate to . In practice, the choice of the compression rate reflects a trade-off between estimation accuracy and sampling rates. Further, the MUSIC spectrum based on CCSS detects two peaks at and , while that based on CCS shows an additional peak around resulted from the colored noise. Evidently, the estimated spectrum based on CCSS outperforms that of CCS in terms of removing colored Gaussian noise, which is particularly useful when the noise spectrum is unavailable or hard to estimate for prewhitening purpose.
VIII Summary
This paper presents novel approaches to the reconstruction of third-order statistics from compressive measurements. Capitalizing on the structures of third-order cumulants resulting from signal stationarity and cumulant symmetry, an over-determined linear system is constructed that explicitly links the cross-cumulants of compressive samples with the cumulants of the uncompressed signal. Guideline on compressive sampler design is provided, and the conditions for lossless recovery guarantee are delineated. It is proven that the proposed techniques are able to achieve significant compression, even when no sparsity constraint is imposed on the signal or cumulants. For applications where only the diagonal cumulant slice is required, a computationally simple approach is developed to directly reconstruct the desired cumulant slices from compressive samples without having to recover the entire cumulant tensor, which turns out to subsume the existing compressive covariance sensing framework as a special case. Extensive simulations are conducted corroborating the effectiveness of the proposed techniques.
Appendix A Mapping Matrix
This is to show how the mapping matrix in (42) is determined. Note that and contain the same set of unique cumulant values, and thus the mapping matrix we aim at finding out is actually a tall deterministic binary matrix . We first introduce a basic fact that will facilitate our derivation: Denote the th entry of the cumulant tensor as , then after vectorization this entry is mapped into the vector with index being , say .
Recall that is defined as
(54) | ||||
Observe that , the elements in , can be found in at indices and indices due to the symmetry Property 1. Then equivalently, according to the fact introduced above, can also be found in at indices
(55) | ||||
Concisely, we have
(56) |
for .
Next we need to find out the indices of located in . It is easy to validate that and , or equivalently we say the indices of in are
(57) | ||||
Concisely, we have
(58) |
for .
References
- [1] Yanbo Wang and Zhi Tian, “Cumulant slice reconstruction from compressive measurements and its application to line spectrum estimation,” in ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2020, pp. 5520–5524.
- [2] Yanbo Wang and Zhi Tian, “Third-order cumulants reconstruction from compressive measurements,” in 2020 54th Asilomar Conference on Signals, Systems, and Computers. IEEE, 2020.
- [3] David L Donoho et al., “Compressed sensing,” IEEE Transactions on information theory, vol. 52, no. 4, pp. 1289–1306, 2006.
- [4] Emmanuel J Candès, Justin Romberg, and Terence Tao, “Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information,” IEEE Transactions on information theory, vol. 52, no. 2, pp. 489–509, 2006.
- [5] Zhi Tian and Georgios B Giannakis, “Compressed sensing for wideband cognitive radios,” in 2007 IEEE International Conference on Acoustics, Speech and Signal Processing-ICASSP’07. IEEE, 2007, vol. 4, pp. IV–1357.
- [6] Yoram Bresler, “Spectrum-blind sampling and compressive sensing for continuous-index signals,” in 2008 Information Theory and Applications Workshop. IEEE, 2008, pp. 547–554.
- [7] Moshe Mishali and Yonina C Eldar, “From theory to practice: Sub-nyquist sampling of sparse wideband analog signals,” IEEE Journal of selected topics in signal processing, vol. 4, no. 2, pp. 375–391, 2010.
- [8] Geert Leus and Dyonisius Dony Ariananda, “Power spectrum blind sampling,” IEEE Signal Processing Letters, vol. 18, no. 8, pp. 443–446, 2011.
- [9] Zhi Tian, Yohannes Tafesse, and Brian M Sadler, “Cyclic feature detection with sub-nyquist sampling for wideband spectrum sensing,” IEEE Journal of Selected topics in signal processing, vol. 6, no. 1, pp. 58–69, 2011.
- [10] Zhi Tian, “Cyclic feature based wideband spectrum sensing using compressive sampling,” in 2011 IEEE International Conference on Communications (ICC). IEEE, 2011, pp. 1–5.
- [11] Geert Leus and Zhi Tian, “Recovering second-order statistics from compressive measurements,” in 2011 4th IEEE International Workshop on Computational Advances in Multi-Sensor Adaptive Processing (CAMSAP). IEEE, 2011, pp. 337–340.
- [12] Dyonisius Dony Ariananda and Geert Leus, “Compressive wideband power spectrum estimation,” IEEE Transactions on signal processing, vol. 60, no. 9, pp. 4775–4789, 2012.
- [13] Daniel Romero, Dyonisius Dony Ariananda, Zhi Tian, and Geert Leus, “Compressive covariance sensing: Structure-based compressive sensing beyond sparsity,” IEEE signal processing magazine, vol. 33, no. 1, pp. 78–93, 2016.
- [14] Dyonisius Dony Ariananda, Geert Leus, and Zhi Tian, “Multi-coset sampling for power spectrum blind sensing,” in 2011 17th International Conference on Digital Signal Processing (DSP). IEEE, 2011, pp. 1–8.
- [15] Palghat P Vaidyanathan and Piya Pal, “Sparse sensing with co-prime samplers and arrays,” IEEE Transactions on Signal Processing, vol. 59, no. 2, pp. 573–586, 2010.
- [16] Piya Pal and Palghat P Vaidyanathan, “Coprime sampling and the music algorithm,” in 2011 Digital Signal Processing and Signal Processing Education Meeting (DSP/SPE). IEEE, 2011, pp. 289–294.
- [17] Jerry M Mendel, “Tutorial on higher-order statistics (spectra) in signal processing and system theory: Theoretical results and some applications,” Proceedings of the IEEE, vol. 79, no. 3, pp. 278–305, 1991.
- [18] Chrysostomos L Nikias and Jerry M Mendel, “Signal processing with higher-order spectra,” IEEE Signal processing magazine, vol. 10, no. 3, pp. 10–37, 1993.
- [19] Georgios B Giannakis and Anastasios N Delopoulos, “Nonparametric estimation of autocorrelation and spectra using cumulants and polyspectra,” in Advanced Signal Processing Algorithms, Architectures, and Implementations. International Society for Optics and Photonics, 1990, vol. 1348, pp. 503–517.
- [20] Kaitian Cao and Hequn Shao, “Compressive wideband spectrum sensing using high-order statistics for cognitive radio,” in 2013 IEEE Global high tech congress on electronics. IEEE, 2013, pp. 187–190.
- [21] Chia Wei Lim and Michael B Wakin, “Compressive temporal higher order cyclostationary statistics,” IEEE Transactions on Signal Processing, vol. 63, no. 11, pp. 2942–2956, 2015.
- [22] Qiong Wu and Qilian Liang, “Higher-order statistics in co-prime sampling with application to channel estimation,” IEEE Transactions on Wireless Communications, vol. 14, no. 12, pp. 6608–6620, 2015.
- [23] Zhuo Sun, Song Kong, and Wenbo Wang, “Sparse learning of higher-order statistics for communications and sensing,” IEEE Transactions on Emerging Topics in Computational Intelligence, 2018.
- [24] Jason N Laska, Sami Kirolos, Marco F Duarte, Tamer S Ragheb, Richard G Baraniuk, and Yehia Massoud, “Theory and implementation of an analog-to-information converter using random demodulation,” in 2007 IEEE International Symposium on Circuits and Systems. IEEE, 2007, pp. 1959–1962.
- [25] Venera Khoromskaia and Boris N Khoromskij, “Block circulant and toeplitz structures in the linearized hartree–fock equation on finite lattices: Tensor approach,” Computational Methods in Applied Mathematics, vol. 17, no. 3, pp. 431–455, 2017.
- [26] Andrzej Cichocki, Danilo Mandic, Lieven De Lathauwer, Guoxu Zhou, Qibin Zhao, Cesar Caiafa, and Huy Anh Phan, “Tensor decompositions for signal processing applications: From two-way to multiway component analysis,” IEEE signal processing magazine, vol. 32, no. 2, pp. 145–163, 2015.
- [27] Thomas F Andre, Robert D Nowak, and Barry D Van Veen, “Low-rank estimation of higher order statistics,” IEEE transactions on signal processing, vol. 45, no. 3, pp. 673–685, 1997.
- [28] Erik Meijer, “Matrix algebra for higher order moments,” Linear Algebra and its Applications, vol. 410, pp. 112–134, 2005.
- [29] Ananthram Swami and Jerry M Mendel, “Cumulant-based approach to harmonic retrieval and related problems,” IEEE Transactions on Signal Processing, vol. 39, no. 5, pp. 1099–1109, 1991.
- [30] Jing Liang and Zhi Ding, “Fir channel estimation through generalized cumulant slice weighting,” IEEE transactions on signal processing, vol. 52, no. 3, pp. 657–667, 2004.
- [31] Georgios B Giannakis and Jerry M Mendel, “Identification of nonminimum phase systems using higher order statistics,” IEEE Transactions on Acoustics, Speech, and Signal Processing, vol. 37, no. 3, pp. 360–377, 1989.
- [32] Piya Pal and Palghat P Vaidyanathan, “Nested arrays: A novel approach to array processing with enhanced degrees of freedom,” IEEE Transactions on Signal Processing, vol. 58, no. 8, pp. 4167–4181, 2010.
- [33] John Leech, “On the representation of 1, 2,…, n by differences,” Journal of the London Mathematical Society, vol. 1, no. 2, pp. 160–169, 1956.
- [34] Daniel Romero, Roberto López-Valcarce, and Geert Leus, “Compression limits for random vectors with linearly parameterized second-order statistics,” IEEE Transactions on Information Theory, vol. 61, no. 3, pp. 1410–1425, 2015.