This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Direction Finding in Partly Calibrated Arrays Exploiting the Whole Array Aperture

Guangbin Zhang, Tianyao Huang, Yimin Liu, Xiqin Wang and Yonina C. Eldar  This work was supported by the National Natural Science Foundation of China under Grants 62171259. G. Zhang, T. Huang, Y. Liu, and X. Wang are with the Department of Electronic Engineering, Tsinghua University, Beijing 100084, China (e-mail: [email protected]; {huangtianyao, yiminliu, wangxq_ee}@tsinghua.edu.cn). Yonina C. Eldar is with the Department of EE Technion, Israel Institute of Technology, Haifa, Israel (e-mail:[email protected]). T. Huang is the corresponding author.
Abstract

We consider the problem of direction finding using partly calibrated arrays, a distributed subarray with position errors between subarrays. The key challenge is to enhance angular resolution in the presence of position errors. To achieve this goal, existing algorithms, such as subspace separation and sparse recovery, have to rely on multiple snapshots, which increases the burden of data transmission and the processing delay. Therefore, we aim to enhance angular resolution using only a single snapshot. To this end, we exploit the orthogonality of the signals of partly calibrated arrays. Particularly, we transform the signal model into a special multiple-measurement model, show that there is approximate orthogonality between the source signals in this model, and then use blind source separation to exploit the orthogonality. Simulation and experiment results both verify that our proposed algorithm achieves high angular resolution as distributed arrays without position errors, inversely proportional to the whole array aperture.

Index Terms:
Partly calibrated array, high angular resolution, orthogonality, blind source separation.

I Introduction

Direction finding using antenna arrays plays a fundamental role in various fields of signal processing such as radar, sonar, and astronomy [1]. High angular resolution is a key objective in these fields, and requires a large array aperture [2]. However, a large aperture usually means high system complexity, poor mobility, and high cost.

A promising alternative to a single large-aperture array is to partition the whole array into distributed subarrays, and fuse their measurements in a coherent way in order to achieve the same angular resolution as the whole array [3, 4, 5, 6, 7, 8]. A premise of coherently fusing the measurements from subarrays is the accurate position of each array element. Positions of array elements within a subarray are easy to obtain, however there are generally inevitable position errors between subarrays, resulting in challenges to perform coherent signal processing and achieve high angular resolution [9]. For fixed ground-based platforms, these position errors could be calibrated offline. For example, black hole imaging uses atomic clocks to precisely calibrate the position errors between subarrays, and synthesizes a large aperture close to the earth diameter [10]. In this case, the synthetic array, with both intra- and inter- subarray well calibrated, is referred to as a fully calibrated array. However, in mobile platforms such as unmanned aerial vehicles (UAV), the movement of the platform makes it hard to have accurate subarray positions in real time. Such a distributed array is often referred to as partly calibrated array [9], with the intra-subarray elements well calibrated but not the inter-subarray elements. Estimating the directions of arrival (DoA) of radiating sources with a partly calibrated array has received wide attention recently [11, 12, 13], and is the focus of this paper.

Early approaches for partly calibrated arrays first estimate the directions with each subarray separately, and then fuse the estimates to improve accuracy [14, 15, 16]. Since signals from different subarrays are not coherently processed, the angular resolution is limited to each single subarray instead of the synthesized aperture.

More advanced algorithms jointly process received signals of all the subarrays. These techniques can be divided into two categories: correlation domain and direct data domain algorithms. Correlation domain techniques first calculate the covariance matrix of the received signals, separate the signal and noise subspaces from the covariance matrix, and use the subspaces to indicate the DoAs. These subspace algorithms [17, 18, 19, 20] are shown to exploit the whole array aperture and achieve high angular resolution. However, correlation domain approaches require a large number of snapshots to accurately estimate the covariance matrix and usually assume that the radiating sources are independent, which may not be satisfied in practice. Moreover, many snapshots increase the burden of data transmission and the processing delay, which has a significant impact on system performance. Particularly in mobile and time-varying scenarios [21], the scenario changes rapidly during long-time observations, which leads to model mismatch and performance loss. Hence, it is of significance to achieve high angular resolution with fewer snapshots, even a single snapshot. However, few snapshots are not typically sufficient to estimate the covariance matrix correctly.

Direct data domain algorithms are often preferred compared to correlation domain algorithms. This is because direct data domain algorithms directly estimate the DoAs by exploiting some prior information on the sources instead of their statistical properties learned from the received signals. Among these techniques, sparse recovery approaches have attracted interest in recent years [22, 11]. These algorithms exploit the sparsity of sources and estimate DoAs by solving a block- and rank-sparse optimization problem. They have tractable complexity and are shown to be less sensitive than correlation domain techniques to few snapshots and correlated sources. However, they have poor performance in the single-snapshot case [11].

To achieve high-resolution direction finding in partly calibrated arrays with only a single snapshot, we propose to transform the signal model into a multiple-measurement model [23, 24], where each measurement corresponds to the single-snapshot data of a subarray. In this way, we encode the high-resolution capability into the phase relationship between the measurements. We then show that there is approximate orthogonality between the measurements of different sources. By exploiting the orthogonality, the phase offsets between subarrays are recovered to enhance angular resolution.

We use a blind source separation (BSS) [25] algorithm called Joint Approximate Diagonalization of Eigen-matrices (JADE) [26] to exploit the orthogonality in our scenarios. BSS is a typical tool that makes use of signal characteristics (independence) between measurements in multiple-measurement models. Based on BSS ideas, our study illustrates that the cost function of JADE can well characterize the signal characteristics (orthogonality) between the measurements of sources in partly calibrated arrays. We thus apply JADE to exploit the orthogonality, and then use the output of JADE for phase offset recovery between subarrays, which leads to direction finding with high angular resolution. Not only the simulation, but also experiment results verify that our proposed algorithm achieves high angular resolution, inversely proportional to the whole array aperture, with only a single snapshot.

The rest of this paper is organized as follows: Section II introduces the signal model of partly calibrated arrays, and shows how to transform it into a multiple-measurement model. Section III illustrates that there is approximate orthogonality between the source signals in partly calibrated arrays. The proposed direction-finding algorithm exploiting the orthogonality is detailed in Section IV. We discuss the relationship between the orthogonality and angular resolution in Section V. Numerical simulation and experimental results are presented in Section VI, followed by a conclusion in Section VII.

Notation: We use \mathbb{R} and \mathbb{C} to denote the sets of real and complex numbers, respectively. Uppercase boldface letters denote matrices (e.g. 𝑪\bm{C}) and lowercase boldface letters denote vectors (e.g. 𝒄\bm{c}). The (m,n)(m,n)-th element of a matrix 𝑪\bm{C} is denoted by [𝑪]m,n[\bm{C}]_{m,n}, and the nn-th column is represented by [𝑪]n[\bm{C}]_{n}. We use trace(){\rm trace}(\cdot) to indicate the trace of a matrix and diag(𝒂){\rm diag}(\bm{a}) to represent a matrix with diagonal elements given by 𝒂\bm{a}. The conjugate, transpose, and conjugate transpose operators are denoted by ,T,H{}^{*},^{T},^{H}, respectively. The amplitude of a scalar, the l2l_{2} norm of a vector and the Frobenius norm of a matrix are represented by |||\cdot|, 2\|\cdot\|_{2} and \|\cdot\|, respectively. We use vec()\rm{vec}(\cdot) for matrix vectorization. The Hadamard product is written as \odot, and \equiv is the definition symbol. We denote the imaginary unit for complex numbers by ȷ=1\jmath=\sqrt{-1}.

II Signal model

In this section, we first introduce the signals of the partly calibrated model in Subsection II-A. For comparison, we introduce the signals of fully calibrated arrays and discuss angular resolution in Subsection II-B. Next, we transform the single-snapshot signals of partly calibrated arrays into a multiple-measurement model in Subsection II-C. Finally, we explain the challenges and motivation of recovering directions with high angular resolution from this multiple-measurement model in Subsection II-D.

II-A Partly calibrated model

Consider KK isotropic linear antenna subarrays, each composed of MkM_{k} sensors for k=1,,Kk=1,\dots,K, where Mk=1KMkM\equiv\sum_{k=1}^{K}M_{k}. We construct a planar Cartesian coordinate system, where the array is on the x-axis. For these subarrays, the partly calibrated model assumes the precisely known intra-subarray displacement and unknown inter-subarray displacement. Particularly, denote by ξk\xi_{k}\in\mathbb{R} the unknown inter-subarray displacement of the first (reference) sensor in the kk-th subarray relative to the the first sensor in the 1-st subarray for k=1,,Kk=1,\dots,K, thus, ξ1=0\xi_{1}=0. We use 𝝃=[ξ2,,ξK]T(K1)×1\bm{\xi}=[\xi_{2},\dots,\xi_{K}]^{T}\in\mathbb{R}^{(K-1)\times 1} to represent the inter-subarray displacement vector. Denote by ηk,m\eta_{k,m}\in\mathbb{R} the known intra-subarray displacement of the mm-th sensor relative to the 1-st sensor in the kk-th subarray for m=1,,Mk,k=1,,Km=1,\dots,M_{k},k=1,\dots,K, thus, ηk,1=0\eta_{k,1}=0. Assume that the MM sensors share a common sampling clock, or the clocks have been synchronized exactly. We illustrate the partly calibrated model in Fig. 1.

Refer to caption
Figure 1: The partly calibrated model, where 𝝃\bm{\xi} is unknown.

There are LL far-field [27], closely spaced emitters impinging narrow-band signals onto the whole array from different directions θ1,,θL\theta_{1},\dots,\theta_{L}. Denote the direction vector by 𝜽=[θ1,,θL]TL×1\bm{\theta}=[\theta_{1},\dots,\theta_{L}]^{T}\in\mathbb{R}^{L\times 1}. In general cases, we assume 𝜽𝟎\bm{\theta}\geqslant\bm{0} or 𝜽𝟎\bm{\theta}\leqslant\bm{0}. The steering vector of the kk-th subarray corresponding to an emitter at direction θ\theta is given by

𝒂k(θ)=𝒃k(θ)ϕk(θ,ξ)Mk×1,\bm{a}_{k}(\theta)=\bm{b}_{k}(\theta)\phi_{k}(\theta,\xi)\in\mathbb{C}^{M_{k}\times 1}, (1)

where ϕk(θ,ξ)=exp(ȷ2πλξksinθ)\phi_{k}(\theta,\xi)={\rm exp}\left(\jmath\frac{2\pi}{\lambda}\xi_{k}\sin\theta\right) is an unknown phase offset between the kk-th and the 1-st subarray, λ\lambda denotes the wavelength of the transmitted signals by emitters and the vector 𝒃k(θ)Mk×1\bm{b}_{k}(\theta)\in\mathbb{C}^{M_{k}\times 1} is defined by

[𝒃k(θ)]m=exp(ȷ2πληk,msinθ),m=1,,Mk.\displaystyle[\bm{b}_{k}(\theta)]_{m}={\rm exp}\left(\jmath\frac{2\pi}{\lambda}\eta_{k,m}\sin\theta\right),m=1,\dots,M_{k}. (2)

Here 𝒃k(θ)\bm{b}_{k}(\theta) is a known function of θ\theta in contrast to the unknown phase offset ϕk(θ,ξ)\phi_{k}(\theta,\xi) for k=1,,Kk=1,\dots,K.

In this paper, we assume that only a single snapshot is available and all the subarrays sample at the same time. Therefore, the signal received by the kk-th subarray is a summation of signals transmitted by the LL emitters, given by

𝒙k=l=1Lsl𝒂k(θl)+𝒏kMk×1,\bm{x}_{k}=\sum_{l=1}^{L}s_{l}\bm{a}_{k}(\theta_{l})+\bm{n}_{k}\in\mathbb{C}^{M_{k}\times 1}, (3)

where sls_{l} represents an unknown complex coefficient and 𝒏k\bm{n}_{k} represents noise for k=1,,Kk=1,\dots,K. By substituting (1) into (3), we rewrite the received signals (3) as

𝒙k=l=1Lsl𝒃k(θl)ϕk(θl,ξ)+𝒏k.\bm{x}_{k}=\sum_{l=1}^{L}s_{l}\bm{b}_{k}(\theta_{l})\phi_{k}(\theta_{l},\xi)+\bm{n}_{k}. (4)

II-B Fully calibrated model

The fully calibrated arrays assume that 𝝃\bm{\xi} in (4) is exactly known or well calibrated. By substituting (2) and ϕk(θ,ξ)=exp(ȷ2πλξksinθ)\phi_{k}(\theta,\xi)={\rm exp}\left(\jmath\frac{2\pi}{\lambda}\xi_{k}\sin\theta\right) to (4), we have

𝒙k=𝑪k(𝜽)𝒔+𝒏k,\displaystyle\bm{x}_{k}=\bm{C}_{k}(\bm{\theta})\bm{s}+\bm{n}_{k}, (5)

where [𝑪k(𝜽)]m,l=exp(ȷ2πλ(ηk,m+ξk)sinθl)[\bm{C}_{k}(\bm{\theta})]_{m,l}={\rm exp}\left(\jmath\frac{2\pi}{\lambda}(\eta_{k,m}+\xi_{k})\sin\theta_{l}\right) for m=1,,Mkm=1,\dots,M_{k} and 𝒔=[s1,,sL]TL×1\bm{s}=[s_{1},\dots,s_{L}]^{T}\in\mathbb{C}^{L\times 1}. Denote 𝒙~c=[𝒙1T,,𝒙KT]TM×1\tilde{\bm{x}}_{c}=[\bm{x}_{1}^{T},\dots,\bm{x}_{K}^{T}]^{T}\in\mathbb{C}^{M\times 1}, 𝑪(𝜽)=[𝑪1T(𝜽),,𝑪KT(𝜽)]TM×L\bm{C}(\bm{\theta})=[\bm{C}_{1}^{T}(\bm{\theta}),\dots,\bm{C}_{K}^{T}(\bm{\theta})]^{T}\in\mathbb{C}^{M\times L} and 𝒏~c=[𝒏1T,,𝒏KT]TM×1\tilde{\bm{n}}_{c}=[\bm{n}_{1}^{T},\dots,\bm{n}_{K}^{T}]^{T}\in\mathbb{C}^{M\times 1}. We then stack the KK signals in (5) together as

𝒙~c=𝑪(𝜽)𝒔+𝒏~c,\displaystyle\tilde{\bm{x}}_{c}=\bm{C}(\bm{\theta})\bm{s}+\tilde{\bm{n}}_{c}, (6)

which is a typical single-snapshot signal model. In (6), ξk\xi_{k} and ηk,m\eta_{k,m} are known and the unknowns in 𝑪(𝜽)\bm{C}(\bm{\theta}) are only 𝜽\bm{\theta}. Direction finding is to estimate 𝜽\bm{\theta} given 𝒙~c\tilde{\bm{x}}_{c}, which can be achieved by existing classical algorithms such as Multiple Signal Classification (MUSIC) [28] and compressed sensing (CS) [29].

Direction finding by fully calibrated arrays can achieve high angular resolution, inversely proportional to the whole array aperture. This is because the fully calibrated model assumes completely known 𝝃\bm{\xi} and the received signals are recast as (6). In this case, the subarrays can be regarded as sparsely distributed array elements in a single array with a large aperture, where the array manifold is denoted by 𝑪(𝜽)\bm{C}(\bm{\theta}) in (6). In this single array, the positions of array elements are in the range from 0 to ηK,MK+ξK\eta_{K,M_{K}}+\xi_{K}, yielding the whole array aperture. Therefore, the angular resolution of (6) is inversely proportional to the whole array aperture [2].

II-C Transform into a multiple-measurement model

However, for the partly calibrated model, an accurate array manifold 𝑪(𝜽)\bm{C}(\bm{\theta}) is not available due to the unknown 𝝃\bm{\xi}, which makes it hard to synthesize the subarrays into a single large-aperture array. The unknown 𝝃\bm{\xi} introduces unknown phase errors ϕk(θl,ξ)\phi_{k}(\theta_{l},\xi), which are related to both subarrays and sources, to the signal model. Particularly, define 𝚽k(𝜽,𝝃)=diag(ϕk(θ1,ξ),,ϕk(θL,ξ))L×L\bm{\Phi}_{k}(\bm{\theta},\bm{\xi})={\rm diag}(\phi_{k}(\theta_{1},\xi),\dots,\phi_{k}(\theta_{L},\xi))\in\mathbb{C}^{L\times L} and 𝑩k(𝜽)=[𝒃k(θ1),,𝒃k(θL)]Mk×L\bm{B}_{k}(\bm{\theta})=[\bm{b}_{k}(\theta_{1}),\dots,\bm{b}_{k}(\theta_{L})]\in\mathbb{C}^{M_{k}\times L} for k=1,,Kk=1,\dots,K. We then rewrite (4) as

𝒙k=𝑩k(𝜽)𝚽k(𝜽,𝝃)𝒔+𝒏k,\bm{x}_{k}=\bm{B}_{k}(\bm{\theta})\bm{\Phi}_{k}(\bm{\theta},\bm{\xi})\bm{s}+\bm{n}_{k}, (7)

where 𝚽k(𝜽,𝝃)\bm{\Phi}_{k}(\bm{\theta},\bm{\xi}) denotes the unknown phase error w.r.t. 𝝃\bm{\xi}. Compared with (5), the unknown 𝚽k(𝜽,𝝃)\bm{\Phi}_{k}(\bm{\theta},\bm{\xi}) in (7) makes high-resolution direction finding a challenging problem in single-snapshot cases [11]. For brevity, we denote 𝚽k(𝜽,𝝃)\bm{\Phi}_{k}(\bm{\theta},\bm{\xi}) and 𝑩k(𝜽)\bm{B}_{k}(\bm{\theta}) by 𝚽k\bm{\Phi}_{k} and 𝑩k\bm{B}_{k}, respectively.

There is a key question in partly calibrated arrays:

Can partly calibrated subarrays achieve angular resolution comparable to that of fully calibrated subarrays?

The answer is positive. The feasibility lies in the fact that the unknown phase offsets between subarrays, ϕk(θl,ξ)\phi_{k}(\theta_{l},\xi), can be well recovered by self-calibration on 𝝃\bm{\xi}. In this paper, we provide an algorithm to achieve high angular resolution performance. We leave the theoretical analysis for future work.

To illustrate our approach, we consider a typical case in partly calibrated arrays, where the intra-subarray displacement of each subarray is the same. In this case, we transform the partly calibrated model into a multiple-measurement model [23, 24]. Particularly, {ηk,}\{\eta_{k,\cdot}\} and {Mk}\{M_{k}\} reduce to the same values for different kk, and are denoted by η¯=ηk,\bar{\eta}_{\cdot}=\eta_{k,\cdot} and M¯=Mk\bar{M}=M_{k}, respectively. We use 𝜼=[η¯2,,η¯M¯]T(M¯1)×1\bm{\eta}=[\bar{\eta}_{2},\dots,\bar{\eta}_{\bar{M}}]^{T}\in\mathbb{R}^{(\bar{M}-1)\times 1} to represent the intra-subarray displacement vector. The steering matrices 𝑩k\bm{B}_{k} become the same for all the subarrays, thus we denote 𝑩¯=𝑩kM¯×L\bar{\bm{B}}=\bm{B}_{k}\in\mathbb{C}^{\bar{M}\times L} for k=1,,Kk=1,\dots,K. We assume that for each subarray, M¯>L\bar{M}>L, such that emitters are identifiable [30]. Under these assumptions, 𝑩¯\bar{\bm{B}} is a full column rank matrix. Then, the received signals (7) are reorganized as

[𝒙1,,𝒙K]\displaystyle[\bm{x}_{1},\dots,\bm{x}_{K}] =𝑩¯[𝚽1𝒔,,𝚽K𝒔]+[𝒏1,,𝒏K]\displaystyle=\bar{\bm{B}}\left[\bm{\Phi}_{1}\bm{s},\dots,\bm{\Phi}_{K}\bm{s}\right]+[\bm{n}_{1},\dots,\bm{n}_{K}]
=𝑩¯[𝒔¯1T𝒔¯LT]+[𝒏1,,𝒏K],\displaystyle=\bar{\bm{B}}\begin{bmatrix}\bar{\bm{s}}_{1}^{T}\\ \vdots\\ \bar{\bm{s}}_{L}^{T}\end{bmatrix}+[\bm{n}_{1},\dots,\bm{n}_{K}], (8)

where 𝒔¯l=sl[ϕ1(θl,ξ),,ϕK(θl,ξ)]TK×1\bar{\bm{s}}_{l}=s_{l}\left[\phi_{1}(\theta_{l},\xi),\dots,\phi_{K}(\theta_{l},\xi)\right]^{T}\in\mathbb{C}^{K\times 1} are viewed as the received signals from the ll-th emitter, sampled KK times by different subarrays. Denote 𝑿=[𝒙1,,𝒙K]M¯×K\bm{X}=[\bm{x}_{1},\dots,\bm{x}_{K}]\in\mathbb{C}^{\bar{M}\times K}, 𝑺=[𝒔¯1,,𝒔¯L]TL×K\bm{S}=\left[\bar{\bm{s}}_{1},\dots,\bar{\bm{s}}_{L}\right]^{T}\in\mathbb{C}^{L\times K} and 𝑵=[𝒏1,,𝒏K]M¯×K\bm{N}=[\bm{n}_{1},\dots,\bm{n}_{K}]\in\mathbb{C}^{\bar{M}\times K}. Then (II-C) is recast as

𝑿=𝑩¯𝑺+𝑵,\bm{X}=\bar{\bm{B}}\bm{S}+\bm{N}, (9)

where 𝑩¯\bar{\bm{B}} is related to {𝜽,𝜼}\{\bm{\theta},\bm{\eta}\}, and 𝑺\bm{S} is related to {𝜽,𝝃,𝒔}\{\bm{\theta},\bm{\xi},\bm{s}\}. In (9), {𝑿,𝜼}\{\bm{X},\bm{\eta}\} are known, and {𝜽,𝝃,𝒔,𝑵}\{\bm{\theta},\bm{\xi},\bm{s},\bm{N}\} are unknown. In addition, LL is assumed to be known by existing algorithms [31, 32, 33]. The recast signal model (9) is a multiple-measurement model, where each row of 𝑺\bm{S} represents the multiple measurements of the KK subarrays. Direction finding is to estimate 𝜽\bm{\theta} given 𝑿\bm{X}.

Note that 𝑺\bm{S} in (9) is constructed by multiple measurements of subarrays instead of multiple time samples as typical multiple-measurement models, and has special signal structure. Particularly, the entries of the ll-th row of 𝑺\bm{S} have the same modulus |sl||s_{l}| and different phase offsets ϕk(θl,ξ)\phi_{k}(\theta_{l},\xi). In (9), high-resolution capability of partly calibrated arrays is encoded in this special signal structure of 𝑺\bm{S}, mainly in ϕk(θl,ξ)\phi_{k}(\theta_{l},\xi). The key issue of enhancing angular resolution lies in how to exactly recover ϕk(θl,ξ)\phi_{k}(\theta_{l},\xi) by exploiting the signal structure of 𝑺\bm{S}.

II-D Challenges and motivation

Existing algorithms encounter difficulties in achieving high angular resolution based on (9). This is because 𝑺\bm{S} is a complex function of {𝜽,𝝃,𝒔}\{\bm{\theta},\bm{\xi},\bm{s}\} and exploiting the signal structure of 𝑺\bm{S} to recover ϕk(θl,ξ)\phi_{k}(\theta_{l},\xi) is not a trivial problem. Particularly, MUSIC algorithms [34, 28] suffer from imprecise subspace structure of signals due to the unknown 𝝃\bm{\xi}. Typical sparse recovery algorithms [35] divide the parameter space of 𝜽\bm{\theta} into finite grids, construct a over-complete, grid-based dictionary of 𝑩¯\bar{\bm{B}}, and solve a convex lasso [36] problem. However, these algorithms ignore the phase relationship between subarrays ϕk(θl,ξ)\phi_{k}(\theta_{l},\xi) and assume 𝑺\bm{S} to be completely unknown, which corresponds to non-coherent processing which cannot achieve high angular resolution. Therefore, we do not use gridding on (9). To enhance angular resolution, more efficient constraints on 𝑺\bm{S} are required in the problem formulation. However, it is hard to propose a proper constraint that fully characterizes the signal structure, and the introduction of such constraints increase the difficulty of the solution.

Our goal is to find a way that both makes full use of the signal structure and is tractable to solve. As mentioned above, the key of high angular resolution lies in the exact recovery of ϕk(θl,ξ)\phi_{k}(\theta_{l},\xi), which are the phases of 𝑺\bm{S} in (9). Therefore, we propose to exploit a specific characteristic of 𝑺\bm{S} to recover ϕk(θl,ξ)\phi_{k}(\theta_{l},\xi) first, and then find the directions based on the phase estimates. The characteristic we consider is approximate orthogonality between the source signals of partly calibrated arrays, which is detailed next.

III Approximate orthogonality between the rows of 𝑺\bm{S}

In this section, we first analyze the statistics of the sample covariance of 𝑺\bm{S}, 𝑹^𝒔¯1K𝑺𝑺H\hat{\bm{R}}_{\bar{\bm{s}}}\equiv\frac{1}{K}\bm{S}\bm{S}^{H}, in Subsection III-A, where 𝑺\bm{S} denotes the source signals in (9). Then, we show that there is approximate orthogonality between the rows of 𝑺\bm{S} in partly calibrated arrays in Subsection III-B.

III-A Statistics of 𝐑^𝐬¯1K𝐒𝐒H\hat{\bm{R}}_{\bar{\bm{s}}}\equiv\frac{1}{K}\bm{S}\bm{S}^{H}

First, we calculate the elements of 𝑹^𝒔¯1K𝑺𝑺H\hat{\bm{R}}_{\bar{\bm{s}}}\equiv\frac{1}{K}\bm{S}\bm{S}^{H}. By substituting the definitions of 𝑺\bm{S}, 𝒔¯l\bar{\bm{s}}_{l} and ϕk(θ,ξ)\phi_{k}(\theta,\xi) into 𝑹^𝒔¯\hat{\bm{R}}_{\bar{\bm{s}}}, the (i,j)(i,j)-th element of 𝑹^𝒔¯\hat{\bm{R}}_{\bar{\bm{s}}}, denoted by Ri,jR_{i,j} for abbreviation, is

Ri,j[𝑹^𝒔¯]i,j\displaystyle R_{i,j}\equiv[\hat{\bm{R}}_{\bar{\bm{s}}}]_{i,j} =sisjKk=1Kϕk(θi,ξ)ϕk(θj,ξ),\displaystyle=\frac{s_{i}s_{j}^{*}}{K}\sum_{k=1}^{K}\phi_{k}(\theta_{i},\xi)\phi_{k}^{*}(\theta_{j},\xi),
=sisjKk=1Keȷ2πλξk(sinθisinθj),\displaystyle=\frac{s_{i}s_{j}^{*}}{K}\sum_{k=1}^{K}e^{\jmath\frac{2\pi}{\lambda}\xi_{k}\left(\sin\theta_{i}-\sin\theta_{j}\right)}, (10)

where Ri,j=Rj,iR_{i,j}=R_{j,i}^{*} for i,j=1,,Li,j=1,\dots,L. The diagonal elements of 𝑹^𝒔¯\hat{\bm{R}}_{\bar{\bm{s}}} are Ri,i=|si|2R_{i,i}=|s_{i}|^{2}. We assume that the source intensities are all equal, |si|=1|s_{i}|=1, i=1,,Li=1,\dots,L in the sequel. Under this assumption, the intensities of the off-diagonal elements rely on the inter-subarray displacements 𝝃\bm{\xi} and the directions 𝜽\bm{\theta}.

Denote the whole array aperture and the intra-subarray element spacing by DD and dd, respectively. To analyze the statistics of 𝑹^𝒔¯\hat{\bm{R}}_{\bar{\bm{s}}}, we consider a common practical scenario that the subarrays are randomly, uniformly placed in [0,D][0,D], i.e., ξk𝒰[0,D]\xi_{k}\sim\mathcal{U}\left[0,D\right] for k=1,,Kk=1,\dots,K. The corresponding geometry is shown in Fig. 2.

Refer to caption
Figure 2: The geometry of the randomly, uniformly distributed subarrays.

This typical case leads to the following proposition, a similar analysis can be found in [37].

Proposition 1.

When ξk𝒰[0,D]\xi_{k}\sim\mathcal{U}\left[0,D\right] for k=1,,Kk=1,\dots,K, we have

|𝔼[Ri,j]|=|sinρi,jρi,j|,\left|\mathbb{E}\left[R_{i,j}\right]\right|=\left|\frac{\sin\rho_{i,j}}{\rho_{i,j}}\right|, (11)
𝔼[|Ri,j|2]=1K+(11K)|sinρi,jρi,j|2,\mathbb{E}\left[\left|R_{i,j}\right|^{2}\right]=\frac{1}{K}+\left(1-\frac{1}{K}\right)\left|\frac{\sin\rho_{i,j}}{\rho_{i,j}}\right|^{2}, (12)

where ρi,j=πDλ(sinθisinθj)\rho_{i,j}=\frac{\pi D}{\lambda}(\sin\theta_{i}-\sin\theta_{j}), i,j=1,,Li,j=1,\dots,L and iji\neq j.

Proof.

See Appendix A. ∎

The term |sinρi,jρi,j|\left|\frac{\sin\rho_{i,j}}{\rho_{i,j}}\right| plays an important role in (11) and (12). We draw the curve of |sinρi,jρi,j|\left|\frac{\sin\rho_{i,j}}{\rho_{i,j}}\right| w.r.t. (sinθisinθj)/Δ(\sin\theta_{i}-\sin\theta_{j})/\Delta in Fig. 3, where Δλ/D\Delta\equiv\lambda/D denotes the empirical angular resolution of the whole array.

Refer to caption
Figure 3: The off-diagonal elements |Ri,j||R_{i,j}| in a statistical sense.

From Fig. 3, we see that the curve is composed of a main lobe and several side lobes. When the interval sinθisinθj\sin\theta_{i}-\sin\theta_{j} equals Δ\Delta, the intensity |Ri,j|\left|R_{i,j}\right| reaches its first zero point, which is related to the angular resolution Δ\Delta. When the interval is in the region [0,Δ)[0,\Delta), the distributed array cannot distinguish sources at such close interval and |Ri,j|\left|R_{i,j}\right| has large values. When the interval belongs to the region [Δ,)[\Delta,\infty), |Ri,j|\left|R_{i,j}\right| is typically small. In this region, the zero points of |Ri,j|\left|R_{i,j}\right| occur in a period of Δ\Delta and there are side lobes between these zero points.

III-B Approximate orthogonality in 𝐑^𝐬¯\hat{\bm{R}}_{\bar{\bm{s}}}

Based on the statistics of 𝑹^𝒔¯\hat{\bm{R}}_{\bar{\bm{s}}}, there is approximate orthogonality in partly calibrated arrays. Here, orthogonality is defined as the orthogonality between the rows of 𝑺\bm{S} in (9). Particularly, we say there is orthogonality in partly calibrated arrays if the off-diagonal elements of 𝑹^𝒔¯\hat{\bm{R}}_{\bar{\bm{s}}} are 0, i.e.,

𝑹^𝒔¯=[|s1|200|sL|2].\displaystyle\hat{\bm{R}}_{\bar{\bm{s}}}=\begin{bmatrix}|s_{1}|^{2}&&0\\ &\ddots&\\ 0&&|s_{L}|^{2}\end{bmatrix}. (13)

In Proposition 1 and Fig. 3, we show that when the source separation exceeds one resolution cell, the off-diagonal elements of 𝑹^𝒔¯\hat{\bm{R}}_{\bar{\bm{s}}} are small relative to the diagonal elements in a statistical sense. In other words, there is approximate orthogonality between the rows of 𝑺\bm{S} in (9), and

𝑹^𝒔¯[|s1|200|sL|2].\displaystyle\hat{\bm{R}}_{\bar{\bm{s}}}\approx\begin{bmatrix}|s_{1}|^{2}&&0\\ &\ddots&\\ 0&&|s_{L}|^{2}\end{bmatrix}. (14)

We may also understand the orthogonality in partly calibrated arrays using the concept of coherence [38], given by

μmaxij|𝐬¯jH𝐬¯i|𝐬¯i𝐬¯j,\displaystyle\mu\equiv\underset{i\neq j}{\rm max}\frac{|\bar{\bm{s}}_{j}^{H}\bar{\bm{s}}_{i}|}{\|\bar{\bm{s}}_{i}\|\|\bar{\bm{s}}_{j}\|}, (15)

where 0μ10\leqslant\mu\leqslant 1. In partly calibrated arrays, the coherence μ\mu corresponds to the largest modulus of the off-diagonal elements of 𝑹^𝒔¯\hat{\bm{R}}_{\bar{\bm{s}}}. Orthogonality means that μ=0\mu=0 and approximate orthogonality means that 0<μ10<\mu\ll 1.

In summary, when the source separation is within a resolution cell, there is no obvious orthogonality between the source signals. When the source separation exceeds a resolution cell, the modulus of the off-diagonal Ri,jR_{i,j} is small in a statistical sense, yielding approximate orthogonality. This phenomenon not only illustrates the orthogonality in partly calibrated arrays, but also builds a connection between orthogonality and angular resolution, which will be discussed in Section V.

IV Joint Approximate Diagonalization

In this section, we introduce our algorithm. First, we propose to use BSS ideas to exploit the orthogonality, and recover the phase offsets between subarrays in Subsection IV-A. Then, based on the estimated phase offsets, beamforming algorithms for direction finding are discussed in Subsection IV-B.

IV-A Phase offset recovery

In (9), when 𝑩¯\bar{\bm{B}} is unknown, how to exploit the orthogonality between the rows of 𝑺\bm{S} is a challenging problem. Here, we propose a solution to this problem. Particularly, we introduce BSS techniques to the self-calibration problems in distributed arrays. To this end, we first review basic BSS ideas. Then, we show how a specific BSS algorithm exploits the orthogonality of 𝑺\bm{S} in (9). Finally, we show how we recover the phase offsets between subarrays.

IV-A1 Review of BSS

BSS problems usually consider the following model:

𝒀=𝑪𝑯+𝑵,\bm{Y}=\bm{C}\bm{H}+\bm{N}, (16)

where 𝒀N×Ts\bm{Y}\in\mathbb{C}^{N\times T_{s}} contains the received signals, 𝑪N×L\bm{C}\in\mathbb{C}^{N\times L} is an unknown, full-rank measurement matrix, 𝑯=[𝒉1,,𝒉L]TL×Ts\bm{H}=[\bm{h}_{1},\dots,\bm{h}_{L}]^{T}\in\mathbb{C}^{L\times T_{s}} are the unknown original source signals, 𝑵N×Ts\bm{N}\in\mathbb{C}^{N\times T_{s}} are the unknown white noises, TsT_{s}, NN and LL denote the number of samples, measurements and sources, respectively, and L<NL<N. The purpose of BSS is to recover 𝑯\bm{H} from 𝒀\bm{Y} with unknown 𝑪\bm{C}.

In order to make the recovery of 𝑯\bm{H} possible, BSS usually imposes assumptions on the independence between the original source signals 𝒉l,l=1,,L\bm{h}_{l},l=1,\dots,L. Under these assumptions, BSS formulates an optimization problem exploiting the independence of 𝒉l\bm{h}_{l} to find the linear ‘inverse’ matrix of 𝑪\bm{C} in (16) as 𝑪^1\widehat{\bm{C}}^{-1}, where 𝑪^1𝒀\widehat{\bm{C}}^{-1}\bm{Y} yields the recovery of 𝑯\bm{H}. Note that there are unavoidable ambiguity between 𝑯\bm{H} and 𝑪\bm{C} in BSS problems. For example, for any nonzero diagonal matrix 𝑸L×L\bm{Q}\in\mathbb{C}^{L\times L}, we have 𝑪𝑯=(𝑪𝑸)(𝑸1𝑯)\bm{C}\bm{H}=\left(\bm{C}\bm{Q}\right)\left(\bm{Q}^{-1}\bm{H}\right). Therefore, without loss of generality, it is typically assumed that the source signal has unit power, 𝔼[hl(t)hl(t)]=1\mathbb{E}[h_{l}(t)h_{l}^{*}(t)]=1, where hl(t)h_{l}(t) denotes the continuous signals. More details on BSS can be found in [25].

IV-A2 How BSS exploits the orthogonality

Here we introduce a specific BSS algorithm, called JADE [26], and show how JADE exploits the orthogonality in partly calibrated arrays from its cost function. JADE assumes unknown measurement matrix 𝑪\bm{C} in (16), which is compatible with the unknown measurement matrix 𝑩¯\bar{\bm{B}} in (9).

The cost function of JADE is given by [26]

fs(𝑺)=r,p,q=1,,L|Cum^([𝑺T]r,[𝑺H]r,[𝑺T]p,[𝑺H]q)|2,f_{s}(\bm{S})=\sum_{r,p,q=1,\dots,L}\Big{|}\widehat{\text{Cum}}\left([\bm{S}^{T}]_{r},[\bm{S}^{H}]_{r},[\bm{S}^{T}]_{p},[\bm{S}^{H}]_{q}\right)\Big{|}^{2}, (17)

where the high-order cumulant Cum^()\widehat{\text{Cum}}(\cdot) is denoted by

Cum^(ϵa,ϵb,ϵc,ϵd)=1K(ϵaϵb)T(ϵcϵd)\displaystyle\widehat{\text{Cum}}(\bm{\epsilon}_{a},\bm{\epsilon}_{b},\bm{\epsilon}_{c},\bm{\epsilon}_{d})=\frac{1}{K}(\bm{\epsilon}_{a}\odot\bm{\epsilon}_{b})^{T}(\bm{\epsilon}_{c}\odot\bm{\epsilon}_{d})
1K2(ϵaTϵbϵcTϵd+ϵaTϵcϵbTϵd+ϵaTϵdϵbTϵc),\displaystyle-\frac{1}{K^{2}}\left(\bm{\epsilon}_{a}^{T}\bm{\epsilon}_{b}\bm{\epsilon}_{c}^{T}\bm{\epsilon}_{d}+\bm{\epsilon}_{a}^{T}\bm{\epsilon}_{c}\bm{\epsilon}_{b}^{T}\bm{\epsilon}_{d}+\bm{\epsilon}_{a}^{T}\bm{\epsilon}_{d}\bm{\epsilon}_{b}^{T}\bm{\epsilon}_{c}\right), (18)

and ϵ\bm{\epsilon} corresponds to [𝑺T]l=𝒔¯l[\bm{S}^{T}]_{l}=\bar{\bm{s}}_{l} in (II-C). Based on the definition of Ri,jR_{i,j} in (III-A), we have

Ri,j1K𝒔¯jH𝒔¯i=1K[𝑺T]jH[𝑺T]i,\displaystyle R_{i,j}\equiv\frac{1}{K}\bar{\bm{s}}_{j}^{H}\bar{\bm{s}}_{i}=\frac{1}{K}[\bm{S}^{T}]^{H}_{j}[\bm{S}^{T}]_{i}, (19)

which corresponds to ϵjTϵi\bm{\epsilon}_{j}^{T}\bm{\epsilon}_{i} in (IV-A2) for i,j=1,,Li,j=1,\dots,L.

The expression of fs(𝑺)f_{s}(\bm{S}) in (17) w.r.t high-order cumulants is complex, however, it can be simplified due to the special signal structure of partly calibrated arrays. Particularly, since 𝒔¯l\bar{\bm{s}}_{l} in (II-C) has constant modulus |sl||s_{l}|, Rl,l=1K𝒔¯lH𝒔¯l=|sl|2R_{l,l}=\frac{1}{K}\bar{\bm{s}}_{l}^{H}\bar{\bm{s}}_{l}=|s_{l}|^{2} and 𝒔¯l𝒔¯lH=|sl|2[1,1,,1]T\bar{\bm{s}}_{l}\odot\bar{\bm{s}}_{l}^{H}=|s_{l}|^{2}\cdot[1,1,\dots,1]^{T}. Based on the above properties, substituting (19) to (17), we have

fs(𝑺)\displaystyle f_{s}(\bm{S}) =r,p,q||sr|2Rp,q|sr|2Rp,qR~r,pR~r,qRr,qRp,r|2,\displaystyle=\sum_{r,p,q}\left||s_{r}|^{2}R_{p,q}-|s_{r}|^{2}R_{p,q}-\tilde{R}_{r,p}\tilde{R}_{r,q}^{*}-R_{r,q}R_{p,r}\right|^{2},
=r,p,q|R~r,pR~r,q+Rr,qRp,r|2,\displaystyle=\sum_{r,p,q}\left|\tilde{R}_{r,p}\tilde{R}_{r,q}^{*}+R_{r,q}R_{p,r}\right|^{2}, (20)

where

R~i,j1K𝒔¯jT𝒔¯i=sisjKk=1Keȷ2πλξk(sinθi+sinθj),\displaystyle\tilde{R}_{i,j}\equiv\frac{1}{K}\bar{\bm{s}}_{j}^{T}\bar{\bm{s}}_{i}=\frac{s_{i}s_{j}}{K}\sum_{k=1}^{K}e^{\jmath\frac{2\pi}{\lambda}\xi_{k}\left(\sin\theta_{i}+\sin\theta_{j}\right)}, (21)

for i,j=1,,Li,j=1,\dots,L. Note that R~i,j\tilde{R}_{i,j} is similar to Ri,jR_{i,j}, and the property of Ri,jR_{i,j} in Proposition 1 when |sinθjsinθi|Δ|\sin\theta_{j}-\sin\theta_{i}|\geqslant\Delta can be directly extended to R~i,j\tilde{R}_{i,j} when |sinθj+sinθi|Δ|\sin\theta_{j}+\sin\theta_{i}|\geqslant\Delta.

Based on (IV-A2), we now provide an intuitive explanation to how JADE exploits the orthogonality in partly calibrated arrays. In Proposition 1 and Fig. 3, we illustrate that when the source separation exceeds one resolution cell, we have small Ri,jR_{i,j} in a statistical sense, yielding approximate orthogonality. We extend this property to R~i,j\tilde{R}_{i,j} and have

|sinθisinθj|Δ,|Ri,j|0|sinθi+sinθj|Δ,|R~i,j|0}fs(𝑺)0,\displaystyle\left.\begin{aligned} |\sin\theta_{i}-\sin\theta_{j}|\geqslant\Delta,|R_{i,j}|\approx 0\\ |\sin\theta_{i}+\sin\theta_{j}|\geqslant\Delta,|\tilde{R}_{i,j}|\approx 0\end{aligned}\right\}\Rightarrow f_{s}(\bm{S})\approx 0, (22)

which means that smaller |Ri,j||R_{i,j}| and |R~i,j||\tilde{R}_{i,j}| yield smaller fs(𝑺)f_{s}(\bm{S}) approximately. When 𝜽𝟎\bm{\theta}\geqslant\bm{0} or 𝜽𝟎\bm{\theta}\leqslant\bm{0} as assumed, |sinθisinθj|Δ|\sin\theta_{i}-\sin\theta_{j}|\geqslant\Delta implies |sinθj+sinθi|Δ|\sin\theta_{j}+\sin\theta_{i}|\geqslant\Delta. Therefore, when the sources are resolvable (|sinθisinθj|Δ|\sin\theta_{i}-\sin\theta_{j}|\geqslant\Delta), minimizing fs(𝑺)f_{s}(\bm{S}) is likely to have an optimal result of 𝑺\bm{S} with |Ri,j|0|R_{i,j}|\approx 0. This illustrates that the cost function of JADE is feasible to exploit the orthogonality in partly calibrated arrays. The optimal solution of JADE is not guaranteed to be global from (IV-A2). However, JADE is verified to achieve good performance by simulations in Section VI. This shows the potential of using BSS techniques to enhance angular resolution in partly calibrated arrays.

In most practical scenarios, the orthogonality is approximately satisfied, and the performance of JADE is affected by the level of the approximation. Intuitively speaking, higher level of orthogonality corresponds to a lower cost function of JADE, which likely yields better performance of JADE in phase recovery. This conclusion is verified by simulations in Fig. 4 and Fig. 7 in Section VI. The specific algorithm of JADE is detailed in Appendix C for reference.

IV-A3 Phase offset estimation

We showed how JADE exploits the orthogonality. Next we discuss how to use JADE for phase offset estimation.

To this end, we input the received signals of partly calibrated arrays in (9) to JADE, which outputs an estimate of 𝑺\bm{S}, denoted by 𝑺^\hat{\bm{S}}. Due to the unavoidable ambiguity in the recovery of BSS, it is assumed that the source signal has unit power in JADE. However, the estimate 𝑺^\hat{\bm{S}} usually has fluctuating modulus. Therefore, we normalize 𝑺^\hat{\bm{S}} and estimate the phase offsets as

ϕ^l,k=[𝑺^]l,k|[𝑺^]l,k|,\hat{\phi}_{l,k}=\frac{[\hat{\bm{S}}]_{l,k}}{\left|[\hat{\bm{S}}]_{l,k}\right|}, (23)

for l=1,,Ll=1,\dots,L and k=1,,Kk=1,\dots,K. In the sequel, we denote 𝚽^k=diag(ϕ^1,k,,ϕ^l,k)L×L\hat{\bm{\Phi}}_{k}={\rm diag}\left(\hat{\phi}_{1,k},\dots,\hat{\phi}_{l,k}\right)\in\mathbb{C}^{L\times L}.

IV-B Direction finding

When the phase offsets between subarrays is recovered by (23), direction finding reduces to a coherent array signal processing problem: Recover 𝜽\bm{\theta} from the following signals,

𝑿=𝑩¯(𝜽)[𝚽^1𝒔,,𝚽^K𝒔]+𝑵.\bm{X}=\bar{\bm{B}}(\bm{\theta})\left[\hat{\bm{\Phi}}_{1}\bm{s},\dots,\hat{\bm{\Phi}}_{K}\bm{s}\right]+\bm{N}. (24)

This problem can be solved by many classical algorithms. In this subsection, we introduce two representative techniques for direction finding.

IV-B1 Matched filtering

We begin with matched filtering (MF) for direction finding. Consider the L=1L=1 case, where (24) is simplified as

𝑿=𝒃¯(θ1)[ϕ^1,1,,ϕ^1,K]s+𝑵,\bm{X}=\bar{\bm{b}}(\theta_{1})\left[\hat{\phi}_{1,1},\dots,\hat{\phi}_{1,K}\right]s+\bm{N}, (25)

and the steering vector 𝒃¯(θ)M¯×1\bar{\bm{b}}(\theta)\in\mathbb{C}^{\bar{M}\times 1} is denoted by 𝒃¯(θ)=[1,exp(ȷ2πλη2sinθ),,exp(ȷ2πληM¯sinθ)]T\bar{\bm{b}}(\theta)=\left[1,{\rm exp}\left(\jmath\frac{2\pi}{\lambda}\eta_{2}\sin\theta\right),\dots,{\rm exp}\left(\jmath\frac{2\pi}{\lambda}\eta_{\bar{M}}\sin\theta\right)\right]^{T}. For (25), the MF algorithm is to find θ1\theta_{1} by maximizing the following cost function,

θ^1\displaystyle\hat{\theta}_{1} =argmaxθ(π2,π2)|vecH(𝐗)vec(𝐛¯(θ)[ϕ^1,1,,ϕ^1,K])|\displaystyle=\underset{\theta\in(-\frac{\pi}{2},\frac{\pi}{2})}{\rm argmax}\ \left|{\rm vec}^{H}(\bm{X}){\rm vec}\left(\bar{\bm{b}}(\theta)\left[\hat{\phi}_{1,1},\dots,\hat{\phi}_{1,K}\right]\right)\right|
=argmaxθ(π2,π2)|k=1K𝐱kH(𝐛¯(θ)ϕ^1,k)|.\displaystyle=\underset{\theta\in(-\frac{\pi}{2},\frac{\pi}{2})}{\rm argmax}\ \left|\sum_{k=1}^{K}\bm{x}_{k}^{H}\left(\bar{\bm{b}}(\theta)\hat{\phi}_{1,k}\right)\right|. (26)

The MF algorithm is directly applicable to the case when L2L\geqslant 2. Particularly, directions 𝜽\bm{\theta} are estimated as

θ^l=argmaxθ(π2,π2)|k=1K𝐱kH(𝐛¯(θ)ϕ^l,k)|,\hat{\theta}_{l}=\underset{\theta\in(-\frac{\pi}{2},\frac{\pi}{2})}{\rm argmax}\ \left|\sum_{k=1}^{K}\bm{x}_{k}^{H}\left(\bar{\bm{b}}(\theta)\hat{\phi}_{l,k}\right)\right|, (27)

for l=1,,Ll=1,\dots,L. The optimization problem (27) can be directly solved by conducting LL individual one-dimensional searches on the direction range (π/2,π/2)(-\pi/2,\pi/2).

The MF algorithm (27) is easy to implement. However, in the multi-source (L2L\geqslant 2) case, the estimation accuracy of θ^l\hat{\theta}_{l} by (27) is affected by the other L1L-1 source signals in 𝒙k\bm{x}_{k}, i.e., rlLsr𝒃¯(θr)ϕk(θr,ξ)\sum_{r\neq l}^{L}s_{r}\bar{\bm{b}}(\theta_{r})\phi_{k}(\theta_{r},\xi) in (II-C) for l=1,,Ll=1,\dots,L. Therefore, for better performance, we introduce an alternative algorithm to solve this problem.

IV-B2 Nonlinear least squares

Here, we introduce the nonlinear least squares (NLS) algorithm [39] that jointly estimates {𝜽,𝒔}\{\bm{\theta},\bm{s}\}. Particularly, we consider the following NLS optimization problem w.r.t. {𝜽,𝒔}\{\bm{\theta},\bm{s}\},

min𝜽,𝒔C(𝜽,𝐬)=k=1K𝐱k𝐁¯(𝜽)𝚽^k𝐬22.\underset{\bm{\theta},\bm{s}}{\rm min}\ C(\bm{\theta},\bm{s})=\sum_{k=1}^{K}\left\|\bm{x}_{k}-\bar{\bm{B}}(\bm{\theta})\hat{\bm{\Phi}}_{k}\bm{s}\right\|_{2}^{2}. (28)

To solve (28), we use gradient descend algorithms and alternatively carry out the following two steps until convergence:

𝒔t+1\displaystyle\bm{s}^{t+1} =𝒔tκ𝒔t𝒔C(𝜽t,𝒔t),\displaystyle=\bm{s}^{t}-\kappa_{\bm{s}}^{t}\nabla_{\bm{s}}C(\bm{\theta}^{t},\bm{s}^{t}), (29)
𝜽t+1\displaystyle\bm{\theta}^{t+1} =𝜽tκ𝜽t𝜽C(𝜽t,𝒔t+1),\displaystyle=\bm{\theta}^{t}-\kappa_{\bm{\theta}}^{t}\nabla_{\bm{\theta}}C(\bm{\theta}^{t},\bm{s}^{t+1}), (30)

where tt denotes the iteration index, {κ𝒔t,κ𝜽t}\{\kappa_{\bm{s}}^{t},\kappa_{\bm{\theta}}^{t}\} denote the step size of {𝜽,𝒔}\{\bm{\theta},\bm{s}\} in the tt-th repetition, determined via Armijo line search [40], and \nabla_{\cdot} denotes the gradient operator w.r.t. \cdot.

We refer to the methods above as direction finding in the partly calibrated model using BSS and MF (BSS-MF) or NLS (BSS-NLS), respectively. The framework of the two algorithms are summarized in Algorithm 1 together, where they differ in step 3).

Algorithm 1 BSS-MF / BSS-NLS

Input: The received signals 𝑿\bm{X}, the intra-subarray displacement 𝜼\bm{\eta} and the number of sources LL.

1) Apply the JADE algorithm to (9) to obtain 𝑺^\hat{\bm{S}} .
2) Estimate the phase offsets with (23).
3) [BSS-MF]: Estimate directions with (27), or
  [BSS-NLS]: Estimate directions by solving (28).

Output: Direction estimation θ^l\hat{\theta}_{l} for l=1,,Ll=1,\dots,L.

V Resolution and orthogonality

In this section, we discuss the angular resolution of the partly and fully calibrated arrays from the aspect of orthogonality in Section III, respectively. Since a rigorous analysis of angular resolution, such as [41], is usually complex, here we give a heuristic explanation instead and leave the theoretical analysis to future work.

Angular resolution of partly calibrated arrays. As discussed in Proposition 1, when the intervals between sources are greater than Δ\Delta, i.e., sinθisinθjΔ\sin\theta_{i}-\sin\theta_{j}\geqslant\Delta, |Ri,j||R_{i,j}| is small. In this case, the phase offsets ϕk(θl,ξ)\phi_{k}(\theta_{l},\xi) can be well recovered by the algorithms exploiting orthogonality, yielding high-resolution direction finding. However, when the intervals between sources are less than Δ\Delta, i.e., sinθisinθj<Δ\sin\theta_{i}-\sin\theta_{j}<\Delta, there is no obvious orthogonality between the source signals. This suggests that the achievable angular resolution of our proposed algorithms is on the order of Δ=λ/D\Delta=\lambda/D, inversely proportional to the whole array aperture DD. This will be verified by simulation results in Section VI.

Angular resolution of fully calibrated arrays. We also compare |Ri,j||R_{i,j}| of partly calibrated arrays with the counterpart of fully calibrated arrays. Particularly, the angular correlation coefficient [42], which can be used to indicate the angular resolution, is defined as

𝒢i,j=𝒈jH𝒈i𝒈i𝒈j,\mathcal{G}_{i,j}=\frac{\bm{g}_{j}^{H}\bm{g}_{i}}{\|\bm{g}_{i}\|\|\bm{g}_{j}\|}, (31)

where [𝒈]n=eȷ2πλζnsinθ[\bm{g}_{\cdot}]_{n}=e^{\jmath\frac{2\pi}{\lambda}\zeta_{n}\sin\theta_{\cdot}}, \cdot represents ii or jj, and ζn\zeta_{n} denotes the displacement of the nn-th sensor relative to the 1-st sensor in the whole array, n=1,,KM¯n=1,\dots,K\bar{M}. Similar to Proposition 1, when ξk𝒰[0,D]\xi_{k}\sim\mathcal{U}\left[0,D\right], |𝒢i,j||\mathcal{G}_{i,j}| is given by Proposition 2.

Proposition 2.

When ξk𝒰[0,D]\xi_{k}\sim\mathcal{U}\left[0,D\right] for k=1,,Kk=1,\dots,K, we have

|𝔼[𝒢i,j]|\displaystyle\left|\mathbb{E}[\mathcal{G}_{i,j}]\right| =|i,j||sinρi,jρi,j|,\displaystyle=|\mathcal{M}_{i,j}|\cdot\left|\frac{\sin\rho_{i,j}}{\rho_{i,j}}\right|, (32)
𝔼[|𝒢i,j|2]\displaystyle\mathbb{E}\left[|\mathcal{G}_{i,j}|^{2}\right] =|i,j|2(1K+(11K)|sinρi,jρi,j|2),\displaystyle=|\mathcal{M}_{i,j}|^{2}\left(\frac{1}{K}+\left(1-\frac{1}{K}\right)\left|\frac{\sin\rho_{i,j}}{\rho_{i,j}}\right|^{2}\right), (33)

where i,j=sin(M¯φi,j)M¯sinφi,j\mathcal{M}_{i,j}=\frac{\sin(\bar{M}\varphi_{i,j})}{\bar{M}\sin\varphi_{i,j}}, φi,j=πdλ(sinθisinθj)\varphi_{i,j}=\frac{\pi d}{\lambda}(\sin\theta_{i}-\sin\theta_{j}) and ρi,j=πDλ(sinθisinθj)\rho_{i,j}=\frac{\pi D}{\lambda}(\sin\theta_{i}-\sin\theta_{j}).

Proof.

See Appendix B. ∎

From Proposition 2, we take |𝔼[𝒢i,j]|\left|\mathbb{E}[\mathcal{G}_{i,j}]\right| in (32) for example and find that it has two parts, which are derived from the intra-subarray and inter-subarray parts of the steer vector 𝒈\bm{g}. The first part is |i,j||\mathcal{M}_{i,j}|, which is related to the subarray aperture. The second part is the same as (11) in Proposition 1. Since we assume the same intra-subarray displacements η¯\bar{\eta} for all the subarrays, this part of |𝔼[𝒢i,j]|\left|\mathbb{E}[\mathcal{G}_{i,j}]\right| can be extracted and calculated as i,j\mathcal{M}_{i,j}, and the inter-subarray part is then calculated as (11).

Based on the orthogonality in Proposition 1 and Proposition 2, we heuristically see that the fully and partly calibrated arrays have the same angular resolution. Particularly, since the first zero point can be used to indicate the angular resolution as shown in Fig. 3, we compare the first zero points of (11) and (32). In (32), the first zero points of |i,j||\mathcal{M}_{i,j}| and |sinρi,jρi,j|\left|\frac{\sin\rho_{i,j}}{\rho_{i,j}}\right| are Δ1=λM¯d\Delta_{1}=\frac{\lambda}{\bar{M}d} and Δ2=λD\Delta_{2}=\frac{\lambda}{D}, respectively. Since M¯dD\bar{M}d\ll D, we have Δ2Δ1\Delta_{2}\ll\Delta_{1}, i.e., the first zero point of |𝒢i,j||\mathcal{G}_{i,j}| in (32) is λD\frac{\lambda}{D}, which is the same as |Ri,j||R_{i,j}| in (11). This implies that the angular resolution of the partly calibrated arrays is the same as the fully calibrated arrays’. A rigorous proof of this conclusion is left to future work.

VI Experimental results

In this section, we first present two simulation results: The first verifies the ability of phase offset estimation by JADE in Subsection VI-A. The second compares the performance of our proposed algorithm with existing algorithms and the Crame´{\rm\acute{e}}r-Rao lower bound (CRB) [18] in Subsection VI-B. Then, we carry out hardware experiments to further verify the feasibility of our algorithms in Subsection VI-C.

VI-A Phase offset estimation by JADE

Here, we show the performance of phase offset estimation with JADE by comparing the true |Ri,j||R_{i,j}| with the estimated one by JADE. This is because more accurate phase offset estimation yields better recovery of |Ri,j||R_{i,j}|. We also verify that higher level of orthogonality likely yields better performance of JADE.

Particularly, we set L=2L=2, s1=s2=1s_{1}=s_{2}=1, M¯=10\bar{M}=10, K=10K=10, D=450λD=450\lambda, d=λ2d=\frac{\lambda}{2} and fix θ1=1.2\theta_{1}=1.2^{\circ}. With sinθ2sinθ1\sin\theta_{2}-\sin\theta_{1} as the variable, we first calculate |R2,1|\left|R_{2,1}\right| of (III-A) in the equidistant and random distributed arrays. Then, we substitute the phase offset estimation as (23) by JADE to (III-A) and calculate the corresponding |R2,1|\left|R_{2,1}\right| as |R^2,1|=|1Kk=1Kϕ^2,kϕ^1,k|\left|\hat{R}_{2,1}\right|=\left|\frac{1}{K}\sum_{k=1}^{K}\hat{\phi}_{2,k}\hat{\phi}_{1,k}^{*}\right|. The simulation results are shown in Fig. 4.

Refer to caption
(a) Equidistant case
Refer to caption
(b) Random case
Figure 4: |R2,1||R_{2,1}| and |R^2,1||\hat{R}_{2,1}| w.r.t. (sinθ2sinθ1)/Δ(\sin\theta_{2}-\sin\theta_{1})/\Delta in the equidistant and random cases, where Δ\Delta denotes angular resolution.

From Fig. 4, first we find that the estimation |R^2,1||\hat{R}_{2,1}| is close to the truth |R2,1||R_{2,1}| when |R2,1||R_{2,1}| is small, yielding the feasibility of phase recovery in partly calibrated arrays using JADE. Second, larger |R2,1||R_{2,1}| usually corresponds to a bigger gap between |R^2,1||\hat{R}_{2,1}| and |R2,1||R_{2,1}| and vice versus, which verifies that high level of orthogonality (smaller |R2,1||R_{2,1}|) likely means better performance of JADE (smaller gap between |R^2,1||\hat{R}_{2,1}| and |R2,1||R_{2,1}|). Third, |R2,1||R_{2,1}| in the equidistant case has apparent grating lobes at (sinθ2sinθ1)/Δ=10(\sin\theta_{2}-\sin\theta_{1})/\Delta=10, which do not appear in the random case due to the randomness of the displacement.

VI-B Method comparison

In this subsection, we compare our proposed algorithms, BSS-MF and BSS-NLS, with the root-RARE [18], spectral-RARE [19], COBRAS [11], and group sparse [43] methods in the scenarios with only a single snapshot. We compare both the angular resolution and direction estimation accuracy of our proposed techniques with other methods and the corresponding CRBs. The simulations are carried out under various signal-to-noise rate (SNR)s and angle differences.

The CRBs for partly calibrated arrays, denoted by PC-CRB [18], is both given as the benchmarks. We use the root mean square error (RMSE) of 𝜽\bm{\theta} to indicate the estimation performance of directions. We carry out TmT_{m} Monte Carlo trials and denote the RMSE of 𝜽\bm{\theta} by

RMSE(𝜽)=1Tmt=1Tm𝜽^t𝜽22,{\rm RMSE}(\bm{\theta})=\sqrt{\frac{1}{T_{m}}\sum_{t=1}^{T_{m}}\left\|\hat{\bm{\theta}}_{t}-\bm{\theta}^{*}\right\|_{2}^{2}}, (34)

where 𝜽^t\hat{\bm{\theta}}_{t} is the direction estimate in the tt-th trial and 𝜽\bm{\theta}^{*} is the true value.

We consider that there are L=2L=2 emitters in the directions of 𝜽\bm{\theta}, transmitting signals with wavelength λ\lambda. There are K=10K=10 uniform linear subarrays receiving the signals and each subarray has M¯\bar{M} array elements. The intra-subarray and inter-subarray displacements are respectively set as η¯m=(m1)λ/2\bar{\eta}_{m}=(m-1)\lambda/2 for m=1,,M¯m=1,\dots,\bar{M} and ξk=(k1)DK1\xi_{k}=\frac{(k-1)D}{K-1} for k=1,,Kk=1,\dots,K, where the whole array aperture is D=450λD=450\lambda. In this case, the angular resolution of a single subarray and the whole array are λ(M¯1)dcosθ12.73\frac{\lambda}{(\bar{M}-1)d\cos\theta}\approx 12.73^{\circ} and λDcosθ0.13\frac{\lambda}{D\cos\theta}\approx 0.13^{\circ}, respectively. The received signal amplitudes are set as 𝒔=[eȷπ5,3eȷ3π5]T\bm{s}=[e^{\jmath\frac{\pi}{5}},3e^{\jmath\frac{3\pi}{5}}]^{T}. In the scenario, there are additive noises 𝒏k\bm{n}_{k} for k=1,,Kk=1,\dots,K, which are i.i.d. white Gaussian with mean 𝟎\bm{0} and variance σ2𝑰\sigma^{2}\bm{I}. Here SNR is defined as 1/σ2{1}/{\sigma^{2}}.

VI-B1 Performance w.r.t. SNR

Here we show the performance of algorithms w.r.t. SNR in the angular resolution and estimation accuracy. Particularly, we set M¯=10\bar{M}=10 and SNR =20=20 dB. Since the angular resolution of the subarray and whole array are 12.7312.73^{\circ} and 1.31.3^{\circ} respectively, we consider two cases that 𝜽=[1.2,14.2]T\bm{\theta}=[1.2^{\circ},14.2^{\circ}]^{T} and [1.2,1.4]T[1.2^{\circ},1.4^{\circ}]^{T}. The former can be distinguished by the subarray, and the latter can be distinguished by the whole array, but not by the subarray.

In the group sparse algorithm, we set the grids of 𝜽\bm{\theta} as {1,1.1,,15}\{1,1.1,\dots,15\} for 𝜽=[1.2,14.2]T\bm{\theta}=[1.2^{\circ},14.2^{\circ}]^{T} and {1,1.01,,2}\{1,1.01,\dots,2\} for [1.2,1.4]T[1.2^{\circ},1.4^{\circ}]^{T}, such that the true values are on the grids, and hence there is no grid mismatch problem [44]. For COBRAS, we set the grids of 𝜽\bm{\theta} as {0.2,1.2,,15.2}\{0.2,1.2,\dots,15.2\} for 𝜽=[1.2,14.2]T\bm{\theta}=[1.2^{\circ},14.2^{\circ}]^{T} and {1,1.1,,2}\{1,1.1,\dots,2\} for [1.2,1.4]T[1.2^{\circ},1.4^{\circ}]^{T}, since more grids lead to much higher computation for existing solvers, such as cvx [45]. For example, the direction finding results of a single trial are shown in Fig. 5. Then, we change SNR and carry out Tm=300T_{m}=300 trials for each SNR. The RMSEs of BSS-NLS, BSS-MF, root-RARE, spectral-RARE, COBRAS and the group sparse (GS) algorithms, as well as PC-CRB, are shown in Fig. 6 with a logarithmic coordinate.

Refer to caption
(a) 𝜽=[1.2,14.2]T\bm{\theta}=[1.2^{\circ},14.2^{\circ}]^{T}
Refer to caption
(b) 𝜽=[1.2,1.4]T\bm{\theta}=[1.2^{\circ},1.4^{\circ}]^{T}
Figure 5: Comparison in the angular resolution.
Refer to caption
(a) 𝜽=[1.2,14.2]T\bm{\theta}=[1.2^{\circ},14.2^{\circ}]^{T}
Refer to caption
(b) 𝜽=[1.2,1.4]T\bm{\theta}=[1.2^{\circ},1.4^{\circ}]^{T}
Figure 6: RMSE w.r.t. SNR.

From Fig. 5, we first find that root-RARE and spectral-RARE algorithms only find the source at θ=14.2/1.4\theta=14.2^{\circ}/1.4^{\circ} and fail to identify the two sources in both cases. This is because only a single snapshot is available in our scenarios, but these algorithms require enough snapshots for subspace separation. Then in Fig. 5(a), we find that the group sparse and COBRAS algorithms separate the two sources and achieve good estimation when 𝜽=[1.2,14.2]T\bm{\theta}=[1.2^{\circ},14.2^{\circ}]^{T}. But they can only identify one emitter when 𝜽=[1.2,1.4]T\bm{\theta}=[1.2^{\circ},1.4^{\circ}]^{T} as shown in Fig. 5(b). This indicates that these sparse recovery algorithms cannot achieve the angular resolution corresponding to the whole array aperture. However, our proposed algorithms, BSS-MF and BSS-NLS, are shown to identify the two emitters and estimate the directions accurately, which indicates the ability of exploiting the whole array aperture and achieving high angular resolution.

In Fig. 6, the statistical results match with the results of a single trial. The root-RARE and spectral-RARE algorithms have poor performance in both cases. The group sparse and COBRAS algorithms estimate 𝜽\bm{\theta} well when 𝜽=[1.2,14.2]T\bm{\theta}=[1.2^{\circ},14.2^{\circ}]^{T}. For high SNR (SNR 20\geqslant 20dB) in Fig. 6(a), these sparse recovery algorithms almost exactly estimate the directions, because the true values are assumed on the grids (Due to space limit, we use -25dB to denote the completely exact estimation). But when 𝜽=[1.2,1.4]T\bm{\theta}=[1.2^{\circ},1.4^{\circ}]^{T}, they fail in high SNRs. This further shows that these sparse recovery algorithms cannot exploit the whole array aperture. However, the RMSEs of our proposed algorithms are close to PC-CRB and outperform other algorithms when 𝜽=[1.2,1.4]T\bm{\theta}=[1.2^{\circ},1.4^{\circ}]^{T}. This further verifies that our algorithms can achieve high angular resolution, inversely proportional to the whole array aperture.

VI-B2 Performance w.r.t. the angle difference

Here we show the performance of the algorithms w.r.t. the angle difference to have a detailed analysis on the angular resolution. Particularly, we set M¯=30\bar{M}=30, SNR =20=20 dB, fix θ1=1.2\theta_{1}=1.2^{\circ}, and change θ2\theta_{2} from 1.221.22^{\circ} to 1.601.60^{\circ}. To compare with the feature of |R2,1||R_{2,1}| in Fig. 4, we consider the RMSEs of algorithms w.r.t. (sinθ2sinθ1)/Δ(\sin\theta_{2}-\sin\theta_{1})/\Delta. The simulation results are shown in Fig. 7 with a logarithmic coordinate.

Refer to caption
Figure 7: RMSE w.r.t. the angle difference.

In Fig. 7, first we find that the RMSEs of our proposed algorithm are smaller than other algorithms and close to PC-CRB when (sinθ2sinθ1)/Δ(\sin\theta_{2}-\sin\theta_{1})/\Delta is an integer, yielding better estimation accuracy. Second, it is worth noting that the RMSEs of our proposed algorithms periodically get worse w.r.t. the angle difference. We use JADE for phase offset estimation and verify that smaller |Ri,j||R_{i,j}| likely yields better JADE performance in Subsection VI-A. The increase of the angle difference changes |Ri,j||R_{i,j}| periodically, which affects the JADE performance in phase offset estimation, and eventually leads to periodic variations in the direction finding performance. Particularly, the curves of BSS-NLS and BSS-MF in Fig. 7 approximate to the minimums when (sinθ2sinθ1)/Δ=1,2,3,(\sin\theta_{2}-\sin\theta_{1})/\Delta=1,2,3,\dots, which is consistent with the minimums of |R2,1||R_{2,1}| in Fig. 4(a). These phenomena support that smaller |Ri,j||R_{i,j}| corresponds to higher level of orthogonality, yielding better performance.

VI-C Experiment results

We consider a scenario of far-field targets impinging narrow-band signals onto the partly calibrated array. This scenario is achieved based on a vector network analyzer (VNA). Particularly, the two ports of the VNA are connected to the transmit (Tx) and receive (Rx) antennas, respectively. We view the transmit antennas as the sources. We construct the received signals of the partly calibrated array by sequentially moving the receive antenna to the position of each element and then sampling the received signals. The received signals of the two sources, T1 and T2, are sampled separately, denoted by 𝒕1M×1\bm{t}_{1}\in\mathbb{C}^{M\times 1} and 𝒕2M×1\bm{t}_{2}\in\mathbb{C}^{M\times 1}. We view 𝒕1+𝒕2\bm{t}_{1}+\bm{t}_{2} as the received signals of the two sources. The geometry is shown in Fig. 8 and the practical scenario is shown in Fig. 9.

Refer to caption
Figure 8: The geometry of the experiment.
Refer to caption
Figure 9: Photo of the experiment scenario.

The parameter settings of this experiment are shown as follows: The frequency of the transmitted signals is fc=15f_{c}=15 GHz, and the wavelength is λ=2\lambda=2 cm. There are K=5K=5 uniform linear subarrays and each has M¯=5\bar{M}=5 array elements with d=1d=1 cm. Particularly, we construct a planar XOYXOY coordinate system and the unit is centimeter (cm). The elements are located on the X axis and the coordinates are (12,0),(11,0),,(12,0)(-12,0),(-11,0),\dots,(12,0). Therefore, the angular resolution of a single subarray and the whole array are λ(M¯1)dcosθ28.6\frac{\lambda}{(\bar{M}-1)d\cos\theta}\approx 28.6^{\circ} and λDcosθ5.7\frac{\lambda}{D\cos\theta}\approx 5.7^{\circ}, respectively. The positions of the sources are (5.46,500)(5.46,500) and (58.27,500)(58.27,500). We take the array center 𝑶(0,0)\bm{O}(0,0) as the reference point and the directions of the sources are 𝜽=[0.63,6.65]T\bm{\theta}=[0.63^{\circ},6.65^{\circ}]^{T}. The array positions are exactly measured using a grid paper with 11 cm interval, but the source positions are not because they are far from the receiver and measurement errors are inevitable. To obtain the true directions 𝜽\bm{\theta} in practice, we use the beamforming results of 𝒕1\bm{t}_{1} and 𝒕2\bm{t}_{2} with the exact 𝝃\bm{\xi} as the ground truth instead.

Under the above settings, we have the received signals 𝒕1\bm{t}_{1} and 𝒕2\bm{t}_{2}. Recall that the fully and partly calibrated model assume exact and biased inter-subarray displacements 𝝃\bm{\xi}, respectively. The exact 𝝃\bm{\xi} is set above and the biased 𝝃^\hat{\bm{\xi}} is constructed by adding errors to the exact 𝝃\bm{\xi} artificially. Particularly, we take λ/2\lambda/2 as unit and add a Gaussian error to exact 𝝃\bm{\xi} to construct biased 𝝃^\hat{\bm{\xi}} as 𝝃^=𝝃+Δ𝝃\hat{\bm{\xi}}=\bm{\xi}+\Delta\bm{\xi}, Δ𝝃N(𝟎,σξ2𝑰)\Delta\bm{\xi}\sim N(\bm{0},\sigma_{\xi}^{2}\bm{I}). Here we set 10log(1/σξ2)=1010{\rm log}(1/\sigma_{\xi}^{2})=10 dB. Then, we calculate the beamforming results of 𝒕1+𝒕2\bm{t}_{1}+\bm{t}_{2} with biased 𝝃^\hat{\bm{\xi}} and compare it with the estimation of BSS-MF, BSS-NLS and the group sparse algorithm (We do not consider the root-RARE and spectral-RARE algorithms here since they fail in the single-snapshot cases). The comparison is shown in Fig. 10.

Refer to caption
(a) Beamforming (BF) results with 𝝃\bm{\xi}
Refer to caption
(b) DoA estimates with biased 𝝃^\hat{\bm{\xi}}
Figure 10: Comparison between beamforming and other algorithms on experimental data 𝒕1\bm{t}_{1}, 𝒕2\bm{t}_{2} and 𝒕1+𝒕2\bm{t}_{1}+\bm{t}_{2}.

From Fig. 10(a), we find that the main lobes of both sources are clear and the side lobes are low. The peaks of the main lobes are corresponding to 0.26-0.26^{\circ} and 5.905.90^{\circ}, which are close to our settings. These phenomenons imply that we get high-quality real measured data.

From Fig. 10(b), when the inter-subarray displacements are erroneous, we find that the beamforming results have many high side lobes and it is hard to identify both sources. This phenomenon illustrates that the subarray position errors in the partly calibrated model have a severe effect on the direction finding performance. However, our proposed algorithms are shown to identify the two sources, with the estimates close to the truth, outperforming the group sparse and COBRAS techniques.

VII Conclusion

In this paper, we consider achieving high angular resolution using distributed arrays in the presence of array displacement errors. We propose a novel algorithm with only a single snapshot, yielding less data transmission burden and processing delay compared with existing works. This is achieved by exploiting the orthogonality between source signals, using BSS. Based on BSS ideas, we propose a direction finding algorithm, and discuss the relationship between orthogonality and angular resolution. Simulation and experiment results both verify the feasibility of our algorithms in achieving high angular resolution, inversely proportional to the whole array aperture.

Appendix A Proof of Proposition 1

Here we prove Proposition 1, i.e., calculate |𝔼[Ri,j]|\left|\mathbb{E}\left[R_{i,j}\right]\right| and 𝔼[|Ri,j|2]\mathbb{E}\left[\left|R_{i,j}\right|^{2}\right] when ξk𝒰[0,D]\xi_{k}\sim\mathcal{U}\left[0,D\right] for k=1,,Kk=1,\dots,K. We denote ρi,j=πDλ(sinθisinθj)\rho_{i,j}=\frac{\pi D}{\lambda}(\sin\theta_{i}-\sin\theta_{j}).

For |𝔼[Ri,j]|\left|\mathbb{E}\left[R_{i,j}\right]\right|, based on (III-A), we have

|𝔼[Ri,j]|\displaystyle\left|\mathbb{E}\left[R_{i,j}\right]\right| =1K|𝔼[k=1Keȷ2ρi,jξk/D]|\displaystyle=\frac{1}{K}\left|\mathbb{E}\left[\sum_{k=1}^{K}e^{\jmath 2\rho_{i,j}\xi_{k}/D}\right]\right|
=1K|k=1K𝔼[eȷ2ρi,jξk/D]|.\displaystyle=\frac{1}{K}\left|\sum_{k=1}^{K}\mathbb{E}\left[e^{\jmath 2\rho_{i,j}\xi_{k}/D}\right]\right|. (35)

Since ξk𝒰[0,D]\xi_{k}\sim\mathcal{U}\left[0,D\right], k=1,,Kk=1,\dots,K, the KK expectations in (A) are the same. Therefore, we rewrite (A) as

|𝔼[Ri,j]|\displaystyle\left|\mathbb{E}\left[R_{i,j}\right]\right| =1KK|𝔼[eȷ2ρi,jξk/D]|\displaystyle=\frac{1}{K}\cdot K\cdot\left|\mathbb{E}\left[e^{\jmath 2\rho_{i,j}\xi_{k}/D}\right]\right|
=|1D0Deȷ2ρi,jξ/D𝑑ξ|\displaystyle=\left|\frac{1}{D}\int_{0}^{D}e^{\jmath 2\rho_{i,j}\xi/D}d\xi\right|
=|sinρi,jρi,j|.\displaystyle=\left|\frac{\sin\rho_{i,j}}{\rho_{i,j}}\right|. (36)

For 𝔼[|Ri,j|2]\mathbb{E}\left[\left|R_{i,j}\right|^{2}\right], based on (III-A), we have

𝔼[|Ri,j|2]\displaystyle\mathbb{E}\left[\left|R_{i,j}\right|^{2}\right] =1K2𝔼[|k=1Keȷ2ρi,jξk/D|2].\displaystyle=\frac{1}{K^{2}}\mathbb{E}\left[\left|\sum_{k=1}^{K}e^{\jmath 2\rho_{i,j}\xi_{k}/D}\right|^{2}\right]. (37)

Due to |x|2=xHx|x|^{2}=x^{H}x, we write (37) as

𝔼[|Ri,j|2]\displaystyle\mathbb{E}\left[\left|R_{i,j}\right|^{2}\right] =1K2𝔼[p=1Keȷ2ρi,jξp/Dq=1Keȷ2ρi,jξq/D],\displaystyle=\frac{1}{K^{2}}\mathbb{E}\left[\sum_{p=1}^{K}e^{-\jmath 2\rho_{i,j}\xi_{p}/D}\sum_{q=1}^{K}e^{\jmath 2\rho_{i,j}\xi_{q}/D}\right],

where the expectation term further equals

𝔼[p=1Kq=1Keȷ2ρi,j(ξqξp)/D]=p=1Kq=1K𝔼[eȷ2ρi,j(ξqξp)/D].\small\mathbb{E}\left[\sum_{p=1}^{K}\sum_{q=1}^{K}e^{\jmath 2\rho_{i,j}(\xi_{q}-\xi_{p})/D}\right]=\sum_{p=1}^{K}\sum_{q=1}^{K}\mathbb{E}\left[e^{\jmath 2\rho_{i,j}(\xi_{q}-\xi_{p})/D}\right].

When p=qp=q, eȷ2ρi,j(ξqξp)/D=1e^{\jmath 2\rho_{i,j}(\xi_{q}-\xi_{p})/D}=1, we have

𝔼[|Ri,j|2]\displaystyle\mathbb{E}\left[\left|R_{i,j}\right|^{2}\right] =1K+1K2p=1Kq=1,qpK𝔼[eȷ2ρi,j(ξqξp)/D].\displaystyle=\frac{1}{K}+\frac{1}{K^{2}}\sum_{p=1}^{K}\sum_{q=1,q\neq p}^{K}\mathbb{E}\left[e^{\jmath 2\rho_{i,j}(\xi_{q}-\xi_{p})/D}\right]. (38)

Since ξk𝒰[0,D]\xi_{k}\sim\mathcal{U}\left[0,D\right], k=1,,Kk=1,\dots,K, the K2KK^{2}-K expectations in (38) are the same. Hence, we write (38) as

𝔼[|Ri,j|2]\displaystyle\mathbb{E}\left[\left|R_{i,j}\right|^{2}\right] =1K+K2KK2D20D0Deȷ2ρi,j(ξqξp)/D𝑑ξp𝑑ξq\displaystyle=\frac{1}{K}+\frac{K^{2}-K}{K^{2}D^{2}}\int_{0}^{D}\int_{0}^{D}e^{\jmath 2\rho_{i,j}(\xi_{q}-\xi_{p})/D}d\xi_{p}d\xi_{q}
=1K+K2KK2D2D2sin2ρi,jρi,j2\displaystyle=\frac{1}{K}+\frac{K^{2}-K}{K^{2}D^{2}}\cdot\frac{D^{2}\sin^{2}\rho_{i,j}}{\rho_{i,j}^{2}}
=1K+(11K)sin2ρi,jρi,j2,\displaystyle=\frac{1}{K}+\left(1-\frac{1}{K}\right)\frac{\sin^{2}\rho_{i,j}}{\rho_{i,j}^{2}}, (39)

completing the proof.

Appendix B Proof of Proposition 2

Based on the definition of 𝒢i,j\mathcal{G}_{i,j}, we have

|𝒢i,j|\displaystyle\left|\mathcal{G}_{i,j}\right| =|𝒈iH𝒈j𝒈i𝒈j|=1KM¯|n=1KM¯eȷ2πλζn(sinθisinθj)|.\displaystyle=\left|\frac{\bm{g}_{i}^{H}\bm{g}_{j}}{\|\bm{g}_{i}\|\|\bm{g}_{j}\|}\right|=\frac{1}{K\bar{M}}\left|\sum_{n=1}^{K\bar{M}}e^{\jmath\frac{2\pi}{\lambda}\zeta_{n}(\sin\theta_{i}-\sin\theta_{j})}\right|. (40)

Note that ζn,n=1,,KM¯\zeta_{n},n=1,\dots,K\bar{M} is related to both the intra-subarray 𝜼¯\bar{\bm{\eta}} and inter-subarray displacement 𝝃\bm{\xi}. Therefore, we can represent ζn\zeta_{n} of the mm-th sensor in the kk-th subarray as ζn=ξk+η¯m\zeta_{n}=\xi_{k}+\bar{\eta}_{m}. Then, we rewrite (40) as

|𝒢i,j|=\displaystyle\left|\mathcal{G}_{i,j}\right|=
|1M¯m=1M¯eȷ2πλη¯m(sinθisinθj)||1Kk=1Keȷ2πλξk(sinθisinθj)|.\displaystyle\left|\frac{1}{\bar{M}}\sum_{m=1}^{\bar{M}}e^{\jmath\frac{2\pi}{\lambda}\bar{\eta}_{m}(\sin\theta_{i}-\sin\theta_{j})}\right|\cdot\left|\frac{1}{K}\sum_{k=1}^{K}e^{\jmath\frac{2\pi}{\lambda}\xi_{k}(\sin\theta_{i}-\sin\theta_{j})}\right|. (41)

Since 𝜼¯m=(m1)d\bar{\bm{\eta}}_{m}=(m-1)d is a constant, we calculate the first summation of (B) as

|i,j|\displaystyle|\mathcal{M}_{i,j}| |1M¯m=1M¯eȷ2πλη¯m(sinθisinθj)|\displaystyle\equiv\left|\frac{1}{\bar{M}}\sum_{m=1}^{\bar{M}}e^{\jmath\frac{2\pi}{\lambda}\bar{\eta}_{m}(\sin\theta_{i}-\sin\theta_{j})}\right|
=|sin(M¯φi,j)M¯sinφi,j|,\displaystyle=\left|\frac{\sin(\bar{M}\varphi_{i,j})}{\bar{M}\sin\varphi_{i,j}}\right|, (42)

where φi,j=πdλ(sinθisinθj)\varphi_{i,j}=\frac{\pi d}{\lambda}(\sin\theta_{i}-\sin\theta_{j}).

Next we consider the second summation of (B). Note that it is the same as the calculation of (III-A). Therefore, based on Proposition 1, we directly have

|𝔼[𝒢i,j]|\displaystyle\left|\mathbb{E}[\mathcal{G}_{i,j}]\right| =|i,j||sinρi,jρi,j|,\displaystyle=|\mathcal{M}_{i,j}|\cdot\left|\frac{\sin\rho_{i,j}}{\rho_{i,j}}\right|, (43)
𝔼[|𝒢i,j|2]\displaystyle\mathbb{E}\left[|\mathcal{G}_{i,j}|^{2}\right] =|i,j|2(1K+(11K)|sinρi,jρi,j|2),\displaystyle=|\mathcal{M}_{i,j}|^{2}\left(\frac{1}{K}+\left(1-\frac{1}{K}\right)\left|\frac{\sin\rho_{i,j}}{\rho_{i,j}}\right|^{2}\right), (44)

where ρi,j=πDλ(sinθisinθj)\rho_{i,j}=\frac{\pi D}{\lambda}(\sin\theta_{i}-\sin\theta_{j}), completing the proof.

Appendix C The JADE algorithm

In (16), the JADE algorithm considers recovering 𝑯\bm{H} from 𝒀\bm{Y} with unknown 𝑪\bm{C}. The main framework of JADE is:

  • (1)

    Compute a whitening matrix 𝑾L×N\bm{W}\in\mathbb{C}^{L\times N} based on 𝒀\bm{Y}.

  • (2)

    Calculate the whitened data 𝒁=𝑾𝒀L×Ts\bm{Z}=\bm{W}\bm{Y}\in\mathbb{C}^{L\times T_{s}}.

  • (3)

    Find a unitary matrix 𝑽L×L\bm{V}\in\mathbb{C}^{L\times L}, such that fs(𝑽H𝒁)f_{s}(\bm{V}^{H}\bm{Z}) reaches its maximum, where fs()f_{s}(\cdot) is a function characterizing the level of independence.

  • (4)

    Output the estimation of 𝑯\bm{H} as 𝑯^=𝑽H𝑾𝒀\hat{\bm{H}}=\bm{V}^{H}\bm{W}\bm{Y}.

Here we detail the specific procedures of JADE. In step (1), to calculate the whitening matrix, JADE first calculates the sample covariance matrix of 𝒀\bm{Y}, denoted by 𝑹^𝒚1Ts𝒀𝒀H\hat{\bm{R}}_{\bm{y}}\equiv\frac{1}{T_{s}}\bm{Y}\bm{Y}^{H}. In 𝑹^𝒚\hat{\bm{R}}_{\bm{y}}, the noise variance σ^\hat{\sigma} is estimated as the average of the NLN-L smallest eigenvalues of 𝑹^𝒚\hat{\bm{R}}_{\bm{y}}, given by σ^=1NLl=L+1N[𝚺^𝒚]l,l\hat{\sigma}=\frac{1}{N-L}\sum_{l=L+1}^{N}[\hat{\bm{\Sigma}}_{\bm{y}}]_{l,l}, where 𝚺^𝒚\hat{\bm{\Sigma}}_{\bm{y}} is obtained from the singular value decomposition (SVD) of 𝑹^𝒚\hat{\bm{R}}_{\bm{y}}, i.e., 𝑹^𝒚=𝑼^𝒚𝚺^𝒚𝑼^𝒚H\hat{\bm{R}}_{\bm{y}}=\hat{\bm{U}}_{\bm{y}}\hat{\bm{\Sigma}}_{\bm{y}}\hat{\bm{U}}_{\bm{y}}^{H}. Therefore, the white noise covariance is estimated as σ^𝑰\hat{\sigma}\bm{I}. Subtracting the noise covariance from 𝚺^𝒚\hat{\bm{\Sigma}}_{\bm{y}} implies noiseless estimation of 𝚺𝒚\bm{\Sigma}_{\bm{y}}. Then, the whitening matrix 𝑾\bm{W} is estimated as

𝑾^=[𝚺^𝒚Tσ^𝑰]1:L12𝑼^𝒚H,\hat{\bm{W}}=[\hat{\bm{\Sigma}}_{\bm{y}}^{T}-\hat{\sigma}\bm{I}]_{1:L}^{-\frac{1}{2}}\hat{\bm{U}}_{\bm{y}}^{H}, (45)

where []1:L[\cdot]_{1:L} denotes the first LL columns of matrix \cdot.

In step (3), JADE minimizes the following cost function characterizing independence, given by

fs(𝑮)=r,p,q=1,,L|Cum^([𝑮T]r,[𝑮H]r,[𝑮T]p,[𝑮H]q)|2,f_{s}(\bm{G})=\sum_{r,p,q=1,\dots,L}\Big{|}\widehat{\text{Cum}}\left([\bm{G}^{T}]_{r},[\bm{G}^{H}]_{r},[\bm{G}^{T}]_{p},[\bm{G}^{H}]_{q}\right)\Big{|}^{2}, (46)

where Cum^()\widehat{\text{Cum}}(\cdot) is denoted by

Cum^(ϵa,ϵb,ϵc,ϵd)=1Ts(ϵaϵb)T(ϵcϵd)\displaystyle\widehat{\text{Cum}}(\bm{\epsilon}_{a},\bm{\epsilon}_{b},\bm{\epsilon}_{c},\bm{\epsilon}_{d})=\frac{1}{T_{s}}(\bm{\epsilon}_{a}\odot\bm{\epsilon}_{b})^{T}(\bm{\epsilon}_{c}\odot\bm{\epsilon}_{d})
1Ts2(ϵaTϵbϵcTϵd+ϵaTϵcϵbTϵd+ϵaTϵdϵbTϵc),\displaystyle-\frac{1}{T_{s}^{2}}\left(\bm{\epsilon}_{a}^{T}\bm{\epsilon}_{b}\bm{\epsilon}_{c}^{T}\bm{\epsilon}_{d}+\bm{\epsilon}_{a}^{T}\bm{\epsilon}_{c}\bm{\epsilon}_{b}^{T}\bm{\epsilon}_{d}+\bm{\epsilon}_{a}^{T}\bm{\epsilon}_{d}\bm{\epsilon}_{b}^{T}\bm{\epsilon}_{c}\right), (47)

for ϵa=[𝑮T]a,ϵb=[𝑮H]b,ϵc=[𝑮T]c,ϵd=[𝑮H]d\bm{\epsilon}_{a}=[\bm{G}^{T}]_{a},\bm{\epsilon}_{b}=[\bm{G}^{H}]_{b},\bm{\epsilon}_{c}=[\bm{G}^{T}]_{c},\bm{\epsilon}_{d}=[\bm{G}^{H}]_{d} for a,b,c,d=1,,La,b,c,d=1,\dots,L. In [26], the minimization of (46) w.r.t. 𝑮\bm{G} is proven to be equivalent to a joint approximate diagonalization of some cumulant matrices and solved efficiently.

The JADE algorithm is summarized as Algorithm 2. As mentioned above, the JADE algorithm transforms the maximization of (46) to a joint approximate diagonalization of some cumulant matrices. The steps (2) and (3) in Algorithm 2 are to construct the cumulant matrices, and step (4) is the joint approximate diagonalization.

Algorithm 2 JADE [26]

Input: The received signals 𝒀\bm{Y} and the number of sources LL.

(1) Calculate 𝑾^\hat{\bm{W}} as (45) and let 𝒁𝑾^𝒀\bm{Z}\equiv\hat{\bm{W}}\bm{Y}.
(2) Calculate 𝑸^z(𝒁)\hat{\bm{Q}}_{z}(\bm{Z}), given by,
[𝑸^z]p,q=Cum^([𝒁T]a,[𝒁H]b,[𝒁T]c,[𝒁H]d),[\hat{\bm{Q}}_{z}]_{p,q}=\widehat{\text{Cum}}([\bm{Z}^{T}]_{a},[\bm{Z}^{H}]_{b},[\bm{Z}^{T}]_{c},[\bm{Z}^{H}]_{d}), (48)
where Cum^()\widehat{\text{Cum}}(\cdot) is denoted by (C), p=a+(b1)Lp=a+(b-1)L and q=d+(c1)Lq=d+(c-1)L for a,b,c,d=1,,La,b,c,d=1,\dots,L.
(3) Let ϱl=[𝚺Q]l,l\varrho_{l}=[\bm{\Sigma}_{Q}]_{l,l} and 𝒓l=[𝑼Q]l\bm{r}_{l}=[\bm{U}_{Q}]_{l}, where 𝚺Q,𝑼Q\bm{\Sigma}_{Q},\bm{U}_{Q} are from the SVD of 𝑸^z\hat{\bm{Q}}_{z}, i.e., 𝑸^z=𝑼Q𝚺Q𝑼QH\hat{\bm{Q}}_{z}=\bm{U}_{Q}\bm{\Sigma}_{Q}\bm{U}_{Q}^{H} and define 𝑹lL×L\bm{R}_{l}\in\mathbb{C}^{L\times L}, s.t., vec(𝑹l)=𝒓l\text{vec}(\bm{R}_{l})=\bm{r}_{l} for l=1,,Ll=1,\dots,L.
(4) Jointly diagonalize {𝑹~l=ϱl𝑹l1lL}\mathcal{R}\equiv\left\{\widetilde{\bm{R}}_{l}=\varrho_{l}\bm{R}_{l}\mid 1\leqslant l\leqslant L\right\} by a unitary matrix 𝑽\bm{V}, given in Algorithm 3.
(5) Estimate 𝑯\bm{H} as 𝑯^=𝑽H𝑾^𝒀\hat{\bm{H}}=\bm{V}^{H}\hat{\bm{W}}\bm{Y}.

Output: The waveform estimation 𝑯^\hat{\bm{H}}.

Algorithm 3 Joint approximate diagonalization

Input: The matrix set \mathcal{R} and the number of sources LL.

For m=1,,L1m=1,\dots,L-1: For n=m+1,,Ln=m+1,\dots,L: 1) Let 𝑶m,n=[𝒐m,n1,,𝒐m,nL]T\bm{O}_{m,n}=[\bm{o}_{m,n}^{1},\dots,\bm{o}_{m,n}^{L}]^{T}, where
𝒐m,nl=[\displaystyle\bm{o}_{m,n}^{l}=\big{[} [𝑹~l]m,m[𝑹~l]n,n,[𝑹~l]m,n+[𝑹~l]n,m,\displaystyle[\tilde{\bm{R}}_{l}]_{m,m}-[\tilde{\bm{R}}_{l}]_{n,n},[\tilde{\bm{R}}_{l}]_{m,n}+[\tilde{\bm{R}}_{l}]_{n,m},
ȷ([𝑹~l]n,m[𝑹~l]m,n)]T.\displaystyle\jmath([\tilde{\bm{R}}_{l}]_{n,m}-[\tilde{\bm{R}}_{l}]_{m,n})\big{]}^{T}. (49)
2) Let 𝜼m,n=[𝑼o]1\bm{\eta}_{m,n}=[\bm{U}_{o}]_{1}, where 𝑼o\bm{U}_{o} is from the SVD as Re(𝑶m,nH𝑶m,n)=𝑼o𝚺o𝑼oH{\rm Re}(\bm{O}_{m,n}^{H}\bm{O}_{m,n})=\bm{U}_{o}\bm{\Sigma}_{o}\bm{U}_{o}^{H}. 3) Calculate αm,n\alpha_{m,n} and βm,n\beta_{m,n} as
αm,n\displaystyle\alpha_{m,n} =1+[𝜼m,n]1,12,\displaystyle=\sqrt{\frac{1+[\bm{\eta}_{m,n}]_{1,1}}{2}},
βm,n\displaystyle\beta_{m,n} =[𝜼m,n]2,1ȷ[𝜼m,n]3,12αm,n.\displaystyle=\frac{[\bm{\eta}_{m,n}]_{2,1}-\jmath[\bm{\eta}_{m,n}]_{3,1}}{2\alpha_{m,n}}. (50)
4) Define 𝑮m,n\bm{G}_{m,n} by
𝑮m,n=\displaystyle\bm{G}_{m,n}= 𝑰+βm,n𝒆n𝒆mTβm,n𝒆m𝒆nT\displaystyle\bm{I}+\beta_{m,n}\bm{e}_{n}\bm{e}_{m}^{T}-\beta_{m,n}^{*}\bm{e}_{m}\bm{e}_{n}^{T}
+(αm,n1)(𝒆m𝒆mT𝒆n𝒆nT),\displaystyle+(\alpha_{m,n}-1)(\bm{e}_{m}\bm{e}_{m}^{T}-\bm{e}_{n}\bm{e}_{n}^{T}), (51)
where 𝒆L×1\bm{e}_{\cdot}\in\mathbb{C}^{L\times 1} is the unit vector. 5) Update the set \mathcal{R} as
𝑹~l𝑮m,nH𝑹~l𝑮m,n,l=1,,L.\tilde{\bm{R}}_{l}\leftarrow\bm{G}_{m,n}^{H}\tilde{\bm{R}}_{l}\bm{G}_{m,n},\ l=1,\dots,L. (52)
end
end Denote 𝑽\bm{V} by
𝑽=m=1L1n=m+1L𝑮m,n.\bm{V}=\prod_{m=1}^{L-1}\prod_{n=m+1}^{L}\bm{G}_{m,n}. (53)

Output: The unitary matrix 𝑽\bm{V}.

References

  • [1] H. Krim and M. Viberg, “Two decades of array signal processing research: the parametric approach,” IEEE Signal Processing Magazine, vol. 13, no. 4, pp. 67–94, 1996.
  • [2] H. L. Van Trees, Optimum array processing: Part IV of detection, estimation, and modulation theory.   John Wiley & Sons, 2004.
  • [3] R. Heimiller, J. Belyea, and P. Tomlinson, “Distributed array radar,” IEEE Transactions on Aerospace and Electronic Systems, vol. AES-19, no. 6, pp. 831–839, 1983.
  • [4] Q. H. Wang, T. Ivanov, and P. Aarabi, “Acoustic robot navigation using distributed microphone arrays,” Information Fusion, vol. 5, no. 2, pp. 131–140, 2004, robust Speech Processing. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S156625350300085X
  • [5] D. Jenn, Y. Loke, T. C. H. Matthew, Y. E. Choon, O. C. Siang, and Y. S. Yam, “Distributed phased arrays and wireless beamforming networks,” International Journal of Distributed Sensor Networks, vol. 5, no. 4, pp. 283–302, 2009.
  • [6] A. Anton, I. Garcia-Rojo, A. Girón, E. Morales, and R. Martínez, “Analysis of a distributed array system for satellite acquisition,” IEEE Transactions on Aerospace and Electronic Systems, vol. 53, no. 3, pp. 1158–1168, 2017.
  • [7] S. Sun, H. Lin, J. Ma, and X. Li, “Multi-sensor distributed fusion estimation with applications in networked systems: A review paper,” Information Fusion, vol. 38, pp. 122–134, 2017. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S1566253517300362
  • [8] S. Mghabghab and J. A. Nanzer, “A self-mixing receiver for wireless frequency synchronization in coherent distributed arrays,” in 2020 IEEE/MTT-S International Microwave Symposium (IMS).   IEEE, 2020, pp. 1137–1140.
  • [9] P. Stoica, M. Viberg, Kon Max Wong, and Qiang Wu, “Maximum-likelihood bearing estimation with partly calibrated arrays in spatially correlated noise fields,” IEEE Transactions on Signal Processing, vol. 44, no. 4, pp. 888–899, 1996.
  • [10] E. H. T. Collaboration, K. Akiyama, A. Alberdi, W. Alef, K. Asada, R. AZULY et al., “First m87 event horizon telescope results. i. the shadow of the supermassive black hole,” Astrophys. J. Lett, vol. 875, no. 1, p. L1, 2019.
  • [11] C. Steffens and M. Pesavento, “Block- and rank-sparse recovery for direction finding in partly calibrated arrays,” IEEE Transactions on Signal Processing, vol. 66, no. 2, pp. 384–399, 2018.
  • [12] A. Adler and M. Wax, “Direct localization by partly calibrated arrays: A relaxed maximum likelihood solution,” in 2019 27th European Signal Processing Conference (EUSIPCO).   IEEE, 2019, pp. 1–5.
  • [13] W. Suleiman, P. Parvazi, M. Pesavento, and A. M. Zoubir, “Non-coherent direction-of-arrival estimation using partly calibrated arrays,” IEEE Transactions on Signal Processing, vol. 66, no. 21, pp. 5776–5788, 2018.
  • [14] M. Wax and T. Kailath, “Decentralized processing in sensor arrays,” IEEE Transactions on Acoustics, Speech, and Signal Processing, vol. 33, no. 5, pp. 1123–1129, 1985.
  • [15] P. Stoica, A. Nehorai, and T. Söderström, “Decentralized array processing using the mode algorithm,” Circuits, Systems and Signal Processing, vol. 14, no. 1, pp. 17–38, 1995.
  • [16] S. Sarvotham, D. Baron, M. Wakin, M. F. Duarte, and R. G. Baraniuk, “Distributed compressed sensing of jointly sparse signals,” in Asilomar conference on signals, systems, and computers, 2005, pp. 1537–1541.
  • [17] A. L. Swindlehurst, B. Ottersten, R. Roy, and T. Kailath, “Multiple invariance esprit,” IEEE Transactions on Signal Processing, vol. 40, no. 4, pp. 867–881, 1992.
  • [18] M. Pesavento, A. B. Gershman, and K. M. Wong, “Direction finding in partly calibrated sensor arrays composed of multiple subarrays,” IEEE Transactions on Signal Processing, vol. 50, no. 9, pp. 2103–2115, 2002.
  • [19] C. M. S. See and A. B. Gershman, “Direction-of-arrival estimation in partly calibrated subarray-based sensor arrays,” IEEE Transactions on Signal Processing, vol. 52, no. 2, pp. 329–338, 2004.
  • [20] P. Parvazi, M. Pesavento, and A. B. Gershman, “Direction-of-arrival estimation and array calibration for partly-calibrated arrays,” in 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2011, pp. 2552–2555.
  • [21] H.-L. Ding, W.-B. Zhao, and L.-Z. Zhang, “Tracking number time-varying nonlinear targets based on squf-gmphda in radar networking,” in Intelligent Computing Theories and Application, D.-S. Huang and K.-H. Jo, Eds.   Cham: Springer International Publishing, 2016, pp. 825–837.
  • [22] C. Steffens, P. Parvazi, and M. Pesavento, “Direction finding and array calibration based on sparse reconstruction in partly calibrated arrays,” in 2014 IEEE 8th Sensor Array and Multichannel Signal Processing Workshop (SAM), 2014, pp. 21–24.
  • [23] Y. C. Eldar and M. Mishali, “Robust recovery of signals from a structured union of subspaces,” IEEE Transactions on Information Theory, vol. 55, no. 11, pp. 5302–5316, 2009.
  • [24] Y. C. Eldar, P. Kuppinger, and H. Bolcskei, “Block-sparse signals: Uncertainty relations and efficient recovery,” IEEE Transactions on Signal Processing, vol. 58, no. 6, pp. 3042–3054, 2010.
  • [25] S. Choi, A. Cichocki, H.-M. Park, and S.-Y. Lee, “Blind source separation and independent component analysis: A review,” Neural Information Processing-Letters and Reviews, vol. 6, no. 1, pp. 1–57, 2005.
  • [26] J.-F. Cardoso and A. Souloumiac, “Blind beamforming for non-gaussian signals,” in IEE proceedings F (radar and signal processing), vol. 140, no. 6.   IET, 1993, pp. 362–370.
  • [27] Y. Wang and K. C. Ho, “Unified near-field and far-field localization for aoa and hybrid aoa-tdoa positionings,” IEEE Transactions on Wireless Communications, vol. 17, no. 2, pp. 1242–1254, 2018.
  • [28] W. Liao and A. Fannjiang, “Music for single-snapshot spectral estimation: Stability and super-resolution,” Applied and Computational Harmonic Analysis, vol. 40, no. 1, pp. 33–67, 2016. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S1063520314001432
  • [29] Y. C. Eldar and G. Kutyniok, Compressed sensing: theory and applications.   Cambridge university press, 2012.
  • [30] P. Stoica, P. Babu, and J. Li, “Spice: A sparse covariance-based estimation method for array processing,” IEEE Transactions on Signal Processing, vol. 59, no. 2, pp. 629–638, 2011.
  • [31] H. Akaike, “A new look at the statistical model identification,” IEEE transactions on automatic control, vol. 19, no. 6, pp. 716–723, 1974.
  • [32] G. Schwarz, “Estimating the dimension of a model,” The annals of statistics, pp. 461–464, 1978.
  • [33] M. Wax and T. Kailath, “Detection of signals by information theoretic criteria,” IEEE Transactions on acoustics, speech, and signal processing, vol. 33, no. 2, pp. 387–392, 1985.
  • [34] R. Schmidt, “Multiple emitter location and signal parameter estimation,” IEEE transactions on antennas and propagation, vol. 34, no. 3, pp. 276–280, 1986.
  • [35] Y. C. Eldar, Sampling theory: Beyond bandlimited systems.   Cambridge University Press, 2015.
  • [36] R. Tibshirani, “Regression shrinkage and selection via the lasso,” Journal of the royal statistical society series b-methodological, vol. 58, no. 1, pp. 267–288, 1996.
  • [37] M. Rossi, A. M. Haimovich, and Y. C. Eldar, “Spatial compressive sensing for mimo radar,” IEEE Transactions on Signal Processing, vol. 62, no. 2, pp. 419–430, 2014.
  • [38] J. A. Tropp, “A mathematical introduction to compressive sensing [book review],” Bulletin of the American Mathematical Society, vol. 54, no. 1, pp. 151–165, 2017.
  • [39] P. E. Gill and W. Murray, “Algorithms for the solution of the nonlinear least-squares problem,” SIAM Journal on Numerical Analysis, vol. 15, no. 5, pp. 977–992, 1978.
  • [40] N. Boumal and P.-A. Absil, “Low-rank matrix completion via preconditioned optimization on the grassmann manifold,” Linear Algebra and its Applications, vol. 475, pp. 200–239, 2015.
  • [41] S. Smith, “Statistical resolution limits and the complexified cramér-rao bound,” IEEE Transactions on Signal Processing, vol. 53, no. 5, pp. 1597–1609, 2005.
  • [42] A. G. Asuero, A. Sayago, and A. Gonzalez, “The correlation coefficient: An overview,” Critical reviews in analytical chemistry, vol. 36, no. 1, pp. 41–59, 2006.
  • [43] J. Huang, T. Zhang et al., “The benefit of group sparsity,” The Annals of Statistics, vol. 38, no. 4, pp. 1978–2004, 2010.
  • [44] Y. Chi, L. L. Scharf, A. Pezeshki, and A. R. Calderbank, “Sensitivity to basis mismatch in compressed sensing,” IEEE Transactions on Signal Processing, vol. 59, no. 5, pp. 2182–2195, 2011.
  • [45] M. Grant and S. Boyd, “Cvx: Matlab software for disciplined convex programming, version 2.1,” 2014.