This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Online unsupervised deep unfolding
for MIMO channel estimation

Luc Le Magoarou, Stéphane Paquelet b<><>com
Rennes, France
Abstract

Channel estimation is a difficult problem in MIMO systems. Using a physical model allows to ease the problem, injecting a priori information based on the physics of propagation. However, such models rest on simplifying assumptions and require to know precisely the system configuration, which is unrealistic. In this paper, we propose to perform online learning for channel estimation in a massive MIMO context, adding flexibility to physical models by unfolding a channel estimation algorithm (matching pursuit) as a neural network. This leads to a computationally efficient neural network that can be trained online when initialized with an imperfect model. The method allows a base station to automatically correct its channel estimation algorithm based on incoming data, without the need for a separate offline training phase. It is applied to realistic channels and shows great performance, achieving channel estimation error almost as low as one would get with a perfectly calibrated system.

Index Terms:
Autoencoders, deep unfolding, MIMO channel estimation, online learning.

I Introduction

Data processing techniques are often based on the manifold assumption: Meaningful data (signals) lie near a low dimensional manifold, although their apparent dimension is much larger [1, 2].

In MIMO channel estimation, using a physical model amounts to parameterize a manifold by physical parameters such as the directions, delays and gains of the propagation paths, the dimension of the manifold being equal to the number of real parameters considered in the model. Physical models allow to inject strong a priori knowledge based on solid principles [3, 4], but necessarily make simplifying assumptions (e.g. the plane wave assumption [5]) and require exact knowledge of the system configuration (positions of the antennas, gains, etc.).

On the other hand, machine learning techniques have recently led to tremendous successes in various domains [6]. Their main feature is to learn the data representation (manifold) directly on training data, without requiring any specific a priori knowledge. This flexibility in the manifold construction comes at the price of computationally heavy learning and difficulties to inject knowledge on the problem at hand.

Recently, it has been proposed to unfold iterative inference algorithms so as to express them as neural networks that can be optimized [7, 8]. This has the advantage of adding flexibility to algorithms based on classical models, and amounts to constrain the search for the appropriate manifold with a priori knowledge on the problem at hand. Moreover, this leads to inference algorithms of reduced complexity [9].

Contributions. In this letter, we propose to perform online learning for channel estimation in a massive MIMO context. Starting from an imperfect physical channel model, our method allows a base station to automatically correct its channel estimation algorithm based on incoming data, without the need for a separate offline training phase. It is based on the unfolding of the matching pursuit algorithm, which is simple and computationally efficient. The obtained neural network is trained in an unsupervised way. The overall complexity of the forward and backward passes in the network is of the same order as performing channel estimation only (without any learning), which makes online learning feasible. Such a method is particularly suited to imperfectly known or non-calibrated systems. Note that since this paper has been written, we further developed the introduced ideas in a longer paper [10] (still in a preprint state, not submitted anywhere). This longer paper introduces an automatic adaptation to the signal to noise ratio (SNR) and demonstrates several potential applications of the method.

Related work. Machine learning holds promise for wireless communications (see [11, 12] for exhaustive surveys). It has recently been proposed to use adaptive data representations for MIMO channel estimation using dictionary learning techniques [13]. However, dictionary learning with algorithms such as K-SVD [14] as proposed in [13] is very computationally heavy, and thus not suited to online learning.

Deep unfolding has also been considered by communication researchers (see [15] and references therein). It has been proposed in [16] to perform channel estimation in a massive MIMO context, based on the unfolding of a sparse recovery algorithm (namely denoising-based approximate message passing [17]). However, the method is directly adapted from image processing and does not make use of a physical channel model as initialization. A recent work also proposes to use deep unfolding for channel estimation [18], but using a physical model to optimize the shrinkage function. However, previously proposed methods based on unfolding all require an offline training phase and are of high complexity compared to classical methods [19].

The main novelty of this letter is the online nature of the method, that does not require a separate offline learning phase, since learning is done while using the channel estimation algorithm that corrects itself over time. This is made possible by the very low complexity of the considered estimation algorithm.

II Problem formulation

System settings. We consider in this letter a massive MIMO system, also known as multi-user MIMO (MU-MIMO) system [20], in which a base station equipped with NN antennas communicates with KK single antenna users (K<NK<N). The system operates in time division duplex (TDD) mode, so that channel reciprocity holds and the channel is estimated in the uplink: each user sends a pilot sequence 𝐩k\mathbf{p}_{k} (orthogonal to the sequences of the other users, 𝐩kH𝐩l=δkl\mathbf{p}_{k}^{H}\mathbf{p}_{l}=\delta_{kl}) for the base station to estimate the channel. The received signal is thus expressed 𝐑=k=1K𝐡k𝐩kH+𝐍\mathbf{R}=\sum_{k=1}^{K}\mathbf{h}_{k}\mathbf{p}_{k}^{H}+\mathbf{N}, where 𝐍\mathbf{N} is noise. After correlating the received signal with the pilot sequences, and assuming no pilot contamination from adjacent cells for simplicity, the base station gets noisy measurements of the channels of all users, each taking the canonical form

𝐱=𝐡+𝐧,\mathbf{x}=\mathbf{h}+\mathbf{n}, (1)

where 𝐡\mathbf{h} is the channel of the considered user and 𝐧\mathbf{n} is the noise, with 𝐧𝒞𝒩(0,σ2𝐈𝐝)\mathbf{n}\sim\mathcal{CN}(0,\sigma^{2}\mathbf{Id}). We drop the user index kk here and in the following, since our approach treats the channels of all users the same way. Note that 𝐱\mathbf{x} is already an unbiased estimator of the channel, and we call it the least squares (LS) estimator in the sequel. Its performance can be assessed by the signal to noise ratio (SNR)

SNRin𝐡22Nσ2.\text{SNR}_{\text{in}}\triangleq\frac{\left\|\mathbf{h}\right\|_{2}^{2}}{N\sigma^{2}}.

However, one can get better channel estimates using a physical model, as is explained in the next paragraph.

Physical model.Let us denote {g1,,gN}\{g_{1},\dots,g_{N}\} the complex gains of the base stations antennas and {a1,,aN}\{\overrightarrow{a_{1}},\dots,\overrightarrow{a_{N}}\} their positions with respect to the centroid of the antenna array. Then, under the plane wave assumption and assuming omnidirectional antennas (isotropic radiation patterns), the channel resulting from a single propagation path with direction of arrival (DoA) u\overrightarrow{u} is proportional to the steering vector

𝐞(u)(g1ej2πλa1.u,,gNej2πλaN.u)T\mathbf{e}(\overrightarrow{u})\triangleq(g_{1}\mathrm{e}^{-\mathrm{j}\frac{2\pi}{\lambda}\overrightarrow{a_{1}}.\overrightarrow{u}},\dots,g_{N}\mathrm{e}^{-\mathrm{j}\frac{2\pi}{\lambda}\overrightarrow{a_{N}}.\overrightarrow{u}})^{T}

which reads 𝐡=β𝐞(u),\mathbf{h}=\beta\mathbf{e}(\overrightarrow{u}), with β\beta\in\mathbb{C}. In that case, a sensible estimation strategy [3, 4] is to build a dictionary of steering vectors corresponding to AA potential DoAs: 𝐄(𝐞(u1),,𝐞(uA))\mathbf{E}\triangleq\begin{pmatrix}\mathbf{e}(\overrightarrow{u_{1}}),\dots,\mathbf{e}(\overrightarrow{u_{A}})\end{pmatrix} and to compute a channel estimate with the procedure

v=argmaxui|𝐞(ui)H𝐱|,𝐡^=𝐞(v)𝐞(v)H𝐱.\displaystyle\begin{split}&\overrightarrow{v}=\text{argmax}_{\overrightarrow{u_{i}}}\,\,\,|\mathbf{e}(\overrightarrow{u_{i}})^{H}\mathbf{x}|,\\ &\hat{\mathbf{h}}=\mathbf{e}(\overrightarrow{v})\mathbf{e}(\overrightarrow{v})^{H}\mathbf{x}.\end{split} (2)

The first step of this procedure amounts to find in the dictionary the most correlated column with the observation to estimate the DoA and the second step amounts to project the observation on the corresponding steering vector. The SNR at the output of this procedure reads

SNRout𝐡22𝔼[𝐡𝐡^22],\text{SNR}_{\text{out}}\triangleq\frac{\|\mathbf{h}\|_{2}^{2}}{\mathbb{E}\big{[}\|\mathbf{h}-\hat{\mathbf{h}}\|_{2}^{2}\big{]}},

and we have at best SNRout=NSNRin\text{SNR}_{\text{out}}=N\text{SNR}_{\text{in}} (neglecting the discretization error).

Note that the evoked strategy can be generalized to multipath channels of the form 𝐡=p=1Pβp𝐞(up),\mathbf{h}=\sum\nolimits_{p=1}^{P}\beta_{p}\mathbf{e}(\overrightarrow{u_{p}}), using greedy sparse recovery algorithms such as matching pursuit (MP) or orthogonal matching pursuit (OMP) [21].

Refer to caption
Figure 1: SNR loss in decibels (dB) due to imperfect knowledge of the system.

III Impact of imperfect models

The estimation strategy based on a physical model requires knowing the system configuration (antenna gains and positions) and necessarily relies on hypotheses. What happens if the configuration is imperfectly known or if some hypotheses are not valid? In order to answer this simple question, let us perform an experiment. Consider an antenna array of N=64N=64 antennas at the base station, whose known nominal configuration is an uniform linear array (ULA) of unit gain antennas separated by half-wavelengths and aligned with the xx-axis. This nominal configuration corresponds to gains and positions {g~i,ai~}i=1N\{\tilde{g}_{i},\tilde{\overrightarrow{a_{i}}}\}_{i=1}^{N}. Now, suppose the knowledge of the system configuration is imperfect, meaning that the unknown true configuration of the system is given by the gains and positions {gi,ai}i=1N\{g_{i},\overrightarrow{a_{i}}\}_{i=1}^{N}, with

gi=g~i+ng,i,ng,i𝒞𝒩(0,σg2),ai=ai~+λ𝐧p,i,𝐧p,i=(ep,i,0,0)T,ep,i𝒩(0,σp2).\displaystyle\begin{split}&g_{i}=\tilde{g}_{i}+n_{g,i},\,n_{g,i}\sim\mathcal{CN}(0,\sigma_{g}^{2}),\\ &\overrightarrow{a_{i}}=\tilde{\overrightarrow{a_{i}}}+\lambda\mathbf{n}_{p,i},\,\mathbf{n}_{p,i}={\small\begin{pmatrix}e_{p,i},&0,&0\end{pmatrix}^{T}},e_{p,i}\sim\mathcal{N}(0,\sigma_{p}^{2}).\end{split} (3)

This way, σg\sigma_{g} (resp. σp\sigma_{p}) quantifies the uncertainty about the antenna gains (resp. spacings). Moreover, let

𝐞~(u)(g~1ej2πλa1~.u,,g~Nej2πλaN~.u)T\tilde{\mathbf{e}}(\overrightarrow{u})\triangleq(\tilde{g}_{1}\mathrm{e}^{-\mathrm{j}\frac{2\pi}{\lambda}\tilde{\overrightarrow{a_{1}}}.\overrightarrow{u}},\dots,\tilde{g}_{N}\mathrm{e}^{-\mathrm{j}\frac{2\pi}{\lambda}\tilde{\overrightarrow{a_{N}}}.\overrightarrow{u}})^{T}

be the nominal steering vector and 𝐄~(𝐞~(u1),,𝐞~(uA))\tilde{\mathbf{E}}\triangleq\begin{pmatrix}\tilde{\mathbf{e}}(\overrightarrow{u_{1}}),\dots,\tilde{\mathbf{e}}(\overrightarrow{u_{A}})\end{pmatrix} be a dictionary of nominal steering vectors. The experiment consists in comparing the estimation strategy of (2) using the true (perfect but unknown) dictionary 𝐄\mathbf{E} with the exact same strategy using the nominal (imperfect but known) dictionary 𝐄~\tilde{\mathbf{E}}. To do so, we generate measurements according to (1) with channels of the form 𝐡=𝐞(u)\mathbf{h}=\mathbf{e}(\overrightarrow{u}) where u\overrightarrow{u} corresponds to azimuth angles chosen uniformly at random, and SNRin\text{SNR}_{\text{in}} is set to 10dB10\,\text{dB}. Then, the dictionaries 𝐄\mathbf{E} and 𝐄~\tilde{\mathbf{E}} are built by choosing A=32NA=32N directions corresponding to evenly spaced azimuth angles. Let 𝐡^𝐄\hat{\mathbf{h}}_{\mathbf{E}} be the estimate obtained using 𝐄\mathbf{E} in (2), and 𝐡^𝐄~\hat{\mathbf{h}}_{\tilde{\mathbf{E}}} the estimate obtained using 𝐄~\tilde{\mathbf{E}}. The SNR loss caused by using 𝐄~\tilde{\mathbf{E}} instead of 𝐄\mathbf{E} is measured by the quantity 𝐡^𝐄~𝐡22/𝐡^𝐄𝐡22\|\hat{\mathbf{h}}_{\tilde{\mathbf{E}}}-\mathbf{h}\|_{2}^{2}/\|\hat{\mathbf{h}}_{\mathbf{E}}-\mathbf{h}\|_{2}^{2}. Results in terms of SNR loss, in average over 1010 antenna array realizations and 10001000 channel realizations per antenna array realization are shown on figure 1. From the figure, it is obvious that even a relatively small uncertainty about the system configuration can cause a great SNR loss. For example, an uncertainty of 0.03λ0.03\lambda on the antenna spacings and of 0.090.09 on the antenna gains leads to an SNR loss of more than 10dB10\,\text{dB}, which means that the mean squared error is increased more than ten times. This experiment highlights the fact that using imperfect models can severely harm estimation performance. The main contribution of this letter is to propose a way to correct imperfect physical models using machine learning.

IV Deep unfolding strategy

Let us now propose a strategy based on deep unfolding allowing to correct a channel estimation algorithm based on an imperfect physical model incrementally, via online learning.

IV-A Basic principle

Unfolding. The estimation strategy of (2) can be unfolded as a neural network taking the observation 𝐱\mathbf{x} as input and outputting a channel estimate 𝐡^\hat{\mathbf{h}}. Indeed, the first step amounts to perform a linear transformation (multiplying the input by the matrix 𝐄H\mathbf{E}^{H}) followed by a nonlinear one (finding the inner product of maximum amplitude and setting all the others to zero) and the second step corresponds to a linear transformation (multiplying by the matrix 𝐄\mathbf{E}). Such a strategy is parameterized by the dictionary of steering vectors 𝐄\mathbf{E}. In the case where the dictionary 𝐄\mathbf{E} is unknown (or imperfectly known), we propose to learn the matrix used in (2) directly on data via backpropagation [22], using as initialization the matrix 𝐄~\tilde{\mathbf{E}} corresponding to the imperfect physical model.

Neural network structure. Such a neural network structure corresponds to the kk-sparse autoencoder [23], which has originally been introduced for image classification. The deep unfolding of channel estimation using a physical model as in (2) corresponds to use a kk-sparse autoencoder, setting the sparsity parameter to k=1k=1. This neural network structure is shown on figure 2, where HT1\text{HT}_{1} refers to the hard thresholding operator which keeps only the entry of greatest modulus of its input and sets all the others to zero. The parameters of this neural network are the weights 𝐖N×A\mathbf{W}\in\mathbb{C}^{N\times A}. Note that complex weights and inputs are handled classically by stacking the real and imaginary parts for vectors and using the real representation for matrices.

Refer to caption
Figure 2: Deep unfolding for single path channel estimation.

Training. The method we propose to jointly estimate channels while simultaneously correcting an imperfect physical model amounts to initialize the network of figure 2 with a dictionary of nominal steering vectors 𝐄~\tilde{\mathbf{E}} and then to perform a minibatch gradient descent [24] on the cost function 12𝐱𝐡^22\tfrac{1}{2}\|\mathbf{x}-\hat{\mathbf{h}}\|_{2}^{2} to update the weights 𝐖\mathbf{W} in order to correct the model. It operates online, on streaming observations 𝐱i,i=1,,\mathbf{x}_{i},\,i=1,\dots,\infty of the form (1) acquired over time (coming from all users simultaneously). Note that, as opposed to the classical unfolding strategies [7, 8], the proposed method is totally unsupervised, meaning that it requires only noisy channel observations and no clean channels to run.

Implementation details. In all the experiments performed in this letter, we use minibatches of 200200 observations and the Adam optimization algorithm [25] with an exponentially decreasing learning rate starting at 0.0010.001 and being multiplied by 0.90.9 every 200200 minibatchs. Moreover, the method was found to perform better with the input data being normalized. To simplify notation, we denote abusively also 𝐱i,i=1,,\mathbf{x}_{i},\,i=1,\dots,\infty the data after normalization.

IV-B Generalization to multipath channels

Real channels are often not made of a single path, in which case the proposed method becomes suboptimal. Indeed, it uses a kk-sparse autoencoder with k=1k=1, implicitly assuming a single path. However, real world channels are often sparse (well approximated by only a few paths). This is particularly true at millimeter wave frequencies [26]. In order to adapt the unfolding strategy to such channels, we propose to apply recursively the structure of figure 2, subtracting at each step the current output from the observation, exactly mimicking the matching pursuit (MP) algorithm [21]. The number KK of times the structure is replicated (depth of the network) corresponds to the number of estimated paths. The neural network corresponding to such a strategy is schematized on figure 3, we call it mpNet (for matching pursuit network). It is trained exactly as the network of figure 2, with tied weights across iterations (we tried to untie the weights but observed no improvement) and cost function 12𝐱𝐡^22=12𝐫K22\tfrac{1}{2}\|\mathbf{x}-\hat{\mathbf{h}}\|_{2}^{2}=\tfrac{1}{2}\|\mathbf{r}_{K}\|_{2}^{2}.

Refer to caption
Figure 3: mpNet: Unfolding for multipath channel estimation.
Refer to caption
Refer to caption
Refer to caption
Figure 4: Channel estimation performance on synthetic realistic channels for various SNRs and model imperfections.

Computational complexity. Note that the forward pass in mpNet costs 𝒪(KNA)\mathcal{O}(KNA) arithmetic operations and the backpropagation step cotst 𝒪(KN)\mathcal{O}(KN) arithmetic operations (AA times less). This means that jointly learning the model and estimating the channel (computing the forward and backward passes) is done at a cost that is the same order as the one of simply estimating the channel with a greedy algorithm (MP or OMP) without adapting the model at all to data (which corresponds to computing only the forward pass). This very light computational cost makes the method adapted to online learning, as opposed to previously proposed channel estimation strategies based on deep unfolding [16, 18, 19].

IV-C Experiment

Setting. Let us now assess mpNet on realistic channels. To do so, we consider the SSCM channel model [26] in order to generate non-line-of-sight (NLOS) channels at 28GHz28\,\text{GHz} (see [26, table IV]) corresponding to all users. We consider the same setting as in section III, namely a base station equipped with an ULA of 6464 antennas, with an half-wavelength nominal spacing and unit nominal gains used to build the imperfect nominal dictionary 𝐄~\tilde{\mathbf{E}} (with A=8NA=8N) which serves as an initialization for mpNet. The actual antenna arrays are generated the same way as in section III, using (3), and are kept fixed for the whole experiment. We consider two model imperfections: σp=0.05,σg=0.15\sigma_{p}=0.05,\,\sigma_{g}=0.15 (small uncertainty) and σp=0.1,σg=0.3\sigma_{p}=0.1,\,\sigma_{g}=0.3 (large uncertainty) to build the unknown ideal dictionary 𝐄\mathbf{E}. The input SNR takes the values {5,10}dB\{5,10\}\,\text{dB} while the parameter KK (controlling the depth of mpNet) is set to {6,8}\{6,8\} respectively (determined by cross validation). The proposed method is compared to the least squares estimator and to the OMP algorithm with KK iterations using either the imperfect nominal dictionary 𝐄~\tilde{\mathbf{E}} or the unknown ideal dictionary 𝐄\mathbf{E}. In order to show the interest of the imperfect model initialization, we also compare the proposed method to mpNet using a random (Gaussian) initialization. This baseline correspond to a classical online dictionary learning method [27].

Results. The results of this experiment are shown on figure 4 as a function of the number of channels of the form (1) seen by the base station over time. The performance measure is the relative mean squared error (rMSE=𝐡^𝐡22/𝐡22\text{rMSE}=\|\hat{\mathbf{h}}-\mathbf{h}\|_{2}^{2}/\|\mathbf{h}\|_{2}^{2}) averaged over minibatches of 200200 channels. First of all, the imperfect model is shown to be well corrected by mpNet, the blue curve being very close to the green one (ideal unknown dictionary) after a certain amount of time. This is true both for a small uncertainty and for a large one and at all tested SNRs. Note that using the nominal dictionary (initialization of mpNet) may be even worse than the least squares method, showing the interest of correcting the model, since with learning mpNet always ends up outperforming the least squares. Second, comparing the leftmost and center figures, it is interesting to notice that learning is faster and the attained performance is better with a large SNR (the blue and green curves get closer, faster), which can be explained by the better quality of data used to train the model. Third, comparing the leftmost and rightmost figures, it is apparent that a smaller uncertainty, which means a better initialization since the nominal dictionary is closer to the ideal unknown dictionary, leads to a faster convergence, but obviously also to a smaller improvement. Finally, comparing the blue and orange curves on all figures, it is apparent that initialization matters. Indeed, the random initialization performs much worse than the initialization with the nominal dictionary and takes longer to converge. These conclusions are very promising and highlight the applicability of the proposed method.

V Conclusion and perspectives

In this paper, we proposed a method to add flexibility to physical models used for MIMO channel estimation. It is based on the deep unfolding strategy that views classical algorithms as neural networks. The proposed method was shown to correct incrementally (via online learning) an imperfect or imperfectly known physical model in order to make channel estimation as efficient as if the unknown ideal model were known. This claim was empirically validated on realistic millimeter wave outdoor channels, for various SNRs and model imperfections.

We used here uncertainty on the antenna gains and positions to illustrate physical models imperfection, but the presented method applies in principle to any imperfection (be it linear or not). For example, it could correct models in cases where the radiation pattern of the antennas differs from the nominal one, or if the plane wave assumption is not perfectly valid. Moreover, we chose to unfold the MP algorithm, but more sophisticated sparse recovery algorithm could be unfolded the same way (such as approximate message passing [28]).

References

  • [1] Gunnar Carlsson. Topology and data. Bulletin of the American Mathematical Society, 46(2):255–308, 2009.
  • [2] Gabriel Peyré. Manifold models for signals and images. Computer Vision and Image Understanding, 113(2):249–260, 2009.
  • [3] Akbar M Sayeed. Deconstructing multiantenna fading channels. IEEE Transactions on Signal Processing, 50(10):2563–2579, 2002.
  • [4] Luc Le Magoarou and Stéphane Paquelet. Parametric channel estimation for massive MIMO. In IEEE Statistical Signal Processing Workshop (SSP), 2018.
  • [5] Luc Le Magoarou, Antoine Le Calvez, and Stéphane Paquelet. Massive mimo channel estimation taking into account spherical waves. In 2019 IEEE 20th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC), pages 1–5. IEEE, 2019.
  • [6] Yann LeCun, Yoshua Bengio, and Geoffrey Hinton. Deep learning. nature, 521(7553):436–444, 2015.
  • [7] Karol Gregor and Yann LeCun. Learning fast approximations of sparse coding. In Proceedings of the 27th International Conference on International Conference on Machine Learning, pages 399–406. Omnipress, 2010.
  • [8] John R Hershey, Jonathan Le Roux, and Felix Weninger. Deep unfolding: Model-based inspiration of novel deep architectures. arXiv preprint arXiv:1409.2574, 2014.
  • [9] Vishal Monga, Yuelong Li, and Yonina C Eldar. Algorithm unrolling: Interpretable, efficient deep learning for signal and image processing. arXiv preprint arXiv:1912.10557, 2019.
  • [10] Taha Yassine and Luc Le Magoarou. mpnet: variable depth unfolded neural network for massive mimo channel estimation. arXiv preprint arXiv: 2008.04088, 2020.
  • [11] Timothy O’Shea and Jakob Hoydis. An introduction to deep learning for the physical layer. IEEE Transactions on Cognitive Communications and Networking, 3(4):563–575, 2017.
  • [12] Tianqi Wang, Chao-Kai Wen, Hanqing Wang, Feifei Gao, Tao Jiang, and Shi Jin. Deep learning for wireless physical layer: Opportunities and challenges. China Communications, 14(11):92–111, 2017.
  • [13] Yacong Ding and Bhaskar D Rao. Dictionary learning-based sparse channel representation and estimation for fdd massive mimo systems. IEEE Transactions on Wireless Communications, 17(8):5437–5451, 2018.
  • [14] M. Aharon, M. Elad, and A. Bruckstein. K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation. Signal Processing, IEEE Transactions on, 54(11):4311–4322, Nov 2006.
  • [15] Alexios Balatsoukas-Stimming and Christoph Studer. Deep unfolding for communications systems: A survey and some new directions. arXiv preprint arXiv:1906.05774, 2019.
  • [16] Hengtao He, Chao-Kai Wen, Shi Jin, and Geoffrey Ye Li. Deep learning-based channel estimation for beamspace mmwave massive mimo systems. IEEE Wireless Communications Letters, 7(5):852–855, 2018.
  • [17] Christopher A Metzler, Arian Maleki, and Richard G Baraniuk. From denoising to compressed sensing. IEEE Transactions on Information Theory, 62(9):5117–5144, 2016.
  • [18] Xiuhong Wei, Chen Hu, and Linglong Dai. Knowledge-aided deep learning for beamspace channel estimation in millimeter-wave massive mimo systems. arXiv preprint arXiv:1910.12455, 2019.
  • [19] Michel van Lier, Alexios Balatsoukas-Stimming, Henk Corporaaal, and Zoran Zivkovic. Optcomnet: Optimized neural networks for low-complexity channel estimation. arXiv preprint arXiv:2002.10493, 2020.
  • [20] Fredrik Rusek, Daniel Persson, Buon Kiong Lau, Erik G Larsson, Thomas L Marzetta, Ove Edfors, and Fredrik Tufvesson. Scaling up mimo: Opportunities and challenges with very large arrays. IEEE Signal Processing Magazine, 30(1):40–60, 2013.
  • [21] S.G. Mallat and Z. Zhang. Matching pursuits with time-frequency dictionaries. Signal Processing, IEEE Transactions on, 41(12):3397–3415, Dec 1993.
  • [22] D. E. Rumelhart, G. E. Hinton, and R. J. Williams. Learning Internal Representations by Error Propagation, page 318–362. MIT Press, Cambridge, MA, USA, 1986.
  • [23] Alireza Makhzani and Brendan Frey. K-sparse autoencoders. arXiv preprint arXiv:1312.5663, 2013.
  • [24] Léon Bottou. Large-scale machine learning with stochastic gradient descent. In Proceedings of COMPSTAT’2010, pages 177–186. Springer, 2010.
  • [25] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  • [26] Mathew K Samimi and Theodore S Rappaport. 3-d millimeter-wave statistical channel model for 5g wireless system design. IEEE Transactions on Microwave Theory and Techniques, 64(7):2207–2225, 2016.
  • [27] Julien Mairal, Francis Bach, Jean Ponce, and Guillermo Sapiro. Online learning for matrix factorization and sparse coding. Journal of Machine Learning Research, 11(1):19–60, January 2010.
  • [28] David L Donoho, Arian Maleki, and Andrea Montanari. Message-passing algorithms for compressed sensing. Proceedings of the National Academy of Sciences, 106(45):18914–18919, 2009.