This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Deep Residual Learning for Channel Estimation in Intelligent Reflecting Surface-Assisted Multi-User Communications

Chang Liu, Member, IEEE, Xuemeng Liu,
Derrick Wing Kwan Ng, Fellow, IEEE, and Jinhong Yuan, Fellow, IEEE
C. Liu, D. W. K. Ng, and J. Yuan are with the School of Electrical Engineering and Telecommunications, The University of New South Wales, Sydney, Australia. X. Liu is with The University of Sydney, Sydney, Australia.This work has been accepted in part to present at the IEEE International Conference on Communications (ICC), 2021 [1].
Abstract

Channel estimation is one of the main tasks in realizing practical intelligent reflecting surface-assisted multi-user communication (IRS-MUC) systems. However, different from traditional communication systems, an IRS-MUC system generally involves a cascaded channel with a sophisticated statistical distribution. In this case, the optimal minimum mean square error (MMSE) estimator requires the calculation of a multidimensional integration which is intractable to be implemented in practice. To further improve the channel estimation performance, in this paper, we model the channel estimation as a denoising problem and adopt a deep residual learning (DReL) approach to implicitly learn the residual noise for recovering the channel coefficients from the noisy pilot-based observations. To this end, we first develop a versatile DReL-based channel estimation framework where a deep residual network (DRN)-based MMSE estimator is derived in terms of Bayesian philosophy. As a realization of the developed DReL framework, a convolutional neural network (CNN)-based DRN (CDRN) is then proposed for channel estimation in IRS-MUC systems, in which a CNN denoising block equipped with an element-wise subtraction structure is specifically designed to exploit both the spatial features of the noisy channel matrices and the additive nature of the noise simultaneously. In particular, an explicit expression of the proposed CDRN is derived and analyzed in terms of Bayesian estimation to characterize its properties theoretically. Finally, simulation results demonstrate that the performance of the proposed method approaches that of the optimal MMSE estimator requiring the availability of the prior probability density function of channel.

Index Terms:
Intelligent reflecting surface (IRS), channel estimation, deep learning, Bayesian estimation.

I Introduction

Recently, intelligent reflecting surface (IRS), which has the capability of shaping the wireless channels between the users and the base station (BS) to enhance the system performance, has been proposed as a promising technology for the future smart radio environment [2, 3, 4, 5, 6]. In particular, thanks to the development of advanced materials, a reconfigurable and passive metasurface made of electromagnetic materials has been introduced to design the IRS to make it deployable and sustainable in practice [7, 8]. Generally, an IRS is composed of a large number of passive reflecting elements while each element is reconfigurable and can be independently controlled to adapt its phase shift to the actual environment to alter the reflection of the incident signals. As such, by jointly adjusting the phase shifts of all the passive elements, a desirable reflection pattern can be obtained which establishes a favourable wireless channel to improve the transmission quality with a low system power consumption [9, 10, 11]. Based on this, an IRS can be adopted in communication systems to enhance the energy efficiency or the spectral efficiency of communication networks through passive beamforming techniques, i.e., designing an efficient configuration of phase shifts to improve the received signal-to-noise ratio (SNR) at the desired receivers [12]. Therefore, IRS-assisted communication systems and the related studies such as the network capacity or received SNR maximization [13, 14], the energy efficiency or spectral efficiency maximization [15, 16], and the IRS-assisted physical layer security [17, 18], etc., have drawn vast attention from both the academia and the industry.

In practice, the promised performance gain brought by an IRS relies on accurate channel state information (CSI) for beamforming. Yet, the aforementioned studies were conducted under the assumption of perfect knowledge of CSI, which is usually not available. In fact, an indispensable task for realizing IRS-assisted communication systems is to perform accurate channel estimation. In contrast to the channel estimation in traditional systems, the IRS is passive and cannot perform training sequence transmission/reception or signal processing, i.e., the channel of IRS to user/BS is generally not available separately and only a cascaded channel of user-to-IRS-to-BS can be estimated. More importantly, the cascaded channel brings two main challenges to IRS-assisted communication systems which are listed as follows: (i) Limited channel estimation accuracy: Note that the cascaded user-to-IRS-to-BS channel does not follow the conventional Rayleigh fading model. In this case, the optimal minimum mean square error (MMSE) estimator involves a multidimensional integration which is overly computationally intensive for practical implementation. Meanwhile, the performance of the available linear MMSE (LMMSE) and least squares (LS) estimators still has a large gap compared with that of the optimal MMSE estimator. Thus, the channel estimation accuracy is unsatisfactory for practical IRS-assisted communication systems. (ii) Large channel estimation training overhead: The IRS generally consists of a large number of elements. Thus, the cascaded channel is with a high dimension and the corresponding channel estimation via conventional methods, e.g., the LS method and the LMMSE method, is computationally costly.

To overcome the challenge of limited channel estimation performance in IRS-assisted systems (i.e., challenge (i)), a variety of effective algorithms and efficient schemes have been proposed recently. For example, in [19], a binary reflection controlled least squares (LS) channel estimation scheme was developed for single-user systems by switching on only one reflecting element of the IRS and switching off the rest reflecting elements for each time slot. In this case, the BS only receives interference from the direct link and does not receive any interference from the other reflection elements [20]. Thus, the BS can estimate the cascaded channel successively, which paves the way for the channel estimation in IRS-assisted systems. However, since the binary reflection scheme only active one element each time, only a small received SNR can be obtained at the BS for estimation. Besides, an exceeding long delay may introduce to the system if there is a large number of IRS elements. To further improve the received SNR and shorten the required time for channel estimation, [21] proposed to switch on all the reflecting elements of the IRS for each time slot. In particular, the authors proposed a discrete Fourier transform (DFT) training sequence-based minimum variance unbiased estimator which can achieve satisfactory estimation accuracy. Moreover, the authors in [22] developed two parallel factor (PARAFAC)-based methods where the LS estimator was first adopted to obtain coarse channel coefficients and then more accurate estimation results can be obtained through exploiting the least squares Khatri-Rao factorization and the bilinear estimation techniques. On the other hand, some initial attempts have been devoted to the design of efficient schemes to reduce the required training overhead of channel estimation (i.e., challenge (ii)). For instance, [23] and [24] developed a subsurface-based channel estimation scheme where the IRS is divided into several independent subsurfaces and each subsurface is composed of multiple adjacent reflecting elements applying one common phase shift. Hence, by adopting the sharing strategy, the training overhead is dramatically reduced. Besides, [25] introduced a cascaded channel estimation framework and proposed the related algorithm based on the sparse matrix factorization and the matrix completion. Moreover, [26] studied the channel estimation in IRS-aided multiple-input multiple-output (MIMO) systems. Specifically, the authors first formulated the channel estimation as a problem of recovering a sparse channel matrix and then proposed a compressed sensing-based scheme to explore the sparsity of the cascaded channel. Also, in [27] and [28], the authors investigated the application of an IRS to a multi-user system and developed a three-phase channel estimation framework to shorten the required time duration for channel estimation by exploiting the redundancy in the reflecting channels of different users. Different from the previous works, [29] developed a novel IRS architecture equipped with a single active radio frequency (RF) chain for channel estimation at the IRS side, which further shortens the training overhead. However, despite the significant research efforts devoted in the aforementioned literature, the problem of limited channel estimation performance in (i) has not been well addressed, thus an effective and practical algorithm which can further improve the estimation accuracy is expected.

Recently, in contrast to the traditional model-driven approaches, data-driven deep learning (DL) techniques [30, 31, 32] have proved their effectiveness in IRS-assisted communication systems, such as the DL-based passive beamforming design [33, 34, 35, 36], the deep reinforcement learning (DRL)-based phase shift optimization [37], and the DRL-based secure wireless communications for IRS-MUC systems [38]. Although these DL-based methods are promising, they still require the availability of perfect CSI for implementation in IRS-assisted systems. Note that the channel estimation problem is essentially a denoising problem. In particular, the DL technique, especially the deep residual learning (DReL) which adopts a deep residual network (DRN) to improve the network performance and accelerate the network training speed, has been recognized by its powerful capability in denoising in various research areas [39, 40, 41]. Motivated by these facts, in this paper, we focus on IRS-assisted multi-user communication (IRS-MUC) systems and adopt a DReL approach to intelligently exploit the channel features to further improve the channel estimation accuracy. Note that different from the existing work in [33] where the IRS requires the installation of additional active channel sensors, our work focuses on a general IRS without deploying any active sensors to address the challenging channel estimation problem in IRS-assisted communication systems. Besides, in contrast to the MMSE channel estimation [42, 43, 44], our proposed method adopts the LS-based channel estimation value as the initial coarse estimated value, models the channel estimation as a denoising problem, and exploits the neural network to design a denoiser. Thus, our proposed method is model-free and does not require any prior statistical information. The main contributions of this work are listed as follows111 Note that this paper mainly focuses on improving the estimation accuracy of CSI, but not on reducing the training overhead. As such, the proposed method in this paper adopts the LS-based channel estimation value as the network input and exploits a neural network to further improve the estimation accuracy. In this case, the training overhead of our proposed method comes from the LS method which can be addressed by adopting the low overhead grouping scheme for IRS elements [23, 24].:

  1. (1)

    In contrast to existing channel estimation methods, e.g., [19, 21, 22, 23, 24, 25, 26, 27, 28, 42, 43, 44], we model the channel estimation problem in IRS-MUC systems as a denoising problem and develop a DReL-based channel estimation framework which adopts a DRN to implicitly learn the residual noise for recovering the channel coefficients from the noisy pilot-based observations. Specifically, according to the MMSE criterion, a DRN-based MMSE (DRN-MMSE) estimator is derived in terms of Bayesian philosophy which enables the design of an efficient estimator.

  2. (2)

    To realize the developed framework, we adopt a convolutional neural network (CNN) to facilitate the DReL and propose a CNN-based DRN (CDRN) for channel estimation, in which a CNN-based denoising block with an element-wise subtraction structure is specifically designed to exploit both the spatial features of the noisy channel matrices and the additive nature of the noise simultaneously. Inheriting from the superiorities of CNN and DReL in feature extraction and denoising, the proposed CDRN method could further improve the estimation accuracy.

  3. (3)

    Although it is generally intractable to analyze the performance of a neural network, we formulate the proposed CDRN as a mathematical function and derive an explicit expression of the proposed CDRN estimator to explain the mechanism of the proposed algorithm theoretically, which proves that the proposed CDRN estimator can achieve the same performance as that of the LMMSE estimator.

  4. (4)

    Extensive simulations have been conducted to verify the efficiency of the proposed method in terms of the impacts of SNR and channel size on normalized MSE (NMSE), respectively. Additionally, visualizations of the proposed CDRN are also provided to illustrate the denoising process. Our results show that the performance of the proposed method approaches that of the optimal MMSE estimator requiring the computation of a prior probability density function (PDF) of cascaded channel.

The reminder of our paper is organized as follows. Section \@slowromancapii@ introduces the system model of the IRS-MUC system and derives the optimal MMSE estimator, the LMMSE estimator, and the LS estimator as benchmarks. In Section \@slowromancapiii@, we model the channel estimation as a denoising problem and develop a versatile DReL-based channel estimation framework. As a realization of the developed framework, a CDRN architecture and a CDRN-based channel estimation algorithm are designed in Section \@slowromancapiv@. Extensive simulation results are provided in Section \@slowromancapv@ to verify the effectiveness of the proposed scheme, and finally Section \@slowromancapvi@ concludes the work of this paper.

The notations used in our paper are listed as follows. The superscripts TT and HH are used to represent the transpose and conjugate transpose, respectively. Terms \mathbb{R} and \mathbb{C} denote the sets of real numbers and complex numbers, respectively. Term \mathbb{Z} denotes the set of integers. 𝒞𝒩(𝝁,𝚺){\mathcal{CN}}(\bm{\mu},\mathbf{\Sigma}) is the circularly symmetric complex Gaussian (CSCG) distribution where 𝝁\bm{\mu} and 𝚺\mathbf{\Sigma} are the mean vector and the covariance matrix, respectively. The matrix 𝐈p{\mathbf{I}}_{p} represents the pp-by-pp identity matrix and the vector 𝟎{\mathbf{0}} represents a zero vector. ()1(\cdot)^{-1} denotes the operation of the matrix inverse. Re{}\mathrm{Re}\{\cdot\} and Im{}\mathrm{Im}\{\cdot\} are used to represent the operations of extracting the real part and the imaginary part of a complex-valued matrix, respectively. tr[]\mathrm{tr}[\cdot] is the trace of a matrix and diag()\mathrm{diag}(\cdot) represents the construction of a diagonal matrix. F\|\cdot\|_{F} are used to denote the Frobenius norm of a matrix. In addition, exp()\exp(\cdot) indicates the exponential function, E()E(\cdot) represents the statistical expectation operation, and 𝟙Ω()\mathbbm{1}_{\Omega}(\cdot) denotes the indicator function of an event Ω\Omega.

Refer to caption
Figure 1: The uplink of the considered IRS-MUC system.

II System Model

In this paper, we consider an IRS-MUC system adopting the time division duplex (TDD) protocol, which consists of one base station (BS), one IRS, and KK users, as shown in Fig. 1. The BS is equipped with an MM-element antenna array and NN passive reflecting elements are installed at the IRS to assist the BS to serve the KK single-antenna users. By joint beamforming design at both the BS and the IRS, the throughput of the IRS-MUC system could be further enhanced. However, the downlink beamforming design critically depends on the availability of accurate CSI. Due to the property of channel reciprocity in TDD systems, the downlink CSI could be acquired from the uplink channel estimation in the IRS-MUC system. The uplink system model is illustrated in Fig. 1, where UkU_{k}, k{1,2,,K}k\in\{1,2,\cdots,K\}, denotes the kk-th user. The channels of the UkU_{k}-BS link, the UkU_{k}-IRS link, and the IRS-BS link are represented by 𝐝kM×1\mathbf{d}_{k}\in\mathbb{C}^{M\times 1}, 𝐟kN×1\mathbf{f}_{k}\in\mathbb{C}^{N\times 1}, and 𝐆M×N\mathbf{G}\in\mathbb{C}^{M\times N}, respectively. The reflecting link UkU_{k}-IRS-BS can be regarded as a dyadic backscatter channel [2, 7], where each element at IRS combines all the arriving signals and re-scatters them to the BS behaving as a single point source. Denote 𝐑=diag(𝐫)N×N\mathbf{R}=\mathrm{diag}(\mathbf{r})\in\mathbb{C}^{N\times N} with 𝐫=[βejφ1,βejφ2,,βejφN]T\mathbf{r}=[\beta e^{j\varphi_{1}},\beta e^{j\varphi_{2}},\cdots,\beta e^{j\varphi_{N}}]^{T} as the phase-shift matrix, where 0β10\leq\beta\leq 1 and 0φn2π0\leq\varphi_{n}\leq 2\pi are the amplitude and the phase shift at the nn-th, n{1,2,,N}n\in\{1,2,\cdots,N\}, element of IRS, respectively. The channel response from UkU_{k} to the BS via the IRS can be expressed as 𝐆𝐑𝐟kM×1\mathbf{G}\mathbf{R}\mathbf{f}_{k}\in\mathbb{C}^{M\times 1}. In this case, the channel response of the reflecting link UkU_{k}-IRS-BS can be altered through adjusting the phase-shift matrix 𝐑\mathbf{R} at the IRS. Since it is obvious that diag(𝐫)𝐟k=diag(𝐟k)𝐫\mathrm{diag}(\mathbf{r})\mathbf{f}_{k}=\mathrm{diag}(\mathbf{f}_{k})\mathbf{r}, the channel of reflecting link UkU_{k}-IRS-BS can be expressed as 𝐆diag(𝐫)𝐟k=𝐆diag(𝐟k)𝐫\mathbf{G}\mathrm{diag}(\mathbf{r})\mathbf{f}_{k}=\mathbf{G}\mathrm{diag}(\mathbf{f}_{k})\mathbf{r}. Therefore, the objective of the uplink channel estimation is to estimate

𝐇k=[𝐝k,𝐁k]M×(N+1),k,\mathbf{H}_{k}=[\mathbf{d}_{k},\mathbf{B}_{k}]\in\mathbb{C}^{M\times(N+1)},\forall k, (1)

where

𝐁k=𝐆diag(𝐟k)M×N,k,\mathbf{B}_{k}=\mathbf{G}\mathrm{diag}(\mathbf{f}_{k})\in\mathbb{C}^{M\times N},\forall k, (2)

is a cascaded channel. Therefore, the downlink channels can be obtained by the conjugate transpose of the uplink channels via exploiting the channel reciprocity in TDD systems.

Refer to caption
Figure 2: Channel estimation protocol for the considered IRS-MUC system.

Based on these, we then develop a channel estimation protocol for the considered IRS-MUC system. As shown in Fig. 2, each frame in the data stream has an identical frame structure which consists of a channel estimation phase and a data transmission phase. In this paper, we mainly focus on the channel estimation phase to estimate 𝐇k\mathbf{H}_{k} as defined in (1). In the channel estimation phase, the IRS generates CC, CN+1C\geq N+1, different reflection patterns through CC different phase-shift matrices, denoted by

𝚲=[𝐑1,𝐑2,,𝐑C]T.\mathbf{\Lambda}=[\mathbf{R}_{1},\mathbf{R}_{2},\cdots,\mathbf{R}_{C}]^{T}. (3)

Here, 𝐑c=diag(𝐫c)\mathbf{R}_{c}=\mathrm{diag}(\mathbf{r}_{c}) is the cc-th, c{1,2,,C}c\in\{1,2,\cdots,C\}, phase-shift matrix with 𝐫c=[βcejφc,1,βcejφc,2,\mathbf{r}_{c}=[\beta_{c}e^{j\varphi_{c,1}},\beta_{c}e^{j\varphi_{c,2}}, ,βcejφc,N]\cdots,\beta_{c}e^{j\varphi_{c,N}}], where βc[0,1]\beta_{c}\in[0,1] and 0φc,n2π0\leq\varphi_{c,n}\leq 2\pi are the amplitude and the phase shift at the nn-th element of the IRS under the cc-th phase-shift matrix, respectively. In addition, KK orthogonal pilot sequences are adopted to distinguish different users, i.e., for UkU_{k}, k\forall k, a pilot sequence with a length of LL, LKL\geq K, is adopted for each IRS phase-shift matrix, denoted by 𝐮k=[uk,1,uk,2,,uk,L]T\mathbf{u}_{k}=[u_{k,1},u_{k,2},\cdots,u_{k,L}]^{T} with 𝐮kH𝐮k=𝒫L\mathbf{u}_{k}^{H}\mathbf{u}_{k}=\mathcal{P}L and 𝐮aH𝐮b=0\mathbf{u}_{a}^{H}\mathbf{u}_{b}=0, where 𝒫\mathcal{P} is the power of each user, a,b{1,2,,K}a,b\in\{1,2,\cdots,K\}, and aba\neq b. According to Fig. 2, the channel estimation phase consists of CC sub-frames and each sub-frame consists of LL pilot symbols. For the IRS, the phase-shift matrix keeps unchanged within one sub-frame and it switches to different phase-shift matrices for different sub-frames. As for the kk-th user, UkU_{k}, k\forall k, it sends its identity pilot sequence 𝐮k\mathbf{u}_{k} in each sub-frame. Therefore, the ll-th, l{1,2,,L}l\in\{1,2,\cdots,L\}, received pilot signal vector at the BS in the cc-th sub-frame can be expressed as

𝐬c,l\displaystyle\mathbf{s}_{c,l} =k=1K(𝐝k+𝐆diag(𝐫c)𝐟k)uk,l+𝐯c,l\displaystyle=\sum\limits_{k=1}^{K}(\mathbf{d}_{k}+\mathbf{G}\mathrm{diag}(\mathbf{r}_{c})\mathbf{f}_{k}){u}_{k,l}+\mathbf{v}_{c,l} (4)
=k=1K𝐇k𝐩cuk,l+𝐯c,l.\displaystyle=\sum\limits_{k=1}^{K}\mathbf{H}_{k}\mathbf{p}_{c}{u}_{k,l}+\mathbf{v}_{c,l}. (5)

Here, 𝐇k\mathbf{H}_{k} is the channel matrix needed to be estimated as defined in (1) and 𝐩c=[1,𝐫c]T(N+1)×1\mathbf{p}_{c}=[1,\mathbf{r}_{c}]^{T}\in\mathbb{C}^{(N+1)\times 1}. Equation (5) is due to the property of diag(𝐫c)𝐟k=diag(𝐟k)𝐫c\mathrm{diag}(\mathbf{r}_{c})\mathbf{f}_{k}=\mathrm{diag}(\mathbf{f}_{k})\mathbf{r}_{c}. In addition, 𝐯c,lM×1\mathbf{v}_{c,l}\in\mathbb{C}^{M\times 1} is the ll-th sampling noise vector at the BS in the cc-th sub-frame. Generally, 𝐯c,l\mathbf{v}_{c,l} is assumed to be a CSCG random vector, i.e., 𝐯c,l𝒞𝒩(𝟎,σv2𝐈M)\mathbf{v}_{c,l}\sim\mathcal{CN}(\mathbf{0},\sigma_{v}^{2}\mathbf{I}_{M}), where σv2\sigma_{v}^{2} denotes the noise variance of each antenna at the BS. Stacking the LL received pilot signal vectors at the BS during the cc-th sub-frame into a matrix form:

𝐒c=k=1K𝐇k𝐩c𝐮kH+𝐕c,\mathbf{S}_{c}=\sum\limits_{k=1}^{K}\mathbf{H}_{k}\mathbf{p}_{c}\mathbf{u}_{k}^{H}+\mathbf{V}_{c}, (6)

where 𝐒c=[𝐬c,1,𝐬c,2,,𝐬c,L]\mathbf{S}_{c}=[\mathbf{s}_{c,1},\mathbf{s}_{c,2},\cdots,\mathbf{s}_{c,L}] and 𝐕c=[𝐯c,1,𝐯c,2,,𝐯c,L]\mathbf{V}_{c}=[\mathbf{v}_{c,1},\mathbf{v}_{c,2},\cdots,\mathbf{v}_{c,L}]. Since the pilot sequences of each two users are orthogonal, the received signal from UkU_{k} can be separated by multiplying a sequence 𝐮k\mathbf{u}_{k} with 𝐒c\mathbf{S}_{c} in (6), i.e.,

𝐱c,k=𝐇k𝐩c+𝐳c,k.{\mathbf{x}}_{c,k}=\mathbf{H}_{k}\mathbf{p}_{c}+{\mathbf{z}}_{c,k}. (7)

Here, 𝐱c,k=1𝒫L𝐒c𝐮kM×1{\mathbf{x}}_{c,k}=\frac{1}{\mathcal{P}L}\mathbf{S}_{c}\mathbf{u}_{k}\in\mathbb{C}^{M\times 1} is the received signal vector at the BS from UkU_{k} within the cc-th sub-frame. 𝐳c,k=1𝒫L𝐕c𝐮kM×1{\mathbf{z}}_{c,k}=\frac{1}{\mathcal{P}L}\mathbf{V}_{c}\mathbf{u}_{k}\in\mathbb{C}^{M\times 1} with 𝐳c,k𝒞𝒩(𝟎,σz2𝐈M){\mathbf{z}}_{c,k}\sim\mathcal{CN}(\mathbf{0},\sigma_{z}^{2}\mathbf{I}_{M}), where σz2\sigma_{z}^{2} is the variance of each element of 𝐳c,k{\mathbf{z}}_{c,k}. Based on this and after CC subframes, we can then obtain the matrix form of (7): 𝐗k=[𝐱1,k,𝐱2,k,,𝐱C,k]\mathbf{X}_{k}=[\mathbf{x}_{1,k},\mathbf{x}_{2,k},\cdots,\mathbf{x}_{C,k}] which can be expressed as

𝐗k=𝐇k𝐏+𝐙k,\mathbf{X}_{k}=\mathbf{H}_{k}\mathbf{P}+\mathbf{Z}_{k}, (8)

where 𝐏=[𝐩1,𝐩2,,𝐩C](N+1)×C\mathbf{P}=[\mathbf{p}_{1},\mathbf{p}_{2},\cdots,\mathbf{p}_{C}]\in\mathbb{C}^{(N+1)\times C} with 𝐩c=[1,𝐫c]T\mathbf{p}_{c}=[1,\mathbf{r}_{c}]^{T} as defined in (5) and 𝐙k=[𝐳1,k,𝐳2,k,,𝐳C,k]M×C\mathbf{Z}_{k}=[\mathbf{z}_{1,k},\mathbf{z}_{2,k},\cdots,\mathbf{z}_{C,k}]\in\mathbb{C}^{M\times C}. According to [21], an optimal scheme of 𝐏\mathbf{P} in terms of improving received signal power at the BS is to design 𝐏\mathbf{P} as a discrete Fourier transform (DFT), i.e.,

𝐏=[1111WCWCC11WCNWCN(C1)](N+1)×C,\mathbf{P}=\left[\begin{matrix}1&1&\cdots&1\\ 1&W_{C}&\cdots&W_{C}^{C-1}\\ \vdots&\vdots&\ddots&\vdots\\ 1&W_{C}^{N}&\cdots&W_{C}^{N(C-1)}\\ \end{matrix}\right]\in\mathbb{C}^{(N+1)\times C}, (9)

where WC=ej2π/CW_{C}=e^{j2\pi/C} and 𝐏𝐏H=C𝐈N+1\mathbf{P}\mathbf{P}^{H}=C\mathbf{I}_{N+1}. The CC different phase-shift matrices 𝚲=[𝐑1,𝐑2,\mathbf{\Lambda}=[\mathbf{R}_{1},\mathbf{R}_{2}, ,𝐑C]T\cdots,\mathbf{R}_{C}]^{T} defined in (3) are set based on 𝐏\mathbf{P} in (9).

Therefore, the task of channel estimation in IRS-MUC is to recover 𝐇k\mathbf{H}_{k}, k\forall k, by exploiting 𝐗k\mathbf{X}_{k} and 𝐏\mathbf{P}. Based on the models in (8) and (9), we then first introduce three commonly adopted methods as three benchmarks, namely the LS method, the MMSE method, and the LMMSE method.

(a) LS Channel Estimator:

According to [45], the estimated 𝐇k\mathbf{H}_{k} using LS estimator is given by

𝐇~kLS=𝐗k𝐏,\tilde{\mathbf{H}}_{k}^{\mathrm{LS}}=\mathbf{X}_{k}\mathbf{P}^{\dagger}, (10)

where 𝐏=𝐏H(𝐏𝐏H)1\mathbf{P}^{\dagger}=\mathbf{P}^{H}(\mathbf{P}\mathbf{P}^{H})^{-1} denotes the pseudoinverse of 𝐏\mathbf{P}.

(b) MMSE Channel Estimator:

Different from the LS method where 𝐇k\mathbf{H}_{k} is assumed to be an unknown but deterministic constant, the Bayesian approach assumes that 𝐇k\mathbf{H}_{k} is a random variable with a prior PDF p(𝐇k)p(\mathbf{H}_{k}). Therefore, the Bayesian approach can take advantage of prior knowledge to further improve the estimation accuracy. According to [46], the optimal Bayesian estimator is the MMSE estimator which can be expressed as

𝐇~kMMSE=E[𝐇k|𝐗k]=𝐇kp(𝐇k|𝐗k)𝑑𝐇k,\tilde{\mathbf{H}}_{k}^{\mathrm{MMSE}}=E[{\mathbf{H}_{k}}|{\mathbf{X}_{k}}]=\int{{\mathbf{H}_{k}}p({\mathbf{H}_{k}}|{\mathbf{X}_{k}})d{\mathbf{H}_{k}}}, (11)

where p(𝐇k|𝐗k)=p(𝐗k|𝐇k)p(𝐇k)p(𝐗k|𝐇k)p(𝐇k)𝑑𝐇kp({\mathbf{H}_{k}}|{\mathbf{X}_{k}})=\frac{{p({\mathbf{X}_{k}}|{\mathbf{H}_{k}})p({\mathbf{H}_{k}})}}{{\int{p({\mathbf{X}_{k}}|{\mathbf{H}_{k}})p({\mathbf{H}_{k}})d{\mathbf{H}_{k}}}}} is the posterior PDF and p(𝐗k|𝐇k)p({\mathbf{X}_{k}}|{\mathbf{H}_{k}}) is the conditional PDF.

(c) LMMSE Channel Estimator:

Note that the optimal MMSE estimator in (11) is computationally intensive to be implemented in practice when the channel distribution is complicated. Thus, the LMMSE estimator can be selected as a more practical approach with a simple explicit expression222Note that the proposed joint optimization approach in [47] is an effective scheme for realizing the LMMSE estimation in IRS-assisted systems. However, the system model in [47] is a downlink channel estimation model requiring the design of a new training sequence while our work focuses on the uplink channel estimation model for a given training sequence. Thus, the developed approach in [47] cannot be directly adopted in our work and the extension of our work to the downlink scenario is left for our future work due to the page limitation. [46]:

𝐇~kLMMSE=𝐗k(𝐏H𝐑𝐇k𝐏+Mσz2𝐈C)1𝐏H𝐑𝐇k,\tilde{\mathbf{H}}_{k}^{\mathrm{LMMSE}}=\mathbf{X}_{k}(\mathbf{P}^{H}\mathbf{R}_{\mathbf{H}_{k}}\mathbf{P}+M\sigma_{z}^{2}\mathbf{I}_{C})^{-1}\mathbf{P}^{H}\mathbf{R}_{\mathbf{H}_{k}}, (12)

where 𝐑𝐇k=E(𝐇kH𝐇k)\mathbf{R}_{\mathbf{H}_{k}}=E(\mathbf{H}_{k}^{H}\mathbf{H}_{k}) denotes the statistical channel correlation matrix [45].

Note that the LS estimator requires no prior knowledge of channel and is widely used in practice. If the channel follows the Rayleigh fading model and the statistical channel correlation matrix is available, the optimal MMSE estimator is equivalent to the LMMSE estimator [46] and it can be adopted to capture the statistical characteristics of the channel to further improve the estimation performance. However, for IRS-MUC systems, 𝐇k\mathbf{H}_{k} includes a cascaded channel which generally does not follow the Rayleigh fading model. In this case, the optimal MMSE estimator involves the calculation of a multidimensional integration and thus it is intractable to be implemented in practice due to its intensive computational cost. Meanwhile, the LMMSE estimator still has a performance gap compared with the optimal MMSE estimator [46].

To further improve the estimation performance, in the next section, we still retain the MSE criterion but adopt a data driven DL approach to design a practical channel estimation method, i.e., a DReL-based channel estimation scheme.

III Deep Residual Learning-based Channel Estimation Framework

Note that the additive noise in the system model is a key obstruction for recovering the channel coefficients. In this section, we first model the channel estimation as a denoising problem and then develop a DReL-based framework to learn the residual noise from the noisy observations for paving the way in recovering the channel coefficients. In the developed DReL framework, a DRN with a subtraction architecture is designed to exploit both the features of received observations and the additive nature of the noise to further improve the estimation performance. In the following, we will introduce the derived denoising model and the developed DReL-based channel estimation framework, respectively.

III-A Denoising Model for Channel Estimation

Since a salient feature of the LS estimator is that it does not require any statistical characterization of the data, it is widely adopted in practice due to its convenience of implementation. According to (10), the estimation error of the LS estimator is [45]

εLS\displaystyle\varepsilon_{\mathrm{LS}} =E(𝐇k𝐇~kLSF2)\displaystyle=E(\|\mathbf{H}_{k}-\tilde{\mathbf{H}}_{k}^{\mathrm{LS}}\|_{F}^{2}) (13)
=Mσz2tr[(𝐏𝐏H)1]\displaystyle=M\sigma_{z}^{2}\mathrm{tr}[(\mathbf{P}\mathbf{P}^{H})^{-1}]
=Mσz2(N+1)C.\displaystyle=\frac{M\sigma_{z}^{2}}{(N+1)C}.

It can be observed that εLS\varepsilon_{\mathrm{LS}} is essentially a function of the noise power and its value decreases with the increase of CC.

Therefore, if we exploit the LS-based channel estimation value as the initial coarse estimated value, then the channel estimation in the IRS-MUC system can be regarded as a denoising problem: recovering 𝐇k\mathbf{H}_{k} from a noisy observation

𝐗~k=𝐇k+𝐙~k,\tilde{\mathbf{X}}_{k}={\mathbf{H}}_{k}+\tilde{\mathbf{Z}}_{k}, (14)

where 𝐗~k=𝐗k𝐏M×(N+1)\tilde{\mathbf{X}}_{k}=\mathbf{X}_{k}\mathbf{P}^{\dagger}\in\mathbb{C}^{M\times(N+1)} is the noisy observation based on the LS estimator and 𝐙~k=𝐙k𝐏M×(N+1)\tilde{\mathbf{Z}}_{k}=\mathbf{Z}_{k}\mathbf{P}^{\dagger}\in\mathbb{C}^{M\times(N+1)} denotes the noise component. According to (14), 𝐇k\mathbf{H}_{k} includes a cascaded channel 𝐁k\mathbf{B}_{k} of which the prior PDF is not a Gaussian PDF, thus, the data model in (14) is not a Bayesian general linear model [46]. In this case, it is intractable to derive a Bayesian estimator with an explicit expression as the one adopted in the model driven approach. In contrast, we will adopt a data-driven approach to develop a DReL-based channel estimation framework for IRS-MUC systems in the following section.

Refer to caption
Figure 3: The developed DReL based channel estimation framework for IRS-MUC systems.

III-B The Developed DReL-based Channel Estimation Framework

Based on the denoising model, we develop a DReL-based channel estimation framework which consists of an offline training phase and an online estimation phase, as shown in Fig. 3. In the offline training phase, a DRN training process is operated to obtain a well-trained DRN, i.e., the DRN-based channel estimator. For the online estimation phase, the test data is sent to the well-trained DRN to obtain the estimated channel coefficients directly. In the following, we will introduce the phases of offline training and online estimation, respectively.

1) Offline Training:

Denote by

(𝒳,)={(𝐗k(1),𝐇k(1)),(𝐗k(2),𝐇k(2)),,(𝐗k(Nt),𝐇k(Nt))}(\mathcal{X},\mathcal{H})=\{(\mathbf{X}_{k}^{(1)},\mathbf{H}_{k}^{(1)}),(\mathbf{X}_{k}^{(2)},\mathbf{H}_{k}^{(2)}),\cdots,(\mathbf{X}_{k}^{(N_{t})},\mathbf{H}_{k}^{(N_{t})})\} (15)

the raw training set, where 𝐗k(i)\mathbf{X}_{k}^{(i)} and 𝐇k(i)\mathbf{H}_{k}^{(i)} are the input and the ground truth of the ii-th, i{1,2,,Nt}i\in\{1,2,\cdots,N_{t}\}, training example, respectively. To obtain the coarse estimated channel coefficients as defined in the denoising problem of (14), we then send (𝒳,)(\mathcal{X},\mathcal{H}) to the LS estimator in (10) and obtain the LS-based training set:

(𝒳~,)={(𝐗~k(1),𝐇k(1)),(𝐗~k(2),𝐇k(2)),,(𝐗~k(Nt),𝐇k(Nt))},(\tilde{\mathcal{X}},\mathcal{H})=\{(\tilde{\mathbf{X}}_{k}^{(1)},\mathbf{H}_{k}^{(1)}),(\tilde{\mathbf{X}}_{k}^{(2)},\mathbf{H}_{k}^{(2)}),\cdots,(\tilde{\mathbf{X}}_{k}^{(N_{t})},\mathbf{H}_{k}^{(N_{t})})\}, (16)

where (𝐗~k(i),𝐇k(i))(\tilde{\mathbf{X}}_{k}^{(i)},\mathbf{H}_{k}^{(i)}) is the ii-th training example of (𝒳~,)(\tilde{\mathcal{X}},\mathcal{H}). According to Fig. 3, (𝒳~,)(\tilde{\mathcal{X}},\mathcal{H}) is then sent to DRN for training. The DRN consists of a DNN and an element-wise subtraction operator between the DNN input and the DNN output. Note that the DNN can be any kind of neural networks, e.g., multi-layer perception [48], CNN [49], etc. Denote by f~θ()\tilde{f}_{\theta}(\cdot) the expression of the DNN where θ\theta denotes the total parameters of DNN, the output of the DRN can be expressed as

𝐎=fθ(𝐀)=𝐀f~θ(𝐀),\mathbf{O}=f_{\theta}(\mathbf{A})=\mathbf{A}-\tilde{f}_{\theta}(\mathbf{A}), (17)

where 𝐀\mathbf{A} and 𝐎\mathbf{O} are the input and output of DRN, and fθ()f_{\theta}(\cdot) denotes the expression of DRN.

In the following, we select the cost function for DRN training. According to the MMSE criterion, the optimal MMSE estimator is derived by minimizing the Bayesian MSE, i.e.,

Bmse=E[𝐇~k𝐇kF2],B_{\mathrm{mse}}=E[\|\tilde{\mathbf{H}}_{k}-\mathbf{H}_{k}\|_{F}^{2}], (18)

where 𝐇~k\tilde{\mathbf{H}}_{k} and 𝐇k\mathbf{H}_{k} denote the estimated channel matrix and the ground truth of the channel matrix. To achieve the optimal performance, an intuitive selection of the cost function is the Bayesian MSE expression. However, the number of training examples is finite in practice and thus the statistical Bayesian MSE is not available and only the empirical MSE can be adopted which is defined as

Emse(Nt)=12Nti=1Nt𝐇~k(i)𝐇k(i)F2,E_{\mathrm{mse}}(N_{t})=\frac{1}{2N_{t}}\sum_{i=1}^{N_{t}}\left\|\tilde{\mathbf{H}}_{k}^{(i)}-\mathbf{H}_{k}^{(i)}\right\|_{F}^{2}, (19)

where 𝐇~k(i)=fθ(𝐗~k(i))\tilde{\mathbf{H}}_{k}^{(i)}=f_{\theta}(\tilde{\mathbf{X}}_{k}^{(i)}) is the ii-th estimated channel matrix based on 𝐗~k(i)\tilde{\mathbf{X}}_{k}^{(i)} by the DRN. In this case, we can only select the empirical MSE as the cost function which can be expressed as

JMSE(θ)=12Nti=1Ntfθ(𝐗~k(i))𝐇k(i)F2.J_{\mathrm{MSE}}(\theta)=\frac{1}{2N_{t}}\sum_{i=1}^{N_{t}}\left\|f_{\theta}(\tilde{\mathbf{X}}_{k}^{(i)})-\mathbf{H}_{k}^{(i)}\right\|_{F}^{2}. (20)

Based on this, the DRN can adopt the backpropagation (BP) algorithm to obtain a well-trained DRN by progressively updating the network parameters. Therefore, the well-trained DRN, i.e., the DRN-based MMSE (DRN-MMSE) estimator can be expressed as

𝐇~kDRN=fθ(𝐗~),\tilde{\mathbf{H}}_{k}^{\mathrm{DRN}}=f_{\theta^{*}}(\tilde{\mathbf{X}}), (21)

where 𝐗~\tilde{\mathbf{X}} represents an arbitrary LS-based input and θ\theta^{*} denotes the well-trained network parameters by minimizing JMSE(θ)J_{\mathrm{MSE}}(\theta).

Refer to caption
Figure 4: A flowchart illustration of the derivation of the DRN-based MMSE estimator. The upper half represents the proposed data driven DReL approach while the lower half is the model driven approach.

For further understanding, we then analyze the performance of the DRN-based estimator in terms of Bayesian philosophy. As shown in Fig. 4, there are two approaches to design the MMSE estimator: model driven approach and data driven approach. Traditionally, given a prior PDF, the MMSE estimator can be derived by minimizing the Bayesian MSE in a model driven manner. However, for the cascaded channel, it is intractable to implement the MMSE estimator due to a lack of closed-form expression and the associated high-cost in computation. Alternatively, the proposed DRN-MMSE estimator is obtained by minimizing the empirical MSE based on the training set in a data driven approach. Note that it is obvious that

limNtEmse(Nt)=limNt12Nti=1Nt𝐇~k(i)𝐇k(i)F2=Bmse.\lim_{N_{t}\to\infty}E_{\mathrm{mse}}(N_{t})=\lim_{N_{t}\to\infty}\frac{1}{2N_{t}}\sum_{i=1}^{N_{t}}\left\|\tilde{\mathbf{H}}_{k}^{(i)}-\mathbf{H}_{k}^{(i)}\right\|_{F}^{2}=B_{\mathrm{mse}}. (22)

Hence, the DRN estimator converges to the MMSE estimator when NtN_{t}\rightarrow\infty. As a summary, the DRN estimator is a data driven approach by minimizing the empirical MSE based on a training set while the optimal MMSE is a model driven approach by minimizing Bayesian MSE based on a prior PDF. In particular, the performance of the DRN estimator converges to the optimal MMSE estimator when the training set is sufficiently large. Next, we will introduce the online estimation phase of the developed framework.

2) Online Estimation:

Given the test data 𝐗˙k\dot{\mathbf{X}}_{k}, we first adopt the LS estimator to obtain the initial coarse channel estimation 𝐗˙k𝐏\dot{\mathbf{X}}_{k}{\mathbf{P}}^{\dagger} and then send it to the well-trained DRN to operate the channel estimation, as shown in Fig. 3. Finally, the estimated channel coefficients by the proposed DRN estimator can be expressed as

𝐇~kDRN=fθ(𝐗˙𝐏)=𝐗˙𝐏f~θ(𝐗˙𝐏).\tilde{\mathbf{H}}_{k}^{\mathrm{DRN}}=f_{\theta^{*}}({\dot{\mathbf{X}}}{\mathbf{P}}^{\dagger})={\dot{\mathbf{X}}}{\mathbf{P}}^{\dagger}-\tilde{f}_{\theta^{*}}({\dot{\mathbf{X}}}{\mathbf{P}}^{\dagger}). (23)

Remark 1: Note that the DRN in the developed DReL framework can be realized by any kind of DNN architecture, e.g., the dense neural network [48] and the CNN [31]. Therefore, the developed DReL based channel estimation framework is a universal channel estimation framework which can strike a balance between system performance and computational complexity via the selection of DRN. Specifically, the DRN makes the best use of the DNN with an element-subtraction operator to exploit both the channel features and additive nature of noise to further improve the channel estimation performance for IRS-MUC systems.

IV CNN-based Deep Residual Network for Channel Estimation

In this section, we introduce a practical approach for implementing the developed DReL framework via a CNN. Note that CNN has powerful capability in extracting features from the noisy observation matrices, meanwhile the subtraction structure of DReL contributes to exploiting the additive nature of the noise [39, 40, 41]. Therefore, we adopt a CNN to facilitate the application of DReL, resulting a CDRN for channel estimation. In the following, we will introduce the CDRN architecture and the CDRN-based channel estimation algorithm, respectively.

IV-A CDRN Architecture

As shown in Fig. 5, the CDRN consists of one input layer, DD denoising blocks, and one output layer. The denoising block utilizes the CNN with a subtraction structure to learn the residual noise from the noisy channel matrix for denoising. In addition, considering the channel matrix is a complex-valued matrix, we divide the input into two different parts: the real part and the imaginary part for the convenience of extracting features333Since the real part and the imaginary part of the input channel matrix are generally orthogonal [50], we divide the neural network input into two independent parts and adopt the real-valued convolution to exploit features from the real part and the imaginary part of channel matrix, respectively.. The hyperparameters of CDRN are summarized in Table I and the details of each layer of CDRN are introduced as follows.

Refer to caption
Figure 5: The network architecture of the proposed CDRN for channel estimation in the IRS-MUC system.
TABLE I: Hyperparameters of the proposed CDRN
Input: Noisy channel matrix with the size of M×(N+1)×2M\times(N+1)\times 2
Denoising Block:
       Layers Operations        Filter Size
       1 Conv + BN + ReLU        64×(3×3×2)64\times(3\times 3\times 2)
       2Nl12\sim N_{l}-1 Conv + BN + ReLU        64×(3×3×64)64\times(3\times 3\times 64)
       NlN_{l} Conv        2×(3×3×64)2\times(3\times 3\times 64)
Output: Denoised channel matrix with the size of M×(N+1)×2M\times(N+1)\times 2

a) Input Layer:

To extract features from the complex-valued noisy channel matrix, it is necessary to adopt two neural network channels for the real part and imaginary part of the noisy channel matrix, respectively. Thus, the input layer can be expressed as

𝐀=([Re{𝐗~k(i)},Im{𝐗~k(i)}]),\mathbf{A}=\mathcal{F}([\mathrm{Re}\{\tilde{\mathbf{X}}_{k}^{(i)}\},\mathrm{Im}\{\tilde{\mathbf{X}}_{k}^{(i)}\}]), (24)

where 𝐀M×(N+1)×2\mathbf{A}\in\mathbb{R}^{M\times(N+1)\times 2} is the input of CDRN, ():M×(2N+2)M×(N+1)×2\mathcal{F}(\cdot):\mathbb{R}^{M\times(2N+2)}\mapsto\mathbb{R}^{M\times(N+1)\times 2} represents the mapping function, and 𝐗~k(i)\tilde{\mathbf{X}}_{k}^{(i)} is the LS-based input from an arbitrary training example.

b) Denoising Blocks:

In CDRN, we adopt DD denoising blocks to gradually enhance the denoising performance and all denoising blocks have an identical structure. Specifically, a denoising block consists of a residual subnetwork and an element-wise subtraction operator, as depicted in Fig. 5. The residual subnetwork of each denoising block has NlN_{l} layers. For the first Nl1N_{l}-1 layers, the “Conv+BN+ReLU” operations denoted by the green arrows are adopted for each of the first Nl1N_{l}-1 layers. The convolution (Conv) operation and the rectified linear unit (ReLU) are adopted together to exploit the spatial features of the channel matrix. To improve the stability of the network and to accelerate the training speed of CDRN, a batch normalization (BN) [51] is added between Conv and ReLU. For the last layer, a convolution operation denoted by the yellow arrows is finally adopted to obtain the residual noise matrix for the subsequent element-wise subtraction. Since the noise in the received signals is of additive nature, an element-wise subtraction is added between the input and the output of the residual subnetwork to obtain the denoised channel matrix.

Denote by θd()\mathcal{R}_{\theta_{d}}(\cdot) the function expression of the dd-th, d{1,2,,D}d\in\{1,2,\cdots,D\}, residual subnetwork where θd\theta_{d} is the network parameters. The dd-th denoising block can be expressed as

𝐀d=𝐀d1θd(𝐀d1),d,\mathbf{A}_{d}=\mathbf{A}_{d-1}-\mathcal{R}_{\theta_{d}}(\mathbf{A}_{d-1}),\forall d, (25)

where 𝐀0=𝐀\mathbf{A}_{0}=\mathbf{A}, 𝐀d1\mathbf{A}_{d-1} and 𝐀d\mathbf{A}_{d} are the input and the output of the dd-th denoising block, respectively.

c) Output Layer:

The output layer is the the denoised channel matrix based on DD denoising blocks, i.e., the output of the DD-th denoising block:

𝐀D=𝐀D1θD(𝐀D1).\mathbf{A}_{D}=\mathbf{A}_{D-1}-\mathcal{R}_{\theta_{D}}(\mathbf{A}_{D-1}). (26)

According to (25) and (26), the output of the CDRN can be finally expressed as

gθ(𝐀)=𝐀d=1Dθd(𝐀d1),g_{\theta}(\mathbf{A})=\mathbf{A}-\sum\limits_{d=1}^{D}\mathcal{R}_{\theta_{d}}\left(\mathbf{A}_{d-1}\right), (27)

where gθ()g_{\theta}(\cdot) denotes the function expression of CDRN with parameters θ=[θ1,θ2,,θD]\theta=[\theta_{1},\theta_{2},\cdots,\theta_{D}] and d=1Dθd(𝐀d1)\sum_{d=1}^{D}\mathcal{R}_{\theta_{d}}\left(\mathbf{A}_{d-1}\right) represents the residual noise component. Therefore, the denoised channel matrix is the result of the element-wise subtraction between the noisy channel matrix 𝐀\mathbf{A} and the residual noise component d=1Dθd(𝐀d1)\sum_{d=1}^{D}\mathcal{R}_{\theta_{d}}\left(\mathbf{A}_{d-1}\right).

Remark 2: Note that the neural network is of powerful scalability [48]. Although the channel size changes with the settings of the IRS-MUC systems, the proposed CDRN architecture is still scalable and can be easily extended to different shapes accordingly, as will be shown in the simulation results of Section \@slowromancapv@.

Refer to caption
Figure 6: The developed CDRN-based framework for channel estimation in the considered IRS-MUC system.

IV-B CDRN-based Channel Estimation Algorithm

Based on the designed CDRN, we apply it to the developed DReL framework and propose a CDRN-based channel estimation algorithm, as shown in Fig. 6. The proposed CDRL algorithm consists of an offline training and an online estimation.

a) Offline Training of CDRN Algorithm:

Based on the LS-based training set defined in (16), we can obtain the following training set:

(𝒴~,)={(𝐘~k(1),𝐇k(1)),(𝐘~k(2),𝐇k(2)),,(𝐘~k(Nt),𝐇k(Nt))},(\tilde{\mathcal{Y}},\mathcal{H})=\{(\tilde{\mathbf{Y}}_{k}^{(1)},\mathbf{H}_{k}^{(1)}),(\tilde{\mathbf{Y}}_{k}^{(2)},\mathbf{H}_{k}^{(2)}),\cdots,(\tilde{\mathbf{Y}}_{k}^{(N_{t})},\mathbf{H}_{k}^{(N_{t})})\}, (28)

where 𝐘~k(i)=([Re{𝐗~k(i)},Im{𝐗~k(i)}])M×(N+1)×2\tilde{\mathbf{Y}}_{k}^{(i)}=\mathcal{F}([\mathrm{Re}\{\tilde{\mathbf{X}}_{k}^{(i)}\},\mathrm{Im}\{\tilde{\mathbf{X}}_{k}^{(i)}\}])\in\mathbb{R}^{M\times(N+1)\times 2} and (𝐘~k(i),𝐇k(i))(\tilde{\mathbf{Y}}_{k}^{(i)},\mathbf{H}_{k}^{(i)}) represents the ii-th training example of (𝒴~,)(\tilde{\mathcal{Y}},\mathcal{H}). According to (20) and (27), the cost function of CDRN can be expressed as

JCDRN(θ)=12Nti=1Nthθ(𝐘~k(i))𝐇k(i)F2.J_{\mathrm{CDRN}}(\theta)=\frac{1}{2N_{t}}\sum_{i=1}^{N_{t}}\left\|h_{\theta}(\tilde{\mathbf{Y}}_{k}^{(i)})-\mathbf{H}_{k}^{(i)}\right\|_{F}^{2}. (29)

Here, hθ(𝐘~(i))=(gθ(𝐘~(i)))h_{\theta}(\tilde{\mathbf{Y}}^{(i)})=\mathcal{M}(g_{\theta}(\tilde{\mathbf{Y}}^{(i)})) where ():M×(N+1)×2M×(N+1)\mathcal{M}(\cdot)\!\!:\mathbb{R}^{M\times(N+1)\times 2}\mapsto\mathbb{C}^{M\times(N+1)} represents a function that builds the complex-valued matrix based on the real-valued matrices. As shown in Fig. 6, we can then use the BP algorithm to progressively update the network parameters to finally obtain the well-trained CDRN, i.e., the CDRN estimator:

𝐇~kCDRN=hθ(𝐘~)=(gθ(𝐘~)),\tilde{\mathbf{H}}_{k}^{\mathrm{CDRN}}=h_{\theta^{*}}(\tilde{\mathbf{Y}})=\mathcal{M}(g_{\theta^{*}}(\tilde{\mathbf{Y}})), (30)

where 𝐘~\tilde{\mathbf{Y}} denotes the input matrix and gθ()g_{\theta^{*}}(\cdot) is the function expression of the well-trained CDRN with the well-trained network parameters θ=[θ1,θ2,,θD]\theta^{*}=[\theta^{*}_{1},\theta^{*}_{2},\cdots,\theta^{*}_{D}] and hθ()h_{\theta^{*}}(\cdot) is the complex-valued form of gθ()g_{\theta^{*}}(\cdot).

b) Online Estimation of CDRN Algorithm:

Given a test data denoted by 𝐗˙k\dot{\mathbf{X}}_{k}, we first send it to the LS estimator to obtain a coarse channel estimation denoted by 𝐗˙k𝐏\dot{\mathbf{X}}_{k}\mathbf{P}^{\dagger}, and then send it to the CDRN estimator to obtain a refined channel estimate matrix:

𝐇~kCDRN=hθ(𝐘˙)=(𝐘˙d=1Dθd(𝐘˙d1)),\tilde{\mathbf{H}}_{k}^{\mathrm{CDRN}}=h_{\theta^{*}}(\dot{\mathbf{Y}})=\mathcal{M}\left(\dot{\mathbf{Y}}-\sum\limits_{d=1}^{D}\mathcal{R}_{\theta^{*}_{d}}\left(\dot{\mathbf{Y}}_{d-1}\right)\right), (31)

where 𝐘˙=([Re{𝐗˙k𝐏},Im{𝐗˙k𝐏}])\dot{\mathbf{Y}}=\mathcal{F}([\mathrm{Re}\{\dot{\mathbf{X}}_{k}\mathbf{P}^{\dagger}\},\mathrm{Im}\{\dot{\mathbf{X}}_{k}\mathbf{P}^{\dagger}\}]), 𝐘˙0=𝐘˙\dot{\mathbf{Y}}_{0}=\dot{\mathbf{Y}}, and 𝐘˙d1\dot{\mathbf{Y}}_{d-1} and 𝐘˙d\dot{\mathbf{Y}}_{d} are the input and output of the dd-th denoising block of CDRN as defined in (25).

c) CDRN-based Channel Estimation Algorithm Steps:

Based on the above analysis, we then summarize the proposed CDRN-based channel estimation algorithm in Algorithm 1, where we use iti_{t} and ItI_{t} to denote the iteration index and the maximum iteration number, respectively.

Algorithm 1 CDRN-based Channel Estimation Algorithm
Initialization: it=0i_{t}=0, real-valued training set (𝒴~,)(\tilde{\mathcal{Y}},\mathcal{H})
Offline Training:
1:        Input: Training set (𝒴~,)(\tilde{\mathcal{Y}},\mathcal{H})
2:           while itIti_{t}\leq I_{t} do
3:                Update θ\theta by BP algorithm to minimize JCDRN(θ)J_{\mathrm{CDRN}}(\theta)
                    it=it+1i_{t}=i_{t}+1
4:           end while
5:        Output: Well-trained hθ(){h}_{\theta^{*}}(\cdot)
Online Estimation:
6:      Input: Test data 𝐘˙=([Re{𝐗˙k𝐏},Im{𝐗˙k𝐏}])\dot{\mathbf{Y}}=\mathcal{F}([\mathrm{Re}\{\dot{\mathbf{X}}_{k}\mathbf{P}^{\dagger}\},\mathrm{Im}\{\dot{\mathbf{X}}_{k}\mathbf{P}^{\dagger}\}])
7:         do Channel Estimation using CDRN estimator in (30)
8:      Output: 𝐇~kCDRN=hθ(𝐘˙)\tilde{\mathbf{H}}_{k}^{\mathrm{CDRN}}=h_{\theta^{*}}(\dot{\mathbf{Y}}).

IV-C Theoretical Analysis

To offer more insight of the proposed estimation algorithm, we then analyze the proposed CDRN and characterize its properties theoretically. Although it is difficult to analyze the neural network which consists of a large number of parameters, we formulate the proposed CDRN as an explicit mathematical function and analyze the output of the CDRN theoretically.

As shown in Fig. 5, the CDRN involves DD identical denoising blocks and the dd-th denoising block can be expressed as the subtraction between the input 𝐀d1\mathbf{A}_{d-1} and the output of the residual subnetwork θd()\mathcal{R}_{\theta_{d}}(\cdot). For the convenience of analysis, a simplified real-valued channel matrix is considered as the network input and it can be easily extended to a complex-valued model via a similar approach as in [46]. In this case, the input matrix becomes two-dimension matrix and the function ()\mathcal{M}(\cdot) should be removed for (29)-(31). Note that θd()\mathcal{R}_{\theta_{d}}(\cdot) consists of NlN_{l} convolutional layers and both BN and ReLU operations have been shown to be able to improve the network stability and the training speed [48], [52]. For simplicity, we mainly focus on the effect of the convolution and ReLU operations on the network output. Since the convolution operation can be expressed as the production of two matrices [48], for each residual subnetwork, we have

θd(𝐀d1)\displaystyle\mathcal{R}_{\theta_{d}}(\mathbf{A}_{d-1}) =𝐐NlfR(𝐐Nl1fR(𝐐2fR(𝐐1𝐀d1)))\displaystyle=\mathbf{Q}_{N_{l}}f_{R}(\mathbf{Q}_{N_{l-1}}\cdots f_{R}(\mathbf{Q}_{2}f_{R}(\mathbf{Q}_{1}\mathbf{A}_{d-1}))\cdots) (32)
=𝐀d1𝐀d1𝐐NlfR(𝐐Nl1fR(𝐐2fR(𝐐1𝐀d1)))\displaystyle=\mathbf{A}_{d-1}\mathbf{A}_{d-1}^{{\dagger}}\mathbf{Q}_{N_{l}}f_{R}(\mathbf{Q}_{N_{l-1}}\cdots f_{R}(\mathbf{Q}_{2}f_{R}(\mathbf{Q}_{1}\mathbf{A}_{d-1}))\cdots)
=𝐀d1𝐖d.\displaystyle=\mathbf{A}_{d-1}\mathbf{W}_{d}.

Here, 𝐖d=𝐀d1𝐐NlfR(𝐐Nl1fR(𝐐2fR(𝐐1𝐀d1)))\mathbf{W}_{d}=\mathbf{A}_{d-1}^{{\dagger}}\mathbf{Q}_{N_{l}}f_{R}(\mathbf{Q}_{N_{l-1}}\cdots f_{R}(\mathbf{Q}_{2}f_{R}(\mathbf{Q}_{1}\mathbf{A}_{d-1}))\cdots), where 𝐀d1=𝐀d1H(𝐀d1𝐀d1H)1\mathbf{A}_{d-1}^{{\dagger}}=\mathbf{A}_{d-1}^{H}(\mathbf{A}_{d-1}\mathbf{A}_{d-1}^{H})^{-1} is the pseudoinverse of 𝐀d1\mathbf{A}_{d-1} and 𝐐l\mathbf{Q}_{l} represents the network weights at the ll-th, l{1,2,,Nl}l\in\{1,2,\cdots,N_{l}\}, layer. In addition, fR(x)=max(0,x)f_{R}(x)=\max(0,x) denotes the ReLU activation function. Note that although the output of the dd-th residual subnetwork can be expressed as the product of two matrices, the elements in 𝐖d\mathbf{W}_{d} are obtained through the non-linear operation and the residual subnetwork is still a non-linear network. Thus, the output of the dd-th denoising block can be expressed as

𝐀d=𝐀d1θd(𝐀d1)=𝐀d1𝐖~d=i=1d𝐀𝐖~i,\displaystyle\mathbf{A}_{d}=\mathbf{A}_{d-1}-\mathcal{R}_{\theta_{d}}(\mathbf{A}_{d-1})=\mathbf{A}_{d-1}\tilde{\mathbf{W}}_{d}=\prod_{i=1}^{d}\mathbf{A}\tilde{\mathbf{W}}_{i}, (33)

where 𝐀0=𝐀\mathbf{A}_{0}=\mathbf{A} as defined in (25) and 𝐖~d=𝐈N+1𝐖d,d{1,2,,D}\tilde{\mathbf{W}}_{d}=\mathbf{I}_{N+1}-\mathbf{W}_{d},\forall d\in\{1,2,\cdots,D\}. Based on (33), the equation (32) can be rewritten as

θd(𝐀d1)=i=1d1𝐀𝐖~i𝐖d.\mathcal{R}_{\theta_{d}}(\mathbf{A}_{d-1})=\prod_{i=1}^{d-1}\mathbf{A}\tilde{\mathbf{W}}_{i}\mathbf{W}_{d}. (34)

Thus, substituting (34) into (27), the expression of CDRN can be written as

gθ(𝐀)=𝐀d=1Dθd(𝐀d1)=𝐀𝐀𝐖1d=2Di=1d1𝐀𝐖~i𝐖d=𝐀𝐀𝐖,g_{\theta}(\mathbf{A})=\mathbf{A}-\sum\limits_{d=1}^{D}\mathcal{R}_{\theta_{d}}\left(\mathbf{A}_{d-1}\right)=\mathbf{A}-\mathbf{A}\mathbf{W}_{1}-\sum_{d=2}^{D}\prod_{i=1}^{d-1}\mathbf{A}\tilde{\mathbf{W}}_{i}\mathbf{W}_{d}=\mathbf{A}-\mathbf{A}\mathbf{W}, (35)

where 𝐖=𝐖1+𝟙Ω(D)d=2Di=1d1𝐖~i𝐖d\mathbf{W}=\mathbf{W}_{1}+\mathbbm{1}_{\Omega}(D)\sum_{d=2}^{D}\prod_{i=1}^{d-1}\tilde{\mathbf{W}}_{i}\mathbf{W}_{d} with Ω={a|a2,a}\Omega=\{a|a\geq 2,a\in\mathbb{Z}\}. Note that we adopt the coarse estimated value obtained by the LS method as the input of the proposed CDRN. Therefore, through learning from a large amount of LS-based training examples, the proposed CDRN estimator can exploit distinguishable features for denoising the LS-based input, i.e., our proposed method can further improve the estimation accuracy compared with the LS method. Based on this, we have 𝐀=𝐇~kLS\mathbf{A}=\tilde{\mathbf{H}}_{k}^{\mathrm{LS}}. In this case, the CDRN estimator, i.e., the well-trained CDRN can be formulated as

𝐇~kCDRN=gθ(𝐇~kLS)=𝐇~kLS𝐇~kLS𝐖=𝐇~kLS(𝐈N+1𝐖),\tilde{\mathbf{H}}_{k}^{\mathrm{CDRN}}=g_{\theta^{*}}(\tilde{\mathbf{H}}_{k}^{\mathrm{LS}})=\tilde{\mathbf{H}}_{k}^{\mathrm{LS}}-\tilde{\mathbf{H}}_{k}^{\mathrm{LS}}\mathbf{W}^{*}=\tilde{\mathbf{H}}_{k}^{\mathrm{LS}}\left(\mathbf{I}_{N+1}-\mathbf{W}^{*}\right), (36)

where 𝐖=𝐖1+𝟙Ω(D)d=1Di=1d1𝐖~i𝐖d\mathbf{W}^{*}=\mathbf{W}^{*}_{1}+\mathbbm{1}_{\Omega}(D)\sum_{d=1}^{D}\prod_{i=1}^{d-1}\tilde{\mathbf{W}}_{i}^{*}\mathbf{W}_{d}^{*} denotes the well-trained matrix obtained from a well-trained network with parameter θ\theta^{*}. Thus, the proposed CDRN is a data-driven non-linear estimator which can exploit more distinguishable features from a large amount of training examples to further improve estimation accuracy. On the other hand, according to (10) and (12), the LMMSE estimator can be rewritten as [46]

𝐇~kLMMSE\displaystyle\tilde{\mathbf{H}}_{k}^{\mathrm{LMMSE}} =𝐇kLS(𝐑𝐇k+Mσz2C𝐈N+1)1𝐑𝐇k\displaystyle=\mathbf{H}_{k}^{\mathrm{LS}}\left(\mathbf{R}_{\mathbf{H}_{k}}+\frac{M\sigma_{z}^{2}}{C}\mathbf{I}_{N+1}\right)^{-1}\mathbf{R}_{\mathbf{H}_{k}} (37)
=𝐇kLS(𝐈N+1𝐑𝐇k1(𝐑𝐇k1+CMσz2𝐈N+1)1).\displaystyle=\mathbf{H}_{k}^{\mathrm{LS}}\left(\mathbf{I}_{N+1}-\mathbf{R}_{\mathbf{H}_{k}}^{-1}\left(\mathbf{R}_{\mathbf{H}_{k}}^{-1}+\frac{C}{M\sigma_{z}^{2}}\mathbf{I}_{N+1}\right)^{-1}\right).

Based on (36) and (37), the proposed CDRN estimator is equivalent to the LMMSE estimator when 𝐖=𝐑𝐇k1(𝐑𝐇k1+CMσz2𝐈N+1)1\mathbf{W}^{*}=\mathbf{R}_{\mathbf{H}_{k}}^{-1}\left(\mathbf{R}_{\mathbf{H}_{k}}^{-1}+\frac{C}{M\sigma_{z}^{2}}\mathbf{I}_{N+1}\right)^{-1} is adopted. That is, the proposed CDRN estimator is a generalized framework which subsumes an LMMSE estimator in a data driven manner as a subcase.

According to the universal approximation theorem [53], the neural network has powerful capability in algorithmic learning and has been proven to be a universal function approximator. Therefore, based on (36), (37), and the universal approximation theorem, the proposed CDRN estimator is capable of approximating the LMMSE estimator through exploiting distinguishable features from a large amount of training examples, i.e., the proposed CDRN estimator can achieve the same performance as that of the LMMSE estimator. Specifically, since the optimal MMSE estimator behaves the same as the LMMSE estimator when the prior PDF of 𝐇k\mathbf{H}_{k} is a Gaussian PDF, the proposed CDRN estimator may achieve the optimal performance for the case of a Rayleigh fading channel. On the other hand, when the prior PDF of 𝐇k\mathbf{H}_{k} is not a Gaussian PDF, the optimal MMSE estimator does not admit a closed-form. Then, the proposed CDRN estimator can exploit more distinguishable features from a large amount of training examples to achieve satisfactory estimation performance, while the LMMSE estimator can only obtain the limited estimation performance which is merely based on a second order channel correlation matrix. These observations will be verified through simulations in the following section.

TABLE II: Simulation Settings
Parameters      Default Values
Path loss exponent of UkU_{k}-BS γ1=3.6\gamma_{1}=3.6
Path loss exponent of IRS-BS γ2=2.3\gamma_{2}=2.3
Path loss exponent of UkU_{k}-IRS γ3=2\gamma_{3}=2
Rician factor of UkU_{k}-BS βUB=0\beta_{\mathrm{UB}}=0
Rician factor of IRS-BS βIB=10\beta_{\mathrm{IB}}=10
Rician factor of UkU_{k}-IRS βUI=0\beta_{\mathrm{UI}}=0

V Numerical Results

In this section, extensive simulations are provided to verify the efficientness of the proposed algorithm. In the simulations, an IRS-MUC system consists of a BS equipped with MM antennas, an IRS with NN reflecting elements, and KK single-antenna users is considered, as defined in Fig. 1. Unless further specified, the IRS operates with continuous phase shifts and the default settings of the considered IRS-MUC system are M=8M=8, N=32N=32, K=6K=6, and C=N+1=33C=N+1=33. A path loss model is adopted in the simulations and the path losses of each channel can be modeled as αkUB=α0(λkUB/λ0)γ1\alpha_{k}^{\mathrm{UB}}=\alpha_{0}(\lambda_{k}^{\mathrm{UB}}/\lambda_{0})^{-\gamma_{1}}, αkUI=α0(λkUI/λ0)γ2\alpha_{k}^{\mathrm{UI}}=\alpha_{0}(\lambda_{k}^{\mathrm{UI}}/\lambda_{0})^{-\gamma_{2}}, and αIB=α0(λIB/λ0)γ3\alpha^{\mathrm{IB}}=\alpha_{0}(\lambda^{\mathrm{IB}}/\lambda_{0})^{-\gamma_{3}}, where γi\gamma_{i}, i{1,2,3}i\in\{1,2,3\}, is the path loss exponent, λ0=10m\lambda_{0}=10~{}\mathrm{m} is the reference distance, α0=15dB\alpha_{0}=-15~{}\mathrm{dB} is the path loss at the reference distance, and λkUB\lambda_{k}^{\mathrm{UB}}, λkUI\lambda_{k}^{\mathrm{UI}}, and λIB\lambda^{\mathrm{IB}} represent the distances of UkU_{k}-BS, UkU_{k}-IRS, and IRS-BS, respectively. Specifically, we set λkUB=100\lambda_{k}^{\mathrm{UB}}=100 m, λkUI=16\lambda_{k}^{\mathrm{UI}}=16 m, and λIB=90\lambda^{\mathrm{IB}}=90 m. In addition, a Rician fading model is considered for all the channels to characterize the small-scale fading. In this case, the channel of IRS-BS link can be expressed as

𝐆=βIBβIB+1𝐆LOS+1βIB+1𝐆NLOS,\mathbf{G}=\sqrt{\frac{\beta_{\mathrm{IB}}}{\beta_{\mathrm{IB}}+1}}\mathbf{G}_{\mathrm{LOS}}+\sqrt{\frac{1}{\beta_{\mathrm{IB}}+1}}\mathbf{G}_{\mathrm{NLOS}}, (38)

where βIB\beta_{\mathrm{IB}} represents the Rician factor of IRS-BS channel, and 𝐆LOS\mathbf{G}_{\mathrm{LOS}} and 𝐆NLOS\mathbf{G}_{\mathrm{NLOS}} are the line-of-sight (LOS) component and the Rayleigh component, respectively. Note that the above model expression is a general model which involves both the LOS channel model (when βIB\beta_{\mathrm{IB}}\to\infty) and the Rayleigh channel model (when βIB=0\beta_{\mathrm{IB}}=0). The channels of BS-UkU_{k} and IRS-UkU_{k} are generated similarly as defined in (38). Similarly, we define the Rician factors of channels of UkU_{k}-BS and UkU_{k}-IRS as βUB\beta_{\mathrm{UB}} and βUI\beta_{\mathrm{UI}}, respectively. To consider the impacts of path loss factors, the elements in 𝐆\mathbf{G} should then be multiplied by αIB\sqrt{\alpha^{\mathrm{IB}}}. The detailed settings of the simulations are summarized in Table \@slowromancapii@, where the settings of UkU_{k}-BS link, i.e., βUB=0\beta_{\mathrm{UB}}=0 and γ1=3.6\gamma_{1}=3.6, are due to the rich scattering environment and the relatively large distance between UkU_{k} and the BS. Specifically, a normalized MSE (NMSE) is adopted as the estimation performance metric: NMSE=E(𝐇~𝐇F2)E(𝐇F2)\mathrm{NMSE}=\frac{E(\|\tilde{\mathbf{H}}-\mathbf{H}\|_{F}^{2})}{E(\|\mathbf{H}\|_{F}^{2})} where 𝐇~\tilde{\mathbf{H}} and 𝐇\mathbf{H} denote the estimated channel and the ground truth of the channel, respectively. In addition, the signal-to-noise ratio (SNR) used in the simulation results refers to the transmit SNR, i.e., SNR=𝒫/σz2\mathrm{SNR}=\mathcal{P}/\sigma_{z}^{2}. To evaluate the estimation performance, we also provide the simulation of the following algorithms for comparisons: the optimal MMSE method [46], the LMMSE method [46], the LS method [45], the bilinear alternating least squares (BALS) method [22], and the binary reflection controlled LMMSE (B-LMMSE) method. For the B-LMMSE method, we exploit the binary reflection controlled strategy proposed in [19] and then adopt the LMMSE estimator for channel estimation. In addition, a CDRN equipped with D=3D=3 denoising blocks is adopted for our proposed method and the hyperparameters of CDRN are set according to Table \@slowromancapi@. Each simulation point of the presented results is obtained by averaging the performance over 100,000100,000 Monte Carlo realizations444The source code is available at https://github.com/XML124/CDRN-channel-estimation-IRS..

Next, we will verify the channel estimation performance of the proposed CDRN method in terms of NMSE and unveil its relationship with SNR and channel dimensionality, respectively. Besides, visualizations of our proposed CDRN are provided to illustrate the denoising process. Finally, the complexity analysis is also investigated to evaluate the computational complexity.

Refer to caption
Figure 7: NMSE performance with different SNRs under MM = 8, N=32N=32, and C=33C=33.

V-A NMSE versus SNR

The results of NMSE under different SNRs are presented in Fig. 7. Since the cascaded channel does not follow the Rayleigh fading model, it is generally intractable to design the optimal MMSE estimator and derive the closed-form of LMMSE estimator. Hence, only an empirical LMMSE (denoted by E-LMMSE) can be obtained where the required statistical channel correlation matrix is replaced by a Monte Carlo based empirical channel correlation matrix. As expected, the NMSEs of all the algorithms decrease with the increase of SNR since a higher transmit power can effectively mitigate the impact of noise. Now, we first investigate the estimation performance in terms of different IRS control strategies. It can be seen that the DFT controlled methods (denoted by LS, E-LMMSE, and CDRN) outperform the B-LMMSE method significantly, especially under the high SNR regions. This is because for each time slot, all the NN elements are switched on for reflecting the incident signals when adopting the DFT controlled strategy, while only one element is switched on under the binary controlled strategy, leading to insufficient received signal power for channel estimation. We then discuss the performance of the schemes adopting the DFT controlled methods. It is noticed that the LS method has a performance gap compared with the E-LMMSE method since the E-LMMSE can capture the statistical knowledge of the channel, while the LS method is designed by treating the channel as a deterministic but unknown constant. Compared with LMMSE, the proposed CDRN method further improves the estimation performance and thus achieves the best performance among all the considered algorithms. For example, the proposed CDRN method achieves a SNR gain of 55 dB in terms of the NMSE 101\approx 10^{-1} compared with the E-LMMSE method. The reason is that the E-LMMSE method is developed based on the constrained linear estimator. In contrast, the proposed method adopts the non-linear CDRN to recover the channels by intelligently exploiting the spatial features of channels in a data driven approach. Besides, it can be observed from Fig. 7 that our proposed method can achieve a better NMSE performance compared with the BALS method. This is because the BALS method does not explore the statistical features of channels. In contrast, our proposed method can further improve the estimation performance by exploiting the channel statistical features from a large amount of LS-based training examples. On the other hand, low-quality phase shifters with a small number of quantization levels may be deployed to reduce the implementation cost [8, 12, 24]. Therefore, it is necessary to investigate if the proposed channel estimation scheme is robust against the quantization of IRS phase shifts. As shown in Fig. 7, the proposed CDRN method with a 2-bit uniformly quantized IRS phase shift [8, 24] (denoted by CDRN with 2-bit quantization) almost behaves the same as the CDRN method with a continuous IRS phase shift. This is because through offline training, the proposed CDRN can learn distinguishable features to address the introduced noise terms by the phase quantization errors.

Refer to caption

(a) 𝐁k\mathbf{B}_{k} for reflecting link.                     (b) 𝐝k\mathbf{d}_{k} for direct link.    

Figure 8: NMSE performance with different SNRs under MM = 8, N=32N=32, and C=33C=33.

To further evaluate the estimation performance, we present the NMSE results of the direct channel 𝐁k\mathbf{B}_{k} and the reflection channel 𝐝k\mathbf{d}_{k} in Fig. 8(a) and Fig. 8(b), respectively. Similar to the results in Fig. 7, Fig. 8(a) shows that the proposed CDRN still achieves the best performance in terms of the channel estimation of the reflection channel 𝐁k\mathbf{B}_{k}. Since the direct channel follows Rayleigh fading model, the optimal MMSE method and the LMMSE method achieve the same performance as they have the same mathematical expression as in (12). Therefore, we compare three methods in Fig. 8(b): the LS method, the optimal MMSE method, and the proposed CDRN method. It is shown that the CDRN method outperforms the LS method and achieves almost the same estimation performance as that of the optimal MMSE method, which presents the optimality of the proposed method.

V-B NMSE versus Channel Dimensionality

Note that the channel dimensionality is also an important parameter which affects the estimation performance. To illustrate the scalability of the proposed CDRN, we conduct simulations to evaluate the estimation performance with different channel dimensionality. Specifically, the number of elements of IRS, denoted by NN, directly affects the number of columns of channel matrix 𝐇k\mathbf{H}_{k} and thus Fig. 9 plots the curves of NMSE versus NN for different algorithms. It can be seen that the NMSE of DFT controlled methods (denoted by LS, E-LMMSE, and CDRN) decrease with the increase of NN, while the NMSE binary controlled strategy based method, i.e., the B-MMSE method, remains almost a constant value for different NN. This is because the B-MMSE method only switches on one element each time slot despite the total number of reflecting elements at the IRS. In addition, the proposed CDRN method still achieves the best performance among all the considered algorithms since it is able to exploit more distinguishable features of channels from a larger size of input channel matrix. Besides NN, the number of antennas at the BS, denoted by MM, determines the number of rows of 𝐇k\mathbf{H}_{k} which also plays an important role in estimation performance. Therefore, Fig. 10 investigates the effect of MM on the NMSE performance. It is obvious that the NMSE is insensitive to the change of MM. This is because in the metric of NMSE, the array gain due to the increase of MM achieved in the numerator is neutralized by the one in the denominator.

Refer to caption
Figure 9: NMSE performance with different numbers of reflecting elements of IRS under MM = 8, C=N+1C=N+1, and SNR = 5 dB.
Refer to caption
Figure 10: NMSE performance with different numbers of antennas at the BS under N=16N=16, C=17C=17, and SNR = 10 dB.
Refer to caption
Figure 11: NMSE performance with different numbers of pilots under M=8M=8, N=16N=16, and SNR = 5 dB.

Moreover, the number of pilots CN+1C\geq N+1 has also impacts on the NMSE. It is shown in Fig. 11 that the NMSEs of all the algorithms decrease with the increase of the number of pilots and different methods have different slopes with respect to CC. In particular, the NMSE of B-LMMSE method decreases gently with the increase of CC, while the NMSEs of LS, E-LMMSE, and CDRN methods decrease sharply. The reason is that for the B-LMMSE method, the power of the received signal at the BS is very limited which is only 1/N1/N times of that of the other methods. In addition, the proposed CDRN could exploit more distinguishable spatial features from a larger scale input matrix to enhance the estimation accuracy, and thus achieves the best performance among all the considered algorithms, as presented in Fig. 11.

V-C Visualization of CDRN

To offer more insights of our proposed CDRN, we present the visualization of the outputs of each denoising block of a well-trained CDRN. For the convenience of understanding, we only present the denoising process of ten points of a channel matrix, i.e., only these ten points are the received signals at the BS while the remaining points are the noise samples. For visualization, we transform the input matrix and the output matrices of each denoising block to the grayscale images. In particular, the elements of each matrix are normalized to an interval of [0,1][0,1], where 0 and 11 are represented by black color and white color, respectively.

According to Fig. 5, the input of CDRN, the output of the first, second, and last denoising blocks are represented by 𝐀\mathbf{A}, 𝐀1\mathbf{A}_{1}, 𝐀2\mathbf{A}_{2}, and 𝐀3\mathbf{A}_{3}, respectively. The grayscaled image of 𝐀\mathbf{A} is shown in Fig. 12, where the ten distinguishable bright pixels are the interest channel coefficients while the remaining pixels are the noise samples. The cluttered noise pixels indicate that the noise samples have a large variance. Subsequently, Fig. 12 and Fig. 12 are the visualizations of 𝐀1\mathbf{A}_{1} and 𝐀2\mathbf{A}_{2}, where we find that the noise pixels become tidy, i.e., the noisy channel matrix are denoised progressively. Finally, the output of 𝐀3\mathbf{A}_{3}, i.e., the output of CDRN is visualized in Fig. 12, where the noise pixels are largely eliminated and thus a denoised channel matrix is obtained. In fact, the proposed CDRN method adopts a stage-by-stage denoising mechanism and achieves a satisfactory performance.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 12: A visualization of denoising blocks of a well-trained CDRN with D=3D=3 denoising blocks under SNR = 16 dB. (a) 𝐀\mathbf{A}, the input of CDRN; (b) 𝐀1\mathbf{A}_{1}, the output of the first denoising block; (c) 𝐀2\mathbf{A}_{2}, the output of the second denoising block; (d) 𝐀3\mathbf{A}_{3}, the output of the last denoising block, i.e., the output of CDRN.

V-D Complexity Analysis

Finally, we analyze the computational complexities of the considered algorithms and summarize the results in Table III. Here, n0n_{0} denotes the input dimension, nln_{l} and sls_{l} represent the number of neural network channels and the spatial size (length of side) of the filter of the ll-th, l{1,2,,Nl}l\in\{1,2,\cdots,N_{l}\}, convolutional layer, respectively. In addition, NtN_{t} and ItI_{t} are both defined in Algorithm 1. Specifically, both the B-LMMSE method and the E-LMMSE method have the same computational complexity, thus we use LMMSE to denote the B-LMMSE/E-LMMSE method. Note that the computational complexities of the LS and LMMSE methods only arise from the online estimation, while the proposed CDRN method has an additional computational complexity due to the offline training. Specifically, the LS method can be realized via the inverse fast Fourier transform (IFFT). While the computational cost of the LMMSE method mainly comes from the matrix inversions. As for the proposed method, its computational cost is dominated by the computations of the CDRN [54] and the LS-based input. For simplicity, we only present the highest order terms of the complexities in Table III, as commonly adopted in e.g., [55]. It can be seen from Table III that compared with the LS and LMMSE methods, the proposed CDRN method can achieve the best system performance at the expense of a higher computational complexity. On the other hand, although the proposed CDRN involves a large number of parameters, the required actual online estimation time can be greatly reduced by exploiting the parallelization of a graphics processing unit (GPU) [48]. This conclusion can be verified in Table IV, where each result is obtained by executing the algorithms on a desktop computer with an i7-8700 3.20 GHz central processing unit (CPU) and a Nvidia GeForce RTX 2070 GPU. Therefore, although the proposed CDRN method has a relatively higher computational complexity compared with the LS and LMMSE methods, the associated computation time can be greatly reduced through exploiting the parallel computing of GPU.

TABLE III: Computational complexities of different estimation algorithms
Algorithm Online Estimation Offline Training
LS O(MClog2(MC))O(MC\log_{2}(MC)) -
LMMSE O(C3+MC2)O(C^{3}+MC^{2}) -
CDRN O(DM(N+1)l=1Nlnl1sl2nl)O\left(DM(N+1)\sum\limits_{l=1}^{N_{l}}n_{l-1}\cdot s_{l}^{2}\cdot n_{l}\right) O(NtItDM(N+1)l=1Lnl1sl2nl)O\left(N_{t}I_{t}DM(N+1)\sum\limits_{l=1}^{L}n_{l-1}\cdot s_{l}^{2}\cdot n_{l}\right)
TABLE IV: Computation time (in seconds) of different estimation algorithms
Algorithm Online Estimation Offline Training
LS 1.26×1031.26\times 10^{-3} -
LMMSE 4.56×1034.56\times 10^{-3} -
CDRN 2.66×1032.66\times 10^{-3} 492.58492.58

VI Conclusion

This paper proposed a data-driven DReL approach to address the channel estimation problem for IRS-MUC systems operating with TDD protocol. Firstly, we formulated the channel estimation problem in IRS-MUC systems as a denoising problem and developed a versatile DReL-based channel estimation framework. Specifically, according to the Bayesian MMSE criterion, a DRN-based MMSE was derived and analyzed in terms of Bayesian philosophy. Under the developed DReL-based framework, we then designed a CDRN to denoise the noisy channel matrix for channel recovery. For the proposed CDRN, a CNN-based denoising block with an element-wise subtraction structure was specifically designed to exploit both the spatial features of the noisy channel matrices and the additive nature of the noise simultaneously. Taking advantages of the superiorities of CNN and DReL in feature extraction and denoising, the proposed CDRN estimation algorithm can further improve the estimation accuracy. Simulation results demonstrated that the proposed method can achieve almost the same estimation accuracy as that of the optimal MMSE estimator based on a prior PDF of channel.

References

  • [1] C. Liu, X. Liu, D. W. K. Ng, and J. Yuan, “Deep residual network empowered channel estimation for IRS-assisted multi-user communication systems,” in Proc. IEEE Int. Conf. Commun. (ICC), 2021 [To appear].
  • [2] S. Gong, X. Lu, D. T. Hoang, D. Niyato, L. Shu, D. I. Kim, and Y.-C. Liang, “Towards smart wireless communications via intelligent reflecting surfaces: A contemporary survey,” IEEE Commun. Surveys Tuts., vol. 22, no. 4, 4th Quart. 2020.
  • [3] J. Zhao and Y. Liu, “A survey of intelligent reflecting surfaces (IRSs): Towards 6G wireless communication networks,” arXiv preprint arXiv:1907.04789, 2019.
  • [4] C. Liaskos, S. Nie, A. Tsioliaridou, A. Pitsillides, S. Ioannidis, and I. Akyildiz, “A new wireless communication paradigm through software-controlled metasurfaces,” IEEE Comm. Mag., vol. 56, no. 9, pp. 162–169, Sept. 2018.
  • [5] V. W. S. Wong, R. Schober, D. W. K. Ng, and L.-C. Wang, Key technologies for =5G wireless systems. Cambridge University Press, 2017.
  • [6] J. Zhang, E. Björnson, M. Matthaiou, D. W. K. Ng, H. Yang, and D. J. Love, “Prospective multiple antenna technologies for beyond 5G,” IEEE J. Sel. Areas Commun., vol. 38, no. 8, pp. 1637–1660, Aug. 2020.
  • [7] H. Yang, X. Cao, F. Yang, J. Gao, S. Xu, M. Li, X. Chen, Y. Zhao, Y. Zheng, and S. Li, “A programmable metasurface with dynamic polarization, scattering and focusing control,” Sci. Rep., vol. 6, no. 35692, pp. 1–11, Oct. 2016.
  • [8] Q. Wu and R. Zhang, “Towards smart and reconfigurable environment: Intelligent reflecting surface aided wireless network,” IEEE Comm. Mag., vol. 58, no. 1, pp. 106–112, Jan. 2020.
  • [9] M. Di Renzo et al., “Smart radio environments empowered by reconfigurable AI meta-surfaces: An idea whose time has come,” EURASIP J. Wireless Commun. Netw., vol. 2019, no. 129, pp. 1–20, May 2019.
  • [10] C. Huang, S. Hu, G. C. Alexandropoulos, A. Zappone, C. Yuen, R. Zhang, M. Di Renzo, and M. Debbah, “Holographic MIMO surfaces for 6G wireless networks: Opportunities, challenges, and trends,” IEEE Wireless Commun., vol. 27, no. 5, pp. 118–125, Oct. 2020.
  • [11] G. C. Alexandropoulos, G. Lerosey, M. Debbah, and M. Fink, “Reconfigurable intelligent surfaces and metamaterials: The potential of wave propagation control for 6G wireless communications,” IEEE ComSoc TCCN Newsl., vol. 6, no. 1, pp. 25–37, Jun. 2020.
  • [12] Q. Wu and R. Zhang, “Intelligent reflecting surface enhanced wireless network via joint active and passive beamforming,” IEEE Trans. Wireless Commun., vol. 18, no. 11, pp. 5394–5409, Nov. 2019.
  • [13] S. Zhang and R. Zhang, “Capacity characterization for intelligent reflecting surface aided MIMO communication,” IEEE J. Sel. Areas Commun., vol. 38, no. 8, pp. 1823–1838, Aug. 2020.
  • [14] H. Guo, Y.-C. Liang, J. Chen, and E. G. Larsson, “Weighted sum-rate maximization for reconfigurable intelligent surface aided wireless networks,” IEEE Trans. Wireless Commun., vol. 19, no. 5, pp. 3064–3076, May 2020.
  • [15] C. Huang, A. Zappone, G. C. Alexandropoulos, M. Debbah, and C. Yuen, “Reconfigurable intelligent surfaces for energy efficiency in wireless communication,” IEEE Trans. Wireless Commun., vol. 18, no. 8, pp. 4157–4170, Aug. 2019.
  • [16] S. Zhou, W. Xu, K. Wang, M. Di Renzo, and M.-S. Alouini, “Spectral and energy efficiency of IRS-assisted MISO communication with hardware impairments,” IEEE Wireless Commun. Lett., vol. 9, no. 9, Sept. 2020.
  • [17] Z. Chu, W. Hao, P. Xiao, and J. Shi, “Intelligent reflecting surface aided multi-antenna secure transmission,” IEEE Wireless Commun. Lett., vol. 9, no. 1, pp. 108–112, Jan. 2020.
  • [18] X. Guan, Q. Wu, and R. Zhang, “Intelligent reflecting surface assisted secrecy communication: Is artificial noise helpful or not ?,” IEEE Wireless Commun. Lett., vol. 9, no. 6, pp. 778–782, Jan. 2020.
  • [19] D. Mishra and H. Johansson, “Channel estimation and low-complexity beamforming design for passive intelligent surface assisted MISO wireless energy transfer,” in Proc. IEEE Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 4659–4663, Brighton, United Kingdom, May 2019.
  • [20] Y. Yang, B. Zheng, S. Zhang, and R. Zhang, “Intelligent reflecting surface meets OFDM: Protocol design and rate maximization,” IEEE Trans. Commun., vol. 68, no. 7, pp. 4522–4535, Jul. 2020.
  • [21] T. L. Jensen and E. De Carvalho, “An optimal channel estimation scheme for intelligent reflecting surfaces based on a minimum variance unbiased estimator,” in Proc. IEEE Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 5000–5004, Barcelona, Spain, May 2020.
  • [22] G. T. de Araújo and A. L. F. de Almeida, “PARAFAC-based channel estimation for intelligent reflective surface assisted MIMO system,” in Proc. IEEE Sens. Array Multichannel Signal Process. Workshop (SAM), pp. 1–5, Hangzhou, China, Jun. 2020.
  • [23] B. Zheng and R. Zhang, “Intelligent reflecting surface-enhanced OFDM: Channel estimation and reflection optimization,” IEEE Wireless Commun. Lett., vol. 9, no. 4, pp. 518–522, Apr. 2020.
  • [24] C. You, B. Zheng, and R. Zhang, “Intelligent reflecting surface with discrete phase shifts: Channel estimation and passive beamforming,” in Proc. IEEE Int. Conf. Commun. (ICC), pp. 1–6, Dublin, Ireland, Jun. 2020.
  • [25] J. Chen, Y.-C. Liang, H. V. Cheng, and W. Yu, “Channel estimation for reconfigurable intelligent surface aided multi-user MIMO systems,” arXiv preprint arXiv:1912.03619, 2019.
  • [26] Z.-Q. He and X. Yuan, “Cascaded channel estimation for large intelligent metasurface assisted massive MIMO,” IEEE Wireless Commun. Lett., vol. 9, no. 2, pp. 210–214, Feb. 2020.
  • [27] Z. Wang, L. Liu, and S. Cui, “Channel estimation for intelligent reflecting surface assisted multiuser communications,” in Proc. IEEE Wireless. Commun. Netw. Conf. (WCNC), Seoul, Korea (South), May 2020.
  • [28] Z. Wang, L. Liu, and S. Cui, “Channel estimation for intelligent reflecting surface assisted multiuser communications: Framework, algorithms, and analysis,” IEEE Trans. Wireless Commun., vol. 19, no. 10, pp. 6607–6620, Oct. 2020.
  • [29] G. C. Alexandropoulos and E. Vlachos, “A hardware architecture for reconfigurable intelligent surfaces with minimal active elements for explicit channel estimation,” in Proc. IEEE Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 9175–9179, Barcelona, Spain, May 2020.
  • [30] G. E. Hinton and R. R. Salakhutdinov, “Reducing the dimensionality of data with neural networks,” Science, vol. 313, no. 5786, pp. 504–507, Jul. 2006.
  • [31] Y. Lecun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521, no. 7553, pp. 436–444, May 2015.
  • [32] T. Wang, C.-K. Wen, H. Wang, F. Gao, T. Jiang, and S. Jin, “Deep learning for wireless physical layer: Opportunities and challenges,” China Commun., vol. 14, no. 11, pp. 92–111, Nov. 2017.
  • [33] A. Taha, M. Alrabeiah, and A. Alkhateeb, “Enabling large intelligent surfaces with compressive sensing and deep learning,” arXiv preprint arXiv:1904.10136, 2019.
  • [34] A. Taha, M. Alrabeiah, and A. Alkhateeb, “Deep learning for large intelligent surfaces in millimeter wave and massive mimo systems,” in Proc. IEEE Global Telecommun. Conf. (GLOBECOM), pp. 1–6, Waikoloa, HI, USA, Dec. 2019.
  • [35] C. Huang, G. C. Alexandropoulos, C. Yuen, and M. Debbah, “Indoor signal focusing with deep learning designed reconfigurable intelligent surfaces,” in Proc. IEEE Int. Workshop Signal Process. Adv. Wireless Commun. (SPAWC), pp. 1–5, Cannes, France, Jul. 2019.
  • [36] G. C. Alexandropoulos, S. Samarakoon, M. Bennis, and M. Debbah, “Phase configuration learning in wireless networks with multiple reconfigurable intelligent surfaces,” arXiv preprint arXiv:2010.04376, 2020.
  • [37] K. Feng, Q. Wang, X. Li, and C.-K. Wen, “Deep reinforcement learning based intelligent reflecting surface optimization for MISO communication systems,” IEEE Wireless Commun. Lett., vol. 9, no. 5, pp. 745–749, May 2020.
  • [38] H. Yang, Z. Xiong, J. Zhao, D. Niyato, and L. Xiao, “Deep reinforcement learning based intelligent reflecting surface for secure wireless communications,” IEEE Trans. Wireless Commun., Sept. 2020, [Early Access].
  • [39] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 770–778, Las Vegas, NV, USA, Jun. 2016.
  • [40] K. Zhang, W. Zuo, Y. Chen, D. Meng, and L. Zhang, “Beyond a =Gaussian denoiser: =Residual learning of deep =CNN for image denoising,” IEEE Trans. Image Proc., vol. 26, no. 7, pp. 3142–3155, Jul. 2017.
  • [41] H. He, C.-K. Wen, S. Jin, and G. Y. Li, “Deep learning-based channel estimation for beamspace mmwave massive MIMO systems,” IEEE Wireless Commun. Lett., vol. 7, no. 5, pp. 852–855, Oct. 2018.
  • [42] D. Neumann, T. Wiese, and W. Utschick, “Learning the MMSE channel estimator,” IEEE Trans. Signal Proc., vol. 66, no. 11, pp. 2905–2917, Jun. 2018.
  • [43] X. Chen and M. Jiang, “Adaptive statistical bayesian MMSE channel estimation for visible light communication,” IEEE Trans. Signal Proc., vol. 65, no. 5, pp. 1287–1299, Mar. 2017.
  • [44] J. Mirzaei, F. Sohrabi, R. Adve, and S. ShahbazPanahi, “MMSE-based channel estimation for hybrid beamforming massive MIMO with correlated channels,” in Proc. IEEE Int. Conf. Acoust. Speech Signal Process. (ICASSP), pp. 5015–5019, Barcelona, Spain, May 2020.
  • [45] M. Biguesh and A. B. Gershman, “Training-based MIMO channel estimation: A study of estimator tradeoffs and optimal training signals,” IEEE Trans. Signal Proc., vol. 54, no. 3, pp. 884–893, Mar. 2006.
  • [46] S. M. Kay, Fundamentals of statistical signal processing, volume =I: Estimation theory. Prentice Hall, 1993.
  • [47] J.-M. Kang, “Intelligent reflecting surface: Joint optimal training sequence and refection pattern,” IEEE Commun. Lett., vol. 24, no. 8, pp. 1784–1788, Aug. 2020.
  • [48] I. Goodfellow, Y. Bengio, A. Courville, and Y. Bengio, Deep learning. MIT Press Cambridge, 2016.
  • [49] C. Liu, J. Wang, X. Liu, and Y.-C. Liang, “Deep =CM-CNN for spectrum sensing in cognitive radio,” IEEE J. Sel. Areas Commun., vol. 37, no. 10, pp. 2306–2321, Oct. 2019.
  • [50] D. Tse and P. Viswanath, Fundamentals of wireless communication. Cambridge University Press, 2005.
  • [51] S. Ioffe and C. Szegedy, “Batch normalization: Accelerating deep network training by reducing internal covariate shift,” in Proc. IEEE Int. Conf. Mach. Learn. (ICML), pp. 448–456, Lille, France, Jul. 2015.
  • [52] X. Glorot, A. Bordes, and Y. Bengio, “Deep sparse rectifier neural networks,” in Proc. Int. Conf. Artif. Intell. Stat., pp. 315–323, Fort Lauderdale, USA, Apr. 2011.
  • [53] K. Hornik, M. Stinchcombe, and H. White, “Multilayer feedforward networks are universal approximators,” Neural Networks, vol. 2, no. 5, pp. 359–366, 1989.
  • [54] K. He and J. Sun, “Convolutional neural networks at constrained time cost,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 5353–5360, Boston, MA, USA, Jun. 2015.
  • [55] S. Arora and B. Barak, Computational complexity: A modern approach. Cambridge University Press, 2009.