This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Deep Learning Phase Compression for MIMO CSI Feedback by Exploiting FDD Channel Reciprocity

Yu-Chien Lin, Zhenyu Liu, Ta-Sung Lee, and Zhi Ding Y.-C Lin is with the Department of Electrical and Computer Engineering, University of California at Davis, Davis, CA, USA, and was affiliated with National Yang Ming Chiao Tung University, Taiwan (e-mail: [email protected]). Z. Liu was with University of California, Davis, CA, USA and is now with Beijing University of Posts and Telecommunications, China (e-mail: [email protected]). T.-S Lee is with the Institute of Communications Engineering, National Yang Ming Chiao Tung University, Taiwan (e-mail: [email protected]). Z. Ding is with the Department of Electrical and Computer Engineering, University of California, Davis, CA, USA (e-mail: [email protected]).This work was supported by the Center for Open Intelligent Connectivity from The Featured Areas Research Center Program within the framework of the Higher Education Sprout Project by the Ministry of Education of Taiwan.
Abstract

Large scale MIMO FDD systems are often hampered by bandwidth required to feedback downlink CSI. Previous works have made notable progresses in efficient CSI encoding and recovery by taking advantage of FDD uplink/downlink reciprocity between their CSI magnitudes. Such framework separately encodes CSI phase and magnitude. To further enhance feedback efficiency, we propose a new deep learning architecture for phase encoding based on limited CSI feedback and magnitude-aided information. Our contribution features a framework with a modified loss function to enable end-to-end joint optimization of CSI magnitude and phase recovery. Our test results show superior performance in indoor/outdoor scenarios.

Index Terms:
CSI feedback, massive MIMO, deep learning

I Introduction

Massive multiple-input multiple-output (MIMO) transceiver systems have demonstrated significant success in achieving high spectrum and energy efficiency for 5G and future wireless communication systems. These high achievable benefits require sufficiently accurate downlink (DL) channel state information (CSI) at the gNB (i.e., gNodeB). Frequency-division duplexing (FDD) systems, however, can only estimate DL CSI through feedback from UEs because DL and uplink (UL) channels occupy different frequency bands and may exhibit different channel characteristics. Since the feedback resource overhead grows proportionally with increasing MIMO size and spectrum, reducing CSI feedback overhead is vital to the widespread deployment of FDD MIMO systems. To improve feedback efficiency, previous works in [1, 2] developed a deep neural network with an autoencoder structure whose encoder at UEs and base stations, respectively for CSI compression and recovery. Related works and variants [3] have demonstrated performance advantages over traditional compressive sensing approaches.

Recent works have revealed the importance of exploiting correlated channel information such as UL CSI [4, 5], past CSI [6], and CSI adjacent UEs [7] for improving the accuracy of DL CSI recovery at base stations. In particular, important physical insights regarding FDD reciprocity, slow changes in propagation environment in time, and similar propagation conditions within short geographical distance, underscore respectively the strong spectral, temporal, and spatial correlations between magnitudes of different CSI in delay-angle (DA) domain. Since, side information from correlated CSI lowers the conditional entropy (uncertainty) of the DL CSI, their effective utilization reduces encoded feedback payload required from UEs [6].

Unlike the DA domain CSI magnitudes which tend to clearly exhibit strong temporal, spectral, and spatial correlations, CSI phases are very sensitive to changes in time, frequency, and location. To exploit the side CSI information, existing solutions have adopted dual feedback framework which separately encodes and recovers CSI phases from their corresponding magnitudes. These studies have utilized an isolated autoencoder to compress and recover the CSI magnitudes. For phase recovery, the basic principle in [4, 5, 7] is to expend more feedback resources (bandwidth) to encode the significant phases according to the corresponding magnitudes. For example, the authors in [7] designed a deep learning model with a magnitude-dependent polar-phase (MDPP) loss function to compress the significant CSI phases depending on the CSI magnitude.

Presently, these existing magnitude-dependent CSI feedback frameworks tend to train two learning models to encode and recover DL CSI magnitudes and phases, respectively. Intuitively, both CSI magnitudes and phases depend on the RF propagation environment including multipath delays, Doppler spread, bandwidth, and scatter distribution. CSI magnitude and phase encoding and recovery should be jointly instead of individually optimized. In fact, the structural sparsity of CSI phases and their joint distribution with the CSI magnitude are generally unknown and under explored.

In this work, we develop a deep learning based CSI feedback framework which jointly optimizes the magnitude and phase encoding. We propose a new loss function, namely sinusoidal magnitude-adjust phase error (SMAPE), that directly corresponds to the MSE of DL CSI recovery. Furthermore, we take advantage of the circular properties of CSI matrices in DA domain and propose a novel circularly convolutional neural network (C-CNN) layers proved to significantly enhance CSI compression efficiency and recovery performance.

II System Model

Without loss of generality, we consider a single-cell MIMO FDD link in which a gNB with NbN_{b} antennas communicates with a single antenna UE. The OFDM signal spans NfN_{f} DL subcarriers. The DL received signal of the kkth subcarrier is

yDL(k)=𝐑DL(k)​H​𝐰T(k)​xDL(k)+nDL(k),y_{\text{DL}}^{(k)}=\mathbf{h}_{\text{DL}}^{(k)H}\mathbf{w}_{\text{T}}^{(k)}x_{\text{DL}}^{(k)}+n_{\text{DL}}^{(k)}, (1)

where (β‹…)H(\cdot)^{H} denotes conjugate transpose. Here for the kβˆ’k-th subcarrier, 𝐑DL(k)βˆˆβ„‚NbΓ—1\mathbf{h}_{\text{DL}}^{(k)}\in\mathbb{C}^{N_{b}\times 1} denotes the CSI vector and 𝐰T(k)βˆˆβ„‚NbΓ—1\mathbf{w}_{\text{T}}^{(k)}\in\mathbb{C}^{N_{b}\times 1} denotes the corresponding precoding vector111gNB calculates precoding vectors at subcarriers with DL CSI matrix. whereas xDL(k)βˆˆβ„‚x_{\text{DL}}^{(k)}\in\mathbb{C} and nDL(k)βˆˆβ„‚n_{\text{DL}}^{(k)}\in\mathbb{C} denote the DL source signal and additive noise, respectively. With the same antennas, gNb receives UL signal

𝐲UL(k)=𝐑UL(k)​xUL(k)+𝐧UL(k)βˆˆβ„‚NbΓ—1,\mathbf{y}_{\text{UL}}^{(k)}=\mathbf{h}_{\text{UL}}^{(k)}x_{\text{UL}}^{(k)}+\mathbf{n}_{\text{UL}}^{(k)}\in\mathbb{C}^{N_{b}\times 1}, (2)

where subscript UL is used to denote uplink channel 𝐑UL(k)βˆˆβ„‚NbΓ—1\mathbf{h}_{\text{UL}}^{(k)}\in\mathbb{C}^{N_{b}\times 1}, signal xUL(k)x_{\text{UL}}^{(k)}, and noise 𝐧UL(k)βˆˆβ„‚NbΓ—1\mathbf{n}_{\text{UL}}^{(k)}\in\mathbb{C}^{N_{b}\times 1}. DL and UL channel vectors can be written as spatial-frequency CSI (SF-CSI) matrices 𝐇DLSF=[𝐑DL(1),…,𝐑DL(Nf)]Hβˆˆβ„‚NfΓ—Nb{\mathbf{H}}_{\text{DL}}^{\text{SF}}=[\mathbf{h}_{\text{DL}}^{(1)},...,\mathbf{h}_{\text{DL}}^{(N_{f})}]^{H}\in\mathbb{C}^{N_{f}\times N_{b}} and 𝐇ULSF=[𝐑UL(1),…,𝐑UL(Nf)]Hβˆˆβ„‚NfΓ—Nb{\mathbf{H}}_{\text{UL}}^{\text{SF}}=[\mathbf{h}_{\text{UL}}^{(1)},...,\mathbf{h}_{\text{UL}}^{(N_{f})}]^{H}\in\mathbb{C}^{N_{f}\times N_{b}}, respectively. Typically in FDD systems, DL CSI 𝐇DLSF{\mathbf{H}}_{\text{DL}}^{\text{SF}} is estimated by UE for feedback to gNB. However, the number (NfΓ—NbN_{f}\times N_{b}) of unknowns in 𝐇DLSF{\mathbf{H}}_{\text{DL}}^{\text{SF}} occupies substantial feedback spectrum in large or massive MIMO systems. To exploit CSI sparsity to reduce feedback overhead, we apply IDFT 𝐅Dβˆˆβ„‚NfΓ—Nf\mathbf{F}_{D}\in\mathbb{C}^{N_{f}\times N_{f}} and DFT 𝐅Aβˆˆβ„‚NbΓ—Nb\mathbf{F}_{A}\in\mathbb{C}^{N_{b}\times N_{b}} on 𝐇SF\mathbf{H}^{\text{SF}} to generate DA domain CSI matrix

𝐇DA=𝐅D​𝐇SF​𝐅A,\mathbf{H}^{\text{DA}}=\mathbf{F}_{D}\mathbf{H}^{\text{SF}}\mathbf{F}_{A}, (3)

which demonstrates sparsity. Note that 𝐇SF\mathbf{H}^{\text{SF}} denotes either 𝐇ULSF\mathbf{H}^{\text{SF}}_{\text{UL}} or 𝐇ULSF\mathbf{H}^{\text{SF}}_{\text{UL}}. Owing to limited multipath delay spread and limited number of scatters, most elements in 𝐇DA\mathbf{H}^{\text{DA}} are found to be near insignificant, except for the first QfQ_{f} and the last QlQ_{l} rows. Therefore, we shorten the CSI matrix 𝐇DA\mathbf{H}^{\text{DA}} in DA domain to Qt=Qf+QlQ_{t}=Q_{f}+Q_{l} rows that contain sizable non-zero values and utilize 𝐇DL\mathbf{H}_{\text{DL}} and 𝐇UL\mathbf{H}_{\text{UL}} to denote the correspondingly truncated matrices of 𝐇DLDA{\mathbf{H}}_{\text{DL}}^{\text{DA}} and 𝐇ULDA{\mathbf{H}}_{\text{UL}}^{\text{DA}}, respectively. For simplicity, we shall denote 𝐇DL\mathbf{H}_{\text{DL}} as 𝐇\mathbf{H} in the rest of this work except for cases when ambiguity may arise.

Subsequently, to further reduce the feedback overhead, the DL CSI matrix 𝐇\mathbf{H} is encoded at the UE and recovered by the gNB. The recovered DL CSI matrix can be expressed as

𝐇^=fde​(fen​(𝐇)),\widehat{\mathbf{H}}=f_{\text{de}}(f_{\text{en}}(\mathbf{H})), (4)

where fen​(β‹…)f_{\text{en}}(\cdot) and fde​(β‹…)f_{\text{de}}(\cdot) denote encoding/decoding operations.

III Magnitude-aided CSI feedback Framework

Most deep-learning works on CSI compression leverage the success of real-valued deep learning network (DLN) in image processing by separating CSI matrices into real and imaginary parts that are analogous to image files [1, 2, 6], as shown in Fig.Β 1(a). Recent studies [4, 6, 7], however, uncovered the benefit of separately encoding magnitudes and phases of 𝐇\mathbf{H} instead in order to better exploit other correlated CSI magnitudes as auxiliary magnitude information (AMI). Such architecture, illustrated in Fig.Β 1(b), requires substantially lower feedback overhead for the magnitudes of 𝐇\mathbf{H} and allocate more feedback resources for phase feedback of 𝐇\mathbf{H}.

Refer to caption
Figure 1: General network architecture. (a) Conventional CSI feedback framework, (b) conventional magnitude-aided CSI feedback framework, and (c) proposed magnitude-aided CSI feedback framework.

Fig.Β 1(c) illustrates our proposed new DLN framework, consisting of magnitude and phase branches. The gNB further contains a combining network to estimate the full CSI based on results from magnitude and phase decoders. We optimize encoders, decoders, and combining network jointly by minimizing a single loss function for end-to-end learning during offline training. Note that the magnitude branch can be independently optimized. To facilitate convergence [8], the training of the DLN has two stages. In stage-1, the CSI magnitude encoder/decoder branch is pre-trained for magnitude recovery. In stage-2, both the CSI phase branch and the combining network are optimized with the help of the magnitude branch, while the parameters of the magnitude branch are fixed.

III-A DualNet-MP

We now present a new DLN called DualNet-MP. As shown in Fig. 2, DualNet-MP splits each complex CSI matrix into

𝐇=|𝐇|βŠ™ejβ€‹βˆ‘β€‹π‡,\mathbf{H}=|\mathbf{H}|\odot e^{j\measuredangle{\mathbf{H}}},

where βŠ™\odot represents Hadamard product. Denote the (m,n)(m,n)-th entry of 𝐇\mathbf{H} as 𝐇m,n=|𝐇m,n|​ejβ€‹βˆ‘β€‹π‡m,n\mathbf{H}_{m,n}=|\mathbf{H}_{m,n}|e^{j\measuredangle{\mathbf{H}_{m,n}}}. The magnitude matrix |𝐇||\mathbf{H}| and consists of entries |𝐇m,n||\mathbf{H}_{m,n}| and phase matrix ejβ€‹βˆ‘β€‹π‡e^{j\measuredangle{\mathbf{H}}} consists of entries ejβ€‹βˆ‘β€‹π‡m,ne^{j\measuredangle{\mathbf{H}_{m,n}}}.

Similar to [5], we forward the CSI magnitudes to the magnitude encoder network, including four 7Γ—77\times 7 circular convolutional layers with 16, 8, 4, 2 and 1 channels and activation functions. Given the circular characteristic of CSI matrices, we introduce circular convolutional layers to replace the traditional linear ones. Subsequently, a fully connected (FC) layer with ⌈CRMAG​Qt​NbβŒ‰\left\lceil{\text{CR}_{\text{MAG}}Q_{t}N_{b}}\right\rceil elements is included for dimension reduction after reshaping. CRMAG\text{CR}_{\text{MAG}} denotes the magnitude compression ratio. The output of the FC layer is then fed into the quantization module, called the sum-of-sigmoid (SSQ) [5] to generate magnitude codewords for feedback.

At the gNB, a magnitude decoder uses magnitude codewords and available UL CSI magnitudes222 UL CSI is estimated at the gNB and assumed to be perfectly estimated. as AMI to jointly recover DL CSI magnitudes. The DL magnitude branch is first optimized by updating the network parameters ΘMAG\Theta_{\text{MAG}}

arg⁑minΘen,MAG,Θde,MAG{βˆ₯|𝐇^|βˆ’|𝐇|βˆ₯F2}{\mathop{\arg\min}_{\Theta_{\text{en,MAG}},\Theta_{\text{de,MAG}}}\left\{\mathinner{\!\left\lVert|\mathbf{\hat{H}}|-|\mathbf{H}|\right\rVert}^{2}_{\text{F}}\right\}}\\ (5)

to minimize the MSE of recovered MIMO CSI magnitude

|𝐇^|=fde,MAG​(fen,MAG​(|𝐇|,Θen,MAG),Θde,MAG,𝐇UL),{|\mathbf{\hat{H}}|=f_{\text{de,MAG}}(f_{\text{en,MAG}}(|\mathbf{H}|,\Theta_{\text{en,MAG}}),\Theta_{\text{de,MAG}},\mathbf{H}_{\text{UL}}),} (6)

in which Θ\Theta denotes DLN parameters and subscripts en, de, UL, and MAG of the f​(β‹…)f(\cdot) denote the encoder, decoder, UL, and magnitude branch, respectively.

For CSI recovery of MIMO channels, we are only interested in their wrapped phases (i.e., βˆ‘β€‹π‡\measuredangle{\mathbf{H}}). There is 1-to-1 relationship between a phase value Ο•\phi and (cos⁑(Ο•),sign​[sin⁑(Ο•)])\left(\cos(\phi),\mbox{sign}[\sin(\phi)]\right). For these reasons, we propose to form a ”cosine” matrix whose entries are cosines of entries from 𝐇\mathbf{H} denoted by

𝐂𝐨𝐬=cos⁑(βˆ‘β€‹π‡).\mathbf{Cos}=\cos(\measuredangle{\mathbf{H}}). (7a)
Let 𝐀m,n=sign​[sin⁑[βˆ‘β€‹(𝐇m,n)]]\mathbf{A}_{m,n}=\mbox{sign}[\sin[\measuredangle(\mathbf{H}_{m,n})]]. We also form a sign matrix
𝐀=[𝐀m,n].{\mathbf{A}=\left[\mathbf{A}_{m,n}\right].} (7b)

Thus (𝐂𝐨𝐬,𝐀)(\mathbf{Cos},\mathbf{A}) uniquely determines βˆ‘β€‹π‡\measuredangle{\mathbf{H}}.

Since 𝐂𝐨𝐬\mathbf{Cos} matrix is real, we can adopt a phase encoder similar to the magnitude encoder. Let CRPHA\text{CR}_{\text{PHA}} denotes the phase compression ratio. Each 𝐂𝐨𝐬\mathbf{Cos} generates a ⌈CRPHA​Qt​NbβŒ‰\lceil\text{CR}_{\text{PHA}}Q_{t}N_{b}\rceil-element codeword. Our DLN uses tanh activation function in each circular convolutional layer of the phase encoder to capture the underlying features of significant phases associated with large magnitudes. Upon completion of encoder training, the UE processes each CSI 𝐇\mathbf{H}, and feeds back the CSI magnitude codeword fen,MAG​(|𝐇|,Θen,MAG)f_{\text{en,MAG}}(|\mathbf{H}|,\Theta_{\text{en,MAG}}), the phase codeword fen,PHA​(𝐂𝐨𝐬,Θen,PHA)f_{\text{en,PHA}}(\mathbf{Cos},\Theta_{\text{en,PHA}}) and the sign matrix 𝐀\mathbf{A} to gNB.

At the gNB receiver, the phase codeword fen,PHA​(𝐂𝐨𝐬,Θen,PHA)f_{\text{en,PHA}}(\mathbf{Cos},\Theta_{\text{en,PHA}}) and the feedback sign matrix 𝐀\mathbf{A} are sent to the phase decoder with the tanh activation function as the last layer to constrain the entries of DL CSI cosine matrix 𝐂𝐨𝐬^\widehat{\mathbf{Cos}} within [βˆ’1,1][-1,1]. The magnitude codeword and side information are used by the magnitude decoder to obtain an estimated CSI magnitude matrix |𝐇^||\widehat{\mathbf{H}}|. Based on the relationship sin⁑(Ο•)=sign​(sin⁑[Ο•])​1βˆ’cos2⁑(Ο•)\sin(\phi)=\mbox{sign}(\sin[\phi])\sqrt{1-\cos^{2}(\phi)}, we form

𝐒𝐒𝐧^=π€βŠ™(πŸβˆ’π‚π¨π¬^βŠ™π‚π¨π¬^)1/2.\widehat{\mathbf{Sin}}=\mathbf{A}\odot(\mathbf{1}-\widehat{\mathbf{Cos}}\odot\widehat{\mathbf{Cos}})^{1/2}.

Therefore, we can directly generate a preliminary CSI estimate

𝐇^=[|𝐇^|βŠ™π‚π¨π¬^,|𝐇^|βŠ™π’π’π§^]\widehat{\mathbf{H}}=\left[|\widehat{\mathbf{H}}|\odot\widehat{\mathbf{Cos}},|\widehat{\mathbf{H}}|\odot\widehat{\mathbf{Sin}}\right]

from locally available 𝐂𝐨𝐬^\widehat{\mathbf{Cos}}, 𝐀\mathbf{A}, |𝐇|^\widehat{|\mathbf{H}|}. The combining network is trainable and can include two residual blocks containing four circular convolutional layers to refine the DL CSI matrix.

For end-to-end optimization, our training criterion relies on

minimizeΘen,PHA,Θde,PHA,ΘC{βˆ₯𝐇^βˆ’π‡βˆ₯F2},{\mathop{\text{minimize}}_{\Theta_{\text{en,PHA}},\Theta_{\text{de,PHA}},\Theta_{\text{C}}}\left\{\mathinner{\!\left\lVert\mathbf{\hat{H}}-\mathbf{H}\right\rVert}^{2}_{\text{F}}\right\},}\\ (8)

to optimize the parameters Θen,PHA\Theta_{\text{en,PHA}} of phase encoder fen,PHAf_{\text{en,PHA}} and parameters Θde,PHA\Theta_{\text{de,PHA}} of phase decoder fen,PHAf_{\text{en,PHA}} to estimate

𝐂𝐨𝐬^=fde,PHA​(fen,PHA​(𝐂𝐨𝐬,Θen,PHA),𝐀,Θde,PHA).\widehat{\mathbf{Cos}}=f_{\text{de,PHA}}(f_{\text{en,PHA}}(\mathbf{Cos},\Theta_{\text{en,PHA}}),\mathbf{A},\Theta_{\text{de,PHA}}). (9)

Using the same loss function (8), we also train the combining network fCf_{\text{C}} by optimizing parameters ΘC\Theta_{\text{C}} to generate

𝐇^=fC​(|𝐇^|,𝐂𝐨𝐬^,ΘC).\mathbf{\hat{H}}=f_{\text{C}}(|\widehat{\mathbf{H}}|,\widehat{\mathbf{Cos}},\Theta_{\text{C}}). (10)

Since the training of the magnitude learning branch can be decoupled, our framework optimizes the entire architecture by minimizing the overall CSI MSE of (8). It is possible, however, to also partially incorporate the MSE of (8) to further refine the magnitude DLN branch by adopting a slower learning rate.

Refer to caption
Figure 2: Network architecture of DualNet-MAG-PHA.

III-B Loss Function Redesign

Considering the MSE loss function, it may be intuitive to simply rewrite the loss function as follows:

L​o​s​s0=MSECSI​(|𝐇|^,βˆ‘β€‹π‡^)=βˆ₯π‡βˆ’π‡^βˆ₯F2\displaystyle Loss_{\text{0}}=\text{MSE}_{\text{CSI}}(\widehat{|\mathbf{H}|},\measuredangle{\hat{\mathbf{H}}})=\mathinner{\!\left\lVert\mathbf{H}-\mathbf{\hat{H}}\right\rVert}^{2}_{\text{F}} (11)
=βˆ₯|𝐇|βŠ™cos​(βˆ‘β€‹π‡)βˆ’|𝐇^|βŠ™cos​(βˆ‘β€‹π‡^)βˆ₯F2\displaystyle=\mathinner{\!\left\lVert|\mathbf{H}|\odot\text{cos}(\measuredangle{\mathbf{H}})-|\hat{\mathbf{H}}|\odot\text{cos}(\measuredangle{\hat{\mathbf{H}}})\right\rVert}^{2}_{\text{F}}
+βˆ₯|𝐇|βŠ™sin​(βˆ‘β€‹π‡)βˆ’|𝐇^|βŠ™sin​(βˆ‘β€‹π‡^)βˆ₯F2.\displaystyle+\mathinner{\!\left\lVert|\mathbf{H}|\odot\text{sin}(\measuredangle{\mathbf{H}})-|\hat{\mathbf{H}}|\odot\text{sin}(\measuredangle{\hat{\mathbf{H}}})\right\rVert}^{2}_{\text{F}}.

This means that |𝐇||\mathbf{H}| and βˆ‘β€‹π‡\measuredangle{\mathbf{H}} are used as encoder network input variables whereas their estimates are the decoder network output variables. However, the presence of infinitely many and shallow local minima of sinusoidal functions sin⁑(β‹…)\sin(\cdot) and cos⁑(β‹…)\cos(\cdot) often lead to training difficulties [9]. To overcome this problem, the authors in [7] recently proposed a weighted MDPP loss function

MSEMDPP=βˆ₯|βˆ‘β€‹π‡βˆ’βˆ‘β€‹π‡^|βŠ™|𝐇|βˆ₯F2\text{MSE}_{\text{MDPP}}=\mathinner{\!\left\lVert|\measuredangle{\mathbf{H}}-\measuredangle{\mathbf{\hat{H}}}|{\odot}|\mathbf{H}|\right\rVert}^{2}_{\text{F}}\\ (12)

which still uses |𝐇||\mathbf{H}| and βˆ‘β€‹π‡\measuredangle{\mathbf{H}} as input and output variables. where βˆ‘β€‹π‡\measuredangle{\mathbf{H}} and βˆ‘β€‹π‡^\measuredangle{\mathbf{\hat{H}}} denote the true and estimated phases, respectively. By weighting the original phase discrepancy with the true CSI magnitude, this new loss function helps capture the underlying features of the critical phases associated with CSI coefficents with dominant magnitudes. However, the loss function is not equivalent to our final goal for minimizing MSE of DL CSI. We now propose a reparamterization of the same MSE loss function during training. Instead of changing the loss function, we can overcome the training problem of directly parameterization in Eq. (17). Instead, recognizing that only the wrapped phases of βˆ‘β€‹π‡\measuredangle{\mathbf{H}} are of interest, we replace βˆ‘β€‹π‡\measuredangle{\mathbf{H}} with 𝐂𝐨𝐬\mathbf{Cos} and 𝐀\mathbf{A} via the following reparameterization:

MSESMAPE​(|𝐇|^,𝐂𝐨𝐬^,𝐀)\displaystyle\text{MSE}_{\text{SMAPE}}(\widehat{|\mathbf{H}|},\widehat{\mathbf{Cos}},\mathbf{A}) =βˆ₯π‡βˆ’π‡^βˆ₯F2\displaystyle=\mathinner{\!\left\lVert\mathbf{H}-\mathbf{\hat{H}}\right\rVert}^{2}_{\text{F}} (13)
=βˆ₯|𝐇|βŠ™π‚π¨π¬βˆ’|𝐇^|βŠ™π‚π¨π¬^βˆ₯F2\displaystyle=\mathinner{\!\left\lVert|\mathbf{H}|\odot\mathbf{Cos}-|\hat{\mathbf{H}}|\odot\widehat{\mathbf{Cos}}\right\rVert}^{2}_{\text{F}}
+βˆ₯|𝐇|βŠ™π’π’π§βˆ’|𝐇^|βŠ™π’π’π§^βˆ₯F2,\displaystyle+\mathinner{\!\left\lVert|\mathbf{H}|\odot\mathbf{Sin}-|\hat{\mathbf{H}}|\odot\widehat{\mathbf{Sin}}\right\rVert}^{2}_{\text{F}},

where we have used the sign matrix 𝐀\mathbf{A} feedback to generate

𝐒𝐒𝐧\displaystyle{\mathbf{Sin}} =π€βŠ™(πŸβˆ’π‚π¨π¬βŠ™π‚π¨π¬)1/2\displaystyle=\mathbf{A}\odot(\mathbf{1}-{\mathbf{Cos}}\odot{\mathbf{Cos}})^{1/2} (14a)
𝐒𝐒𝐧^\displaystyle\widehat{\mathbf{Sin}} =π€βŠ™(πŸβˆ’π‚π¨π¬^βŠ™π‚π¨π¬^)1/2.\displaystyle=\mathbf{A}\odot(\mathbf{1}-\widehat{\mathbf{Cos}}\odot\widehat{\mathbf{Cos}})^{1/2}. (14b)

This formulation saves about half the bandwidth by sending the sign matrix 𝐀\mathbf{A} without encoding matrix 𝐒𝐒𝐧\mathbf{Sin}.

Moreover, the sparsity of 𝐇\mathbf{H} means that we only need to feed back partial entries of 𝐀\mathbf{A} associated with a swath of entries with dominant magnitudes. If we define a reduction ratio RsR_{\text{s}} to further reduce feedback overhead333Usually, the reconstruction performance can remain approximately the same even if the sign ratio RsR_{\text{s}} is less than 0.250.25 due to the sparsity.. The total phase feedback overhead (in bits) is summarized as follows:

BSMAPE=CRPHA​(KPHA​Qt​Nb+Rs​Qt​Nb)​(bits),B_{\text{SMAPE}}=\text{CR}_{\text{PHA}}(K_{\text{PHA}}Q_{t}N_{b}+R_{\text{s}}Q_{t}N_{b})(\text{bits}), (15)

where KPHAK_{\text{PHA}} denotes the number of encoding bits for each entry of the compressed cosine matrix fen,PHA​(𝐂𝐨𝐬,Θen,PHA)f_{\text{en,PHA}}(\mathbf{Cos},\Theta_{\text{en,PHA}}).

To summarize our training strategy of DualNet-MP, we use Eq. (5) as the loss function during the first training stage. In the second training stage, we used Eqs. (β€’ β€£ IV-A) as the loss function to build an end-to-end learning architecture.

IV Experimental Evaluations

IV-A Experiment Setup

In our experiments444The source codes are online available in https://github.com/max821002/DualNet-MP/, we let the UL and DL bandwidths be 20 MHz and the subcarrier number be NfN_{f} = 1024. We consider both indoor and outdoor cases. We place the gNB with a height of 20 m at the center of a circular cell coverage with a radius of 20 m for indoor and 200 m for outdoor. The number of gNB antennas is Nb=32N_{b}=32 whereas each UE has a single antenna. A half-wavelength inter-antenna spacing is considered. For each trained model, the number of epochs and batch size were set to 1,000 and 200, respectively. We generate two datasets each containing 100,000 random channels for both indoor and outdoor cases from two different channel models. 60,000 and 20,000 random channels are for training and validation. The remaining 20,000 random channels are test data for performance evaluation.

In the first (indoor) dataset, we used the COST 2100 [10] simulator and select the scenario IndoorHall at 5GHz to generate indoor channels at 5.1-GHz UL and 5.3-GHz DL with LOS paths. The antenna and band types are set as MIMO VLA omni and wideband, respectively. As for the outdoor dataset, we utilized QuaDRiGa simulator [11] with the scenario features of 3GPP 38.901 UMi. We considered the UMi scenario at 2 and 2.1 GHz of UL and DL bands, respectively, without LOS paths. The number of cluster paths was set as 13. The antenna type is set to omni. By DFT/IDFT operations and truncating described in Eqs. (3) and (4) of the revised manuscript, the resulting 32Γ—3232\times 32 UL/DL CSI matrices in DA domain are obtained and be fed as the encoder input after proper processes.

The performance metric is the normalized MSE

NMSE=1Dβ€‹βˆ‘d=1Dβˆ₯𝐇^DL,dSFβˆ’π‡DL,dSFβˆ₯F2/βˆ₯𝐇DL,dSFβˆ₯F2,\text{NMSE}=\frac{1}{D}\sum^{D}_{d=1}\mathinner{\!\left\lVert\widehat{\mathbf{H}}_{\text{DL},d}^{\text{SF}}-{\mathbf{H}}_{\text{DL},d}^{\text{SF}}\right\rVert}^{2}_{\text{F}}/\mathinner{\!\left\lVert{\mathbf{H}}_{\text{DL},d}^{\text{SF}}\right\rVert}^{2}_{\text{F}}, (16)

where the number DD and subscript dd denote the total number and index of channel realizations, respectively. Instead of evaluating the estimated DL CSI matrix 𝐇^DL\widehat{\mathbf{H}}_{\text{DL}}, we evaluate the estimated SFCSI matrix 𝐇^DLSF\widehat{\mathbf{H}}_{\text{DL}}^{\text{SF}} that can be obtained by reversing the Fourier processing and padding zero matrix. Note that this NMSE includes both the errors caused by truncation at the encoder and the overall recovery error. Thus, it is practically more meaningful.

In the following section, we evaluate the performance of CSI recovery by adopting the proposed optimization method and encoder/decoder architecture. Thus, we trained DualNet-MP with the same core network design for magnitude recovery. However, we test different methods to reconstruct the CSI phases for two phase compression ratios of CRPHA=1/8\text{CR}_{\text{PHA}}=1/8 and 1/161/16 555All alternate approaches consume 1.21.2 and 0.6250.625 bits/phase entry:

  • β€’

    SMAPE: the network architecture follows DualNet-MP as illustrated in Fig. 2 of the revised manuscript. The sign ratio RsR_{\text{s}} varies between [0.25,0.125][0.25,0.125] and we use KPHA=8K_{\text{PHA}}=8 bits for both CRPHA={1/8,1/16}\text{CR}_{\text{PHA}}=\{1/8,1/16\}. The loss function for phase reconstruction is given as follows:

    MSESMAPE​(|𝐇|^,𝐂𝐨𝐬^,𝐀)\displaystyle\text{MSE}_{\text{SMAPE}}(\widehat{|\mathbf{H}|},\widehat{\mathbf{Cos}},\mathbf{A}) =βˆ₯π‡βˆ’π‡^βˆ₯F2\displaystyle=\mathinner{\!\left\lVert\mathbf{H}-\mathbf{\hat{H}}\right\rVert}^{2}_{\text{F}}
    =βˆ₯|𝐇|βŠ™π‚π¨π¬βˆ’|𝐇^|βŠ™π‚π¨π¬^βˆ₯F2\displaystyle=\mathinner{\!\left\lVert|\mathbf{H}|\odot\mathbf{Cos}-|\hat{\mathbf{H}}|\odot\widehat{\mathbf{Cos}}\right\rVert}^{2}_{\text{F}}
    +βˆ₯|𝐇|βŠ™π’π’π§βˆ’|𝐇^|βŠ™π’π’π§^βˆ₯F2.\displaystyle+\mathinner{\!\left\lVert|\mathbf{H}|\odot\mathbf{Sin}-|\hat{\mathbf{H}}|\odot\widehat{\mathbf{Sin}}\right\rVert}^{2}_{\text{F}}.
  • β€’

    MDPQ [4]: the design assigns more encoding bits to encode the significant phases according the corresponding magnitudes. It assigns [0,0,0,3,7][0,0,0,3,7] and [0,0,0,0,5][0,0,0,0,5] bits for CRPHA=[1/8,1/16]\text{CR}_{\text{PHA}}=[1/8,1/16], respectively, to encode the CSI phases corresponding to [0,0.5,0.7,0.8,0.9][0,0.5,0.7,0.8,0.9] of the cumulative distribution of CSI magnitude.

  • β€’

    MSE0: instead of cosine, CSI phases are fed directly to the phase encoder. Both cosine and sine functions are appended as the final layer of the phase decoder to transform the decoded phase into cosine and sine, respectively, to recover the real and imaginary parts of CSI. We set KPHAK_{\text{PHA}} to 88 bits. The loss function for phase reconstruction is given as follows:

    Loss0=MSECSI​(|𝐇|^,βˆ‘β€‹π‡^)=βˆ₯π‡βˆ’π‡^βˆ₯F2\displaystyle\mbox{Loss}_{\text{0}}=\text{MSE}_{\text{CSI}}(\widehat{|\mathbf{H}|},\measuredangle{\hat{\mathbf{H}}})=\mathinner{\!\left\lVert\mathbf{H}-\mathbf{\hat{H}}\right\rVert}^{2}_{\text{F}} (17)
    =βˆ₯|𝐇|βŠ™cos​(βˆ‘β€‹π‡)βˆ’|𝐇^|βŠ™cos​(βˆ‘β€‹π‡^)βˆ₯F2\displaystyle=\mathinner{\!\left\lVert|\mathbf{H}|\odot\text{cos}(\measuredangle{\mathbf{H}})-|\hat{\mathbf{H}}|\odot\text{cos}(\measuredangle{\hat{\mathbf{H}}})\right\rVert}^{2}_{\text{F}}
    +βˆ₯|𝐇|βŠ™sin​(βˆ‘β€‹π‡)βˆ’|𝐇^|βŠ™sin​(βˆ‘β€‹π‡^)βˆ₯F2.\displaystyle+\mathinner{\!\left\lVert|\mathbf{H}|\odot\text{sin}(\measuredangle{\mathbf{H}})-|\hat{\mathbf{H}}|\odot\text{sin}(\measuredangle{\hat{\mathbf{H}}})\right\rVert}^{2}_{\text{F}}.
  • β€’

    MDPP [7]: we reuse the loss function Eq.(18) with the same network architecture. We set KPHAK_{\text{PHA}} to 1010 bits.

    MSEMDPP=βˆ₯|βˆ‘β€‹π‡βˆ’βˆ‘β€‹π‡^|βŠ™|𝐇|βˆ₯F2\text{MSE}_{\text{MDPP}}=\mathinner{\!\left\lVert|\measuredangle{\mathbf{H}}-\measuredangle{\mathbf{\hat{H}}}|{\odot}|\mathbf{H}|\right\rVert}^{2}_{\text{F}}\\ (18)

IV-B Different Phase Compression Designs

To demonstrate the superiority of the proposed SMAPE loss function, we applied different phase reconstruction approaches to DualNet-MP for different phase compression ratios CRPHA\text{CR}_{\text{PHA}}. Figs. 3 (a) and (b) show the NMSE performance of different approaches under indoor and outdoor scenarios, respectively, at different compression ratios. As expected, DaulNet-MP encounters training difficulties when using the simple loss function L​o​s​s0Loss_{0}. By adopting MDPP loss functions, DualNet-MP performs much better than the simple loss function L​o​s​s0Loss_{0}. Although DualNet-MP appears to be better when using MDPQ instead of MDPP, encoding bit-assignment require careful tuning to achieve a satisfactory result. Finally, DualNet-MP based on the proposed SMAPE loss function achieves 4-dB performance improvement in terms of NMSE reduction for CRPHA\text{CR}_{\text{PHA}}=1/8 at outdoor and 7-dB improvement for CRPHA\text{CR}_{\text{PHA}}=1/8 at outdoor.

Refer to caption
Figure 3: NMSE performance for different loss functions in (a) indoor and (b) outdoor scenarios.

IV-C Different Core Layer Designs

To investigate the appropriate core layer designs of DualNet-MP in order to efficiently extract the underlying features of CSI phases, we provide a performance evaluation using FC, linear convolutional, and circular convolutional layers, respectively, for the core network. Denoted respectively as DNN, CNN and C-CNN, these networks adopted SSQ [5] and binary-level quantization (BLQ) as the quantization module at the encoder. Denote that the DNN design follows the recent work [7]. We consider the phase compression ratio of CRPHA=1/8\text{CR}_{\text{PHA}}=1/8. For SSQ, we assign KPHA=8K_{\text{PHA}}=8 bits for each codeword. That is, there are CRPHA​Qt​Nb=128\text{CR}_{\text{PHA}}Q_{t}N_{b}=128 8-bit codewords sent to the gNB. In contrast, there are KPHA​CRPHA​Qt​Nb=1024K_{\text{PHA}}\text{CR}_{\text{PHA}}Q_{t}N_{b}=1024 1-bit codewords when applying BLQ.

Figs. 4.(a) and (b) show the NMSE performance for the considered core layer designs. For both indoor and outdoor scenarios, DualNet-MP demonstrates superiority when adopting SSQ and C-CNN, which can be attributed to two possible reasons. Firstly, unlike BLQ, SSQ is differentiable such that it is easier to train. Secondly, there are many structural and circular features of CSI phases in the angle-delay domain that can be extracted better with the proposed structural changes.

In terms of storage and complexity of the proposed architecture, we note that C-CNN with only 826K parameters is considerably simpler than DNN of 11.6M parameters, using comparable floating point operations. Thus, the newly proposed DualNet-MP architecture combining SSQ and C-CNN provides both performance gain and cost benefits.

Refer to caption
Figure 4: NMSE performance for different core layer designs in (a) indoor and (b) outdoor scenarios.

IV-D Accessment of FLOPs and parameter numbers

We evaluated the numbers of floating point operations (FLOPs) and parameters of DualNet-MP with different phase encoding approaches and core layer designs in Tables I and II, respectively. The comparison shows similarly complexity in terms of FLOPS. Since the complexity and storage issues are both crucial in the design of encoders deployed at UEs, we provide Table III to compare the encoder complexity between the proposed framework and two state-of-the-art alternatives.

TABLE I: FLOP and parameter numbers of DualNet-MP with different phase encoding schemes.
MDPQ MSE0\text{MSE}_{\text{0}}
CRPHA\text{CR}_{\text{PHA}} 1/8,1/16 1/8 1/16
FLOPs 39.5M 77.2M 76.9M
Parameters 282.4K 825.9K 694.7K
MDPP SMAPE
CRPHA\text{CR}_{\text{PHA}} 1/8 1/16 1/8 1/16
FLOPs 77.2M 79.6M 79.2M 78M
Parameters 825.9K 694.7K 826K 695.3K
TABLE II: FLOP and parameter numbers of DualNet-MP with different core layer designs when CRPHA=1/8\text{CR}_{\text{PHA}}=1/8.
DNN C-CNN CNN
Quantization Type BLQ SSQ BLQ SSQ SSQ
FLOPs 118M 81.7M 82.8M 79.2M 79.2M
Parameters 34.2M 11.8M 2.6M 826K 826K
TABLE III: FLOP and parameter numbers at the encoders of DualNet-MP and state-of-the-art alternatives when CR=1/8\text{CR}=1/8.
Method Parameters FLOPs
CsiNet+ 525K 1.86M
CoCsiNet 14.9M 39.1M
DualNet-MP 280K 38M

Since the FC layers are responsible for most of the parameters in many models, CoCsiNet requires the most parameters as it uses multiple FC layers at the encoder. Compared with CsiNet+, the proposed framework requires about half storage by separately encoding CSI phase and magnitude. To be specific, the FC layers require (2048Γ—256+256)(2048\times 256+256) and 2​(1024Γ—128+128)2(1024\times 128+128) parameters at the encoders of CsiNet+ and DualNet-MP, respectively.

Often, convolutional layers contribute most FLOPs in a model due to the considerable number of two-dimensional convolution operations. However, DualNet-MP uses more convolutional layers at the encoder but still requires fewer FLOPs than CoCsiNet. Thus the proposed new architecture exhibits improvement in complexity reduction.

V Conclusions

This work presents a new deep-learning (DL) framework for large scale CSI estimation that leverages feedback compression and auxiliary CSI magnitude information in FDD systems. Utilizing vital domain knowledge in DL for CSI estimation to overcome known training issues, our new framework provides a novel loss function to enable efficient end-to-end learning and improves CSI recovery performance. We further exploit the circular characteristics of the underlying CSI in DA domain to propose an innovative circular convolution neural network (C-CNN). Our test results reveal significant improvement of overall CSI recovery performance for both indoor and outdoor scenarios and complexity reduction in comparison with a number of published alternative DL compression designs for MIMO CSI feedback.

References

  • [1] C.Β Wen, W.Β Shih, and S.Β Jin, β€œDeep Learning for Massive MIMO CSI Feedback,” IEEE Wirel. Commun. Lett., vol.Β 7, no.Β 5, pp. 748–751, 2018.
  • [2] J.Β Guo et al., β€œConvolutional Neural Network-Based Multiple-Rate Compressive Sensing for Massive MIMO CSI Feedback: Design, Simulation, and Analysis,” IEEE Trans. Wirel. Commun., vol.Β 19, no.Β 4, pp. 2827–2840, 2020.
  • [3] Z.Β Qin et al., β€œDeep Learning in Physical Layer Communications,” IEEE Wirel. Commun., vol.Β 26, no.Β 2, pp. 93–99, 2019.
  • [4] Z.Β Liu, L.Β Zhang, and Z.Β Ding, β€œExploiting Bi-Directional Channel Reciprocity in Deep Learning for Low Rate Massive MIMO CSI Feedback,” IEEE Wirel. Commun. Lett., vol.Β 8, no.Β 3, pp. 889–892, 2019.
  • [5] β€”β€”, β€œAn Efficient Deep Learning Framework for Low Rate Massive MIMO CSI Reporting,” IEEE Trans. Commun., vol.Β 68, no.Β 8, pp. 4761–4772, 2020.
  • [6] Z.Β Liu, M.Β Rosario, and Z.Β Ding, β€œA Markovian Model-Driven Deep Learning Framework for Massive MIMO CSI Feedback,” arXiv preprint arXiv:2009.09468, 2020.
  • [7] J.Β Guo et al., β€œDL-based CSI Feedback and Cooperative Recovery in Massive MIMO,” arXiv preprint arXiv:2003.03303, 2020.
  • [8] Y.Β Lin et al., β€œLearning-Based Phase Compression and Quantization for Massive MIMO CSI Feedback with Magnitude-Aided Information,” arXiv preprint arXiv:2103.00432, 2021.
  • [9] G.Β Parascandolo, H.Β Huttunen, and T.Β Virtanen, β€œTaming the Waves: Sine as Activation Function in Deep Neural Networks,” 2017.
  • [10] L.Β Liu et al., β€œThe COST 2100 MIMO Channel Model,” IEEE Wirel. Commun., vol.Β 19, no.Β 6, pp. 92–99, 2012.
  • [11] S.Β Jaeckel et al., β€œQuaDRiGa: A 3-D Multi-Cell Channel Model with Time Evolution for Enabling Virtual Field Trials,” IEEE Trans. Antennas and Propag., vol.Β 62, no.Β 6, pp. 3242–3256, 2014.