This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Joint Activity Detection, Channel Estimation, and Data Decoding for Grant-free Massive Random Access

Xinyu Bian,  Yuyi Mao,  and Jun Zhang,  Manuscript received 13 April 2022; revised 28 December 2022; accepted 3 February 2023. This work was supported in part by the General Research Fund (Project No. 15207220) from the Hong Kong Research Grants Council and in part by a start-up fund of the Hong Kong Polytechnic University (Project ID P0038174). This paper was presented in part at the 2021 IEEE International Workshop on Signal Processing Advances in Wireless Communications (SPAWC) [1]. (Corresponding author: Yuyi Mao.) X. Bian and J. Zhang are with the Department of Electronic and Computer Engineering, the Hong Kong University of Science and Technology, Hong Kong (E-mail: [email protected], [email protected]). Y. Mao is with the Department of Electronic and Information Engineering, the Hong Kong Polytechnic University, Hong Kong (E-mail: [email protected]).
Abstract

In the massive machine-type communication (mMTC) scenario, a large number of devices with sporadic traffic need to access the network on limited radio resources. Recently, grant-free random access has emerged as a promising mechanism for this challenging scenario, but its potential has not been fully unleashed. In particular, the available auxiliary information has not been fully exploited, including the common sparsity pattern in the received pilot and data signal, as well as the channel decoding information. This paper develops advanced receivers in a holistic manner to improve the massive access performance by jointly designing activity detection, channel estimation, and data decoding. To tackle the algorithmic and computational challenges, a turbo structure is adopted at the joint receiver. For performance enhancement, all the received symbols are utilized to jointly estimate the channel state, user activity, and soft data symbols, which effectively exploits the common sparsity pattern. Meanwhile, the extrinsic information from the channel decoder will assist the joint channel estimation and data detection. To reduce the complexity, a low-cost side information (SI)-aided receiver is also proposed, where the channel decoder provides side information to update the estimates on whether a user is active or not. Simulation results show that the turbo receiver is able to reduce the activity detection, channel estimation, and data decoding errors effectively, supporting twice as many active users compared with a separate design that disregards the common sparsity. In addition, the SI-aided receiver notably outperforms the conventional methods with a relatively low complexity.

Index Terms:
Grant-free massive random access, massive machine-type communication (mMTC), user activity detection, channel estimation, channel coding, approximate message passing (AMP), turbo receiver.

I Introduction

The upsurge of numerous Internet of Things (IoT) applications, e.g., autonomous vehicles, intelligent robots, smart city, and industry 4.0, is boosting a rapid paradigm shift of wireless communications from connecting people to connecting things [2]. It is estimated by Cisco that the share of machine-to-machine (M2M) communications will increase from 33% in 2018 to 50% in 2023 [3], leading to an unprecedented requirement of ubiquitous and scalable connectivity. In order to fulfill such an urgent need, massive machine-type communications (mMTC) has been identified as one of the key application scenarios of the fifth-generation (5G) cellular networks by the third-generation partnership project (3GPP) [4]. The most distinctive feature of mMTC is that a huge amount of devices are simultaneously connected to a base station (BS), while only a small proportion of them are active for transmitting a short data packet at each time [5]. This brings forth significant challenges that cannot be catered with existing wireless technologies. In particular, uplink access in cellular networks is traditionally controlled by grant-based random access (RA) mechanisms, where each user first initiates a RA procedure by transmitting a scheduling request (SR) to its serving BS, and it cannot start data transmission until the request is granted [6]. Nevertheless, due to the limited preamble sequences for SR, grant-based RA suffers from potential collisions of the connection requests when two or more users pick the same sequence, especially with massive concurrent requests [7]. Although complementary techniques such as power ramping, back-off, and access class barring [8], can be utilized to relieve access collision, long access latency and significant signalling overhead will be incurred, which are unfavorable for mMTC [9, 10]. Therefore, more efficient RA mechanisms are needed to meet the stringent latency and reliability requirements of mMTC.

Grant-free RA, which allows users to directly send messages without waiting for access permissions from the BS [11], is widely acknowledged as a promising alternative for massive RA. In contrast to grant-based RA where each user selects a random pilot sequence at each time slot, grant-free RA assigns a fixed and unique pilot sequence to each user to enable contention-free uplink access. Nevertheless, the BS does not have knowledge of the set of transmitting (i.e., active) users in grant-free RA, making it arduous to perform accurate channel estimation and data reception. Consequently, detecting the set of transmitting users (i.e., user activity detection) at the BS, becomes a new and critical task for grant-free massive RA [12].

I-A Related Works and Motivations

The popular pilot-based grant-free RA protocol is considered in this paper, where the active users transmit their unique pilot sequences using dedicated radio resources followed by data symbols, and the BS identifies the set of active users and decodes their data. Attributed to the massive number of devices and the limited resources for pilot transmission, it is infeasible to assign orthogonal pilot sequences to all the devices, rendering conventional collision avoidance mechanisms not readily applicable. Fortunately, user activity detection in mMTC turns out to be a compressive sensing (CS) problem [13] thanks to the sporadic data traffic pattern, for which, many efficient algorithms are available [14, 15, 16, 17]. For instance, by formulating user activity detection as a maximum likelihood (ML) estimation problem, low-complexity algorithms based on sample covariance matrices of the received pilot signal were proposed for massive multi-input multi-output (MIMO) systems in [18]. The user activity detection accuracies in massive and cooperative MIMO systems were analyzed in [19] with the approximate message passing (AMP) algorithm. Nevertheless, early studies on grant-free massive RA only focused on user activity detection but neglected the data reception performance, which motivates the investigations on multi-user detection (MUD) for massive connectivity [20, 21]. In particular, based on the orthogonal matching pursuit (OMP), a support detection algorithm was proposed for joint activity and data detection in grant-free non-orthogonal multiple access (NOMA) systems in [20]. A similar problem was tackled by fusing the expectation maximization (EM) algorithm with AMP in [21], which leverages prior information of the transmitted data symbols in addition to the sparse user activity pattern. However, these works assume full channel state information (CSI) available at the BS, which is idealized and impractical for grant-free massive RA.

To bridge this gap, joint activity detection and channel estimation (JADCE) becomes an emerging theme for grant-free massive RA. It was shown in [22] that JADCE can be formulated as a single measurement vector (SMV) and multiple measurement vector (MMV) CS problem with single- and multi-antenna BS, respectively, and both problems can be efficiently solved by AMP. The false alarm and missed detection probabilities were also characterized in [22]. Interestingly, a subsequent investigation [23] revealed that the activity detection error can be made arbitrarily small with sufficient BS antennas. Besides, characteristics of the wireless channels have been utilized together with the sparse user activity pattern to promote the accuracy of JADCE [24, 25]. Moreover, in reliance on a spatial and angular domain channel model respectively, user activity detection and channel estimation algorithms were developed for orthogonal frequency division multiplexing (OFDM) massive MIMO systems in [24] to achieve considerably improved access performance. Beyond those, low-complexity JADCE algorithms were proposed based on dimension reduction [26] and deep learning [27].

While the aforementioned JADCE algorithms exploit the sparsity pattern in the received pilot signal, a common sparsity pattern inherently embedded in both the received pilot and data signal could be further utilized to enhance the performance. The benefits of exploiting such a common sparsity pattern was first witnessed in [28] for massive RA systems with a single-antenna BS, where a joint activity detection, channel estimation and MUD algorithm was proposed under the framework of MMV CS. An extension for multi-antenna BSs was conducted in [29] via the bilinear generalized AMP (BiG-AMP) algorithm [17]. These preliminary attempts, however, were limited to uncoded transmissions and failed to take advantages of channel coding in modern digital communication systems. Specifically, the error detection mechanism of channel codes can be utilized to determine a subset of active users with high channel quality [30]. Besides, the soft decoding results, which carry posterior information of the transmitted data symbols, are valuable for improving the accuracy of activity detection, channel estimation and MUD. However, the joint detection/decoding problem is computationally infeasible for channel codes even with reasonable block lengths [31]. Thus, it necessitates a holistic investigation of how channel decoders can be effectively integrated with other critical components in a massive RA receiver, including user activity detection, channel estimation, and MUD, which will be pursued in this paper.

I-B Contributions

In this paper, we endeavor to push the performance limit of uplink receivers for grant-free massive RA by leveraging both the common sparsity pattern and channel decoding results. Our main contributions lie in developing effective methods to tackle the algorithmic and computational challenges of the joint design of activity detection, channel estimation, and data decoding, as summarized below.

  • We propose a turbo receiver for joint activity detection, channel estimation, and data decoding, which iterates between a joint estimator and a channel decoder. In order to exploit the common sparsity pattern in the received pilot and data signal, the joint estimator for joint activity detection, channel estimation, and data symbol detection, is developed by solving a bilinear inference problem based on the BiG-AMP algorithm. To boost its performance, we further leverage the posterior log-likelihood ratios (LLRs) of the data bits from the channel decoder to derive the extrinsic information, which serves as the updated estimates of the user activity as well as data symbol distribution and are used as prior information for the next turbo iteration.

  • Albeit the turbo receiver is effective in exploiting the common sparsity pattern, the BiG-AMP-based joint estimator incurs significant computation overhead. To facilitate fast execution while retaining the performance gain of the iterative receiver, we develop a side information (SI)-aided receiver that executes a sequential estimator and a channel decoder alternatively. The sequential estimator is developed for JADCE based on the AMP algorithm, which processes only the received pilot signal, leaving data symbol detection to a minimum mean square error (MMSE) equalizer. To effectively leverage the common sparsity and channel decoding results, the estimates on whether a user is active or not are used as SI for the sequential estimator, which are derived according to the parity check results and posterior LLRs.

  • Simulation results show that the turbo receiver significantly reduces the activity detection, channel estimation, and data decoding errors compared with the baseline schemes. Remarkably, in the simulated setting, assuming the block error rate (BLER) requirement is 10310^{-3}, the turbo receiver is able to support 40 active users while the separate design can only support 20. Meanwhile, the SI-aided receiver saves more than 60% of the execution time compared with the turbo receiver, while maintaining a noticeable performance improvement compared against a data-assisted design that only leverages the common sparsity pattern.

I-C Organization

The rest of this paper is organized as follows. We introduce the system model in Section II. A turbo receiver for joint activity detection, channel estimation, and data decoding is developed in Section III. In Section IV, we propose a low-complexity SI-aided receiver. Simulation results are presented in Section V, and Section VI concludes this paper.

I-D Notations

We use lower-case letters, bold-face lower-case letters, bold-face upper-case letters, and math calligraphy letters to denote scalars, vectors, matrices, and sets, respectively. The entry in the ii-th row and jj-th column of matrix 𝐌\mathbf{M} is denoted as mijm_{ij}, and the matrix transpose, complex conjugate, and conjugate transpose operators are denoted as ()T(\cdot)^{\mathrm{T}}, ()(\cdot)^{\mathrm{*}}, and ()H(\cdot)^{\mathrm{H}}, respectively. Besides, 𝐌\i,j\mathbf{M}_{\backslash i,j} represents all the elements in matrix 𝐌[mij]\mathbf{M}\triangleq[m_{ij}] except mijm_{ij}. In addition, exp()\exp\left(\cdot\right) denotes the exponential function, δ()\delta\left(\cdot\right) denotes the Dirac delta function, \lfloor\cdot\rfloor denotes the floor function, and 𝒞𝒩(x;μ,v)\mathcal{CN}(x;\mu,v) denotes the probability density function (PDF) of a complex Gaussian random variable xx with mean μ\mu and variance vv.

II System Model

We consider an uplink cellular system as shown in Fig. 1, where a BS with MM antennas serves NN single-antenna users. The users are assumed to have short data packets to transmit occasionally, and at each time instant, KK (KNK\leq N) among the NN users become active for transmission. Denote un{0,1}u_{n}\in\{0,1\} as the user activity indicator, where un=1u_{n}=1 means user nn is active and un=0u_{n}=0 if it is inactive. The sets of system users and active users are represented by 𝒩{1,,N}\mathcal{N}\triangleq\{1,\cdots,N\} and Ξ{n𝒩|un=1}\Xi\triangleq\left\{n\in\mathcal{N}|u_{n}=1\right\}, respectively, and the set of BS antennas is denoted as {1,,M}\mathcal{M}\triangleq\{1,\cdots,M\}. Besides, the number of BS antennas is assumed to be no less than the number of active users, i.e., MKM\geq K, to avoid the system from being overloaded [32].

Each transmission block contains TT symbol intervals and we assume quasi-static block fading channels, i.e., the channel state remains unchanged within a transmission block, but varies independently across multiple blocks. The uplink channel vector from user nn to the BS is modeled as 𝐟n=βn𝜶n\mathbf{f}_{n}=\sqrt{\beta_{n}}\bm{\alpha}_{n}, where 𝜶n\bm{\alpha}_{n} and βn\beta_{n} denote the small-scale and large-scale fading coefficients, respectively. We focus on Rayleigh fading channels, i.e., 𝜶n𝒞𝒩(𝟎,𝐈M)\bm{\alpha}_{n}\sim\mathcal{CN}(\bm{0},\mathbf{I}_{M})111With slight abuse of notation, 𝒞𝒩(𝝁,𝚺)\mathcal{CN}\left(\bm{\mu},\bm{\Sigma}\right) also denotes the complex Gaussian distribution with mean 𝝁\bm{\mu} and covariance matrix 𝚺\bm{\Sigma}., and assume that the users are static with {βn}\{\beta_{n}\}’s known by the BS [22].

A grant-free RA scheme where each transmission block is divided into two phases, as shown in Fig. 1, is adopted for uplink transmission. Specifically, in the first phase, LL symbols, denoted as 𝐓p\mathbf{T}_{p}, are reserved for pilot transmission, which are essential for user activity detection and channel estimation at the BS; Whereas the remaining LdTLL_{d}\triangleq T-L symbols, denoted as 𝐓d\mathbf{T}_{d}, are used for payload delivery in the second phase. It is important to note that although orthogonal pilot sequences are most beneficial for accurate channel estimation, it is strictly prohibitive in mMTC since the number of users can be much larger than the pilot length, i.e., NLN\gg L. As a result, we assign the users with a set of non-orthogonal and unique pilot sequences {𝐱pn}\{\mathbf{x}_{pn}\}’s by sampling a complex Gaussian distribution, i.e., 𝐱pn[xn1,,xnL]\mathbf{x}_{pn}\triangleq\left[x_{n1},\cdots,x_{nL}\right] with xnl𝒞𝒩(0,1)x_{nl}\sim\mathcal{C}\mathcal{N}\left(0,1\right), which achieves asymptotic orthogonality when LL is sufficiently large. Define 𝐗p[𝐱p1,,𝐱pN]T\mathbf{X}_{p}\triangleq\left[\mathbf{x}_{p1},\cdots,\mathbf{x}_{pN}\right]^{\mathrm{T}} as the collection of pilot sequences.

Refer to caption
Figure 1: System model and the transmission block structure.

In each transmission block, NbN_{b} payload bits, denoted as 𝒃n[bn1,,bnNb],nΞ\bm{b}_{n}\triangleq\left[b_{n1},\cdots,b_{nN_{b}}\right],n\in\Xi, need to be transmitted for each active user, which are encoded for error detection and correction. Following contemporary communication standards such as the long term evolution (LTE) [33] and 5G new radio (NR) [34], cyclic redundancy check (CRC) bits are generated and attached to the payload bits to form a code block. We represent the CRC generation and attachment procedures by function Υ\Upsilon:{0,1}Nb{0,1}Nd:\{0,1\}^{N_{b}}\rightarrow\{0,1\}^{N_{d}}, where NdN_{d} denotes the size of a code block. Thus, the code blocks of the active users can be expressed as follows:

𝒅n[dn1,,dnNd]=Υ(𝒃n),nΞ.\displaystyle\bm{d}_{n}\triangleq\left[d_{n1},\cdots,d_{nN_{d}}\right]=\Upsilon(\bm{b}_{n}),n\in\Xi. (1)

Each code block is then encoded by a channel encoder, which is abstracted as function Φ\Phi:{0,1}Nd{0,1}Nc:\{0,1\}^{N_{d}}\rightarrow\{0,1\}^{N_{c}}, and the coded bits can be represented as follows:

𝒄n[cn1,,cnNc]=Φ(𝒅n),nΞ.\displaystyle\bm{c}_{n}\triangleq\left[c_{n1},\cdots,c_{nN_{c}}\right]=\Phi(\bm{d}_{n}),n\in\Xi. (2)

Note that NcN_{c} is the number of coded bits and the code rate ϕ\phi is defined as the ratio between NdN_{d} and NcN_{c}, i.e., ϕNdNc\phi\triangleq\frac{N_{d}}{N_{c}}.

The coded bits are modulated to a set of constellation points 𝒳\mathcal{X} with normalized average power via an invertible mapping μ\mu:{0,1}log2|𝒳|𝒳:\{0,1\}^{\log_{2}|\mathcal{X}|}\rightarrow\mathcal{X}, i.e., for an arbitrary bit sequence with length log2|𝒳|\log_{2}|\mathcal{X}|, μ([c1,,clog2|𝒳|])=s\mu\left([c_{1},\cdots,c_{\log_{2}|\mathcal{X}|}]\right)=s if and only if μ1(s)=[c1,,clog2|𝒳|]\mu^{-1}\left(s\right)=\left[c_{1},\cdots,c_{\log_{2}|\mathcal{X}|}\right], where s𝒳s\in\mathcal{X} is a constellation point. The modulated symbols for the active users are denoted as follows:

𝐱dn[xn(L+1),,xnT],nΞ.\displaystyle\mathbf{x}_{dn}\triangleq\left[x_{n(L+1)},\cdots,x_{nT}\right],n\in\Xi. (3)

We assume Nc=Ldlog2|𝒳|N_{c}=L_{d}\log_{2}|\mathcal{X}| for simplicity and assign zero vectors to 𝐱dn\mathbf{x}_{dn} for the set of inactive users. Let 𝐗d[𝐱d1,,𝐱dN]T\mathbf{X}_{d}\triangleq\left[\mathbf{x}_{d1},\cdots,\mathbf{x}_{dN}\right]^{\mathrm{T}} denote the transmitted data symbols from all the users. As a result, the received signal of the transmission block 𝐘~M×T\tilde{\mathbf{Y}}\in\mathbb{C}^{M\times T} at the BS can be expressed as follows:

𝐘~[𝐘~p,𝐘~d]=γ𝐇[𝐗p,𝐗d]𝐗+[𝐍~p,𝐍~d]𝐍~,\displaystyle\tilde{\mathbf{Y}}\triangleq\left[\tilde{\mathbf{Y}}_{p},\tilde{\mathbf{Y}}_{d}\right]=\sqrt{\gamma}\mathbf{H}\underbrace{\left[\mathbf{X}_{p},\mathbf{X}_{d}\right]}_{\triangleq\mathbf{X}}+\underbrace{\left[\tilde{\mathbf{N}}_{p},\tilde{\mathbf{N}}_{d}\right]}_{\triangleq\tilde{\mathbf{N}}}, (4)

where γ\gamma is the uplink transmit power, 𝐘~pM×L\tilde{\mathbf{Y}}_{p}\in\mathbb{C}^{M\times L} and 𝐘~dM×Ld\tilde{\mathbf{Y}}_{d}\in\mathbb{C}^{M\times L_{d}} are the received pilot and data signal, respectively, and 𝐇[𝐡1,,𝐡N]M×N\mathbf{H}\triangleq\left[\mathbf{h}_{1},...,\mathbf{h}_{N}\right]\in\mathbb{C}^{M\times N} with 𝐡nun𝐟n\mathbf{h}_{n}\triangleq u_{n}\mathbf{f}_{n} denotes the effective channel coefficient matrix. Besides, 𝐍~=[𝐧~1,,𝐧~T]M×T\tilde{\mathbf{N}}=\left[\tilde{\mathbf{n}}_{1},...,\tilde{\mathbf{n}}_{T}\right]\in\mathbb{C}^{M\times T} is the additive white Gaussian noise (AWGN) with zero mean and variance σ2\sigma^{2} for each element, and 𝐍~pM×L\tilde{\mathbf{N}}_{p}\in\mathbb{C}^{M\times L} and 𝐍~dM×Ld\tilde{\mathbf{N}}_{d}\in\mathbb{C}^{M\times L_{d}} are the noise of the received pilot and data signal, respectively. The noise variance σ2\sigma^{2} is assumed known, which can be accurately estimated at the BS [35]. Define 𝐘𝐘~/γ\mathbf{Y}\triangleq\tilde{\mathbf{Y}}/\sqrt{\gamma}, 𝐘p𝐘~p/γ\mathbf{Y}_{p}\triangleq\tilde{\mathbf{Y}}_{p}/\sqrt{\gamma}, 𝐘d𝐘~d/γ\mathbf{Y}_{d}\triangleq\tilde{\mathbf{Y}}_{d}/\sqrt{\gamma}, 𝐍𝐍~/γ\mathbf{N}\triangleq\tilde{\mathbf{N}}/\sqrt{\gamma}, 𝐍p𝐍~p/γ\mathbf{N}_{p}\triangleq\tilde{\mathbf{N}}_{p}/\sqrt{\gamma}, and 𝐍d𝐍~d/γ\mathbf{N}_{d}\triangleq\tilde{\mathbf{N}}_{d}/\sqrt{\gamma} as the normalized received signals and noise for the ease of notation. TABLE I summarizes the key notations in this paper and their definitions.

TABLE I: Key Notations and Their Definitions
Notation Definition
MM, NN, KK Number of BS antennas, system users, and active users
\mathcal{M}, 𝒩\mathcal{N}, Ξ\Xi Set of BS antennas, system users, and active users
TT, LL, LdL_{d} Number of symbols, pilot symbols, and data symbols in a transmission block
𝜶n\bm{\alpha}_{n}, βn\beta_{n} Small-scale and large-scale fading coefficients
unu_{n} User activity indicator
𝐡n\mathbf{h}_{n}, 𝐇\mathbf{H} Effective channel coefficient vector and matrix
𝒃n\bm{b}_{n}, 𝒅n\bm{d}_{n}, 𝒄n\bm{c}_{n} Payload bits, code block, and coded bits
NbN_{b}, NdN_{d}, NcN_{c} Number of payload bits, code block bits, and coded bits
𝒳\mathcal{X} Set of constellation points
𝐗p\mathbf{X}_{p}, 𝐗d\mathbf{X}_{d} Pilot sequences and modulated data symbols
𝐘~p\tilde{\mathbf{Y}}_{p}, 𝐘~d\tilde{\mathbf{Y}}_{d} Received pilot and data signals
𝐘p\mathbf{Y}_{p}, 𝐘d\mathbf{Y}_{d} Normalized received pilot and data signals
γ\gamma, σ2\sigma^{2} Uplink transmit power and noise variance
Ξ^\hat{\Xi}, Ξ^c\hat{\Xi}_{c} Estimated set of active users and users that pass CRC
𝒅^n\hat{\bm{d}}_{n}, 𝒃^n\hat{\bm{b}}_{n} Detected code blocks and payload bits

In the following sections, we will develop efficient algorithms to detect the set of active users, estimate their channel coefficients and the transmitted payload bits.

III Joint Estimation via a Turbo Receiver

In this section, we develop a turbo receiver for joint estimation of the user activity, channel coefficients, and payload data of the active users. It is noteworthy that while the turbo principle has achieved great success in conventional multi-user MIMO systems [36, 37], its applications in grant-free massive RA are still unchartered due to the new requirement of user activity detection. Besides, in contrast to most state-of-the-art approaches that follow a sequential user activity detection and data detection/decoding pipeline [9], our design exploits the common sparsity pattern in both the received pilot and data signal. Meanwhile, it takes advantages of the soft decoding information in order to optimize the activity detection and data reception performance.

III-A Overview of the Turbo Receiver

The proposed turbo receiver iterates between a joint estimator and a channel decoder as shown in Fig. 2, which is inspired by the turbo decoding principle [38] that leverages multiple concatenated elementary decoders with the aid of the extrinsic information. In particular, responsible for user activity detection, channel estimation, and soft data symbol detection, the joint estimator is designed based on the BiG-AMP algorithm [17]. It also estimates the posterior probabilities of the transmitted data symbols in each turbo iteration, which are converted as extrinsic information of the coded bits. On the other hand, the channel decoder is developed based on the belief propagation (BP) algorithm [39], which accepts the extrinsic information of the coded bits as input to generate their posteriors. The extrinsic LLRs of the coded bits, i.e., the logarithm of ratio between the probabilities that a coded bit is “0” or “1”, are obtained accordingly and translated to priors of the transmitted data symbols for the use of the joint estimator in the next turbo iteration. The turbo iteration terminates after Q1Q_{1} rounds or when an exit condition is achieved, after which, hard decision is performed to obtain the code block, followed by a cyclic redundancy check. The workflow of the turbo receiver is summarized in Algorithm 1 with details of the joint estimator and channel decoder to be elaborated in the following subsections. Note that an initial estimate of the effective channel coefficients and their variances, as well as the average sparsity levels derived from the estimated effective channel coefficients, are obtained via the AMP algorithm developed in [24]222With prior knowledge of the user active probability, the AMP algorithm [24] estimates the effective channel coefficients and their variances based on the received pilot signal, and a set of belief indicators {ρ~mn}\{\tilde{\rho}_{mn}\}’s are derived as the posterior probabilities of the effective channel coefficients to be non-zero. In this paper, we term ρ~mn\tilde{\rho}_{mn} as the posterior sparsity level of user nn at the mm-th BS antenna, and define ρ¯n1Mmρ~mn\bar{\rho}_{n}\triangleq\frac{1}{M}\sum_{m\in\mathcal{M}}\tilde{\rho}_{mn} as the average sparsity level of user nn, which is a reliable statistic of the activity status and updated iteratively by the joint estimator of the proposed turbo receiver..

Refer to caption
Figure 2: The proposed turbo receiver for massive RA.
Algorithm 1 The Proposed Turbo Receiver for Massive RA

Input: The normalized received signal 𝐘\mathbf{Y}, pilot symbols 𝐗p\mathbf{X}_{p}, maximum number of iterations Q1Q_{1}, and accuracy tolerance ϵ1\epsilon_{1}.
Output: The estimated set of active users Ξ^\hat{\Xi}, the set of users that pass CRC Ξ^c\hat{\Xi}_{c} and their detected payload bits {𝒃^n}\{\hat{\bm{b}}_{n}\}’s.
Initialize: j0j\leftarrow 0, n𝒩n\in\mathcal{N}, x^nt(0)0\hat{x}_{nt}^{(0)}\leftarrow 0, t𝐓dt\in\mathbf{T}_{d}, λn(0)=KN\lambda_{n}^{(0)}=\frac{K}{N}, n𝒩n\in\mathcal{N}, ηnt,s(1)1|𝒳|\eta_{nt,s}^{(1)}\leftarrow\frac{1}{|\mathcal{X}|}, t𝐓dt\in\mathbf{T}_{d}, LEa(cnjc(1))0L_{E}^{a}\left(c_{nj_{c}}^{(1)}\right)\leftarrow 0, jc=1,,Ncj_{c}=1,\cdots,N_{c}.

1:Execute the AMP algorithm in [24] to obtain the initial estimates of the effective channel coefficients {h^mn(0)}\{\hat{h}_{mn}^{(0)}\}’s and their variances {Vmnh(0)}\{V_{mn}^{h(0)}\}’s, and the average sparsity levels {ρ¯n(0)}\{\bar{\rho}_{n}^{(0)}\}’s.
2:while j<Q1j<Q_{1} and Σn,t|x^nt(j)x^nt(j1)|2Σn,t|x^nt(j1)|2>ϵ1\frac{\Sigma_{n,t}|\hat{x}_{nt}^{(j)}-\hat{x}_{nt}^{(j-1)}|^{2}}{\Sigma_{n,t}|\hat{x}_{nt}^{(j-1)}|^{2}}>\epsilon_{1} do
3:   jj+1j\leftarrow j+1
4: //The Joint Estimator//
5:   Based on {h^mn(j1)}\{\hat{h}_{mn}^{(j-1)}\}’s, {Vmnh(j1)}\{V_{mn}^{h(j-1)}\}’s, and {ρ¯n(j1)}\{\bar{\rho}_{n}^{(j-1)}\}’s, the
6: joint estimator executes Algorithm 2 to estimate the set
7: of active users Ξ^(j)\hat{\Xi}^{(j)}, the posterior probabilities that xntx_{nt}
8: equals ss, i.e., η~nt,s(j)\tilde{\eta}_{nt,s}^{(j)}, nΞ^(j)n\in\hat{\Xi}^{(j)}, t𝐓dt\in\mathbf{T}_{d}, and the soft data
9: symbols x^nt(j)\hat{x}_{nt}^{(j)}, t𝐓dt\in\mathbf{T}_{d}.
10:   Convert η~nt,s(j)\tilde{\eta}_{nt,s}^{(j)} to the posterior LLRs of coded bits
11:LEp(cnjc(j))L_{E}^{p}\left(c_{nj_{c}}^{(j)}\right), nΞ^(j)n\in\hat{\Xi}^{(j)} according to (21).
12:   Calculate the extrinsic information LEe(cnjc(j))L_{E}^{e}\left(c_{nj_{c}}^{(j)}\right), nΞ^(j)n\in\hat{\Xi}^{(j)}
13: as input of the channel decoder LDa(cnjc(j))L_{D}^{a}\left(c_{nj_{c}}^{(j)}\right), nΞ^(j)n\in\hat{\Xi}^{(j)}
14: according to (22).
15: //The Channel Decoder//
16:   Perform soft data decoding via a BP-based channel
17: decoder and obtain the posterior LLRs of the coded bits
18:LDp(cnjc(j))L_{D}^{p}\left(c_{nj_{c}}^{(j)}\right), nΞ^(j)n\in\hat{\Xi}^{(j)}.
19:   Calculate the extrinsic information LDe(cnjc(j))L_{D}^{e}\left(c_{nj_{c}}^{(j)}\right), nΞ^(j)n\in\hat{\Xi}^{(j)}
20: via (24) as input of joint estimator LEa(cnjc(j+1))L_{E}^{a}\left(c_{nj_{c}}^{(j+1)}\right) for the
21: next turbo iteration, and obtain the prior probabilities
22: that xntx_{nt} equals ss, i.e., ηnt,s(j+1){\eta}_{nt,s}^{(j+1)}, t𝐓dt\in\mathbf{T}_{d} according to (26).
23:end while
24:Determine the set of active users Ξ^\hat{\Xi} as Ξ^(j)\hat{\Xi}^{(j)}.
25:Perform hard decision based on LDp(cnjc(j))L_{D}^{p}\left(c_{nj_{c}}^{(j)}\right) via (27) to obtain the code blocks 𝒅^n\hat{\bm{d}}_{n}, nΞ^n\in\hat{\Xi}.
26:Perform CRC to determine Ξ^c\hat{\Xi}_{c} and detach the CRC bits from 𝒅^n\hat{\bm{d}}_{n} to obtain 𝒃^n\hat{\bm{b}}_{n}, nΞ^cn\in\hat{\Xi}_{c}.

III-B The Joint Estimator

The joint estimator is designed to detect the set of active users, estimate their channel coefficients and the transmitted data symbols. Since the user activity pattern is encapsulated in 𝐇\mathbf{H} and can be determined accordingly, it remains for the joint estimator to estimate the effective channel coefficients and soft data symbols. We resort to the MMSE estimators, which can be expressed for the effective channel coefficients and soft data symbols respectively as follows [40]:

h^mn𝔼[hmn|𝐘]=hmnp(hmn|𝐘)𝑑hmn,m,n𝒩,\displaystyle\hat{h}_{mn}\triangleq\mathbb{E}\left[h_{mn}|\mathbf{Y}\right]=\int h_{mn}p(h_{mn}|\mathbf{Y})dh_{mn},\forall m\in\mathcal{M},n\in\mathcal{N}, (5)
x^nt𝔼[xnt|𝐘]=xntp(xnt|𝐘),n𝒩,t𝐓d,\displaystyle\hat{x}_{nt}\triangleq\mathbb{E}\left[x_{nt}|\mathbf{Y}\right]=\sum x_{nt}p(x_{nt}|\mathbf{Y}),\forall n\in\mathcal{N},t\in\mathbf{T}_{d}, (6)

where h^mn\hat{h}_{mn} (x^nt\hat{x}_{nt}) is the estimate of hmnh_{mn} (xntx_{nt}), and p(hmn|𝐘)p(h_{mn}|\mathbf{Y}) (p(xnt|𝐘)p(x_{nt}|\mathbf{Y})) denotes the marginal posterior distribution of hmnh_{mn} (xntx_{nt}) given the normalized received signal 𝐘\mathbf{Y}. The marginal posterior distributions can be rewritten in terms of the joint posterior distribution p(𝐇,𝐗|𝐘)p(\mathbf{H},\mathbf{X}|\mathbf{Y}) as follows:

p(hmn|𝐘)=𝐇\m,n𝐗p(𝐇,𝐗|𝐘)d𝐇,\displaystyle p\left(h_{mn}|\mathbf{Y}\right)=\int_{\mathbf{H}_{\backslash m,n}}\sum_{\mathbf{X}}p(\mathbf{H},\mathbf{X}|\mathbf{Y})d\mathbf{H}, (7)
p(xnt|𝐘)=𝐗\n,t𝐇p(𝐇,𝐗|𝐘)𝑑𝐇,\displaystyle p\left(x_{nt}|\mathbf{Y}\right)=\sum_{\mathbf{X}_{\backslash n,t}}\int_{\mathbf{H}}p(\mathbf{H},\mathbf{X}|\mathbf{Y})d\mathbf{H}, (8)

where p(𝐇,𝐗|𝐘)p\left(\mathbf{H},\mathbf{X}|\mathbf{Y}\right) can be factorized via the Bayes’ rule:

p(𝐇,𝐗|𝐘)=p(𝐘|𝐇,𝐗)p(𝐇)p(𝐗)p(𝐘)=(a)1p(𝐘)p(𝐘|𝐇,𝐗)p(𝐇|𝐔)p(𝐔)p(𝐗)=(b)1p(𝐘)m=1Mt=1Tp(ymt|n=1Nhmnxnt)×n=1N[p(un)m=1Mp(hmn|un)t=1Tp(xnt)].\displaystyle\begin{split}p(\mathbf{H},\mathbf{X}|\mathbf{Y})&=\frac{p(\mathbf{Y}|\mathbf{H},\mathbf{X})p(\mathbf{H})p(\mathbf{X})}{p(\mathbf{Y})}\\ &\overset{(a)}{=}\frac{1}{p(\mathbf{Y})}p(\mathbf{Y}|\mathbf{H},\mathbf{X})p(\mathbf{H|U})p(\mathbf{U})p(\mathbf{X})\\ &\overset{(b)}{=}\frac{1}{p(\mathbf{Y})}\prod_{m=1}^{M}\prod_{t=1}^{T}p\left(y_{mt}|\sum_{n=1}^{N}h_{mn}x_{nt}\right)\\ &\times\prod_{n=1}^{N}\left[p\left(u_{n}\right)\prod_{m=1}^{M}p\left(h_{mn}|u_{n}\right)\prod_{t=1}^{T}p\left(x_{nt}\right)\right].\end{split} (9)

In (9), (a) holds since p(𝐇)=p(𝐇,𝐔)=p(𝐇|𝐔)p(𝐔)p(\mathbf{H})=p(\mathbf{H,U})=p(\mathbf{H|U})p(\mathbf{U}), as the user activity pattern is deterministic given 𝐇\mathbf{H}, and (b) is attributed to the conditional independence of random variables.

Nevertheless, the marginal distributions in (7) and (8) are intractable due to high-dimensional integrals and summations. Fortunately, the factorization in (9) implies efficient approximations via the BP algorithm operating on factor graphs [39]. As shown in Fig. 3, a factor graph consists of variable nodes (as indicated by circles), factor nodes (correspond to PDFs as indicated by squares), and edges connecting variable nodes and factor nodes. In the formats of PDFs, messages are propagated in the factor graph and updated iteratively. Specifically, the message from a variable node to a factor node is the product of messages from other adjacent factor nodes of that variable node, while the message from a factor node to a variable node is the integral of the product of that factor and messages from other adjacent variable nodes of that factor node. The posterior PDF of a variable is approximated by the belief of the corresponding variable node, which is the product of messages from all its adjacent factor nodes. For instance, let IxntfymtI_{x_{nt}\rightarrow f_{y_{mt}}} and IfymtxntI_{f_{y_{mt}}\rightarrow x_{nt}} be the messages from variable node xntx_{nt} to factor node p(ymt|n𝒩hmnxnt)p(y_{mt}|\sum_{n\in\mathcal{N}}h_{mn}x_{nt}) and from factor node p(ymt|n𝒩hmnxnt)p(y_{mt}|\sum_{n\in\mathcal{N}}h_{mn}x_{nt}) to variable node xntx_{nt}, respectively. They are updated in each iteration of the BP algorithm as follows:

IxntfymtIfxntxntk{m}Ifyktxnt,\displaystyle I_{x_{nt}\rightarrow f_{y_{mt}}}\leftarrow I_{f_{x_{nt}}\rightarrow x_{nt}}\prod_{k\in\mathcal{M}\setminus\{m\}}I_{f_{y_{kt}}\rightarrow x_{nt}}, (10)
Ifymtxntp(ymtk=1Nhmkxkt)×r𝒩{n}(Ixrtfymt)k𝒩(Ihmkfymt)d𝐡md𝐱t\n,\displaystyle\begin{split}I_{f_{y_{mt}}\rightarrow x_{nt}}&\leftarrow\int p\left(y_{mt}\mid\sum_{k=1}^{N}h_{mk}x_{kt}\right)\\ &\times\prod_{r\in\mathcal{N}\setminus\{n\}}\left(I_{x_{rt}\rightarrow f_{y_{mt}}}\right)\prod_{k\in\mathcal{N}}\left(I_{h_{mk}\rightarrow f_{y_{mt}}}\right)d\mathbf{h}_{m}d\mathbf{x}_{t\backslash n},\end{split} (11)

where IfxntxntI_{f_{x_{nt}}\rightarrow x_{nt}} denotes the message from factor node p(xnt)p(x_{nt}) to variable node xntx_{nt} that is used to approximate the prior distribution of xntx_{nt}, and IhmnfymtI_{h_{mn}\rightarrow f_{y_{mt}}} is the message from variable node hmnh_{mn} to factor node p(ymt|n𝒩hmnxnt)p(y_{mt}|\sum_{n\in\mathcal{N}}h_{mn}x_{nt}). The belief of variable node xntx_{nt} and the approximated posterior distribution of xntx_{nt}, i.e., BxntB_{x_{nt}} and rxntr_{x_{nt}}, are updated via the following expressions, respectively:

BxntIfxntxntmIfymtxnt,\displaystyle B_{x_{nt}}\leftarrow I_{f_{x_{nt}}\rightarrow x_{nt}}\prod_{m\in\mathcal{M}}I_{f_{y_{mt}}\rightarrow x_{nt}}, (12)
rxntBxntBxnt𝑑xnt.\displaystyle r_{x_{nt}}\leftarrow\frac{B_{x_{nt}}}{\int B_{x_{nt}}dx_{nt}}. (13)
Refer to caption
Figure 3: The factor graph of the joint posterior distribution p(𝐇,𝐗|𝐘)p(\mathbf{H,X|Y}), where zmtn𝒩hmnxntz_{mt}\triangleq\sum_{n\in\mathcal{N}}h_{mn}x_{nt}.

Unfortunately, although the BP algorithm is efficient in calculating marginal distributions, computations of the high-dimensional integrals in (11) still exhibit excessive complexity since NN is very large in massive RA systems. To develop a joint estimator with affordable complexity, we turn to the framework of AMP, which is a variant of BP that provides more tractable approximations for marginal distributions [16]. It adopts the Central Limit Theorem to approximate the product of some messages as a complex Gaussian distribution so that only the mean and variance need to be propagated. Besides, high-order terms of messages are omitted in deriving the means and variances to further reduce the computation complexity. Since the joint estimation of effective channel coefficients and soft data symbols belongs to a bilinear inference problem, the BiG-AMP algorithm [17] offers a viable solution by estimating 𝐇\mathbf{H} and 𝐗\mathbf{X} alternatively. Key steps of the BiG-AMP-based joint estimator are summarized in Algorithm 2, which is an iterative algorithm that estimates three sets of variables in each iteration, including: 1) The linear mixing variables {zmt}\{z_{mt}\}’s (zmtn𝒩hmnxntz_{mt}\triangleq\sum_{n\in\mathcal{N}}h_{mn}x_{nt}); 2) The effective channel coefficients {hmn}\{h_{mn}\}’s; and 3) The soft data symbols {xnt}\{x_{nt}\}’s, as elaborated in the following.

Algorithm 2 The Joint Estimator based on BiG-AMP

Input: The normalized received signal 𝐘\mathbf{Y}, pilot symbols 𝐗p\mathbf{X}_{p}, the estimates of the likelihood that each user is active {λn(j)}\{\lambda_{n}^{(j)}\}’s, the estimates of the effective channel coefficients {h^mn(j1)}\{\hat{h}_{mn}^{(j-1)}\}’s and their variances {Vmnh(j1)}\{V_{mn}^{h(j-1)}\}’s, the prior probabilities that xntx_{nt} equals ss, i.e., {ηnt,s(j)}\{{\eta}_{nt,s}^{(j)}\}’s, the threshold of determining the active user θ\theta, maximum number of iterations Q2Q_{2}, and accuracy tolerance ϵ2\epsilon_{2}.
Output: The estimated set of active users Ξ^(j)\hat{\Xi}^{(j)}, and the posterior probabilities that xntx_{nt} equals ss, i.e., η~nt,s(j)\tilde{\eta}_{nt,s}^{(j)}, nΞ^(j)n\in\hat{\Xi}^{(j)}, t𝐓dt\in\mathbf{T}_{d}.
Initialize: i0i\leftarrow 0, h^mn(j)(0)h^mn(j1)\hat{h}_{mn}^{(j)}(0)\leftarrow\hat{h}_{mn}^{(j-1)}, Vmnh(j)(0)Vmnh(j1)V_{mn}^{h(j)}(0)\leftarrow V_{mn}^{h(j-1)}, s^mt(j)(0)0\hat{s}_{mt}^{(j)}(0)\leftarrow 0, x^nt(j)(0)0\hat{x}_{nt}^{(j)}(0)\leftarrow 0, t𝐓dt\in\mathbf{T}_{d}, Vntx(j)(0)1V_{nt}^{x(j)}(0)\leftarrow 1, t𝐓dt\in\mathbf{T}_{d}.

1:while i<Q2i<Q_{2} and Σm,t|z^mt(j)(i)z^mt(j)(i1)|2Σm,t|z^mt(j)(i1)|2>ϵ2\frac{\Sigma_{m,t}|\hat{z}_{mt}^{(j)}(i)-\hat{z}_{mt}^{(j)}(i-1)|^{2}}{\Sigma_{m,t}|\hat{z}_{mt}^{(j)}(i-1)|^{2}}>\epsilon_{2} do
2:   ii+1i\leftarrow i+1
3: //Estimate the Linear Mixing Variable zmt{z}_{mt}//
4:   m,t:Mmtp(j)(i)=nh^mn(j)(i1)x^nt(j)(i1)s^mt(j)(i1)\forall m,t:M_{mt}^{p(j)}(i)=\sum_{n}\hat{h}_{mn}^{(j)}(i-1)\hat{x}_{nt}^{(j)}(i-1)-\hat{s}_{mt}^{(j)}(i-1)
5:n(|x^nt(j)(i1)|2Vmnh(j)(i1)+|h^mn(j)(i1)|2Vntx(j)(i1))\sum_{n}\Big{(}|\hat{x}_{nt}^{(j)}(i-1)|^{2}V_{mn}^{h(j)}(i-1)+|\hat{h}_{mn}^{(j)}(i-1)|^{2}V_{nt}^{x(j)}(i-1)\Big{)}
6:   m,t:Vmtp(j)(i)=n(|x^nt(j)(i1)|2Vmnh(j)(i1)\forall m,t:V_{mt}^{p(j)}(i)=\sum_{n}\Big{(}|\hat{x}_{nt}^{(j)}(i-1)|^{2}V_{mn}^{h(j)}(i-1)
7:+|h^mn(j)(i1)|2Vntx(j)(i1))+nVmnh(j)(i1)Vntx(j)(i1)+|\hat{h}_{mn}^{(j)}(i-1)|^{2}V_{nt}^{x(j)}(i-1)\Big{)}+\sum_{n}V_{mn}^{h(j)}(i-1)V_{nt}^{x(j)}(i-1)
8:   m,t:z^mt(j)(i)=𝔼[zmt|Mmtp(j)(i),Vmtp(j)(i)]\forall m,t:\hat{z}_{mt}^{(j)}(i)=\mathbb{E}\left[z_{mt}|M_{mt}^{p(j)}(i),V_{mt}^{p(j)}(i)\right]
9:=ymtVmtp(j)(i)+(σ2/γ)Mmtp(j)(i)(σ2/γ)+Vmtp(j)(i)=\frac{y_{mt}V_{mt}^{p(j)}(i)+(\sigma^{2}/\gamma)M_{mt}^{p(j)}(i)}{(\sigma^{2}/\gamma)+V_{mt}^{p(j)}(i)}
10:   m,t:Vmtz(j)(i)=Var[zmt|Mmtp(j)(i),Vmtp(j)(i)]\forall m,t:V_{mt}^{z(j)}(i)=\text{Var}\left[z_{mt}|M_{mt}^{p(j)}(i),V_{mt}^{p(j)}(i)\right]
11:=(σ2/γ)Vmtp(j)(i)(σ2/γ)+Vmtp(j)(i)=\frac{(\sigma^{2}/\gamma)V_{mt}^{p(j)}(i)}{(\sigma^{2}/\gamma)+V_{mt}^{p(j)}(i)}
12:   m,t:s^mt(j)(i)=(z^mt(j)(i)Mmtp(j)(i))/Vmtp(j)(i)\forall m,t:\hat{s}_{mt}^{(j)}(i)=\left(\hat{z}_{mt}^{(j)}(i)-M_{mt}^{p(j)}(i)\right)/V_{mt}^{p(j)}(i)
13:   m,t:Vmts(j)(i)=(1Vmtz(j)(i)/Vmtp(j)(i))/Vmtp(j)(i)\forall m,t:V_{mt}^{s(j)}(i)=\left(1-V_{mt}^{z(j)}(i)/{V_{mt}^{p(j)}(i)}\right)/V_{mt}^{p(j)}(i)
14: //Estimate the Effective Channel Coefficients //
15:   m,n:Qp,mnh(j)(i)=(t𝐓p|xnt|2Vmts(j)(i))1\forall m,n:Q_{p,mn}^{h(j)}(i)=\left(\sum_{t\in\mathbf{T}_{p}}|x_{nt}|^{2}V_{mt}^{s(j)}(i)\right)^{-1}
16:   m,n:Pp,mnh(j)(i)=h^mn(j)(i1)+Qp,mnh(j)(i)\forall m,n:P_{p,mn}^{h(j)}(i)=\hat{h}_{mn}^{(j)}(i-1)+Q_{p,mn}^{h(j)}(i)
17:t𝐓pxnts^mt(j)(i)\sum_{t\in\mathbf{T}_{p}}x_{nt}^{*}\hat{s}_{mt}^{(j)}(i)\cdot
18:   m,n:Qd,mnh(j)(i)=(t𝐓d|x^nt(j)(i1)|2Vmts(j)(i))1\forall m,n:Q_{d,mn}^{h(j)}(i)=\left(\sum_{t\in\mathbf{T}_{d}}\left|\hat{x}_{nt}^{(j)}(i-1)\right|^{2}V_{mt}^{s(j)}(i)\right)^{-1}
19:   m,n:Pd,mnh(j)(i)=h^mn(j)(i1)(1Qd,mnh(j)(i)t𝐓d\forall m,n:P_{d,mn}^{h(j)}(i)=\hat{h}_{mn}^{(j)}(i-1)\Big{(}1-Q_{d,mn}^{h(j)}(i)\sum_{t\in\mathbf{T}_{d}}
20:Vntx(j)(i1)Vmts(j)(i))+Qd,mnh(j)(i)t𝐓dx^nt(j)(i1)s^mt(j)(i)V_{nt}^{x(j)}(i-1)V_{mt}^{s(j)}(i)\Big{)}+Q_{d,mn}^{h(j)}(i)\sum_{t\in\mathbf{T}_{d}}\hat{x}_{nt}^{(j)*}(i-1)\hat{s}_{mt}^{(j)}(i)
21:   m,n:Pmnh(j)(i)=(Pp,mnh(j)(i)Qd,mnh(j)(i)+Pd,mnh(j)(i)\forall m,n:P_{mn}^{h(j)}(i)=\Big{(}P_{p,mn}^{h(j)}(i)Q_{d,mn}^{h(j)}(i)+P_{d,mn}^{h(j)}(i)\cdot
22:Qp,mnh(j)(i))/(Qp,mnh(j)(i)+Qd,mnh(j)(i))Q_{p,mn}^{h(j)}(i)\Big{)}\Big{/}\left(Q_{p,mn}^{h(j)}(i)+Q_{d,mn}^{h(j)}(i)\right)
23:   m,n:Qmnh(j)(i)=Qp,mnh(j)(i)Qd,mnh(j)(i)/(Qp,mnh(j)(i)\forall m,n:Q_{mn}^{h(j)}(i)=Q_{p,mn}^{h(j)}(i)Q_{d,mn}^{h(j)}(i)\Big{/}\Big{(}Q_{p,mn}^{h(j)}(i)
24:+Qd,mnh(j)(i))+Q_{d,mn}^{h(j)}(i)\Big{)}
25:   m,n:Kmn(j)(i)=ln(𝒞𝒩(0;Pmnh(j)(i),Qmnh(j)(i)+βn)𝒞𝒩(0;Pmnh(j)(i),Qmnh(j)(i)))\forall m,n:K_{mn}^{(j)}(i)=\ln\left(\frac{\mathcal{CN}\left(0;P_{mn}^{h(j)}(i),Q_{mn}^{h(j)}(i)+\beta_{n}\right)}{\mathcal{CN}\left(0;P_{mn}^{h(j)}(i),Q_{mn}^{h(j)}(i)\right)}\right)
26:=ln(Qmnh(j)(i)Qmnh(j)(i)+βn)+|Pmnh(j)(i)|2βn(Qmnh(j)(i)+βn)Qmnh(j)(i)=\ln\left(\frac{Q_{mn}^{h(j)}(i)}{Q_{mn}^{h(j)}(i)+\beta_{n}}\right)+\frac{\left|P_{mn}^{h(j)}(i)\right|^{2}\beta_{n}}{\left(Q_{mn}^{h(j)}(i)+\beta_{n}\right)Q_{mn}^{h(j)}(i)}
27:   m,n:Lmn(j)(i)=ln(λn(j)1λn(j))+k\{m}(Kkn(j)(i))\forall m,n:L_{mn}^{(j)}(i)=\ln\left(\frac{\lambda_{n}^{(j)}}{1-\lambda_{n}^{(j)}}\right)+\sum_{k\in\mathcal{M}\backslash\{m\}}\left(K_{kn}^{(j)}(i)\right)
28:   m,n:ρmn(j)(i)=exp(Lmn(j)(i))/(1+exp(Lmn(j)(i)))\forall m,n:\rho_{mn}^{(j)}(i)=\exp\left(L_{mn}^{(j)}(i)\right)\Big{/}\left(1+\exp\left(L_{mn}^{(j)}(i)\right)\right)
29:   m,n:μmn(j)(i)=βnPmnh(j)(i)/(βn+Qmnh(j)(i))\forall m,n:\mu_{mn}^{(j)}(i)=\beta_{n}P_{mn}^{h(j)}(i)\Big{/}\left(\beta_{n}+Q_{mn}^{h(j)}(i)\right),
30:τmn(j)(i)=βnQmnh(j)(i)/(βn+Qmnh(j)(i))\tau_{mn}^{(j)}(i)=\beta_{n}Q_{mn}^{h(j)}(i)\Big{/}\left(\beta_{n}+Q_{mn}^{h(j)}(i)\right)
31:   m,n:ρ~mn(j)(i)=ρmn(j)(i)/(ρmn(j)(i)+(1ρmn(j)(i))\forall m,n:\tilde{\rho}_{mn}^{(j)}(i)=\rho_{mn}^{(j)}(i)\Big{/}\Big{(}\rho_{mn}^{(j)}(i)+\left(1-\rho_{mn}^{(j)}(i)\right)\cdot
32:exp(ln(Qmnh(j)(i)Qmnh(j)(i)+βn)|Pmnh(j)(i)|2βn(Qmnh(j)(i)+βn)Qmnh(j)(i)))\exp\Big{(}-\ln\left(\frac{Q_{mn}^{h(j)}(i)}{Q_{mn}^{h(j)}(i)+\beta_{n}}\right)-\frac{|P_{mn}^{h(j)}(i)|^{2}\beta_{n}}{\left(Q_{mn}^{h(j)}(i)+\beta_{n}\right)Q_{mn}^{h(j)}(i)}\Big{)}\Big{)}
33:   m,n:h^mn(j)(i)=𝔼[hmn|Pmnh(j)(i),Qmnh(j)(i)]\forall m,n:\hat{h}_{mn}^{(j)}(i)=\mathbb{E}\left[h_{mn}|P_{mn}^{h(j)}(i),Q_{mn}^{h(j)}(i)\right]
34:=ρ~mn(j)(i)μmn(j)(i)=\tilde{\rho}_{mn}^{(j)}(i)\mu_{mn}^{(j)}(i)
35:   m,n:Vmnh(j)(i)=Var[hmn|Pmnh(j)(i),Qmnh(j)(i)]\forall m,n:V_{mn}^{h(j)}(i)=\text{Var}\left[h_{mn}|P_{mn}^{h(j)}(i),Q_{mn}^{h(j)}(i)\right]
36:=ρ~mn(j)(i)(|μmn(j)(i)|2+τmn(j)(i))|h^mn(j)(i)|2=\tilde{\rho}_{mn}^{(j)}(i)\left(\left|\mu_{mn}^{(j)}(i)\right|^{2}+\tau_{mn}^{(j)}(i)\right)-\left|\hat{h}_{mn}^{(j)}(i)\right|^{2}
37: //Estimate the Soft Data Symbols//
38:   n,t𝐓d:Qntx(j)(i)=(m|h^mn(j)(i1)|2Vmts(j)(i))1\forall n,t\in\mathbf{T}_{d}:Q_{nt}^{x(j)}(i)=\left(\sum_{m}\left|\hat{h}_{mn}^{(j)}(i-1)\right|^{2}V_{mt}^{s(j)}(i)\right)^{-1}
39:   n,t𝐓d:Pntx(j)(i)=x^nt(j)(i1)(1Qntx(j)(i)\forall n,t\in\mathbf{T}_{d}:P_{nt}^{x(j)}(i)=\hat{x}_{nt}^{(j)}(i-1)\Big{(}1-Q_{nt}^{x(j)}(i)\cdot
40:mVmnh(j)(i1)Vmts(j)(i))+Qntx(j)(i)mh^mn(j)(i1)s^mt(j)(i)\sum_{m}V_{mn}^{h(j)}(i-1)V_{mt}^{s(j)}(i)\Big{)}+Q_{nt}^{x(j)}(i)\sum_{m}\hat{h}_{mn}^{(j)*}(i-1)\hat{s}_{mt}^{(j)}(i)
41:   n,t𝐓d:η~nt,s(j)(i)=ηnt,s(j)𝒞𝒩(s;Pntx(j)(i),Qntx(j)(i))s𝒳ηnt,s(j)𝒞𝒩(s;Pntx(j)(i),Qntx(j)(i))\forall n,t\in\mathbf{T}_{d}:\tilde{\eta}_{nt,s}^{(j)}(i)=\frac{\eta_{nt,s}^{(j)}\mathcal{C}\mathcal{N}\left(s;P_{nt}^{x(j)}(i),Q_{nt}^{x(j)}(i)\right)}{\sum_{s^{\prime}\in\mathcal{X}}\eta_{nt,s^{\prime}}^{(j)}\mathcal{C}\mathcal{N}\left(s^{\prime};P_{nt}^{x(j)}(i),Q_{nt}^{x(j)}(i)\right)}
42:   n,t𝐓d:x^nt(j)(i)=𝔼[xnt|Pntx(j)(i),Vntx(j)(i)]\forall n,t\in\mathbf{T}_{d}:\hat{x}_{nt}^{(j)}(i)=\mathbb{E}\left[x_{nt}|P_{nt}^{x(j)}(i),V_{nt}^{x(j)}(i)\right]
43:=ρ¯n(j1)s𝒳η~nt,s(j)(i)s=\bar{\rho}_{n}^{(j-1)}\sum_{s\in\mathcal{X}}\tilde{\eta}_{nt,s}^{(j)}(i)s
44:   n,t𝐓d:Vntx(j)(i)=Var[xnt|Pntx(j)(i),Vntx(j)(i)]\forall n,t\in\mathbf{T}_{d}:V_{nt}^{x(j)}(i)=\text{Var}\left[x_{nt}|P_{nt}^{x(j)}(i),V_{nt}^{x(j)}(i)\right]
45:=s𝒳η~nt,s(j)(i)|ρ¯n(j1)sx^nt(j)(i)|2=\sum_{s\in\mathcal{X}}\tilde{\eta}_{nt,s}^{(j)}(i)\left|\bar{\rho}_{n}^{(j-1)}s-\hat{x}_{nt}^{(j)}(i)\right|^{2}
46:end while
47:Update ρ¯n(j)=1Mmρ~mn(j)\bar{\rho}_{n}^{(j)}=\frac{1}{M}\sum_{m\in\mathcal{M}}\tilde{\rho}_{mn}^{(j)}, h^mn(j)\hat{h}_{mn}^{(j)}, and Vmnh(j)V_{mn}^{h(j)} for the next turbo iteration.
48:Update λn(j+1)=κρ¯n(j)+(1κ)λn(j)\lambda_{n}^{(j+1)}=\kappa\bar{\rho}_{n}^{(j)}+(1-\kappa)\lambda_{n}^{(j)}, n𝒩\forall n\in\mathcal{N}.
49:Determine the estimated set of active users as Ξ^(j){n𝒩ρ¯n(j)θ}\hat{\Xi}^{(j)}\triangleq\{n\in\mathcal{N}\mid\bar{\rho}_{n}^{(j)}\geq\theta\}.

III-B1 Estimate the linear mixing variables

In each iteration, the joint estimator first estimates the linear mixing variable zmtz_{mt} from ymty_{mt}. The basic principle is that if the prior distribution of a variable and its likelihood function are available, the posterior probability can be derived for MMSE estimation by using the Bayes’ rule. Since ymt=zmt+nmty_{mt}=z_{mt}+n_{mt}, the likelihood function is given as follows:

p(ymt|n𝒩hmnxnt)=γπσ2exp(γσ2|ymtn𝒩hmnxnt|2).\displaystyle p\left(y_{mt}|\sum_{n\in\mathcal{N}}h_{mn}x_{nt}\right)=\frac{\gamma}{\pi\sigma^{2}}\text{exp}\left(-\frac{\gamma}{\sigma^{2}}\left|y_{mt}-\sum_{n\in\mathcal{N}}h_{mn}x_{nt}\right|^{2}\right). (14)

The prior distribution of zmtz_{mt} is approximated as a complex Gaussian distribution with mean Mmtp(j)(i)M_{mt}^{p(j)}(i) and variance Vmtp(j)(i)V_{mt}^{p(j)}(i) in the ii-th iteration of the joint estimator, as shown in Line 3 and 4 in Algorithm 2, respectively, where the superscript “(j)” denotes the turbo iteration index, h^mn(j)(i1)\hat{h}_{mn}^{(j)}(i-1) and Vmnh(j)(i1)V_{mn}^{h(j)}(i-1) are the most updated estimate of the effective channel coefficient and its variance, x^nt(j)(i1)\hat{x}_{nt}^{(j)}(i-1) and Vntx(j)(i1)V_{nt}^{x(j)}(i-1) are the latest estimate of the soft data symbol and its variance, and s^mt(j)(i1)\hat{s}_{mt}^{(j)}(i-1) denotes the scaled residual of zmtz_{mt}. Note that for t𝐓pt\in\mathbf{T}_{p}, we have x^nt(j)(i1)=xnt\hat{x}_{nt}^{(j)}(i-1)=x_{nt} and Vntx(j)(i1)=0V_{nt}^{x(j)}(i-1)=0 as the pilot symbols are known at the BS.

Since both the approximated prior distribution of zmtz_{mt} and the likelihood function of zmtz_{mt} are complex Gaussian, the posterior distribution p(zmt|ymt)p\left(z_{mt}|y_{mt}\right) can also be approximated by a complex Gaussian distribution with mean z^mt(j)(i)\hat{z}_{mt}^{(j)}(i) and variance Vmtz(j)(i)V_{mt}^{z(j)}(i) given in Line 5 and 6 in Algorithm 2, respectively. Detailed derivations are provided in Appendix A. Note that the posterior mean of zmtz_{mt} also gives the MMSE estimate z^mt(j)(i)\hat{z}_{mt}^{(j)}(i). Besides, the scaled residual s^mt(j)(i)\hat{s}_{mt}^{(j)}(i) of zmtz_{mt} and the corresponding inverse-residual-variance Vmts(j)(i)V_{mt}^{s(j)}(i) are updated in Line 7 and 8, respectively, which are useful for approximating the likelihood functions of the effective channel coefficients and soft data symbols.

III-B2 Estimate the effective channel coefficients

The effective channel coefficients and their variances are estimated by incorporating both the received pilot and data signals. In order to approximate the posterior distribution of hmnh_{mn} in the ii-th iteration of the joint estimator, we first obtain the belief of variable node hmnh_{mn} as it only differs from the posterior distribution of hmnh_{mn} by a normalizing constant, which can be derived based on the BP algorithm as follows:

Bhmn(j)(i)=Ifhmnhmn(j)(i)t𝐓pIfymthmn(j)(i)t𝐓dIfymthmn(j)(i),\displaystyle B_{h_{mn}}^{(j)}(i)=I_{f_{h_{mn}\rightarrow h_{mn}}}^{(j)}(i)\prod_{t\in\mathbf{T}_{p}}I_{f_{y_{mt}\rightarrow h_{mn}}}^{(j)}(i)\prod_{t\in\mathbf{T}_{d}}I_{f_{y_{mt}\rightarrow h_{mn}}}^{(j)}(i), (15)

where Ifhmnhmn(j)(i)I_{f_{h_{mn}\rightarrow h_{mn}}}^{(j)}(i) denotes the message from factor node p(hmn|un)p(h_{mn}|u_{n}) to variable node hmnh_{mn} that serves as the prior distribution of hmnh_{mn}, and Ifymthmn(j)(i)I_{f_{y_{mt}\rightarrow h_{mn}}}^{(j)}(i) represents the message from factor node p(ymt|zmt)p(y_{mt}|z_{mt}) to variable node hmnh_{mn}. Thus, the term t𝐓pIfymthmn(j)(i)\prod_{t\in\mathbf{T}_{p}}I_{f_{y_{mt}\rightarrow h_{mn}}}^{(j)}(i) t𝐓dIfymthmn(j)(i)\prod_{t\in\mathbf{T}_{d}}I_{f_{y_{mt}\rightarrow h_{mn}}}^{(j)}(i) can be interpreted as the likelihood function of hmnh_{mn} in the ii-th iteration. Specifically, the term t𝐓pIfymthmn(j)(i)\prod_{t\in\mathbf{T}_{p}}I_{f_{y_{mt}\rightarrow h_{mn}}}^{(j)}(i), which corresponds to the received pilot symbols, is approximated as a complex Gaussian PDF with mean Pp,mnh(j)(i)P_{p,mn}^{h(j)}(i) and variance Qp,mnh(j)(i)Q_{p,mn}^{h(j)}(i), and the term t𝐓dIfymthmn(j)(i)\prod_{t\in\mathbf{T}_{d}}I_{f_{y_{mt}\rightarrow h_{mn}}}^{(j)}(i) that relates to the received data symbols, is approximated as another complex Gaussian PDF with mean Pd,mnh(j)(i)P_{d,mn}^{h(j)}(i) and variance Qd,mnh(j)(i)Q_{d,mn}^{h(j)}(i). Consequently, the term t𝐓pIfymthmn(j)(i)t𝐓dIfymthmn(j)(i)\prod_{t\in\mathbf{T}_{p}}I_{f_{y_{mt}\rightarrow h_{mn}}}^{(j)}\!(i)\!\prod_{t\in\mathbf{T}_{d}}I_{f_{y_{mt}\rightarrow h_{mn}}}^{(j)}\!(i) is also approximated as a complex Gaussian PDF with mean Pmnh(j)(i)P_{mn}^{h(j)}(i) and variance Qmnh(j)(i)Q_{mn}^{h(j)}(i) given in Line 13 and 14 of Algorithm 2, respectively, which are derived in Appendix B.

To derive Bhmn(j)(i)B_{h_{mn}}^{(j)}(i), we obtain Ifhmnhmn(j)(i)I_{f_{h_{mn}\rightarrow h_{mn}}}^{(j)}(i) as follows:

Ifhmnhmn(j)(i)=(1ρmn(j)(i))δ(hmn)+ρmn(j)(i)𝒞𝒩(hmn;0,βn),\displaystyle I_{f_{h_{mn}\rightarrow h_{mn}}}^{(j)}(i)=\left(1-\rho_{mn}^{(j)}(i)\right)\delta(h_{mn})\!+\!\rho_{mn}^{(j)}(i)\mathcal{CN}(h_{mn};0,\beta_{n}), (16)

where ρmn(j)(i)\rho_{mn}^{(j)}(i) approximates the probability that hmnh_{mn} is non-zero, and it is defined as the sparsity level of user nn at the mm-th BS antenna. Detailed derivations of (16) are deferred to Appendix C. Note that estimates of the likelihood that each user is active or not in the considered transmission block, i.e., λn(j)\lambda_{n}^{\left(j\right)}, are required for the calculations of {ρmn(j)(i)}\{\rho_{mn}^{(j)}\left(i\right)\}’s in Lines 15-17 of Algorithm 2. We propose to update {λn(j)}\{\lambda_{n}^{\left(j\right)}\}’s in each turbo iteration for more accurate estimation of the BiG-AMP algorithm, as will be elaborated shortly. Therefore, the posterior distribution of hmnh_{mn} is approximated in the ii-th iteration as follows:

rhmn(j)(i)=Bhmn(j)(i)Bhmn(j)(i)𝑑hmn=(1ρ~mn(j)(i))δ(hmn)+ρ~mn(j)(i)𝒞𝒩(hmn;μmn(j)(i),τmn(j)(i)),\displaystyle\begin{split}r_{h_{mn}}^{(j)}(i)&=\frac{B_{h_{mn}}^{(j)}(i)}{\int B_{h_{mn}}^{(j)}(i)dh_{mn}}=\left(1-\tilde{\rho}_{mn}^{(j)}(i)\right)\delta(h_{mn})\\ &+\tilde{\rho}_{mn}^{(j)}(i)\mathcal{C}\mathcal{N}\left(h_{mn};\mu_{mn}^{(j)}(i),\tau_{mn}^{(j)}(i)\right),\end{split} (17)

where μmn(j)(i)\mu_{mn}^{(j)}(i) and τmn(j)(i)\tau_{mn}^{(j)}(i) are given in Line 18 of Algorithm 2, and ρ~mn(j)(i)\tilde{\rho}_{mn}^{(j)}(i) presented in Line 19 is defined as the posterior sparsity level of user nn at the mm-th BS antenna. Based on the posterior distribution, the MMSE estimate of the effective channel coefficient and its variance are obtained in Line 20 and 21 of Algorithm 2, respectively.

III-B3 Estimate the soft data symbols

Due to the symmetry of xntx_{nt} and hmnh_{mn} in the bilinear inference problem of the joint estimator, we similarly obtain the conditional mean Pntx(j)(i)P_{nt}^{x(j)}(i) and variance Qntx(j)(i)Q_{nt}^{x(j)}(i) given xntx_{nt} in Line 23 and 22 of Algorithm 2, respectively. Prior distributions of the transmitted data symbols can be estimated as follows:

p(xnt)=Ifxntxnt(j)(i)ρ¯n(j1)s𝒳ηnt,s(j)δ(xnts),t𝐓d,\displaystyle p(x_{nt})=I_{f_{x_{nt}}\rightarrow x_{nt}}^{(j)}(i)\approx\bar{\rho}_{n}^{(j-1)}\sum_{s\in\mathcal{X}}\eta_{nt,s}^{(j)}\delta(x_{nt}-s),t\in\mathbf{T}_{d}, (18)

where ρ¯n(j1)1Mmρ~mn(j1)\bar{\rho}_{n}^{(j-1)}\triangleq\frac{1}{M}\sum_{m\in\mathcal{M}}\tilde{\rho}_{mn}^{(j-1)}, and ηnt,s(j)\eta_{nt,s}^{(j)} denotes the probability that xntx_{nt} belongs to constellation point ss. As will be introduced in the next subsection, {ηnt,s(j)}\{\eta_{nt,s}^{(j)}\}’s are obtained from the channel decoder in the last turbo iteration. Thus, the approximated posterior distributions of the transmitted data symbols in the ii-th iteration of the joint estimator can be expressed as follows:

rxnt(j)(i)=ρ¯n(j1)s𝒳η~nt,s(j)(i)δ(xnts),t𝐓d,\displaystyle r_{x_{nt}}^{(j)}(i)=\bar{\rho}_{n}^{(j-1)}\sum_{s\in\mathcal{X}}\tilde{\eta}_{nt,s}^{(j)}(i)\delta(x_{nt}-s),t\in\mathbf{T}_{d}, (19)

where η~nt,s(j)(i)\tilde{\eta}_{nt,s}^{(j)}(i) denotes the posterior probability that xntx_{nt} belongs to constellation point ss as derived using the Bayes’ rule in Line 24 of Algorithm 2. The soft data symbols and the corresponding posterior variances are estimated via Line 25 and 26, respectively.

Once the while loop of Algorithm 2 is terminated, ρ¯n(j)\bar{\rho}_{n}^{(j)}, h^mn(j)\hat{h}_{mn}^{(j)} and Vmnh(j)V_{mn}^{h(j)} are updated for the next turbo iteration. Accordingly, we update {λn(j)}\{\lambda_{n}^{(j)}\}’s in Line 28 using the average sparsity level ρ¯n(j)\bar{\rho}_{n}^{(j)}, i.e., p(un=1)λn(j+1)=κρ¯n(j)+(1κ)λn(j)p(u_{n}=1)\triangleq\lambda_{n}^{(j+1)}=\kappa\bar{\rho}_{n}^{(j)}+(1-\kappa)\lambda_{n}^{(j)}, n𝒩n\in\mathcal{N}, where κ[0,1]\kappa\in[0,1] is the learning rate. This is inspired by the idea of exploration and exploitation from reinforcement learning [41], which avoids using the average sparsity levels exclusively to eliminate the potential estimation errors caused by inaccurate prior information of the BiG-AMP algorithm. The set of active users is determined as Ξ^(j){n𝒩|ρ¯n(j)θ}\hat{\Xi}^{(j)}\triangleq\{n\in\mathcal{N}|\bar{\rho}_{n}^{(j)}\geq\theta\}, where θ\theta is an empirical threshold [24].

III-B4 Derive the extrinsic information of the joint estimator

With the estimated set of active users Ξ^(j)\hat{\Xi}^{(j)} and soft data symbols {x^nt}\{\hat{x}_{nt}\}’s, extrinsic information of the joint estimator is derived as input of the channel decoder, which aims at minimizing the data decoding error by eliminating some redundancy from the prior information of the coded bits. In particular, the posterior probability of a transmitted data symbols is translated to the posterior probabilities of the corresponding coded bits as follows:

p(cnjc(j)=b|𝐘)=s𝒳j^cbη~nt,s(j),nΞ^(j),\displaystyle p\left(c_{nj_{c}}^{(j)}=b|\mathbf{Y}\right)=\sum\nolimits_{s\in\mathcal{X}_{\hat{j}_{c}}^{b}}\tilde{\eta}_{nt,s}^{(j)},n\in\hat{\Xi}^{(j)}, (20)

where j^cmod(jc,log2|𝒳|)\hat{j}_{c}\triangleq\mod({j}_{c},\log_{2}|\mathcal{X}|), t=L+1+jclog2|𝒳|t=L+1+\left\lfloor\frac{j_{c}}{\log_{2}|\mathcal{X}|}\right\rfloor, and 𝒳lb\mathcal{X}_{l}^{b} represents the set of constellation points with the ll-th position (l=0,,log2|𝒳|1l=0,\cdots,\log_{2}|\mathcal{X}|-1) of the corresponding bit sequence as bb. For example, suppose the bit sequences “00”, “01”, “10” and “11” are modulated to constellation points s0s_{0}, s1s_{1}, s2s_{2}, and s3s_{3} respectively in quadrature phase shift keying (QPSK), we have 𝒳00={s0,s1}\mathcal{X}_{0}^{0}=\{s_{0},s_{1}\}, 𝒳01={s2,s3}\mathcal{X}_{0}^{1}=\{s_{2},s_{3}\}, 𝒳10={s0,s2}\mathcal{X}_{1}^{0}=\{s_{0},s_{2}\}, and 𝒳11={s1,s3}\mathcal{X}_{1}^{1}=\{s_{1},s_{3}\}. Other modulation schemes such as the 16-Quadrature Amplitude Modulation (QAM) can be applied similarly. The posterior probabilities of the coded bits are used to derive the posterior LLRs as follows:

LEp(cnjc(j))ln(p(cnjc(j)=0|𝐘)p(cnjc(j)=1|𝐘)),nΞ^(j),\displaystyle L_{E}^{p}\left(c_{nj_{c}}^{(j)}\right)\triangleq\ln\left(\frac{p(c_{nj_{c}}^{(j)}=0|\mathbf{Y})}{p(c_{nj_{c}}^{(j)}=1|\mathbf{Y})}\right),n\in\hat{\Xi}^{(j)}, (21)

which are converted to the extrinsic information as defined below [42]:

LEe(cnjc(j))LEp(cnjc(j))LEa(cnjc(j)),nΞ^(j),\displaystyle L_{E}^{e}\left(c_{nj_{c}}^{(j)}\right)\triangleq L_{E}^{p}\left(c_{nj_{c}}^{(j)}\right)-L_{E}^{a}\left(c_{nj_{c}}^{(j)}\right),n\in\hat{\Xi}^{(j)}, (22)

where LEa(cnjc(j))ln(p(cnjc(j)=0)p(cnjc(j)=1))L_{E}^{a}\left(c_{nj_{c}}^{(j)}\right)\triangleq\ln\left(\frac{p(c_{nj_{c}}^{(j)}=0)}{p(c_{nj_{c}}^{(j)}=1)}\right) is the prior information obtained from the channel decoder in the last turbo iteration.

III-C The Channel Decoder

The channel decoder determines the most probable code block for each user that is determined as active by the joint estimator, which can be formulated as the following maximum a posteriori probability (MAP) estimation problem:

𝒄^n=argmax𝒄n{0,1}Ncp(𝒄n{LEe(cnjc(j))}),nΞ^(j).\displaystyle\bm{\hat{c}}_{n}=\arg\max_{\bm{c}_{n}\in\{0,1\}^{N_{c}}}p\left(\bm{c}_{n}\mid\{L_{E}^{e}(c_{nj_{c}}^{(j)})\}\right),n\in\hat{\Xi}^{(j)}. (23)

Known for its effectiveness in calculating marginal distributions, the BP algorithm has a long history of applications for channel decoder designs [42]. In this paper, we adopt a BP-based channel decoder to solve (23), which should be able to accept the extrinsic information of the coded bits derived from the joint estimator as input, and calculate the posterior LLRs of the coded bits LDp(cnjc(j))L_{D}^{p}(c_{nj_{c}}^{(j)}), nΞ^(j)n\in\hat{\Xi}^{(j)} as the soft decoding results. We emphasize that this is a mild requirement that can be satisfied by a variety of off-the-shelf BP-based channel decoders, e.g., the decoders developed in [43], [44] and [38], [45] for low-density parity-check (LDPC) code and turbo code, respectively.

Similar to the joint estimator, extrinsic information of the channel decoder is derived as follows:

LDe(cnjc(j))LDp(cnjc(j))LDa(cnjc(j)),nΞ^(j),\displaystyle L_{D}^{e}\left(c_{nj_{c}}^{(j)}\right)\triangleq L_{D}^{p}\left(c_{nj_{c}}^{(j)}\right)-L_{D}^{a}\left(c_{nj_{c}}^{(j)}\right),n\in\hat{\Xi}^{(j)}, (24)

which is adopted as prior information LEa(cnjc(j+1))L_{E}^{a}(c_{nj_{c}}^{(j+1)}), nΞ^(j)n\in\hat{\Xi}^{(j)} for the use of the joint estimator in the next turbo iteration. Therefore, the prior distribution of a coded bit is given as

p(cnjc(j))={11+exp(LDe(cnjc(j))),cnjc(j)=1,exp(LDe(cnjc(j)))1+exp(LDe(cnjc(j))),cnjc(j)=0,nΞ^(j),\displaystyle p\left(c_{nj_{c}}^{(j)}\right)=\left\{\begin{aligned} \frac{1}{1+\text{exp}\left(L_{D}^{e}\left(c_{nj_{c}}^{(j)}\right)\right)},c_{nj_{c}}^{(j)}=1,\\ \frac{\text{exp}\left(L_{D}^{e}\left(c_{nj_{c}}^{(j)}\right)\right)}{1+\text{exp}\left(L_{D}^{e}\left(c_{nj_{c}}^{(j)}\right)\right)},c_{nj_{c}}^{(j)}=0,\end{aligned}\right.\ n\in\hat{\Xi}^{(j)}, (25)

and prior distributions of the transmitted data symbols can be estimated according to the following expression:

ηnt,s(j+1)=jc=v1v2p(cnjc(j)),t𝐓d,nΞ^(j).\displaystyle{\eta}_{nt,s}^{(j+1)}=\prod\nolimits_{j_{c}=v_{1}}\nolimits^{v_{2}}p\left(c_{nj_{c}}^{(j)}\right),t\in\mathbf{T}_{d},n\in\hat{\Xi}^{(j)}. (26)

where v1(tL1)log2|𝒳|v_{1}\triangleq(t-L-1)\text{log}_{2}|\mathcal{X}|, v2(tL)log2|𝒳|1v_{2}\triangleq(t-L)\text{log}_{2}|\mathcal{X}|-1, and μ([cnv1,,cnv2])=s\mu\left([c_{nv_{1}},\cdots,c_{nv_{2}}]\right)=s. Note that for the users that are determined as inactive, we reuse the prior information of the transmitted data symbols from the last turbo iteration by setting LEa(cnjc(j+1))=LEa(cnjc(j))L_{E}^{a}\left(c_{nj_{c}}^{(j+1)}\right)=L_{E}^{a}\left(c_{nj_{c}}^{(j)}\right) and ηnt,s(j+1)=ηnt,s(j){\eta}_{nt,s}^{(j+1)}={\eta}_{nt,s}^{(j)}, n𝒩Ξ^(j)n\in\mathcal{N}\setminus\hat{\Xi}^{(j)}.

The values of {LDp(cnjc)}\{L_{D}^{p}\left(c_{nj_{c}}\right)\}’s are also utilized to obtain the code block 𝒅^n\hat{\bm{d}}_{n}, nΞ^n\in\hat{\Xi} after the last turbo iteration by performing hard decision as follows:

d^njc={0,LDp(cnjc)0,1,LDp(cnjc)<0.\displaystyle\hat{d}_{nj_{c}}=\left\{\begin{aligned} 0,L_{D}^{p}\left(c_{nj_{c}}\right)\geq 0,\\ 1,L_{D}^{p}\left(c_{nj_{c}}\right)<0.\end{aligned}\right. (27)

CRC is then performed for 𝒅^n,nΞ^\hat{\bm{d}}_{n},n\in\hat{\Xi} and the CRC bits are detached to obtain the payload bits 𝒃^n\hat{\bm{b}}_{n} for the users that pass parity check, which is denoted as Ξ^c\hat{\Xi}_{c} in Algorithm 1.

IV A Low-complexity Side Information-aided Receiver

Assisted by prior information of the transmitted data symbols, the proposed turbo receiver effectively exploits the common sparsity pattern in the received pilot and data signal via BiG-AMP. Nevertheless, such a design incurs significant computation overhead even with a reasonable size of the payload data, since the joint estimation of effective channel coefficients and soft data symbols is performed iteratively by incorporating all the received symbols in each turbo iteration. Besides, in order to estimate the prior information, the channel decoder needs to be executed in each turbo iteration for all users in Ξ^(j)\hat{\Xi}^{\left(j\right)} (See Line 7 of Algorithm 1), which brings additional computation overhead. These observations necessitate low-complexity receivers for massive RA that can leverage both the common sparsity pattern and channel decoding results more efficiently. In this section, we develop a low-complexity side information (SI)-aided receiver without relying on BiG-AMP.

IV-A Overview of the SI-aided Receiver

The SI-aided receiver iterates between a sequential estimator and a channel decoder as shown in Fig. 4. Unlike the turbo receiver developed in Section III, it estimates the effective channel coefficients and soft data symbols sequentially in each iteration to reduce the computation overhead. Specifically, the sequential estimator cascades the AMP algorithm [24] for JADCE, and an MMSE-based soft demodulator to compute the prior LLRs of the coded bits. A BP-based channel decoder is adopted to obtain the posterior LLRs of the coded bits similar as the turbo receiver, while hard decision is required for parity check in each iteration. By updating the SI, i.e., the estimates on whether a user is active, in each iteration jointly based on the average sparsity levels, the posterior LLRs of the coded bits, and the parity check results, the receiver progresses with more precise prior knowledge for the sequential estimator so that more accurate JADCE can be achieved [46]. The workflow of the SI-aided receiver is summarized in Algorithm 3. We introduce details of the sequential estimator and the channel decoder in Section IV-B, and elaborate the design of the SI in Section IV-C.

Refer to caption
Figure 4: The proposed SI-aided receiver for massive RA.
Algorithm 3 The Proposed SI-aided Receiver for Massive RA

Input: The normalized received signal 𝐘\mathbf{Y}, pilot symbols 𝐗p\mathbf{X}_{p}, maximum number of iterations Q3Q_{3}, and accuracy tolerance ϵ3\epsilon_{3}.
Output: The estimated set of active users Ξ^\hat{\Xi}, the set of users that pass CRC Ξ^c\hat{\Xi}_{c} and their detected payload bits 𝒃^n\hat{\bm{b}}_{n}.
Initialize: j0\!j\!\leftarrow\!0, λn(1)KN\lambda_{n}^{(1)}\!\leftarrow\!\frac{K}{N}, n𝒩n\!\in\!\mathcal{N}, x^nt(0)0\hat{x}_{nt}^{(0)}\leftarrow 0, t𝐓dt\in\mathbf{T}_{d}, Ξ^c\hat{\Xi}_{c}\leftarrow\emptyset.

1:while j<Q3j<Q_{3} and Σn,t|x^nt(j)x^nt(j1)|2Σn,t|x^nt(j1)|2>ϵ3\frac{\Sigma_{n,t}|\hat{x}_{nt}^{(j)}-\hat{x}_{nt}^{(j-1)}|^{2}}{\Sigma_{n,t}|\hat{x}_{nt}^{(j-1)}|^{2}}>\epsilon_{3} do
2:   jj+1j\leftarrow j+1
3: //The Sequential Estimator//
4:   Execute the AMP algorithm [24] with the SI {λn(j)}\{\lambda_{n}^{(j)}\}’s
5: as the prior knowledge of the user activity to estimate
6: the effective channel coefficients {h^mn(j)}\{\hat{h}_{mn}^{(j)}\}’s and set of
7: active users Ξ^(j)\hat{\Xi}^{(j)}.
8:   Estimate the transmitted data symbols x^nt(j)\hat{x}_{nt}^{(j)}, t𝐓dt\in\mathbf{T}_{d} via
9: an MMSE equalizer, i.e., 𝐗^d,a(j)=((𝐇^a(j))H𝐇^a(j)+\hat{\mathbf{X}}_{d,a}^{(j)}=\Big{(}(\hat{\mathbf{H}}_{a}^{(j)})^{\mathrm{H}}\hat{\mathbf{H}}_{a}^{(j)}+
10:(σ2/γ)𝐈)1(𝐇^a(j))H𝐘d(\sigma^{2}/\gamma)\mathbf{I}\Big{)}^{-1}(\hat{\mathbf{H}}_{a}^{(j)})^{\mathrm{H}}\mathbf{Y}_{d}, where 𝐇a(j)[{𝐡^k(j)}kΞ^(j)]\mathbf{H}_{a}^{(j)}\triangleq\left[\left\{\hat{\mathbf{h}}_{k}^{(j)}\right\}_{k\in\hat{\Xi}^{(j)}}\right]
11: stacks the effective channel coefficients of all
12: the estimated active users, and x^nt(j)\hat{x}_{nt}^{(j)} is the entry of 𝐗^d,a(j)\hat{\mathbf{X}}_{d,a}^{(j)}.
13:   Compute the prior LLRs of the coded bits as LDa(cnjc(j))L_{D}^{a}\left(c_{nj_{c}}^{(j)}\right),
14:nΞ^(j)n\in\hat{\Xi}^{(j)} via soft demodulation according to (28).
15: //The Channel Decoder//
16:   Perform soft data decoding via a BP-based channel
17: decoder to obtain the posterior LLRs of the coded bits
18:LDp(cnjc(j))L_{D}^{p}\left(c_{nj_{c}}^{(j)}\right), nΞ^(j)Ξ^cn\in\hat{\Xi}^{(j)}\setminus\hat{\Xi}_{c}.
19:   Perform hard decision, determine the set of users in
20:Ξ^(j)\hat{\Xi}^{(j)} that pass Ξ^c(j)\hat{\Xi}_{c}^{(j)} and obtain their payload bits {𝒃^n}\{\hat{\bm{b}}_{n}\}’s.
21:   Ξ^cΞ^cΞ^c(j)\hat{\Xi}_{c}\leftarrow\hat{\Xi}_{c}\bigcup\hat{\Xi}_{c}^{(j)}
22:   Update the SI {λn(j+1)}\{\lambda_{n}^{(j+1)}\}’s according to (29).
23:end while

IV-B The Sequential Estimator and the Channel Decoder

Based on the normalized received pilot signal 𝐘p\mathbf{Y}_{p}, the sequential estimator adopts the AMP algorithm [24] to estimate the effective channel coefficients {h^mn(j)}\{\hat{h}_{mn}^{(j)}\}’s and the set of active users Ξ^(j)\hat{\Xi}^{(j)}. We also derive the sparsity levels {ρmn(j)}\{\rho_{mn}^{(j)}\}’s from the AMP algorithm following similar steps in Lines 15-17 of Algorithm 2. Soft data symbol detection is then performed using an MMSE-based soft demodulator based on the results of the sequential estimator. In particular, signal distortion caused by wireless fading is first removed from the normalized received data signal 𝐘d\mathbf{Y}_{d} to estimate the transmitted data symbols x^nt\hat{x}_{nt}, nΞ^(j)n\in\hat{\Xi}^{(j)}, t𝐓dt\in\mathbf{T}_{d} via an MMSE equalizer. The prior LLRs of the coded bits are obtained via soft demodulation as follows:

LDa(cnjc(j))ln(p(cnjc(j)=0|x^nt(j))p(cnjc(j)=1|x^nt(j)))=ln(s𝒳j^c0exp(γx^nt(j)s22/σ2)s𝒳j^c1exp(γx^nt(j)s22/σ2)),nΞ^(j),\displaystyle\begin{split}L_{D}^{a}\left(c_{nj_{c}}^{(j)}\right)&\triangleq\ln\left(\frac{p\left(c_{nj_{c}}^{(j)}=0|\hat{x}_{nt}^{(j)}\right)}{p\left(c_{nj_{c}}^{(j)}=1|\hat{x}_{nt}^{(j)}\right)}\right)\\ &=\ln\left(\frac{\sum_{s\in\mathcal{X}_{\hat{j}_{c}}^{0}}\exp\left(-\gamma||\hat{x}_{nt}^{(j)}-s||_{2}^{2}/\sigma^{2}\right)}{\sum_{s\in\mathcal{X}_{\hat{j}_{c}}^{1}}\exp\left(-\gamma||\hat{x}_{nt}^{(j)}-s||_{2}^{2}/\sigma^{2}\right)}\right),n\in\hat{\Xi}^{(j)},\end{split} (28)

where j^cmod(jc,log2|𝒳|)\hat{j}_{c}\triangleq\mod({j}_{c},\log_{2}|\mathcal{X}|) and t=L+1+jclog2|𝒳|t=L+1+\left\lfloor\frac{j_{c}}{\log_{2}|\mathcal{X}|}\right\rfloor.

With the knowledge of {LDa(cnjc(j))}\{L_{D}^{a}\left(c_{nj_{c}}^{(j)}\right)\}’s, the BP-based channel decoder calculates the posterior LLRs of the coded bits LDP(cnjc(j))L_{D}^{P}\left(c_{nj_{c}}^{(j)}\right), nΞ^(j)Ξ^cn\in\hat{\Xi}^{(j)}\setminus\hat{\Xi}_{c} and decides the code blocks according to (27). CRC is performed for all users in Ξ^(j)Ξ^c\hat{\Xi}^{(j)}\setminus\hat{\Xi}_{c} to obtain their payload bits {𝒃^n}\{\hat{\bm{b}}_{n}\}’s. Note that channel decoding is not performed for the users that have already passed the parity check, which differs the turbo receiver and helps to save the computations. We would like to point out that the SI-aided receiver can be extended by incorporating the idea of successful interference cancellation, i.e., subtracting the user data that have passed CRC from the received signal. However, the performance improvement is not guaranteed due to the potentially large channel estimation error. A thorough investigation on such an extension will be left for future works.

IV-C The Side Information

The AMP algorithm is a key component of the sequential estimator, which determines the average sparsity levels and effective channel coefficients based on the framework of Bayesian estimation [40]. As a consequence, prior knowledge of the user activity, i.e., the SI for the sequential estimator {λn}\{\lambda_{n}\}’s, also has significant impacts on the estimation accuracy, similar with the case of the BiG-AMP algorithm. In order to obtain more precise estimates through multiple iterations of Algorithm 3, we propose to update the SI by jointly utilizing the results of the sequential estimator and channel decoder, according to three different cases depending on the estimated set of active users and their parity check results via the following update rule:

λn(j+1)={1,nΞ^c,κ1ρ¯n(j)+1κ1Ncjc|LDP(cnjc(j))|1+|LDP(cnjc(j))|,nΞ^(j)Ξ^c,κ2ρ¯n(j)+(1κ2)λn(j),n𝒩(Ξ^(j)Ξ^c).\displaystyle\lambda_{n}^{(j+1)}=\left\{\begin{aligned} &1,\ n\in\hat{\Xi}_{c},\\ &\kappa_{1}\bar{\rho}_{n}^{(j)}+\frac{1-\kappa_{1}}{N_{c}}\sum_{j_{c}}\frac{\left|L_{D}^{P}(c_{nj_{c}}^{(j)})\right|}{1+\left|L_{D}^{P}(c_{nj_{c}}^{(j)})\right|},\ n\in\hat{\Xi}^{(j)}\setminus\hat{\Xi}_{c},\\ &\kappa_{2}\bar{\rho}_{n}^{(j)}+(1-\kappa_{2})\lambda_{n}^{(j)},\ n\in\mathcal{N}\setminus(\hat{\Xi}^{(j)}\cup\hat{\Xi}_{c}).\end{aligned}\right. (29)

In particular, in the first case of (29), we set λn(j+1)=1\lambda_{n}^{(j+1)}=1, nΞ^cn\in\hat{\Xi}_{c} since the users that have passed the parity check in the current or previous iterations can be safely determined as active. In the second case, we handle the users that are estimated as active but fail to pass the parity check in the current iteration, i.e., nΞ^(j)Ξ^cn\in\hat{\Xi}^{(j)}\setminus\hat{\Xi}_{c}. For this set of users, the average sparsity levels and the posterior LLRs of the coded bits are jointly utilized to update the SI since both of them are informative on users’ activity. Specifically, the term 1Ncjc|LDP(cnjc(j))|1+|LDP(cnjc(j))|[0,1)\frac{1}{N_{c}}\sum_{j_{c}}\frac{\left|L_{D}^{P}(c_{nj_{c}}^{(j)})\right|}{1+\left|L_{D}^{P}(c_{nj_{c}}^{(j)})\right|}\in[0,1) indicates the decoding reliability of user nn as its complement, i.e., 1Ncjc11+|LDP(cnjc(j))|\frac{1}{N_{c}}\sum_{j_{c}}\frac{1}{1+\left|L_{D}^{P}(c_{nj_{c}}^{(j)})\right|}, provides an accurate estimate of the bit error rate [47]. Besides, parameter κ1[0,1]\kappa_{1}\in[0,1] is an empirical weighting factor balancing the contributions of the channel estimation and data decoding results. In the third case, we update the SI for the users that neither pass the CRC in any iteration nor being determined as inactive in the current iteration using the average sparsity levels, using similar methodology as that for the turbo receiver in Section III, where κ2[0,1]\kappa_{2}\in[0,1] denotes the learning rate.

To demonstrate the rationality of the SI update rule in (29), we provide numerical examples on the evolutions of {λn(j)}\{\lambda_{n}^{(j)}\}’s, considering two scenarios with K=40K=40 and 8080 in Fig. 5(a) and (b), respectively. From these figures, we see that as the iteration of Algorithm 3 proceeds, the SI evolves from the initial values, which is set to be λn(0)=KN\lambda_{n}^{(0)}=\frac{K}{N}, n𝒩n\in\mathcal{N}, to the perfect estimates, i.e., λn(0)=1\lambda_{n}^{(0)}=1, n=1,,Kn=1,\cdots,K and λn(0)=0\lambda_{n}^{(0)}=0, n=K+1,,Nn=K+1,\cdots,N. This validates the effectiveness of the proposed SI update rule. Besides, we observe that a larger number of active users leads to a slower convergence rate and higher estimation variance, implying the need of more iterations in Algorithm 3.

Refer to caption
Figure 5: Illustrations on the SI update rule in (29). We set N=200N=200 and κ1=κ2=0.5\kappa_{1}=\kappa_{2}=0.5. The first KK users are active in the transmission block.

IV-D Computational Complexity Analysis

The computational complexity of the two proposed receivers is summarized in TABLE II, where the number of complex-valued multiplications is adopted as the metric, and the complexity of a real-valued multiplication is assumed to be one quarter of a complex-valued multiplication. We use OdO_{d} to denote the complexity of the decoder. Since the overall computational complexity of the two proposed receivers is determined by the iteration numbers, i.e., Q1Q_{1} and Q3Q_{3} for the turbo and SI-aided receiver respectively, we only focus on the complexity of one iteration in the following discussions, which is contributed by the estimator and channel decoder.

TABLE II: Complexity of the proposed receivers in one iteration
Receiver Number of complex multiplications
Turbo (94MNL+134MNLd+154MN(\frac{9}{4}MNL+\frac{13}{4}MNL_{d}+\frac{15}{4}MN +32M(L+Ld)+34NLd|𝒳|)Tj+K1Od+\frac{3}{2}M(L+L_{d})+\frac{3}{4}NL_{d}|\mathcal{X}|)T_{j}+K_{1}^{\prime}O_{d}
SI-aided (4MNL+74MN+192NL)Ts+52K22M(4MNL+\frac{7}{4}MN+\frac{19}{2}NL)T_{s}+\frac{5}{2}K_{2}^{2}M +34MK2Ld+54K2+12K2Ld|𝒳|+K2Od+\frac{3}{4}MK_{2}L_{d}+\frac{5}{4}K_{2}+\frac{1}{2}K_{2}L_{d}|\mathcal{X}|+K_{2}^{\prime}O_{d}

As summarized in TABLE II, the complexity of the channel decoder in each turbo iteration is given as K1OdK_{1}^{\prime}O_{d} for the turbo receiver (K2OdK_{2}^{\prime}O_{d} for the SI-aided receiver), where K1K_{1}^{\prime} (K2K_{2}^{\prime}) is the number of users that need to be decoded in this iteration. The other terms shown in the table correspond to the complexity of the joint (sequential) estimator. Compared with the joint estimator that requires (94MNL+134MNLd+154MN+32M(L+(\frac{9}{4}MNL+\frac{13}{4}MNL_{d}+\frac{15}{4}MN+\frac{3}{2}M(L+ Ld)+34NLd|𝒳|)TjL_{d})+\frac{3}{4}NL_{d}|\mathcal{X}|)T_{j} complex-valued multiplications for the BiG-AMP algorithm in each turbo iteration, the sequential estimator first performs the AMP algorithm, followed by an one-shot data detection procedure, which respectively require (4MNL+74MN+192NL)Ts(4MNL+\frac{7}{4}MN+\frac{19}{2}NL)T_{s} and 52K22M+34MK2Ld+54K2+12K2Ld|𝒳|\frac{5}{2}K_{2}^{2}M+\frac{3}{4}MK_{2}L_{d}+\frac{5}{4}K_{2}+\frac{1}{2}K_{2}L_{d}|\mathcal{X}| complex-valued multiplications. Please be noted that K2K_{2} denotes the number of users that are detected as active, and TjT_{j} and TsT_{s} stand for the actual iteration numbers of the BiG-AMP and AMP algorithm, respectively.

We use the simulation setting as will be detailed in Section V to give a more intuitive idea on the computational complexity of the two proposed receivers by assuming K=20K=20. The values of TjT_{j} and TsT_{s}, given respectively by 41 and 52, are obtained by averaging the results over 100100 independent channel realizations. The numbers of complex-valued multiplications of the turbo and SI-aided receiver in one iteration (excluding those of the channel decoders) are given by 3.2×1083.2\times 10^{8} and 1.39×1081.39\times 10^{8}, respectively. In addition, since hard decision is made in each iteration of the SI-aided receiver, K2K_{2}^{\prime} is typically smaller than K1K_{1}^{\prime}, especially in later iterations. These two factors jointly imply that the SI-aided receiver has a much lower complexity compared with the turbo receiver. In the next section, we will compare the computational complexity of different receivers numerically using the measured execution time in simulations.

V Simulation Results

V-A Simulation Setting and Baseline Schemes

A single-cell uplink cellular network is simulated, where 200 users are randomly distributed in a circle with a radius of 500 m centered at the BS equipped with 64 antennas. The path loss of user nn is calculated as βn=128.136.7log10(rn)\beta_{n}=-128.1-36.7\text{log}_{10}(r_{n}) (dB), where rnr_{n} (km) is the distance to the BS. The system bandwidth is 1 MHz, and the user transmit power is 23 dBm. Without otherwise specified, QPSK is employed as the modulation scheme and LDPC code is used for channel coding. Besides, we select CRC-8 to show the effectiveness of the proposed receivers, which is one of the CRC options for the physical uplink control channel (PUCCH) in 3GPP standards [48].

In order to achieve more stable convergence behavior, a damping factor ω(0,1]\omega\in(0,1] [17] is applied to moderate the updates of MmtpM_{mt}^{p}, VmtpV_{mt}^{p}, h^mn\hat{h}_{mn}, and x^nt\hat{x}_{nt}. For instance, the damped version of the estimated soft data symbol can be expressed as x¯nt(i)=ωx^nt(i)+(1ω)x¯nt(i1),t𝐓d\bar{x}_{nt}(i)=\omega\hat{x}_{nt}(i)+(1-\omega)\bar{x}_{nt}(i-1),t\in\mathbf{T}_{d}. In particular, the mean and variance of zmtz_{mt} in Line 3 and 4 of Algorithm 2 are replaced with the damped versions M¯mtp\bar{M}_{mt}^{p} and V¯mtp\bar{V}_{mt}^{p}, respectively, while the damped versions of h^mn\hat{h}_{mn} and x^nt\hat{x}_{nt} are used in Lines 10-12 and Lines 22-23 of Algorithm 2. The simulation results are averaged over 10510^{5} independent channel realizations, and other critical simulation parameters are summarized in TABLE III.

TABLE III: Simulation parameters
Parameters Values Parameters Values
MM 64 TT 200
LL 50 LdL_{d} 150
NbN_{b} 142 NdN_{d} 150
NcN_{c} 300 CRC type CRC-8
ω\omega 0.6 NN 200
θ\theta 0.4 Q1Q_{1}, Q3Q_{3} 6
Q2Q_{2} 100 ϵ1\epsilon_{1}, ϵ2\epsilon_{2}, ϵ3\epsilon_{3} 10510^{-5}
κ\kappa, κ2\kappa_{2} 0.5 κ1\kappa_{1} 0.5
Noise power density -169 dBm/Hz Code rate 1/21/2

We adopt two baseline schemes and a performance upper bound for comparisons:

  • Separate design [24]: This scheme first performs JADCE via the AMP algorithm, after which, data symbols are detected using an MMSE equalizer. The detected soft data symbols are then converted to prior LLRs of the coded bits for data decoding via soft demodulation using (28). This can be viewed as an instance of the SI-aided receiver by setting Q3=1Q_{3}=1.

  • Data-assisted design with BiG-AMP [29]: This scheme exploits the common sparsity pattern using the BiG-AMP algorithm for joint activity detection, channel estimation, and soft data symbol detection. The detected soft data symbols are converted to prior LLRs of the coded bits for data decoding using (28). This is a special case of the turbo receiver when Q1=1Q_{1}=1.

  • Turbo receiver with known user activity: This scheme assumes perfect knowledge of the user activity and consequently, channel estimation and data decoding are performed via the proposed turbo receiver by setting λn(j)=1\lambda_{n}^{(j)}=1, nΞn\in\Xi, and λn(j)=0\lambda_{n}^{(j)}=0, n𝒩Ξn\in\mathcal{N}\setminus\Xi. This scheme serves as a performance upper bound.

Note that all the simulated schemes adopt the same BP-based LDPC decoder [43] for fair comparisons.

V-B Results

We first evaluate the activity detection error probability (including the missed detection and false alarm probability) and the normalized mean square error (NMSE) of channel estimation in Fig. 6 and Fig. 7, respectively. It is observed that a large number of active users degrade both the activity detection and channel estimation performance due to the limited radio resource reserved for pilot transmission. Compared with the separate design, the data-assisted design achieves much lower activity detection and channel estimation errors, validating the benefits of incorporating the received data symbols. It is also seen that the proposed turbo receiver significantly outperforms the data-assisted design as the soft channel decoding results are further utilized to refine the prior distributions of the transmitted data symbols through multiple turbo iterations. Besides, despite with some performance loss compared with the turbo receiver, the low-cost SI-aided receiver secures noticeable performance improvement compared with the data-assisted design, which can be credited to the use of the customized SI as prior knowledge of the user activity.

Refer to caption
Figure 6: Activity detection error probability vs. the number of active users.
Refer to caption
Figure 7: NMSE of channel estimation vs. the number of active users.

Fig. 8 shows the BLER of all the simulated schemes versus the number of active users. Similar to activity detection and channel estimation, the turbo receiver achieves the best BLER performance. Assuming the block error rate requirement is 10310^{-3}, the turbo receiver is able to support 40 active users while the separate design can only support 20, which is a remarkable 100% increase. Compared with the baseline schemes, it also greatly narrows the performance gap to the upper bound scheme with perfect knowledge of the user activity, owing to the more accurate activity detection and channel estimation. Because of the same reason, the SI-aided receiver brings notable BLER reduction compared with the data-assisted design.

Refer to caption
Figure 8: BLER vs. the number of active users.

Since both the turbo receiver and the SI-aided receiver iterate between an estimator and a channel decoder, we further investigate the impacts of the number of iterations, i.e., Q1Q_{1} for the turbo receiver and Q3Q_{3} for the SI-aided receiver, as shown in Fig. 9. We examine the computation complexity of different schemes by measuring their average execution time on the same computing server. Since the average execution time is platform-specific, it is normalized with respect to that of the separate design. As shown in the figure, the separate design has the lowest complexity but the highest BLER, as it ignores both the common sparsity pattern and the information offered by the channel decoder. Besides, it is observed that the performance achieved by both of the proposed receivers improves with the number of iterations (i.e., Q1Q_{1} for the turbo receiver and Q3Q_{3} for the data-assisted design), which again corroborates the effectiveness of the iterative estimation on the prior information of the user activity and the transmitted data symbols. However, such performance improvement is accompanied with increased computation complexity. Compared with the turbo receiver, the SI-aided receiver enjoys 66%\sim74% average execution time reduction since the sequential estimator only processes the pilot signal for JADCE, and channel decoding is performed just for the users that have not passed the parity check. In addition, we notice that the major performance gains of the proposed receivers come from the first few iterations, e.g., seven in the considered scenario, and the subsequent iterations only contribute to marginal further improvement. In other words, there is no need to execute a large number of iterations, and wise choices of hyper-parameters for the proposed receivers are critical to balance the performance gain and the computation cost.

Refer to caption
Figure 9: BLER vs. the normalized average execution time (K=20K=20).

To show the effectiveness of the two proposed receivers with different modulation schemes, we also simulate a grant-free massive RA system with 16-QAM while keeping LdL_{d} unchanged. As shown in Fig. 10, when 16-QAM is used, the BLER performance of different receivers degrades, which is in accordance with intuition. However, we see the two proposed receivers still achieve significant performance improvements compared with the separate design, which again demonstrates the benefit of utilizing the common sparsity and data decoding results in designing grant-free massive RA receivers.

Refer to caption
Figure 10: BLER vs. the number of active users with QPSK and 16-QAM.

Furthermore, we investigate the impacts of the threshold value θ\theta for determining the set of active users in the AMP-based receivers. A covariance-based receiver is also simulated, which applies the covariance-based method [18] for activity detection together with an MMSE channel estimator. Similar to the AMP-based receivers, a threshold ν\nu (ν>0\nu>0) was introduced to detect the set of active users in [18]. It is observed from Fig. 11 and Fig. 12 that there is a tradeoff between the missed detection and false alarm probability, and an optimal threshold value brings the best BLER performance for a given receiver. The two proposed receivers achieve better performance compared with the baselines with various values of θ\theta. Although the covariance-based receiver outperforms the conventional separate and data-assisted designs, it is defeated by the proposed turbo receiver by a large margin. On the other hand, while the SI-aided receiver achieves comparable activity detection performance as the covariance-based receiver, it outperforms the covariance-based receiver in terms of BLER with an optimized threshold value. It is also worthwhile to note that the covariance-based receiver has a very high complexity and its execution time is 6.5 times of the separate design, but that of the SI-aided receiver is only 4.24.2 times. These results demonstrate the competence of the proposed receivers over the covariance-based receiver.

Refer to caption
Figure 11: Probability of missed detection vs. probability of false alarm (K=30K=30, θ{0.1,0.2,0.3,0.4,0.5,0.6,0.7,0.8,0.9}\theta\in\{0.1,0.2,0.3,0.4,0.5,0.6,0.7,0.8,0.9\}, and ν{3,10,20,35,50,75,100,200,300}\nu\in\{3,10,20,35,50,75,100,200,300\}).
Refer to caption
Figure 12: BLER vs. threshold index (K=30K=30, θ{0.1,0.2,0.3\theta\in\{0.1,0.2,0.3, 0.4,0.5,0.6,0.7,0.8,0.9}0.4,0.5,0.6,0.7,0.8,0.9\}, and ν{3,10,20,35,50,75,100,200,300}\nu\in\{3,10,20,35,50,75,100,200,300\}.

VI Conclusions

This paper carried out the first holistic investigation that jointly considered activity detection, channel estimation, and data decoding for grant-free massive random access (RA). A turbo receiver was proposed to exploit the common sparsity pattern in the received pilot and data signal, and its performance is enhanced by the extrinsic information from the channel decoder. To reduce the complexity, we also developed a low-cost side information (SI)-aided receiver, where the SI is updated iteratively to take advantages of the common sparsity pattern and the channel decoding results. Simulation results demonstrated that substantial performance gains can be obtained with advanced receivers for a given protocol.

Our study demonstrates the benefits of exploiting the structured information from the pilot and data symbols, and the importance of incorporating the interplay among activity detection, channel estimation, and data decoding to the design of massive RA receivers. In other words, treating activity detection and data decoding separately, as in many previous studies, leads to highly suboptimal receivers. For future investigations, it would be interesting to extend the proposed receivers to scenarios with spatial-temporal correlation of the user activity and investigate more complex massive RA systems supported by ultra-massive MIMO and reconfigurable meta-surfaces. Meanwhile, since the proposed receivers are iterative algorithms by nature, it is critical to further reduce their computational complexity via, for examples, deep learning based methods [49, 50] to facilitate practical implementation. In addition, optimally controlling the uplink transmit power is also very important.

Appendix A Derivations of z^mt\hat{z}_{mt} and Its Variance

Since p(ymt|zmt)=𝒞𝒩(zmt;ymt,σ2γ)p(y_{mt}|z_{mt})=\mathcal{CN}\left(z_{mt};y_{mt},\frac{\sigma^{2}}{\gamma}\right) and the prior distribution of p(zmt)p(z_{mt}) is approximated as 𝒞𝒩(zmt;Mmtp(j)(i),\mathcal{CN}\Big{(}z_{mt};M_{mt}^{p(j)}(i), Vmtp(j)(i))V_{mt}^{p(j)}(i)\Big{)}, the joint PDF p(ymt,zmt)p(y_{mt},z_{mt}) is approximated in the ii-th iteration of Algorithm 2 by

Jzmt(j)(i)\displaystyle J_{z_{mt}}^{(j)}(i) =𝒞𝒩(zmt;ymt,σ2γ)𝒞𝒩(zmt;Mmtp(j)(i),Vmtp(j)(i))\displaystyle=\mathcal{CN}\left(z_{mt};y_{mt},\frac{\sigma^{2}}{\gamma}\right)\mathcal{CN}\left(z_{mt};M_{mt}^{p(j)}(i),V_{mt}^{p(j)}(i)\right) (30)
=A1𝒞𝒩(zmt;C(j)(i),D(j)(i)),\displaystyle=A_{1}\cdot\mathcal{CN}\left(z_{mt};C^{(j)}(i),D^{(j)}(i)\right),

where C(j)(i)ymtVmtp(j)(i)γ+σ2Mmtp(j)(i)σ2+Vmtp(j)(i)γC^{(j)}(i)\triangleq\frac{y_{mt}V_{mt}^{p(j)}(i)\gamma+\sigma^{2}M_{mt}^{p(j)}(i)}{\sigma^{2}+V_{mt}^{p(j)}(i)\gamma}, D(j)(i)σ2Vmtp(j)(i)σ2+Vmtp(j)(i)γD^{(j)}(i)\triangleq\frac{\sigma^{2}V_{mt}^{p(j)}(i)}{\sigma^{2}+V_{mt}^{p(j)}(i)\gamma}, and A1𝒞𝒩(0;ymtMmtp(j)(i),σ2γ+Vmtp(j)(i))A_{1}\triangleq\mathcal{CN}\Big{(}0;y_{mt}-M_{mt}^{p(j)}(i),\frac{\sigma^{2}}{\gamma}+V_{mt}^{p(j)}(i)\Big{)}. Thus, the posterior distribution of zmtz_{mt} can be approximated in the ii-th iteration via the Bayes’ rule as follows:

rzmt(j)(i)=Jzmt(j)(i)Jzmt(j)(i)𝑑zmt=𝒞𝒩(zmt;C(j)(i),D(j)(i)).\displaystyle r_{z_{mt}}^{(j)}(i)=\frac{J_{z_{mt}}^{(j)}(i)}{\int J_{z_{mt}}^{(j)}(i)dz_{mt}}=\mathcal{CN}\left(z_{mt};C^{(j)}(i),D^{(j)}(i)\right). (31)

Since the MMSE estimate of zmtz_{mt} is the posterior mean, we have z^mt(j)(i)=ymtVmtp(j)(i)γ+σ2Mmtp(j)(i)σ2+Vmtp(j)(i)γ\hat{z}_{mt}^{(j)}(i)=\frac{y_{mt}V_{mt}^{p(j)}(i)\gamma+\sigma^{2}M_{mt}^{p(j)}(i)}{\sigma^{2}+V_{mt}^{p(j)}(i)\gamma}. Accordingly, the variance of the MMSE estimate is given by the posterior variance as Vmtz(j)(i)=σ2Vmtp(j)(i)σ2+Vmtp(j)(i)γV_{mt}^{z(j)}(i)=\frac{\sigma^{2}V_{mt}^{p(j)}(i)}{\sigma^{2}+V_{mt}^{p(j)}(i)\gamma}.

Appendix B Derivations of Pmnh(j)(i)P_{mn}^{h\left(j\right)}\left(i\right) and Qmnh(j)(i)Q_{mn}^{h\left(j\right)}\left(i\right)

According to the principles of the BiG-AMP algorithm, t=1LIfymthmn(j)(i)\prod_{t=1}^{L}I_{f_{y_{mt}}\rightarrow h_{mn}}^{(j)}\left(i\right) and t=L+1TIfymthmn(j)(i)\prod_{t=L+1}^{T}I_{f_{y_{mt}}\rightarrow h_{mn}}^{(j)}\left(i\right) are approximated as complex Gaussian distributions 𝒞𝒩(hmn;Pp,mnh(j)(i),\mathcal{CN}\Big{(}h_{mn};P_{p,mn}^{h(j)}(i), Qp,mnh(j)(i))Q_{p,mn}^{h(j)}(i)\Big{)} and 𝒞𝒩(hmn;Pd,mnh(j)(i),Qd,mnh(j)(i))\mathcal{CN}\Big{(}h_{mn};P_{d,mn}^{h(j)}(i),Q_{d,mn}^{h(j)}(i)\Big{)}, respectively. By substituting these two complex Gaussian PDFs into the term t=1LIfymthmn(j)(i)t=L+1TIfymthmn(j)(i)\prod_{t=1}^{L}I_{f_{y_{mt}}\rightarrow h_{mn}}^{(j)}\left(i\right)\prod_{t=L+1}^{T}I_{f_{y_{mt}}\rightarrow h_{mn}}^{(j)}\left(i\right), we have:

t=1LIfymthmn(j)(i)t=L+1TIfymthmn(j)(i)\displaystyle\prod_{t=1}^{L}I_{f_{y_{mt}}\rightarrow h_{mn}}^{(j)}\left(i\right)\prod_{t=L+1}^{T}I_{f_{y_{mt}}\rightarrow h_{mn}}^{(j)}\left(i\right) (32)
=𝒞𝒩(hmn;Pp,mnh(j)(i),Qp,mnh(j)(i))𝒞𝒩(hmn;Pd,mnh(j)(i),Qd,mnh(j)(i))\displaystyle\!=\!\mathcal{CN}\!\left(h_{mn};\!P_{p,mn}^{h(j)}(i),Q_{p,mn}^{h(j)}(i)\!\right)\!\mathcal{CN}\!\left(h_{mn};\!P_{d,mn}^{h(j)}(i),Q_{d,mn}^{h(j)}(i)\!\right)
=A2𝒞𝒩(hmn;Pmnh(j)(i),Qmnh(j)(i)),\displaystyle\!=\!A_{2}\cdot\mathcal{CN}\!\left(h_{mn};P_{mn}^{h(j)}(i),Q_{mn}^{h(j)}(i)\right),

where A2𝒞𝒩(0;Pp,mnh(j)(i)Pd,mnh(j)(i),Pp,mnh(j)(i)+Pd,mnh(j)(i))A_{2}\!\triangleq\!\mathcal{CN}\left(0;P_{p,mn}^{h(j)}(i)\!-\!P_{d,mn}^{h(j)}(i),P_{p,mn}^{h(j)}(i)\!+\!P_{d,mn}^{h(j)}(i)\right), Pmnh(j)(i)=Pp,mnh(j)(i)Qd,mnh(j)(i)+Pd,mnh(j)(i)Qp,mnh(j)(i)Qp,mnh(j)(i)+Qd,mnh(j)(i)P_{mn}^{h(j)}(i)=\frac{P_{p,mn}^{h(j)}(i)Q_{d,mn}^{h(j)}(i)+P_{d,mn}^{h(j)}(i)Q_{p,mn}^{h(j)}(i)}{Q_{p,mn}^{h(j)}(i)+Q_{d,mn}^{h(j)}(i)}, and Qmnh(j)(i)=Q_{mn}^{h(j)}(i)= Qp,mnh(j)(i)Qd,mnh(j)(i)Qp,mnh(j)(i)+Qd,mnh(j)(i)\frac{Q_{p,mn}^{h(j)}(i)Q_{d,mn}^{h(j)}(i)}{Q_{p,mn}^{h(j)}(i)+Q_{d,mn}^{h(j)}(i)}.

Appendix C Derivation of (16)

It is straightforward that p(hmn|un=1)=𝒞𝒩(hmn;p\left(h_{mn}|u_{n}=1\right)=\mathcal{CN}(h_{mn}; 0,βn)0,\beta_{n}) and p(hmn|un=0)=δ(hmn)p\left(h_{mn}|u_{n}=0\right)=\delta(h_{mn}). Thus, based on the BP algorithm, the term Ifhmnhmn(j)(i)I_{f_{h_{mn}\rightarrow h_{mn}}}^{(j)}(i) can be determined as

Ifhmnhmn(j)(i)\displaystyle I_{f_{h_{mn}}\rightarrow h_{mn}}^{(j)}(i) =un{0,1}p(hmn|un)Iunfhmn(j)(i)\displaystyle=\sum_{u_{n}\in\{0,1\}}p(h_{mn}|u_{n})I_{u_{n}\rightarrow f_{h_{mn}}}^{(j)}(i) (33)
=Iunfhmn(j)(i)|un=0δ(hmn)\displaystyle=I_{u_{n}\rightarrow f_{h_{mn}}}^{\left(j\right)}\left(i\right)\Big{|}_{u_{n}=0}\delta(h_{mn})
+Iunfhmn(j)(i)|un=1𝒞𝒩(hmn;0,βn),\displaystyle+I_{u_{n}\rightarrow f_{h_{mn}}}^{\left(j\right)}\left(i\right)\Big{|}_{u_{n}=1}\mathcal{CN}(h_{mn};0,\beta_{n}),

where Iunfhmn(j)(i)I_{u_{n}\rightarrow f_{h_{mn}}}^{(j)}(i) is the message from variable node unu_{n} to factor node p(hmn|un)p(h_{mn}|u_{n}) that can be obtained as follows:

Iunfhmn(j)(i)=p(un)k\{m}Ifhknun(j)(i).\displaystyle I_{u_{n}\rightarrow f_{h_{mn}}}^{(j)}(i)=p(u_{n})\prod_{k\in\mathcal{M}\backslash\{m\}}I_{f_{h_{kn}}\rightarrow u_{n}}^{(j)}(i). (34)

In (34), p(un)p\left(u_{n}\right) stands for the likelihood that user nn is active or not in the considered transmission block, and Ifhknun(j)(i)I_{f_{h_{kn}}\rightarrow u_{n}}^{(j)}(i) is the message from factor node p(hkn|un)p(h_{kn}|u_{n}) (k{m}k\in\mathcal{M}\setminus\{m\}) to variable node unu_{n} that can be expanded as follows:

Ifhknun(j)(i)=p(hkn|un)Ihknfhkn(j)(i)𝑑hkn.\displaystyle I_{f_{h_{kn}}\rightarrow u_{n}}^{(j)}(i)=\int p\left(h_{kn}|u_{n}\right)I_{h_{kn}\rightarrow f_{h_{kn}}}^{(j)}(i)dh_{kn}. (35)

In (35), Ihknfhkn(j)(i)I_{h_{kn}\rightarrow f_{h_{kn}}}^{(j)}(i) is the message from variable node hknh_{kn} to factor node p(hkn|un)p(h_{kn}|u_{n}) that can be obtained based on the BP algorithm as follows:

Ihknfhkn(j)(i)=t=1LIfykthkn(j)(i)t=L+1TIfykthkn(j)(i)\displaystyle I_{h_{kn}\rightarrow f_{h_{kn}}}^{(j)}(i)=\prod_{t=1}^{L}I_{f_{y_{kt}}\rightarrow h_{kn}}^{(j)}\left(i\right)\prod_{t=L+1}^{T}I_{f_{y_{kt}}\rightarrow h_{kn}}^{(j)}\left(i\right) (36)
=A3𝒞𝒩(hkn;Pknh(j)(i),Qknh(j)(i)),\displaystyle=A_{3}\cdot\mathcal{CN}\left(h_{kn};P_{kn}^{h(j)}(i),Q_{kn}^{h(j)}(i)\right),

where the second equality in (36) adopts the same approximation as the one used in (32), and A3A_{3} is a constant given as 𝒞𝒩(0;\mathcal{CN}\Big{(}0; Pp,knh(j)(i)Pd,knh(j)(i),Qp,knh(j)(i)+Qd,knh(j)(i))P_{p,kn}^{h(j)}(i)-P_{d,kn}^{h(j)}(i),Q_{p,kn}^{h(j)}(i)+Q_{d,kn}^{h(j)}(i)\Big{)}. By substituting the right-hand side of (36) into (35), we have:

Ifhknun(j)(i)\displaystyle I_{f_{h_{kn}}\rightarrow u_{n}}^{(j)}(i) =A3p(hkn|un)\displaystyle=A_{3}\int p(h_{kn}|u_{n}) (37)
×𝒞𝒩(hkn;Pknh(j)(i),Qknh(j)(i))dhkn\displaystyle\times\mathcal{CN}\left(h_{kn};P_{kn}^{h(j)}(i),Q_{kn}^{h(j)}(i)\right)dh_{kn}
=A3{𝒞𝒩(0;Pknh(j)(i),Qknh(j)(i)+βn),un=1,𝒞𝒩(0;Pknh(j)(i),Qknh(j)(i)),un=0.\displaystyle=A_{3}\left\{\begin{array}[]{ll}\mathcal{CN}\left(0;P_{kn}^{h(j)}(i),Q_{kn}^{h(j)}(i)+\beta_{n}\right),u_{n}=1,\\ \mathcal{CN}\left(0;P_{kn}^{h(j)}(i),Q_{kn}^{h(j)}(i)\right),u_{n}=0.\end{array}\right. (40)

Define

Lmn(j)(i)ln(Iunfhmn(j)(i)|un=1Iunfhmn(j)(i)|un=0),\displaystyle L_{mn}^{(j)}(i)\triangleq\ln\left(\frac{I_{u_{n}\rightarrow f_{h_{mn}}}^{(j)}(i)|_{u_{n}=1}}{I_{u_{n}\rightarrow f_{h_{mn}}}^{(j)}(i)|_{u_{n}=0}}\right), (41)

which uniquely determines the values of Iunfhmn(j)(i)|un=1I_{u_{n}\rightarrow f_{h_{mn}}}^{(j)}(i)\Big{|}_{u_{n}=1} and Iunfhmn(j)(i)|un=0I_{u_{n}\rightarrow f_{h_{mn}}}^{(j)}(i)\Big{|}_{u_{n}=0}, since un{0,1}Iunfhmn(j)(i)=1\sum_{u_{n}\in\{0,1\}}I_{u_{n}\rightarrow f_{h_{mn}}}^{(j)}(i)=1. We further define

Kkn(j)(i)\displaystyle K_{kn}^{(j)}(i) ln(Ifhknun(j)(i)|un=1Ifhknun(j)(i)|un=0)\displaystyle\triangleq\ln\left(\frac{I_{f_{h_{kn}}\rightarrow u_{n}}^{(j)}(i)|_{u_{n}=1}}{I_{f_{h_{kn}}\rightarrow u_{n}}^{(j)}(i)|_{u_{n}=0}}\right) (42)
=ln(Qmnh(j)(i)Qmnh(j)(i)+βn)+|Pmnh(j)(i)|2βn(Qmnh(j)(i)+βn)Qmnh(j)(i)\displaystyle=\ln\left(\frac{Q_{mn}^{h(j)}(i)}{Q_{mn}^{h(j)}(i)+\beta_{n}}\right)+\frac{|P_{mn}^{h(j)}(i)|^{2}\beta_{n}}{\left(Q_{mn}^{h(j)}(i)+\beta_{n}\right)Q_{mn}^{h(j)}(i)}

and Unln(p(un=1)p(un=0))U_{n}\triangleq\ln\left(\frac{p\left(u_{n}=1\right)}{p\left(u_{n}=0\right)}\right). With some basic mathematical manipulations, Lmn(j)(i)L_{mn}^{(j)}(i) can be simplified as follows:

Lmn(j)(i)=Un+k\{m}Kkn(j)(i).\displaystyle L_{mn}^{(j)}(i)=U_{n}+\sum\nolimits_{k\in\mathcal{M}\backslash\{m\}}K_{kn}^{(j)}(i). (43)

As a result, the terms Iunfhmn(j)(i)|un=1I_{u_{n}\rightarrow f_{h_{mn}}}^{(j)}(i)\Big{|}_{u_{n}=1} and Iunfhmn(j)(i)|un=0I_{u_{n}\rightarrow f_{h_{mn}}}^{(j)}(i)\Big{|}_{u_{n}=0} can be determined as exp(Lmn(j)(i))1+exp(Lmn(j)(i))\frac{\exp{\left(L_{mn}^{(j)}(i)\right)}}{1+\exp{\left(L_{mn}^{(j)}(i)\right)}} and 11+exp(Lmn(j)(i))\frac{1}{1+\exp{\left(L_{mn}^{(j)}(i)\right)}}, respectively. We complete the derivation by defining ρmn(j)(i)exp(Lmn(j)(i))1+exp(Lmn(j)(i))\rho_{mn}^{(j)}(i)\triangleq\frac{\exp{\left(L_{mn}^{(j)}(i)\right)}}{1+\exp{\left(L_{mn}^{(j)}(i)\right)}}.

References

  • [1] X. Bian, Y. Mao, and J. Zhang, “Joint activity detection and data decoding in massive random access via a turbo receiver,” in Proc. IEEE Int. Workshop Signal Process. Adv. Wireless Commun. (SPAWC), Sep. 2021.
  • [2] A. Al-Fuqaha, M. Guizani, M. Mohammadi, M. Aledhari, and M. Ayyash, “Internet of Things: A survey on enabling technologies, protocols, and applications,” IEEE Commun. Surveys Tut., vol. 17, no. 4, pp. 2347-2376, Fourth Quart. 2015.
  • [3] Cisco, “Cisco annual Internet report (2018–2023),” Cisco White Paper, Mar. 2020.
  • [4] ITU-R, “Framework and overall objectives of the future development of IMT for 2020 and beyond Recommendation,” ITU-R M.2083-0, 2015.
  • [5] C. Bockelmann, N. Pratas, H. Nikopour, K. Au, T. Svensson, C. Stefanovic, P. Popovski, and A. Dekorsy, “Massive machine-type communications in 5G: Physical and MAC-layer solutions,” IEEE Commun. Mag., vol. 54, no. 9, pp. 59–65, Sep. 2016.
  • [6] M. Hasan, E. Hossain, and D. Niyato, “Random access for machine-to-machine communication in LTE-advanced networks: Issues and approaches,” IEEE Commun. Mag., vol. 51, no. 6, pp. 86-93, Jun. 2013.
  • [7] E. Björnson, E. Carvalho, J. H. Sørensen, E. G. Larsson, and P. Popovski, “A random access protocol for pilot allocation in crowded massive MIMO systems,” IEEE Trans. Wireless Commun., vol. 16, no. 4, pp. 2220-2234, Apr. 2017.
  • [8] N. Jiang, Y. Deng, A. Nallanathan, X. Kang, and T. Q. S. Quek, “Analyzing random access collisions in massive IoT networks,” IEEE Trans. Wireless Commun., vol. 17, no. 10, pp. 6853-6879, Oct. 2018.
  • [9] L. Liu, E. G. Larsson, W. Yu, P. Popovski, C. Stefanović, and E. Carvalh, “Sparse signal processing for grant-free massive connectivity: A future paradigm for random access protocols in the Internet of Things,” IEEE Signal Process. Mag., vol. 35, no. 5, pp. 88-99, Sep. 2018.
  • [10] P. Schulz et al., “Latency critical IoT applications in 5G: Perspective on the design of radio interface and network architecture,” IEEE Commun. Mag., vol. 55, no. 2, pp. 70-78, Feb. 2017.
  • [11] X. Chen, D. Ng, W. Yu, E. G. Larsson, N. Al-Dhahir, and R. Schober, “Massive access for 5G and beyond,” IEEE J. Sel. Areas Commun., vol. 39, no. 3, pp. 615-637, Mar. 2021.
  • [12] Y. Wu, X. Gao, S. Zhou, W. Yang, Y. Polyanskiy, and G. Caire, “Massive access for future wireless communication systems,” IEEE Wireless Commun., vol. 27, no. 4, pp. 148-156, Aug. 2020.
  • [13] D. L. Donoho, “Compressed sensing,” IEEE Trans. Inf. Theory, vol. 52, no. 4, pp. 1289–1306, Apr. 2006.
  • [14] T. Robert, “Regression shrinkage and selection via the Lasso,” J. Roy. Statist. Soc., vol. 58, no. 1, pp. 267–288, Jan. 1996.
  • [15] J. Tropp, “Greed is good: Algorithmic results for sparse approximation,” IEEE Trans. Inf. Theory, vol. 50, no. 10, pp. 2231– 2242, Oct. 2004.
  • [16] D. L. Donoho, A. Maleki, and A. Montanari, “Message passing algorithms for compressed sensing: I. motivation and construction,” in Proc. IEEE Inf. Theory Workshop (ITW), Cairo, Egypt, Jan. 2010.
  • [17] J. T. Parker, P. Schniter, and V. Cevher, “Bilinear generalized approximate message passing – Part I: Derivation,” IEEE Trans. Signal Process., vol. 62, no. 22, pp. 5839–5853, Nov. 2014.
  • [18] S. Haghighatshoar, P. Jung, and G. Caire, “Improved scaling law for activity detection in massive MIMO systems,” in Proc. IEEE Int. Symp. Inf. Theory (ISIT), Vail, CO, USA, Jun. 2018.
  • [19] Z. Chen, F. Sohrabi, and W. Yu, “Multi-cell sparse activity detection for massive random access: Massive MIMO versus cooperative MIMO,” IEEE Trans. Wireless Commun., vol. 18, no. 8, pp. 4060-4072, Aug. 2019.
  • [20] B. Wang, L. Dai, Y. Zhang, T. Mir, and J. Li, “Dynamic compressive sensing-based multi-user detection for uplink grant-Free NOMA,” IEEE Commun. Lett., vol. 20, no. 11, pp. 2320-2323, Nov. 2016.
  • [21] C. Wei, H. Liu, Z. Zhang, J. Dang, and L. Wu, “Approximate message passing-based joint user activity and data detection for NOMA,” IEEE Commun. Lett., vol. 21, no. 3, pp. 640–643, Mar. 2017.
  • [22] Z. Chen, F. Sohrabi, and W. Yu, “Sparse activity detection for massive connectivity,” IEEE Trans. Signal Process., vol. 66, no. 7, pp. 1890–1904, Apr. 2018.
  • [23] L. Liu and W. Yu, “Massive connectivity with massive MIMO – Part I: Device activity detection and channel estimation,” IEEE Trans. Signal Process., vol. 66, no. 11, pp. 2933–2946, Jun. 2018.
  • [24] M. Ke, Z. Gao, Y. Wu, X. Gao, and R. Schober, “Compressive sensing based adaptive active user detection and channel estimation: Massive access meets massive MIMO,” IEEE Trans. Signal Process., vol. 68, pp. 764–779, 2020.
  • [25] Y. Cheng, L. Liu, and P. Li, “Orthogonal AMP for massive access in channels with spatial and temporal correlations,” IEEE J. Sel. Areas Commun., vol. 39, no. 3, pp. 726-740, Mar. 2021.
  • [26] X. Shao, X. Chen and R. Jia, “A dimension reduction-based joint activity detection and channel estimation algorithm for massive access,” IEEE Trans. Signal Process., vol. 68, pp. 420-435, Dec. 2019.
  • [27] Y. Cui, S. Li and W. Zhang, “Jointly sparse signal recovery and support recovery via deep learning with applications in MIMO-based grant-free random access,” IEEE J. Sel. Areas Commun., vol. 39, no. 3, pp. 788-803, Mar. 2021.
  • [28] Y. Du, et al., “Joint channel estimation and multiuser detection for uplink grant-free NOMA,” IEEE Wireless Commun. Lett., vol. 7, no. 4, pp. 682–685, Feb. 2018.
  • [29] Q. Zou, H. Zhang, D. Cai and H. Yang, “A low-complexity joint user activity, channel and data estimation for grant-free massive MIMO systems,” IEEE Signal Process. Lett., vol. 27, pp. 1290-1294, 2020.
  • [30] X. Bian, Y. Mao, and J. Zhang, “Supporting more active users for massive access via data-assisted activity detection,” in Proc. IEEE Int. Conf. Commun. (ICC), Montreal, QC, Canada, Jun. 2021.
  • [31] B. M. Hochwald and S. t. Brink, “Achieving near-capacity on a multiple-antenna channel,” IEEE Trans. Commun., vol. 51, no. 3, pp. 389-399, Mar. 2003.
  • [32] K. Wong, A. Paulraj, and R. Murch, “Efficient high-performance decoding for overloaded MIMO antenna systems,” IEEE Trans. Wireless Commun., vol. 6, no. 5, pp. 1833–1843, May 2007.
  • [33] 3GPP, “3GPP TS 36.212 version 10.0.0 Release 10,” Jan. 2011.
  • [34] 3GPP, “3GPP TS 38.212 version 15.10.0 Release 15,” Nov. 2020.
  • [35] T. Cui and C. Tellambura, “Power delay profile and noise variance estimation for OFDM,” IEEE Commun. Lett., vol. 10, no. 1, pp. 25-27, Jan. 2006.
  • [36] S. Haykin, M. Sellathurai, Y. de Jong, and T. Willink, “Turbo-MIMO for wireless communications,” IEEE Commun. Mag., vol. 42, no. 10, pp. 48-53, Oct. 2004.
  • [37] X. Wautelet, A. Dejonghe, and L. Vandendorpe, “MMSE-based fractional turbo receiver for space-time BICM over frequency-selective MIMO fading channels,” IEEE Trans. Signal Process., vol. 52, no. 6, pp. 1804-1809, Jun. 2004.
  • [38] C. Berrou and A. Glavieux, “Near optimum error correcting coding and decoding: Turbo-codes,” IEEE Trans. Commun., vol. 44, no. 10, pp. 1261-1271, Oct. 1996.
  • [39] F. R. Kschischang, B. J. Frey and H. A. Loeliger, “Factor graphs and the sum-product algorithm,” IEEE Trans. Inf. Theory, vol. 47, no. 2, pp. 498-519, Feb. 2001.
  • [40] S. Kay,​ Fundamentals of Statistical Signal Processing:​ Estimation Theory. Prentice Hall, Englewood Cliffs, NJ, USA, 1993.
  • [41] R. S. Sutton and A. G. Barto, Reinforcement Learning: An Introduction. MIT Press, 2018.
  • [42] B. Vucetic and J. Yuan, Turbo Codes: Principles and Applications. Springer, 2001.
  • [43] R. G. Gallager, “Low density parity check codes,” IRE Trans. Inf. Theory, vol. IT-8, no. 1, pp. 21-28, Jan. 1962.
  • [44] D. E. Hocevar, “A reduced complexity decoder architecture via layered decoding of LDPC codes,” in Proc. IEEE Workshop Signal Process. Syst. (SiPS), Austin, TX, USA, Oct. 2004.
  • [45] J. Hagenauer, E. Offer, and L. Papke, “Iterative decoding of binary block and convolutional codes,” IEEE Trans. Inf. Theory, vol. 42, no. 2, pp. 429-445, Mar. 1996.
  • [46] A. Ma, Y. Zhou, C. Rush, D. Baron, and D. Needell, “An approximate message passing framework for side information,” IEEE Trans. Signal Process., vol. 67, no. 7, pp. 1875-1888, Apr. 2019.
  • [47] B. Goektepe, S. Faehse, L. Thiele, T. Schierl, and C. Hellge, “Subcode-based early HARQ for 5G,” in Proc. IEEE Int. Conf. Commun. (ICC), Kansas City, MO, USA, May 2018.
  • [48] 3GPP, “3GPP TS 36.211 version 15.3.0 Release 15,” Oct. 2018.
  • [49] V. Satorras and M. Welling, “Neural enhanced belief propagation on factor graphs,” in Proc. AISTATS-21, pp. 685–693, Apr. 2021.
  • [50] Y. Shen, Y. Shi, J. Zhang, and K. B. Letaief, “Graph neural networks for scalable radio resource management: architecture design and theoretical analysis,” IEEE J. Select. Areas Commun., vol. 39, no. 1, pp. 101–115, Jan. 2021.