This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Soft-input, soft-output joint detection and GRAND

Hadi Sarieddeen Research Laboratory of Electronics
Massachusetts Institute of Technology
Cambridge, MA 02139, USA
[email protected]
   Muriel Médard Research Laboratory of Electronics
Massachusetts Institute of Technology
Cambridge, MA 02139, USA
[email protected]
   Ken. R. Duffy The project or effort depicted was or is sponsored by the Defense Advanced Research Projects Agency under Grant number HR00112120008. The content of the information does not necessarily reflect the position or policy of the Government, and no official endorsement should be inferred. To appear in the IEEE GLOBECOM 2022 proceedings. Hamilton Institute
Maynooth University
Ireland
[email protected]
Abstract

Guessing random additive noise decoding (GRAND) is a maximum likelihood (ML) decoding method that identifies the noise effects corrupting code-words of arbitrary code-books. In a joint detection and decoding framework, this work demonstrates how GRAND can leverage crude soft information in received symbols and channel state information to generate, through guesswork, soft bit reliability outputs in log-likelihood ratios (LLRs). The LLRs are generated via successive computations of Euclidean-distance metrics corresponding to candidate noise-recovered words. Noting that the entropy of noise is much smaller than that of information bits, a small number of noise effect guesses generally suffices to hit a code-word, which allows generating LLRs for critical bits; LLR saturation is applied to the remaining bits. In an iterative (turbo) mode, the generated LLRs at a given soft-input, soft-output GRAND iteration serve as enhanced a priori information that adapts noise-sequence guess ordering in a subsequent iteration. Simulations demonstrate that a few turbo-GRAND iterations match the performance of ML-detection-based soft-GRAND in both AWGN and Rayleigh fading channels at a complexity cost that, on average, grows linearly (instead of exponentially) with the number of symbols.

Index Terms:
GRAND, soft-GRAND, turbo-GRAND

I Introduction

The upcoming sixth generation (6G) of wireless communications promises to support a plethora of data-demanding and delay-sensitive applications [1], requiring both ultra-broadband high-frequency connectivity [2] and ultra-reliable low-latency communication (URLLC) [3]. Such variety in requirements compels a paradigm shift from structured, code-specific channel-code decoding to universal and practical decoding that is efficient for different code rates and lengths. Early universal near-maximum-likelihood (ML) decoders for linear codes adopted a list-decoding principle [4, 5]. Recently, guessing random additive noise decoding (GRAND) [6, 7] is celebrated as a novel and practical universal decoder suited for block-code constructions of moderate redundancy.

GRAND is a capacity-achieving channel-code decoder that has demonstrated ML decoding performance on arbitrary (even unstructured) code-books. Instead of directly identifying the transmitted code-word, GRAND aims at identifying the noise effect that corrupts the code-word; it successively reverses the noise effects from the received signal to recover candidate transmitted words. By leveraging information on channel and noise models, the candidate noise sequences are ordered and queried in decreasing likelihood, guaranteeing the first recovered code-word to be the ML decoding solution, even for channels with memory in the absence of interleaving. The guesswork literature [8, 9, 10] establishes the computational feasibility of GRAND for all moderate redundancy codes, where the Shannon entropy rate of noise is typically less than that of information symbols [7]. Furthermore, GRAND’s computational efficiency and modularity have resulted in highly efficient circuit design, as demonstrated in a recent 65 nm [11] synthesis and a 40 nm [12] CMOS implementation.

Incorporating soft-detection symbol reliability information into decoding decisions enhances decoding accuracy [13, 14]. At one end, GRAND can leverage as soft information a one-bit mask designating whether a channel use is reliable or not [15]; specifying reliable bits via a channel-fading-induced mask is also proposed in [16]. At the other end, the complete information in continuous channel outputs serves as soft symbol-reliability information in the soft-GRAND (SGRAND) scheme [17]. A compromise between one-bit soft-GRAND and SGRAND is ordered reliability bits GRAND (ORBGRAND) [18], which matches the decoding accuracy of SGRAND through code-book-independent quantization of soft information in a hardware-friendly algorithm.

With fading channels, generating high-resolution soft information (as opposed to masks [16]) through exhaustive ML detection is computationally demanding, and low-complexity soft-output detectors only generate sub-optimal LLRs. In the absence of soft information (or with low-quality information), iterative decoding schemes can intrinsically generate soft-decoding reliability information to be fed as soft-input decoding information in subsequent iterations [19]. Such information can be updated over both the detection and decoding iterations [20], introducing more degrees of freedom in versatile, adaptive communication systems. However, how to generate soft-reliability outputs through GRAND remains unclear. Consequently, iterative soft-input, soft-output (SISO) GRAND (turbo-GRAND) has not yet been investigated.

In this work, we propose a variation of GRAND that does not avail of input soft-detection bit reliability information but leverages complex received symbols (soft-information in a crude, unprocessed form), channel state information (CSI), and demapped bits (linear detector outputs) to generate soft-decoding bit-reliability information in log-likelihood ratios (LLRs). We compute the LLRs by populating Euclidean distance metrics corresponding to a list of candidate words–not necessarily code-words–resulting from noise guesswork. Because the number of guesses before hitting a code-word is limited, we cannot compute LLRs for all bits; we propose LLR thresholding. We propose a turbo-GRAND scheme in which the generated extrinsic LLRs are fed as a priori LLRs in subsequent SISO decoding iterations. With access to complex received vectors and continuous CSI, the Euclidean distance computations are common to both detection and decoding; turbo-GRAND thus realizes joint detection and decoding.

The paper is organized as follows. The problem formulation is first presented in Sec. II. Then, the proposed turbo-GRAND scheme is detailed in Sec. III. Performance and complexity results are reported in Sec. IV; conclusions are drawn in Sec. V. Regarding notation, bold upper case, bold lower case, and lower case letters correspond to matrices, vectors, and scalars, respectively. Scalar and vector L2\text{L}_{2} norms are denoted by ||\left|{\cdot}\right| and \left\|{\cdot}\right\|, respectively. 𝖤[]\mathsf{E}[\cdot], ()T(\cdot)^{T}, and ()(\cdot)^{*}, stand for the expected value, transpose, and conjugate transpose, respectively. 𝐈M\mathbf{I}_{M} is an identity matrix of size MM, 𝟎N\mathbf{0}_{N} is a vector of zeros of size NN, 𝔽u\mbox{\bb F}_{u} denotes a Galois field with uu elements, Pr()\mathrm{Pr}(\cdot) is the probability function, and \odot is the Hadamard product.

II System Model and Problem Formulation

II-A System Model

We consider a communication system of equivalent baseband input-output relation, 𝐲=𝐇𝐱+𝐧\mathbf{y}=\mathbf{H}\mathbf{x}+{\mathbf{n}}, where 𝐲M×1\mathbf{y}\!\in\!\mbox{\bb C}^{M\times 1} is the received symbol vector, 𝐇M×M\mathbf{H}\!\in\!\mbox{\bb C}^{M\times M} is the channel matrix, 𝐱=[x1xixM]TM×1\mathbf{x}\!=\![x_{1}\cdots x_{i}\cdots x_{M}]^{T}\!\in\!\mbox{\bb C}^{M\times 1} is the transmitted symbol vector, and 𝐧=[n1ninM]TM×1{\mathbf{n}}\!=\![{n}_{1}\cdots{n}_{i}\cdots{n}_{M}]^{T}\!\in\!\mbox{\bb C}^{M\times 1} is the additive–possibly colored–noise vector (𝖤[nini]=σi2)\left(\mathsf{E}[{n}_{i}{n}_{i}^{*}]\!=\!\sigma_{i}^{2}\right). Note that 𝐇\mathbf{H} can be an identity matrix in an additive white Gaussian noise (AWGN) system, a diagonal matrix under point-to-point fading, or a complete matrix under spatial diversity/multiplexing schemes. Furthermore, we assume the information symbol, xix_{i}, to belong to a normalized complex constellation, 𝒳i\mathcal{X}_{i} (𝖤[xixi]=1)\left(\mathsf{E}[x_{i}^{*}x_{i}]\!=\!1\right). Consequently, 𝐱𝒳¯M×1\mathbf{x}\!\in\!\bar{\mathcal{X}}\subset\mbox{\bb C}^{M\times 1}, where 𝒳¯\bar{\mathcal{X}} is the finite MM-dimensional lattice of all possible modulated symbol vectors. The bit-representation of xix_{i} is 𝐜i=[ci,1ci,jci,qi]T𝔽2qi\mathbf{c}_{i}\!=\![c_{i,1}\cdots c_{i,j}\cdots c_{i,q_{i}}]^{T}\!\in\!\mbox{\bb F}_{2}^{q_{i}}, where qi=log2(|𝒳i|)q_{i}\!=\!\lceil{\log_{2}(\left|{\mathcal{X}_{i}}\right|)\rceil}. The bit-representation of 𝐱\mathbf{x} is thus 𝐜=[𝐜1𝐜i𝐜M]T𝔽2N\mathbf{c}\!=\![\mathbf{c}_{1}\cdots\mathbf{c}_{i}\cdots\mathbf{c}_{M}]^{T}\!\in\!\mbox{\bb F}_{2}^{N}, where N=i=1MqiN\!=\!\sum_{i=1}^{M}q_{i}. We assume 𝐜\mathbf{c} to be a code-word encoded with an error correcting code α:𝔽2K𝔽2N\alpha\!:\!\mbox{\bb F}_{2}^{K}\!\rightarrow\!\mbox{\bb F}_{2}^{N}, of code-rate R=K/NR\!=\!K/N. A code-book 𝒞{𝐜:𝐜=α(𝐛),𝐛𝔽2K}\mathcal{C}\!\triangleq\!\{\mathbf{c}\!:\mathbf{c}\!=\!\alpha(\mathbf{b}),\mathbf{b}\!\in\!\mbox{\bb F}_{2}^{K}\} includes all possible code-words, where 𝐛\mathbf{b} is the uncoded bit vector. We denote by 𝐯+N\mathbf{v}\!\in\!{\mbox{\bb R}^{+}}^{N} a vector of σi,j2\sigma_{i,j}^{2} values, the second-order noise statistics per bit. At the receiver side, assuming perfect CSI, a hard detector, β:M𝒳¯\beta:\mbox{\bb C}^{M}\!\rightarrow\!\bar{\mathcal{X}}, equalizes the channel and recovers a symbol vector, 𝐱^\hat{\mathbf{x}}, from 𝐲\mathbf{y}; a demapper recovers a word, 𝐜^\hat{\mathbf{c}}, from 𝐱^\hat{\mathbf{x}}.

II-B Problem Formulation

Algorithm 1 Hard GRAND
1:Demapped bits 𝐜^\hat{\mathbf{c}}; ordered noise-generating function Π\Pi; abandonment threshold BB
2:Decoded 𝐜¯GRAND\bar{\mathbf{c}}^{\mathrm{GRAND}}
3:k0k\leftarrow 0
4:while k<2Nk<2^{N} do
5:     kk+1k\leftarrow k+1
6:     𝐰Π(k)\mathbf{w}\leftarrow\Pi(k) \triangleright kth{k}{\text{th}} likely noise sequence
7:     if 𝐜^𝐰𝒞\hat{\mathbf{c}}\ominus\mathbf{w}\in\mathcal{C} or k=Bk=B then
8:         𝐜¯GRAND𝐜^𝐰\bar{\mathbf{c}}^{\mathrm{GRAND}}\leftarrow\hat{\mathbf{c}}\ominus\mathbf{w}
9:         return 𝐜¯GRAND\bar{\mathbf{c}}^{\mathrm{GRAND}}
10:     end if
11:end while

The ML decoder computes the conditional probability of the demapped word, 𝐜^\hat{\mathbf{c}}, for each of the 2K2^{K} code-words, 𝐜\mathbf{c}, in code-book, 𝒞\mathcal{C}. The 𝐜\mathbf{c} with the highest conditional likelihood of transmission given what was received is the ML solution, 𝐜¯ML=argmax{Pr(𝐜^𝐜):𝐜𝒞}\bar{\mathbf{c}}^{\mathrm{ML}}\!=\!\operatorname*{arg\,max}\{\mathrm{Pr}\left(\hat{\mathbf{c}}\mid\mathbf{c}\right)\!:\!\mathbf{c}\!\in\!\mathcal{C}\}. Instead of searching code-words, GRAND searches putative, not necessarily memoryless, noise effect sequences that corrupt 𝐜\mathbf{c}, 𝐰=[w1wi,jwN]T𝔽2N\mathbf{w}\!=\![w_{1}\cdots w_{i,j}\cdots w_{N}]^{T}\!\in\!\mbox{\bb F}_{2}^{N}, with non-increasing probability. We express the channel’s action at the bit level through function \oplus, where 𝐜¯=𝐜𝐰\bar{\mathbf{c}}=\mathbf{c}\oplus\mathbf{w}. We can write Pr(𝐜¯𝐜)=Pr(𝐜¯=𝐜𝐰)\mathrm{Pr}\left(\bar{\mathbf{c}}\!\mid\!\mathbf{c}\right)=\mathrm{Pr}\left(\bar{\mathbf{c}}=\mathbf{c}\oplus\mathbf{w}\right), and it follows that

𝐜¯GRAND=argmax{Pr(𝐰=𝐜¯𝐜):𝐜𝒞}.\bar{\mathbf{c}}^{\mathrm{GRAND}}=\operatorname*{arg\,max}\{\mathrm{Pr}\left(\mathbf{w}=\bar{\mathbf{c}}\ominus\mathbf{c}\right):\mathbf{c}\in\mathcal{C}\}. (1)

The receiver creates a list of noise effect sequences of decreasing order of likelihood through a noise generating function Π:{1,,2N}𝐰𝔽2N\Pi\!:\!\{1,\!\cdots\!,2^{N}\}\!\rightarrow\!\mathbf{w}\!\in\!\mbox{\bb F}_{2}^{N}\!; the sequences are queried until the first code-word hit, 𝐰=𝐜^𝐜\mathbf{w}\!=\!\hat{\mathbf{c}}\!\ominus\!\mathbf{c} (block-code syndrome computations). GRAND is thus a ML decoder that executes Algorithm 1 and returns 𝐜¯GRAND\bar{\mathbf{c}}^{\mathrm{GRAND}}; information bits are retrieved as 𝐛¯GRAND=α1(𝐜¯GRAND)\bar{\mathbf{b}}^{\mathrm{GRAND}}\!=\!\alpha^{-1}(\bar{\mathbf{c}}^{\mathrm{GRAND}}). Because the entropy of noise is small in most communication systems, GRAND is low-complexity. GRAND’s efficiency is further guaranteed by abandoning guessing after a computational cut-off [7], BB.

Soft GRAND accepts, in addition to 𝐜^\hat{\mathbf{c}}, bit-reliability information in a vector 𝚲=[λ1,1λi,jλM,q]TN\mathbf{\Lambda}\!=\![\lambda_{1,1}\cdots\lambda_{i,j}\cdots\lambda_{M,q}]^{T}\!\in\!\mbox{\bb R}^{N} (assuming qi=q,iq_{i}\!=\!q,\forall i). We can generate a weight metric per putative noise sequence by multiplying the noise sequences by |𝚲|\left|{\mathbf{\Lambda}}\right|; noise sequences with smaller weights are more likely to occur. However, 𝚲\mathbf{\Lambda} is not always available (or available but with bad quality), as soft-output detectors are computationally demanding. We aim at generating soft-reliability outputs within GRAND, proposing a low-complexity noise-centric soft-output decoding algorithm, a function α¯:{𝔽2K,M}N\bar{\alpha}:\{\mbox{\bb F}_{2}^{K},\mbox{\bb C}^{M}\}\!\rightarrow\!\mbox{\bb R}^{N}, that accepts 𝐜^\hat{\mathbf{c}} and 𝐲\mathbf{y} and generates output LLRs, 𝚲¯=[λ¯1,1λ¯i,jλ¯M,q]TN\bar{\mathbf{\Lambda}}\!=\![\bar{\lambda}_{1,1}\cdots\bar{\lambda}_{i,j}\cdots\bar{\lambda}_{M,q}]^{T}\!\in\!\mbox{\bb R}^{N}, with the knowledge of 𝐇\mathbf{H}. By further integrating knowledge of noise statistics in LLR computations (scaling by noise variance), our proposal can account for noise bursts (Markovian channel noise, for example), foregoing interleaves, and whitening filters [7, 21]. The extrinsic LLRs, 𝚲¯\bar{\mathbf{\Lambda}}, can then be fed as intrinsic LLRs, 𝚲\mathbf{\Lambda}, in a subsequent soft GRAND; an iteration that is repeated in the proposed turbo-GRAND.

III Proposed SISO Turbo-GRAND

We propose a SISO variation of GRAND that leverages, in addition to the resources of Algorithm 1, the received symbols, CSI, and optional noise statistics, to generate extrinsic bit-reliability LLRs through joint detection and decoding. The LLRs are fed as input soft-decoding information to a subsequent SISO GRAND iteration (up to TT iterations), resulting in turbo-GRAND (Fig. 1). Turbo-GRAND aims to approach the performance of a soft GRAND with soft-input information from an exhaustive soft-output ML detector. Therefore, turbo-GRAND is helpful in the absence of soft-input information or the presence of sub-optimal soft information.

The LLR of bit jj of symbol ii is defined as

λi,j=log(Pr(ci,j=1,𝐲,𝐇)/Pr(ci,j=0,𝐲,𝐇)).\lambda_{i,j}=\log\left(\mathrm{Pr}\left(c_{i,j}\!=\!1,\mathbf{y},\mathbf{H}\right)/\mathrm{Pr}\left(c_{i,j}\!=\!0,\mathbf{y},\mathbf{H}\right)\right). (2)

Near-optimal ML-detection log-max LLRs [22] are computed by exhaustively searching the lattice 𝒳¯\bar{\mathcal{X}}, computing |𝒳1|×|𝒳i|××|𝒳M|\left|{\mathcal{X}_{1}}\right|\times\left|{\mathcal{X}_{i}}\right|\times\cdots\times\left|{\mathcal{X}_{M}}\right| Euclidean distance metrics to solve for [23]

λi,jML1σ2(min𝐱𝒳¯i,j,1𝐲𝐇𝐱2min𝐱𝒳¯i,j,0𝐲𝐇𝐱2),\lambda_{i,j}^{\mathrm{ML}}\approx\frac{1}{\sigma^{2}}\left(\min_{\mathbf{x}\in\bar{\mathcal{X}}^{i,j,1}}{\left\|{\mathbf{y}-\mathbf{H}\mathbf{x}}\right\|^{2}}-\min_{\mathbf{x}\in\bar{\mathcal{X}}^{i,j,0}}{\left\|{\mathbf{y}-\mathbf{H}\mathbf{x}}\right\|^{2}}\right), (3)

where 𝒳¯i,j,1{𝐱𝒳¯:ci,j=1}\bar{\mathcal{X}}^{i,j,1}\!\triangleq\!\{\mathbf{x}\!\in\!\bar{\mathcal{X}}:c_{i,j}\!=\!1\} and 𝒳¯i,j,0{𝐱𝒳¯:ci,j=0}\bar{\mathcal{X}}^{i,j,0}\!\triangleq\!\{\mathbf{x}\!\in\!\bar{\mathcal{X}}:c_{i,j}\!=\!0\} are subsets of symbol vectors in 𝒳¯\bar{\mathcal{X}}, having in the corresponding jth{j}{\text{th}} bit of the ith{i}{\text{th}} symbol a value of 11 and 0, respectively. We have further assumed the case of white noise, σi,j=σ,i,j\sigma_{i,j}\!=\!\sigma,\forall i,j. For colored noise, 𝐲𝐇𝐱2\left\|{\mathbf{y}-\mathbf{H}\mathbf{x}}\right\|^{2} is replaced by (𝐲𝐇𝐱)Γ1(𝐲𝐇𝐱)\left(\mathbf{y}-\mathbf{H}\mathbf{x}\right)^{*}\Gamma^{-1}\left(\mathbf{y}-\mathbf{H}\mathbf{x}\right), where Γ=diag(𝐯)\Gamma\!=\!\text{diag}\left(\mathbf{v}\right). The hard ML detection output is 𝐱^ML=argmin𝐱𝒳¯𝐲𝐇𝐱2\hat{\mathbf{x}}^{\mathrm{ML}}\!=\!\operatorname*{arg\,min}_{\mathbf{x}\in\bar{\mathcal{X}}}\left\|{\mathbf{y}-\mathbf{H}\mathbf{x}}\right\|^{2}. Furthermore, soft information can be extracted per symbol in linear detectors [24], which are near-optimal in point-to-point systems at a high signal-to-noise ratio (SNR) but sub-optimal under symbol interference/correlation. In particular, a zero-forcing (ZF) detector equalizes the channel by multiplying by its pseudo-inverse, 𝐲^ZF=(𝐇𝐇)1𝐇𝐲\mathbf{\hat{y}}^{\mathrm{ZF}}\!=\!\left(\mathbf{H}^{*}\mathbf{H}\right)^{-1}\mathbf{H}^{*}\mathbf{y}. The per-symbol ZF LLRs in 𝚲ZF\mathbf{\Lambda}^{\mathrm{ZF}} are

λi,jZF=1σiZF2(minxi𝒳ij,1|y^iZFxi|2minxi𝒳ij,0|y^iZFxi|2),\lambda_{i,j}^{\mathrm{ZF}}=\frac{1}{{\sigma_{i}^{\mathrm{ZF}}}^{2}}\left(\min_{x_{i}\in\mathcal{X}_{i}^{j,1}}\left|{\hat{y}_{i}^{\mathrm{ZF}}-x_{i}}\right|^{2}-\min_{x_{i}\in\mathcal{X}_{i}^{j,0}}\left|{\hat{y}_{i}^{\mathrm{ZF}}-x_{i}}\right|^{2}\right), (4)

where 𝒳ij,1{xi𝒳i:ci,j=1}\mathcal{X}_{i}^{j,1}\!\triangleq\!\{x_{i}\!\in\!\mathcal{X}_{i}:c_{i,j}\!=\!1\} and 𝒳ij,0{xi𝒳i:ci,j=0}\mathcal{X}_{i}^{j,0}\!\triangleq\!\{x_{i}\!\in\!\mathcal{X}_{i}:c_{i,j}\!=\!0\} are subsets of symbols in the one-dimensional constellation 𝒳i\mathcal{X}_{i}, having a jth{j}{\text{th}} bit of 11 and 0, respectively, and σiZF2=σi2(hihi)1{\sigma_{i}^{\mathrm{ZF}}}^{2}\!=\!\sigma_{i}^{2}\!\left(h_{i}^{*}h_{i}\right)^{-1} is a scaled noise variance. The ZF hard-outputs are computed per symbol as xi^ZF=y^iZFxi𝒳i\hat{x_{i}}^{\mathrm{ZF}}\!=\!\left\lfloor\hat{y}^{\mathrm{ZF}}_{i}-x_{i}\right\rceil_{\mathcal{X}_{i}}, where η𝒳iargminx𝒳i|ηx|\lfloor\eta\rceil_{\mathcal{X}_{i}}\triangleq\operatorname*{arg\,min}_{x\in\mathcal{X}_{i}}\left|{\eta-x}\right| is the slicing operator over 𝒳i\mathcal{X}_{i}. The demapped version of 𝐱^ZF\hat{\mathbf{x}}^{\mathrm{ZF}}, 𝐜^ZF\hat{\mathbf{c}}^{\mathrm{ZF}}, is the initial vector over which noise effects are guessed in turbo-GRAND.

Refer to caption
Figure 1: A block diagram of turbo-GRAND.
Algorithm 2 Joint detection and decoding - turbo-GRAND
1:Demapped bits 𝐜^=𝐜^ZF\hat{\mathbf{c}}\!=\!\hat{\mathbf{c}}^{\mathrm{ZF}}; received symbols 𝐲\mathbf{y}; channel matrix 𝐇\mathbf{H}; noise matrix 𝐖\mathbf{W}; noise statistics 𝐯\mathbf{v}; noise generating/sorting function Π/Π´\Pi/\acute{\Pi}; input LLRs 𝚲\mathbf{\Lambda} (𝚲=𝚲ZF\mathbf{\Lambda}\!=\!\mathbf{\Lambda}^{\mathrm{ZF}} or 𝚲=𝟎N\mathbf{\Lambda}\!=\!\mathbf{0}_{N}); abandonment threshold BB
2:Output LLRs 𝚲¯GRAND\bar{\mathbf{\Lambda}}^{\mathrm{GRAND}}; demapped 𝐜^GRAND\hat{\mathbf{c}}^{\mathrm{GRAND}}; decoded 𝐜¯GRAND\bar{\mathbf{c}}^{\mathrm{GRAND}}
3:dMLd^{\mathrm{ML}}\!\leftarrow\!\infty; d¯ML\bar{d}^{\mathrm{ML}}\!\leftarrow\!\infty; 𝐝cMLN(1/𝐯)\mathbf{d}^{\mathrm{cML}}\!\leftarrow\!N\odot(1/\mathbf{v})
4:𝐜^GRAND𝐜^\hat{\mathbf{c}}^{\mathrm{GRAND}}\!\leftarrow\!\hat{\mathbf{c}}; 𝐜¯GRAND𝐜^\bar{\mathbf{c}}^{\mathrm{GRAND}}\!\leftarrow\!\hat{\mathbf{c}}; 𝚲¯GRAND𝚲\bar{\mathbf{\Lambda}}^{\mathrm{GRAND}}\!\leftarrow\!\mathbf{\Lambda}
5:t0t\leftarrow 0;
6:while t<Tt<T do
7:     𝐬Π´(𝐖×|𝚲¯GRAND|)\mathbf{s}\leftarrow\acute{\Pi}\left(\mathbf{W}\times\left|{\bar{\mathbf{\Lambda}}^{\mathrm{GRAND}}}\right|\right)
8:     k0k\leftarrow 0
9:     while k<Bk<B do
10:         kk+1k\leftarrow k+1; 𝐰Π(𝐬(k))\mathbf{w}\leftarrow\Pi(\mathbf{s}(k)) \triangleright kth{k}{\text{th}} likely noise
11:         𝐜¯𝐜^GRAND𝐰\bar{\mathbf{c}}\leftarrow\hat{\mathbf{c}}^{\mathrm{GRAND}}\ominus\mathbf{w}; 𝐱¯mod(𝐜¯)\bar{\mathbf{x}}\leftarrow\text{mod}\left(\bar{\mathbf{c}}\right)
12:         d(𝐲𝐇𝐱¯)Γ1(𝐲𝐇𝐱¯)d\leftarrow\left(\mathbf{y}-\mathbf{H}\bar{\mathbf{x}}\right)^{*}\Gamma^{-1}\left(\mathbf{y}-\mathbf{H}\bar{\mathbf{x}}\right) \triangleright Γ=diag(𝐯)\Gamma\!=\!\text{diag}\left(\mathbf{v}\right)
13:         n0n\leftarrow 0
14:         while n<Nn<N do
15:              nn+1n\leftarrow n+1
16:              if c¯nc^nGRAND\bar{c}_{n}\neq\hat{c}^{\mathrm{GRAND}}_{n} then\triangleright complementary bits
17:                  if d<dMLd<d^{\mathrm{ML}} then
18:                       dncMLdMLd^{\mathrm{cML}}_{n}\leftarrow d^{\mathrm{ML}}\triangleright update dcMLd^{\mathrm{cML}} values
19:                  else if d<dncMLd<d^{\mathrm{cML}}_{n} then
20:                       dncMLdd^{\mathrm{cML}}_{n}\leftarrow d
21:                  end if
22:              end if
23:         end while
24:         if d<dMLd<d^{\mathrm{ML}} then \triangleright detection
25:              𝐜^GRAND𝐜¯\hat{\mathbf{c}}^{\mathrm{GRAND}}\leftarrow\bar{\mathbf{c}}; dMLdd^{\mathrm{ML}}\leftarrow d
26:         end if
27:         if 𝐜¯𝒞\bar{\mathbf{c}}\in\mathcal{C} or k=Bk=B then \triangleright code-word hit
28:              if d<d¯MLd<\bar{d}^{\mathrm{ML}} then \triangleright decoding
29:                  𝐜¯GRAND𝐜¯\bar{\mathbf{c}}^{\mathrm{GRAND}}\leftarrow\bar{\mathbf{c}}; d¯MLd\bar{d}^{\mathrm{ML}}\leftarrow d
30:              end if
31:              𝚲¯GRAND(2𝐜^GRAND1)(dML𝐝cML)\bar{\mathbf{\Lambda}}^{\mathrm{GRAND}}\leftarrow\left(2\hat{\mathbf{c}}^{\mathrm{GRAND}}-1\right)\odot\left(d^{\mathrm{ML}}-\mathbf{d}^{\mathrm{cML}}\right)
32:              tt+1t\leftarrow t+1
33:              break \triangleright go to next turbo iteration - line 4
34:         end if
35:     end while
36:end while
37:return 𝚲¯GRAND\bar{\mathbf{\Lambda}}^{\mathrm{GRAND}}, 𝐜^GRAND\hat{\mathbf{c}}^{\mathrm{GRAND}}, and 𝐜¯GRAND\bar{\mathbf{c}}^{\mathrm{GRAND}}

Contrary to conventional ML and list decoders that query all or multiple code-words, the basic implementation of GRAND achieves ML decoding performance by querying noise sequences and recovering a single code-word. This scarcity in code-word hits across turbo-GRAND iterations is mitigated in joint detection and decoding. We propose generating soft-output LLRs by populating a number of Euclidean distance computations equal to the number of noise guesses, as opposed to the exponential number of distance computations in (3). In particular, through guesswork, we aim at extracting an updated hard-detected vector, 𝐜^GRAND\hat{\mathbf{c}}^{\mathrm{GRAND}}, and a reliability metric for each bit in 𝐜^GRAND\hat{\mathbf{c}}^{\mathrm{GRAND}}, accumulated in the soft-decoding LLR vector, 𝚲¯GRAND\bar{\mathbf{\Lambda}}^{\mathrm{GRAND}}, all while recovering the decoded output, 𝐜¯GRAND\bar{\mathbf{c}}^{\mathrm{GRAND}}.

Algorithm 2 illustrates turbo-GRAND in more detail. For detection, we keep track of an ML-detection distance metric, dMLd^{\mathrm{ML}}; for decoding, we keep track of an ML-decoding distance metric, d¯ML\bar{d}^{\mathrm{ML}}. After each turbo-GRAND iteration, the candidate vector corresponding to dMLd^{\mathrm{ML}} is the updated detected vector, 𝐜^GRAND\hat{\mathbf{c}}^{\mathrm{GRAND}}, and the candidate vector corresponding to d¯ML\bar{d}^{\mathrm{ML}} is the updated decoded vector, 𝐜¯GRAND\bar{\mathbf{c}}^{\mathrm{GRAND}}. Furthermore, for LLR computations, we accumulate a vector of counter-ML-detection distance metrics, 𝐝cML=[d1,1cMLdi,jcMLdM,qcML]TN\mathbf{d}^{\mathrm{cML}}\!=\![d^{\mathrm{cML}}_{1,1}\cdots d^{\mathrm{cML}}_{i,j}\cdots d^{\mathrm{cML}}_{M,q}]^{T}\!\in\!\mbox{\bb R}^{N}, which tracks vectors closest to 𝐜^GRAND\hat{\mathbf{c}}^{\mathrm{GRAND}}, but with bit-flips at corresponding indices. Searching the entire lattice 𝒳¯\bar{\mathcal{X}} results in dML=min𝐱𝒳¯𝐲𝐇𝐱2d^{\mathrm{ML}}\!=\!\min_{\mathbf{x}\in\bar{\mathcal{X}}}\left\|{\mathbf{y}-\mathbf{H}\mathbf{x}}\right\|^{2}; we can re-express (3) as

λi,jML{1σ2dML1σ2min𝐱𝒳¯i,j,0𝐲𝐇𝐱2if c^i,jML=11σ2min𝐱𝒳¯i,j,1𝐲𝐇𝐱21σ2dMLif c^i,jML=0,\lambda_{i,j}^{\mathrm{ML}}\approx\begin{cases}\frac{1}{\sigma^{2}}d^{\mathrm{ML}}-\frac{1}{\sigma^{2}}\min_{\mathbf{x}\in\bar{\mathcal{X}}^{i,j,0}}{\left\|{\mathbf{y}-\mathbf{H}\mathbf{x}}\right\|^{2}}&\text{if $\hat{c}^{\mathrm{ML}}_{i,j}=1$}\\ \frac{1}{\sigma^{2}}\min_{\mathbf{x}\in\bar{\mathcal{X}}^{i,j,1}}{\left\|{\mathbf{y}-\mathbf{H}\mathbf{x}}\right\|^{2}}-\frac{1}{\sigma^{2}}d^{\mathrm{ML}}&\text{if $\hat{c}^{\mathrm{ML}}_{i,j}=0$},\end{cases} (5)

where di,jcML=min𝐱𝒳¯i,j,0𝐲𝐇𝐱2d^{\mathrm{cML}}_{i,j}=\min_{\mathbf{x}\in\bar{\mathcal{X}}^{i,j,0}}{\left\|{\mathbf{y}-\mathbf{H}\mathbf{x}}\right\|^{2}} if c^i,jML=1\hat{c}^{\mathrm{ML}}_{i,j}=1 and di,jcML=min𝐱𝒳¯i,j,1𝐲𝐇𝐱2d^{\mathrm{cML}}_{i,j}=\min_{\mathbf{x}\in\bar{\mathcal{X}}^{i,j,1}}{\left\|{\mathbf{y}-\mathbf{H}\mathbf{x}}\right\|^{2}} if c^i,jML=0\hat{c}^{\mathrm{ML}}_{i,j}=0 (note that colored noise scaling is embedded into 𝐝ML\mathbf{d}^{\mathrm{ML}} and 𝐝cML\mathbf{d}^{\mathrm{cML}} in Alg. 2). However, turbo-GRAND only searches a limited number of 𝐱¯\bar{\mathbf{x}} vectors that are extracted from the recovered 𝐜¯=𝐜^𝐰\bar{\mathbf{c}}\!=\!\hat{\mathbf{c}}\ominus\mathbf{w} words via guesswork, and that can be accumulated in a set 𝒮\mathcal{S} (|𝒮|<T×B|𝒳¯|)\left(\left|{\mathcal{S}}\right|<T\!\times\!B\!\ll\!\left|{\bar{\mathcal{X}}}\right|\right).

We first initialize dMLd^{\mathrm{ML}} to \infty and 𝐝cML\mathbf{d}^{\mathrm{cML}} to saturation noise-scaled thresholds. Then, dMLd^{\mathrm{ML}} and 𝐝cML\mathbf{d}^{\mathrm{cML}} are updated iteratively, upon every new noise guess (up to BB guesses per iteration tt in GRAND with abandonment). For every guessed word, 𝐜¯\bar{\mathbf{c}}, we re-generate a modulated vector, 𝐱¯=mod(𝐜¯)\bar{\mathbf{x}}\!=\!\text{mod}\left(\bar{\mathbf{c}}\right), and add it to 𝒮\mathcal{S}. Hence, for turbo-GRAND, dML=min𝐱¯𝒮𝐲𝐇𝐱¯2d^{\mathrm{ML}}\!=\!\min_{\bar{\mathbf{x}}\in\mathcal{S}}\left\|{\mathbf{y}-\mathbf{H}\bar{\mathbf{x}}}\right\|^{2}, and

λ¯i,jGRAND{1σ2dML1σ2min𝐱¯𝒮i,j,0𝐲𝐇𝐱¯2if c^i,jGRAND=11σ2min𝐱¯𝒮i,j,1𝐲𝐇𝐱¯21σ2dMLif c^i,jGRAND=0,\bar{\lambda}_{i,j}^{\mathrm{GRAND}}\!\approx\!\begin{cases}\frac{1}{\sigma^{2}}d^{\mathrm{ML}}\!-\!\frac{1}{\sigma^{2}}\min_{\bar{\mathbf{x}}\in\mathcal{S}^{i,j,0}}{\left\|{\mathbf{y}\!-\!\mathbf{H}\bar{\mathbf{x}}}\right\|^{2}}&\text{if $\hat{c}^{\mathrm{GRAND}}_{i,j}\!=\!1$}\\ \frac{1}{\sigma^{2}}\min_{\bar{\mathbf{x}}\in\mathcal{S}^{i,j,1}}{\left\|{\mathbf{y}\!-\!\mathbf{H}\bar{\mathbf{x}}}\right\|^{2}}\!-\!\frac{1}{\sigma^{2}}d^{\mathrm{ML}}&\text{if $\hat{c}^{\mathrm{GRAND}}_{i,j}\!=\!0$},\end{cases} (6)

where di,jcML=min𝐱¯𝒮i,j,0𝐲𝐇𝐱¯2d^{\mathrm{cML}}_{i,j}\!=\!\min_{\bar{\mathbf{x}}\in\mathcal{S}^{i,j,0}}{\left\|{\mathbf{y}-\mathbf{H}\bar{\mathbf{x}}}\right\|^{2}} if c^i,jGRAND=1\hat{c}^{\mathrm{GRAND}}_{i,j}\!=\!1 and di,jcML=min𝐱¯𝒮i,j,1𝐲𝐇𝐱¯2d^{\mathrm{cML}}_{i,j}\!=\!\min_{\bar{\mathbf{x}}\in\mathcal{S}^{i,j,1}}{\left\|{\mathbf{y}-\mathbf{H}\bar{\mathbf{x}}}\right\|^{2}} if c^i,jGRAND=0\hat{c}^{\mathrm{GRAND}}_{i,j}\!=\!0. For the i,j{i,j} indices where di,jcMLd^{\mathrm{cML}}_{i,j} cannot be computed over 𝒮\mathcal{S}, owing to the absence of a corresponding 𝐱¯\bar{\mathbf{x}}, the initial saturated value of N/σi2N/\sigma_{i}^{2} is retained. Note that we use a separate d¯ML\bar{d}^{\mathrm{ML}} for the decoded 𝐜¯GRAND\bar{\mathbf{c}}^{\mathrm{GRAND}} because not every 𝐱¯\bar{\mathbf{x}} in 𝒮\mathcal{S} corresponds to a code-word. Each turbo-GRAND iteration tt ends with a single code-word hit 𝐜¯GRAND(t)\bar{\mathbf{c}}^{\mathrm{GRAND}}(t); the final decoded vector after TT iterations is

𝐜¯GRAND=argmin𝐜¯GRAND(t);t{1,,T}𝐲𝐇mod(𝐜¯GRAND(t))2.\bar{\mathbf{c}}^{\mathrm{GRAND}}=\operatorname*{arg\,min}_{\bar{\mathbf{c}}^{\mathrm{GRAND}}(t);\ t\in\{1,\cdots,T\}}\left\|{\mathbf{y}-\mathbf{H}\ \text{mod}\left(\bar{\mathbf{c}}^{\mathrm{GRAND}}(t)\right)}\right\|^{2}. (7)

Note that Algorithm 2 does not explicitly construct 𝒮\mathcal{S} and store it in memory for post-processing but instead updates the ML and counter-ML distance metrics on the fly.

The core of turbo-GRAND is a soft-input decoding mechanism which rank-orders candidate noise sequences according to 𝚲¯GRAND\bar{\mathbf{\Lambda}}^{\mathrm{GRAND}}. Let 𝐖𝔽22N×N\mathbf{W}\!\in\!\mbox{\bb F}_{2}^{2^{N}\times N} be a matrix containing in its rows all possible noise sequences, and let Π´:2N{1,,2N}2N\acute{\Pi}\!:\!\mbox{\bb R}^{2^{N}}\rightarrow\{1,\cdots,2^{N}\}^{2^{N}} be a sorting function (increasing order). Then, 𝐬=Π´(𝐖×|𝚲|)\mathbf{s}=\acute{\Pi}\left(\mathbf{W}\times\left|{\mathbf{\Lambda}}\right|\right) is a vector of sorted noise-sequence indices, and 𝐰=Π(𝐬(k))\mathbf{w}=\Pi(\mathbf{s}(k)) retrieves the kth{k}{\text{th}} likely noise sequence. However, populating noise sequences in a single matrix is not hardware-friendly nor computationally efficient. Alternatively, SGRAND [17] recursively constructs a max-heap for each combination of reliabilities in 𝚲\mathbf{\Lambda} to dynamically generate 𝐰\mathbf{w} vectors with increasing likelihoods. Also, ORBGRAND [18] builds a bit permutation map based on the decreasing rank order of bit reliability to generate a pre-determined series of putative noise queries. Most probably, we have 𝐬(i)𝐬(j)\mathbf{s}(i)\!\leq\!\mathbf{s}(j) when Pr(𝐰=𝐰i)Pr(𝐰=𝐰j)\mathrm{Pr}(\mathbf{w}\!=\!\mathbf{w}_{i})\!\geq\!\mathrm{Pr}(\mathbf{w}\!=\!\mathbf{w}_{j}). In SGRAND [17], the latter is an “if and only if” condition. Thus, with either SGRAND or ORBGRAND, turbo-GRAND always queries the all-zeros noise sequence first and is biased to give higher priority to noise sequences of smaller Hamming weight, but not necessarily so, depending on the reliability information in 𝚲¯GRAND\bar{\mathbf{\Lambda}}^{\mathrm{GRAND}}. In the absence of soft information or further channel knowledge, noise query follows increasing Hamming weights for a probability of bit flip less than 1/21/2. The latter is captured by an SGRAND core of turbo-GRAND, which reduces to hard-GRAND in the absence of soft information. Hence, SGRAND can be used in the first turbo-GRAND iteration when 𝚲=𝟎N\mathbf{\Lambda}\!=\!\mathbf{0}_{N}. ORBGRAND requires some soft input information to match the performance of SGRAND, so it can be adopted when 𝚲=𝚲ZF\mathbf{\Lambda}\!=\!\mathbf{\Lambda}^{\mathrm{ZF}} in the first iteration. However, ORBGRAND entails more noise guesses on average, so it has the potential to generate richer LLRs.

Turbo-GRAND can be modified to support iterative disjoint detection and decoding, where 𝚲¯GRAND\bar{\mathbf{\Lambda}}^{\mathrm{GRAND}} is fed all the way back to a separate detector. Towards this end, the detector should itself be SISO. For instance, the modified distance metric of a turbo MAP detector can be [25]

φ(𝐱)=𝐲𝐇𝐱2σ2+i=1Nj=1q(2𝐜𝐱(i,j)1)𝚲¯GRAND(i,j),\varphi(\mathbf{x})=-\frac{\left\|{\mathbf{y}-\mathbf{H}\mathbf{x}}\right\|^{2}}{\sigma^{2}}+\sum_{i=1}^{N}\sum_{j=1}^{q}{\left(2\mathbf{c_{x}}(i,j)-1\right)\bar{\mathbf{\Lambda}}^{\mathrm{GRAND}}(i,j)}, (8)

where 𝐜𝐱\mathbf{c_{x}} is the bit representation of 𝐱\mathbf{x}. The a posteriori LLRs can be calculated as

λi,jMAP=(max𝐱𝒳i,j,1φ(𝐱)max𝐱𝒳i,j,0φ(𝐱)).\lambda_{i,j}^{\mathrm{MAP}}=\left(\max_{\mathbf{x}\in\mathcal{X}^{i,j,1}}{\varphi(\mathbf{x})}-\max_{\mathbf{x}\in\mathcal{X}^{i,j,0}}{\varphi(\mathbf{x})}\right). (9)

Disjoint detection and GRAND offers good modularity, where on every detection iteration, the noise can be filtered towards new signal subspaces [26], yielding more efficient GRAND.

Refer to caption
(a) AWGN - SGRAND turbo-GRAND with nil soft input - BPSK.
Refer to caption
(b) Rayleigh - SGRAND turbo-GRAND with nil soft input - BPSK.
Refer to caption
(c) Rayleigh - SGRAND turbo-GRAND with nil soft input - 16QAM.
Refer to caption
(d) Rayleigh - ORBGRAND turbo-GRAND (ZF soft input) - 90%90\% CSI - BPSK.
Figure 2: BLER performance evaluation of the proposed turbo-GRAND schemes with BCH [ ​127,113] ​ codes.

IV Performance and Complexity Evaluation

Following the system model of Sec. II, the block-error rate (BLER) performance of turbo-GRAND in decoding Bose–Chaudhuri–Hocquenghem (BCH) [ ​127,113] ​ codes is compared to reference GRAND/ORBGRAND schemes (ORBGRAND exhibited superior performance and complexity tradeoffs compared to other code-specific decoders [18]). Assuming AWGN and normalized transmit power (SNR of 1/σ21/\sigma^{2}), with the absence of input soft information, Fig. 2a illustrates that SGRAND-based turbo-GRAND generates soft information that matches ML-detection soft information. Turbo-GRAND can outperform ML-detection-based ORBGRAND because: 1) turbo iterations can introduce a list-decoding gain and 2) distance computations are over an entire code-word of many channel uses; feasible ML detection search routines only span a few channel uses. A similar relative performance is noted under Rayleigh fading in Fig. 2b and Fig. 2c, using binary phase-shift keying (BPSK) and 16-quadrature amplitude modulation (16-QAM) (Gray mapping), respectively. The gains in soft-GRAND schemes are larger under fading (more than 6dB6\,\mathrm{dB} SNR gains at a BLER of 10210^{-2}), highlighting the importance of turbo-GRAND in the absence of soft information. We further note that the turbo-GRAND gains are captured in two iterations, beyond which diminishing returns are expected, as both SGRAND and ORBGRAND converge faster on every new iteration, tt, guessing over an enhanced initial vector, 𝐜^GRAND(t)\hat{\mathbf{c}}^{\mathrm{GRAND}}(t). The achievable gains of turbo-GRAND can be improved by tuning a fixed guess budget, BB. Alternatively, some sort of reactive taboo search [27] can be adopted, in which every iteration starts with a pseudo-random initial vector upon which noise is guessed.

Several communication system scenarios further highlight the turbo-GRAND gains, especially under symbol interference in spatial/path diversity schemes, where ML soft information is significantly better than ZF soft information. In such scenarios, starting from ZF soft information, ORBGRAND-based turbo-GRAND can bridge the gap to the much more complex ML-detection-based soft-GRAND (this paper only covers uncorrelated point-to-point channels). In another scenario, under imperfect CSI, joint detection and decoding in turbo-GRAND outperforms conventional soft decoding. We assume 10%10\% CSI error in Fig. 2d, where 𝐇err=0.9𝐇+0.1𝐇~\mathbf{H}_{\text{err}}\!=\!0.9\mathbf{H}\!+\!0.1\tilde{\mathbf{H}}, and 𝐇~\tilde{\mathbf{H}} has the same distribution as 𝐇\mathbf{H} but is independently and randomly generated. Starting from ZF soft information as input LLRs, ORBGRAND-based turbo-GRAND outperforms both ML-soft- and ZF-soft-ORBGRAND. Note, however, that our ML-detection implementation in these simulations only undergoes an exhaustive search over subsets of symbols in 𝐱\mathbf{x} of size four.

We next analyze the complexity in terms of floating-point operations in complex multiplication (CMT\mathrm{CMT}) and complex addition (CAD\mathrm{CAD}). We compare the additional processing in turbo-GRAND over soft-GRAND to the complexity of soft-output ZF and ML detection. The search complexities are dominated by Euclidian distance computations and the complex matrix multiplications they entail. The complexity is exponential (in the symbol-vector length, MM) with ML detection ((3) and (5)), polynomial with turbo-GRAND (6), and linear with ZF (4).

Table I illustrates the approximate worst-case complexity of generating soft information in one channel use. The average complexity of turbo-GRAND is much less because the guess budget, BB, is not exploited on every iteration; with more iteration and higher SNR, a few or a single guess can recover a code-word. Turbo-GRAND is thus much less complex than conventional iterative list-based detection and decoding. Even with larger guess budgets, the recovered words on different iterations often overlap, and redundant computations can be saved. Furthermore, because noise-sequence guessing typically follows increasing Hamming weights, the vector Euclidean distance computations can reduce to simple symbol-based scalar distance computations/updates, resulting in an average linear turbo-GRAND complexity of T×B(MCMT+MCAD)T\times B\left(M\mathrm{CMT}+M\mathrm{CAD}\right). Hence, for large modulation orders, a hardware-optimized turbo-GRAND can even prove to be less complex than soft-output ZF detection followed by soft-GRAND; much less complex than ML detection. Further simplifications can be made if the channel remains static over multiple uses.

TABLE I: Soft-output data detection complexity comparison
Detector Complexity
ML |𝒳|M((M2+M)CMT+(M2+M)CAD)\left|{\mathcal{X}}\right|^{M}\left((M^{2}\!+\!M)\mathrm{CMT}+(M^{2}\!+\!M)\mathrm{CAD}\right)
ZF |𝒳|(MCMT+MCAD)\left|{\mathcal{X}}\right|\left(M\mathrm{CMT}+M\mathrm{CAD}\right)
turbo-GRAND T×B((M2+M)CMT+(M2+M)CAD)T\times B\left((M^{2}\!+\!M)\mathrm{CMT}+(M^{2}\!+\!M)\mathrm{CAD}\right)

V Conclusions

We proposed a mechanism for using GRAND to extract soft information in a joint detection and decoding framework. By leveraging access to complex received symbols, hard demapped bits, CSI, and possibly noise statistics, we generate LLRs by populating a shortlist of Euclidean distance computations. Such LLRs can be used in subsequent soft-decoding GRAND iterations, giving rise to turbo-GRAND. Compared to hard GRAND, a few iterations of turbo-GRAND introduce an excess of 6dB6\,\mathrm{dB} SNR gains at a BLER of 10210^{-2} (much higher gains at lower BLERs), under practical communication system scenarios of Rayleigh fading channels. Furthermore, turbo-GRAND can match and even outperform exhaustive ML-detection-based soft-GRAND at a much-reduced average linear complexity. This work can extend into a generic joint detection and decoding framework for future investigations.

References

  • [1] N. Rajatheva, I. Atzeni, E. Bjornson, A. Bourdoux, S. Buzzi, J.-B. Dore, S. Erkucuk, M. Fuentes, K. Guan et al., “White paper on broadband connectivity in 6G,” arXiv preprint arXiv:2004.14247, 2020.
  • [2] H. Sarieddeen, M.-S. Alouini, and T. Y. Al-Naffouri, “An overview of signal processing techniques for terahertz communications,” Proceedings of the IEEE, vol. 109, no. 10, pp. 1628–1665, 2021.
  • [3] G. Durisi, T. Koch, and P. Popovski, “Toward massive, ultrareliable, and low-latency wireless communication with short packets,” Proceedings of the IEEE, vol. 104, no. 9, pp. 1711–1726, 2016.
  • [4] D. Gazelle and J. Snyders, “Reliability-based code-search algorithms for maximum-likelihood decoding of block codes,” IEEE Trans. Inf. Theory, vol. 43, no. 1, pp. 239–249, 1997.
  • [5] A. Valembois and M. Fossorier, “Box and match techniques applied to soft-decision decoding,” IEEE Trans. Inf. Theory, vol. 50, no. 5, pp. 796–810, 2004.
  • [6] K. R. Duffy, J. Li, and M. Médard, “Guessing noise, not code-words,” in Proc. IEEE Int. Symp. Inf. Theory (ISIT), 2018, pp. 671–675.
  • [7] ——, “Capacity-achieving guessing random additive noise decoding,” IEEE Trans. Inf. Theory, vol. 65, no. 7, pp. 4023–4040, 2019.
  • [8] M. M. Christiansen and K. R. Duffy, “Guesswork, large deviations, and shannon entropy,” IEEE Trans. Inf. Theory, vol. 59, no. 2, pp. 796–802, 2013.
  • [9] A. Beirami, R. Calderbank, M. M. Christiansen, K. R. Duffy, and M. Médard, “A characterization of guesswork on swiftly tilting curves,” IEEE Trans. Inf. Theory, vol. 65, no. 5, pp. 2850–2871, 2019.
  • [10] E. Arikan, “An inequality on guessing and its application to sequential decoding,” IEEE Trans. Inf. Theory, vol. 42, no. 1, pp. 99–105, 1996.
  • [11] S. M. Abbas, T. Tonnellier, F. Ercan, and W. J. Gross, “High-throughput VLSI architecture for GRAND,” in Proc. IEEE Workshop on Signal Process. Syst., 2020, pp. 1–6.
  • [12] A. Riaz, V. Bansal, A. Solomon, W. An, Q. Liu, K. Galligan, K. R. Duffy, M. Medard, and R. T. Yazicigil, “Multi-code multi-rate universal maximum likelihood decoder using GRAND,” in IEEE 47th European Solid State Circuits Conf. (ESSCIRC), 2021, pp. 239–246.
  • [13] T. Kaneko, T. Nishijima, and S. Hirasawa, “An improvement of soft-decision maximum-likelihood decoding algorithm using hard-decision bounded-distance decoding,” IEEE Trans. Inf. Theory, vol. 43, no. 4, pp. 1314–1319, 1997.
  • [14] M. Fossorier and S. Lin, “Soft-decision decoding of linear block codes based on ordered statistics,” IEEE Trans. Inf. Theory, vol. 41, no. 5, pp. 1379–1396, 1995.
  • [15] K. R. Duffy and M. Médard, “Guessing random additive noise decoding with soft detection symbol reliability information - SGRAND,” in Proc. IEEE Int. Symp. Inf. Theory (ISIT), 2019, pp. 480–484.
  • [16] S. M. Abbas, M. Jalaleddine, and W. J. Gross, “GRAND for Rayleigh fading channels,” arXiv preprint arXiv:2205.00030, 2022.
  • [17] A. Solomon, K. R. Duffy, and M. Médard, “Soft maximum likelihood decoding using GRAND,” in Proc. IEEE Int. Conf. Commun. (ICC), 2020, pp. 1–6.
  • [18] K. R. Duffy, “Ordered reliability bits guessing random additive noise decoding,” in Proc. IEEE Int. Conf. Acoustics, Speech, and Signal Process. (ICASSP), 2021, pp. 8268–8272.
  • [19] C. Berrou and A. Glavieux, “Near optimum error correcting coding and decoding: turbo-codes,” IEEE Trans. Commun., vol. 44, no. 10, pp. 1261–1271, 1996.
  • [20] A. Tomasoni, M. Siti, M. Ferrari, and S. Bellini, “Low complexity, quasi-optimal MIMO detectors for iterative receivers,” IEEE Trans. Commun., vol. 9, no. 10, pp. 3166–3177, 2010.
  • [21] W. An, M. Médard, and K. R. Duffy, “Keep the bursts and ditch the interleavers,” in Proc. IEEE Global Conmun. Conf. (GLOBECOM), 2020, pp. 1–6.
  • [22] M. Ivanov, C. Häger, F. Brännström, A. G. i Amat, A. Alvarado, and E. Agrell, “On the information loss of the max-log approximation in BICM systems,” IEEE Trans. Inf. Theory, vol. 62, no. 6, pp. 3011–3025, 2016.
  • [23] H. Sarieddeen, M. M. Mansour, and A. Chehab, “Large MIMO detection schemes based on channel puncturing: Performance and complexity analysis,” IEEE Trans. Commun., no. 99, pp. 1–1, Dec. 2017.
  • [24] C. Studer, S. Fateh, and D. Seethaler, “ASIC implementation of soft-input soft-output MIMO detection using MMSE parallel interference cancellation,” IEEE J. Solid-State Circuits, vol. 46, no. 7, pp. 1754–1765, 2011.
  • [25] A. Tomasoni, M. Siti, M. Ferrari, and S. Bellini, “Hardware oriented, quasi-optimal detectors for iterative and non-iterative MIMO receivers,” EURASIP Journal on Wireless Communications and Networking, vol. 2012, no. 1, p. 62, 2012.
  • [26] H. Sarieddeen, M. M. Mansour, and A. Chehab, “Channel-punctured large MIMO detection,” in Proc. IEEE Int. Symp. Inf. Theory (ISIT), Jun. 2018, pp. 2147–2151.
  • [27] T. Datta, N. Srinidhi, A. Chockalingam, and B. S. Rajan, “Random-restart reactive tabu search algorithm for detection in large-MIMO systems,” IEEE Commun. Lett., vol. 14, no. 12, pp. 1107–1109, Dec. 2010.