Papers
Topics
Authors
Recent
Search
2000 character limit reached

GRAND for Fading Channels using Pseudo-soft Information

Published 22 Jul 2022 in cs.IT and math.IT | (2207.10842v2)

Abstract: Guessing random additive noise decoding (GRAND) is a universal maximum-likelihood decoder that recovers code-words by guessing rank-ordered putative noise sequences and inverting their effect until one or more valid code-words are obtained. This work explores how GRAND can leverage additive-noise statistics and channel-state information in fading channels. Instead of computing per-bit reliability information in detectors and passing this information to the decoder, we propose leveraging the colored noise statistics following channel equalization as pseudo-soft information for sorting noise sequences. We investigate the efficacy of pseudo-soft information extracted from linear zero-forcing and minimum mean square error equalization when fed to a hardware-friendly soft-GRAND (ORBGRAND). We demonstrate that the proposed pseudo-soft GRAND schemes approximate the performance of state-of-the-art decoders of CA-Polar and BCH codes that avail of complete soft information. Compared to hard-GRAND, pseudo-soft ORBGRAND introduces up to 10dB SNR gains for a target 10-3 block-error rate.

Citations (14)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.