Papers
Topics
Authors
Recent
Search
2000 character limit reached

Efficient Quantization-Aware Neural Receivers: Beyond Post-Training Quantization

Published 17 Sep 2025 in eess.SP | (2509.13786v1)

Abstract: As wireless communication systems advance toward Sixth Generation (6G) Radio Access Networks (RAN), Deep Learning (DL)-based neural receivers are emerging as transformative solutions for Physical Layer (PHY) processing, delivering superior Block Error Rate (BLER) performance compared to traditional model-based approaches. Practical deployment on resource-constrained hardware, however, requires efficient quantization to reduce latency, energy, and memory without sacrificing reliability. We extend Post-Training Quantization (PTQ) baselines with Quantization-Aware Training (QAT), which incorporates low-precision simulation during training for robustness at ultra-low bitwidths. Our study applies QAT/PTQ to a neural receiver architecture and evaluates across 3GPP Clustered Delay Line (CDL)-B/D channels in LoS and NLoS environments at user velocities up to 40 m/s. Results show that 4-bit and 8-bit QAT models achieve BLERs similar to that of FP32 models at 10% target BLER. QAT models are also shown to outperform PTQ models by up to 3 dB, and yield 8x compression. These results demonstrate that QAT is a key enabler of low-complexity and latency-constrained inference at the PHY layer, facilitating real-time processing in 6G edge devices

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.