Zero-Shot CSI Lossless Compressors
- The paper introduces a zero-shot framework that leverages large language models as probability oracles to achieve near-entropy lossless CSI compression without domain-specific fine-tuning.
- It presents a hybrid LM–FM architecture that adaptively partitions quantized CSI features between context-aware autoregressive coding and parallel factorized coding.
- The framework effectively manages rate-distortion-complexity trade-offs, demonstrating competitive compression rates and improved NMSE performance on both real and simulated datasets.
Zero-shot CSI lossless compressors refer to a class of schemes that exploit LLMs as general-purpose, zero-shot probability oracles for arithmetic coding of quantized channel state information (CSI) features, achieving lossless compression rates close to the entropy bound with no CSI-specific fine-tuning. The TCLNet framework exemplifies this approach, combining an LLM-driven, context-aware coding path with a parallel factorized model under adaptive complexity control, and operationalizing lossless compression through prompt engineering and token sequence manipulation (Yang et al., 10 Jan 2026).
1. Hybrid LM–FM Architecture for Lossless CSI Compression
Lossless CSI compression in TCLNet begins with a quantized latent vector , obtained by quantizing the output of a lossy encoder with -bit scalar quantization. The lossless stage integrates two probability models:
- Autoregressive LLM (LM): Models the ASCII-tokenized CSI sequence as an autoregressive process, estimating , where for each position and ASCII symbol .
- Factorized Model (FM): Provides a marginal probability distribution , representing , .
A complexity-controlled symbol-selection module assigns each position to either the LM branch (context-aware) or the FM branch (parallel), partitioning into disjoint index sets and . Separate arithmetic coding is performed on these subsets using the respective probability models, resulting in bitstreams and .
2. Adaptive Switching Between Context-Aware and Parallel Coding
For each quantized symbol index , TCLNet calculates:
- Context-based entropy:
- Marginal (FM) entropy:
- Entropy gain:
A user-defined complexity variable determines the proportion of positions assigned to FM coding. After sorting all indices by in descending order, the most gainful positions are encoded via the LM, and the remaining via the FM. Adjusting enables fine-tuning of the rate-distortion-complexity (RDC) trade-off:
- Rate (): Number of bits per element, .
- Distortion (): Fixed by the lossy module, measured in NMSE.
- Complexity ():
since .
This mechanism allows trade-off between coding efficiency and computational cost, central to deployment in practical systems.
3. Zero-Shot Prompting and Operational Workflow
The key innovation is using general-purpose LLMs, such as ChatGPT-5, in a zero-shot mode with no domain tuning. The process comprises:
- ASCII Tokenization: Each quantized integer () maps bijectively to an ASCII character, producing a token sequence interpretable by the LLM.
- Prompt Engineering: The LLM is instructed via an in-context prompt to "act as a probability estimator," providing, for each presented context, a probability vector over the ASCII vocabulary in JSON format.
1 2 3
System: You are given a sequence of ASCII tokens representing quantized CSI coefficients. For each next token, you must output a probability vector over the entire ASCII vocabulary in JSON array format, without additional commentary. User: Input tokens so far: "Gf#…Z" LLM: [0.001, 0.0005, …, 0.002]
- Compression: At each time , the current context is submitted along with the prompt to the LLM; the returned vector is passed to the arithmetic encoder for the next token.
- Decompression: The arithmetic decoder reconstructs the ASCII sequence by querying the LLM with mirrored context and prompt, synchronizing symbol-by-symbol decoding with the encoder's probability estimates. FM-decoded symbols are merged via the indicator matrix. The ASCII sequence is then detokenized, dequantized as , and forwarded to the lossy CSI decoder.
4. Mathematical Formulation of Compression Modules
The formal framework underpinning zero-shot CSI lossless compression includes:
- Quantization:
- Entropy Measures:
- Entropy Gain:
- Coding Complexity:
- Coding Rate:
where is the total bit count and is the length of the original CSI feature vector.
- Distortion Metric (NMSE in dB):
5. Empirical Performance and Comparative Baselines
Experiments were conducted on both the real Argos 2.4 GHz indoor dataset and the COST2100 simulated channel dataset. The full pipeline included conversion to the angle-delay domain, a lossy encoder operating at $1/128$ compression ratio, -bit quantization (typically ), and final lossless compression.
Comparison metrics: Bits per element () and NMSE in dB.
Baselines:
- Fixed-length coding at 7 bits per symbol ().
- FM-only entropy coding ().
- CSI-trained Transformer decoder LM (), approximating the entropy lower bound ().
Zero-shot LLM results:
| Model | Parameters | (bits/element) | CSI-tuning? |
|---|---|---|---|
| ChatGPT-5 | 1T | $0.0513$ | No |
| ChatGPT-4o | 200B | $0.0515$ | No |
These zero-shot LLM compressors outperform fixed-length and FM-only coding, and without CSI-specific fine-tuning, capture of the achievable entropy-coding gain.
6. Ablation Insights and Impact of Prompt and Model Selection
Systematic ablation studies reveal several key phenomena:
- Complexity Control () Sweep: Decreasing (increasing LM usage) drives toward the entropy bound, but increases linearly with . Figure 1 in (Yang et al., 10 Jan 2026) displays the RDC trade-off for various quantization precisions.
- Swin-Transformer Window Size: Optimal NMSE is achieved at window size $4$ in the lossy module, with performance degrading for smaller windows.
- Prompt Design: Use of in-context demonstrations vs. plain instruction yields a minor difference ( bits/element) in for ChatGPT-5 compared to a domain-trained LM.
- Model Choice: A domain-specific LM with $12.5$M parameters outperforms general-purpose LLMs by bits/element. However, general LLMs still realize over 98% of the maximal entropy-coding gain in zero-shot deployment.
7. Significance and Practical Considerations
Flattening quantized CSI features into ASCII tokens and orchestrating general LLMs as zero-shot probability oracles—augmented by a factorized model and an explicitly tunable complexity parameter—enables near-entropy optimal, lossless CSI coding without customized model training or large-scale retraining. This makes practical, runtime deployment of high-efficiency lossless CSI feedback achievable on commodity hardware with on-device LLM capabilities (Yang et al., 10 Jan 2026). A plausible implication is broad applicability in bandwidth-constrained massive MIMO systems where domain-specific re-training is infeasible or undesirable.