Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning End-to-End Channel Coding with Diffusion Models (2302.01714v2)

Published 3 Feb 2023 in cs.IT, cs.LG, and math.IT

Abstract: It is a known problem that deep-learning-based end-to-end (E2E) channel coding systems depend on a known and differentiable channel model, due to the learning process and based on the gradient-descent optimization methods. This places the challenge to approximate or generate the channel or its derivative from samples generated by pilot signaling in real-world scenarios. Currently, there are two prevalent methods to solve this problem. One is to generate the channel via a generative adversarial network (GAN), and the other is to, in essence, approximate the gradient via reinforcement learning methods. Other methods include using score-based methods, variational autoencoders, or mutual-information-based methods. In this paper, we focus on generative models and, in particular, on a new promising method called diffusion models, which have shown a higher quality of generation in image-based tasks. We will show that diffusion models can be used in wireless E2E scenarios and that they work as good as Wasserstein GANs while having a more stable training procedure and a better generalization ability in testing.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (20)
  1. “Learning to communicate: Channel auto-encoders, domain specific regularizers, and attention,” in Int. Symp. Signal Process. Inf. Technol. IEEE, 2016, pp. 223–228.
  2. T. O’Shea and J. Hoydis, “An introduction to deep learning for the physical layer,” Trans. Cogn. Commun. and Netw., vol. 3, no. 4, pp. 563–575, 2017.
  3. “Turbo autoencoder: Deep learning based channel codes for point-to-point communication channels,” in Adv. Neural Inf. Process. Syst., Vancouver, Canada, Dec. 2019, pp. 2754–2764.
  4. “ProductAE: Towards training larger channel codes based on neural product codes,” arXiv preprint arXiv:2110.04466, Oct. 2021.
  5. “Channel agnostic end-to-end learning based communication systems with conditional gan,” in Globecom Workshops, Abu Dhabi, United Arab Emirates, 2018, IEEE, pp. 1–5.
  6. “WGAN-based autoencoder training over-the-air,” in Int. Workshop Signal Process. Adv. Wireless Commun., Atlanta, GA, May 2020, IEEE, pp. 1–5.
  7. “Physical layer communications system design over-the-air using adversarial networks,” in Eur. Signal Process. Conf., Rome, Italy, Sep. 2018, pp. 529–532.
  8. F. A. Aoudia and J. Hoydis, “End-to-end learning of communications systems without a channel model,” in Asilomar Conf. Signals, Syst., and Comput., Pacific Grove, CA, Oct. 2018, pp. 298–303.
  9. “Diffusion models in vision: A survey,” arXiv preprint arXiv:2209.04747, Oct. 2022.
  10. “Deep unsupervised learning using non-equilibrium thermodynamics,” in Int. Conf. Mach. Learn., Lille, France, July 2015, PMLR, pp. 2256–2265.
  11. “Denoising diffusion probabilistic models,” Adv. Neural Inf. Process. Syst., vol. 33, pp. 6840–6851, Dec. 2020.
  12. Y. Song and S. Ermon, “Generative modeling by estimating gradients of the data distribution,” Adv. Neural Inf. Process. Syst., vol. 32, Dec. 2019.
  13. “Score-based generative modeling through stochastic differential equations,” in Int. Conf. Learn. Representations, May 2021.
  14. “Residual-aided end-to-end learning of communication system without known channel,” Trans. Cogn. Commun. and Netw., 2022.
  15. “Generative-AI methods for channel impulse response generation,” in Int. Workshop Smart Antennas, Biot, France, Nov. 2021, VDE, pp. 1–6.
  16. M. Arvinte and J. Tamir, “Score-based generative models for wireless channel modeling and estimation,” in Int. Conf. Learn. Representations Workshop, Chicago, IL, Apr. 2022.
  17. “Generative neural network channel modeling for millimeter-wave uav communication,” Trans. Wireless Commun., 2022.
  18. C. Luo, “Understanding diffusion models: A unified perspective,” arXiv preprint arXiv:2208.11970, Aug. 2022.
  19. W. Feller, “On the theory of stochastic processes, with particular reference to applications,” in Sel. Papers I, pp. 769–798. Springer, 2015.
  20. J. R. Siddiqui, “Denoising diffusion model,” \urlhttps://github.com/azad-academy/denoising-diffusion-model.git, 2022.
Citations (12)

Summary

We haven't generated a summary for this paper yet.