Papers
Topics
Authors
Recent
Search
2000 character limit reached

PACE: A Pragmatic Agent for Enhancing Communication Efficiency Using Large Language Models

Published 30 Jan 2024 in cs.CL and cs.AI | (2402.01750v1)

Abstract: Current communication technologies face limitations in terms of theoretical capacity, spectrum availability, and power resources. Pragmatic communication, leveraging terminal intelligence for selective data transmission, offers resource conservation. Existing research lacks universal intention resolution tools, limiting applicability to specific tasks. This paper proposes an image pragmatic communication framework based on a Pragmatic Agent for Communication Efficiency (PACE) using LLMs (LLM). In this framework, PACE sequentially performs semantic perception, intention resolution, and intention-oriented coding. To ensure the effective utilization of LLM in communication, a knowledge base is designed to supplement the necessary knowledge, dedicated prompts are introduced to facilitate understanding of pragmatic communication scenarios and task requirements, and a chain of thought is designed to assist in making reasonable trade-offs between transmission efficiency and cost. For experimental validation, this paper constructs an image pragmatic communication dataset along with corresponding evaluation standards. Simulation results indicate that the proposed method outperforms traditional and non-LLM-based pragmatic communication in terms of transmission efficiency.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (40)
  1. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
  2. End-to-end object detection with transformers. In European conference on computer vision, pages 213–229. Springer, 2020.
  3. Per-pixel classification is not all you need for semantic segmentation. Advances in Neural Information Processing Systems, 34:17864–17875, 2021.
  4. 6g wireless communication systems: Applications, requirements, technologies, challenges, and research directions. IEEE Open Journal of the Communications Society, 1:957–975, 2020.
  5. Faithful reasoning using large language models. arXiv preprint arXiv:2208.14271, 2022.
  6. Chatlaw: Open-source legal large language model with integrated external knowledge bases. arXiv preprint arXiv:2306.16092, 2023.
  7. Beyond transmitting bits: Context, semantics, and task-oriented communications. IEEE Journal on Selected Areas in Communications, 41(1):5–41, 2022.
  8. Framing image description as a ranking task: Data, models and evaluation metrics. Journal of Artificial Intelligence Research, 47:853–899, 2013.
  9. Sionna: An open-source library for next-generation physical layer research. arXiv preprint, Mar. 2022.
  10. A robust and opponent-aware league training method for starcraft ii. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
  11. Deep joint source-channel coding for wireless image retrieval. In ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 5070–5074. IEEE, 2020.
  12. Task-oriented image transmission for scene classification in unmanned aerial systems. IEEE Transactions on Communications, 70(8):5181–5192, 2022.
  13. Large language models are zero-shot reasoners. Advances in neural information processing systems, 35:22199–22213, 2022.
  14. Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation. In International Conference on Machine Learning, pages 12888–12900. PMLR, 2022.
  15. Microsoft coco: Common objects in context. In European conference on computer vision, pages 740–755. Springer, 2014.
  16. Generative agents: Interactive simulacra of human behavior. In Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology, pages 1–22, 2023.
  17. Flickr30k entities: Collecting region-to-phrase correspondences for richer image-to-sentence models. In Proceedings of the IEEE international conference on computer vision, pages 2641–2649, 2015.
  18. Learning transferable visual models from natural language supervision. In International conference on machine learning, pages 8748–8763. PMLR, 2021.
  19. Design of low-density parity check codes for 5g new radio. IEEE Communications Magazine, 56(3):28–34, 2018.
  20. Bo Rong. 6g: The next horizon: From connected people and things to connected intelligence. IEEE Wireless Communications, 28(5):8–8, 2021.
  21. Hugginggpt: Solving ai tasks with chatgpt and its friends in huggingface. arXiv preprint arXiv:2303.17580, 2023.
  22. From semantic communication to semantic-aware networking: Model, architecture, and open problems. IEEE Communications Magazine, 59(8):44–50, 2021.
  23. Mededit: Model editing for medical question answering with external knowledge bases. arXiv preprint arXiv:2309.16035, 2023.
  24. 6g networks: Beyond shannon towards semantic and goal-oriented communications. Computer Networks, 190:107930, 2021.
  25. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, 2023.
  26. Context-aware effective communications. In 2021 55th Asilomar Conference on Signals, Systems, and Computers, pages 334–339. IEEE, 2021.
  27. Effective communications: A joint learning and communication framework for multi-agent reinforcement learning over noisy channels. IEEE Journal on Selected Areas in Communications, 39(8):2590–2603, 2021.
  28. Lopa J Vora. Evolution of mobile generation technology: 1g to 5g and review of upcoming wireless technology 5g. International journal of modern trends in engineering and research, 2(10):281–290, 2015.
  29. Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing, 13(4):600–612, 2004.
  30. Minilm: Deep self-attention distillation for task-agnostic compression of pre-trained transformers. Advances in Neural Information Processing Systems, 33:5776–5788, 2020.
  31. Self-consistency improves chain of thought reasoning in language models. arXiv preprint arXiv:2203.11171, 2022.
  32. Voyager: An open-ended embodied agent with large language models. arXiv preprint arXiv:2305.16291, 2023.
  33. Warren Weaver. Recent contributions to the mathematical theory of communication. ETC: a review of general semantics, pages 261–281, 1953.
  34. A prompt pattern catalog to enhance prompt engineering with chatgpt. arXiv preprint arXiv:2302.11382, 2023.
  35. The rise and potential of large language model based agents: A survey. arXiv preprint arXiv:2309.07864, 2023.
  36. Task-oriented multi-user semantic communications for vqa. IEEE Wireless Communications Letters, 11(3):553–557, 2021.
  37. Semantic communications with ai tasks. arXiv preprint arXiv:2109.14170, 2021.
  38. Superimposed semantic communication for iot-based real-time ecg monitoring. IEEE Journal of Biomedical and Health Informatics, pages 1–12, 2024.
  39. Glm-130b: An open bilingual pre-trained model. In The Eleventh International Conference on Learning Representations, 2022.
  40. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 586–595, 2018.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.