Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Communication-Efficient Edge AI Inference Over Wireless Networks (2004.13351v1)

Published 28 Apr 2020 in eess.SP, cs.IT, cs.LG, and math.IT

Abstract: Given the fast growth of intelligent devices, it is expected that a large number of high-stake AI applications, e.g., drones, autonomous cars, tactile robots, will be deployed at the edge of wireless networks in the near future. As such, the intelligent communication networks will be designed to leverage advanced wireless techniques and edge computing technologies to support AI-enabled applications at various end devices with limited communication, computation, hardware and energy resources. In this article, we shall present the principles of efficient deployment of model inference at network edge to provide low-latency and energy-efficient AI services. This includes the wireless distributed computing framework for low-latency device distributed model inference as well as the wireless cooperative transmission strategy for energy-efficient edge cooperative model inference. The communication efficiency of edge inference systems is further improved by building up a smart radio propagation environment via intelligent reflecting surface.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Kai Yang (187 papers)
  2. Yong Zhou (156 papers)
  3. Zhanpeng Yang (4 papers)
  4. Yuanming Shi (119 papers)
Citations (2)