Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
158 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Power Minimizer Symbol-Level Precoding: A Closed-Form Sub-Optimal Solution (1807.10619v1)

Published 27 Jul 2018 in eess.SP, cs.IT, and math.IT

Abstract: In this letter, we study the optimal solution of the multiuser symbol-level precoding (SLP) for minimization of the total transmit power under given signal-to-interference-plus-noise ratio (SINR) constraints. Adopting the distance preserving constructive interference regions (DPCIR), we first derive a simplified reformulation of the problem. Then, we analyze the structure of the optimal solution using the Karush-Kuhn-Tucker (KKT) optimality conditions, thereby we obtain the necessary and sufficient condition under which the power minimizer SLP is equivalent to the conventional zero-forcing beamforming (ZFBF). This further leads us to a closed-form sub-optimal SLP solution (CF-SLP) for the original problem. Simulation results show that CF-SLP provides significant gains over ZFBF, while performing quite close to the optimal SLP in scenarios with rather small number of users. The results further indicate that the CF-SLP method has a reduction of order $103$ in computational time compared to the optimal solution.

Citations (30)

Summary

We haven't generated a summary for this paper yet.