Reduce FRQD-learning communication overhead without sacrificing convergence guarantees
Develop a communication-efficient variant of the Fully Resilient QD (FRQD)-learning algorithm that reduces the worst-case per-agent communication complexity from O(|N_i(t)|^2) per time step while preserving the almost sure convergence to the optimal value functions stated in Theorem 1 under F-total Byzantine edge attacks on (6F+1,0)-redundant communication graphs.
References
Reducing this overhead while preserving the same convergence guarantees (Theorem 1) remains a future work.
— Fully Byzantine-Resilient Distributed Multi-Agent Q-Learning
(2604.02791 - Lee et al., 3 Apr 2026) in Remark following Algorithm 1, Subsection "Fully Resilient QD (FRQD)-Learning" (Section 3)