Optimization of Linear Multi-Agent Dynamical Systems via Feedback Distributed Gradient Descent Methods
Abstract: Feedback optimization is an increasingly popular control paradigm to optimize dynamical systems, accounting for control objectives that concern the system operation at steady-state. Existing feedback optimization techniques heavily rely on centralized systems and controller architectures, and thus suffer from scalability and privacy issues when systems become large-scale. In this paper, we propose a distributed architecture for feedback optimization inspired by distributed gradient descent, whereby each agent updates its local control variable by combining the average of its neighbors with a local negative gradient step. Under convexity and smoothness assumptions for the cost, we establish convergence of the control method to a critical optimization point. By reinforcing the assumptions to restricted strong convexity, we show that our algorithm converges linearly to a neighborhood of the optimal point, where the size of the neighborhood depends on the choice of the stepsize. Simulations corroborate the theoretical results.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.