Linear-Quadratic Partially Observed Mean Field Stackelberg Stochastic Differential Game (2503.15803v1)
Abstract: This paper is concerned with a linear-quadratic partially observed mean field Stackelberg stochastic differential game, which contains a leader and a large number of followers. Specifically, the followers confront a large-population Nash game subsequent to the leader's initial announcement of his strategy. In turn, the leader optimizes his own cost functional, taking into account the anticipated reactions of the followers. The state equations of both the leader and the followers are general stochastic differential equations, where the drift terms contain both the state average term and the state expectation term. However, the followers' average state terms enter into the drift term of the leader's state equation and the state expectation term of the leader enters into the state equation of the follower, reflecting the mutual influence between the leader and the followers. By utilizing the techniques of state decomposition and backward separation principle, we deduce the open-loop adapted decentralized strategies and feedback decentralized strategies of this leader-followers system, and demonstrate that the decentralized strategies are the corresponding $\varepsilon$-Stackelberg-Nash equilibrium.