TY - JOUR
T1 - Linear quadratic mean field Stackelberg differential games
AU - Moon, Jun
AU - Başar, Tamer
N1 - Funding Information:
This research was supported in part by the National Research Foundation of Korea (NRF) Grant funded by the Ministry of Science and ICT, South Korea ( NRF-2017R1A5A1015311 ), in part by Institute for Information & communications Technology Promotion (IITP) grant funded by the Korea government (MSIT), South Korea (No. 2018-0-00958 ), and in part by the Office of Naval Research (ONR), USA, MURI Grant N00014-16-1-2710 . The material in this paper was partially presented at the 54th IEEE Conference on Decision and Control, Dec. 15–18, 2015. This paper was recommended for publication in revised form by Associate Editor Dario Bauso under the direction of Editor Ian R. Petersen.
Publisher Copyright:
© 2018 Elsevier Ltd
Copyright:
Copyright 2018 Elsevier B.V., All rights reserved.
PY - 2018/11
Y1 - 2018/11
N2 - We consider linear–quadratic mean field Stackelberg differential games with the adapted open-loop information structure of the leader. There are one leader and N followers, where N is arbitrarily large. The leader holds a dominating position in the game in the sense that the leader first chooses and then announces the optimal strategy to which the N followers respond by playing a Nash game, i.e., choosing their optimal strategies noncooperatively and simultaneously based on the leader's observed strategy. In our setting, the followers are coupled with each other through the mean field term included in their cost functions, and are strongly influenced by the leader's open-loop strategy included in their cost functions and dynamics. From the leader's perspective, he is coupled with the N followers through the mean field term included in his cost function. To circumvent the complexity brought about by the coupling nature among the leader and the followers with large N, which makes the use of the direct approach almost impossible, our approach in this paper is to characterize an approximated stochastic mean field process by solving a local optimal control problem of the followers with leader's control taken as an exogenous stochastic process. We show that for each fixed strategy of the leader, the followers’ local optimal decentralized strategies lead to an ϵ-Nash equilibrium. The paper then solves the leader's local optimal control problem, as a nonstandard constrained optimization problem, with constraints being induced by the approximated mean field process determined by Nash followers (which also depend on the leader's control). We show that the local optimal decentralized controllers for the leader and the followers constitute an (ϵ1,ϵ2)-Stackelberg–Nash equilibrium for the original game, where ϵ1 and ϵ2 both converge to zero as N→∞. Numerical examples are provided to illustrate the theoretical results.
AB - We consider linear–quadratic mean field Stackelberg differential games with the adapted open-loop information structure of the leader. There are one leader and N followers, where N is arbitrarily large. The leader holds a dominating position in the game in the sense that the leader first chooses and then announces the optimal strategy to which the N followers respond by playing a Nash game, i.e., choosing their optimal strategies noncooperatively and simultaneously based on the leader's observed strategy. In our setting, the followers are coupled with each other through the mean field term included in their cost functions, and are strongly influenced by the leader's open-loop strategy included in their cost functions and dynamics. From the leader's perspective, he is coupled with the N followers through the mean field term included in his cost function. To circumvent the complexity brought about by the coupling nature among the leader and the followers with large N, which makes the use of the direct approach almost impossible, our approach in this paper is to characterize an approximated stochastic mean field process by solving a local optimal control problem of the followers with leader's control taken as an exogenous stochastic process. We show that for each fixed strategy of the leader, the followers’ local optimal decentralized strategies lead to an ϵ-Nash equilibrium. The paper then solves the leader's local optimal control problem, as a nonstandard constrained optimization problem, with constraints being induced by the approximated mean field process determined by Nash followers (which also depend on the leader's control). We show that the local optimal decentralized controllers for the leader and the followers constitute an (ϵ1,ϵ2)-Stackelberg–Nash equilibrium for the original game, where ϵ1 and ϵ2 both converge to zero as N→∞. Numerical examples are provided to illustrate the theoretical results.
KW - Forward–backward stochastic differential equations
KW - Mean field theory
KW - Stackelberg and Nash games
KW - Stochastic optimal control
UR - http://www.scopus.com/inward/record.url?scp=85051821593&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85051821593&partnerID=8YFLogxK
U2 - 10.1016/j.automatica.2018.08.008
DO - 10.1016/j.automatica.2018.08.008
M3 - Article
AN - SCOPUS:85051821593
VL - 97
SP - 200
EP - 213
JO - Automatica
JF - Automatica
SN - 0005-1098
ER -