File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
  • Find via Find It@HKUL
Supplementary

Conference Paper: Global Convergence of Policy Gradient for Linear-Quadratic Mean-Field Control/Game in Continuous Time

TitleGlobal Convergence of Policy Gradient for Linear-Quadratic Mean-Field Control/Game in Continuous Time
Authors
Issue Date2021
PublisherML Research Press. The Journal's web site is located at http://proceedings.mlr.press/
Citation
Proceedings of the 38 th International Conference on Machine Learning (ICML 2021), Virtual Conference,18-24 July 2021. In Proceedings of Machine Learning Research, v. 139: Proceedings of ICML 2021, p. 10772-10782 How to Cite?
AbstractRecent years have witnessed the success of multiagent reinforcement learning, which has motivated new research directions for mean-field control (MFC) and mean-field game (MFG), as the multi-agent system can be well approximated by a mean-field problem when the number of agents grows to be very large. In this paper, we study the policy gradient (PG) method for the linearquadratic mean-field control and game, where we assume each agent has identical linear state transitions and quadratic cost functions. While most recent works on policy gradient for MFC and MFG are based on discrete-time models, we focus on a continuous-time model where some of our analyzing techniques could be valuable to the interested readers. For both the MFC and the MFG, we provide PG update and show that it converges to the optimal solution at a linear rate, which is verified by a synthetic simulation. For the MFG, we also provide sufficient conditions for the existence and uniqueness of the Nash equilibrium.
DescriptionPoster Session 3
Persistent Identifierhttp://hdl.handle.net/10722/306031
ISSN

 

DC FieldValueLanguage
dc.contributor.authorWang, W-
dc.contributor.authorHan, J-
dc.contributor.authorYang, Z-
dc.contributor.authorWang, Z-
dc.date.accessioned2021-10-20T10:17:49Z-
dc.date.available2021-10-20T10:17:49Z-
dc.date.issued2021-
dc.identifier.citationProceedings of the 38 th International Conference on Machine Learning (ICML 2021), Virtual Conference,18-24 July 2021. In Proceedings of Machine Learning Research, v. 139: Proceedings of ICML 2021, p. 10772-10782-
dc.identifier.issn2640-3498-
dc.identifier.urihttp://hdl.handle.net/10722/306031-
dc.descriptionPoster Session 3-
dc.description.abstractRecent years have witnessed the success of multiagent reinforcement learning, which has motivated new research directions for mean-field control (MFC) and mean-field game (MFG), as the multi-agent system can be well approximated by a mean-field problem when the number of agents grows to be very large. In this paper, we study the policy gradient (PG) method for the linearquadratic mean-field control and game, where we assume each agent has identical linear state transitions and quadratic cost functions. While most recent works on policy gradient for MFC and MFG are based on discrete-time models, we focus on a continuous-time model where some of our analyzing techniques could be valuable to the interested readers. For both the MFC and the MFG, we provide PG update and show that it converges to the optimal solution at a linear rate, which is verified by a synthetic simulation. For the MFG, we also provide sufficient conditions for the existence and uniqueness of the Nash equilibrium.-
dc.languageeng-
dc.publisherML Research Press. The Journal's web site is located at http://proceedings.mlr.press/-
dc.relation.ispartofProceedings of Machine Learning Research (PMLR)-
dc.relation.ispartofThe 38th International Conference on Machine Learning (ICML 2021)-
dc.titleGlobal Convergence of Policy Gradient for Linear-Quadratic Mean-Field Control/Game in Continuous Time-
dc.typeConference_Paper-
dc.identifier.emailWang, W: weichenw@hku.hk-
dc.identifier.authorityWang, W=rp02849-
dc.identifier.hkuros327594-
dc.identifier.volume139: Proceedings of ICML 2021-
dc.identifier.spage10772-
dc.identifier.epage10782-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats