Journal article
Policy gradient methods find the Nash equilibrium in N-player general-sum linear-quadratic games
- Abstract:
- We consider a general-sum N-player linear-quadratic game with stochastic dynamics over a finite horizon and prove the global convergence of the natural policy gradient method to the Nash equilibrium. In order to prove convergence of the method we require a certain amount of noise in the system. We give a condition, essentially a lower bound on the covariance of the noise in terms of the model parameters, in order to guarantee convergence. We illustrate our results with numerical experiments to show that even in situations where the policy gradient method may not converge in the deterministic setting, the addition of noise leads to convergence.
- Publication status:
- Published
- Peer review status:
- Peer reviewed
Actions
Access Document
- Files:
-
-
(Preview, Version of record, pdf, 3.9MB, Terms of use)
-
- Publication website:
- http://jmlr.org/papers/v24/21-0842.html
Authors
- Publisher:
- Journal of Machine Learning Research
- Journal:
- Journal of Machine Learning Research More from this journal
- Volume:
- 24
- Issue:
- 139
- Pages:
- 1−56
- Article number:
- 21-0842
- Publication date:
- 2023-04-01
- Acceptance date:
- 2023-03-24
- EISSN:
-
1533-7928
- ISSN:
-
1532-4435
- Language:
-
English
- Keywords:
- Pubs id:
-
1334894
- Local pid:
-
pubs:1334894
- Deposit date:
-
2023-03-29
Terms of use
- Copyright holder:
- Hambly et al.
- Copyright date:
- 2023
- Rights statement:
- © 2023 Ben Hambly, Renyuan Xu and Huining Yang. This is an open access article distributed under the terms of the CC-BY 4.0 License.
- Licence:
- CC Attribution (CC BY)
If you are the owner of this record, you can report an update to it here: Report update to this record