Global optimality of softmax policy gradient with single hidden layer neural networks in the mean-field regime
Publication
, Journal Article
Agazzi, A; Lu, J
October 22, 2020
We study the problem of policy optimization for infinite-horizon discounted Markov Decision Processes with softmax policy and nonlinear function approximation trained with policy gradient algorithms. We concentrate on the training dynamics in the mean-field regime, modeling e.g., the behavior of wide single hidden layer neural networks, when exploration is encouraged through entropy regularization. The dynamics of these models is established as a Wasserstein gradient flow of distributions in parameter space. We further prove global optimality of the fixed points of this dynamics under mild conditions on their initialization.
Duke Scholars
Publication Date
October 22, 2020
Citation
APA
Chicago
ICMJE
MLA
NLM
Agazzi, A., & Lu, J. (2020). Global optimality of softmax policy gradient with single hidden layer
neural networks in the mean-field regime.
Agazzi, Andrea, and Jianfeng Lu. “Global optimality of softmax policy gradient with single hidden layer
neural networks in the mean-field regime,” October 22, 2020.
Agazzi A, Lu J. Global optimality of softmax policy gradient with single hidden layer
neural networks in the mean-field regime. 2020 Oct 22;
Agazzi, Andrea, and Jianfeng Lu. Global optimality of softmax policy gradient with single hidden layer
neural networks in the mean-field regime. Oct. 2020.
Agazzi A, Lu J. Global optimality of softmax policy gradient with single hidden layer
neural networks in the mean-field regime. 2020 Oct 22;
Publication Date
October 22, 2020