On The Convergence Rate Of Entropy-regularized Natural Policy Gradient With Linear Function
Presenter
August 3, 2021
Abstract
We study the convergence rate of entropy-regularized Natural Policy Gradient (NPG) algorithms with linear function approximation. We show that NPG exhibits O(1/T) within an approximation error under mild assumptions on the distribution mismatch and the representation power of the feature vectors, and linear convergence under stronger assumptions. Joint work with Semih Cayci and Niao He.