A Policy Gradient Framework for Stochastic Optimal Control Problems with Global Convergence Guarantee

02/11/2023
by   Mo Zhou, et al.
0

In this work, we consider the stochastic optimal control problem in continuous time and a policy gradient method to solve it. In particular, we study the gradient flow for the control, viewed as a continuous time limit of the policy gradient. We prove the global convergence of the gradient flow and establish a convergence rate under some regularity assumptions. The main novelty in the analysis is the notion of local optimal control function, which is introduced to compare the local optimality of the iterate.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/22/2022

Linear convergence of a policy gradient method for finite horizon continuous time stochastic control problems

Despite its popularity in the reinforcement learning community, a provab...
research
06/10/2020

Gradient Flows for Regularized Stochastic Control Problems

This work is motivated by a desire to extend the theoretical underpinnin...
research
03/16/2023

Controlled Descent Training

In this work, a novel and model-based artificial neural network (ANN) tr...
research
01/01/2023

Policy iteration for the deterministic control problems – a viscosity approach

This paper is concerned with the convergence rate of policy iteration fo...
research
06/05/2019

Global Optimality Guarantees For Policy Gradient Methods

Policy gradients methods are perhaps the most widely used class of reinf...
research
03/24/2022

Non-Parametric Stochastic Policy Gradient with Strategic Retreat for Non-Stationary Environment

In modern robotics, effectively computing optimal control policies under...
research
01/17/2022

Optimisation of Structured Neural Controller Based on Continuous-Time Policy Gradient

This study presents a policy optimisation framework for structured nonli...

Please sign up or login with your details

Forgot password? Click here to reset