Near-Optimal Regret Bounds for Model-Free RL in Non-Stationary Episodic MDPs
We consider model-free reinforcement learning (RL) in non-stationary Markov decision processes (MDPs). Both the reward functions and the state transition distributions are allowed to vary over time, either gradually or abruptly, as long as their cumulative variation magnitude does not exceed certain budgets. We propose an algorithm, named Restarted Q-Learning with Upper Confidence Bounds (RestartQ-UCB), for this setting, which adopts a simple restarting strategy and an extra optimism term. Our algorithm outperforms the state-of-the-art (model-based) solution in terms of dynamic regret. Specifically, RestartQ-UCB with Freedman-type bonus terms achieves a dynamic regret of O(S^1/3 A^1/3Δ^1/3 H T^2/3), where S and A are the numbers of states and actions, respectively, Δ>0 is the variation budget, H is the number of steps per episode, and T is the total number of steps. We further show that our algorithm is near-optimal by establishing an information-theoretical lower bound of Ω(S^1/3 A^1/3Δ^1/3 H^2/3 T^2/3), which to the best of our knowledge is the first impossibility result in non-stationary RL in general.
READ FULL TEXT