pith. machine review for the scientific record. sign in

arxiv: 1803.02348 · v3 · submitted 2018-03-06 · 💻 cs.LG · cs.AI

Recognition: unknown

Smoothed Action Value Functions for Learning Gaussian Policies

Authors on Pith no claims yet
classification 💻 cs.LG cs.AI
keywords smoothedgaussianpolicyq-valuevalueactionalgorithmscovariance
0
0 comments X
read the original abstract

State-action value functions (i.e., Q-values) are ubiquitous in reinforcement learning (RL), giving rise to popular algorithms such as SARSA and Q-learning. We propose a new notion of action value defined by a Gaussian smoothed version of the expected Q-value. We show that such smoothed Q-values still satisfy a Bellman equation, making them learnable from experience sampled from an environment. Moreover, the gradients of expected reward with respect to the mean and covariance of a parameterized Gaussian policy can be recovered from the gradient and Hessian of the smoothed Q-value function. Based on these relationships, we develop new algorithms for training a Gaussian policy directly from a learned smoothed Q-value approximator. The approach is additionally amenable to proximal optimization by augmenting the objective with a penalty on KL-divergence from a previous policy. We find that the ability to learn both a mean and covariance during training leads to significantly improved results on standard continuous control benchmarks.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.