Feature Selection for Value Function Approximation Using Bayesian Model Selection

01/31/2012
by   Tobias Jung, et al.
0

Feature selection in reinforcement learning (RL), i.e. choosing basis functions such that useful approximations of the unkown value function can be obtained, is one of the main challenges in scaling RL to real-world applications. Here we consider the Gaussian process based framework GPTD for approximate policy evaluation, and propose feature selection through marginal likelihood optimization of the associated hyperparameters. Our approach has two appealing benefits: (1) given just sample transitions, we can solve the policy evaluation problem fully automatically (without looking at the learning task, and, in theory, independent of the dimensionality of the state space), and (2) model selection allows us to consider more sophisticated kernels, which in turn enable us to identify relevant subspaces and eliminate irrelevant state variables such that we can achieve substantial computational savings and improved prediction performance.

READ FULL TEXT
research
01/16/2019

Representation Learning on Graphs: A Reinforcement Learning Application

In this work, we study value function approximation in reinforcement lea...
research
01/02/2023

Learning to Maximize Mutual Information for Dynamic Feature Selection

Feature selection helps reduce data acquisition costs in ML, but the sta...
research
01/31/2012

Learning RoboCup-Keepaway with Kernels

We apply kernel-based methods to solve the difficult reinforcement learn...
research
10/15/2017

Manifold Regularization for Kernelized LSTD

Policy evaluation or value function or Q-function approximation is a key...
research
06/17/2021

Taming Nonconvexity in Kernel Feature Selection—Favorable Properties of the Laplace Kernel

Kernel-based feature selection is an important tool in nonparametric sta...
research
10/05/2021

Feature Selection by a Mechanism Design

In constructing an econometric or statistical model, we pick relevant fe...
research
08/18/2022

Hybrid Learning with New Value Function for the Maximum Common Subgraph Problem

Maximum Common induced Subgraph (MCS) is an important NP-hard problem wi...

Please sign up or login with your details

Forgot password? Click here to reset