A Dantzig Selector Approach to Temporal Difference Learning

06/27/2012
by   Matthieu Geist, et al.
0

LSTD is a popular algorithm for value function approximation. Whenever the number of features is larger than the number of samples, it must be paired with some form of regularization. In particular, L1-regularization methods tend to perform feature selection by promoting sparsity, and thus, are well-suited for high-dimensional problems. However, since LSTD is not a simple regression algorithm, but it solves a fixed--point problem, its integration with L1-regularization is not straightforward and might come with some drawbacks (e.g., the P-matrix assumption for LASSO-TD). In this paper, we introduce a novel algorithm obtained by integrating LSTD with the Dantzig Selector. We investigate the performance of the proposed algorithm and its relationship with the existing regularized approaches, and show how it addresses some of their drawbacks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/28/2014

Efficient Regularized Regression for Variable Selection with L0 Penalty

Variable (feature, gene, model, which we use interchangeably) selections...
research
01/31/2023

Adaptive sparseness for correntropy-based robust regression via automatic relevance determination

Sparseness and robustness are two important properties for many machine ...
research
10/16/2012

Value Function Approximation in Noisy Environments Using Locally Smoothed Regularized Approximate Linear Programs

Recently, Petrik et al. demonstrated that L1Regularized Approximate Line...
research
06/04/2021

Top-k Regularization for Supervised Feature Selection

Feature selection identifies subsets of informative features and reduces...
research
10/05/2016

ℓ_1 Regularized Gradient Temporal-Difference Learning

In this paper, we study the Temporal Difference (TD) learning with linea...
research
04/16/2014

An Analysis of State-Relevance Weights and Sampling Distributions on L1-Regularized Approximate Linear Programming Approximation Accuracy

Recent interest in the use of L_1 regularization in the use of value fun...
research
10/23/2016

Inertial Regularization and Selection (IRS): Sequential Regression in High-Dimension and Sparsity

In this paper, we develop a new sequential regression modeling approach ...

Please sign up or login with your details

Forgot password? Click here to reset