Cross-Entropic Learning of a Machine for the Decision in a Partially Observable Universe

05/18/2006
by   Frederic Dambreville, et al.
0

Revision of the paper previously entitled "Learning a Machine for the Decision in a Partially Observable Markov Universe" In this paper, we are interested in optimal decisions in a partially observable universe. Our approach is to directly approximate an optimal strategic tree depending on the observation. This approximation is made by means of a parameterized probabilistic law. A particular family of hidden Markov models, with input and output, is considered as a model of policy. A method for optimizing the parameters of these HMMs is proposed and applied. This optimization is based on the cross-entropic principle for rare events simulation developed by Rubinstein.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/17/2018

Hidden Markov Model Estimation-Based Q-learning for Partially Observable Markov Decision Process

The objective is to study an on-line Hidden Markov model (HMM) estimatio...
research
02/25/2016

Reinforcement Learning of POMDPs using Spectral Methods

We propose a new reinforcement learning algorithm for partially observab...
research
07/13/2016

The Partially Observable Hidden Markov Model and its Application to Keystroke Dynamics

The partially observable hidden Markov model is an extension of the hidd...
research
11/02/2012

Learning classifier systems with memory condition to solve non-Markov problems

In the family of Learning Classifier Systems, the classifier system XCS ...
research
12/17/2020

Repairing dynamic models: a method to obtain identifiable and observable reparameterizations with mechanistic insights

Mechanistic dynamic models allow for a quantitative and systematic inter...
research
05/20/2020

Hidden Markov Models and their Application for Predicting Failure Events

We show how Markov mixed membership models (MMMM) can be used to predict...

Please sign up or login with your details

Forgot password? Click here to reset