Learning Representations in Reinforcement Learning:An Information Bottleneck Approach

11/12/2019
by   Pei Yingjun, et al.
0

The information bottleneck principle is an elegant and useful approach to representation learning. In this paper, we investigate the problem of representation learning in the context of reinforcement learning using the information bottleneck framework, aiming at improving the sample efficiency of the learning algorithms. information when the analytically derive the optimal conditional distribution of the representation, and provide a variational lower bound. Then, we maximize this lower bound with the Stein variational (SV) gradient method. We incorporate this framework in the advantageous actor critic algorithm (A2C) and the proximal policy optimization algorithm (PPO). Our experimental results show that our framework can improve the sample efficiency of vanilla A2C and PPO significantly. Finally, we study the information bottleneck (IB) perspective in deep RL with the algorithm called mutual information neural estimation(MINE) . We experimentally verify that the information extraction-compression process also exists in deep RL and our framework is capable of accelerating this process. We also analyze the relationship between MINE and our method, through this relationship, we theoretically derive an algorithm to optimize our IB framework without constructing the lower bound.

READ FULL TEXT
research
06/04/2020

A Maximum Mutual Information Framework for Multi-Agent Reinforcement Learning

In this paper, we propose a maximum mutual information (MMI) framework f...
research
03/01/2023

A Variational Approach to Mutual Information-Based Coordination for Multi-Agent Reinforcement Learning

In this paper, we propose a new mutual information framework for multi-a...
research
12/31/2019

On the Difference Between the Information Bottleneck and the Deep Information Bottleneck

Combining the Information Bottleneck model with deep learning by replaci...
research
03/07/2020

The Variational InfoMax Learning Objective

Bayesian Inference and Information Bottleneck are the two most popular o...
research
05/28/2019

Variational Information Bottleneck for Unsupervised Clustering: Deep Gaussian Mixture Embedding

In this paper, we develop an unsupervised generative clustering framewor...
research
05/24/2016

Relevant sparse codes with variational information bottleneck

In many applications, it is desirable to extract only the relevant aspec...
research
07/11/2018

Distributed Variational Representation Learning

The problem of distributed representation learning is one in which multi...

Please sign up or login with your details

Forgot password? Click here to reset