Learning Task-Driven Control Policies via Information Bottlenecks

02/04/2020
by   Vincent Pacelli, et al.
0

This paper presents a reinforcement learning approach to synthesizing task-driven control policies for robotic systems equipped with rich sensory modalities (e.g., vision or depth). Standard reinforcement learning algorithms typically produce policies that tightly couple control actions to the entirety of the system's state and rich sensor observations. As a consequence, the resulting policies can often be sensitive to changes in task-irrelevant portions of the state or observations (e.g., changing background colors). In contrast, the approach we present here learns to create a task-driven representation that is used to compute control actions. Formally, this is achieved by deriving a policy gradient-style algorithm that creates an information bottleneck between the states and the task-driven representation; this constrains actions to only depend on task-relevant information. We demonstrate our approach in a thorough set of simulation results on multiple examples including a grasping task that utilizes depth images and a ball-catching task that utilizes RGB images. Comparisons with a standard policy gradient approach demonstrate that the task-driven policies produced by our algorithm are often significantly more robust to sensor noise and task-irrelevant changes in the environment.

READ FULL TEXT

page 1

page 7

research
09/20/2018

Task-Driven Estimation and Control via Information Bottlenecks

Our goal is to develop a principled and general algorithmic framework fo...
research
12/13/2022

Scalable and Sample Efficient Distributed Policy Gradient Algorithms in Multi-Agent Networked Systems

This paper studies a class of multi-agent reinforcement learning (MARL) ...
research
09/10/2020

A framework for reinforcement learning with autocorrelated actions

The subject of this paper is reinforcement learning. Policies are consid...
research
08/17/2020

Learning to Actively Reduce Memory Requirements for Robot Control Tasks

Robots equipped with rich sensing modalities (e.g., RGB-D cameras) perfo...
research
02/26/2021

Robust Deep Reinforcement Learning via Multi-View Information Bottleneck

Deep reinforcement learning (DRL) agents are often sensitive to visual c...
research
05/30/2017

Learning End-to-end Multimodal Sensor Policies for Autonomous Navigation

Multisensory polices are known to enhance both state estimation and targ...
research
08/30/2023

DBNet: Leveraging DBMS for Network Automation

We present DBNet, a data-driven network automation framework built on to...

Please sign up or login with your details

Forgot password? Click here to reset