Hierarchical Soft Actor-Critic: Adversarial Exploration via Mutual Information Optimization

06/17/2019
by   Ari Azarafrooz, et al.
0

We describe a novel extension of soft actor-critics for hierarchical Deep Q-Networks (HDQN) architectures using mutual information metric. The proposed extension provides a suitable framework for encouraging explorations in such hierarchical networks. A natural utilization of this framework is an adversarial setting, where meta-controller and controller play minimax over the mutual information objective but cooperate on maximizing expected rewards.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/01/2023

The Point to Which Soft Actor-Critic Converges

Soft actor-critic is a successful successor over soft Q-learning. While ...
research
12/04/2017

Hierarchical Actor-Critic

We present a novel approach to hierarchical reinforcement learning calle...
research
09/11/2019

Mutual-Information Regularization in Markov Decision Processes and Actor-Critic Learning

Cumulative entropy regularization introduces a regulatory signal to the ...
research
12/08/2021

Hyper-parameter optimization based on soft actor critic and hierarchical mixture regularization

Hyper-parameter optimization is a crucial problem in machine learning as...
research
04/08/2020

Learning Discrete Structured Representations by Adversarially Maximizing Mutual Information

We propose learning discrete structured representations from unlabeled d...
research
07/25/2022

Improving Adversarial Robustness via Mutual Information Estimation

Deep neural networks (DNNs) are found to be vulnerable to adversarial no...
research
03/27/2010

Active Testing for Face Detection and Localization

We provide a novel search technique, which uses a hierarchical model and...

Please sign up or login with your details

Forgot password? Click here to reset