Compression and Localization in Reinforcement Learning for ATARI Games

04/20/2019
by   Joel Ruben Antony Moniz, et al.
0

Deep neural networks have become commonplace in the domain of reinforcement learning, but are often expensive in terms of the number of parameters needed. While compressing deep neural networks has of late assumed great importance to overcome this drawback, little work has been done to address this problem in the context of reinforcement learning agents. This work aims at making first steps towards model compression in an RL agent. In particular, we compress networks to drastically reduce the number of parameters in them (to sizes less than 3 pool after the final convolution layer, and propose using Actor-Mimic in the context of compression. Finally, we show that this global max-pool allows for weakly supervised object localization, improving the ability to identify the agent's points of focus.

READ FULL TEXT

page 4

page 6

research
04/11/2021

The Atari Data Scraper

Reinforcement learning has made great strides in recent years due to the...
research
02/12/2020

Data Efficient Training for Reinforcement Learning with Adaptive Behavior Policy Sharing

Deep Reinforcement Learning (RL) is proven powerful for decision making ...
research
10/07/2021

Explaining Deep Reinforcement Learning Agents In The Atari Domain through a Surrogate Model

One major barrier to applications of deep Reinforcement Learning (RL) bo...
research
02/05/2021

GNN-RL Compression: Topology-Aware Network Pruning using Multi-stage Graph Embedding and Reinforcement Learning

Model compression is an essential technique for deploying deep neural ne...
research
02/20/2018

Do Deep Learning Models Have Too Many Parameters? An Information Theory Viewpoint

Deep learning models often have more parameters than observations, and s...
research
12/14/2018

Scaling shared model governance via model splitting

Currently the only techniques for sharing governance of a deep learning ...
research
12/09/2022

Expeditious Saliency-guided Mix-up through Random Gradient Thresholding

Mix-up training approaches have proven to be effective in improving the ...

Please sign up or login with your details

Forgot password? Click here to reset