Reinforcement Learning via Gaussian Processes with Neural Network Dual Kernels

04/10/2020
by   Imène R. Goumiri, et al.
0

While deep neural networks (DNNs) and Gaussian Processes (GPs) are both popularly utilized to solve problems in reinforcement learning, both approaches feature undesirable drawbacks for challenging problems. DNNs learn complex nonlinear embeddings, but do not naturally quantify uncertainty and are often data-inefficient to train. GPs infer posterior distributions over functions, but popular kernels exhibit limited expressivity on complex and high-dimensional data. Fortunately, recently discovered conjugate and neural tangent kernel functions encode the behavior of overparameterized neural networks in the kernel domain. We demonstrate that these kernels can be efficiently applied to regression and reinforcement learning problems by analyzing a baseline case study. We apply GPs with neural network dual kernels to solve reinforcement learning tasks for the first time. We demonstrate, using the well-understood mountain-car problem, that GPs empowered with dual kernels perform at least as well as those using the conventional radial basis function kernel. We conjecture that by inheriting the probabilistic rigor of GPs and the powerful embedding properties of DNNs, GPs using NN dual kernels will empower future reinforcement learning models on difficult domains.

READ FULL TEXT
research
10/13/2019

Deep Probabilistic Kernels for Sample-Efficient Learning

Gaussian Processes (GPs) with an appropriate kernel are known to provide...
research
11/01/2021

End-to-End Learning of Deep Kernel Acquisition Functions for Bayesian Optimization

For Bayesian optimization (BO) on high-dimensional data with complex str...
research
07/08/2017

Adversarial Examples, Uncertainty, and Transfer Testing Robustness in Gaussian Process Hybrid Deep Networks

Deep neural networks (DNNs) have excellent representative power and are ...
research
05/15/2019

Expressive Priors in Bayesian Neural Networks: Kernel Combinations and Periodic Functions

A simple, flexible approach to creating expressive priors in Gaussian pr...
research
07/04/2018

Conditional Neural Processes

Deep neural networks excel at function approximation, yet they are typic...
research
08/31/2022

Light curve completion and forecasting using fast and scalable Gaussian processes (MuyGPs)

Temporal variations of apparent magnitude, called light curves, are obse...

Please sign up or login with your details

Forgot password? Click here to reset