Optimal Connectivity through Network Gradients for the Restricted Boltzmann Machine

09/14/2022
by   A. C. N. de Oliveira, et al.
0

Leveraging sparse networks to connect successive layers in deep neural networks has recently been shown to provide benefits to large scale state-of-the-art models. However, network connectivity also plays a significant role on the learning curves of shallow networks, such as the classic Restricted Boltzmann Machines (RBM). A fundamental problem is efficiently finding connectivity patterns that improve the learning curve. Recent principled approaches explicitly include network connections as parameters that must be optimized in the model, but often rely on continuous functions to represent connections and on explicit penalization. This work presents a method to find optimal connectivity patterns for RBMs based on the idea of network gradients: computing the gradient of every possible connection, given a specific connection pattern, and using the gradient to drive a continuous connection strength parameter that in turn is used to determine the connection pattern. Thus, learning RBM parameters and learning network connections is truly jointly performed, albeit with different learning rates, and without changes to the objective function. The method is applied to the MNIST data set showing that better RBM models are found for the benchmark tasks of sample generation and input classification.

READ FULL TEXT
research
06/01/2013

An Analysis of the Connections Between Layers of Deep Neural Networks

We present an analysis of different techniques for selecting the connect...
research
07/25/2020

From Boltzmann Machines to Neural Networks and Back Again

Graphical models are powerful tools for modeling high-dimensional data, ...
research
11/06/2017

Characterizing Sparse Connectivity Patterns in Neural Networks

We propose a novel way of reducing the number of parameters in the stora...
research
05/11/2022

Deep Architecture Connectivity Matters for Its Convergence: A Fine-Grained Analysis

Advanced deep neural networks (DNNs), designed by either human or AutoML...
research
01/13/2022

Automatic Sparse Connectivity Learning for Neural Networks

Since sparse neural networks usually contain many zero weights, these un...
research
10/30/2017

Log-DenseNet: How to Sparsify a DenseNet

Skip connections are increasingly utilized by deep neural networks to im...
research
02/16/2023

The autoregressive neural network architecture of the Boltzmann distribution of pairwise interacting spins systems

Generative Autoregressive Neural Networks (ARNN) have recently demonstra...

Please sign up or login with your details

Forgot password? Click here to reset