A Branching and Merging Convolutional Network with Homogeneous Filter Capsules

01/24/2020
by   Adam Byerly, et al.
7

We present a convolutional neural network design with additional branches after certain convolutions so that we can extract features with differing effective receptive fields and levels of abstraction. From each branch, we transform each of the final filters into a pair of homogeneous vector capsules. As the capsules are formed from entire filters, we refer to them as filter capsules. We then compare three methods for merging the branches–merging with equal weight and merging with learned weights, with two different weight initialization strategies. This design, in combination with a domain-specific set of randomly applied augmentation techniques, establishes a new state of the art for the MNIST dataset with an accuracy of 99.84 models, as well as establishing a new state of the art for a single model (99.79 the number of parameters and the number of epochs of training relative to the previously best performing capsule network on MNIST. All training was performed using the Adam optimizer and experienced no overfitting.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/09/2023

Revisiting Permutation Symmetry for Merging Models between Different Datasets

Model merging is a new approach to creating a new model by combining the...
research
11/20/2017

Learning Steerable Filters for Rotation Equivariant CNNs

In many machine learning tasks it is desirable that a model's prediction...
research
04/21/2022

Merging of neural networks

We propose a simple scheme for merging two neural networks trained with ...
research
09/30/2022

E-Branchformer: Branchformer with Enhanced merging for speech recognition

Conformer, combining convolution and self-attention sequentially to capt...
research
03/16/2015

Enhanced Image Classification With a Fast-Learning Shallow Convolutional Neural Network

We present a neural network architecture and training method designed to...
research
06/20/2019

Homogeneous Vector Capsules Enable Adaptive Gradient Descent in Convolutional Neural Networks

Capsules are the name given by Geoffrey Hinton to vector-valued neurons....
research
10/07/2022

Understanding the Covariance Structure of Convolutional Filters

Neural network weights are typically initialized at random from univaria...

Please sign up or login with your details

Forgot password? Click here to reset