Utilizing Excess Resources in Training Neural Networks

07/12/2022
by   Amit Henig, et al.
0

In this work, we suggest Kernel Filtering Linear Overparameterization (KFLO), where a linear cascade of filtering layers is used during training to improve network performance in test time. We implement this cascade in a kernel filtering fashion, which prevents the trained architecture from becoming unnecessarily deeper. This also allows using our approach with almost any network architecture and let combining the filtering layers into a single layer in test time. Thus, our approach does not add computational complexity during inference. We demonstrate the advantage of KFLO on various network models and datasets in supervised learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/04/2022

Mixup for Test-Time Training

Test-time training provides a new approach solving the problem of domain...
research
08/31/2018

Nightmare at test time: How punctuation prevents parsers from generalizing

Punctuation is a strong indicator of syntactic structure, and parsers tr...
research
10/17/2022

Test-Time Training for Graph Neural Networks

Graph Neural Networks (GNNs) have made tremendous progress in the graph ...
research
10/20/2016

A Multidimensional Cascade Neuro-Fuzzy System with Neuron Pool Optimization in Each Cascade

A new architecture and learning algorithms for the multidimensional hybr...
research
02/18/2022

Transfer and Marginalize: Explaining Away Label Noise with Privileged Information

Supervised learning datasets often have privileged information, in the f...
research
10/11/2022

The Unreasonable Effectiveness of Fully-Connected Layers for Low-Data Regimes

Convolutional neural networks were the standard for solving many compute...
research
04/13/2020

Cascade Neural Ensemble for Identifying Scientifically Sound Articles

Background: A significant barrier to conducting systematic reviews and m...

Please sign up or login with your details

Forgot password? Click here to reset