Trusting SVM for Piecewise Linear CNNs

11/07/2016
by   Leonard Berrada, et al.
0

We present a novel layerwise optimization algorithm for the learning objective of Piecewise-Linear Convolutional Neural Networks (PL-CNNs), a large class of convolutional neural networks. Specifically, PL-CNNs employ piecewise linear non-linearities such as the commonly used ReLU and max-pool, and an SVM classifier as the final layer. The key observation of our approach is that the problem corresponding to the parameter estimation of a layer can be formulated as a difference-of-convex (DC) program, which happens to be a latent structured SVM. We optimize the DC program using the concave-convex procedure, which requires us to iteratively solve a structured SVM problem. This allows to design an optimization algorithm with an optimal learning rate that does not require any tuning. Using the MNIST, CIFAR and ImageNet data sets, we show that our approach always improves over the state of the art variants of backpropagation and scales to large data and large network settings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/16/2020

Unified SVM algorithm based LS-DC Loss

Over the past two decades, Support Vector Machine (SVM) has been a popul...
research
07/05/2020

Piecewise Linear Regression via a Difference of Convex Functions

We present a new piecewise linear regression methodology that utilizes f...
research
02/15/2019

SVM-based Deep Stacking Networks

The deep network model, with the majority built on neural networks, has ...
research
06/26/2020

Implicit Convex Regularizers of CNN Architectures: Convex Optimization of Two- and Three-Layer Networks in Polynomial Time

We study training of Convolutional Neural Networks (CNNs) with ReLU acti...
research
10/08/2018

Diagnosing Convolutional Neural Networks using their Spectral Response

Convolutional Neural Networks (CNNs) are a class of artificial neural ne...
research
01/23/2019

Decoupled Greedy Learning of CNNs

A commonly cited inefficiency of neural network training by back-propaga...
research
05/27/2022

Optimizing Objective Functions from Trained ReLU Neural Networks via Sampling

This paper introduces scalable, sampling-based algorithms that optimize ...

Please sign up or login with your details

Forgot password? Click here to reset