Deep Neural Networks as 0-1 Mixed Integer Linear Programs: A Feasibility Study

12/17/2017
by   Matteo Fischetti, et al.
0

Deep Neural Networks (DNNs) are very popular these days, and are the subject of a very intense investigation. A DNN is made by layers of internal units (or neurons), each of which computes an affine combination of the output of the units in the previous layer, applies a nonlinear operator, and outputs the corresponding value (also known as activation). A commonly-used nonlinear operator is the so-called rectified linear unit (ReLU), whose output is just the maximum between its input value and zero. In this (and other similar cases like max pooling, where the max operation involves more than one input value), one can model the DNN as a 0-1 Mixed Integer Linear Program (0-1 MILP) where the continuous variables correspond to the output values of each unit, and a binary variable is associated with each ReLU to model its yes/no nature. In this paper we discuss the peculiarity of this kind of 0-1 MILP models, and describe an effective bound-tightening technique intended to ease its solution. We also present possible applications of the 0-1 MILP model arising in feature visualization and in the construction of adversarial examples. Preliminary computational results are reported, aimed at investigating (on small DNNs) the computational performance of a state-of-the-art MILP solver when applied to a known test case, namely, hand-written digit recognition.

READ FULL TEXT

page 6

page 8

page 9

research
07/06/2019

ReLU Networks as Surrogate Models in Mixed-Integer Linear Programs

We consider the embedding of piecewise-linear deep neural networks (ReLU...
research
08/19/2020

ReLU activated Multi-Layer Neural Networks trained with Mixed Integer Linear Programs

This paper is a case study to demonstrate that, in principle, multi-laye...
research
02/20/2023

Model-based feature selection for neural networks: A mixed-integer programming approach

In this work, we develop a novel input feature selection framework for R...
research
11/20/2018

Strong mixed-integer programming formulations for trained neural networks

We present an ideal mixed-integer programming (MIP) formulation for a re...
research
01/03/2023

Improving Performance in Neural Networks by Dendrites-Activated Connections

Computational units in artificial neural networks compute a linear combi...
research
07/12/2019

Deep Learning-powered Iterative Combinatorial Auctions

In this paper, we study the design of deep learning-powered iterative co...
research
02/10/2020

Deep Gated Networks: A framework to understand training and generalisation in deep learning

Understanding the role of (stochastic) gradient descent in training and ...

Please sign up or login with your details

Forgot password? Click here to reset