How Could Polyhedral Theory Harness Deep Learning?

06/17/2018
by   Thiago Serra, et al.
0

The holy grail of deep learning is to come up with an automatic method to design optimal architectures for different applications. In other words, how can we effectively dimension and organize neurons along the network layers based on the computational resources, input size, and amount of training data? We outline promising research directions based on polyhedral theory and mixed-integer representability that may offer an analytical approach to this question, in contrast to the empirical techniques often employed.

READ FULL TEXT

page 1

page 2

research
06/09/2019

Quadratic Suffices for Over-parametrization via Matrix Chernoff Bound

We improve the over-parametrization size over two beautiful results [Li ...
research
08/24/2021

Adversarial Robustness of Deep Learning: Theory, Algorithms, and Applications

This tutorial aims to introduce the fundamentals of adversarial robustne...
research
03/02/2021

Sparse Training Theory for Scalable and Efficient Agents

A fundamental task for artificial intelligence is learning. Deep Neural ...
research
08/08/2023

Deep Learning based Image Watermarking: A Brief Survey

The act of secretly embedding and extracting a watermark on a cover imag...
research
04/16/2022

Visual Attention Methods in Deep Learning: An In-Depth Survey

Inspired by the human cognitive system, attention is a mechanism that im...
research
02/17/2020

Identifying Critical Neurons in ANN Architectures using Mixed Integer Programming

We introduce a novel approach to optimize the architecture of deep neura...
research
06/12/2018

Learning to Automatically Generate Fill-In-The-Blank Quizzes

In this paper we formalize the problem automatic fill-in-the-blank quest...

Please sign up or login with your details

Forgot password? Click here to reset