Efficient Deep Learning of GMMs

02/15/2019
by   Shirin Jalali, et al.
0

We show that a collection of Gaussian mixture models (GMMs) in R^n can be optimally classified using O(n) neurons in a neural network with two hidden layers (deep neural network), whereas in contrast, a neural network with a single hidden layer (shallow neural network) would require at least O((n)) neurons or possibly exponentially large coefficients. Given the universality of the Gaussian distribution in the feature spaces of data, e.g., in speech, image and text, our result sheds light on the observed efficiency of deep neural networks in practical classification problems.

READ FULL TEXT
research
07/24/2023

Rates of Approximation by ReLU Shallow Neural Networks

Neural networks activated by the rectified linear unit (ReLU) play a cen...
research
08/29/2016

Why does deep and cheap learning work so well?

We show how the success of deep learning could depend not only on mathem...
research
06/06/2018

A Peek Into the Hidden Layers of a Convolutional Neural Network Through a Factorization Lens

Despite their increasing popularity and success in a variety of supervis...
research
12/14/2020

Perceptron Theory for Predicting the Accuracy of Neural Networks

Many neural network models have been successful at classification proble...
research
02/01/2023

Multi-Grade Deep Learning

The current deep learning model is of a single-grade, that is, it learns...
research
06/01/2015

Learning with hidden variables

Learning and inferring features that generate sensory input is a task co...
research
11/14/2021

A layer-stress learning framework universally augments deep neural network tasks

Deep neural networks (DNN) such as Multi-Layer Perception (MLP) and Conv...

Please sign up or login with your details

Forgot password? Click here to reset