Separability is not the best goal for machine learning

07/08/2018
by   Wlodzislaw Duch, et al.
0

Neural networks use their hidden layers to transform input data into linearly separable data clusters, with a linear or a perceptron type output layer making the final projection on the line perpendicular to the discriminating hyperplane. For complex data with multimodal distributions this transformation is difficult to learn. Projection on k≥ 2 line segments is the simplest extension of linear separability, defining much easier goal for the learning process. Simple problems are 2-separable, but problems with inherent complex logic may be solved in a simple way by k-separable projections. The difficulty of learning non-linear data distributions is shifted to separation of line intervals, simplifying the transformation of data by hidden network layers. For classification of difficult Boolean problems, such as the parity problem, linear projection combined with is sufficient and provides a powerful new target for learning. More complex targets may also be defined, changing the goal of learning from linear discrimination to creation of data distributions that can easily be handled by specialized models selected to analyze output distributions. This approach can replace many layers of transformation required by deep learning models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/09/2023

Hidden Classification Layers: a study on Data Hidden Representations with a Higher Degree of Linear Separability between the Classes

In the context of classification problems, Deep Learning (DL) approaches...
research
09/29/2021

A Comprehensive Survey and Performance Analysis of Activation Functions in Deep Learning

Neural networks have shown tremendous growth in recent years to solve nu...
research
10/19/2016

Learning to Learn Neural Networks

Meta-learning consists in learning learning algorithms. We use a Long Sh...
research
02/01/2023

Simplicity Bias in 1-Hidden Layer Neural Networks

Recent works have demonstrated that neural networks exhibit extreme simp...
research
03/22/2022

Exploring Linear Feature Disentanglement For Neural Networks

Non-linear activation functions, e.g., Sigmoid, ReLU, and Tanh, have ach...
research
05/03/2015

Making Sense of Hidden Layer Information in Deep Networks by Learning Hierarchical Targets

This paper proposes an architecture for deep neural networks with hidden...
research
05/19/2021

Obstructing Classification via Projection

Machine learning and data mining techniques are effective tools to class...

Please sign up or login with your details

Forgot password? Click here to reset