Learning towards Minimum Hyperspherical Energy

05/23/2018
by   Weiyang Liu, et al.
0

Neural networks are a powerful class of nonlinear functions that can be trained end-to-end on various applications. While the over-parametrization nature in many neural networks renders the ability to fit complex functions and the strong representation power to handle challenging tasks, it also leads to highly correlated neurons that can hurt the generalization ability and incur unnecessary computation cost. As a result, how to regularize the network to avoid undesired representation redundancy becomes an important issue. To this end, we draw inspiration from a well-known problem in physics -- Thomson problem, where one seeks to find a state that distributes N electrons on a unit sphere as even as possible with minimum potential energy. In light of this intuition, we reduce the redundancy regularization problem to generic energy minimization, and propose a minimum hyperspherical energy (MHE) objective as generic regularization for neural networks. We also propose a few novel variants of MHE, and provide some insights from a theoretical point of view. Finally, we apply networks with MHE regularization to several challenging tasks. Extensive experiments demonstrate the effectiveness of our method, by showing the superior performance with MHE regularization.

READ FULL TEXT

page 2

page 17

research
06/12/2019

Compressive Hyperspherical Energy Minimization

Recent work on minimum hyperspherical energy (MHE) has demonstrated its ...
research
06/06/2020

MMA Regularization: Decorrelating Weights of Neural Networks by Maximizing the Minimal Angles

The strong correlation between neurons or filters can significantly weak...
research
03/02/2021

Learning with Hyperspherical Uniformity

Due to the over-parameterization nature, neural networks are a powerful ...
research
11/20/2018

Gradient-Coherent Strong Regularization for Deep Neural Networks

Deep neural networks are often prone to over-fitting with their numerous...
research
10/27/2022

On the Approximation and Complexity of Deep Neural Networks to Invariant Functions

Recent years have witnessed a hot wave of deep neural networks in variou...
research
08/08/2023

Minimizing Quotient Regularization Model

Quotient regularization models (QRMs) are a class of powerful regulariza...
research
03/23/2022

Out of Distribution Detection, Generalization, and Robustness Triangle with Maximum Probability Theorem

Maximum Probability Framework, powered by Maximum Probability Theorem, i...

Please sign up or login with your details

Forgot password? Click here to reset