Phantom Embeddings: Using Embedding Space for Model Regularization in Deep Neural Networks

04/14/2023
by   Mofassir ul Islam Arif, et al.
0

The strength of machine learning models stems from their ability to learn complex function approximations from data; however, this strength also makes training deep neural networks challenging. Notably, the complex models tend to memorize the training data, which results in poor regularization performance on test data. The regularization techniques such as L1, L2, dropout, etc. are proposed to reduce the overfitting effect; however, they bring in additional hyperparameters tuning complexity. These methods also fall short when the inter-class similarity is high due to the underlying data distribution, leading to a less accurate model. In this paper, we present a novel approach to regularize the models by leveraging the information-rich latent embeddings and their high intra-class correlation. We create phantom embeddings from a subset of homogenous samples and use these phantom embeddings to decrease the inter-class similarity of instances in their latent embedding space. The resulting models generalize better as a combination of their embedding and regularize them without requiring an expensive hyperparameter search. We evaluate our method on two popular and challenging image classification datasets (CIFAR and FashionMNIST) and show how our approach outperforms the standard baselines while displaying better training behavior.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/15/2015

A Comparative Study on Regularization Strategies for Embedding-based Neural Networks

This paper aims to compare different regularization strategies to addres...
research
11/22/2022

Adaptive Prototypical Networks

Prototypical network for Few shot learning tries to learn an embedding f...
research
09/14/2022

Learning Deep Optimal Embeddings with Sinkhorn Divergences

Deep Metric Learning algorithms aim to learn an efficient embedding spac...
research
03/16/2022

Is it all a cluster game? – Exploring Out-of-Distribution Detection based on Clustering in the Embedding Space

It is essential for safety-critical applications of deep neural networks...
research
06/12/2023

Augmenting Zero-Shot Detection Training with Image Labels

Zero-shot detection (ZSD), i.e., detection on classes not seen during tr...
research
09/27/2022

Why neural networks find simple solutions: the many regularizers of geometric complexity

In many contexts, simpler models are preferable to more complex models a...
research
10/15/2019

MUTE: Data-Similarity Driven Multi-hot Target Encoding for Neural Network Design

Target encoding is an effective technique to deliver better performance ...

Please sign up or login with your details

Forgot password? Click here to reset