Learning Discrete Representations via Information Maximizing Self-Augmented Training

02/28/2017
by   Weihua Hu, et al.
0

Learning discrete representations of data is a central machine learning task because of the compactness of the representations and ease of interpretation. The task includes clustering and hash learning as special cases. Deep neural networks are promising to be used because they can model the non-linearity of data and scale to large datasets. However, their model complexity is huge, and therefore, we need to carefully regularize the networks in order to learn useful representations that exhibit intended invariance for applications of interest. To this end, we propose a method called Information Maximizing Self-Augmented Training (IMSAT). In IMSAT, we use data augmentation to impose the invariance on discrete representations. More specifically, we encourage the predicted representations of augmented data points to be close to those of the original data points in an end-to-end fashion. At the same time, we maximize the information-theoretic dependency between data and their predicted discrete representations. Extensive experiments on benchmark datasets show that IMSAT produces state-of-the-art results for both clustering and unsupervised hash learning.

READ FULL TEXT

page 8

page 13

page 15

research
01/03/2018

ScreenerNet: Learning Curriculum for Neural Networks

We propose to learn a curriculum or a syllabus for supervised learning w...
research
10/18/2019

Automatic Data Augmentation by Learning the Deterministic Policy

Aiming to produce sufficient and diverse training samples, data augmenta...
research
05/10/2020

Variational Clustering: Leveraging Variational Autoencoders for Image Clustering

Recent advances in deep learning have shown their ability to learn stron...
research
11/25/2020

Squared ℓ_2 Norm as Consistency Loss for Leveraging Augmented Data to Learn Robust and Invariant Representations

Data augmentation is one of the most popular techniques for improving th...
research
03/07/2022

Regularising for invariance to data augmentation improves supervised learning

Data augmentation is used in machine learning to make the classifier inv...
research
10/08/2021

Learning a Self-Expressive Network for Subspace Clustering

State-of-the-art subspace clustering methods are based on self-expressiv...
research
10/15/2022

Data-Efficient Augmentation for Training Neural Networks

Data augmentation is essential to achieve state-of-the-art performance i...

Please sign up or login with your details

Forgot password? Click here to reset