Mixup Training as the Complexity Reduction

06/11/2020
by   Masanari Kimura, et al.
0

Machine learning has achieved remarkable results in recent years due to the increase in the number of data and the development of computational resources. However, despite such excellent performance, machine learning models often suffer from the problem of over-fitting. Many data augmentation methods have been proposed to tackle such a problem, and one of them is called Mixup. Mixup is a recently proposed regularization procedure, which linearly interpolates a random pair of training examples. This regularization method works very well experimentally, but its theoretical guarantee is not fully discussed. In this study, we aim to find out why Mixup works well from the aspect of computational learning theory. In addition, we reveal how the effect of Mixup changes in each situation. Furthermore, we also investigated the effects of changes in the Mixup's parameter. This contributes to the search for the optimal parameters and to estimate the effects of the parameters currently used. The results of this study provide a theoretical clarification of when and how effective regularization by Mixup is.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/11/2018

Data augmentation instead of explicit regularization

Modern deep artificial neural networks have achieved impressive results ...
research
07/08/2020

Density Fixing: Simple yet Effective Regularization Method based on the Class Prior

Machine learning models suffer from overfitting, which is caused by a la...
research
04/12/2023

Towards Understanding How Data Augmentation Works with Imbalanced Data

Data augmentation forms the cornerstone of many modern machine learning ...
research
03/23/2018

SEGEN: Sample-Ensemble Genetic Evolutional Network Model

Deep learning, a rebranding of deep neural network research works, has a...
research
11/19/2015

A Unified Gradient Regularization Family for Adversarial Examples

Adversarial examples are augmented data points generated by imperceptibl...
research
05/27/2021

Drawing Multiple Augmentation Samples Per Image During Training Efficiently Decreases Test Error

In computer vision, it is standard practice to draw a single sample from...
research
12/13/2017

Regularization and Optimization strategies in Deep Convolutional Neural Network

Convolution Neural Networks, known as ConvNets exceptionally perform wel...

Please sign up or login with your details

Forgot password? Click here to reset