OSLNet: Deep Small-Sample Classification with an Orthogonal Softmax Layer

04/20/2020
by   Xiaoxu Li, et al.
0

A deep neural network of multiple nonlinear layers forms a large function space, which can easily lead to overfitting when it encounters small-sample data. To mitigate overfitting in small-sample classification, learning more discriminative features from small-sample data is becoming a new trend. To this end, this paper aims to find a subspace of neural networks that can facilitate a large decision margin. Specifically, we propose the Orthogonal Softmax Layer (OSL), which makes the weight vectors in the classification layer remain orthogonal during both the training and test processes. The Rademacher complexity of a network using the OSL is only 1/K, where K is the number of classes, of that of a network using the fully connected classification layer, leading to a tighter generalization error bound. Experimental results demonstrate that the proposed OSL has better performance than the methods used for comparison on four small-sample benchmark datasets, as well as its applicability to large-sample datasets. Codes are available at: https://github.com/dongliangchang/OSLNet.

READ FULL TEXT
research
06/27/2020

ReMarNet: Conjoint Relation and Margin Learning for Small-Sample Image Classification

Despite achieving state-of-the-art performance, deep learning methods ge...
research
05/30/2016

Stochastic Function Norm Regularization of Deep Networks

Deep neural networks have had an enormous impact on image analysis. Stat...
research
03/21/2023

Equiangular Basis Vectors

We propose Equiangular Basis Vectors (EBVs) for classification tasks. In...
research
04/25/2019

Learning Discriminative Features Via Weights-biased Softmax Loss

Loss functions play a key role in training superior deep neural networks...
research
04/24/2019

Deep Sparse Representation-based Classification

We present a transductive deep learning-based formulation for the sparse...
research
08/05/2022

Almost-Orthogonal Layers for Efficient General-Purpose Lipschitz Networks

It is a highly desirable property for deep networks to be robust against...
research
07/30/2020

Improving Sample Efficiency with Normalized RBF Kernels

In deep learning models, learning more with less data is becoming more i...

Please sign up or login with your details

Forgot password? Click here to reset