Effectiveness of Hierarchical Softmax in Large Scale Classification Tasks

12/13/2018
by   Abdul Arfat Mohammed, et al.
0

Typically, Softmax is used in the final layer of a neural network to get a probability distribution for output classes. But the main problem with Softmax is that it is computationally expensive for large scale data sets with large number of possible outputs. To approximate class probability efficiently on such large scale data sets we can use Hierarchical Softmax. LSHTC datasets were used to study the performance of the Hierarchical Softmax. LSHTC datasets have large number of categories. In this paper we evaluate and report the performance of normal Softmax Vs Hierarchical Softmax on LSHTC datasets. This evaluation used macro f1 score as a performance measure. The observation was that the performance of Hierarchical Softmax degrades as the number of classes increase.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/02/2023

Global Hierarchical Neural Networks using Hierarchical Softmax

This paper presents a framework in which hierarchical softmax is used to...
research
03/22/2018

Unbiased scalable softmax optimization

Recent neural network and language models rely on softmax distributions ...
research
03/07/2023

Predicted Embedding Power Regression for Large-Scale Out-of-Distribution Detection

Out-of-distribution (OOD) inputs can compromise the performance and safe...
research
01/30/2019

Doubly Sparse: Sparse Mixture of Sparse Experts for Efficient Softmax Inference

Computations for the softmax function are significantly expensive when t...
research
09/23/2016

One-vs-Each Approximation to Softmax for Scalable Estimation of Probabilities

The softmax representation of probabilities for categorical variables pl...
research
11/23/2020

Effectiveness of MPC-friendly Softmax Replacement

Softmax is widely used in deep learning to map some representation to a ...
research
05/02/2022

Simple Techniques Work Surprisingly Well for Neural Network Test Prioritization and Active Learning (Replicability Study)

Test Input Prioritizers (TIP) for Deep Neural Networks (DNN) are an impo...

Please sign up or login with your details

Forgot password? Click here to reset