Macro F1 and Macro F1

11/08/2019
by   Juri Opitz, et al.
0

The 'macro F1' metric is frequently used to evaluate binary, multi-class and multi-label classification problems. Yet, we find that there exist two different formulas to calculate this quantity. In this note, we show that only under rare circumstances, the two computations can be considered equivalent. More specifically, one formula well 'rewards' classifiers which produce a skewed error type distribution. In fact, the difference in outcome of the two computations can be as high as 0.5. Finally, we show that the two computations may not only diverge in their scalar result but also lead to different classifier rankings.

READ FULL TEXT
research
09/03/2016

An Online Universal Classifier for Binary, Multi-class and Multi-label Classification

Classification involves the learning of the mapping function that associ...
research
04/14/2022

Multi-label topic classification for COVID-19 literature with Bioformer

We describe Bioformer team's participation in the multi-label topic clas...
research
11/27/2021

Exploring Transformer Based Models to Identify Hate Speech and Offensive Content in English and Indo-Aryan Languages

Hate speech is considered to be one of the major issues currently plagui...
research
05/19/2022

Why only Micro-F1? Class Weighting of Measures for Relation Classification

Relation classification models are conventionally evaluated using only a...
research
10/16/2019

Using Supervised Learning to Classify Metadata of Research Data by Discipline of Research

Automated classification of metadata of research data by their disciplin...
research
02/08/2014

Thresholding Classifiers to Maximize F1 Score

This paper provides new insight into maximizing F1 scores in the context...
research
05/09/2023

Towards Understanding Generalization of Macro-AUC in Multi-label Learning

Macro-AUC is the arithmetic mean of the class-wise AUCs in multi-label l...

Please sign up or login with your details

Forgot password? Click here to reset