An Interpretable Approach to Hateful Meme Detection

08/09/2021
by   Tanvi Deshpande, et al.
0

Hateful memes are an emerging method of spreading hate on the internet, relying on both images and text to convey a hateful message. We take an interpretable approach to hateful meme detection, using machine learning and simple heuristics to identify the features most important to classifying a meme as hateful. In the process, we build a gradient-boosted decision tree and an LSTM-based model that achieve comparable performance (73.8 validation and 72.7 test auROC) to the gold standard of humans and state-of-the-art transformer models on this challenging task.

READ FULL TEXT
research
07/21/2022

Learning Physics from the Machine: An Interpretable Boosted Decision Tree Analysis for the Majorana Demonstrator

The Majorana Demonstrator is a leading experiment searching for neutrino...
research
02/09/2023

FrameBERT: Conceptual Metaphor Detection with Frame Embedding Learning

In this paper, we propose FrameBERT, a RoBERTa-based model that can expl...
research
05/14/2018

Rogue Drone Detection: A Machine Learning Approach

The emerging, practical and observed issue of how to detect rogue drones...
research
08/17/2023

Factuality Detection using Machine Translation – a Use Case for German Clinical Text

Factuality can play an important role when automatically processing clin...
research
05/06/2019

RSL19BD at DBDC4: Ensemble of Decision Tree-based and LSTM-based Models

RSL19BD (Waseda University Sakai Laboratory) participated in the Fourth ...
research
08/20/2019

TabNet: Attentive Interpretable Tabular Learning

We propose a novel high-performance interpretable deep tabular data lear...
research
06/06/2019

Blockwise Based Detection of Local Defects

Print quality is an important criterion for a printer's performance. The...

Please sign up or login with your details

Forgot password? Click here to reset