One-step and Two-step Classification for Abusive Language Detection on Twitter

06/05/2017
by   Ji Ho Park, et al.
0

Automatic abusive language detection is a difficult but important task for online social media. Our research explores a two-step approach of performing classification on abusive language and then classifying into specific types and compares it with one-step approach of doing one multi-class classification for detecting sexist and racist languages. With a public English Twitter corpus of 20 thousand tweets in the type of sexism and racism, our approach shows a promising performance of 0.827 F-measure by using HybridCNN in one-step and 0.824 F-measure by using logistic regression in two-steps.

READ FULL TEXT
POST COMMENT

Comments

There are no comments yet.

Authors

page 1

page 2

page 3

page 4

07/19/2016

Discriminating between similar languages in Twitter using label propagation

Identifying the language of social media messages is an important first ...
05/17/2016

Tweet Acts: A Speech Act Classifier for Twitter

Speech acts are a way to conceptualize speech as action. This holds true...
08/30/2016

Language Detection For Short Text Messages In Social Media

With the constant growth of the World Wide Web and the number of documen...
01/03/2019

Event detection in Twitter: A keyword volume approach

Event detection using social media streams needs a set of informative fe...
07/25/2020

Duluth at SemEval-2020 Task 12: Offensive Tweet Identification in English with Logistic Regression

This paper describes the Duluth systems that participated in SemEval–202...
04/24/2017

Turing at SemEval-2017 Task 8: Sequential Approach to Rumour Stance Classification with Branch-LSTM

This paper describes team Turing's submission to SemEval 2017 RumourEval...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.