Designing Toxic Content Classification for a Diversity of Perspectives

06/04/2021
by   Deepak Kumar, et al.
0

In this work, we demonstrate how existing classifiers for identifying toxic comments online fail to generalize to the diverse concerns of Internet users. We survey 17,280 participants to understand how user expectations for what constitutes toxic content differ across demographics, beliefs, and personal experiences. We find that groups historically at-risk of harassment - such as people who identify as LGBTQ+ or young adults - are more likely to to flag a random comment drawn from Reddit, Twitter, or 4chan as toxic, as are people who have personally experienced harassment in the past. Based on our findings, we show how current one-size-fits-all toxicity classification algorithms, like the Perspective API from Jigsaw, can improve in accuracy by 86 personalized model tuning. Ultimately, we highlight current pitfalls and new design directions that can improve the equity and efficacy of toxic content classifiers for all users.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/11/2018

Mitigating Confirmation Bias on Twitter by Recommending Opposing Views

In this work, we propose a content-based recommendation approach to incr...
research
05/17/2023

Personalizing Content Moderation on Social Media: User Perspectives on Moderation Choices, Interface Design, and Labor

Social media platforms moderate content for each user by incorporating t...
research
07/05/2017

Like trainer, like bot? Inheritance of bias in algorithmic content moderation

The internet has become a central medium through which `networked public...
research
07/30/2023

Anatomy of an AI-powered malicious social botnet

Large language models (LLMs) exhibit impressive capabilities in generati...
research
04/03/2019

Health and Kinship Matter: Learning About Direct-To-Consumer Genetic Testing User Experiences via Online Discussions

Direct-to-consumer (DTC) genetic testing has gained in popularity over t...
research
07/23/2018

Influence of Selective Exposure to Viewing Contents Diversity

Personalization, including both self-selected and pre-selected, is inevi...
research
01/17/2023

The Feasibility of Algorithmic Detection and Decentralised Moderation for Protecting Women from Online Abuse

Online abuse is becoming an increasingly prevalent issue in modern-day s...

Please sign up or login with your details

Forgot password? Click here to reset