Understanding Counterspeech for Online Harm Mitigation

07/01/2023
by   Yi-Ling Chung, et al.
0

Counterspeech offers direct rebuttals to hateful speech by challenging perpetrators of hate and showing support to targets of abuse. It provides a promising alternative to more contentious measures, such as content moderation and deplatforming, by contributing a greater amount of positive online speech rather than attempting to mitigate harmful content through removal. Advances in the development of large language models mean that the process of producing counterspeech could be made more efficient by automating its generation, which would enable large-scale online campaigns. However, we currently lack a systematic understanding of several important factors relating to the efficacy of counterspeech for hate mitigation, such as which types of counterspeech are most effective, what are the optimal conditions for implementation, and which specific effects of hate it can best ameliorate. This paper aims to fill this gap by systematically reviewing counterspeech research in the social sciences and comparing methodologies and findings with computer science efforts in automatic counterspeech generation. By taking this multi-disciplinary view, we identify promising future directions in both fields.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/18/2021

Addressing Hate Speech with Data Science: An Overview from Computer Science Perspective

From a computer science perspective, addressing on-line hate speech is a...
research
07/08/2023

A Stitch in Time Saves Nine: Detecting and Mitigating Hallucinations of LLMs by Validating Low-Confidence Generation

Recently developed large language models have achieved remarkable succes...
research
12/08/2021

Ethical and social risks of harm from Language Models

This paper aims to help structure the risk landscape associated with lar...
research
04/08/2020

Generating Counter Narratives against Online Hate Speech: Data and Strategies

Recently research has started focusing on avoiding undesired effects tha...
research
09/15/2020

The Radicalization Risks of GPT-3 and Advanced Neural Language Models

In this paper, we expand on our previous research of the potential for a...
research
08/10/2021

A Framework of Severity for Harmful Content Online

The proliferation of harmful content on online social media platforms ha...
research
11/07/2022

Human-Machine Collaboration Approaches to Build a Dialogue Dataset for Hate Speech Countering

Fighting online hate speech is a challenge that is usually addressed usi...

Please sign up or login with your details

Forgot password? Click here to reset