Handling and Presenting Harmful Text

04/29/2022
by   Leon Derczynski, et al.
1

Textual data can pose a risk of serious harm. These harms can be categorised along three axes: (1) the harm type (e.g. misinformation, hate speech or racial stereotypes) (2) whether it is elicited as a feature of the research design from directly studying harmful content (e.g. training a hate speech classifier or auditing unfiltered large-scale datasets) versus spuriously invoked from working on unrelated problems (e.g. language generation or part of speech tagging) but with datasets that nonetheless contain harmful content, and (3) who it affects, from the humans (mis)represented in the data to those handling or labelling the data to readers and reviewers of publications produced from the data. It is an unsolved problem in NLP as to how textual harms should be handled, presented, and discussed; but, stopping work on content which poses a risk of harm is untenable. Accordingly, we provide practical advice and introduce HarmCheck, a resource for reflecting on research into textual harms. We hope our work encourages ethical, responsible, and respectful research in the NLP community.

READ FULL TEXT
research
10/09/2020

Case Study: Deontological Ethics in NLP

Recent work in natural language processing (NLP) has focused on ethical ...
research
10/09/2019

Exploring Hate Speech Detection in Multimodal Publications

In this work we target the problem of hate speech detection in multimoda...
research
09/14/2021

Just What do You Think You're Doing, Dave?' A Checklist for Responsible Data Use in NLP

A key part of the NLP ethics movement is responsible use of data, but ex...
research
01/20/2021

The Challenges of Persian User-generated Textual Content: A Machine Learning-Based Approach

Over recent years a lot of research papers and studies have been publish...
research
04/08/2020

Generating Counter Narratives against Online Hate Speech: Data and Strategies

Recently research has started focusing on avoiding undesired effects tha...
research
10/14/2022

The State of Profanity Obfuscation in Natural Language Processing

Work on hate speech has made the consideration of rude and harmful examp...

Please sign up or login with your details

Forgot password? Click here to reset