"HOT" ChatGPT: The promise of ChatGPT in detecting and discriminating hateful, offensive, and toxic comments on social media

04/20/2023
by   Lingyao Li, et al.
0

Harmful content is pervasive on social media, poisoning online communities and negatively impacting participation. A common approach to address this issue is to develop detection models that rely on human annotations. However, the tasks required to build such models expose annotators to harmful and offensive content and may require significant time and cost to complete. Generative AI models have the potential to understand and detect harmful content. To investigate this potential, we used ChatGPT and compared its performance with MTurker annotations for three frequently discussed concepts related to harmful content: Hateful, Offensive, and Toxic (HOT). We designed five prompts to interact with ChatGPT and conducted four experiments eliciting HOT classifications. Our results show that ChatGPT can achieve an accuracy of approximately 80 displays a more consistent classification for non-HOT comments than HOT comments compared to human annotations. Our findings also suggest that ChatGPT classifications align with provided HOT definitions, but ChatGPT classifies "hateful" and "offensive" as subsets of "toxic." Moreover, the choice of prompts used to interact with ChatGPT impacts its performance. Based on these in-sights, our study provides several meaningful implications for employing ChatGPT to detect HOT content, particularly regarding the reliability and consistency of its performance, its understand-ing and reasoning of the HOT concept, and the impact of prompts on its performance. Overall, our study provides guidance about the potential of using generative AI models to moderate large volumes of user-generated content on social media.

READ FULL TEXT

page 19

page 24

research
03/27/2022

bitsa_nlp@LT-EDI-ACL2022: Leveraging Pretrained Language Models for Detecting Homophobia and Transphobia in Social Media Comments

Online social networks are ubiquitous and user-friendly. Nevertheless, i...
research
04/03/2023

Detection of Homophobia Transphobia in Dravidian Languages: Exploring Deep Learning Methods

The increase in abusive content on online social media platforms is impa...
research
11/01/2021

AutoShard – Declaratively Managing Hot Spot Data Objects in NoSQL Document Stores

NoSQL document stores are becoming increasingly popular as backends in w...
research
02/01/2023

Netizens, Academicians, and Information Professionals' Opinions About AI With Special Reference To ChatGPT

This study aims to understand the perceptions and opinions of academicia...
research
08/12/2020

Social Media and Health Misinformation during the US COVID Crisis

Health misinformation has been found to be prevalent on social media, pa...
research
10/19/2022

Automated Content Moderation Increases Adherence to Community Guidelines

Online social media platforms use automated moderation systems to remove...
research
11/11/2020

Policing Chronic and Temporary Hot Spots of Violent Crime: A Controlled Field Experiment

Hot-spot-based policing programs aim to deter crime through increased pr...

Please sign up or login with your details

Forgot password? Click here to reset