BERT-Beta: A Proactive Probabilistic Approach to Text Moderation

09/18/2021
by   Fei Tan, et al.
0

Text moderation for user generated content, which helps to promote healthy interaction among users, has been widely studied and many machine learning models have been proposed. In this work, we explore an alternative perspective by augmenting reactive reviews with proactive forecasting. Specifically, we propose a new concept text toxicity propensity to characterize the extent to which a text tends to attract toxic comments. Beta regression is then introduced to do the probabilistic modeling, which is demonstrated to function well in comprehensive experiments. We also propose an explanation method to communicate the model decision clearly. Both propensity scoring and interpretation benefit text moderation in a novel manner. Finally, the proposed scaling mechanism for the linear model offers useful insights beyond this work.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/19/2012

Alpha/Beta Divergences and Tweedie Models

We describe the underlying probabilistic interpretation of alpha and bet...
research
09/16/2019

On the Hurwitz zeta function with an application to the exponential-beta distribution

We prove a monotonicity property of the Hurwitz zeta function which, in ...
research
06/07/2019

Real or Fake? Learning to Discriminate Machine from Human Generated Text

Recent advances in generative modeling of text have demonstrated remarka...
research
12/17/2021

Explain, Edit, and Understand: Rethinking User Study Design for Evaluating Model Explanations

In attempts to "explain" predictions of machine learning models, researc...
research
12/01/2019

JNET: Learning User Representations via Joint Network Embedding and Topic Embedding

User representation learning is vital to capture diverse user preference...
research
06/07/2023

A New Family of Regression Models for [0,1] Outcome Data: Expanding the Palette

Beta regression is a popular methodology when the outcome variable y is ...
research
04/13/2020

Reverse Engineering Configurations of Neural Text Generation Models

This paper seeks to develop a deeper understanding of the fundamental pr...

Please sign up or login with your details

Forgot password? Click here to reset