Prompting users to avoid harmful language

We experimented with Google Perspective API Plugin, but ran into similar issues as described in the article you linked… it struggled on more nuanced language, and would sometimes pick up on some non-offensive language as offensive… and these false positives can be offensive in their own right!

However, Twitter’s early tests ran into some problems. It found its systems and algorithms sometimes struggled to understand the nuance that occurs in many conversations. For example, it couldn’t always differentiate between offensive replies and sarcasm or, sometimes, even friendly banter. It also struggled to account for those situations in which language is being reclaimed by underrepresented communities, and then used in nonharmful ways.

9 Likes