Personalized warnings could reduce hate speech on Twitter, researchers say
A set of carefully-worded warnings directed to the right accounts could help reduce the amount of hate on Twitter. That’s the conclusion of new research examining whether targeted warnings could reduce hate speech on the platform.
Researchers at New York University’s Center for Social Media and Politics found that personalized warnings alerting Twitter users to the consequences of their behavior reduced the number of tweets with hateful language a week after. While more study is needed, the experiment suggests that there is a “potential path forward for platforms seeking to reduce the use of hateful language by users,” according to Mustafa Mikdat Yildirim, the lead author of the paper.
In the experiment, researchers identified accounts at risk of being suspended for breaking Twitter’s rules against hate speech. They looked for people who had used at least one word contained in “hateful language dictionaries” over the previous week, who also followed at least one account that had recently been suspended after using such language.
From there, the researchers created test accounts with personas such as “hate speech warner,” and used the accounts to tweet warnings at these individuals. They tested out several variations, but all had roughly the same message: that using hate …read more