Contact Us

Instagram unveils new tools to reduce abuse, racist comments

Published August 11,2021
Subscribe
Instagram on Wednesday announced new measures to curb abusive and racist content on the image-centric social network, following an onslaught of hateful comments directed at UK footballers after the Euro 2020 final.

The Facebook-owned platform said it would start using "stronger warnings" when people post potentially offensive comments, and a new "hidden words" feature that enables users to filter abusive messages.

Instagram will also allow users to limit comments and message requests during "spikes of increased attention."

The move comes weeks after social platforms scrambled to take down racially abusive comments directed at members of the England football team following a heartbreaking loss in the Euro 2020 final, which was played last month after a delay due to the coronavirus pandemic.

The racist and hateful content, which had prompted condemnation from British political leaders and the public, also led to pressure on the large US social networks such as Facebook and Twitter.

Instagram chief Adam Mosseri said the new measures were designed to reduce the spread of racist, sexist and homophobic content.

"Our research shows that a lot of negativity towards public figures comes from people who don't actually follow them, or who have only recently followed them, and who simply pile on in the moment," Mosseri said in a blog post.

"We saw this after the recent Euro 2020 final, which resulted in a significant -- and unacceptable -- spike in racist abuse towards players."

Mosseri said users "tell us they don't want to switch off comments and messages completely; they still want to hear from their community and build those relationships," and that the new policy "allows you to hear from your long-standing followers, while limiting contact from people who might only be coming to your account to target you."

Mosseri said the warnings already in place reduce the frequency of abusive comments by as much as 50 percent.

"We already show a warning when someone tries to post a potentially offensive comment," he wrote.

"And if they try to post potentially offensive comments multiple times, we show an even stronger warning... Now, rather than waiting for the second or third comment, we'll show this stronger message the first time."