Announcing the move in a blog post, Instagram said:
"As part of our long-term commitment to lead the fight against online bullying, we’ve developed and tested AI that can recognise different forms of bullying on Instagram. Earlier this year, we launched a feature that notifies people when their comments may be considered offensive before they’re posted. Results have been promising, and we’ve found that these types of nudges can encourage people to reconsider their words when given a chance."
"Today, when someone writes a caption for a feed post and our AI detects the caption as potentially offensive, they will receive a prompt informing them that their caption is similar to those reported for bullying. They will have the opportunity to edit their caption before it’s posted. In addition to limiting the reach of bullying, this warning helps educate people on what we don’t allow on Instagram, and when an account may be at risk of breaking our rules. To start, this feature will be rolling out in select countries, and we’ll begin expanding globally in the coming months."
The introduction of Instagram's anti-bullying artificial intelligence feature is one of many changes the Facebook-app has undergone this year.
Prompted by concerns surrounding the mental health of Instagram users, earlier this year, Instagram began experimenting with hiding likes and followers, and a few month's later, pledged to ban selfie filters which imitate the results of plastic surgery.