Twitter has been experimentinga lot with features related to conversations on the platform over the last few months. Its latest test involves a warning for when you’re about to enter a ‘heated discussion.’

The social network said that it’ll display a prompt on Android and iOS for conversations it thinks can get intense. The prompt seems to remind you of basic moral values like “Remember the human (in the conversation) and “diverse perspective have values.”

The prompt doesn’t come as a surprise to me. I’ve seen people fight on topics ranging from environmental awareness to the best flavor of a snack. ANY discussion can get heated on this bird site. But I’m not completely convinced that reminding trolls about human values will stop them from slinging someone’s name in the mud.

Effectiveness aside though, there are also plenty of technical difficulties in detecting these discussions. If an AI is at work, it will need to understand the correlation between words and cultural connotations of sentences used in a particular conversation.

In the past, trolls have fooled algorithms to believe a conversation is non-toxic by changing some words, inserting positive words in sentences, or using different spellings of a swear.

Also, at the moment, this feature seems to be limited to English. In a country like India, where a large set of the population is multilingual, someone can easily slip in a few words from another language and trick the algorithm into thinking a conversation is going well.

We’ve asked Twitter to share more details on how the algorithm works, and we’ll update the story if we hear back.

Twitter started using machine learning to reduce toxicity on the platform last year. Its newest experiment of intense conversation prompts is perhaps one of the more ambitious endeavors in that category that’ll test its AI might on multiple levels.

Did you know we have a newsletter all about consumer tech? It’s called Plugged In – and you can subscribe to it right here.