The next time you encounter an unusually polite reply on social media, you might want to check twice. It could be an AI model trying (and failing) to blend in with the crowd.
On Wednesday, researchers from the University of Zurich, University of Amsterdam, Duke University, and New York University released a study revealing that AI models remain easily distinguishable from humans in social media conversations. The key factor that gives them away? An overly friendly emotional tone. The research tested nine open-weight models across platforms like Twitter/X, Bluesky, and Reddit, finding that classifiers developed by the researchers detected AI-generated replies with an impressive 70 to 80 percent accuracy.
Introducing the Computational Turing Test
The study introduces a concept termed the “computational Turing test” to assess how closely AI models can mimic human language. This framework doesn’t rely solely on subjective opinions regarding the authenticity of text. Instead, it employs automated classifiers and linguistic analysis to identify specific characteristics that distinctively separate machine-generated content from human-written text.
The findings of the research team, led by Nicolò Pagan at the University of Zurich, reveal that even when optimized, the outputs of large language models (LLMs) remain clearly recognizable compared to human text. This distinction is particularly obvious in terms of affective tone and emotional expression. The study indicates that deeper emotional cues persist as reliable indicators that a text interaction online was authored by an AI chatbot rather than a human, no matter how sophisticated the model.
The Toxicity Tell
Researchers evaluated nine large language models: Llama 3.1 8B, Llama 3.1 8B Instruct, Llama 3.1 70B, Mistral 7B v0.1, Mistral 7B Instruct v0.2, Qwen 2.5 7B Instruct, Gemma 3 4B Instruct, DeepSeek-R1-Distill-Llama-8B, and Apertus-8B-2509.
When prompted to generate replies to actual social media posts from real users, these AI models struggled to match the level of casual negativity and spontaneous emotional expression that is common in human interactions online. Notably, the toxicity scores of the AI-generated content were consistently lower than those of authentic human replies across all tested platforms.
To address this deficiency, researchers tried various optimization strategies, including providing writing samples and context retrieval. Although these methods aimed to reduce structural differences such as sentence length or word count, emotional tone variations remained. “Our comprehensive calibration tests challenge the assumption that more sophisticated optimization necessarily yields more human-like output,” the researchers concluded.
This groundbreaking research highlights the current limitations of AI models in mimicking human nuances in social context, reinforcing the need for continued advancements in AI technology.
For a deeper dive into the study and its implications, click Here.
Image Credit: arstechnica.com






