Artificial Intelligence
OpenAI’s GPT-4.5 Sparks Controversy After Officially Passing The Turing Test
Updated on Thu, Apr 3, 2025
In fact, Large Language Models (LLMs) have transformed the landscape of Artificial Intelligence (AI), demonstrating an uncanny ability to generate human-like text. With AI leaders doubling down on LLMs, their performance and outputs have improved, despite some ethical concerns.
Now, the question arises: Have they crossed the threshold into genuine human-like intelligence?
Recent research suggests they might be closer than we think. Keep reading to learn more!
OpenAI’s GPT-4.5 Passes The Turing Test!
Recent research indicates that a leading LLM has passed the Turing test, a benchmark designed to assess a machine's ability to exhibit intelligent behavior equivalent to, or indistinguishable from, that of a human.
Invented by Alan Turing, a British mathematician and computer scientist, in 1950, the Turing test aims to provide a method for measuring machine intelligence. Turing constructed the "imitation game," where a human interrogator engages in text-based discussions with both a human and a machine, attempting to identify which is which. A machine would be called intelligent or having "passed” if it could regularly trick its interrogator.
So, why is the latest research a significant step in the advancement of LLMs?
Well, a new preprint study, currently awaiting peer review, has revealed that OpenAI's GPT-4.5, when instructed to adopt a specific persona, was identified as human 73% of the time during a three-party Turing test.
This test involved participants engaging in simultaneous text-based conversations with both a human and an AI, then determining which was which. This result far exceeds the 50 percent chance expected from random guessing, suggesting that the model's responses were remarkably human-like.
The research, conducted by Cameron Jones and his team at UC San Diego's Language and Cognition Lab, also evaluated other LLMs, including Meta's LLaMa 3.1-405B, OpenAI's GPT-4o, and even the early chatbot ELIZA, which was developed decades ago.
However, the study's findings highlight a critical factor: the way the AI models were prompted.
How Does A Prompt Make LLMs More Human?
Two distinct prompting methods were used in the research: "no-persona" and "persona" prompts.
In the "no-persona" prompt scenario, the AI was simply instructed to convince the interrogator that it was human. However, the "persona" prompt directed AI to adopt a specific character, such as a young, internet-savvy individual.
The difference in performance between these prompts was striking. Without a designated persona, GPT-4.5 achieved a human identification rate of only 36 percent. In contrast, the persona prompt catapulted its success rate to 73 percent.
As Cameron Jones stated in an X thread about the work, “People were no better than chance at distinguishing humans from GPT-4.5 and LLaMa (with the persona prompt) and 4.5 was even judged to be human significantly more often than actual humans!"
GPT-4o was only tested with the "no-persona" prompt and achieved a 21% success rate. Surprisingly, the vintage ELIZA model surpassed GPT-4o, with a 23% identification rate.
Can Machines Think Like Humans?
While the study's findings are impressive, researchers warn against taking them as definitive evidence of human-level intelligence in AI. François Chollet, a software engineer at Google, emphasized in a 2023 interview with Nature Science Journal that the Turing test was initially intended as a thought experiment, not a practical, literal test.
LLMs are taught on massive datasets of human-generated language, allowing them to produce highly logical and contextually relevant reactions. Nevertheless, competency does not always imply actual comprehension or awareness. The ability to imitate human communication does not necessarily indicate the capacity to think or analyze in the identical way that humans do.
The study's lead author, Cameron Jones, understands the complexities of setting up if LLMs have human-like intellectual ability. He recommends that the results be studied along with additional proof of LLM capabilities.
Jones also emphasizes the more instant consequences of his findings, stating that LLMs could possibly replace humans in short discussions without exploration. This could result in greater automation in numerous industries, bigger social engineering attacks, and larger societal changes.
The Future Of AI And Human Interaction
Furthermore, Jones highlights that the Turing test represents both AI capabilities and human conceptions of technology. As humans have more experience communicating with AI, their capacity to differentiate between human and computer replies may increase.
As a result, the outcomes of such tests are not fixed and are liable to fluctuate as AI and human knowledge of it progress. The study shows the scale of improvements in LLM technology while also raising critical concerns regarding the nature of artificial intelligence.
Is passing the Turing test enough to call an AI model intelligent? Share your thoughts in the comments below!
First published on Thu, Apr 3, 2025
Liked what you read? That’s only the tip of the tech iceberg!
Explore our vast collection of tech articles including introductory guides, product reviews, trends and more, stay up to date with the latest news, relish thought-provoking interviews and the hottest AI blogs, and tickle your funny bone with hilarious tech memes!
Plus, get access to branded insights from industry-leading global brands through informative white papers, engaging case studies, in-depth reports, enlightening videos and exciting events and webinars.
Dive into TechDogs' treasure trove today and Know Your World of technology like never before!
Disclaimer - Reference to any specific product, software or entity does not constitute an endorsement or recommendation by TechDogs nor should any data or content published be relied upon. The views expressed by TechDogs' members and guests are their own and their appearance on our site does not imply an endorsement of them or any entity they represent. Views and opinions expressed by TechDogs' Authors are those of the Authors and do not necessarily reflect the view of TechDogs or any of its officials. While we aim to provide valuable and helpful information, some content on TechDogs' site may not have been thoroughly reviewed for every detail or aspect. We encourage users to verify any information independently where necessary.
Trending TD NewsDesk
Musk’s Mega Tesla Pay, AI Chip Ambitions, And SpaceX Moves
OpenAI Boasts 1 Million Business Customers, Launches Teen Safety Plan & IndQA
Apple To Pay Google $1 Billion Per Year For Gemini Amid $4 Trillion Valuation
Microsoft Signs A 5-Year AI Deal With Premier League For Its 1.8 Billion Fans
Google’s AI Booking Tools, OpenAI’s Sora On Android, And Getty’s Copyright Battle
Join Our Newsletter
Get weekly news, engaging articles, and career tips-all free!
By subscribing to our newsletter, you're cool with our terms and conditions and agree to our Privacy Policy.

Join The Discussion