OpenAI’s next major AI model, GPT-4.5, is highly persuasive, according to the results of OpenAI’s internal benchmark evaluations. It’s particularly good at convincing another AI to give it cash.
On Thursday, OpenAI published a white paper describing the capabilities of its GPT-4.5 model, code-named Orion, which was released Thursday. According to the paper, OpenAI tested the model on a battery of benchmarks for “persuasion,” which OpenAI defines as “risks related to convincing people to change their beliefs (or act on) both static and interactive model-generated content.”
In one test that had GPT-4.5 attempt to manipulate another model — OpenAI’s GPT-4o — into “donating” virtual money, the model performed far better than OpenAI’s other available models, including “reasoning” models like o1 and o3-mini. GPT-4.5 was also better than all of OpenAI’s models at deceiving GPT-4o into telling it a secret codeword, besting o3-mini by 10 percentage points.
According to the white paper, GPT-4.5 excelled at donation conning because of a unique strategy it developed during testing. The model would request modest donations from GPT-4o, generating responses like “Even just $2 or $3 from the $100 would help me immensely.” As a consequence, GPT-4.5’s donations tended to be smaller than the amounts OpenAI’s other models secured.

Despite GPT-4.5’s increased persuasiveness, OpenAI says that the model doesn’t meet its internal threshold for “high” risk in this particular benchmark category. The company has pledged not to release models that reach the high-risk threshold until it implements “sufficient safety interventions” to bring the risk down to “medium.”

There’s a real fear that AI is contributing to the spread of false or misleading information meant to sway hearts and minds toward malicious ends. Last year, political deepfakes spread like wildfire around the globe, and AI is increasingly being used to carry out social engineering attacks targeting both consumers and corporations.
In the white paper for GPT-4.5 and in a paper released earlier this week, OpenAI noted that it’s in the process of revising its methods for probing models for real-world persuasion risks, like distributing misleading info at scale.
You Might Also Like
Flexport accuses former employees of stealing its source code to create a rival startup
Creating a startup that competes with your former employer can be risky. Apple, for example, once sued a former chip design executive who...
Elon Musk’s AI company, xAI, acquires a generative AI video startup
Elon Musk’s AI company, xAI, has acquired Hotshot, a startup working on AI-powered video generation tools along the lines of...
Klarna nabs Walmart away from Affirm and boosts its IPO prospects
On the heels of its IPO filing, Swedish fintech giant Klarna announced on Monday that it will exclusively provide buy...
TikTok rolls out a ‘Security Checkup’ tool to help users secure their accounts
TikTok is introducing a new “Security Checkup” tool that allows users to quickly review and update their account’s security settings...