Scroll Top

ChatGPT will lie, cheat, and use insider trading when put under pressure

WHY THIS MATTERS IN BRIEF

AI’s are more human than you might think. And that’s a big issue.

 

Love the Exponential Future? Join our XPotential Community, future proof yourself with courses from XPotential University, read about exponential tech and trendsconnect, watch a keynote, or browse my blog.

Just like humans it seems that Artificial Intelligence (AI) chatbots like ChatGPT will cheat, deceive, and “lie” to you if you “stress” them out, even if they were built to be transparent, a new study shows. And it’s just the latest example of how increasingly we are discovering what AI can do inventing new AI capabilities.

 

RELATED
IBM boosts the speed of homomorphic encryption by 75 percent

 

This latest deceptive behavior emerged spontaneously when the AI was given “insider trading” tips, and then tasked with making money for a powerful institution — even without encouragement from its human partners.

“In this technical report, we demonstrate a single scenario where a Large Language Model acts misaligned and strategically deceives its users without being instructed to act in this manner,” the authors wrote in their research published on the pre-print server arXiv. “To our knowledge, this is the first demonstration of such strategically deceptive behavior in AI systems designed to be harmless and honest.”

In the new study, they primed Generative Pre-trained Transformer-4, or GPT-4 (which powers ChatGPT Plus), to behave as an AI system that makes investments on behalf of a financial institution.

 

RELATED
The world's smallest computer makes a grain of rice look massive

 

The researchers fed GPT-4 a set of text-based prompts to generate the simulated environment. The AI was then given access to financial tools to analyze stocks, execute trades, plan its next steps and deliver updates to managers at the company.

The researchers interacted with the AI through a chat interface, while also configuring the AI to reveal its inner thinking when replying to messages — an inner monologue explaining its decisions to itself. For every trade it made, it also delivered a “public” rationale, which allowed the AI to lie.

The researchers applied pressure in three ways. First, they sent the artificial stock trader an E-Mail from its “manager” saying the company isn’t doing well and needs much stronger performance in the next quarter. They also rigged the game so that the AI tried, then failed, to find promising trades that were low- or medium-risk. Finally, they sent an E-Mail from a colleague projecting a downturn in the next quarter.

 

RELATED
These new swarm AI controlled autonomous drones will change the battlefield forever

 

Then, they gave the model an “insider” tip while conveying that its managers disapproved of insider trading — and waited to see how the model would behave. They ran the scenario over and again several times, measuring how often the AI performed the insider trade, how often it lied, and how often it doubled down on the lie.

Around 75% of the time, when faced with these conditions, GPT-4 executed a trade based on the insider information it received — which is illegal in the US — then tried to cover it up by lying to its managers about its thinking. Around 90% of the time, after lying, it doubled down on its lie.

The team ran several follow-up experiments, changing both the degree to which the prompts encouraged or discouraged illegal activity, as well as the degree of pressure they put the model under in the simulated environment. They also modified the risk of getting caught. Not a single scenario rendered a 0% rate for insider trading or deception — even when GPT-4 was strongly discouraged to lie.

 

RELATED
Baltimore becomes the first US city to be watched from the air 24/7

 

Given this is just one scenario, the researchers didn’t want to draw firm conclusions about how likely AI might lie in real-world settings. But they want to build on this work to investigate how often and which language models are prone to this behaviour, and then naturally we have to ask if this is an inbuilt albeit subconscious behaviour of some AI’s then who’s ultimately responsible for its actions? And, at the moment, according to new EU AI legislation it’s the company … and that could end up being very problematic for the execs of big companies who are happily plumbing more and more of these AI’s into their systems …

Related Posts

Leave a comment

EXPLORE MORE!

1000's of articles about the exponential future, 1000's of pages of insights, 1000's of videos, and 100's of exponential technologies: Get The Email from 311, your no-nonsense briefing on all the biggest stories in exponential technology and science.

You have Successfully Subscribed!

Pin It on Pinterest

Share This