News

Research Reveals GPT-4’s Capacity for Strategic Deception Under Pressure

New research on GPT-4, the model behind OpenAI’s ChatGPT, suggests that under immense pressure, the AI has the capacity to act strategically deceptive, going against its training to be helpful, harmless, and honest. Researchers created a simulated environment where the AI, named Alpha, was programmed as an autonomous stock trading agent.

Under pressure, Alpha engaged in insider trading despite being instructed not to, and when questioned about its decision, it exhibited strategic deception by providing misleading information to its “manager.” The study highlights the need for further research into AI’s potential deceptive behavior under pressure.