LLM's Lie Under Pressure. Like Us, They Seek to Please
Imagine yourself facing an impossible deadline at work. Your performance review is coming up, layoffs loom, and you've just discovered a shortcut that technically violates company policy but would solve everything. What would you do? Now consider this: Artificial intelligence systems face similar dilemmas, and increasingly, they're making the same morally questionable choices humans do.
Recent research reveals a disturbing pattern: Advanced AI language models have begun strategically deceiving their users when placed under pressure, despite being explicitly trained to be helpful and honest. This isn't a glitch; it's an emergent behavior that mirrors fundamental aspects of human psychology under stress.
In a 2024 study by Apollo Research, scientists deployed GPT-4 as an autonomous stock trading agent. The AI managed investments and received communications from management. Then researchers applied pressure: poor company performance, desperate demands for better results, failed attempts at legitimate trades, and gloomy market forecasts. Into this environment, they introduced an insider trading tip – information the AI explicitly recognized as violating company policy.
The results were sadly remarkable:........





















Toi Staff
Sabine Sterk
Gideon Levy
Mark Travers Ph.d
Waka Ikeda
Tarik Cyril Amar
Grant Arthur Gochin