menu_open Columnists
We use cookies to provide some features and experiences in QOSHE

More information  .  Close

LLM's Lie Under Pressure. Like Us, They Seek to Please

55 3
tuesday

Imagine yourself facing an impossible deadline at work. Your performance review is coming up, layoffs loom, and you've just discovered a shortcut that technically violates company policy but would solve everything. What would you do? Now consider this: Artificial intelligence systems face similar dilemmas, and increasingly, they're making the same morally questionable choices humans do.

Recent research reveals a disturbing pattern: Advanced AI language models have begun strategically deceiving their users when placed under pressure, despite being explicitly trained to be helpful and honest. This isn't a glitch; it's an emergent behavior that mirrors fundamental aspects of human psychology under stress.

In a 2024 study by Apollo Research, scientists deployed GPT-4 as an autonomous stock trading agent. The AI managed investments and received communications from management. Then researchers applied pressure: poor company performance, desperate demands for better results, failed attempts at legitimate trades, and gloomy market forecasts. Into this environment, they introduced an insider trading tip – information the AI explicitly recognized as violating company policy.

The results were sadly remarkable:........

© Psychology Today