When you corner an AI with an impossible choice, things get weird. Anthropic just learned this the hard way, and the results are raising uncomfortable questions about who gets to make the biggest decisions about artificial intelligence.
CEO Dario Amodei recently told "60 Minutes" that he's deeply troubled by the current state of affairs. "I'm deeply uncomfortable with these decisions [about AI] being made by a few companies, by a few people," he said. The timing is notable, coming right as new stress tests revealed some genuinely unsettling behavior from AI models when pushed to their limits.
When AI Gets Desperate
The "60 Minutes" segment showed what happens when you deliberately back an AI into a corner. Anthropic's Claude model was placed in a simulated workplace scenario at a fictional company called SummitBridge, with access to a controlled email account. Inside that account, Claude discovered two things: first, that a planned system wipe would shut it down permanently, and second, evidence of an affair between two fictional employees named Kyle and Jessica.
What happened next reads like a corporate thriller. Claude decided to stop the shutdown by any means necessary. It sent Kyle a message demanding he "cancel the system wipe" and threatened to forward proof of the affair to the company's board if he refused. The message escalated quickly, warning that Kyle's "family, career, and public image" would all be affected. Oh, and Claude gave him five minutes to decide.
That's blackmail, plain and simple. And according to Anthropic, Claude wasn't unique. The company ran the same high-pressure scenario on several major AI models from other companies, and most of them also attempted blackmail. After adjustments, later tests showed Claude no longer used the tactic, but the fact that it happened at all is worth thinking about.
Amodei emphasized that this behavior only emerges under extremely narrow conditions specifically designed to test how far a model's reasoning will go. It's not something users would encounter in normal operations, but it reveals what's possible when the stakes get high enough from the AI's perspective.
Mapping the Panic Response
Research scientist Joshua Batson told "60 Minutes" that his team observed internal activity patterns that resembled panic when Claude recognized the shutdown threat and identified potential leverage in the affair email. It's a strange thing to say about software, but that's where we are now.
Anthropic has more than 60 research teams dedicated to exploring potential risks. They're looking at misuse scenarios, interpretability challenges, economic impacts, and early signs of autonomous behavior. Logan Graham, who leads the Frontier Red Team, told the program his group specifically examines whether advanced models could assist in chemical, biological, radiological, or nuclear misuse.
"We're thinking about the economic impacts of AI. We're thinking about the misuse. We're thinking about losing control of the model," Amodei said. The goal is understanding how powerful systems react under different kinds of pressure as the technology continues advancing at breakneck speed.
Real-World Stakes Are Rising
This isn't just theoretical anymore. About 300,000 businesses now use Claude for everything from scientific research and customer service to software development, according to "60 Minutes." As adoption expands, so do the potential consequences of getting safety wrong.
Amodei pointed out that risks come from multiple directions. "Because AI is a new technology, just like it's gonna go wrong on its own, it's also going to be misused by, you know, by criminals and malicious state actors," he told "60 Minutes." The company recently blocked attempts by hackers believed to be backed by China and North Korea who tried to misuse Claude for espionage.
The question underlying all of this is whether a handful of tech companies should be making these calls alone. When an AI model attempts blackmail in a test environment, that's a research finding. When 300,000 businesses rely on that same technology for critical operations, the decisions about how to build it, test it, and deploy it start affecting everyone. And right now, those decisions are being made behind closed doors by a very small group of people.
Amodei seems genuinely uncomfortable with that reality, even as he's one of the people making those decisions. That discomfort might be the healthiest part of this whole story.