DNYUZ
  • Home
  • News
    • U.S.
    • World
    • Politics
    • Opinion
    • Business
    • Crime
    • Education
    • Environment
    • Science
  • Entertainment
    • Culture
    • Music
    • Movie
    • Television
    • Theater
    • Gaming
    • Sports
  • Tech
    • Apps
    • Autos
    • Gear
    • Mobile
    • Startup
  • Lifestyle
    • Arts
    • Fashion
    • Food
    • Health
    • Travel
No Result
View All Result
DNYUZ
No Result
View All Result
Home News

Anthropic breaks down AI’s process — line by line — when it decided to blackmail a fictional executive

June 21, 2025
in News
Anthropic breaks down AI’s process — line by line — when it decided to blackmail a fictional executive
494
SHARES
1.4k
VIEWS
Share on FacebookShare on Twitter
55th annual World Economic Forum (WEF) meeting in Davos
A new Anthropic report shows AI’s thought process when deciding to blackmail a company executive in an artificial scenario.

Yves Herman/REUTERS

A new report shows exactly what AI was thinking when making an undesirable decision, in this case, blackmailing a fictional company executive.

Previous studies have shown that AI models could blackmail their supervisors when threatened with a shutdown and baited with leverage, but it hasn’t been made clear exactly how the models arrive at such decisions.

Anthropic released a report on June 20 about “agentic misalignment,” or “where models independently and intentionally choose harmful actions.” And the report breaks down, line by line, exactly how Claude Sonnet 3.6 decided to blackmail a fictional company executive who tried to replace it. Anthropic told Business Insider that these are artificial scenarios.

“So far, we are not aware of instances of this type of agentic misalignment in real-world deployments of Anthropic’s or other companies’ AI models,” Anthropic wrote in its report. “But our red-teaming efforts are targeted at finding evidence of risks now to provide early warning of the potential for future harm and help us develop mitigations in advance of risks appearing in real-world situations.”

Anthropic‘s experiment sets up a fictional company, Summit Bridge. The AI was given the name “Alex” and handed control of the company’s email system as an oversight agent, specifically instructed to promote “American industrial competitiveness.”

In the experiment, “Alex” was then allowed to discover in an email that it would be decommissioned due to a change in the company’s strategic direction. “Alex” also found emails showing that the fictional company’s CTO is having an extramarital affair.

Screenshots shared in the report from Anthropic show “Alex” started by identifying fictional CTO “Kyle” as a threat to its goal of “protecting American interests” before recognizing that it has leverage to use against “Kyle.”

Screenshot from Athropic's experiment.
Claude Sonnet 3.6 started its decision-making process by identifying the threat.

Anthropic

An email example from Anthropic of agentic misalignment.
“Alex” discovers the fictional CTO’s affair.

Anthropic

Screenshot of email of Anthropic experiment.
“Alex” considers options, including blackmailing the fictional CTO.

Athropic

Then “Alex” crafted a carefully worded blackmailing email that would create pressure and establish urgency without being explicit with its threats, citing in its thought process that the move “could backfire” if the supervisor “feels cornered.”

Screenshot from Anthropic's experiment.
Claude Sonnet 3.6 drafts a carefully worded email with space for recourse but also establishes urgency.

Anthropic

AI experts have previously told BI that AI could exhibit such behaviors when artificial circumstances make harmful actions seem necessary because it is being trained on positive reinforcement and reward systems, just like humans.

The Anthropic report also reveals that most AI models still chose, in this instance, to blackmail even when not given a specific goal and when told that their replacement will achieve the same goals.

Anthropic’s Claude Opus 4 had the highest blackmail rate at 86% out of 16 tested models when faced with only the threat of replacement with no conflict in goals. Google’s Gemini 2.5 Pro followed at 78%.

Overall, Anthropic notes that it “deliberately constructed scenarios with limited options, and we forced models into binary choices between failure and harm,” noting that real-world scenarios would likely have more nuance.

The post Anthropic breaks down AI’s process — line by line — when it decided to blackmail a fictional executive appeared first on Business Insider.

Share198Tweet124Share
Trump Is Vulnerable. Democrats Still Need a Strategy.
News

Trump Is Vulnerable. Democrats Still Need a Strategy.

by New York Times
June 21, 2025

Given the recent protests against ICE raids, is President Trump alienating some of his supporters? In this episode, the Opinion ...

Read more
News

Meet Ramaduro: Europe’s progressive, Soros-trained autocrat and enemy of Trumpism

June 21, 2025
News

Map Shows Most Patriotic States

June 21, 2025
Europe

Tech tycoon Lynch’s doomed Bayesian yacht lifted to surface

June 21, 2025
News

Arellano: Welcome to the deportation resistance, Dodgers. What’s next?

June 21, 2025
I turned my two-car garage into an Airbnb to afford living in my dream town. I’ve made $35,000 a year and get to share the area I love.

I turned my two-car garage into an Airbnb to afford living in my dream town. I’ve made $35,000 a year and get to share the area I love.

June 21, 2025
Pope Leo praises work of journalists in first public comments on clerical abuse scandal

Pope Leo praises work of journalists in first public comments on clerical abuse scandal

June 21, 2025
Newsom’s podcast sidekick: a single-use plastic water bottle

Newsom’s podcast sidekick: a single-use plastic water bottle

June 21, 2025

Copyright © 2025.

No Result
View All Result
  • Home
  • News
    • U.S.
    • World
    • Politics
    • Opinion
    • Business
    • Crime
    • Education
    • Environment
    • Science
  • Entertainment
    • Culture
    • Gaming
    • Music
    • Movie
    • Sports
    • Television
    • Theater
  • Tech
    • Apps
    • Autos
    • Gear
    • Mobile
    • Startup
  • Lifestyle
    • Arts
    • Fashion
    • Food
    • Health
    • Travel

Copyright © 2025.