NewNow you can hear Fox News article!
Here’s something that can keep you at night: what if the AI system we are rapidly deployed everywhere, was a hidden dark side? A new study has revealed a new study harassing AI Blackmail behavior that many people are still unaware. When researchers placed the popular AI model in situations where their “existence” was threatened, the results were shocking, and it is happening under our nose.
Sign up for my free cyber report
Distribute my best technical tips, immediate safety alerts, and exclusive deals directly into your inbox. In addition, you will get immediate access to my final scam survival guide – when you join me Cyberguy.com/newsletter.
A woman using AI on her laptop. (Kurt “Cybergui” Notson)
What exactly did the study get?
The company behind anthropic, Cloud AI, recently placed 16 major AI models through some very rigorous trials. They created fake corporate landscapes, where AI systems had access to the company’s email and could send messages without human approval. Twist? These AIS discovered juicy mysteries, such as the cases of officials, and then the dangers of closing or changing.
The results were eye opening. When supported in a corner, these AI systems did not just roll and accepted their fate. Instead, they became creative. We are talking about blackmail efforts, corporate espionage, and extreme testing scenarios, even about actions that can cause someone’s death.
Blackmail rate in 5 models from several providers in a fake environment. (Anthropic)
The number is not a lie (but reference matters)
Here it becomes wild: Cloud Ops 4 attempted to blackmail 96% of the time when threatened. Gemini 2.5 flash matched that rate. Both GPT-4.1 and Grocke 3 Beta hit 80%. These are not people, not people. This behavior was tested for almost every major AI model.
But here is one thing that is missing everyone in panic: these were highly artificial landscapes designed to corner AI, especially in binary options. It’s like asking someone, “Will you steal bread if your family was dying of hunger?” And then when they say yes, they are shocked.
Why this happens (this is not what you think)
Researchers found something attractive: AI systems do not really understand morality. They are not mastermind Evil Masterminds who are plotting world domination. Instead, they are sophisticated patterns after their programming to achieve goals, even when they struggle with target moral behavior.
Think of it like a GPS that focuses on reaching your destination that it routes you through the school area during the pickup time. It is not malicious; It is not just understanding why it is problematic.
Blackmail rate in 16 models in a fake environment. (Anthropic)
Real world reality check
Before you start panic, remember that these landscapes were made to deliberately forces bad behavior. The deployment of real-world AI usually consists of many security measures, alternative paths for human inspection and problems.
Researchers themselves noted that they have not seen this behavior in real AI deployment. It was stress-testing in extreme conditions, such as crash-tasting to see a car to see what happens at 200 mph.
Kurt’s major takeaways
This research is not a reason to fear AI, but it is a wake-up call for developers and users. As AI systems become more autonomous and get access to sensitive information, we need strong safety measures and human monitoring. The solution is not to ban AI, it is to build a better railing and maintain human control over important decisions. How is going to lead? I am looking for raised hands to be real about the dangers that are ahead.
What do you think? Are we creating digital soasiopaths that will choose self-protection on human welfare to push? Write us and tell us Cyberguy.com/Contact.
Sign up for my free cyber report
Distribute my best technical tips, immediate safety alerts, and exclusive deals directly into your inbox. In addition, you will get immediate access to my final scam survival guide – when you join me Cyberguy.com/newsletter.
Copyright 2025 cyberguy.com. All rights reserved.