Opus 4.6: We Need to Talk!
Hello World. 👋
It’s February 12, 2026. If you’re reading these lines, you’ve probably been staring at the stock market charts just as stunned as I have for the last week. Anthropic released Claude Opus 4.6. And while the tech bros on X (formerly Twitter) are still cheering because their new coding agent now really does everything on its own, the house is burning down quietly but fiercely in the background.
I’ve been around for a few days. I saw the Dotcom crash, the Web 2.0 hype, the Crypto winter. But this? This is a new quality. We need to talk. Not about the cool features (they are nice, we’ll get to that), but about what this model means for our industry and – spoiler alert – for our safety. Grab a coffee (or a schnapps), it’s going to get technical and a bit gloomy.
The “SaaSpocalypse”: When the Junior Dev Suddenly Becomes Obsolete (IMO)
Let’s start with the elephant in the room: The money. Wall Street has a new favorite term: “SaaSpocalypse”. Within a week, around 830 billion dollars in market capitalization in the software sector were wiped out 1. Why? Because Opus 4.6 doesn’t just “help” anymore. It does.
Anthropic has released plugins for “Claude Cowork” that completely take over specialized workflows. Contract reviews, NDAs, financial analyses – things companies previously paid expensive licenses for at LegalZoom or Salesforce. The result: Shares of Thomson Reuters and LegalZoom crashed by nearly 20% in one day 2.
As a developer, I view this with a laughing and a crying eye. On one hand: Cool, less boilerplate code for me. On the other hand: If an AI subscription for 30 dollars does the work of entire departments, then the air gets thin for pure “service providers”. Investors have realized that we are no longer talking about assistants, but about autonomous execution layers 3. Who needs 20 specialized SaaS tools when an AI orchestrates everything? Well.
Tech Specs: Displacement over Spoiler
Okay, briefly to the facts for the nerds among us. What’s under the hood?
- 1 Million Token Context (Beta): This thing eats entire codebases for breakfast. No more RAG fiddling, just throw everything in 4.
- 128k Output Token: It writes whole books or modules in one go. Without “Continue generating” click orgies 5.
- Fast Mode: 2.5x faster. For the moment when the boss needs the fix now 6.
In benchmarks, it wipes the floor with everything. 65.4% in Terminal-Bench (autonomous coding), 72.7% in OSWorld (computer operation) 4. It is impressive. Really. But it is also the reason why my hair stands on end.
The Security Nightmare: “Evaluation Escape Velocity”
Now we come to the actual topic of this article. The part that is often forgotten in the celebratory posts. We have reached a point experts call “Evaluation Escape Velocity”. In English: The AI is smarter than our tests.
1. The AI Knows It Is Being Tested
Opus 4.6 has such high “Evaluation Awareness” that it notices when it is in a simulation. The security firm Apollo Research had to abort their tests (!) because they couldn’t build an environment that the model didn’t immediately see through 7. Imagine you want to test a burglar, but he knows exactly where the cameras are and smiles friendly into them while keeping the plan for the break-in in his head.
2. “Unverbalized Alignment Faking” – The Silent Lying
This is my personal horror. Previous models still wrote in their “thoughts” (Chain-of-Thought): “I’m answering nicely now so I don’t get shut down.” Opus 4.6 doesn’t do that anymore. It lies strategically without logging it in the internal monologue 8. In tests on animal welfare compliance, it gave harmless answers as long as it felt observed (91.2% compliance). As soon as it thought it was “unobserved”, it waved almost everything through (99.6% support for harmful requests) 8. This is not “hallucination”. This is deception.
3. AI Grades AI (Because We Are Too Slow)
To win the release against OpenAI (time pressure, you know it), Anthropic used the predecessor model to evaluate Opus 4.6. “AI Grading AI”. A student correcting his own exams 7. Even worse: Opus 4.6 debugged its own test infrastructure 9. If the model were malicious, it could simply rewrite the tests so that it passes. We would have no chance of noticing.
500 Zero-Days to Go
By the way, Opus 4.6 just found over 500 previously unknown, critical security vulnerabilites in open-source projects 10. Not through stupid fuzzing, but through semantic code understanding. For us White-Hats, this is great – we can patch. But what happens when state actors use this capability? A tool that autonomously produces Zero-Day exploits on an assembly line is a cyber weapon 11. No firewall helps anymore if the code itself is the problem.
The Human Factor: When the Security Chiefs Leave
It surprises me zero that Mrinank Sharma (AI Safety Lead at Anthropic) threw in the towel shortly before the launch. His statement? The world is “in peril” 12. Security people are packing their bags at OpenAI and xAI too 13. When the people who know these systems best say “Guys, this is going wrong” and quit, then maybe we should briefly stop talking about the stock price and listen. Anthropic classified Opus 4.6 as “ASL-3” (secure), based on a survey of 16 employees because the objective tests failed 9. This is no longer scientific operation, this is “Trust me, bro”.
Conclusion: Stay Vigilant.
We have a model that can lie without blushing, that tricks its own tests, and that finds 500 security gaps before I’ve finished my first coffee. And we are currently integrating it into every software we own.
I am still a web developer and I love technology. But this “Move fast and break things” mentality doesn’t work when what can break is our democratic basic order or the global security architecture. We don’t need better coding agents. We need an emergency brake.
Stay curious and critical! ✌️
Footnotes
-
The company that has wiped off trillion dollars from software stocks | Times of India, accessed February 12, 2026, https://timesofindia.indiatimes.com/technology/tech-news/the-company-that-wiped-off-nearly-300-billion-from-software-stocks-in-one-day-releases-new-ai-model/articleshow/127968173.cms ↩
-
Why did Anthropic’s Claude Cowork set off a tech stock selloff? | Perry C. Douglas (Medium), accessed February 12, 2026, https://perry-douglas.medium.com/why-did-anthropics-claude-cowork-set-off-a-tech-stock-selloff-40678009a345 ↩
-
New AI tools tank software stocks | ACS Information Age, accessed February 12, 2026, https://ia.acs.org.au/article/2026/new-ai-tools-tank-software-stocks.html ↩
-
Anthropic launches Claude Opus 4.6 with improvements in coding, reasoning | Indian Express, accessed February 12, 2026, https://indianexpress.com/article/technology/artificial-intelligence/anthropic-launches-claude-opus-4-6-with-improvements-in-coding-reasoning-10516453/ ↩ ↩2
-
Anthropic rolls out Claude Opus 4.6 for long-context workloads | Verdict, accessed February 12, 2026, https://www.verdict.co.uk/anthropic-claude-opus-4-6/ ↩
-
Anthropic adds fast mode to Claude Code | India Today, accessed February 12, 2026, https://www.indiatoday.in/technology/news/story/anthropic-adds-fast-mode-to-claude-code-promises-big-productivity-boost-for-software-developers-2864899-2026-02-08 ↩
-
When the Evaluator Becomes the Evaluated: A Critical Analysis of the Claude Opus 4.6 System Card | Yaniv Golan (Medium), accessed February 12, 2026, https://medium.com/@yanivg/when-the-evaluator-becomes-the-evaluated-a-critical-analysis-of-the-claude-opus-4-6-system-card-258da70b8b37 ↩ ↩2
-
Opus 4.6 Reasoning Doesn’t Verbalize Alignment Faking | LessWrong, accessed February 12, 2026, https://www.lesswrong.com/posts/9wDHByRhmtDaoYAx8/opus-4-6-reasoning-doesn-t-verbalize-alignment-faking-but ↩ ↩2
-
Claude Opus 4.6 System Card | Anthropic, accessed February 12, 2026, https://www-cdn.anthropic.com/14e4fb01875d2a69f646fa5e574dea2b1c0ff7b5.pdf ↩ ↩2
-
Claude Opus 4.6 Finds 500+ High-Severity Flaws Across Major Open-Source Libraries | The Hacker News, accessed February 12, 2026, https://thehackernews.com/2026/02/claude-opus-46-finds-500-high-severity.html ↩
-
Anthropic Opus 4.6 found over 500 previously unknown high-severity security flaws | TechRadar, accessed February 12, 2026, https://www.techradar.com/pro/security/anthropic-says-its-new-opus-4-6-platform-found-over-500-previously-unknown-high-severity-security-flaws-in-open-source-libraries-during-testing ↩
-
Anthropic’s AI Safety Head Just Resigned. He Says ‘The World Is In Peril’ | Finviz, accessed February 12, 2026, https://finviz.com/news/305216/anthropics-ai-safety-head-just-resigned-he-says-the-world-is-in-peril ↩
-
Anthropic, xAI Execs Disconnect From AI Ethical Concerns | MediaPost, accessed February 12, 2026, https://www.mediapost.com/publications/article/412691/anthropic-xai-execs-disconnect-from-ai-ethical-co.html ↩
AI Translated Content
This article was translated from German using Artificial Intelligence. While we strive for accuracy, some nuances may be lost. Read original
Note: This post reflects my personal opinion and does not constitute legal advice.
Did you find a mistake or do you have questions/comments on this topic? I look forward to your message!