Back to Blog

Watch What the Car Thief Drives

Microsoft just told you which AI company to trust — by choosing not to use their own

By Catalin Lichi · Sugau


There is an old piece of consumer wisdom that has survived decades because it is almost always correct: watch what the expert chooses for themselves.

Watch what the mechanic drives. Watch what the chef eats on their day off. Watch what the security researcher runs on their personal machine.

And watch — very carefully — what Microsoft integrates into its own critical infrastructure when it thinks nobody is framing it as a story.

Because in April 2026, Microsoft announced it would embed Anthropic’s Claude Mythos Preview into its Security Development Lifecycle — the framework Microsoft uses to harden its own software against vulnerabilities before it ships to the world.

Microsoft. The company that owns Copilot. The company that has invested billions into OpenAI. The company that has been telling every enterprise customer on earth that AI-assisted development means Microsoft AI.

When it needed AI for the use case that actually mattered — securing its own code — it reached for Claude.

That is not a partnership announcement. That is a verdict.


The Most Consistent Talent Scout in Technology History

To understand what Microsoft’s choice means, you need to understand Microsoft’s history as a technology evaluator.

Microsoft has spent forty years being wrong about almost everything early and right about almost everything eventually. They missed the internet, then built Internet Explorer and briefly owned it. They missed search, then spent $44 billion trying to buy Yahoo. They missed mobile entirely. They missed cloud, then built Azure into a $100 billion business by being a fast and well-capitalised second mover.

The pattern is consistent: Microsoft does not innovate at the frontier. Microsoft watches what works, waits until the market signal is undeniable, and then deploys its distribution and capital advantages to capture the category.

Which makes what Microsoft chooses to use internally the most reliable leading indicator in enterprise technology.

When Microsoft bought GitHub in 2018, it told you that developer infrastructure was the next platform battle. When Microsoft invested $13 billion into OpenAI, it told you that LLM capability was the next productivity layer. When Microsoft built Teams to strangle Slack, it told you that collaboration tooling was too strategically important to cede.

And when Microsoft chose Anthropic’s Claude over its own Copilot for security-critical code analysis — it told you something that no amount of benchmark marketing can convey.

It told you which AI it actually trusts.


The Copilot Paradox

Let us be precise about what Microsoft did and why it is significant.

Microsoft’s Security Development Lifecycle is not a peripheral product. It is the process by which Microsoft evaluates its own code for vulnerabilities before releasing software to hundreds of millions of users and thousands of enterprise customers. Getting this wrong is not a reputational problem. It is an existential one. A vulnerability that ships in Windows or Azure because the SDL missed it is the kind of failure that ends careers, triggers congressional hearings, and costs billions.

Microsoft does not make SDL decisions carelessly.

Microsoft also has GitHub Copilot — its own AI coding assistant, built on OpenAI models, marketed aggressively to every developer and security team on the planet as the AI-powered future of secure software development. It has entire product lines, sales motions, and marketing campaigns built around the premise that Copilot is the right AI for code security.

And yet when Microsoft needed AI inside its own SDL — the highest-stakes code security context that exists inside Microsoft — it did not reach for Copilot.

It reached for Claude.

The official reason, buried in the announcement, is about cybersecurity capability and responsible deployment. Mythos, Anthropic’s frontier model with advanced security research capabilities, brings something to vulnerability detection that Microsoft’s own models apparently do not. But there is a second reason that the press release does not state explicitly and does not need to.

Anthropic’s safety positioning is credible in a way that OpenAI’s is not right now. The constitutional AI approach, the responsible scaling policy, the deliberate restraint on Mythos deployment — these are not just marketing. They represent a verifiable track record of prioritising caution over capability velocity. For a company that needs to defend its SDL decisions to regulators, auditors, and enterprise customers, that track record has real procurement value.

Microsoft chose the AI that it could defend choosing. That is a different calculus than choosing the AI that performs best on a benchmark.


What the Car Thief Knows

The car thief analogy deserves its full elaboration because it is more precise than it first appears.

A car thief is not choosing a vehicle based on the advertisement. They are not swayed by the marketing campaign or the brand positioning or the aspirational lifestyle imagery. They are making a pure functional evaluation under conditions where being wrong has immediate and serious consequences.

They choose based on what actually works.

When you see a pattern in what car thieves prefer — and law enforcement tracks this — it tells you something about engineering quality, security architecture, and real-world reliability that no consumer review captures with the same fidelity. The thief has done the work.

Microsoft is the car thief of enterprise software. Not malicious — but ruthlessly functional in its evaluations, operating under conditions where being wrong is expensive, and entirely unsentimental about brand loyalty when the stakes are high enough.

Microsoft chose Claude for its SDL.

That is the equivalent of the car thief choosing the vehicle for their own getaway. There is no higher endorsement available in that particular evaluation framework.


The Market Signal Nobody Is Reading Correctly

The enterprise AI market is currently drowning in benchmark theatre. Every model release comes with carefully selected evaluation scores, cherry-picked comparisons, and performance claims that are technically accurate and practically meaningless for the use cases that actually drive procurement.

CTOs making real decisions — decisions with budget authority and accountability attached — are not reading benchmark reports. They are watching what the sophisticated buyers do.

And the sophisticated buyers are telling a clear story right now.

Microsoft chose Anthropic for security-critical infrastructure. This happened the same week that GPT-5.5 launched with enormous fanfare and OpenAI’s most aggressive capability claims to date. The timing is not coincidental — it is clarifying. Capability and trustworthiness are not the same variable. The market is beginning to price them separately.

For regulated industries — finance, defence, healthcare, critical infrastructure — trustworthiness is not a soft consideration. It is a hard procurement requirement. You cannot deploy an AI system in a regulated environment on the basis of benchmark scores. You deploy it on the basis of a defensible safety and governance track record that will survive an audit, a regulator’s question, or a board inquiry after something goes wrong.

Anthropic has built that track record deliberately and at the cost of moving slower than its competitors. Microsoft’s SDL integration is the first major signal that the market is beginning to reward that investment.

It will not be the last.


What This Means For Your Infrastructure Decisions

If you are a CTO evaluating AI for anything security-adjacent — code review, vulnerability detection, threat analysis, compliance automation — Microsoft just made your evaluation framework simpler.

The most sophisticated enterprise software organisation on the planet, with unlimited budget to evaluate every available option, chose Anthropic for the use case where being wrong is most expensive. That is a data point worth more than any third-party benchmark.

If you are evaluating AI for sovereign or regulated deployments — where auditability, safety governance, and responsible scaling matter as much as raw capability — the same signal applies with even greater force. Anthropic’s approach to deployment restraint is not a limitation. It is the feature that makes deployment defensible in environments where every decision requires justification.

And if you are building private AI infrastructure — running models on your own hardware, in your own jurisdiction, under your own control — the model layer is only one decision. The infrastructure layer underneath it determines whether your sovereignty claim is real or theatrical.

Watch what Microsoft chose for its most critical deployment.

Then ask whether your own AI infrastructure decisions reflect the same level of seriousness.


Catalin Lichi is the founder of Sugau — a bare-metal Kubernetes consultancy specialising in sovereign infrastructure and private AI for regulated industries. Building the infrastructure layer that makes the model layer defensible.