Editorial front page
FinalAI-edited source brief

Researchers Claim Anthropic's Restricted Mythos AI Was Used to Breach macOS

The allegation arrives weeks after Anthropic disclosed that unauthorized users had accessed the cybersecurity model through a third-party vendor.

Published 7 sources1 Reddit5 web80% confidence
Thumbnail from Engadget

What matters

  • Security researchers claim to have breached macOS using Anthropic's restricted Claude Mythos AI model.
  • Apple, a partner in Anthropic's Project Glasswing cybersecurity initiative, is reportedly taking the claims seriously.
  • Anthropic disclosed in April that unauthorized users accessed Mythos via a third-party vendor, reportedly by guessing its URL using data exposed in a breach of vendor Mercor.
  • Mythos was designed to find high-severity vulnerabilities in major operating systems and was withheld from public release due to its offensive potential.
  • Technical details of the alleged macOS exploit have not been publicly disclosed.

What happened

On May 15, security researchers claimed they had compromised macOS using Claude Mythos, a restricted Anthropic AI model designed to find and exploit software vulnerabilities, according to Engadget. Apple, which is both the maker of macOS and a partner in Anthropic’s Project Glasswing cybersecurity initiative, is reportedly taking the claims seriously.

The allegation comes on the heels of an embarrassing access control failure. In late April, Anthropic confirmed it was investigating unauthorized access to Mythos through a third-party vendor environment. According to The Verge, a small group of unauthorized users gained access to the model—reportedly by guessing its online location using information about Anthropic’s other models exposed in a breach of Mercor, a vendor that works with the company. Anthropic said at the time that it had not detected breaches of its own systems beyond the vendor environment.

Mythos was unveiled in early April as part of Project Glasswing, a coalition that includes Apple, Amazon, Google, Microsoft, NVIDIA, and others. Anthropic said the model had already found thousands of high-severity vulnerabilities in every major operating system and web browser, and it was shared with select partners precisely because it was considered too dangerous for public release.

Why it matters

If the macOS claim holds up, it would mark one of the first confirmed cases of a restricted, offense-capable AI model being used to breach a major consumer operating system. The incident highlights the dual-use dilemma: the same tools built to harden software can be turned against it if they leak.

The episode also spotlights supply-chain risk. Anthropic’s own partners rely on third-party vendors to host and develop models, yet that vector appears to have exposed Mythos within days of its announcement. For Apple, the situation is particularly awkward. The company joined Project Glasswing to get ahead of AI-powered threats, only to find its own platform allegedly compromised by the very model it was meant to help evaluate.

Public reaction

Online discussion has been split between awe at Mythos’s reported capabilities and anxiety about its misuse. A widely viewed Reddit thread in r/OpenAI noted that Mythos appeared to land above trendline predictions for advanced AI, though commenters quickly pushed back on the statistical significance of that claim. Others raised broader concerns about the wisdom of building models powerful enough to exploit “every major operating system” when access controls have already failed once.

What to watch

Several critical details remain unconfirmed. It is unclear whether the researchers who claim the macOS breach obtained Mythos through the April unauthorized access incident, or through legitimate preview channels. Anthropic has not publicly detailed what macOS vulnerability was allegedly exploited, and Apple has not independently confirmed a successful intrusion. The next moves by both Anthropic and Apple—particularly any changes to Mythos access controls or partner agreements—will be telling.

Sources

Public reaction

Reddit discussions show a mix of excitement about Mythos's reported capabilities and skepticism toward unrestrained AI hype. One widely discussed thread noted the model's apparent performance against trendline predictions, while commenters pushed back on statistical extrapolation and raised concerns about dual-use risks.

Signals

  • Excitement about advanced AI security capabilities
  • Skepticism about exponential-progress hype and benchmark claims
  • Concern over dual-use AI risks
  • Anxiety about vendor access controls for restricted models

Open questions

  • Did the researchers use Mythos accessed during the April unauthorized incident, or did they have legitimate preview access?
  • What specific macOS vulnerability was allegedly exploited, and has Apple independently confirmed it?
  • How will Anthropic adjust Mythos access controls after consecutive security incidents?

What to do next

Developers

Audit integrations with AI security scanners and isolate AI-generated exploit code in sandboxed environments.

If offensive-capable AI models circulate outside intended access controls, generated code should be treated as potentially untrusted.

Founders

Evaluate vendor security postures and access logs when using restricted AI models for penetration testing.

A breach of the security tool itself undermines the value proposition of AI-assisted defense and creates liability exposure.

PMs

Review third-party AI tool access controls and update incident response plans for AI-assisted breach scenarios.

Partnership programs for dual-use models require stricter governance than standard SaaS integrations.

Investors

Monitor dual-use AI risk exposure and the stability of vendor access controls in cybersecurity portfolios.

Consecutive breaches involving a restricted model may signal systemic governance risks at the vendor level.

Operators

Ensure AI-assisted security tools operate under strict output controls and human-in-the-loop review.

Autonomous vulnerability discovery without oversight can escalate quickly into unintended system compromise.

Testing notes

Caveats

  • Mythos is a restricted preview model available only to select Project Glasswing partners. Anthropic has not released a public API, download, or trial for the model.