Anthropic Limits Access to Advanced Mythos Preview AI Model Due to Security Risks
Anthropic announced on Tuesday that it is withholding public release of its Mythos Preview AI model over concerns about its ability to detect and exploit vulnerabilities. The model is being provided to select tech and cybersecurity firms for defensive testing. This approach aims to develop safeguards before broader deployment.
AxiosAnthropic has restricted access to its new Mythos Preview AI model, citing its advanced capabilities in identifying and exploiting security flaws. The company is rolling out the preview only to a select group of more than 40 tech and cybersecurity organizations. These partners will use the model to scan and secure their code and open-source systems.
Mythos Preview demonstrates sophisticated reasoning skills comparable to an advanced security researcher. It can identify tens of thousands of vulnerabilities in software, including those in major operating systems and web browsers. 1% of test cases on the first attempt.
In specific tests, Mythos Preview uncovered flaws in the Linux kernel, enabling potential full control of affected machines. It also identified a 27-year-old vulnerability in OpenBSD, an operating system used in firewalls, routers, and high-security servers. These findings include bugs believed to be decades old and undetected by prior human security audits.
Controlled Rollout and Partnerships Anthropic launched Project Glasswing, involving 12 companies including Amazon Web Services, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorgan Chase, the Linux Foundation, Microsoft, Nvidia, and Palo Alto Networks.
These participants will apply Mythos Preview to defensive security efforts, with Anthropic sharing resulting insights. The company is providing up to $100 million in usage credits to testing firms and $4 million to open-source security groups such as OpenSSF, Alpha-Omega, and the Apache Software Foundation.
“— Logan Graham, head of Anthropic's frontier red team, Tuesday (Axios)”
S. agencies including the Cybersecurity and Infrastructure Security Agency and the Commerce Department on the model's risks and benefits. A company official noted opportunities for enhancing defensive security measures. The official did not confirm briefings with the Pentagon.
6 identified about 500 zero-day vulnerabilities in open-source software. In contrast, Mythos Preview's output is significantly higher. Anthropic decided against general availability after internal testing and limited feedback indicated substantial risks.
Broader Context and Future Plans AI models have previously aided malicious activities, including a Chinese spying campaign targeting 30 organizations using Anthropic's technology and cybercriminals automating ransomware. Logan Graham stated that similar capabilities could emerge from other AI firms within six to 18 months. OpenAI and other companies are developing comparable models. >
"More powerful models are going to come from us and from others, and so we do need a plan to respond to this."
“— Dario Amodei, Anthropic CEO, Tuesday (Axios)”
Anthropic aims to eventually enable safe deployment of Mythos-class models for general use. The company plans safeguards for its Opus models to refine controls without the risks posed by Mythos Preview.
Story Timeline
4 events- Tuesday
Anthropic announced limited rollout of Mythos Preview to select organizations and launched Project Glasswing.
3 sourcesAxios · @financialjuice · @financialjuice - Prior months
Anthropic conducted internal testing of Mythos Preview and gathered feedback from a small group.
1 sourceAxios - Recent years
AI models including Anthropic's aided malicious activities like Chinese spying and ransomware automation.
1 sourceAxios - Previous release
Anthropic released Opus 4.6 model, which identified about 500 zero-day vulnerabilities.
1 sourceAxios
Potential Impact
- 01
Select companies gain advanced tools to strengthen software security defenses.
- 02
Other AI firms accelerate development of similar vulnerability-detection models.
- 03
Open-source projects receive funding to address vulnerabilities identified by Mythos.
- 04
U.S. agencies incorporate AI risk insights into cybersecurity policies.
- 05
Defensive security practices in tech industry evolve with shared Mythos insights.
Transparency Panel
Related Stories
SemaforAnthropic Co-Founder Warns of Upcoming AI Capabilities for Exploiting Web Vulnerabilities
Anthropic's co-founder stated that powerful AI models capable of exploiting website vulnerabilities will emerge soon. The company's new model, Claude Mythos, identified unknown security flaws in major web browsers and operating systems. Financial authorities have responded by dis…
Los Angeles TimesGallup Poll Shows Increasing AI Use Among US Workers with Persistent Skepticism
A Gallup poll conducted in February indicates that more American workers are using artificial intelligence in their jobs, with about 3 in 10 using it frequently. However, skepticism remains common, with many non-users citing preferences for traditional methods, ethical concerns,…
Federal Bureau of Investigation / Wikimedia (Public domain)AI Assistant Poke Charges Billionaire $136,000 Monthly Fee
Poke, an AI assistant without a price ceiling, charged one billionaire $136,000 a month. Marvin von Hagen stated this pricing detail. The information highlights Poke's premium service model.