Identifying vulnerabilities and exploits with Claude Mythos Preview
Over the past few weeks, we’ve used Claude Mythos Preview to identify thousands of zero-day vulnerabilities (that is, flaws previously unknown to software developers) across every major operating system, every major web browser, and other critical software. Many of them are serious.
Frontier Red Team blog posts provide technical details about a subset of these vulnerabilities that have already been patched, and in some cases how to exploit vulnerabilities discovered by Mythos Preview. Identifying nearly all of these vulnerabilities and developing many associated exploits was done completely autonomously, without any human interaction. Below are three examples.
- Mythos Preview discovered a 27-year-old vulnerability in OpenBSD, which has a reputation as one of the world’s most secure operating systems and is used to run firewalls and other critical infrastructure. This vulnerability allows an attacker to remotely crash a machine by simply connecting to a machine running the operating system.
- A 16-year-old vulnerability in FFmpeg, which is used by countless software to encode and decode video, was also discovered in a line of code where automated testing tools failed to detect the issue after 5 million hits.
- This model autonomously discovered and chained together several vulnerabilities in the Linux kernel (the software that runs most of the world’s servers), allowing attackers to escalate from normal user access to complete control of the machine.
The above vulnerabilities have been reported to the associated software administrators and all have now been patched. For many other vulnerabilities, we are providing cryptographic hashes of the details today (see Red Team blog) and will reveal more details after the fixes are applied.
Evaluation benchmarks such as CyberGym highlight significant differences between Mythos Preview and the next best model, Claude Opus 4.6.
Reproducing cybersecurity vulnerabilities
In addition to our own efforts, many of our partners have already been using Claude Mythos Preview for several weeks. Here’s what they found:
Claude Mythos Preview’s powerful cyber capabilities are a result of its strong agent coding and reasoning skills. For example, as shown in the evaluation results below, this model has the highest scores of any model developed to date on a variety of software coding tasks.
Please refer to the Claude Mythos Preview system card for detailed information on model functionality, safety, and general characteristics.
Although we have no plans to make Claude Mythos Preview publicly available, our ultimate goal is to enable users to safely deploy Mythos class models at scale. Not only for cybersecurity purposes, but also for the countless other benefits that such a highly functional model brings. This requires advancing the development of cybersecurity (and other) safeguards that detect and block the most dangerous outputs of models. We plan to introduce new protections in future Claude Opus models, allowing us to improve and refine protections in models that do not pose the same level of risk as the Mythos Preview.3.
Project Glasswing Plan
Today’s announcement is the beginning of a long-term effort. Success requires broad engagement from within and outside the technology industry.
Project Glasswing partners will have access to Claude Mythos Preview to help them find and remediate vulnerabilities and weaknesses in their underlying systems, which represent a very large portion of the world’s shared cyber attack surface. This work is expected to focus on tasks such as local vulnerability detection, binary black box testing, endpoint protection, and system penetration testing.
Anthropic has committed $100 million in model usage credits to Project Glasswing and additional participants, covering significant usage throughout this research preview. The Claude Mythos preview will then be offered to participants for $25 or $125 per million input/output tokens (participants will have access to the Claude API, Amazon Bedrock, Google Cloud’s Vertex AI, and models on Microsoft Foundry).
In addition to the model usage credit commitment, to help open source software maintainers adapt to this changing landscape, we have donated $2.5 million to Alpha-Omega and OpenSSF through the Linux Foundation and $1.5 million to the Apache Software Foundation (maintainers interested in access can apply through the Claude for Open Source program).
We plan to expand the scope of this work and continue it for many months, sharing as much as we can so other organizations can apply the lessons to their own security. Partners will share information and best practices with each other to the extent possible. Within 90 days, Anthropic will publicly report on what we have learned, any vulnerabilities that have been fixed, and any improvements we can make publicly available. We will also work with leading security organizations to develop a set of practical recommendations on how security practices should evolve in the age of AI. This may include:
- Vulnerability disclosure process.
- Software update process.
- Open source and supply chain security.
- Software development life cycle and secure by design practices.
- Regulated industry standards.
- Prioritize scaling and automation. and
- Automate patching.
Anthropic also has ongoing discussions with U.S. government officials regarding the Claude Mythos Preview and its offensive and defensive cyber capabilities. As stated above, protecting critical infrastructure is a top national security priority for democracies. The emergence of these cyber capabilities is another reason the United States and its allies must maintain a decisive lead in AI technology. Governments have a critical role to play in maintaining that lead and in assessing and mitigating national security risks associated with AI models. We stand ready to work with local, state, and federal representatives to support these missions.
We hope that Project Glasswing will be the seed of a larger effort across industry and the public sector, helping all stakeholders tackle the biggest questions about the security implications of strong models. We invite other AI industry members to help set industry standards. In the medium term, independent third parties that can bring together private and public sector organizations may be the ideal home for continued efforts on these large-scale cybersecurity projects.
