
Your plants and your people are caught between a game-changing tool and a historic level of risk.
In case you somehow missed it, Anthropic's Project Glasswing recently set off alarms throughout the cybersecurity and AI communities. Glasswing is a security initiative where Anthropic was working with Apple, Microsoft, Google, Amazon, and about 40 other key AI players in a sort of focus group setting to assess an AI model named Mythos Preview that runs on their Claude platform.
The goal of Mythos is to proactively detect and patch software vulnerabilities - ideally preventing them from being exploited by malicious actors and protect critical infrastructure by applying AI-powered offensive security techniques. Due to Mythos' powerful potential, Anthropic meant to limit initial access.
However, unauthorized users were able to gain access to the platform via third-party vendor credentials. These parties are part of a Discord online forum group known to search for information about unreleased AI models. After obtaining access, the group proceeded to publicize the ease at which Mythos is able to identify vulnerabilities.
In the wrong hands, this tool offers hackers the ability to attack at a speed which would be nearly impossible to stop. Or, Mythos could be vital in helping defenders finally operate from a proactive posture, instead of constantly playing catch-up. As you can imagine, there was a passionate response.
Shane Fry, Chief Technology Officer, RunSafe Security: "Unauthorized users were able to access Anthropic's Mythos model, reportedly by just changing a model name. Even if their intent is just to explore, it shows how easily these systems can be exposed.
"The reality is these AI capabilities are already out there, 'hacked' or not, and they're going to accelerate how quickly vulnerabilities are found and exploited. Software teams will need to look at how to harden their code so those vulnerabilities can't be used in the first place."
Agnidipta Sarkar, Chief Evangelist at ColorTokens: "While Anthropic is investigating, the only information publicly available so far is that the attack used the oldest trick in the book, impersonating someone with existing access. The users reportedly guessed the model's URL based on knowledge of Anthropic's patterns for other models. The good news is that Anthropic detected the breach and contained it to that specific vendor's environment.
"One of the key controls that every modern environment needs is microsegmentation, which can effectively reduce the blast radius to specific vendors and leave no elbow room for attackers to navigate. I am hoping Anthropic is using similar controls to keep the attack contained, such as zero-trust mechanisms."
Tim Mackey, Head of Software Supply Chain Risk Strategy at Black Duck: "The unfortunate reality is that while it's great to hear that novel cybersecurity models are being provided to select researchers to evaluate, if your team is on the outside looking in, waiting for the final report might not be top of mind. For defenders, even the specter of unauthorized access to an adversarial model as powerful as Mythos is purported to be, only increases anxiety levels.
"What's clear is that security leaders in organizations of all sizes should take this claim as a call to action focused on the role AI-enabled cybersecurity plays in their operations and how best to scale those efforts to deal with AI enabled adversaries."
John Gallagher, Vice President, Viakoo: "There has always been an arms race between cyber defenders and cyber attackers, and Mythos is currently the most powerful armament available. If we do not know whose hands it is in, it should be viewed no differently than uncontrolled distribution of enriched uranium.
"If true, this deeply undermines Project Glasswing which was setup up explicitly to give cyber defenders early access to Mythos Preview in order to define and mount defenses against it. Threat actors having early access to Mythos Preview puts them on the same footing (or possibly with advantages) versus cyber defenders.
"Uncontrolled access to Mythos Preview will hit hardest on operators of critical OT, IoT, and ICS systems. Already knowing the fifty IT organizations with early access to Mythos would naturally focus threat actors on targets outside of those 50 companies, most likely non-standard operating systems that are prevalent in OT and IoT.
"If the model has escaped Pandora's Box, there should be immediate validation and public notification of it. Since that has not happened here, it is likely that there was not significant exposure. However, there has never been a prize as valuable to cyber criminals before as early access to Mythos Preview; it potentially can open all bank accounts and reveal all secrets.
"Threat actors are highly sophisticated, very well-funded, and determined. We are in a race to harden systems and have rapid patching at high scale in place before threat actors can leverage Mythos Preview."
Nicole Carignan, SVP, Security & AI Strategy, and Field CISO at Darktrace: "This highlights the continued weaponization of commercial tooling. Frontier and near‑frontier models are increasingly dual‑use by default. Capabilities designed to improve software quality and security can be repurposed with minimal friction to accelerate vulnerability discovery for malicious ends. This is not a failure of intent; it is an outcome of scale, accessibility, and capability diffusion.
"These models will continue to be a target for threat actors to gain access to in order to achieve initial access capabilities to organizations. More concerning is access to critical vulnerabilities that have not yet been released to the public. Possession of undisclosed, high‑severity vulnerabilities enables threat actors to facilitate more sophisticated and scaled access to organizations through exploiting an 'unknown' vulnerability.
"It is also important to be realistic about containment. This was never going to be contained to a single model, organization, or access control failure. Threat actors do not need this system; they need a system with sufficient capability. Whether through parallel development, model leakage, fine‑tuning, or the combination of multiple weaker models and tools, similar outcomes can be achieved.
"The strategic mistake would be to treat this as an isolated incident rather than a signal. Advanced vulnerability discovery capabilities will continue to proliferate, and the window between discovery and exploitation will continue to shrink. This reinforces the need for scaled visibility, behavioral analytics, anomaly detection, and autonomous containment across endpoints, cloud, identities, SaaS, and critical infrastructure.
"Finally, this is another reminder that investment in AI adoption without commensurate investment in security and risk management is unsustainable. Resilience will depend less on how quickly vulnerabilities can be patched, and more on how effectively exploitation can be detected and contained when prevention inevitably fails."
Reach Security's Co-founder and CEO, Garrett Hamilton: "There is only one viable response to AI-powered attacks: AI-powered defense.
"If a model can discover and exploit unknown weaknesses at machine speed, the defense playbook must change just as fast. Security teams can't rely on periodic scans and manual hardening; they need always-on visibility of their real exposures and clear prioritization of what to fix first.
"However, vulnerabilities should not be the only concern. These are researched routinely by vendors and the cybersecurity community, with patches regularly released. In short, organizations have a fighting chance when it comes to spotting and fixing software vulnerabilities.
"Misconfigurations, on the other hand, have no patches and can offer direct access into an environment. They arise unnoticed over time as networks, software, users, and policies change. They're also far more pervasive than many teams realize: our research found 97 percent of organizations suffered a breach or near miss in the past year due to a security-tool misconfiguration, and it takes 8.3 days on average to remediate once identified. That is more than enough time for an AI-enabled attacker to take advantage.
"The new standard is simple: fight AI with AI, and close the gaps before they become incidents."