Anthropic's new A.I. model is triggering Global alarms
Company Updates

Anthropic's new A.I. model is triggering Global alarms

End Time Headlines2d ago

Anthropic, the prominent artificial intelligence startup, is grappling with a significant security breach involving its most advanced technology.

A small collective of unauthorized individuals reportedly gained access to Claude Mythos, a new model so potent that the company had initially restricted its release due to concerns over its capabilities.

The breach, which was first reported by Bloomberg News on Wednesday, involved a group of users on a private Discord forum. According to reports, these individuals did not use sophisticated hacking techniques to bypass Anthropic's internal systems.

Instead, they managed to access a "preview" version of the model by correctly identifying its online location within a third-party vendor's environment, utilizing knowledge of the company's previous naming conventions and hosting patterns.

The incident is particularly sensitive because Anthropic has spent weeks describing Mythos as a revolutionary but dangerous tool. The model has demonstrated an unprecedented aptitude for identifying and exploiting "zero-day" software vulnerabilities -- flaws unknown to the software's creators -- that have existed for decades in major operating systems and web browsers.

In an effort to manage these risks, Anthropic had launched "Project Glasswing," a highly controlled initiative that granted access only to a select group of approximately 40 partners, including federal agencies and major financial institutions. The goal was to use the AI defensively to patch critical infrastructure before it could be targeted by malicious actors.

"We are investigating a report claiming unauthorized access to Claude Mythos Preview through one of our third-party vendor environments," an Anthropic spokesperson said in a statement.

While the group that accessed the model reportedly claimed they were motivated by curiosity rather than malice, the breach has intensified the debate over whether such powerful "frontier" models can ever be truly secured.

Some experts suggest that the very act of restricting the model may have inadvertently created a new kind of target for digital hobbyists and hackers alike.

Dario Amodei, the chief executive of Anthropic, has previously warned that the rapid advancement of AI coding capabilities represents a fundamental shift in the cybersecurity landscape. He noted that these systems are now identifying weaknesses "that humans have missed" for years.

The leak has also drawn the attention of the federal government. Recent memos indicate that the Office of Management and Budget has been working to provide federal agencies with access to the model, provided that "the appropriate guardrails and safeguards are in place" to prevent misuse.

As the company works to close the loophole that allowed the unauthorized access, the incident serves as a stark reminder of the challenges inherent in the AI arms race.

For many in the industry, the breach confirms a difficult reality: as AI becomes more capable of breaking into systems, the systems holding the AI themselves become increasingly vulnerable.

Originally published by End Time Headlines

Read original source →
AnthropicDiscord