Anthropic has officially acknowledged a troubling report regarding unauthorized access to Claude Mythos, a highly sensitive AI model deemed too risky for public release. In a carefully crafted statement to Bloomberg, an Anthropic spokesperson confirmed, “We’re investigating a report claiming unauthorized access to Claude Mythos Preview through one of our third-party vendor environments.” This revelation raises significant concerns about the security measures in place to protect advanced AI technologies from potential misuse.
According to Bloomberg, the reported breach has been substantiated through live demonstrations and screenshots shared by a member of the unauthorized group. These findings highlight the vulnerabilities that exist even within tightly controlled environments and emphasize the importance of robust cybersecurity protocols.
In a rather convoluted explanation, Bloomberg revealed that an anonymous source claims to be affiliated with an unspecified group that has exploited their access as a worker at a third-party contractor for Anthropic. They reportedly utilized “commonly used internet sleuthing tools often employed by cybersecurity researchers” to gain some level of access to the highly sensitive AI model, raising questions about the effectiveness of current security measures.
Despite the alarming nature of this breach, the source reassured Bloomberg that this clandestine group is primarily interested in exploring new AI models rather than using them for malicious purposes. The source claimed they are merely “playing around” with the technology, which adds a layer of complexity to the ethical considerations surrounding unauthorized access to potentially dangerous AI systems.
The timeline surrounding this apparent breach unfolds as follows:
- A Discord group has been established that employs bots to scour GitHub for information regarding unreleased AI models.
- A data breach occurred at the AI training startup Mercor.
- The group amalgamated information from the Mercor breach with access available to the Bloomberg source due to their employment by an Anthropic contractor.
- This combination enabled the group to deduce the online location of Claude Mythos.
- Since April 7, coinciding with the announcement of Project Glasswing, the group has been experimenting with Claude Mythos without authorization.
In summary, Anthropic has positioned itself as the custodian of what it claims to be the most formidable AI model globally. Many influential institutions seem to share this belief. If we accept Anthropic‘s assertions, we are placing significant trust in their ability to manage this power responsibly. However, the reality is that an unknown entity has breached access to this formidable AI model. If we are to trust their claims, they insist that their interactions have been limited to non-harmful experimentation and exploration.









