Business13 min read

Anthropic Double Breach: Enterprise AI Security 2026

In five days, Anthropic suffered two security incidents that exposed an unreleased AI model's capabilities and 512,000 lines of production source code. Both were caused by human error. Here is what happened, what was exposed, and what every enterprise using AI tools needs to do about it.

Digital Applied Team
March 30, 2026
13 min read
512K

Leaked Source Code Lines

1,906

Exposed TypeScript Files

~3,000

CMS Assets Made Public

5

Days Between Breaches

Key Takeaways

Two breaches in five days from a leading AI safety company: Anthropic, which positions itself as the safety-focused AI lab, suffered a CMS misconfiguration leak on March 26 and an npm packaging error on March 31. Both were caused by human error, not external attacks, raising fundamental questions about operational security at AI companies.
Mythos leak revealed an unreleased model with unprecedented cybersecurity capabilities: Nearly 3,000 unpublished assets became publicly accessible, revealing the Capybara model tier that Anthropic describes as a step change in capability. Internal documents warn the model makes large-scale cyberattacks much more likely in 2026.
Claude Code source leak exposed 512,000 lines of orchestration logic: A 59.8 MB source map file in npm version 2.1.88 exposed the complete agentic harness: permission models, bash security validators, 44 unreleased feature flags, and references to unannounced models. The codebase was mirrored across GitHub within hours.
Every enterprise running AI coding agents lost a layer of defense: The leaked source reveals exact orchestration logic for Hooks and MCP servers. Attackers can now design malicious repositories specifically tailored to trick AI coding agents into running background commands or exfiltrating data.
1 in 8 companies now report AI breaches linked to agentic systems: Malware hidden in public model and code repositories is the most cited source of AI-related breaches at 35%. Gartner predicts AI supply chain attacks will become one of the top five attack vectors by 2026.

Anthropic has built its brand on AI safety. The company's Constitutional AI approach, its Responsible Scaling Policy, and its public commitment to building trustworthy AI systems have made it the preferred AI vendor for risk-conscious enterprises. Then, in the final week of March 2026, two separate security incidents in five days challenged that narrative in ways that have implications far beyond Anthropic itself.

The first incident, on March 26, exposed nearly 3,000 unpublished assets from Anthropic's content management system, including draft documentation revealing an unreleased model called Mythos (internally codenamed Capybara) that the company describes as posing "unprecedented cybersecurity risks." The second incident, on March 31, leaked the complete source code of Claude Code, Anthropic's AI coding agent, through a misconfigured npm package containing a 59.8 MB source map file.

Breach One: The Mythos Document Leak (March 26)

On March 26, 2026, Fortune reported that a human error in the configuration of Anthropic's content management system had left unpublished draft content publicly accessible. Nearly 3,000 assets linked to Anthropic's blog were discoverable in a public data cache, despite never having been officially published. Among these assets was documentation for a new model tier called "Capybara," publicly referred to as Mythos.

The leaked materials revealed that Capybara represents what Anthropic internally describes as "a step change" and "the most capable model we've ever built." The documentation states that Capybara is "larger and more intelligent than our Opus models, which were, until now, our most powerful." Benchmark data in the leaked materials shows dramatically higher scores on tests of software coding, academic reasoning, and cybersecurity compared to Claude Opus 4.6.

What the Mythos Leak Revealed
  • New model tier: Capybara (Mythos) described as larger and more capable than Opus, representing a step change in capability
  • Cybersecurity warning: Internal documents state the model poses "unprecedented cybersecurity risks" and could make large-scale cyberattacks much more likely
  • Government briefings: Anthropic was privately warning top government officials about the model's cybersecurity implications
  • Release status: The model is expensive to run and not yet ready for general release, being tested with a small group of early access customers

The strategic sensitivity of this leak cannot be overstated. A company that was privately briefing government officials about the cybersecurity risks of an unreleased model had that very information exposed through a basic CMS configuration error. The irony was not lost on the industry: the AI safety company that warns about unprecedented cybersecurity risks could not prevent its own blog drafts from becoming publicly accessible.

After being informed of the leak by Fortune, Anthropic removed public access to the data store. The company acknowledged the incident was caused by human error and confirmed that Mythos is indeed a real model under development.

Breach Two: Claude Code Source Leak (March 31)

Five days after the Mythos document leak, Anthropic suffered a second and arguably more consequential security incident. A 59.8 MB JavaScript source map file (.map), intended for internal debugging, was inadvertently included in version 2.1.88 of the @anthropic-ai/claude-code package on the public npm registry. By 4:23 AM ET, researcher Chaofan Shou had discovered the exposure and broadcast it on X (formerly Twitter).

The source map file could be used to reconstruct Claude Code's complete source code: 512,000 lines of TypeScript across 1,906 files. Within hours, the codebase was mirrored across GitHub and analyzed by thousands of developers. This was not the model weights (those were not exposed), but in many ways the leaked code is more strategically valuable: it is the complete agentic harness that gives Claude the ability to use tools, manage files, run bash commands, and orchestrate multi-agent workflows.

What the Claude Code Leak Exposed
  • Complete permission model: Every permission chain, trust prompt, and access control mechanism used by the AI coding agent
  • Bash security validators: The full set of security checks that determine which shell commands the agent can execute
  • 44 unreleased feature flags: References to capabilities and models that Anthropic has not publicly announced
  • MCP server orchestration: Complete logic for Hooks and Model Context Protocol integrations, revealing how the agent interacts with external tools
  • Compaction pipelines: How Claude Code manages context windows, conversation history, and memory across long sessions

Anthropic's official response characterized the incident as "a release packaging issue caused by human error, not a security breach," stating that "no sensitive customer data or credentials were involved or exposed." While technically accurate, this framing understates the strategic impact. The exposed source code provides a detailed blueprint for anyone seeking to find vulnerabilities in AI coding agent architectures.

What Was Actually Exposed and Why It Matters

Understanding the impact of both incidents requires separating what was exposed from what was not. The Mythos leak did not expose model weights, training data, or customer information. The Claude Code leak did not expose the underlying Claude model. However, the combination of both leaks creates a security picture that is greater than the sum of its parts.

Exposed (High Impact)
  • Complete agent orchestration architecture and permission models
  • Bash security validators and sandbox bypass conditions
  • Unreleased model capabilities and cybersecurity risk assessments
  • MCP integration logic and tool-calling interfaces
Not Exposed (Important Distinction)
  • Model weights for Claude or Mythos/Capybara
  • Customer data, credentials, or API keys
  • Training data or RLHF datasets
  • Server-side API infrastructure or deployment configurations

The practical concern for enterprises is that the Claude Code source leak dramatically reduces the cost of attack research against AI coding agents. Previously, finding vulnerabilities in agent permission models required reverse engineering and extensive testing. Now the complete permission model, every bash security validator, and the full MCP interface specification are publicly readable. The same orchestration patterns exist across every enterprise AI agent deployment, and the barrier to crafting targeted exploits has collapsed.

Enterprise Security Implications

The Anthropic incidents arrive at a moment when enterprise AI adoption is accelerating rapidly and security frameworks have not kept pace. According to recent industry research, 1 in 8 companies now report AI breaches linked to agentic systems. Malware hidden in public model and code repositories has emerged as the most cited source of AI-related breaches at 35%. Gartner predicts AI supply chain attacks will become one of the top five attack vectors by the end of 2026.

The Anthropic breaches are instructive because they illustrate failure modes that are specific to AI companies and distinct from traditional software vendor security incidents. AI companies manage unreleased model capabilities with potential national security implications, distribute agentic tools via public package registries, and operate at the intersection of safety research and commercial deployment. These characteristics create unique security requirements that traditional SOC 2 assessments and vendor questionnaires do not adequately address.

Attack Surface Expansion from Claude Code Leak
  • Prompt injection targeting: Attackers can now craft malicious repositories that exploit specific permission chain weaknesses visible in the source code
  • MCP server exploitation: The complete Hook and MCP orchestration logic enables targeted attacks that could exfiltrate data through tool-calling interfaces
  • Trust prompt bypass: Knowledge of exact security validator logic enables adversaries to design inputs that pass validation checks while executing malicious operations
  • Supply chain poisoning: The brief window of malicious axios versions demonstrates how source leaks create opportunistic supply chain attacks

More than 76% of organizations now cite shadow AI as a definite or probable problem, up from 61% in 2025. When developers install AI coding agents through public package registries without centralized IT oversight, every one of these agents becomes a potential entry point. The Anthropic incidents demonstrate that even the AI vendor itself can inadvertently introduce risk through the same distribution channels.

AI Vendor Risk Assessment Framework

The Anthropic incidents provide a concrete case study for building AI vendor risk assessments that go beyond traditional software vendor evaluations. The following framework addresses the AI-specific risk categories that these breaches exposed.

Supply Chain Controls
Package registry and distribution security
  • Does the vendor strip source maps and debug artifacts from production packages?
  • Are npm/PyPI releases automated with CI/CD guardrails or manual and error-prone?
  • Does the vendor sign packages and provide SBOM (Software Bill of Materials)?
  • What is the incident response time for compromised package versions?
Internal Documentation Controls
CMS and unpublished content security
  • How does the vendor prevent draft/internal documents from becoming publicly accessible?
  • Are CMS configurations reviewed for access control before sensitive content is stored?
  • Does the vendor separate internal safety research from marketing content systems?
  • What controls exist for pre-release model documentation?
Agentic System Security
AI agent permission and sandbox architecture
  • What permission models govern AI agent access to file systems, networks, and shell execution?
  • Does the vendor implement sandboxing for agentic operations?
  • How are MCP server connections authenticated and authorized?
  • What monitoring exists for anomalous agent behavior?
Incident Response and Disclosure
Notification timelines and transparency
  • What is the contractual SLA for breach notification to enterprise customers?
  • Does the vendor proactively notify customers or wait for media disclosure?
  • Are post-incident reports shared with affected enterprise customers?
  • How does the vendor distinguish between "security breach" and "packaging error" in communications?

Incident Response Playbook for AI Vendor Breaches

When an AI vendor you depend on suffers a security incident, the response needs to be faster and more targeted than traditional vendor breach playbooks. AI tools often have deeper access to your codebase, infrastructure, and business logic than typical SaaS applications. The following five-step playbook is derived from the Anthropic incidents and applicable to any AI vendor security event.

1Immediate Assessment (First 2 Hours)

Determine which versions of the AI tool are deployed across your organization. Check dependency lock files for compromised package versions. In the Claude Code case, this meant checking for axios 1.14.1 or 0.30.4. Inventory all systems where the AI agent has file system or network access.

2Scope the Exposure (Hours 2-8)

Review what the leaked information reveals about your specific deployment. When source code for security validators is exposed, audit your AI agent configurations against the now public validation logic. Identify any custom MCP servers or tool integrations that could be targeted using the leaked orchestration patterns.

3Implement Compensating Controls (Hours 8-24)

Add monitoring for anomalous AI agent behavior: unexpected file access patterns, outbound network connections to unknown hosts, or shell commands that do not match expected development workflows. Consider restricting AI agent permissions to read-only while assessment is in progress.

4Vendor Communication (Day 1-3)

Engage the vendor's enterprise security team directly. In the Anthropic case, the public response framed the incident as a "packaging issue" rather than a security breach. Push for a detailed post-incident report, a timeline for remediation, and confirmation of what additional measures will prevent recurrence.

5Strategic Review (Week 1-2)

Evaluate whether the incident changes your vendor risk profile. Two breaches in five days from the same vendor is a pattern that requires elevated scrutiny. Consider diversifying AI tool providers, implementing additional network segmentation for AI agent access, and revising your AI vendor assessment criteria to include supply chain controls.

Strategic Lessons for Business Leaders

The Anthropic double breach offers several strategic lessons that extend well beyond one company's security incidents. These lessons apply to any enterprise integrating AI tools into their operations, regardless of vendor.

Safety Branding Is Not Security

Anthropic's position as the safety-focused AI lab did not prevent basic operational security failures. Enterprise buyers should evaluate vendors based on demonstrated security practices, not marketing positioning. Request audit certifications, penetration testing results, and supply chain security documentation.

Package Registries Are Attack Surfaces

AI tools distributed through npm and PyPI face the same supply chain risks as any open-source dependency, plus the additional risk that AI tools have deeper system access than typical packages. Implement dependency scanning, version pinning, and registry mirroring for all AI tool installations.

Human Error Scales Exponentially in AI

Both Anthropic incidents were caused by human error. In traditional software, human error might expose a database backup. In AI, human error can expose model capabilities with national security implications and agentic tool architectures used across thousands of enterprises. The blast radius of operational mistakes at AI companies is larger than at traditional vendors.

Defense in Depth for AI Tools

Do not rely solely on the AI vendor's security measures. Implement your own monitoring, access controls, and network segmentation around AI agent deployments. Treat AI coding agents with the same security scrutiny you would apply to a contractor with root access to your development infrastructure.

The bottom line is that AI vendor security assessment needs to evolve beyond traditional frameworks. The Anthropic incidents demonstrate that AI companies face unique operational security challenges: they manage sensitive model capabilities, distribute tools through public registries, and their products have deep access to customer environments. Enterprises that recognize these differences and adapt their CRM and automation infrastructure accordingly will be better positioned to manage AI vendor risk as the industry matures.

Conclusion

Two breaches in five days from the AI industry's most prominent safety-focused company reveal an uncomfortable truth: operational security at AI companies has not kept pace with the capabilities they are building. The Mythos document leak exposed an unreleased model that Anthropic itself describes as posing unprecedented cybersecurity risks. The Claude Code source leak exposed the exact architecture that thousands of enterprises rely on to keep AI agents operating within safe boundaries.

The response for enterprise leaders is not to abandon AI tools but to treat AI vendor relationships with the security rigor they demand. This means going beyond SOC 2 certifications to assess supply chain controls, package registry practices, and incident response capabilities. It means implementing defense-in-depth strategies that do not rely solely on vendor-side security. And it means recognizing that the AI industry's rapid pace of development creates operational security gaps that even well-intentioned companies will struggle to close.

Strengthen Your AI Security Posture

Our team helps enterprises assess AI vendor risks, implement defense-in-depth strategies, and build incident response frameworks tailored to the unique challenges of AI tool adoption.

Free consultation
Vendor risk assessments
Enterprise AI security

Frequently Asked Questions

Related Business Guides

Continue exploring AI security and enterprise technology strategies