LIVE THREATS
ATLAS OWASP CRITICAL RELEVANCE ▲ 8.5

US summons bank bosses over cyber risks from Anthropic's latest AI model

The US Treasury convened major bank executives to discuss cybersecurity risks posed by Anthropic's unreleased Claude Mythos model, which the company claims has surpassed nearly all human experts at finding and exploiting software vulnerabilities. A code leak prompted Anthropic to publicly acknowledge the model's unprecedented offensive cyber capability, raising systemic financial sector risk concerns. The meeting signals growing regulatory awareness of AI-enabled cyber threats to critical financial infrastructure.

US summons bank bosses over cyber risks from Anthropic's latest AI model

Overview

The US Treasury Secretary Scott Bessent convened an emergency meeting with the chief executives of America’s most systemically important banks in Washington following the emergence of Anthropic’s Claude Mythos AI model and a preceding leak of its code. Federal Reserve Chair Jerome Powell reportedly attended alongside banking leaders including Goldman Sachs’ David Solomon, Citigroup’s Jane Fraser, and Bank of America’s Brian Moynihan. The meeting reflects acute concern that an AI model capable of autonomously discovering and exploiting software vulnerabilities at scale could pose existential risks to financial sector cybersecurity.

Anthropics’ own public disclosure, prompted by the code leak, stated that Claude Mythos has “surpassed all but the most skilled humans at finding and exploiting software vulnerabilities” — framing the model as a potential force multiplier for adversarial cyber operations against banks and critical infrastructure.

Technical Analysis

Claude Mythos represents a qualitative leap in AI-assisted offensive security capability. The model’s reported ability to autonomously identify thousands of vulnerabilities across software and popular applications suggests it functions as an advanced agentic system capable of end-to-end exploit research — from vulnerability identification to exploitation pathway generation — without meaningful human oversight at each step.

The code leak that preceded Anthropic’s disclosure is itself a significant supply chain security event. Exposure of model weights, architecture details, or fine-tuning data could enable adversaries to:

  • Reproduce or fine-tune the model for unconstrained offensive use
  • Identify and bypass safety guardrails embedded in the released version
  • Use leaked training data to infer the vulnerability classes the model specialises in

This positions Claude Mythos as both a direct threat vector (if accessed or replicated by malicious actors) and an indirect one (via the chilling effect of demonstrated AI offensive capability on defender confidence).

Framework Mapping

MITRE ATLAS:

  • AML.T0047 (ML-Enabled Product or Service): The model is explicitly positioned as an offensive cyber tool capable of real-world exploitation.
  • AML.T0010 (ML Supply Chain Compromise): The code leak introduces supply chain risk, potentially allowing adversarial access to model internals.
  • AML.T0044 (Full ML Model Access): Leaked code may provide adversaries with full model access enabling unconstrained use.

OWASP LLM Top 10:

  • LLM08 (Excessive Agency): An agentic model autonomously conducting exploit research represents a textbook excessive agency risk.
  • LLM05 (Supply Chain Vulnerabilities): The code leak is a direct supply chain event.
  • LLM06 (Sensitive Information Disclosure): Leaked model internals may expose training data or safety mechanism details.

Impact Assessment

The financial sector faces compounded risk: as a primary target for cybercrime and nation-state operations, the availability of an AI system that can rapidly enumerate exploitable vulnerabilities across banking software stacks dramatically lowers the barrier for sophisticated attacks. Systemically important banks — whose disruption could destabilise global finance — are the highest-risk targets. Broader critical infrastructure sectors face equivalent exposure.

Mitigation & Recommendations

  • Accelerate patch cadence: Given AI-assisted vulnerability discovery, assume exploit timelines are now compressed dramatically.
  • AI red-teaming: Commission adversarial evaluations of internal systems using equivalent AI tooling before threat actors do.
  • Supply chain audits: Review all AI model dependencies and third-party integrations for exposure to leaked model capabilities.
  • Regulatory engagement: Engage with Treasury and OCC frameworks emerging from this consultation to align security posture with forthcoming guidance.
  • Threat intelligence sharing: Participate in FS-ISAC channels specifically for AI-enabled threat indicators.

References