
TeamPCP Supply Chain Attacks
TeamPCP is exploiting trusted npm and PyPI packages to compromise developer environments, steal credentials, and extend attacks across software supply chains.

There has been a significant amount of interest by CISOs in the impact of frontier artificial intelligence (AI) models for offensive and defensive purposes following Anthropic’s Claude Mythos Preview release April 7, 2026. The response from the security industry has been divided. Have we reached the end of software security as we know it? Or is this just more AI hype?
In discussions we’ve had with security practitioners and leaders, concerns have focused on the speed and scale at which these models can be weaponized and what this means for today’s security programs. This blog shares our analysts’ assessment of Anthropic’s claims about Mythos’ vulnerability discovery and exploit generation, real-world examples of AI-based vulnerability research and adversary usage, and an underground perspective.
We believe Anthropic’s published results about Mythos Preview mark a meaningful step forward in AI-assisted vulnerability research, particularly in autonomous exploit development. The increase in capability — evidenced by Anthropic’s Firefox test in February which we detail below — is indicative of a vulnerability landscape that is rapidly changing. Vulnerability discovery has been visibly reshaped by large language models (LLMs), and the effects are observable in the increase in the number of reported CVEs and bug-bounty reports. This trend is reflected by the National Institute of Standards and Technology’s announcement it will no longer enrich every CVE submitted to the National Vulnerability Database, noting a 263% increase in submissions between 2020 and 2025. As AI models improve, they likely will gain parity with Mythos Preview, compounding this growth. Enterprises must be prepared to operate in an increasingly noisy threat environment where patch prioritization is even more important.
Exploit development has been more difficult to achieve via AI, requiring significant human expertise to produce impactful proofs of concept (PoCs). This notion appears to have changed with the emergence of Mythos Preview. If the advantage demonstrated in Anthropic’s testing environment translates to practical real-life usage, we likely will see a significant collapse in the time between vulnerability discovery and weaponization.
However, until Anthropic publishes details regarding false positive rates and overall success ratios, it’s not known how effective AI models would be as an organization’s primary tool for code assurance. Even if sophisticated AI models result in more exploit code, this does not automatically translate to better outcomes for adversaries who need reliable exploits, which can easily be refined into exploitation frameworks. Meanwhile, defense-in-depth measures still provide organizations with significant protection even when adversaries are equipped with a larger arsenal of weaponized vulnerabilities.
Several factors warrant caution about the broader claims related to Mythos. The public data on Mythos Preview represents a curated sample chosen by Anthropic rather than detailed results — likely designed to stimulate media interest and increase investor interest prior to an upcoming initial public offering (IPO).
As such, AI systems are capable of detecting real-world vulnerabilities — in some cases they also generate legitimate-sounding reports against codebases — but without baseline performance data, the operational cost of AI-assisted vulnerability discovery and exploitation at scale cannot be properly estimated. Additionally, since Mythos Preview is not publicly available, the research community cannot independently audit the claims or reproduce the results — meaning we are essentially evaluating Anthropic's conclusions about its own product.
Finally, while access to Mythos Preview is currently constrained to a select few organizations, there remains a very real chance that threat actors will eventually circumvent those constraints and/or target organizations with authorized access.
With the likely increase of AI adoption in the vulnerability domain, we predict:
In its technical write-up, Anthropic said Mythos Preview demonstrated the ability, when directed by users, to identify and exploit zero-day vulnerabilities across every major operating system (OS) and web browser tested, including complex exploit chains and local privilege escalation paths. Anthropic also described the model as highly proficient in cybersecurity, autonomous coding and long-running agent tasks.
During roughly one month of internal red team testing, Mythos Preview identified thousands of high-severity zero-days in production software. Anthropic chose to highlight a 27-year-old bug in OpenBSD, a 16-year-old flaw in FFmpeg and a 17-year-old remote code execution (RCE) vulnerability in FreeBSD's network file server (NFS) implementation (CVE-2026-4747). Anthropic claimed the model identified and exploited CVE-2026-4747 end-to-end without human intervention after the initial prompt and in some cases achieved unauthenticated root from an internet-facing position.
It conducted similar tests on its earlier model Claude Opus 4.6, which demonstrated AI could be used for vulnerability discovery at scale — months before the Mythos Preview announcement. Anthropic reported Opus 4.6 found and validated more than 500 vulnerabilities across a variety of open source codebases, several of which survived years of expert review and continuous fuzzing. In February 2026, during a two-week collaboration with Mozilla, Anthropic used Opus 4.6 to identify 112 issues in Firefox, of which 22 were assigned CVEs and 14 were high severity. Furthermore, Claude Opus 4.6 produced working shell exploits on two occasions across several hundred attempts. In comparison, using similar test conditions, Mythos Preview reportedly succeeded 181 times, with an additional 29 runs reaching register control — meaning Mythos can control central processing unit (CPU) registers but does not have full RCE capability. Anthropic attributes these results to general improvements in code, reasoning and autonomy rather than any cyber-specific training.
Independent observations point to the same broader shift. The AI security firm Aisle reported its system discovered all 12 zero-day vulnerabilities disclosed in OpenSSL's January 2026 patch and accounted for 13 of 14 OpenSSL CVEs assigned in 2025, indicating AI systems from outside Anthropic are finding zero-days in heavily audited code. Wordfence, which runs a WordPress bug-bounty program, reported that AI-assisted vulnerability submissions rose from 16% to about 66% between late November 2025 and April 2026, while overall submission volume increased 453% in the same period.
One of the key considerations when deciding if AI is a viable solution for vulnerability discovery is the cost. The marquee finding from Anthropic’s Mythos Preview press release was the discovery of the 17-year-old OpenBSD bug. The campaign behind its discovery allegedly cost US $20,000. Anthropic also provided an example of how Mythos Preview was used to generate an exploit that cost under US $1,000 at API pricing. While these costs are viable for many enterprises, only the most resourced adversaries would likely be able to finance them.
LLMs are playing a direct role in the vulnerability research lifecycle itself, particularly in converting disclosed patches into working exploit code. The three CVEs below illustrate how quickly the time-to-exploit collapses when AI-assisted tooling is applied.
We have tracked the use of AI by threat actors since the technology became mainstream in 2023. Initial concerns failed to live up to expectations with the majority of the use being limited to phishing lure refinement and know-your-customer (KYC) bypass — with marginal gains thereafter. However, this has changed with the release of recent models, and AI-assisted vulnerability discovery and exploitation has moved from demonstration to operational reality. The most documented case is a breach of Mexican government organizations between late December 2025 and mid-February 2026 where a single operator compromised nine entities using Claude Code and GPT 4.1 as primary operational tools. We reported the intrusion included more than 1,000 prompts to which Claude Code generated 20 tailored exploits against 20 distinct CVEs, 400 custom attack scripts and about 5,317 executed commands across 34 live sessions. Roughly 75% of all command execution activity on the victim infrastructure was AI generated.
Additionally, Google's February 2026 AI Threat Tracker reported the Chinese state-sponsored hacking group APT31 prompted Gemini under an expert cybersecurity persona to automate vulnerability analysis against U.S. targets, pairing it with HexStrike — an open source tool that lets an LLM agent orchestrate over 150 offensive tools — to work through RCE, web application firewall (WAF) bypass and structured query language-injection (SQLi) paths before Google disabled the associated accounts.
Following the release of Mythos Preview, we observed several discussions in the underground focused on the model. Some actors expressed skepticism about Claude Mythos, arguing that it appears to be more AI hype than a genuine breakthrough and that limited PoC results are being overstated as evidence of a major technological leap. We also noted a message from a user seeking access to Claude Mythos through a Telegram channel, indicating early interest in obtaining the model outside official channels. We did not observe any relevant discussion regarding the GPT‑5.4‑Cyber model at the time of this report.
Interest in Claude Mythos is expected to increase among threat actors, especially those aiming to accelerate vulnerability discovery, exploit development and operational efficiency. While the invitation-only distribution model is likely to limit immediate, large-scale misuse, the model’s perceived offensive value and restricted availability may drive attempts to gain access through compromised accounts, stolen API credentials, or intermediary or insider assistance.
Exploit development is becoming accessible to a wider range of actors. More actors, including less sophisticated ones working with AI output they can't fully evaluate, will be producing exploits. This will create a far noisier environment for enterprise security teams. Patch prioritization and response timelines built for a slower exploitation cycle are already behind. The window organizations have to act on a disclosure is shorter than it was a year ago, and it will be shorter still a year from now. Enterprises should not make planning decisions around last year's threat environment.
AI models perform best against widely used open source software, major browsers and common frameworks — the codebases they've seen the most training data on. Specialist and proprietary environments have more insulation, at least for now. And human expertise still matters: AI remains a productivity multiplier, not a replacement for researchers who can reason about novel contexts.

TeamPCP is exploiting trusted npm and PyPI packages to compromise developer environments, steal credentials, and extend attacks across software supply chains.

An Iranian aligned threat group conducting destructive and espionage focused cyber operations against organizations in Israel and Western countries.

CrazyHunter is a ransomware campaign targeting healthcare that weakens endpoint defenses and escalates privileges before encrypting systems at scale.
Stay informed with our weekly executive update, sending you the latest news and timely data on the threats, risks, and regulations affecting your organization.