UK AI institute tests Anthropic's Mythos model on cyber attacks

The UK government’s AI Security Institute has released an evaluation of Anthropic's Mythos Preview AI model, confirming its strong performance in multistep cyber infiltration challenges. Mythos became the first model to fully complete a demanding 32-step network attack simulation known as 'The Last Ones.' The institute cautions that real-world defenses may limit such automated threats.

Anthropic last week limited the initial release of its Mythos Preview model to a select group of critical industry partners, citing its advanced computer security capabilities. The UK’s AI Security Institute (AISI) conducted independent tests using Capture the Flag challenges designed to assess AI cyberattack potential. These evaluations, ongoing since early 2023, show Mythos completing over 85 percent of apprentice-level tasks, similar to recent models like GPT-5.4, Opus 4.6, and Codex 5.3. AISI said the model matches competitors on individual tasks but stands out in chaining them for complex operations. Anthropic’s model succeeded in fully solving 'The Last Ones' (TLO), a 32-step data extraction attack simulating 20 hours of human effort across multiple hosts. It completed the challenge from start to finish in 3 out of 10 attempts and averaged 22 steps, far exceeding Claude 4.6's 16-step average. AISI noted this suggests Mythos can autonomously target small, weakly defended enterprise systems where initial network access is gained. Mythos struggled with the 'Cooling Tower' test, a seven-step power plant control disruption scenario. The institute highlighted that tests used a 100 million token budget and lack real-world active defenders or detection mechanisms. AISI warned that well-defended systems may resist such attacks, urging AI use in strengthening protections as models advance.

Связанные статьи

Illustration of Anthropic restricting Claude Mythos AI and launching Project Glasswing consortium with tech giants to address cybersecurity vulnerabilities.
Изображение, созданное ИИ

Anthropic restricts Claude Mythos AI release and launches Project Glasswing over cybersecurity risks

Сообщено ИИ Изображение, созданное ИИ

Anthropic has limited access to its Claude Mythos Preview AI model due to its superior ability to detect and exploit software vulnerabilities, while launching Project Glasswing—a consortium with over 45 tech firms including Apple, Google, and Microsoft—to collaboratively patch flaws and bolster defenses. The announcement follows recent data leaks at the firm.

Anthropic has released a new cyber-focused AI model called Mythos, capable of detecting software flaws faster than humans and generating exploits. The model has raised alarms among governments and companies for potentially turbocharging hacking by exposing vulnerabilities quicker than they can be patched. Officials worldwide are scrambling to assess the risks.

Сообщено ИИ

In the wake of Anthropic's unveiling of its powerful Claude Mythos AI—capable of detecting and exploiting software vulnerabilities—the US Treasury Secretary has convened top bank executives to highlight escalating AI-driven cyber threats. The move underscores growing concerns as the AI is restricted to a tech coalition via Project Glasswing.

The Pentagon has formally notified AI company Anthropic that it is deemed a “supply chain risk,” a rare designation that critics say is typically aimed at adversary-linked technology. The move follows a breakdown in negotiations over whether the U.S. military can use Anthropic’s Claude models for all lawful purposes, versus contractual limits the company says are needed to prevent fully autonomous weapons and mass domestic surveillance.

Сообщено ИИ Проверено фактами

After Anthropic CEO Dario Amodei said in late February that the company would not allow its Claude model to be used for mass domestic surveillance or fully autonomous weapons, senior Pentagon officials said they have no intention of using AI for domestic surveillance and insist that private firms cannot set binding limits on how the U.S. military employs AI tools.

US President Donald Trump has directed all federal agencies to immediately cease using Anthropic's AI tools amid a dispute over military applications. The move follows weeks of clashes between Anthropic and Pentagon officials regarding restrictions on AI for mass surveillance and autonomous weapons. A six-month phase-out period has been announced.

Сообщено ИИ

A crypto security firm used artificial intelligence to detect a high-severity bug in Nethermind, an Ethereum client used by nearly 40% of validators. The flaw, which could have disrupted network operations, was fixed before exploitation. This development highlights AI's growing role in cybersecurity amid recent concerns over AI-generated code vulnerabilities.

 

 

 

Этот сайт использует куки

Мы используем куки для анализа, чтобы улучшить наш сайт. Прочитайте нашу политику конфиденциальности для дополнительной информации.
Отклонить