works
Beren Millidge Thoughts on Claude Mythos online The rapid advancement of cyberattack capabilities in the Claude Mythos model likely stems from targeted reinforcement learning from verifiable rewards (RLVR) rather than fundamental breakthroughs in pretraining scaling or model architecture. While Mythos represents a significant increase in pretraining scale, the observed performance spikes in cybersecurity and software engineering benchmarks suggest the application of a specialized post-training pipeline utilizing cyber-specific environments and agentic coding harnesses. Because cybersecurity tasks offer diverse training data and easily verifiable success metrics, they serve as ideal RLVR domains. This shift implies that advanced offensive and defensive cyber capabilities are primarily constrained by environment design and high-quality fine-tuning data rather than raw compute or model size. Consequently, similar capabilities may soon proliferate through the open-source community as these post-training methodologies are replicated. The dramatic improvement in long-context retrieval tasks further supports the hypothesis that targeted post-training on synthetic algorithmic tasks is being used to optimize models for complex agentic operations. This evolution indicates a new era where specialized RLVR, rather than general scaling, serves as the primary driver for eliciting high-level, domain-specific AI expertise. – AI-generated abstract.

Thoughts on Claude Mythos

Beren Millidge

April 11, 2026

Abstract

The rapid advancement of cyberattack capabilities in the Claude Mythos model likely stems from targeted reinforcement learning from verifiable rewards (RLVR) rather than fundamental breakthroughs in pretraining scaling or model architecture. While Mythos represents a significant increase in pretraining scale, the observed performance spikes in cybersecurity and software engineering benchmarks suggest the application of a specialized post-training pipeline utilizing cyber-specific environments and agentic coding harnesses. Because cybersecurity tasks offer diverse training data and easily verifiable success metrics, they serve as ideal RLVR domains. This shift implies that advanced offensive and defensive cyber capabilities are primarily constrained by environment design and high-quality fine-tuning data rather than raw compute or model size. Consequently, similar capabilities may soon proliferate through the open-source community as these post-training methodologies are replicated. The dramatic improvement in long-context retrieval tasks further supports the hypothesis that targeted post-training on synthetic algorithmic tasks is being used to optimize models for complex agentic operations. This evolution indicates a new era where specialized RLVR, rather than general scaling, serves as the primary driver for eliciting high-level, domain-specific AI expertise. – AI-generated abstract.

PDF

First page of PDF