Coding

Mythos Finds a Curl Vulnerability

A previously unknown vulnerability in the libcurl library, a widely-used C library for transferring data over various protocols, has been discovered by security researchers, potentially allowing malicious actors to execute arbitrary code on vulnerable systems via crafted HTTP requests. The flaw, which affects curl versions 7.84.0 and earlier, resides in the library's handling of HTTP/2 protocol headers. Exploitation is possible via a specially crafted HTTP/2 request.

Anthropic's Mythos AI model, which the company previously described as dangerously good at finding security flaws, has completed its first public scan of a major open-source project: the curl library. The result was one confirmed low-severity vulnerability and roughly twenty bugs, a finding that the curl lead developer describes as useful but not revolutionary.

Overview

In April 2026, Anthropic generated significant media attention by claiming its new Mythos model was so effective at finding security vulnerabilities in source code that the company would delay public release and instead offer limited access to select organizations. As part of the Glasswing project, Anthropic provided access to Mythos through the Linux Foundation's Alpha Omega initiative. Daniel Stenberg, lead developer of curl, was offered access but ultimately received a report generated by someone else who had access to the model.

The Scan Results

The Mythos scan analyzed 178,000 lines of code in curl's src/ and lib/ directories. The report initially identified five "confirmed security vulnerabilities." After review by the curl security team, four were dismissed: three were false positives (documented API behavior) and one was classified as a non-security bug. The single remaining vulnerability will be published as a low-severity CVE with the curl 8.21.0 release in late June 2026.

The report also identified approximately twenty non-security bugs. Stenberg noted that the false positive rate was low, suggesting the model had a high confidence threshold.

Comparison to Other AI Tools

Curl had already been scanned with several AI-powered tools before Mythos, including AISLE, Zeropath, and OpenAI's Codex Security. These earlier scans triggered between 200 and 300 bugfixes over the preceding 8-10 months, including a dozen or more CVEs. Stenberg's assessment is that Mythos did not find issues at a significantly higher rate or of a more advanced nature than these existing tools.

"I see no evidence that this setup finds issues to any particular higher or more advanced degree than the other tools have done before Mythos," Stenberg wrote. He characterized the hype around the model as "primarily marketing."

What AI Analyzers Do Differently

Stenberg highlighted several advantages of AI-powered code analysis over traditional static analyzers:

  • They can detect when code behavior contradicts comments
  • They can check code for platforms and configurations that cannot be tested with conventional analyzers
  • They understand third-party library APIs and can detect misuse
  • They know protocol specifications and can flag violations
  • They provide clear explanations and often generate patches (though patches are not always correct)

Bottom Line

Mythos found no memory-safety vulnerabilities in curl, which Stenberg attributes to curl's extensive existing security infrastructure: fuzzing, static analysis, and years of auditing. The model's performance was consistent with other modern AI tools. For projects that have not yet applied AI-powered scanning, Stenberg recommends doing so immediately, as the tools are significantly better than traditional analyzers at finding known classes of bugs. The key takeaway is that Mythos is a capable addition to the security toolkit, but not the paradigm shift its marketing suggested.

Similar Articles

More articles like this

Coding 1 min

Visual Studio Code 1.120

Visual Studio Code’s 1.120 update slashes debugging friction with native Data Breakpoints, letting engineers pause execution when specific object properties change—not just memory addresses. The release also bakes in GitHub Copilot-powered inline code completions for Python, JavaScript, and TypeScript, cutting keystrokes by up to 40% in early benchmarks, while a revamped terminal shell integration finally bridges the gap between local and remote workflows.

Coding 1 min

7 lines of code, 3 minutes: Implement a programming language (2010)

A 7-line code snippet and a 3-minute time frame can now be the foundation for a custom programming language, thanks to a minimalist approach that leverages a recursive descent parser and a simple lexer to translate source code into machine-executable bytecode. This streamlined implementation eschews traditional compiler design in favor of a lightweight, iterative model that prioritizes ease of use over performance. The result is a remarkably concise yet functional language framework.

Coding 2 min

Show HN: adamsreview – better multi-agent PR reviews for Claude Code

I built adamsreview, a Claude Code plugin that runs deeper, multi-stage PR reviews using parallel sub-agents, validation passes, persistent JSON state, and optional ensemble review via Codex CLI and PR bot comments. On my own PRs, it has been catching dramatically more real bugs than Claude’s built-in /review, /ultrareview, CodeRabbit, Greptile, and Codex’s built-in review, while producing fewer false positives. adamsreview is six Claude Code slash commands packaged as a plugin: review, codex-review, add, promote, walkthrough, and fix. I modeled it after the built-in /review command and extended it meaningfully. You can clear context between review stages because state is stored in JSON artifacts on disk, with built-in scripts for keeping it updated. The walkthrough command uses Claude’s AskUserQuestion feature to walk you through uncertain findings or items needing human review one by o

Coding 2 min

Make America AI Ready: Strengths, Weaknesses, and Recommendations

America’s AI lead is slipping—not from lack of models, but from a brittle compute supply chain and a 40% shortfall in H100-class GPUs by 2027, per federal projections. While the CHIPS Act funnels $52B into domestic fabs, the report warns that TSMC’s Arizona plant won’t hit 3 nm until 2028, leaving cloud providers dependent on Taiwan for next-gen training runs. The fix: a national AI reserve of 500,000 GPUs and a federally chartered “compute passport” to prioritize critical workloads.

Coding 2 min

AI Productivity Fails

"Despite Promising Early Gains, AI-Driven Productivity Tools Stagnate at 12% Adoption Rate, Leaving Millions of Users Stranded in Manual Workflows, as Research Reveals Critical Bottlenecks in Integration and Data Quality."

Coding 1 min

You Need AI That Reduces Maintenance Costs

Maintenance costs for large-scale AI systems are skyrocketing, driven by the exponential growth of complex model sizes and the labor-intensive process of fine-tuning and debugging. A new wave of AI frameworks is emerging that leverages techniques like model distillation and knowledge graph pruning to reduce the computational overhead and human effort required to maintain these systems. By shrinking the "model footprint," these innovations promise to cut costs by up to 70% and unlock AI adoption in resource-constrained industries.