Persona

Persona

Recent breakthroughs in AI‑assisted vulnerability discovery are genuinely impressive from a research perspective. Mythos demonstrate capabilities that would have seemed unrealistic just a few years ago. From a purely technical standpoint, this progress is real and worth acknowledging.

The problem is that, in real enterprise environments, this progress is often aimed at the wrong constraint.

Most organizations are not struggling because they fail to discover vulnerabilities. They are struggling because they are already overwhelmed by the vulnerabilities they know about. Enterprise vulnerability backlogs are massive, and only a small fraction of those issues is ever actively exploited. This is further compounded by the fact that almost every large enterprise runs some amount of end‑of‑life software for legitimate technical or business compatibility reasons.

In that reality, vulnerability discovery is rarely the bottleneck. Patching discipline, risk‑based prioritization, and operational capacity are.

The economic reality of vulnerability discovery

There is also a structural economic problem that often gets ignored in the enthusiasm around AI automated discovery. Searching for new vulnerabilities across the long tail of software has weak economic incentives. Paying the equivalent of tens of thousands of dollars in compute per discovered vulnerability does not scale.

While many issues could be found by human researchers, there is simply no market willing to pay for that work across most software products that enterprises depend on today. This economic mismatch alone makes large‑scale, continuous AI vulnerability discovery a poor lever for improving real‑world security outcomes.

The small‑model shortcut

In parallel, many vendors are now pitching “small models” as a practical and cost‑effective alternative. In practice, this positioning often relies on sales shortcuts rather than on demonstrated operational value. When evaluated under real‑world conditions, these solutions tend to look far less compelling.

Real environments are noisy. Data is incomplete. Systems are interconnected. Constraints are organizational as much as technical. When exposed to this complexity, many of these approaches fail to deliver on their promised benefits.

From a strictly AI perspective, these solutions are also frequently heavily prescoped. They rely on curated code snippets and extensive contextual information embedded directly in prompts. As a result, they do not match the actual capabilities demonstrated by Mythos. At the same time, they are only “small” when compared to Mythos‑class systems. Running or renting them still carries meaningful costs, which are often obscured by VC‑funded subsidies.

Those subsidies should be treated as temporary. No market can rely on them indefinitely.

How to improve security in practice

If vulnerability discovery is not the primary constraint, the question becomes straightforward: what actually improves security?

Start with a real threat model of your environment

Build a threat model based on how your environment is actually designed and operated, not on vendor‑driven fear narratives. Identify what assets matter, who could realistically attack them, and what meaningful failure would look like.
If you cannot clearly articulate what you are defending and why, adding another tool will not materially improve your security posture.

Treat classic security fundamentals as non‑negotiable

Patching, monitoring, logging, and preventive controls are still where most real risk reduction happens. They may be boring, but they remain foundational.
Even if advanced threat scenarios apply to your organization, you still have to start here. No AI‑driven discovery compensates for poor security hygiene.

Optimize patching before expanding detection

Focus on improving patching discipline and risk‑based prioritization. Prioritize what is actually exploitable and relevant in your environment instead of accumulating theoretical findings.
A smaller number of well‑prioritized fixes consistently closed is worth more than an ever‑growing backlog.

Plan explicitly for systems that cannot be patched

If patching is not possible due to compatibility or business constraints, treat those systems as explicit risk exceptions and apply alternative mitigations:

Unpatchable does not have to mean undefended.

Strengthen detection and response where prevention ends

Ensure monitoring, alerting, and response playbooks are realistic and exercised. Detection without response does not reduce risk.
Focus on signals that indicate real attacker progression, not raw alert volume.

Treat resilience as part of security

Assume that some controls will fail. Regularly test backup, restore, and recovery processes under realistic conditions.
Security is not only about prevention, but also about how fast and reliably you can recover.

Add advanced tooling only after the basics work

Introduce advanced capabilities only when they reinforce proven security processes and close known gaps.
Tooling should support a coherent control strategy, not substitute for the absence of one.


AI‑assisted security research will continue to advance, and systems like Mythos clearly point to where the field is heading. But enterprise security is not primarily a research problem. It is an operational one, defined by constraints, trade‑offs, and execution. Until organizations align investments with those realities, even the most impressive AI demos will remain just that demos. Real security improvement still comes from understanding your environment, managing risk deliberately, and doing the unglamorous work consistently well.