Type something to search...
Four Generations of Broken Promises: Why AI SOC Agents Might Actually Be Different

Four Generations of Broken Promises: Why AI SOC Agents Might Actually Be Different

Series: The SIEM & AI Reckoning — Article 1 of 10


Over twenty years and hundreds of vendor pitches, one line never changes:

“This is going to change everything.”

2005, SIEM. 2012, Next-Gen SIEM. 2016, SOAR. And now, 2025–2026, it’s AI SOC Agents. Four generations. Three broke their promises. The fourth is making the same ones.

But here’s what the marketing decks never said. Each of the first three generations made the analyst’s job harder before it made it easier. For many teams, it never got easier at all.

The workforce shortage hit 4.8 million globally by 2025. It’s still growing. Three generations of tools, and we’re further behind than when we started.

AI SOC Agents operate on a different principle than everything that came before them. Not a better way to organize data, but a way to reason over it. The question isn’t whether the fourth generation is technically better. It’s whether it finally breaks the pattern where the cure is as hard to manage as the disease.


Three Generations of Security Tooling, One Pattern

We don’t need a history lesson. We need the pattern.

SIEM (2005): The Fire Hose Problem

Before SIEM, security teams had blind spots but manageable workloads. After SIEM, we had visibility into everything and no framework for deciding what any of it meant.

Alert volume exploded. The fire hose was on, and nobody had a nozzle.

Data without context is noise, and SIEM gave us an ocean of it. Analysts spent more time triaging alerts than investigating threats. We solved the visibility problem and created the overwhelm problem.

Next-Gen SIEM (2012): Context for the Cases We Already Knew About

The industry recognized the problem and responded with behavioral analytics, ML-based anomaly detection, and smarter correlation. Give analysts context, not just data.

The catch: context only for behaviors someone had already anticipated. Novel attack paths, edge cases, and attacker creativity remained invisible. And the alerts that were generated required more analyst judgment to validate, not less. A signature hit is binary. A behavioral anomaly requires interpretation.

We added complexity on top of volume. Better tools for swimming, in a deeper ocean.

SOAR (2015): Automating the Wrong Layer

By this point, analyst burnout was undeniable. SOAR’s answer: automate the response. Close the loop. Let the playbook handle the obvious stuff so analysts can focus on what matters.

This is where it really went wrong, and it’s worth being specific about why.

SOAR required a rare, expensive hybrid. Someone who deeply understood the security use case and could translate it into automation logic. Not a developer; they didn’t know the threat context. Not a tier-1 analyst; they couldn’t build the playbooks. You needed a security engineer with automation skills, and those people were already the most constrained resource in the SOC.

SOAR created a bottleneck at exactly the point where you could least afford one. The people who could have built the playbooks were too busy fighting fires to build them. When playbooks were built, they broke on edge cases, which required the same rare people to fix them.

Gartner retired the SOAR Magic Quadrant. That’s about as polite as analysts get when something hasn’t worked.

GenerationThe PromiseWhat Actually HappenedNet Effect on Analysts
SIEM (2005)Centralized visibilityAlert flood, no context frameworkMore volume, no nozzle
Next-Gen SIEM (2012)Smarter detectionContext only for anticipated scenariosComplexity on top of volume
SOAR (2015)Automated responseSME bottleneck; playbooks broke on edge casesDependency on the rarest people you had

Three generations. Each one added burden before, or instead of, removing it. Do AI SOC Agents finally break this pattern?

![[four-generations-diagram.png]]


The Fourth Generation: Why AI SOC Agents Are Structurally Different

Here’s what I’m seeing right now as a CISO evaluating these tools.

Gartner placed “AI SOC agents” at the peak of inflated expectations on their 2025 Hype Cycle. Only 9% of security professionals report being “very confident” in AI-generated alerts (Gurucul, 2025). The SANS 2025 SOC Survey ranked AI/ML tools dead last in satisfaction across all technology categories, despite being adopted by roughly 40% of SOCs.

We have every reason to be skeptical.

Two things are structurally different about AI SOC Agents:

AI may break the SME bottleneck. Natural language prompting lowers the floor for who can build and maintain automation logic. A tier-1 analyst can describe a workflow in plain English and get something useful back. That was never possible with SOAR playbooks. The dependency on a rare security-engineer-plus-automation-skills hybrid may no longer be the rate limiter.

AI reasons over unstructured data without pre-written rules. Instead of “this IP triggered this signature,” you get “this sequence of behaviors across three systems, in the context of this user’s role and this week’s threat landscape, looks like lateral movement because…” This directly addresses the context gap that SIEM and Next-Gen SIEM couldn’t close.

The data backs this up. Microsoft Copilot for Security produced measurable MTTR reductions across 378 organizations: 25.9% faster investigation, 30% reduction in mean time to respond. Palo Alto’s XSIAM crossed $1B in cumulative bookings with documented ROI.

But we have to say this out loud: AI is making attackers better at the same rate. Generative AI has lowered the barrier to automated phishing, reconnaissance, and exploit development. If we adopt AI defensively without a strategy, we’re running faster toward the wrong mountain. Speed doesn’t help when you haven’t aligned on where you’re going.

There’s also a subtler challenge that doesn’t get enough attention. AI’s reasoning quality depends entirely on the context it receives. Feed it clean, enriched, well-structured security data and it performs. Feed it fragmented logs with inconsistent schemas and missing fields, and it produces confident analysis built on a broken foundation.

The context problem that plagued every previous generation doesn’t disappear with AI. It moves upstream. Your data architecture becomes the bottleneck instead of your analyst headcount.


Four Lessons from Twenty Years of Security Tool Failures

If we’re willing to use them:

Visibility without context is noise. Every generation gave us more data. AI is not exempt. It can hallucinate context that doesn’t exist, which may be worse than no context at all.

Automation without judgment creates gaps. SOAR playbooks broke on edge cases. Agentic AI will break in different, less predictable ways. Human oversight isn’t bureaucratic friction. It’s the mechanism that catches where the tool fails.

Tools expose missing expertise, not replace it. Every generation promised to reduce the analyst burden. Every generation required more skilled people to manage the tools. AI might change this, but only if we deploy it where it solves a real problem, not where it demos well. We still need people who know our data and our threat landscape. AI amplifies that knowledge. It doesn’t create it.

Tool adoption isn’t strategy. It’s a purchase order. Having SOAR didn’t mean using SOAR effectively. Organizations that deploy AI without targeting specific failure modes will repeat the early SIEM experience: an expensive system producing impressive outputs that nobody acts on.


Align on the Mountain Before You Start Climbing

Three generations have played out. We’re living through the beginning of the fourth. If there’s one thing I’d tell a fellow CISO evaluating AI for security operations, it’s this: the technology is finally capable enough. The risk is in the implementation.

Pick the mountain first. What specific problem are we solving? Alert fatigue at tier-1? Slow enrichment during incident response? Inconsistent investigation quality across analysts? Pick one. Measure it before you start. Measure it after. If your team can’t name the failure mode AI is supposed to fix, you’re not ready for the vendor conversation yet.

Keep humans in the loop longer than feels necessary. Vendors will say their system is ready for autonomous action. Maybe it is. But the asymmetry matters. An analyst catching an AI false positive costs minutes. AI taking autonomous action on a false positive can cost significantly more. Trust needs to be earned, not assumed. Even if the first step is just observing what the AI recommends without acting on it, that’s the right starting point.

Invest in your data before you invest in AI. This is the step most teams want to skip, and it’s the one that matters most. Every AI system is downstream of data quality. Duplicate logs, misconfigured asset data, inconsistent source formatting: that’s what you’re feeding the model. The boring infrastructure work is the prerequisite, not the afterthought. (More on this in an upcoming article in this series.)

Treat this as a program, not a deployment. SIEM took years to mature. SOAR took years. AI will too. The organizations that succeed won’t be the ones that bought the most advanced platform on day one. They’ll be the ones that started deliberately, measured rigorously, and built toward maturity with documented outcomes as the goal.

Maturity LevelWhat It Looks LikeTimeline
ExploringEvaluating vendors, running pilots on low-risk use casesMonths 1–3
TargetingAI deployed against 1–2 specific failure modes, measured against baselinesMonths 3–9
IntegratingAI woven into daily SOC workflows, human oversight at defined thresholdsMonths 9–18
OptimizingExpanding autonomy based on earned trust and documented outcomesYear 2+

Where I Land

Three generations of vendor promises. Each one better than the last. Each one harder to work with than what came before it, at least in the early years.

Now we’re in the fourth. AI SOC Agents address the two structural failures the previous generations couldn’t crack: the context gap that SIEM left open, and the SME bottleneck that strangled SOAR. For the first time, the technology operates on a principle that could actually reduce the burden on analysts instead of adding to it.

Could. Not will. That depends on us.

Align on the right mountain. Start with specific problems. Invest in the data. Keep humans accountable. Do those four things, and this generation of tools earns a different ending than the last three.

We’ve been given another shot. Let’s not waste it the same way.

Your next step: Before your next AI vendor meeting, write down three specific problems you want AI to solve. Not “improve our SOC.” Failure modes with measurable baselines. If you can’t name them, the meeting can wait.


Next in this series: The SIEM Cost Trap — Why Your Data Lake + AI Agents Will Win

David O’Neil is a CISO and builder focused on the practical realities of security operations, AI adoption, and building programs that actually work. Find him on Twitter/X.

Related Posts

4 Essentials for Executive & Business Buyin on your Incident Response Plan

4 Essentials for Executive & Business Buyin on your Incident Response Plan

The impact and subsequent fallout from a business-impacting cyber security attack are stressful at the best of times. Experience time and again shows that organizations without the benefit of an Inci

read more
The CyberSecurity & Evolving Threats

The CyberSecurity & Evolving Threats

Cybersecurity is a critical concern in today's world, as more and more of our daily lives are conducted online. The threat landscape is constantly evolving, and it can be challenging to keep up with t

read more
Top 5 things for a Successful Cyber Response 'IR' Plan

Top 5 things for a Successful Cyber Response 'IR' Plan

Incident Response Planning & Strategy How important is an Incident Response Plan? Some studies show that just having a plan, can reduce the cost of a breach [example one](https://insights.integrity36

read more
Pre-Selection Beats Post-Selection: How I Made Claude Code 10-30x Faster

Pre-Selection Beats Post-Selection: How I Made Claude Code 10-30x Faster

Every code navigation costs time. When you multiply 300ms delays across hundreds of searches per day, you're losing hours p

read more
I Ran 849 Tests on AI Context Files. Here's What Actually Works.

I Ran 849 Tests on AI Context Files. Here's What Actually Works.

After 849 controlled tests, $20 in API costs, and a week of experiments, I can tell you exactly how to organize your Claude Code reference files. The short version: Put everything in one flat fol

read more
How I Made Claude Code Safer (And You Can Too)

How I Made Claude Code Safer (And You Can Too)

I've been running Claude Code on real projects for months. It's great at writing code — but it doesn't always understand the consequences of what it writes. Claude Code validates which tools can run.

read more
Claude Code Has Two New CVEs — Here's What They Exploit and How to Harden Your Setup

Claude Code Has Two New CVEs — Here's What They Exploit and How to Harden Your Setup

Your engineers cloned repositories today. Probably dozens. If any of those repos contained a malicious .claude/settings.json, they may have executed arbitrary shell code without a single confirmatio

read more
I Scanned 152 Files of My Own AI-Generated Code for Invisible Unicode Malware

I Scanned 152 Files of My Own AI-Generated Code for Invisible Unicode Malware

Two weeks ago, a supply chain attack called Glassworm compromised 150+ GitHub repositories and 72+ browser extensions by hiding malicious payloads in characters that are literally invisible in every

read more
The Math Problem AI Just Changed for Security Testing

The Math Problem AI Just Changed for Security Testing

Published: 2026-03-22 | RSA 2026 Pre-Conference SeriesHere's the problem every security team lives with but rarely says out loud. Your environment changes every time a developer merges code,

read more
The SIEM Cost Trap — Why Your Data Lake + AI Agents Will Win

The SIEM Cost Trap — Why Your Data Lake + AI Agents Will Win

If you've ever sat across from your CFO, your VP of Engineering, or your board and tried to explain why your SIEM costs what it costs — you already know how this conversation goes. The short version o

read more
Your Data Lake Is Only as Useful as Its Ability to Answer a Question

Your Data Lake Is Only as Useful as Its Ability to Answer a Question

You moved your security data out of the SIEM and into a data lake. Costs dropped. For the first time in years, you had budget to spare. Then an investigation hit — and your team spent two weeks findi

read more