Type something to search...
What AI Is Actually Doing in Your SOC — and What It Shouldn't Be Doing Yet

What AI Is Actually Doing in Your SOC — and What It Shouldn't Be Doing Yet

Series: The SIEM & AI Reckoning | Article 4 of 10


Only 9% of security practitioners say they’re “very confident” in AI-generated alerts and recommendations.

Alerting is just one of a dozen things vendors are asking AI to handle in the SOC right now. If confidence is that low on the use case they’ve been polishing the longest, what does that tell you about the rest?

That stat comes from Gurucul’s 2025 Pulse of the AI SOC — 739 cybersecurity professionals, not vendor analysts, not academics. The other 91% range from “mostly trust with review” to “helpful but requires frequent validation.” And yet adoption is accelerating faster than at any point in security history.

The problem isn’t that AI doesn’t work. It’s that most organizations haven’t been honest about the difference between what AI can do in a demo and what it actually delivers — repeatedly, reliably, and context-aware — in their environment.

Most of these implementations treat AI as a one-shot replacement. Hand it the problem, get back the answer. That’s not how this works. The teams getting real value are using AI at specific decision points within existing workflows — automation with AI augmenting the steps, not replacing the process.

(The separate question of how to manage tokens, context windows, and cost across all of these use cases is real and important — but it’s a different article.)


Four Ways AI Shows Up in the SOC

One thing the vendors won’t name: not all “AI-powered” capabilities work the same way. I’ve seen four distinct patterns in how AI actually operates in security workflows:

PatternWhat It MeansHuman Role
Set and forgetConfigured once, runs continuouslyReview baselines periodically
Monitor and tuneAutomated but drifts without calibrationCheck outputs, adjust thresholds
Human-in-the-loopAI recommends, human approves or actsReview every recommendation
Human-initiatedHuman triggers AI for each use caseDrive each interaction

When a vendor says “AI-powered,” they could mean any of these. The demo doesn’t distinguish between them. Your budget should.

Use CaseStatusPattern
Pipeline health monitoring✅ Do itSet and forget
Threat intel attribution✅ Do itSet and forget
Correlation discovery✅ Do itMonitor and tune
Detection creation✅ Do itMonitor and tune
Alert engineering✅ Do itHuman-in-the-loop
Escalation triage✅ Do itHuman-in-the-loop
Triage with runbooks✅ Do itHuman-in-the-loop
Data model building✅ Do itHuman-initiated
Parsing assistance✅ Do itHuman-initiated
Final decisions across the response chain⛔ Not yetTrust gap
Hypothesis-driven hunting⛔ Not yetHuman intuition
Triage without runbooks⛔ Not yetGeneric fails unpredictably
Real-world actions beyond triage⛔ Not yetContext gap
Operating without accountability⛔ Not yetNo audit trail

AI SOC Capability Spectrum


What AI Shouldn’t Be Doing Yet

I’m starting here deliberately. The mistakes are more expensive than the missed opportunities.

Making Final Decisions Across the Response Chain

Come back to that 9% confidence figure. If 91% of practitioners don’t have high confidence in AI-generated alerts, you don’t yet have the trust foundation to remove humans from the decision loop.

And “decisions” isn’t just detection. It’s the entire response chain — did this actually happen? Should I pull more data? Should I block this account? Should we roll back this change? Every one of those needs different context to get right. Lumping them together under “AI-driven response” is how you end up with an automated system that triages well but contains badly.

AI-assisted detection — where AI surfaces candidates and humans validate — works. What doesn’t work is the next step: AI making the call and you finding out after the fact. Not because it’s always wrong, but because we don’t have reliable ways to know when it’s wrong in ways that matter.

And the threat that doesn’t look like anything in the training data? That’s the one you need to catch — and the one AI is most likely to miss.

Replacing Hypothesis-Driven Threat Hunting

Threat hunting is intentionally ambiguous. You’re looking for things without a signature, in data without obvious patterns, using intuition built from years of incident response.

AI is excellent at finding known patterns faster. It is not good at the “what if” question — the hypothesis a senior analyst forms when something feels wrong but doesn’t match any rule. That instinct draws on context that isn’t in your SIEM: organizational dynamics, recent changes not yet reflected in logs, behavioral knowledge about specific users accumulated over years.

Use AI to accelerate hunting workflows — but build those workflows with structure. What are you looking for? Where should AI search? What does “interesting” look like for this specific hunt? Without that framing, you’re burning tokens on the AI equivalent of grep *bad_thing* across your entire lake. That gets expensive fast and returns noise.

Triage Without Runbooks

Triage is where most vendors lead in their demos. It’s also where the most implementation failures happen.

The problem isn’t AI. It’s that organizations hand AI a pile of alerts and say “prioritize these” without giving it the framework to know what prioritization means in their environment. An AI given a generic triage instruction applies it generically — getting some right and some wrong in ways that are hard to predict. And since the agent’s context resets with each interaction, you can’t even count on it being consistently wrong the same way.

You need good runbooks. Not great ones — good ones. The step-by-step logic your L1 team uses for each alert type, the questions they ask, the context they check. Don’t rely on an AI agent to know your environment well enough to build a runbook on the fly — that’s a different problem with a different failure mode. AI trained on your runbooks triages the way your team triages. Without them, it’s guessing with confidence.

Taking Real-World Actions Beyond Triage

This is the line that matters most — and the one most vendors are actively working to push past.

AI triage, enrichment, correlation, and recommendation — that’s the right boundary today. Notifying your security operations team is escalation, and it’s appropriate. But the moment AI starts isolating endpoints, blocking accounts, or triggering automated containment, you’re in fundamentally different territory. Don’t jump to autonomous action. Build the foundation layers first.

A user downloading an unusual volume of data might be an insider threat — or the VP of Finance closing quarter-end books, or an authorized migration nobody told security about. We’ve done a poor job capturing that business context in a form AI can reason from — it lives in our heads, in Slack threads, in conversations that never made it into a runbook.

Until that foundation exists, the right boundary is: AI recommends, humans act.

Operating Without Accountability

There’s no law today requiring that AI-driven security decisions be observable, defensible, or auditable. No regulator is asking for an AI decision trail the way they ask for firewall logs.

I believe that’s going to change.

The security teams building audit trails for AI decisions now — documenting what context the AI had, what logic it applied, what actions it took and why — are building the infrastructure that will eventually be required. The teams that don’t are accumulating technical debt they’ll pay later, under pressure, in the middle of an incident.

The habit is worth building now. Before it’s mandated.


What AI Should Be Doing in Your SOC

These are the areas where I’ve seen AI actually deliver — when the configuration work gets done.

Parsing Assistance

This one isn’t glamorous. It might still be the highest-leverage starting point.

Writing and maintaining parsers across dozens of log formats — without breaking every time a vendor updates their schema — consumes analyst time at a rate most organizations seriously underestimate. AI can analyze log samples, suggest parsing logic, and flag when source formats change. The real value is the monitoring — it catches schema drift before your analysts discover it in the middle of an investigation. Your team doesn’t need to know the parsing language. They need to validate that AI got it right.

Pattern: Human-initiated. Your team triggers it, AI drafts it, humans validate.

Building the Data Model

This is the work every analyst loves to do.

(No, it isn’t.)

It’s hard, tedious, and if done correctly and automated, one of the biggest wins for your organization. It’s where AI is most useful for the tedious parts — detecting inconsistencies, flagging sources that don’t conform, suggesting mappings against standards like OCSF or OpenTelemetry. It doesn’t have to be a full rewrite of your logs — it can be as simple as aliasing existing field names to a standardized framework: otel_src_ip → firewall.source_ip_address.

Pattern: Human-initiated to human-in-the-loop. AI suggests mappings, humans make the judgment calls.

Pipeline Health Monitoring

Your SIEM has gaps in its data collection right now. You probably don’t know which ones.

AI monitors pipeline health continuously — it knows when sources stop sending, when volumes shift off baseline, when something new shows up unclassified. The absence of this capability is the most consistent gap I see in otherwise mature security programs. Teams with significant SIEM investments that don’t know when their data pipeline is broken.

Pattern: Set and forget. Automated once configured.

Correlation Discovery

If you built out that data model, this becomes a straightforward next step. Standardize on otel_username across every source, and suddenly “show me everything user X did in the last 24 hours” is one query — not a manual pivot across six consoles with six different field names.

This is where AI surfaces attack sequences that cross data source boundaries — patterns your analysts would never have the bandwidth to find. But it can only discover what it has access to discover — without normalized data, it’s working from an incomplete picture. This is why the data architecture work in earlier articles isn’t a nice-to-have. It’s the prerequisite.

Pattern: Monitor and tune. Scales with your data model maturity.

Threat Intel Attribution and Escalation

I see teams waste the most money here: using AI tokens to do work the SIEM already handles. Your SIEM should be matching IOCs — IPs, hashes, domains — against threat feeds. That’s pattern matching. The SIEM does it fine.

Where AI earns its keep is the layer above: attribution and ranking. Looking up whois data on a flagged IP and recognizing it’s a major telecom provider — lower priority. Mapping a behavior pattern to a known threat actor’s TTPs and recognizing the targeting fits your industry — escalate now. Your SOC can codify this judgment in natural language that AI applies consistently at scale.

Pattern: Set and forget (lookup) to human-in-the-loop (escalation judgment).

Triage — With Your Runbooks

Microsoft’s Security Copilot study measured a 30.13% reduction in mean time to resolution in a matched study of 177 organizations. That result came from teams that did the configuration work.

Not “triage by severity.” Your actual analyst playbooks — the step-by-step logic per alert type, the questions they ask, the threshold at which they escalate. AI trained on your runbooks doesn’t just triage faster. It triages the way your team does — every time, at 3 AM, without the coffee.

Pattern: Human-in-the-loop when well-configured. Outcome scales with runbook quality.


The Gap That Determines Your Outcome

The 60% of AI SOC adopters who’ve cut investigation time by at least 25% did the configuration work. The teams still frustrated with their AI tools largely haven’t.

If you don’t have the team to build this yourself, managed security providers are moving in this direction. Be careful — if they own the models, the runbooks, and the pipeline, you can find yourself in vendor lock-in faster than you would with a traditional SIEM contract. Own your data model and your runbooks, even if someone else operates the tooling.

Your Next Moves

Good (this week): Audit your current AI deployment against this framework. Which “should” use cases have you actually configured — not just purchased? Which “shouldn’t” use cases are running without guardrails? A gap list is the starting point.

Better (this month): Pick one use case and do the configuration work. Pipeline health monitoring is the highest-ROI starting point — set and forget once configured, and the most consistent gap I see in otherwise mature programs.

Best (this quarter): Build the runbook library that makes AI triage operational. Document your top 10 alert types with step-by-step logic, escalation thresholds, and context requirements. Good runbooks, not perfect ones. This is the work that turns AI from a capable technology into a consistent one.


Next in this series: The vendor landscape — who’s building real AI SOC capabilities, who’s still showing demos, and how to tell the difference before you sign.

David O’Neil is a CISO with 20+ years in cybersecurity leadership. He writes about the practical realities of security operations, AI adoption, and building security programs that survive budget season at cisoexpert.com.

Related Posts

4 Essentials for Executive & Business Buyin on your Incident Response Plan

4 Essentials for Executive & Business Buyin on your Incident Response Plan

The impact and subsequent fallout from a business-impacting cyber security attack are stressful at the best of times. Experience time and again shows that organizations without the benefit of an Inci

read more
The CyberSecurity & Evolving Threats

The CyberSecurity & Evolving Threats

Cybersecurity is a critical concern in today's world, as more and more of our daily lives are conducted online. The threat landscape is constantly evolving, and it can be challenging to keep up with t

read more
Top 5 things for a Successful Cyber Response 'IR' Plan

Top 5 things for a Successful Cyber Response 'IR' Plan

Incident Response Planning & Strategy How important is an Incident Response Plan? Some studies show that just having a plan, can reduce the cost of a breach [example one](https://insights.integrity36

read more
Pre-Selection Beats Post-Selection: How I Made Claude Code 10-30x Faster

Pre-Selection Beats Post-Selection: How I Made Claude Code 10-30x Faster

Every code navigation costs time. When you multiply 300ms delays across hundreds of searches per day, you're losing hours p

read more
I Ran 849 Tests on AI Context Files. Here's What Actually Works.

I Ran 849 Tests on AI Context Files. Here's What Actually Works.

After 849 controlled tests, $20 in API costs, and a week of experiments, I can tell you exactly how to organize your Claude Code reference files. The short version: Put everything in one flat fol

read more
How I Made Claude Code Safer (And You Can Too)

How I Made Claude Code Safer (And You Can Too)

I've been running Claude Code on real projects for months. It's great at writing code — but it doesn't always understand the consequences of what it writes. Claude Code validates which tools can run.

read more
Claude Code Has Two New CVEs — Here's What They Exploit and How to Harden Your Setup

Claude Code Has Two New CVEs — Here's What They Exploit and How to Harden Your Setup

Your engineers cloned repositories today. Probably dozens. If any of those repos contained a malicious .claude/settings.json, they may have executed arbitrary shell code without a single confirmatio

read more
I Scanned 152 Files of My Own AI-Generated Code for Invisible Unicode Malware

I Scanned 152 Files of My Own AI-Generated Code for Invisible Unicode Malware

Two weeks ago, a supply chain attack called Glassworm compromised 150+ GitHub repositories and 72+ browser extensions by hiding malicious payloads in characters that are literally invisible in every

read more
Four Generations of Broken Promises: Why AI SOC Agents Might Actually Be Different

Four Generations of Broken Promises: Why AI SOC Agents Might Actually Be Different

Series: The SIEM & AI Reckoning — Article 1 of 10Over twenty years and hundreds of vendor pitches, one line never changes: "This is going to change everything." 2005, SIEM. 2012, Next-Gen

read more
The Math Problem AI Just Changed for Security Testing

The Math Problem AI Just Changed for Security Testing

Published: 2026-03-22 | RSA 2026 Pre-Conference SeriesHere's the problem every security team lives with but rarely says out loud. Your environment changes every time a developer merges code,

read more
The SIEM Cost Trap — Why Your Data Lake + AI Agents Will Win

The SIEM Cost Trap — Why Your Data Lake + AI Agents Will Win

If you've ever sat across from your CFO, your VP of Engineering, or your board and tried to explain why your SIEM costs what it costs — you already know how this conversation goes. The short version o

read more
Your Data Lake Is Only as Useful as Its Ability to Answer a Question

Your Data Lake Is Only as Useful as Its Ability to Answer a Question

You moved your security data out of the SIEM and into a data lake. Costs dropped. For the first time in years, you had budget to spare. Then an investigation hit — and your team spent two weeks findi

read more