Train your SOC like a triathlete: improve visibility, standardize data, then apply AI where it measurably reduces investigation time and uncertainty.

Train Your SOC Like a Triathlete (With AI in the Mix)
Most SOC leaders are buying AI before they’ve earned the right to benefit from it.
Not because AI is overrated—because AI in cybersecurity is brutally honest. If your telemetry is spotty, your log fields don’t match, and your analysts can’t quickly prove what happened, AI won’t fix that. It’ll just produce faster confusion, higher alert volume, and a fresh wave of “Why did the model say that?” conversations.
A better way to think about SOC performance is the triathlon metaphor: swim (readiness), bike (consistency), run (confidence and endurance). Triathletes don’t win by buying a nicer bike while skipping training. SOCs don’t win by deploying a new AI assistant while their data foundation is thin.
This post is part of our AI in Cybersecurity series, and it takes a stance: your SOC’s “fitness” depends more on evidence quality than tool count. AI is the training plan and the stopwatch—but only if you can feed it clean, connected, long-lived data.
Swim: Build investigation-ready coverage and retention
If you want AI-powered threat detection to work, start by making sure your SOC can see.
Alerts are the starting gun. The real work begins after an alert fires: scoping impact, reconstructing activity, and proving whether something is malicious. That requires coverage across your environment and retention long enough to match real attacker dwell time.
The most common SOC failure: short memory
Many organizations still retain key packet or high-fidelity network evidence for 7–14 days. Meanwhile, real-world intrusions often sit in environments for 30–180 days before anyone connects the dots. That mismatch creates a predictable outcome:
- Analysts chase alerts without historical context
- AI models “hallucinate” patterns because the timeline is incomplete
- Incident response turns into guesswork, not reconstruction
A practical target I’ve seen work well (and aligns with high-performing teams) is:
- 90–95% coverage of the environment for core telemetry
- 6–12 months of searchable data for investigations and threat hunting
The number isn’t magical; the measuring is. Teams routinely assume they have 90% visibility, then discover it’s closer to 70% once they map real coverage by network segment, cloud account, and identity domain.
What “swim fitness” looks like for an AI-enabled SOC
AI doesn’t just need “logs.” It needs consistent, continuous evidence. If you’re planning AI for SOC automation, validate these inputs first:
- Network visibility for lateral movement, command-and-control patterns, and weird internal DNS
- Endpoint visibility for process lineage and persistence
- Identity signals for privilege changes, logins, and token abuse
- Cloud control-plane logs for configuration attacks and service-to-service abuse
When these are missing, AI systems will still output answers—just not reliable ones. Your analysts will stop trusting the model, which kills adoption faster than any budget cut.
Bike: Standardize your data so AI doesn’t amplify chaos
Consistency is the difference between a fast investigation and a five-hour argument about what “source” means.
Here’s the thing about modern SOC stacks: they’re a patchwork of tools built by different vendors, using different schemas, observing traffic from different points of view. One product’s “source” might mean the local endpoint process. Another’s “source” might mean the external IP talking to your firewall. Put both into an AI pipeline and you get a model that’s confidently wrong.
Fix the schema problem before you automate
If you want AI to summarize incidents, correlate alerts, or recommend actions, you need semantic alignment. That means standardizing definitions and mapping fields across feeds so investigations don’t stall.
Start with a simple internal “SOC dictionary” for:
src_ip,dst_ip(and how NAT is represented)useranddeviceidentity (including service accounts)process,parent_process,command_linesession_id/flow_idto link network events end-to-end
Then enforce it. Not in a policy doc nobody reads—enforce it in pipelines, parsers, and detection engineering reviews.
Treat evidence like a product, not exhaust
A lot of logs weren’t built for security investigations. They were built for debugging systems. That’s why you get records that tell you “allowed/blocked” but can’t tell you what happened after access was granted.
For SOC performance, evidence should behave like a product:
- Designed to be queried (fast, normalized, documented)
- Cross-linked (IP ↔ session ↔ user ↔ device ↔ object)
- Consistent (same field names and meanings across sources)
This is where AI can shine later: if your data is integrated and consistent, AI can do high-confidence enrichment, summarization, clustering, and guided investigation steps.
Snippet-worthy truth: AI doesn’t reduce complexity. It increases the speed at which complexity hurts you.
Run: Use AI to turn evidence into confidence (not noise)
The “run” leg is where SOCs either prove impact—or fold under uncertainty.
In incident response, confidence is a deliverable. Leaders don’t just want “we think it’s contained.” They want evidence-backed statements like:
- What systems were accessed
- What data was actually exfiltrated
- Whether persistence remains
- Whether the attacker achieved objectives
A real example from ransomware response illustrates why this matters: evidence showed that only 10% of what the attacker claimed to have exfiltrated was actually stolen. That gave leadership enough certainty to refuse an eight-figure ransom demand.
AI can help you get to that kind of certainty faster—but only if it can see the same evidence an experienced investigator would use.
A metric that exposes SOC fitness: “cause unknown” closures
If you want one KPI that cuts through vanity metrics, track this:
- How many cases close as “cause unknown” each month
Every reduction means your SOC is improving its ability to reconstruct reality. This metric also pairs well with AI because you can measure whether:
- AI-assisted investigations reduce unknowns
- AI summarization speeds up evidence review
- AI-driven enrichment improves decision quality
Where AI actually helps in the run leg
When the data foundation is solid, AI becomes useful in very specific ways:
- Tier 1 triage automation: clustering, deduplication, alert summarization
- Enrichment at scale: pulling identity context, asset criticality, vulnerability exposure, and prior sightings
- Analyst copilots: drafting investigation narratives, suggesting next queries, generating containment checklists
- Anomaly detection: spotting living-off-the-land behavior that won’t trigger traditional malware rules
The goal isn’t to replace analysts. It’s to reduce the number of times they have to manually assemble context across 10 tools while a timer is running.
How attackers use AI (and why “edge + stealth” is the new default)
Attackers still follow a simple strategy: hit your weakest control plane, then blend in.
Right now, that weakness is often the edge—VPNs, remote access gateways, exposed appliances, and identity seams between SaaS and on-prem. After initial access, many campaigns rely on living-off-the-land techniques that look like normal admin work. Traditional malware alerts won’t fire.
So what changes for an AI-enabled SOC?
You need baselines, not just signatures
To catch stealth, you need to know what “normal” looks like across:
- Authentication patterns (impossible travel is the easy case; token reuse and atypical admin tooling is the hard one)
- East-west movement (new RDP/SMB patterns, unusual internal DNS)
- Data access (new object store reads, abnormal export jobs)
AI can help build and maintain these baselines, but only if your inputs are stable and retained long enough to establish real seasonality. That matters in December, for example, when many orgs see:
- Reduced staffing and slower response times
- Increased change activity (year-end freezes breaking late)
- Higher phishing volume aimed at finance and HR workflows
If your SOC’s data drops out during holiday operations, AI baselines degrade right when you need them.
A practical 90-day “triathlon plan” for an AI-ready SOC
Most teams fail by trying to fix visibility, process, and AI adoption all at once. Don’t.
Here’s a staged approach that fits real budgets and real analyst time.
Days 1–30: Swim (visibility and retention)
- Map telemetry coverage by business-critical segments (cloud accounts, key VPCs/VNETs, data centers, identity)
- Identify your “blind months” (periods where you can’t reconstruct activity)
- Set a retention target for investigation-grade data (often 6–12 months) and prioritize the top environments first
Deliverable: a coverage-and-retention scorecard your SOC can defend.
Days 31–60: Bike (normalize and connect evidence)
- Standardize event schemas and field definitions
- Link identity, endpoint, and network evidence using stable identifiers
- Document what each dataset is good for and where it’s weak
Deliverable: a SOC data catalog that analysts actually use.
Days 61–90: Run (pilot 1–2 AI workflows)
Pick use cases with clear ROI and measurable outcomes. Two that consistently work:
- Tier 1 alert triage (summaries, deduplication, enrichment, routing)
- Compliance/audit evidence automation (collecting proof, mapping controls, reducing manual screenshots)
Guardrails that matter:
- Give the model least-privilege access but enough context to be useful
- Log prompts and outputs for review (treat it like another security system)
- Define “success” in SOC terms: reduced time-to-triage, fewer cause-unknown closures, improved containment time
Deliverable: an AI pilot report with before/after metrics that leadership understands.
What to ask vendors (and your own team) before buying more AI
If your goal is leads and outcomes—not a demo that looks good—these questions surface reality fast:
- What exact data does the AI need, and in what schema?
- How does it handle conflicting fields across tools?
- Can it link identity ↔ endpoint ↔ network evidence without manual work?
- What’s the rollback plan if analysts stop trusting outputs?
- Which SOC metric will improve in 30 days, and how will we measure it?
If you can’t answer these clearly, you’re not “behind.” You’re just not ready to spend yet.
Where this fits in the AI in Cybersecurity story
AI in cybersecurity is heading toward more autonomy—agentic workflows that recommend actions, open tickets, quarantine devices, and generate executive narratives. That future will reward SOCs that treat evidence as a first-class asset.
Train your SOC like a triathlete: build readiness, enforce consistency, then use AI to convert evidence into confidence. If you start there, the AI tools you choose later will actually make you faster.
If you’re planning your 2026 SOC roadmap, here’s the question that tells me whether AI will help you or hurt you: When your next incident hits, will you have six months of consistent, searchable evidence—or a handful of alerts and a lot of opinions?