More than half of the CISOs and CIOs I speak with ask the same two questions: “What should I measure?” and “What should I be looking at?” What’s far less common is a clear answer to what decisions those metrics will drive once they exist. If the Governance Council does not understand the metrics, your boss does not know what “good” looks like, and you cannot manage your team by them, then the metrics are not helping anyone.
Before you collect a single metric, think about what you’re trying to achieve.
The most important question is not what you should measure. It’s why you want to measure anything at all.
If your answer is “because it’s what people in security do,” stop right there. That alone is not good enough. In fact, it’s usually how security teams end up buried in dashboards that no one uses and metrics no one acts on.
I see security programs fall into one of three categories.
The first is the passion-project SOC. The team is motivated, capable, and genuinely cares about cybersecurity, but no one else does. The Governance Council isn’t asking about performance. The CIO isn’t reviewing metrics. Unless there’s an incident, security is largely left alone. If this is your reality, you should not start a metrics program. Not yet. Metrics without demand or authority quickly turn into internal theater. Your time is better spent optimizing tools, fixing asset coverage, improving signal quality, and building automations that actually reduce response time.
The second category is the maturing program. The SOC exists, the basics are in place, and the CISO is pushing for operational excellence. Metrics start to feel attractive here because the leader wants to optimize, tune, and improve. That’s valid, but it’s critical to recognize that these metrics are for internal learning, not external accountability. You are measuring systems so you can improve them, not to prove value to an audience that isn’t asking yet.
The third category is the service-oriented security program. Cybersecurity operates alongside IT operations. Performance matters. SLAs exist. Security teams are expected to demonstrate reliability, responsiveness, and alignment with business needs. In this model, metrics are no longer optional. They are how security earns trust and proves it belongs at the operational table.
Which bucket you’re in determines everything. It determines whether metrics are useful, which ones matter, and how precise they need to be. Treating all SOCs the same is how organizations waste enormous energy measuring things they cannot influence.
With that context set, let’s talk about what SOC metrics are actually for.
Security Operations Center (SOC) Metrics in Public-Sector Cybersecurity
Public-sector cybersecurity is not built for perfection. It’s built for endurance.
Budgets are constrained. Staffing is thin. Compliance requirements are heavy. Citizens expect services to be available regardless of what’s happening behind the scenes. In that environment, SOC metrics are not aspirational goals. They are operational diagnostics.
Metrics like Mean Time to Acknowledge (MTTA), Mean Time to Detect (MTTD), and Mean Time to Respond (MTTR) are often treated as performance grades. That’s a mistake. These metrics don’t tell you how good your SOC is. They tell you how your systems behave under pressure.
If you treat metrics as goals, you create stress and distortion. If you treat them as indicators of system health, you create improvement.
There’s a critical distinction here. A goal says, “Respond in four hours.” A system says, “Here’s how incidents actually move through our environment.” One invites shortcuts. The other invites engineering.
SOC metrics exist to reveal friction. They expose where alerts stall, where handoffs break down, where automation helps, and where it doesn’t. When used correctly they become design inputs. Metrics should expose barriers, not reward shortcuts.
Key SOC Metrics, Organization, and Structure
MTTA: Mean Time to Acknowledge
MTTA measures how long it takes for an alert to be seen and acknowledged by a human. On paper, this looks like a simple speed metric. In practice, it reflects alert prioritization, notification design, and workload distribution.
If MTTA is consistently high, the issue is rarely effort. It is usually unclear ownership, poor alert routing, or excessive noise. If MTTA is consistently low but nothing improves downstream, it may indicate rushed acknowledgment without meaningful triage.
MTTA is most useful when tracked by severity and time of day. Leaders should look at trends, not single data points, and use them to evaluate whether alerts are reaching the right people quickly enough to matter.
MTTD: Mean Time to Detect
MTTD measures the time between an event occurring and it being detected by the SOC. This metric is heavily influenced by data coverage, telemetry reliability, and detection logic.
A long MTTD does not necessarily mean analysts are missing threats. More often, it indicates gaps in logging, broken data pipelines, or detection rules that are too noisy to surface real issues quickly.
MTTD should be paired with metrics that show detection coverage and alert quality. Measuring detection time without understanding what is being detected, or missed, creates false confidence.
MTTR: Mean Time to Respond (or Remediate)
MTTR is often treated as the ultimate SOC metric, and it is frequently misapplied.
MTTR measures how long it takes to contain or remediate an incident after detection. What it actually reveals is how well coordination works across teams. Delays often occur outside the SOC: change approvals, handoffs to infrastructure teams, legal review, or manual remediation steps.
If MTTR is high, leaders should resist the urge to push analysts harder. Instead, they should examine where delays occur and whether expectations are realistic. MTTR becomes meaningful only when the scope of “response” is clearly defined and consistently measured.
Detection Coverage and Quality
Speed without accuracy is meaningless. Metrics that track true positive rates, false positives, and false negatives provide critical context to MTTA, MTTD, and MTTR.
High false positives waste time and inflate workload. High false negatives hide risk. Leaders should expect some tradeoff between speed and precision and should track both explicitly.
Detection quality metrics help determine whether improvements in response time are actually reducing risk or simply moving faster on the wrong things.
Some examples are:
Easy to measure
- In-scope log source coverage (% of required sources actively reporting)
- Sensor/agent health rate (EDR, IAM, email, cloud, network)
- Detection validation freshness (% tested in last 30/60/90 days) if you track tests as tickets
- False positive rate (FPR) by severity or detection family if analysts tag dispositions
- Alert-to-incident conversion rate
- Analyst touch time per alert (median) if case management tracks timers
- Alert ingestion success rate (if measured inside SIEM/SOAR)
Medium (effort/skill) to measure
- True positive rate (TPR) by severity (needs consistent “confirmed” criteria)
- Detection confidence score by alert category (requires scoring model and discipline)
- False negative discovery rate (incidents found outside SOC detection) requires consistent capture
- ATT&CK technique coverage for priority threat scenarios (requires mapping and maintenance)
Hard to measure
- False negative rate (true “miss rate” across the environment) beyond sampled discovery
- “Coverage completeness” that proves you’d catch novel TTPs (requires testing/purple teaming)
- Detection quality normalized across agencies/tools (needs standard taxonomy and normalization)
Operational Hygiene and Backlog
Backlog metrics reveal whether the SOC can keep up with demand. Tracking ticket aging, percentage of cases outside SLA, and after-hours workload helps leaders understand whether capacity is balanced.
These metrics are not about productivity or individual performance. They are early indicators of burnout, under-resourcing, or excessive alert volume.
Consistent backlog growth is a signal that something needs to change; either through automation, tuning, or staffing.
Some examples are:
Easy to measure
- Open case backlog by severity (P1/P2/P3)
- Tickets aging beyond SLA (% outside SLA by severity)
- Case reopen rate
- On-call coverage rate
- Page acceptance time (P1) if paging tool logs it
- After-hours page volume per analyst
- Escalation bounce rate if paging/escalation is logged
- Automation closure rate (% auto-resolved) if SOAR tracks outcomes
- Log/source “dark time” duration if you alert on missing data
Medium to measure
- Time from acknowledge to triage decision (needs consistent timestamps/workflow states)
- Analyst utilization mix (triage vs investigation vs improvement) requires time tagging
- “Alert noise ratio” (benign vs actionable) requires tagging discipline
- Queue volatility (spikes by day/week) tied to cause (campaign vs mis-tuning)
Hard to measure
- Analyst workload fairness (true cognitive load, context switching, fatigue)
- “Operational friction” metrics (handoff delays by team) without strong ITSM integration
- Burnout risk scoring that’s defensible (needs HR and ops data, and careful handling)
Governance and Evidence Metrics
In public-sector environments, governance metrics matter. Patch timelines, evidence freshness, and control validation are often driven by regulatory requirements, but they also provide insight into execution discipline.
These metrics should be reviewed on a regular cadence and tied to accountability. When evidence is stale or remediation is overdue, it should trigger follow-up; not just reporting.
Some examples are:
Easy to measure
- PIR completion rate and timeliness (if PIR is a tracked ticket/task)
- Corrective action closure rate from PIRs (if actions are tracked)
- Control evidence freshness (% updated within 90 days) if evidence is cataloged
- Open policy exception count (with expiry) if exceptions are logged
- Privileged access recertification completion rate if run through IAM/GRC
- Orphaned privileged account count if identity is centralized
Medium to measure
- Regulatory control coverage status (CJIS / NIST / IRS-1075 / HIPAA) mapping and updates
- Internet-exposed asset exception count (requires accurate external attack surface inventory)
- Audit finding recurrence rate (requires normalized categories across audits/years)
Hard to measure
- “Control effectiveness” (not just evidence) tied to real risk reduction
- Proving audit readiness continuously across decentralized agencies
- Third-party risk “truth” (evidence quality, compensating controls validity, real remediation)
How Public-Sector Leaders Use SOC Metrics
The difference between mature and immature programs is not which metrics they track, it is what they do when the numbers move.
When metrics are trending positively, effective leaders use them to validate assumptions. They confirm that investments are paying off, that alert tuning is working, and that response processes are predictable. They do not immediately raise targets; they stabilize performance first.
When metrics degrade, mature leaders use them as entry points for investigation. A spike in MTTR prompts a review of recent incidents. A rise in backlog triggers workload analysis. A drop in detection quality leads to rule tuning or data validation.
Metrics should lead to questions, not conclusions.
The most important action tied to metrics is the post-incident review. For significant incidents, leaders should expect a short, clear explanation of what happened, how long each phase took, and why. Over time, these reviews create institutional knowledge and improve performance more than any dashboard.
Misconceptions and Challenges
One of the biggest mistakes organizations make is treating metrics as proof of maturity. Clean dashboards and downward-trending charts can mask real issues if no one is willing to interrogate them.
Another common pitfall is focusing on single numbers instead of distributions. Averages hide outliers. Tracking medians and high-percentile response times gives a more honest picture of performance.
Inconsistent definitions also undermine credibility. If different teams interpret MTTR differently, comparisons are meaningless. Definitions must be explicit and stable.
Finally, metrics must reflect reality. Public-sector SOCs often operate with limited hours, shared infrastructure, and constrained authority. Metrics should account for those constraints rather than pretending they do not exist.
Conclusion: Build Systems, Not Scoreboards
SOC metrics are not inherently valuable. Their value comes from how they are used.
In public-sector cybersecurity, metrics should help leaders understand performance, identify friction, and make informed decisions. They should be understandable, actionable, and tied to real outcomes.
If a metric does not influence behavior or decision-making, it does not belong in your program.
Rethinking SOC metrics is not about tracking more numbers. It is about tracking the right ones—and having the discipline to act on what they reveal.
Don't miss another article. Subscribe to our blog now.
Included Topics
Justin (he/him) is the founder and CEO of NuHarbor Security, where he continues to advance modern integrated cybersecurity services. He has over 20 years of cybersecurity experience, much of it earned while leading security efforts for multinational corporations, most recently serving as global CISO at Keurig Green Mountain Coffee. Justin serves multiple local organizations in the public interest, including his board membership at Champlain College.