NuHarbor Security
  • Solutions
    Solutions
    Custom cybersecurity solutions that meet you where you are.
    • Overview
    • Our Approach
    • Data Icon Resources
    • Consultation Icon Consult with an expert
    • By Business Need
      • Identify Gaps in My Cybersecurity Plan
      • Detect and Respond to Threats in My Environment
      • Fulfill Compliance Assessments and Requirements
      • Verify Security With Expert-Led Testing
      • Manage Complex Cybersecurity Technologies
      • Realize the Full Value of Microsoft Security
      • Security Monitoring With Splunk
    • By Industry
      • State & Local Government
      • Higher Education
      • Federal
      • Finance
      • Healthcare
      • Insurance
    Guide Defining Whole-of-State Security: Building Resilient States Through Unified Cybersecurity
    Defining Whole-of-State Security: Building Resilient States Through Unified Cybersecurity
    Read Guide
  • Services
    Services
    Outcomes you want from a team of experts you can trust.
    • Overview
    • Data Icon Resources
    • Consultation Icon Consult with an expert
    • Security Testing
      • Penetration Testing
      • Application Penetration Testing
      • Vulnerability Scanning
      • Wireless Penetration Testing
      • Internal Penetration Testing
      • External Penetration Testing
    • Assessment & Compliance
      • ARC-AMPE Compliance
      • CJIS Compliance
      • NIST 800-53
      • HIPAA Security Standards
      • ISO 27001
      • MARS-E Security Standards
      • New York Cybersecurity (23 NYCRR 500)
      • Payment Card Industry (PCI)
    • Advisory & Planning
      • Security Strategy
      • Incident Response Planning
      • Security Program Reviews
      • Security Risk Assessments
      • Virtual CISO
      • Policy Review
    • Managed Services
      • SOC as a Service
      • Microsoft Security Managed Services
      • Splunk Managed Services
      • Tenable Managed Services
      • CrowdStrike Managed Detection and Response (MDR)
      • Vendor Security Assessments
      • Curated Threat Intelligence
      • Vulnerability Management
    Guide Defining Whole-of-State Security: Building Resilient States Through Unified Cybersecurity
    Defining Whole-of-State Security: Building Resilient States Through Unified Cybersecurity
    Read Guide
  • Partners
  • Resources
    Resources
    Explore reports, webinars, case studies, and more.
    • Browse Resources
    • Consultation Icon Consult with an expert
    • Blog icon Blog
    • Podcast icon Podcast
    • Downloadable Assets icon Downloadable Assets
    Guide Defining Whole-of-State Security: Building Resilient States Through Unified Cybersecurity
    Defining Whole-of-State Security: Building Resilient States Through Unified Cybersecurity
    Read Guide
  • Company
    Company
    We do cybersecurity differently – the right way.
    • Overview
    • Data Icon Resources
    • Consultation Icon Consult with an expert
    • Leadership
    • News
    • Careers
    • Contact
    Guide Defining Whole-of-State Security: Building Resilient States Through Unified Cybersecurity
    Defining Whole-of-State Security: Building Resilient States Through Unified Cybersecurity
    Read Guide
  • Consult with an expert
  • Client support
  • Careers
  • Contact
1.800.917.5719
NuHarbor Security Blog
    • Industry Insights
    • Security Operations
    • Compliance
    • Advisory and Planning
    • Cybersecurity Technology
    • Security Testing
    • Application Security
    • Threat Intelligence
    • Managed Detection and Response
    • Managed Services
    • Cyber Talent
    • NuHarbor
January 29, 2026

Rethinking SOC Metrics in Public-Sector Cybersecurity

Justin Fimlaid Justin Fimlaid
Rethinking SOC Metrics in Public-Sector Cybersecurity

More than half of the CISOs and CIOs I speak with ask the same two questions: “What should I measure?” and “What should I be looking at?” What’s far less common is a clear answer to what decisions those metrics will drive once they exist. If the Governance Council does not understand the metrics, your boss does not know what “good” looks like, and you cannot manage your team by them, then the metrics are not helping anyone. 

Before you collect a single metric, think about what you’re trying to achieve. 

The most important question is not what you should measure. It’s why you want to measure anything at all. 

If your answer is “because it’s what people in security do,” stop right there. That alone is not good enough. In fact, it’s usually how security teams end up buried in dashboards that no one uses and metrics no one acts on. 

I see security programs fall into one of three categories. 

The first is the passion-project SOC. The team is motivated, capable, and genuinely cares about cybersecurity, but no one else does. The Governance Council isn’t asking about performance. The CIO isn’t reviewing metrics. Unless there’s an incident, security is largely left alone. If this is your reality, you should not start a metrics program. Not yet. Metrics without demand or authority quickly turn into internal theater. Your time is better spent optimizing tools, fixing asset coverage, improving signal quality, and building automations that actually reduce response time. 

The second category is the maturing program. The SOC exists, the basics are in place, and the CISO is pushing for operational excellence. Metrics start to feel attractive here because the leader wants to optimize, tune, and improve. That’s valid, but it’s critical to recognize that these metrics are for internal learning, not external accountability. You are measuring systems so you can improve them, not to prove value to an audience that isn’t asking yet. 

The third category is the service-oriented security program. Cybersecurity operates alongside IT operations. Performance matters. SLAs exist. Security teams are expected to demonstrate reliability, responsiveness, and alignment with business needs. In this model, metrics are no longer optional. They are how security earns trust and proves it belongs at the operational table. 

Which bucket you’re in determines everything. It determines whether metrics are useful, which ones matter, and how precise they need to be. Treating all SOCs the same is how organizations waste enormous energy measuring things they cannot influence. 

With that context set, let’s talk about what SOC metrics are actually for. 


Security Operations Center (SOC) Metrics in Public-Sector Cybersecurity

Public-sector cybersecurity is not built for perfection. It’s built for endurance. 

Budgets are constrained. Staffing is thin. Compliance requirements are heavy. Citizens expect services to be available regardless of what’s happening behind the scenes. In that environment, SOC metrics are not aspirational goals. They are operational diagnostics. 

Metrics like Mean Time to Acknowledge (MTTA), Mean Time to Detect (MTTD), and Mean Time to Respond (MTTR) are often treated as performance grades. That’s a mistake. These metrics don’t tell you how good your SOC is. They tell you how your systems behave under pressure. 

If you treat metrics as goals, you create stress and distortion. If you treat them as indicators of system health, you create improvement. 

There’s a critical distinction here. A goal says, “Respond in four hours.” A system says, “Here’s how incidents actually move through our environment.” One invites shortcuts. The other invites engineering. 

SOC metrics exist to reveal friction. They expose where alerts stall, where handoffs break down, where automation helps, and where it doesn’t. When used correctly they become design inputs. Metrics should expose barriers, not reward shortcuts. 

Key SOC Metrics, Organization, and Structure 

MTTA: Mean Time to Acknowledge

MTTA measures how long it takes for an alert to be seen and acknowledged by a human. On paper, this looks like a simple speed metric. In practice, it reflects alert prioritization, notification design, and workload distribution. 

If MTTA is consistently high, the issue is rarely effort. It is usually unclear ownership, poor alert routing, or excessive noise. If MTTA is consistently low but nothing improves downstream, it may indicate rushed acknowledgment without meaningful triage. 

MTTA is most useful when tracked by severity and time of day. Leaders should look at trends, not single data points, and use them to evaluate whether alerts are reaching the right people quickly enough to matter. 

MTTD: Mean Time to Detect 

MTTD measures the time between an event occurring and it being detected by the SOC. This metric is heavily influenced by data coverage, telemetry reliability, and detection logic. 

A long MTTD does not necessarily mean analysts are missing threats. More often, it indicates gaps in logging, broken data pipelines, or detection rules that are too noisy to surface real issues quickly. 

MTTD should be paired with metrics that show detection coverage and alert quality. Measuring detection time without understanding what is being detected, or missed, creates false confidence. 

MTTR: Mean Time to Respond (or Remediate) 

MTTR is often treated as the ultimate SOC metric, and it is frequently misapplied. 

MTTR measures how long it takes to contain or remediate an incident after detection. What it actually reveals is how well coordination works across teams. Delays often occur outside the SOC: change approvals, handoffs to infrastructure teams, legal review, or manual remediation steps. 

If MTTR is high, leaders should resist the urge to push analysts harder. Instead, they should examine where delays occur and whether expectations are realistic. MTTR becomes meaningful only when the scope of “response” is clearly defined and consistently measured. 

Detection Coverage and Quality 

Speed without accuracy is meaningless. Metrics that track true positive rates, false positives, and false negatives provide critical context to MTTA, MTTD, and MTTR. 

High false positives waste time and inflate workload. High false negatives hide risk. Leaders should expect some tradeoff between speed and precision and should track both explicitly. 

Detection quality metrics help determine whether improvements in response time are actually reducing risk or simply moving faster on the wrong things. 

Some examples are: 

Easy to measure

  • In-scope log source coverage (% of required sources actively reporting)
  • Sensor/agent health rate (EDR, IAM, email, cloud, network)
  • Detection validation freshness (% tested in last 30/60/90 days) if you track tests as tickets
  • False positive rate (FPR) by severity or detection family if analysts tag dispositions
  • Alert-to-incident conversion rate
  • Analyst touch time per alert (median) if case management tracks timers
  • Alert ingestion success rate (if measured inside SIEM/SOAR) 

Medium (effort/skill) to measure 

  • True positive rate (TPR) by severity (needs consistent “confirmed” criteria)
  • Detection confidence score by alert category (requires scoring model and discipline)
  • False negative discovery rate (incidents found outside SOC detection) requires consistent capture
  • ATT&CK technique coverage for priority threat scenarios (requires mapping and maintenance) 

Hard to measure 

  • False negative rate (true “miss rate” across the environment) beyond sampled discovery
  • “Coverage completeness” that proves you’d catch novel TTPs (requires testing/purple teaming)
  • Detection quality normalized across agencies/tools (needs standard taxonomy and normalization) 

Operational Hygiene and Backlog 

Backlog metrics reveal whether the SOC can keep up with demand. Tracking ticket aging, percentage of cases outside SLA, and after-hours workload helps leaders understand whether capacity is balanced. 

These metrics are not about productivity or individual performance. They are early indicators of burnout, under-resourcing, or excessive alert volume. 

Consistent backlog growth is a signal that something needs to change; either through automation, tuning, or staffing. 

Some examples are: 

Easy to measure

  • Open case backlog by severity (P1/P2/P3)
  • Tickets aging beyond SLA (% outside SLA by severity)
  • Case reopen rate
  • On-call coverage rate
  • Page acceptance time (P1) if paging tool logs it
  • After-hours page volume per analyst
  • Escalation bounce rate if paging/escalation is logged
  • Automation closure rate (% auto-resolved) if SOAR tracks outcomes
  • Log/source “dark time” duration if you alert on missing data 

Medium to measure

  • Time from acknowledge to triage decision (needs consistent timestamps/workflow states)
  • Analyst utilization mix (triage vs investigation vs improvement) requires time tagging
  • “Alert noise ratio” (benign vs actionable) requires tagging discipline
  • Queue volatility (spikes by day/week) tied to cause (campaign vs mis-tuning) 

Hard to measure

  • Analyst workload fairness (true cognitive load, context switching, fatigue)
  • “Operational friction” metrics (handoff delays by team) without strong ITSM integration
  • Burnout risk scoring that’s defensible (needs HR and ops data, and careful handling) 

Governance and Evidence Metrics 

In public-sector environments, governance metrics matter. Patch timelines, evidence freshness, and control validation are often driven by regulatory requirements, but they also provide insight into execution discipline. 

These metrics should be reviewed on a regular cadence and tied to accountability. When evidence is stale or remediation is overdue, it should trigger follow-up; not just reporting. 

Some examples are: 

Easy to measure

  • PIR completion rate and timeliness (if PIR is a tracked ticket/task)
  • Corrective action closure rate from PIRs (if actions are tracked)
  • Control evidence freshness (% updated within 90 days) if evidence is cataloged
  • Open policy exception count (with expiry) if exceptions are logged
  • Privileged access recertification completion rate if run through IAM/GRC
  • Orphaned privileged account count if identity is centralized 

Medium to measure

  • Regulatory control coverage status (CJIS / NIST / IRS-1075 / HIPAA) mapping and updates
  • Internet-exposed asset exception count (requires accurate external attack surface inventory)
  • Audit finding recurrence rate (requires normalized categories across audits/years) 

Hard to measure

  • “Control effectiveness” (not just evidence) tied to real risk reduction
  • Proving audit readiness continuously across decentralized agencies
  • Third-party risk “truth” (evidence quality, compensating controls validity, real remediation) 

How Public-Sector Leaders Use SOC Metrics 

The difference between mature and immature programs is not which metrics they track, it is what they do when the numbers move. 

When metrics are trending positively, effective leaders use them to validate assumptions. They confirm that investments are paying off, that alert tuning is working, and that response processes are predictable. They do not immediately raise targets; they stabilize performance first. 

When metrics degrade, mature leaders use them as entry points for investigation. A spike in MTTR prompts a review of recent incidents. A rise in backlog triggers workload analysis. A drop in detection quality leads to rule tuning or data validation. 

Metrics should lead to questions, not conclusions. 

The most important action tied to metrics is the post-incident review. For significant incidents, leaders should expect a short, clear explanation of what happened, how long each phase took, and why. Over time, these reviews create institutional knowledge and improve performance more than any dashboard. 

Misconceptions and Challenges

One of the biggest mistakes organizations make is treating metrics as proof of maturity. Clean dashboards and downward-trending charts can mask real issues if no one is willing to interrogate them. 

Another common pitfall is focusing on single numbers instead of distributions. Averages hide outliers. Tracking medians and high-percentile response times gives a more honest picture of performance. 

Inconsistent definitions also undermine credibility. If different teams interpret MTTR differently, comparisons are meaningless. Definitions must be explicit and stable. 

Finally, metrics must reflect reality. Public-sector SOCs often operate with limited hours, shared infrastructure, and constrained authority. Metrics should account for those constraints rather than pretending they do not exist. 

Conclusion: Build Systems, Not Scoreboards 

SOC metrics are not inherently valuable. Their value comes from how they are used. 

In public-sector cybersecurity, metrics should help leaders understand performance, identify friction, and make informed decisions. They should be understandable, actionable, and tied to real outcomes. 

If a metric does not influence behavior or decision-making, it does not belong in your program. 

Rethinking SOC metrics is not about tracking more numbers. It is about tracking the right ones—and having the discipline to act on what they reveal. 


Don't miss another article. Subscribe to our blog now. 

Subscribe now

 

Included Topics

  • Security Operations
Justin Fimlaid
Justin Fimlaid

Justin (he/him) is the founder and CEO of NuHarbor Security, where he continues to advance modern integrated cybersecurity services. He has over 20 years of cybersecurity experience, much of it earned while leading security efforts for multinational corporations, most recently serving as global CISO at Keurig Green Mountain Coffee. Justin serves multiple local organizations in the public interest, including his board membership at Champlain College.

Related Posts

Security Operations 5 min read
The Pros and Cons of the Student SOC: Cybersecurity’s Teaching Hospital
The Pros and Cons of the Student SOC: Cybersecurity’s Teaching Hospital
Read More
2 min read
Seven Ways to Secure Remote Access Read More
3 min read
8 Strategies for secure backups Read More

Subscribe via Email

Subscribe to our blog to get insights sent directly to your inbox.

Subscribe Here!

Latest Pwned episodes

Episode 200 - Reflections of Pwned...Until Next Time
April 03, 2024
Episode 200 - Reflections of Pwned...Until Next Time
Listen Now
Episode 199 - When a BlackCat Crosses Your Path...
March 21, 2024
Episode 199 - When a BlackCat Crosses Your Path...
Listen Now
Episode 198 - Heard it Through the Grapevine - Beyond the Beltway, 2024
March 08, 2024
Episode 198 - Heard it Through the Grapevine - Beyond the Beltway, 2024
Listen Now
NuHarbor Security logo
NuHarbor Security

553 Roosevelt Highway
Colchester, VT 05446

1.800.917.5719

  • Solutions
  • Services
  • Partners
  • Resources
  • Company
  • Contact
  • Privacy Policy
Connect
  • Twitter
  • Linkedin
  • YouTube
©2026 NuHarbor Security. All rights reserved.