Back to Blog

Analyzing MITRE ATT&CK Coverage in EDR Solutions

Analyzing MITRE ATT&CK Coverage in EDR Solutions

In the modern SOC, the MITRE ATT&CK framework has become the lingua franca for discussing adversary behavior. When evaluating Endpoint Detection and Response (EDR) solutions, the first thing most stakeholders look for is a "coverage heatmap." These colorful matrices, often provided by vendors, suggest a high degree of maturity by showing a dense saturation of technique coverage.

However, for the detection engineer or security architect, these heatmaps are often dangerously misleading. A "covered" technique in a vendor report does not inherently mean the EDR will alert on a real-world exploit. To truly analyze EDR efficacy, we must move beyond the superficiality of the heatmap and dissect the distinction between telemetry visibility, detection logic, and contextual correlation.

The Three Layers of Coverage

To evaluate an EDR solution rigorously, we must decompose "coverage" into three distinct, measurable layers. A failure in any one of these layers renders the coverage for a specific technique effectively zero.

1. The Telemetry Layer (Visibility)

The foundation of any detection is the raw data. If the EDR agent does not capture the specific artifact associated with a technique, detection is impossible.

For example, consider T1059.001 (PowerShell). An EDR might claim coverage for PowerShell execution. However, if the agent only captures `Process Creation` events (e.g., Sysmon Event ID 1) but does not ingest `Script Block Logging` (Event ID 4104), it is blind to obfuscated, multi-stage, or fileless PowerShell commands. The "visibility" exists at the process level, but the "telemetry" required for deep detection is absent.

2. The Logic Layer (Detection)

Once telemetry is captured, the EDR must apply logic to identify malicious patterns. This layer consists of signatures, heuristics, and machine learning models.

The complexity here lies in the "detection depth." A simple signature-based approach might detect `mimikatz.exe` by its file hash or string. However, a sophisticated attacker using T1003 (OS Credential Dumping) via a renamed, obfuscated tool or a reflective DLL injection will bypass simple logic. True coverage requires logic that identifies the behavior (e.ov., an unauthorized process accessing `lsass.exe` memory) rather than the artifact (the specific file name).

3. The Contextual Layer (Correlation)

Modern attacks are rarely single-event occurrences; they are chains of events. The highest tier of coverage is the ability to correlate disparate telemetry points into a single, high-fidelity alert.

An EDR might detect a suspicious `cmd.exe` execution (Layer 2) and a suspicious network connection to a known C2 IP (Layer 2). However, if the EDR cannot link these two events through a common process lineage or thread injection event, the analyst is left with two low-priority, disconnected alerts. Effective coverage must include the ability to reconstruct the attack graph.

An Empirical Methodology for Validation

To move from theoretical coverage to empirical validation, organizations should adopt a "Detection-as-Code" mindset, utilizing Breach and Attack Simulation (BAS) principles.

Step 1: Atomic Testing

Utilize frameworks like Atomic Red Team to execute specific, controlled sub-techniques. Instead of testing "Persistence," test "T1547.001 (Registry Run Keys)."

Step 2: The Gap Analysis

When a test fails to trigger an alert, perform a root-cause analysis (RCA) to identify which layer failed:

  • Did the event appear in the telemetry? If not, you have a visibility gap (Agent configuration or kernel-level blocking).

able to see the event but no alert was generated? You have a logic gap (Inadequate detection rules).

  • Did an alert trigger, but it was buried in noise? You have a correlation or signal-to-noise gap.

Step 3: Measuring the "Detection Window"

A critical, often overlooked metric is the time delta between the execution of the technique and the generation of the alert. In high-speed ransomware scenarios, a detection that occurs 30 minutes after the initial compromise is effectively a failure, regardless of whether it was "covered" by the MITRE matrix.

Operational Considerations and Trade-offs

Implementing high-coverage detection is not a free lunch. There is an inherent tension between Sensitivity and Specificity.

  • The False Positive Tax: Increasing the sensitivity of detection logic (e.g., alerting on any `ntsd.exe` invocation) increases the coverage of credential dumping techniques but simultaneously floods the SOC with false positives from legitimate administrative tools. This leads to alert fatigue and the eventual "silencing" of critical alerts.
  • Computational Overhead: Deep inspection (e.g., inspecting every buffer in a network stream or every script block) requires significant CPU and memory on the endpoint. In production environments, overly aggressive EDR policies can degrade user experience and break mission-critical applications.
  • The Complexity Trap: Relying on overly complex, multi-stage correlation rules makes the detection engineering pipeline fragile. If an upstream telemetry source (like a specific Sysmon event) is disabled during an OS update, a complex rule dependent on that source will silently fail.

Common Pitfalls in EDR Evaluation

  1. The "Vendor-Provided Heatmap" Fallacy: Never use a vendor's marketing collateral as a baseline for security posture. These maps are often generated based on the presence of a data source, not the efficacy of the alert.
  2. Ignoring the "Blind Spot" of Encrypted Traffic: Many EDRs claim coverage for C2 (Command and Control) but cannot inspect encrypted TLS traffic

Conclusion

As shown across "The Three Layers of Coverage", "An Empirical Methodology for Validation", "Operational Considerations and Trade-offs", a secure implementation for analyzing mitre att&ck coverage in edr solutions depends on execution discipline as much as design.

The practical hardening path is to enforce certificate lifecycle governance with strict chain/revocation checks, host hardening baselines with tamper-resistant telemetry, and behavior-chain detection across process, memory, identity, and network telemetry. This combination reduces both exploitability and attacker dwell time by forcing failures across multiple independent control layers.

Operational confidence should be measured, not assumed: track time from suspicious execution chain to host containment and policy-gate coverage and vulnerable artifact escape rate, then use those results to tune preventive policy, detection fidelity, and response runbooks on a fixed review cadence.

Related Articles

Explore related cybersecurity topics:

Recommended Next Steps

If this topic is relevant to your organisation, use one of these paths: