Using confusion matrices to provide figures of merit for security solutions, given improved measurements of performance matched to needs at each stage of the development to production lifecycle.
what risk we are exposed to by acting or not acting on outputs of a system. • AV • DAST • DLP • EDR • IDS/IPS • MSSP • SAST • SIEM • SOC • TI • UBA • WAF
vulnerability does or does not exist. Where our test claims there’s a vulnerability, this is a positive. Anything which is not a positive given the way our tests work, is assumed to test negative. Where our test reflects the underlying truth, this is a true positive or negative.
our test (e.g. SQL injection, which can be actively exploited) FALSE POSITIVE A finding from our test (e.g. cross-site scripting, which does not exist and is not exploitable.) NEGATIVE FALSE NEGATIVE An exploitable vulnerability which is not found by our test (e.g. unpatched software not detected by monitoring tools) TRUE NEGATIVE Anything which is not exploitable, and is not detected by our tests.
more. When remediation costs are highest, we care most about false positives. When remediation costs are lowest, we care most about false negatives. When determining ground truth is difficult, we care less about truth. When determining ground truth is easy, we care more about truth.
findings can be examined and determined as false or true through testing. Different ‘sample’ systems are applicable at each lifecycle stage, but we can calculate FoMs in the same way. With vulnerability findings, we usually will not have true negatives as testing outcomes. False negatives include any positives not identified. Actual Positive Actual Negative Test Positive 62 38 Test Negative 14 0
Identifying true positives is expensive So: • False positives are cheap • False negatives may be caught in later testing We want: • High sensitivity over all
Identifying true positives is expensive So: • False positives are cheap • False negatives may be caught in later testing We want: • High sensitivity over all
Identifying true positives is middling So: • False positives require effort and create backlog • False negatives may be caught in later testing We want: • Balance of sensitivity and overall accuracy
Identifying true positives is easy So: • False positives are expensive • False negatives create unknown residual risk We want: • Negative and positive predictive value to be high
but undesirable • Inappropriate responses can have significant impact So: • False positives are potentially disastrous • False negatives mean an extended breach We want: • As few false positives and negatives as possible
not require a lot of work. 3.Which measure is important varies with lifecycle. If you’ve enjoyed this, donate to TMHC Isolation Con fundraiser for MSF! and get in touch @coffee_fueled https://linkedin.com/in/jbore [email protected]