--- Begin Message ---
Hello Peter,
We operate a blocking DNS resolver. Our false positives are either customer
escalated to our support team via portal/email/phone, or our analyst team
discover false positives using monitoring metrics such as source IP diversity,
query volume, blocked domain age, recent query trends, recent past performance
of a feed/source, etc.
Very broadly, if we were to block a popular established domain that is
requested by a high number of different source IPs it will probably correlate
with high customer impact… but that’s not a hard and fast rule. We’ve
experienced very popular false positive domains that seem to cause little to no
impact to customer business operations. Perhaps because they’re being queried
by unimportant machines rather than humans in front of browsers. Equally,
sometimes a single blocked website interrupting a senior employee’s workflow
can elicit a fast escalation from their IT department who have been put under
pressure to fix it quickly. In short, it is sensible to collect blocked domain
traffic metrics, but I wouldn’t rely on them exclusively in identifying
customer impact.
We maintain a spreadsheet of all false positives and we join each one back with
the originating threat intelligence that caused it to enter the block list.
Using this we track the performance of each source feed over time. If a
particular source feed, or a sub-section of a source feed, begins to misbehave
we can escalate with the author or remove it from the block list altogether.
So, regarding your question, we find that evaluating groups of false positives
based on their source metadata is significantly easier than playing
whack-a-mole with individual domains as they pop up. As such, we put a lot of
effort into maintaining as much metadata as possible as the data flows through
the system from source to block list to dashboard.
I don’t know how others are planning to do their heuristics work, but perhaps
the AI/ML model could be encouraged to expose some of its decision-making
parameters rather than just giving a black box yes/no answer? Maybe the
different ML stages offer scores, confidence intervals, keyword matches, or
other data flags that describe how a blocking outcome was reached. Such data
alongside a list of false positives might help an engineer determine that ‘ML
feature X’ seems to be a common problem so maybe dial it down a bit in the
model.
James
--- End Message ---
_______________________________________________
dns-operations mailing list
[email protected]
https://lists.dns-oarc.net/mailman/listinfo/dns-operations