On 21 Oct 2016, at 18:48, Greg Owen wrote:
On 2016-10-21 19:53, John Todd wrote:
I’d like to propose an extension to PowerDNS Recursor for
mitigating
(partially) events like we had today where major authoritative
nameservers were put out of commission. This might be a particularly
foolish or error-prone method - it only took me a few minutes to
think
up. But I’d at least like to hear a discussion as to why this
isn’t a
good idea. The comment of “But this might end up giving out the
wrong
answer!” is true, but I view a wrong answer as better than no
answer.
...
If that query fails due to a SERVFAIL, then the TTL timer on this
“old” record is set back to zero and the “old” record is
provided as
a response. If an authoritative server is marked as “down” due to
repeated SERVFAIL responses (see packetcache-servfail-ttl) then the
“old” record is handed back immediately without a new query
attempt,
and the TTL timer is set back to zero to keep the answer in a state
of perpetual validity as long as....
There are security concerns to doing this. Most simply, a wrong
answer is worse than no answer if the "wrong" answer is a maliciously
sourced record.
Consider the two following cases:
1) The attacker poisons the records for a zone - either indirectly, or
via compromise of the actual authoritative servers - and then takes
the actual servers down hard, causing SERVFAIL until the owners either
fix the servers, weather the DDoS, or redirect the root NS records.
2) The attacker poisons the records directly via compromise of the
actual authoritative servers, and the owner takes the servers down
until they can be replaced with clean, secured versions.
In these two cases, the measure you're proposing would persist the
malicious entries past their expiration and for the duration of the
attack's effectiveness on the authoritative servers.
Even if your measure is triggered manually - in today's event, for
example, one says "Gosh, I know records are offline because of a Dyn
DDoS, so I know I can compensate by saving records, throw the switch!"
- let's say that someone DDoSed *ALL* of Dyn after poisoning records
for a single zone. You'd have no way of knowing - until the incident
is over and forensic analysis has hopefully caught that nuance - that
you were doing the attacker's work for them.
These attack vectors are not without precedent. So-called "Dark DDoS"
attacks have been used to distract and mislead defenders, providing a
smoke screen for other more direct attacks:
http://www.infosecurity-magazine.com/opinions/dark-ddos-growing-cyber-security/
So, in short, your proposal has the caveat that it may extend the
damage from an attack in more pernicious ways than simple denial of
service. (I'd rather not get to my bank than get to an impostor
posing as my bank!)
...
I agree it's worth putting some thought into how to increase
redundancy and flexibility to compensate for these infrastructure
attacks. For example, perhaps taking your idea but only applying it
to signed DNSSEC records which have slightly higher data integrity?
It's definitely worth exploring, but let's be careful of known and
reasonable ways attackers could take advantage of this compensation.
Thx,
gowen
--
gowen - Greg Owen - go...@swynwyr.com
CISSP, GCIA, GCFA, GWAPT
I agree with your caveats to a degree, but I can only imagine the
results being “worse” in a few edge cases of not-as-clever attacks.
In both the first and second case you describe above, would it not be
the case that a sophisticated attacker would give an unusually large TTL
to the poisoned record in order to avoid repair attempts? A TTL can be
(if I’m reading the RFC correctly) 68 years. I would expect poisoned
entries to be at least 3600 seconds (which is also the default value in
packetcache-ttl) if not significantly more, but I can’t say I’ve
ever paid attention to that number when looking at forensic data online,
and perhaps I overestimate the baseline sophistication of attackers -
but I don’t think so.
Of course, we’re trying via a number of other methods to eliminate
cache poisoning, so that’s a first step on your case #1. DNSSEC is the
best method I can see at the moment for this, so at a minimum it does
seem that this extended TTL timer would work with reasonable expected
“good” results on those records as you suggest, but I don’t think
it should be limited to just DNSSEC-secured records. In case #2, I
can’t imagine that a domain operator would have their servers offline
intentionally for longer than the TTL of the poisoned record - are there
instances where nameservers are down for several hours intentionally
after a breach? In that case, there’s at least an hour of “bad”
data infecting various recursive servers, and I imagine whatever damage
that is to happen is significantly done after an hour, and one would
hope that alternate methods (SSL, or DANE!) would provide an additional
layer of security. I am not suggesting that no additional damage will
be done during the TTL extension period, but that the cases where that
occurs are few and the benefit of operational continuity during
authoritative server outages outweighs the risk of longer-duration
failure modes.
My assertion is: given an attacker with even the most modestly
intelligent attack method, I would expect the long indefinite extension
of TTL in the case of SERVFAIL will probably result in conditions not
significantly worse for end users than if the SERVFAIL TTL extension
method were not used, even in conditions where a poisoned record is
inserted into the cache. No _new_ failure modes or results are being
introduced by this method.
Implementing the timer of course, could also be an optional method with
a default of “0”, giving the recursor operator the flexibility for
enabling/disabling given the requirements of their user community. I
can also imagine an additional counter option on this method which
limits the maximum number of times a TTL may be overridden on a record,
or an ultimate maximum TTL. There may be other more complex ways of
allowing a domain operator to signal behavior in SERVFAIL conditions for
a particular zone or record, such as TXT tags or possibly SRV record
types, but they imply lookup and caching before a SERVFAIL condition
which has a slew of unsatisfactory traffic and conditional state-keeping
issues and possible self-referential loop failures, and chances of
widespread adoption in a reasonable timeframe are fairly low though
I’m sure it would make for an interesting IETF sub-track. A TTL on
TTL? Ugh. Keeping this simple seems to be the best way forward.
I believe putting this timer override method in the resolver is the
fastest way to give local resiliency to resolvers faced with
authoritative server outages.
JT
_______________________________________________
Pdns-dev mailing list
Pdns-dev@mailman.powerdns.com
https://mailman.powerdns.com/mailman/listinfo/pdns-dev