The original idea was to implement DNSCurve client-side for pre-connect DNS security.
Since no stable windows DNSCurve code is out (yet?), df decided to just use dnscrypt.
So some of the data in the section below this one is now outdated, although in the future we still might switch to DNSCurve for pre-connect stuff if there ever is a stable win32/64 build of it.
The idea is fairly simple, v3 widget will use dnscrypt-proxy to connect to one of the IPs listed in dnscrypt-resolvers.csv so that if your ISP is sniffing your DNS, then your ISP will be unable to see that you're trying to resolve (for example) windows-uswest.cstorm.pw. Might be a little bit overkill, but we like overkill when anonymity/security is involved :-)
We're still using CurveDNS as the internet facing side of DeepDNS, simply because I still think that the code is pretty damn good.
Since v3 of the upcoming widget includes DNSCrypt support, I'm switching the list of public resolvers from the old .txt + DNS A record format (that was rarely updated) to the same dnscrypt-resolvers.csv format that dnscrypt-proxy uses (included in v3 widget). In github, you have to view the file in raw format to see the IPs. They'll be listed as ip:443 in there. Those IPs are the same as the DeepDNS IPs, so if you want to use DeepDNS outside of the widget (linux, openvpn GUI on win, etc.), use those.
And here's the old README, most of it still applies:
deepDNS - wtf?
df_cryptostorm has been threatened, harrassed, abused (okay, maybe just asked nicely ...-ish) for technical details on the internal structure of deepDNS that cryptostorm uses network-wide. So let's see if I can articumulate...
DNS over ddns normally happens as such:
- client gets on CS, OpenVPN pushes the exit node's ddns IP to the client, client then uses that for all DNS requests.
- client tries to resolve whatever..
- DNS request hits the internet facing CurveDNS process.
- Note: we haven't actually implemented DNSCurve at this point (well, not in the full model we're working towards - from client machine all the way through the cstorm network & back, securely end-to-end - although we make use of the network-side elements of the system already), but once we figure out a half-decent way to do so all the way from client machines forward for Windows ppl, we can implement it with this existing structure.
- Another note: The reason we prefer CurveDNS as the public facing daemon, is simply because the code appears to be solid, and we'll be using it's other features very soon.
- So since the request is regular(ish) DNS and not CurveDNS, it gets forwarded to the recursive DNS server running on 127.0.0.1:53 (on the exit node).
- 127.0.0.1:53 is the powerdns-recursor process.
- Our pdns-recursor uses this in it's config:
- That translates to:
- * send DNS requests for whatever.tld to the DNS server at 127.0.0.1:5333
- * * whatever.dns to the DNS server at 127.0.0.1:5333
- * * whatever.eth to the DNS server at 127.0.0.1:5333
- * * whatever.p2p to the DNS server at 127.0.0.1:5333
- * * whatever.onion to the DNS server at 127.0.0.1:5335
- * * whatever.i2p to the DNS server at 127.0.0.1:5399
- 127.0.0.1:5333 is DNSChain, which forwards to namecoind for true blockchain-based DNS for those TLDs (.bit, .dns, .eth, .p2p).
- 127.0.0.1:5335 is tor, more specifically, a tor instance with "DNSPort 127.0.0.1:5335" in it's torrc.
- 127.0.0.1:5399 is for .i2p, but this one is a little tricky...
- * the daemon actually listening on this port is unbound, (only because I don't know a way to add a single, static wildcard A record for a TLD to pdns-recursor, so unbound is used).
- * unbound does:
- * * local-zone: "i2p" redirect
- * * local-data: "i2p A 10.98.0.1"
- * so all *.i2p domains resolve to 10.98.0.1.
- privoxy is then setup to listen on ddns-ip:8118 and to forward all .i2p requests to the http proxy server on port 4444.
- * (the lovely config directive "forward / fuckoff" is used in privoxy to prevent ppl from using privoxy to go anywhere else, which is redundant since this is a VPN, but it does prevent 10.* access, and iptables stops anything external, plus it's funny =P).
- Then iptables comes into play:
- * iptables -t nat -A OUTPUT -d 10.98.0.1 -p tcp --dport 80 -j DNAT --to-destination ddns-ip:8118
- * iptables -t nat -A PREROUTING -d 10.98.0.1 -p tcp --dport 80 -j DNAT --to-destination ddns-ip:8118
- That means DNS requests for .i2p (planet.i2p is my goto test site btw) gets to unbound, which gets to privoxy, which goes to iptables,
which goes to the http proxy the actual i2p daemon uses, which was executed with
And for the paranoid:
- privoxy, i2p, tor, basically everything mentioned above that binds to an internet accessible IP is firewalled with iptables to prevent the internet from touching it. Also, every single instance runs under an extremely restrictive account.
Another fun fact:
The CurveDNS daemons mentioned way above are internet accessible.
It's so that once we implement DNSCurve client-side, CS members won't be susceptible to DNS hijacking (or any of the other horrible problems with the DNS protocol).
And anyone who wants to force their system to use our DNS servers before they even connect (to prevent any degree of ISP level DNS hijacks), can do so. So in the mean-time, if you worry your ISP provided DNS servers are caching, poisoning, whatever. Use CS's ddns. All are welcome :) .
They are public DNS servers, and as of a few weeks ago we now maintain & publish a full list of all the deepDNS public resolver IPs... which can also be enumerated via a lookup of "public.deepdns.net".
It's possible there will be a future split between the public deepDNS resolvers and those use for on-cstorm duties, however currently we've not yet seen a good reason to implement such a split & thus have not done so. Actually, the more we looked at it the more we felt it's important to retain as much public access to our deepDNS system of domain resolution as we can without introducing concomitant security risks or operational weak links because... well because so much about conventional DNS resolution on the internet is so horribly broken/insecure that we'd feel like dicks if we prevented folks from making use of what small improvements we're able to provide via deepDNS/public.
We'll keep adding servers & other resources to maintain ongoing unlimited public deepDNS access, even at beyond-fuckload levels of usage/load, just as long as folks keep buying cstorm network access tokens & thereby supporting stuff like public deepDNS (where "support" is syntactically commutative with "providing revenue - money - so we can pay all the bills & still have money left over to buy cookies for Graze," more or less ;-). Basically it's a "help us help you" sort of thing... but without the creepy Tom Cruise Scientology background that lurks just outside the visual frame of Jerry McGuire. Or something :-P