123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188 |
- /*
- * ZeroTier One - Network Virtualization Everywhere
- * Copyright (C) 2011-2016 ZeroTier, Inc. https://www.zerotier.com/
- *
- * This program is free software: you can redistribute it and/or modify
- * it under the terms of the GNU General Public License as published by
- * the Free Software Foundation, either version 3 of the License, or
- * (at your option) any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- * GNU General Public License for more details.
- *
- * You should have received a copy of the GNU General Public License
- * along with this program. If not, see <http://www.gnu.org/licenses/>.
- */
- #include <stdio.h>
- #include <stdlib.h>
- #include <string.h>
- #include <set>
- #include <vector>
- #include "Constants.hpp"
- #include "SelfAwareness.hpp"
- #include "RuntimeEnvironment.hpp"
- #include "Node.hpp"
- #include "Topology.hpp"
- #include "Packet.hpp"
- #include "Peer.hpp"
- #include "Switch.hpp"
- // Entry timeout -- make it fairly long since this is just to prevent stale buildup
- #define ZT_SELFAWARENESS_ENTRY_TIMEOUT 3600000
- namespace ZeroTier {
- class _ResetWithinScope
- {
- public:
- _ResetWithinScope(uint64_t now,InetAddress::IpScope scope) :
- _now(now),
- _scope(scope) {}
- inline void operator()(Topology &t,const SharedPtr<Peer> &p)
- {
- if (p->resetWithinScope(_scope,_now))
- peersReset.push_back(p);
- }
- std::vector< SharedPtr<Peer> > peersReset;
- private:
- uint64_t _now;
- InetAddress::IpScope _scope;
- };
- SelfAwareness::SelfAwareness(const RuntimeEnvironment *renv) :
- RR(renv),
- _phy(32)
- {
- }
- SelfAwareness::~SelfAwareness()
- {
- }
- void SelfAwareness::iam(const Address &reporter,const InetAddress &receivedOnLocalAddress,const InetAddress &reporterPhysicalAddress,const InetAddress &myPhysicalAddress,bool trusted,uint64_t now)
- {
- const InetAddress::IpScope scope = myPhysicalAddress.ipScope();
- if ((scope != reporterPhysicalAddress.ipScope())||(scope == InetAddress::IP_SCOPE_NONE)||(scope == InetAddress::IP_SCOPE_LOOPBACK)||(scope == InetAddress::IP_SCOPE_MULTICAST))
- return;
- Mutex::Lock _l(_phy_m);
- PhySurfaceEntry &entry = _phy[PhySurfaceKey(reporter,receivedOnLocalAddress,reporterPhysicalAddress,scope)];
- if ( (trusted) && ((now - entry.ts) < ZT_SELFAWARENESS_ENTRY_TIMEOUT) && (!entry.mySurface.ipsEqual(myPhysicalAddress)) ) {
- // Changes to external surface reported by trusted peers causes path reset in this scope
- entry.mySurface = myPhysicalAddress;
- entry.ts = now;
- TRACE("physical address %s for scope %u as seen from %s(%s) differs from %s, resetting paths in scope",myPhysicalAddress.toString().c_str(),(unsigned int)scope,reporter.toString().c_str(),reporterPhysicalAddress.toString().c_str(),entry.mySurface.toString().c_str());
- // Erase all entries in this scope that were not reported from this remote address to prevent 'thrashing'
- // due to multiple reports of endpoint change.
- // Don't use 'entry' after this since hash table gets modified.
- {
- Hashtable< PhySurfaceKey,PhySurfaceEntry >::Iterator i(_phy);
- PhySurfaceKey *k = (PhySurfaceKey *)0;
- PhySurfaceEntry *e = (PhySurfaceEntry *)0;
- while (i.next(k,e)) {
- if ((k->reporterPhysicalAddress != reporterPhysicalAddress)&&(k->scope == scope))
- _phy.erase(*k);
- }
- }
- // Reset all paths within this scope
- _ResetWithinScope rset(now,(InetAddress::IpScope)scope);
- RR->topology->eachPeer<_ResetWithinScope &>(rset);
- // Send a NOP to all peers for whom we forgot a path. This will cause direct
- // links to be re-established if possible, possibly using a root server or some
- // other relay.
- for(std::vector< SharedPtr<Peer> >::const_iterator p(rset.peersReset.begin());p!=rset.peersReset.end();++p) {
- if ((*p)->activelyTransferringFrames(now)) {
- Packet outp((*p)->address(),RR->identity.address(),Packet::VERB_NOP);
- RR->sw->send(outp,true,0);
- }
- }
- } else {
- // Otherwise just update DB to use to determine external surface info
- entry.mySurface = myPhysicalAddress;
- entry.ts = now;
- }
- }
- void SelfAwareness::clean(uint64_t now)
- {
- Mutex::Lock _l(_phy_m);
- Hashtable< PhySurfaceKey,PhySurfaceEntry >::Iterator i(_phy);
- PhySurfaceKey *k = (PhySurfaceKey *)0;
- PhySurfaceEntry *e = (PhySurfaceEntry *)0;
- while (i.next(k,e)) {
- if ((now - e->ts) >= ZT_SELFAWARENESS_ENTRY_TIMEOUT)
- _phy.erase(*k);
- }
- }
- std::vector<InetAddress> SelfAwareness::getSymmetricNatPredictions()
- {
- /* This is based on ideas and strategies found here:
- * https://tools.ietf.org/html/draft-takeda-symmetric-nat-traversal-00
- *
- * In short: a great many symmetric NATs allocate ports sequentially.
- * This is common on enterprise and carrier grade NATs as well as consumer
- * devices. This code generates a list of "you might try this" addresses by
- * extrapolating likely port assignments from currently known external
- * global IPv4 surfaces. These can then be included in a PUSH_DIRECT_PATHS
- * message to another peer, causing it to possibly try these addresses and
- * bust our local symmetric NAT. It works often enough to be worth the
- * extra bit of code and does no harm in cases where it fails. */
- // Gather unique surfaces indexed by local received-on address and flag
- // us as behind a symmetric NAT if there is more than one.
- std::map< InetAddress,std::set<InetAddress> > surfaces;
- bool symmetric = false;
- {
- Mutex::Lock _l(_phy_m);
- Hashtable< PhySurfaceKey,PhySurfaceEntry >::Iterator i(_phy);
- PhySurfaceKey *k = (PhySurfaceKey *)0;
- PhySurfaceEntry *e = (PhySurfaceEntry *)0;
- while (i.next(k,e)) {
- if ((e->mySurface.ss_family == AF_INET)&&(e->mySurface.ipScope() == InetAddress::IP_SCOPE_GLOBAL)) {
- std::set<InetAddress> &s = surfaces[k->receivedOnLocalAddress];
- s.insert(e->mySurface);
- symmetric = symmetric||(s.size() > 1);
- }
- }
- }
- // If we appear to be symmetrically NATed, generate and return extrapolations
- // of those surfaces. Since PUSH_DIRECT_PATHS is sent multiple times, we
- // probabilistically generate extrapolations of anywhere from +1 to +5 to
- // increase the odds that it will work "eventually".
- if (symmetric) {
- std::vector<InetAddress> r;
- for(std::map< InetAddress,std::set<InetAddress> >::iterator si(surfaces.begin());si!=surfaces.end();++si) {
- for(std::set<InetAddress>::iterator i(si->second.begin());i!=si->second.end();++i) {
- InetAddress ipp(*i);
- unsigned int p = ipp.port() + 1 + ((unsigned int)RR->node->prng() & 3);
- if (p >= 65535)
- p -= 64510; // NATs seldom use ports <=1024 so wrap to 1025
- ipp.setPort(p);
- if ((si->second.count(ipp) == 0)&&(std::find(r.begin(),r.end(),ipp) == r.end())) {
- r.push_back(ipp);
- }
- }
- }
- return r;
- }
- return std::vector<InetAddress>();
- }
- } // namespace ZeroTier
|