2015-08-20 17:27:15 +02:00
|
|
|
#!/usr/bin/env python
|
|
|
|
#
|
|
|
|
# This library is free software, distributed under the terms of
|
|
|
|
# the GNU Lesser General Public License Version 3, or any later version.
|
|
|
|
# See the COPYING file included in this archive
|
|
|
|
#
|
|
|
|
# The docstrings in this module contain epytext markup; API documentation
|
|
|
|
# may be created by processing this file with epydoc: http://epydoc.sf.net
|
2016-12-14 03:53:24 +01:00
|
|
|
import binascii
|
|
|
|
import hashlib
|
|
|
|
import operator
|
|
|
|
import struct
|
|
|
|
import time
|
2018-02-21 20:53:12 +01:00
|
|
|
import logging
|
2018-02-20 19:30:56 +01:00
|
|
|
from twisted.internet import defer, error, task
|
2016-12-14 03:53:24 +01:00
|
|
|
|
2018-02-21 20:53:12 +01:00
|
|
|
from lbrynet.core.utils import generate_id
|
2018-02-22 17:29:10 +01:00
|
|
|
from lbrynet.core.call_later_manager import CallLaterManager
|
2018-02-21 20:53:12 +01:00
|
|
|
from lbrynet.core.PeerManager import PeerManager
|
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
import constants
|
|
|
|
import routingtable
|
|
|
|
import datastore
|
|
|
|
import protocol
|
2018-02-20 19:40:02 +01:00
|
|
|
from error import TimeoutError
|
2018-02-15 23:30:14 +01:00
|
|
|
from peerfinder import DHTPeerFinder
|
2015-08-20 17:27:15 +02:00
|
|
|
from contact import Contact
|
|
|
|
from hashwatcher import HashWatcher
|
2017-10-23 19:09:46 +02:00
|
|
|
from distance import Distance
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2017-04-10 16:51:49 +02:00
|
|
|
|
2015-09-08 21:42:56 +02:00
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
def rpcmethod(func):
|
|
|
|
""" Decorator to expose Node methods as remote procedure calls
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
Apply this decorator to methods in the Node class (or a subclass) in order
|
|
|
|
to make them remotely callable via the DHT's RPC mechanism.
|
|
|
|
"""
|
|
|
|
func.rpcmethod = True
|
|
|
|
return func
|
|
|
|
|
2016-12-14 01:11:04 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
class Node(object):
|
|
|
|
""" Local node in the Kademlia network
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
This class represents a single local node in a Kademlia network; in other
|
|
|
|
words, this class encapsulates an Entangled-using application's "presence"
|
|
|
|
in a Kademlia network.
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
In Entangled, all interactions with the Kademlia network by a client
|
2016-12-14 00:08:29 +01:00
|
|
|
application is performed via this class (or a subclass).
|
2015-08-20 17:27:15 +02:00
|
|
|
"""
|
2017-03-31 19:32:43 +02:00
|
|
|
|
2018-03-27 21:12:44 +02:00
|
|
|
def __init__(self, node_id=None, udpPort=4000, dataStore=None,
|
2017-10-10 19:15:25 +02:00
|
|
|
routingTableClass=None, networkProtocol=None,
|
2018-02-20 19:30:56 +01:00
|
|
|
externalIP=None, peerPort=None, listenUDP=None,
|
|
|
|
callLater=None, resolve=None, clock=None, peer_finder=None,
|
|
|
|
peer_manager=None):
|
2015-08-20 17:27:15 +02:00
|
|
|
"""
|
|
|
|
@param dataStore: The data store to use. This must be class inheriting
|
|
|
|
from the C{DataStore} interface (or providing the
|
|
|
|
same API). How the data store manages its data
|
|
|
|
internally is up to the implementation of that data
|
|
|
|
store.
|
|
|
|
@type dataStore: entangled.kademlia.datastore.DataStore
|
|
|
|
@param routingTable: The routing table class to use. Since there exists
|
|
|
|
some ambiguity as to how the routing table should be
|
|
|
|
implemented in Kademlia, a different routing table
|
|
|
|
may be used, as long as the appropriate API is
|
|
|
|
exposed. This should be a class, not an object,
|
|
|
|
in order to allow the Node to pass an
|
|
|
|
auto-generated node ID to the routingtable object
|
2016-12-14 00:08:29 +01:00
|
|
|
upon instantiation (if necessary).
|
2015-08-20 17:27:15 +02:00
|
|
|
@type routingTable: entangled.kademlia.routingtable.RoutingTable
|
|
|
|
@param networkProtocol: The network protocol to use. This can be
|
|
|
|
overridden from the default to (for example)
|
|
|
|
change the format of the physical RPC messages
|
|
|
|
being transmitted.
|
|
|
|
@type networkProtocol: entangled.kademlia.protocol.KademliaProtocol
|
2017-10-27 20:12:52 +02:00
|
|
|
@param externalIP: the IP at which this node can be contacted
|
|
|
|
@param peerPort: the port at which this node announces it has a blob for
|
2015-08-20 17:27:15 +02:00
|
|
|
"""
|
2018-02-20 19:30:56 +01:00
|
|
|
|
|
|
|
if not listenUDP or not resolve or not callLater or not clock:
|
|
|
|
from twisted.internet import reactor
|
|
|
|
listenUDP = listenUDP or reactor.listenUDP
|
|
|
|
resolve = resolve or reactor.resolve
|
|
|
|
callLater = callLater or reactor.callLater
|
|
|
|
clock = clock or reactor
|
2018-02-22 17:29:10 +01:00
|
|
|
self.clock = clock
|
|
|
|
CallLaterManager.setup(callLater)
|
2018-02-20 19:30:56 +01:00
|
|
|
self.reactor_resolve = resolve
|
|
|
|
self.reactor_listenUDP = listenUDP
|
2018-02-22 17:29:10 +01:00
|
|
|
self.reactor_callLater = CallLaterManager.call_later
|
2017-10-10 19:15:25 +02:00
|
|
|
self.node_id = node_id or self._generateID()
|
2015-08-20 17:27:15 +02:00
|
|
|
self.port = udpPort
|
2017-03-31 19:32:43 +02:00
|
|
|
self._listeningPort = None # object implementing Twisted
|
2016-11-30 21:20:45 +01:00
|
|
|
# IListeningPort This will contain a deferred created when
|
|
|
|
# joining the network, to enable publishing/retrieving
|
|
|
|
# information from the DHT as soon as the node is part of the
|
|
|
|
# network (add callbacks to this deferred if scheduling such
|
|
|
|
# operations before the node has finished joining the network)
|
2018-02-20 19:37:02 +01:00
|
|
|
self._joinDeferred = defer.Deferred(None)
|
2017-10-10 19:10:47 +02:00
|
|
|
self.change_token_lc = task.LoopingCall(self.change_token)
|
2018-02-20 19:30:56 +01:00
|
|
|
self.change_token_lc.clock = self.clock
|
2017-10-23 19:13:06 +02:00
|
|
|
self.refresh_node_lc = task.LoopingCall(self._refreshNode)
|
2018-02-20 19:30:56 +01:00
|
|
|
self.refresh_node_lc.clock = self.clock
|
2018-03-27 21:12:44 +02:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
# Create k-buckets (for storing contacts)
|
2017-04-25 20:21:13 +02:00
|
|
|
if routingTableClass is None:
|
2018-02-20 19:30:56 +01:00
|
|
|
self._routingTable = routingtable.OptimizedTreeRoutingTable(self.node_id, self.clock.seconds)
|
2015-08-20 17:27:15 +02:00
|
|
|
else:
|
2018-02-20 19:30:56 +01:00
|
|
|
self._routingTable = routingTableClass(self.node_id, self.clock.seconds)
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
# Initialize this node's network access mechanisms
|
2017-04-25 20:21:13 +02:00
|
|
|
if networkProtocol is None:
|
2015-08-20 17:27:15 +02:00
|
|
|
self._protocol = protocol.KademliaProtocol(self)
|
|
|
|
else:
|
|
|
|
self._protocol = networkProtocol
|
|
|
|
# Initialize the data storage mechanism used by this node
|
|
|
|
self.token_secret = self._generateID()
|
|
|
|
self.old_token_secret = None
|
2017-04-25 20:21:13 +02:00
|
|
|
if dataStore is None:
|
2015-08-20 17:27:15 +02:00
|
|
|
self._dataStore = datastore.DictDataStore()
|
|
|
|
else:
|
|
|
|
self._dataStore = dataStore
|
|
|
|
# Try to restore the node's state...
|
|
|
|
if 'nodeState' in self._dataStore:
|
|
|
|
state = self._dataStore['nodeState']
|
2017-10-10 19:15:25 +02:00
|
|
|
self.node_id = state['id']
|
2015-08-20 17:27:15 +02:00
|
|
|
for contactTriple in state['closestNodes']:
|
2016-11-30 21:20:45 +01:00
|
|
|
contact = Contact(
|
|
|
|
contactTriple[0], contactTriple[1], contactTriple[2], self._protocol)
|
2015-08-20 17:27:15 +02:00
|
|
|
self._routingTable.addContact(contact)
|
|
|
|
self.externalIP = externalIP
|
2017-10-27 20:12:52 +02:00
|
|
|
self.peerPort = peerPort
|
2018-02-20 19:30:56 +01:00
|
|
|
self.hash_watcher = HashWatcher(self.clock)
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2018-02-20 19:33:59 +01:00
|
|
|
self.peer_manager = peer_manager or PeerManager()
|
|
|
|
self.peer_finder = peer_finder or DHTPeerFinder(self, self.peer_manager)
|
2018-02-15 23:30:14 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
def __del__(self):
|
2018-02-15 23:30:14 +01:00
|
|
|
log.warning("unclean shutdown of the dht node")
|
2015-08-20 17:27:15 +02:00
|
|
|
if self._listeningPort is not None:
|
|
|
|
self._listeningPort.stopListening()
|
|
|
|
|
2018-02-20 19:30:56 +01:00
|
|
|
@defer.inlineCallbacks
|
2015-08-20 17:27:15 +02:00
|
|
|
def stop(self):
|
2017-10-23 19:13:06 +02:00
|
|
|
# stop LoopingCalls:
|
|
|
|
if self.refresh_node_lc.running:
|
2018-02-20 19:30:56 +01:00
|
|
|
yield self.refresh_node_lc.stop()
|
2017-10-10 19:10:47 +02:00
|
|
|
if self.change_token_lc.running:
|
2018-02-20 19:30:56 +01:00
|
|
|
yield self.change_token_lc.stop()
|
2015-08-20 17:27:15 +02:00
|
|
|
if self._listeningPort is not None:
|
2018-02-20 19:30:56 +01:00
|
|
|
yield self._listeningPort.stopListening()
|
|
|
|
if self.hash_watcher.lc.running:
|
|
|
|
yield self.hash_watcher.stop()
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2018-02-20 19:37:02 +01:00
|
|
|
def start_listening(self):
|
2018-03-05 19:15:07 +01:00
|
|
|
if not self._listeningPort:
|
|
|
|
try:
|
|
|
|
self._listeningPort = self.reactor_listenUDP(self.port, self._protocol)
|
|
|
|
except error.CannotListenError as e:
|
|
|
|
import traceback
|
|
|
|
log.error("Couldn't bind to port %d. %s", self.port, traceback.format_exc())
|
|
|
|
raise ValueError("%s lbrynet may already be running." % str(e))
|
|
|
|
else:
|
|
|
|
log.warning("Already bound to port %d", self._listeningPort.port)
|
2018-02-20 19:37:02 +01:00
|
|
|
|
|
|
|
def bootstrap_join(self, known_node_addresses, finished_d):
|
2018-03-05 19:14:47 +01:00
|
|
|
"""
|
|
|
|
Attempt to join the dht, retry every 30 seconds if unsuccessful
|
|
|
|
:param known_node_addresses: [(str, int)] list of hostnames and ports for known dht seed nodes
|
|
|
|
:param finished_d: (defer.Deferred) called when join succeeds
|
|
|
|
"""
|
2018-02-20 19:37:02 +01:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def _resolve_seeds():
|
|
|
|
bootstrap_contacts = []
|
|
|
|
for node_address, port in known_node_addresses:
|
|
|
|
host = yield self.reactor_resolve(node_address)
|
|
|
|
# Create temporary contact information for the list of addresses of known nodes
|
|
|
|
contact = Contact(self._generateID(), host, port, self._protocol)
|
|
|
|
bootstrap_contacts.append(contact)
|
|
|
|
if not bootstrap_contacts:
|
|
|
|
if not self.hasContacts():
|
|
|
|
log.warning("No known contacts!")
|
|
|
|
else:
|
|
|
|
log.info("found contacts")
|
|
|
|
bootstrap_contacts = self.contacts
|
|
|
|
defer.returnValue(bootstrap_contacts)
|
|
|
|
|
2018-02-21 20:53:12 +01:00
|
|
|
def _rerun(closest_nodes):
|
|
|
|
if not closest_nodes:
|
|
|
|
log.info("Failed to join the dht, re-attempting in 30 seconds")
|
|
|
|
self.reactor_callLater(30, self.bootstrap_join, known_node_addresses, finished_d)
|
2018-02-20 19:37:02 +01:00
|
|
|
elif not finished_d.called:
|
2018-02-21 20:53:12 +01:00
|
|
|
finished_d.callback(closest_nodes)
|
2018-02-20 19:37:02 +01:00
|
|
|
|
|
|
|
log.info("Attempting to join the DHT network")
|
|
|
|
d = _resolve_seeds()
|
|
|
|
# Initiate the Kademlia joining sequence - perform a search for this node's own ID
|
|
|
|
d.addCallback(lambda contacts: self._iterativeFind(self.node_id, contacts))
|
|
|
|
d.addCallback(_rerun)
|
|
|
|
|
2018-02-01 10:17:17 +01:00
|
|
|
@defer.inlineCallbacks
|
2018-02-15 23:30:14 +01:00
|
|
|
def joinNetwork(self, known_node_addresses=None):
|
2018-02-01 10:17:17 +01:00
|
|
|
""" Causes the Node to attempt to join the DHT network by contacting the
|
|
|
|
known DHT nodes. This can be called multiple times if the previous attempt
|
|
|
|
has failed or if the Node has lost all the contacts.
|
|
|
|
|
2018-02-15 23:30:14 +01:00
|
|
|
@param known_node_addresses: A sequence of tuples containing IP address
|
2018-02-01 10:17:17 +01:00
|
|
|
information for existing nodes on the
|
|
|
|
Kademlia network, in the format:
|
|
|
|
C{(<ip address>, (udp port>)}
|
2018-02-15 23:30:14 +01:00
|
|
|
@type known_node_addresses: list
|
2018-02-01 10:17:17 +01:00
|
|
|
"""
|
|
|
|
|
2018-02-20 19:37:02 +01:00
|
|
|
self.start_listening()
|
2017-10-23 19:13:06 +02:00
|
|
|
# #TODO: Refresh all k-buckets further away than this node's closest neighbour
|
|
|
|
# Start refreshing k-buckets periodically, if necessary
|
2018-02-20 19:37:02 +01:00
|
|
|
self.bootstrap_join(known_node_addresses or [], self._joinDeferred)
|
2017-10-23 19:13:06 +02:00
|
|
|
yield self._joinDeferred
|
2018-02-20 19:30:56 +01:00
|
|
|
self.hash_watcher.start()
|
2018-02-15 23:30:14 +01:00
|
|
|
self.change_token_lc.start(constants.tokenSecretChangeInterval)
|
2017-10-23 19:13:06 +02:00
|
|
|
self.refresh_node_lc.start(constants.checkRefreshInterval)
|
2018-02-15 23:30:14 +01:00
|
|
|
|
2017-10-10 19:28:57 +02:00
|
|
|
@property
|
|
|
|
def contacts(self):
|
|
|
|
def _inner():
|
|
|
|
for i in range(len(self._routingTable._buckets)):
|
|
|
|
for contact in self._routingTable._buckets[i]._contacts:
|
|
|
|
yield contact
|
|
|
|
return list(_inner())
|
|
|
|
|
2018-02-01 10:21:52 +01:00
|
|
|
def hasContacts(self):
|
|
|
|
for bucket in self._routingTable._buckets:
|
|
|
|
if bucket._contacts:
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
2017-10-27 20:12:52 +02:00
|
|
|
def announceHaveBlob(self, key):
|
2018-03-27 23:35:31 +02:00
|
|
|
return self.iterativeAnnounceHaveBlob(
|
|
|
|
key, {
|
|
|
|
'port': self.peerPort,
|
|
|
|
'lbryid': self.node_id,
|
|
|
|
}
|
|
|
|
)
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2017-10-10 19:12:47 +02:00
|
|
|
@defer.inlineCallbacks
|
2015-08-20 17:27:15 +02:00
|
|
|
def getPeersForBlob(self, blob_hash):
|
2017-10-10 19:12:47 +02:00
|
|
|
result = yield self.iterativeFindValue(blob_hash)
|
|
|
|
expanded_peers = []
|
|
|
|
if result:
|
|
|
|
if blob_hash in result:
|
|
|
|
for peer in result[blob_hash]:
|
2017-10-25 02:55:24 +02:00
|
|
|
host = ".".join([str(ord(d)) for d in peer[:4]])
|
|
|
|
port, = struct.unpack('>H', peer[4:6])
|
|
|
|
if (host, port) not in expanded_peers:
|
|
|
|
expanded_peers.append((host, port))
|
2017-10-10 19:12:47 +02:00
|
|
|
defer.returnValue(expanded_peers)
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
def get_most_popular_hashes(self, num_to_return):
|
|
|
|
return self.hash_watcher.most_popular_hashes(num_to_return)
|
|
|
|
|
2018-03-27 21:04:47 +02:00
|
|
|
@defer.inlineCallbacks
|
2017-05-25 20:01:39 +02:00
|
|
|
def iterativeAnnounceHaveBlob(self, blob_hash, value):
|
2015-08-20 17:27:15 +02:00
|
|
|
known_nodes = {}
|
2018-03-27 21:04:47 +02:00
|
|
|
contacts = yield self.iterativeFindNode(blob_hash)
|
|
|
|
# store locally if we're the closest node and there are less than k contacts to try storing to
|
|
|
|
if self.externalIP is not None and contacts and len(contacts) < constants.k:
|
|
|
|
is_closer = Distance(blob_hash).is_closer(self.node_id, contacts[-1].id)
|
|
|
|
if is_closer:
|
|
|
|
contacts.pop()
|
|
|
|
yield self.store(blob_hash, value, originalPublisherID=self.node_id,
|
|
|
|
self_store=True)
|
|
|
|
elif self.externalIP is not None:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
raise Exception("Cannot determine external IP: %s" % self.externalIP)
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2018-03-27 21:04:47 +02:00
|
|
|
contacted = []
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2017-10-23 21:36:50 +02:00
|
|
|
@defer.inlineCallbacks
|
2018-03-27 21:04:47 +02:00
|
|
|
def announce_to_contact(contact):
|
|
|
|
known_nodes[contact.id] = contact
|
|
|
|
try:
|
|
|
|
responseMsg, originAddress = yield contact.findValue(blob_hash, rawResponse=True)
|
|
|
|
if responseMsg.nodeID != contact.id:
|
|
|
|
raise Exception("node id mismatch")
|
|
|
|
value['token'] = responseMsg.response['token']
|
|
|
|
res = yield contact.store(blob_hash, value)
|
|
|
|
if res != "OK":
|
|
|
|
raise ValueError(res)
|
|
|
|
contacted.append(contact)
|
|
|
|
log.debug("Stored %s to %s (%s)", blob_hash.encode('hex'), contact.id.encode('hex'), originAddress[0])
|
|
|
|
except protocol.TimeoutError:
|
|
|
|
log.debug("Timeout while storing blob_hash %s at %s",
|
|
|
|
blob_hash.encode('hex')[:16], contact.id.encode('hex'))
|
|
|
|
except ValueError as err:
|
|
|
|
log.error("Unexpected response: %s" % err.message)
|
|
|
|
except Exception as err:
|
|
|
|
log.error("Unexpected error while storing blob_hash %s at %s: %s",
|
|
|
|
binascii.hexlify(blob_hash), contact, err)
|
|
|
|
|
|
|
|
dl = []
|
|
|
|
for c in contacts:
|
|
|
|
dl.append(announce_to_contact(c))
|
|
|
|
|
|
|
|
yield defer.DeferredList(dl)
|
|
|
|
|
|
|
|
log.debug("Stored %s to %i of %i attempted peers", blob_hash.encode('hex')[:16],
|
|
|
|
len(contacted), len(contacts))
|
|
|
|
|
|
|
|
contacted_node_ids = [c.id.encode('hex') for c in contacted]
|
|
|
|
defer.returnValue(contacted_node_ids)
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
def change_token(self):
|
|
|
|
self.old_token_secret = self.token_secret
|
|
|
|
self.token_secret = self._generateID()
|
|
|
|
|
|
|
|
def make_token(self, compact_ip):
|
|
|
|
h = hashlib.new('sha384')
|
|
|
|
h.update(self.token_secret + compact_ip)
|
|
|
|
return h.digest()
|
|
|
|
|
|
|
|
def verify_token(self, token, compact_ip):
|
|
|
|
h = hashlib.new('sha384')
|
|
|
|
h.update(self.token_secret + compact_ip)
|
|
|
|
if not token == h.digest():
|
|
|
|
h = hashlib.new('sha384')
|
|
|
|
h.update(self.old_token_secret + compact_ip)
|
|
|
|
if not token == h.digest():
|
|
|
|
return False
|
|
|
|
return True
|
|
|
|
|
|
|
|
def iterativeFindNode(self, key):
|
|
|
|
""" The basic Kademlia node lookup operation
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
Call this to find a remote node in the P2P overlay network.
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@param key: the n-bit key (i.e. the node or value ID) to search for
|
|
|
|
@type key: str
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@return: This immediately returns a deferred object, which will return
|
|
|
|
a list of k "closest" contacts (C{kademlia.contact.Contact}
|
|
|
|
objects) to the specified key as soon as the operation is
|
|
|
|
finished.
|
|
|
|
@rtype: twisted.internet.defer.Deferred
|
|
|
|
"""
|
|
|
|
return self._iterativeFind(key)
|
|
|
|
|
2017-05-25 20:01:39 +02:00
|
|
|
@defer.inlineCallbacks
|
2015-08-20 17:27:15 +02:00
|
|
|
def iterativeFindValue(self, key):
|
|
|
|
""" The Kademlia search operation (deterministic)
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
Call this to retrieve data from the DHT.
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@param key: the n-bit key (i.e. the value ID) to search for
|
|
|
|
@type key: str
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@return: This immediately returns a deferred object, which will return
|
|
|
|
either one of two things:
|
|
|
|
- If the value was found, it will return a Python
|
|
|
|
dictionary containing the searched-for key (the C{key}
|
|
|
|
parameter passed to this method), and its associated
|
|
|
|
value, in the format:
|
|
|
|
C{<str>key: <str>data_value}
|
|
|
|
- If the value was not found, it will return a list of k
|
|
|
|
"closest" contacts (C{kademlia.contact.Contact} objects)
|
|
|
|
to the specified key
|
|
|
|
@rtype: twisted.internet.defer.Deferred
|
|
|
|
"""
|
|
|
|
|
|
|
|
# Execute the search
|
2017-05-25 20:01:39 +02:00
|
|
|
iterative_find_result = yield self._iterativeFind(key, rpc='findValue')
|
2018-02-20 19:39:14 +01:00
|
|
|
if isinstance(iterative_find_result, dict):
|
|
|
|
# We have found the value; now see who was the closest contact without it...
|
|
|
|
# ...and store the key/value pair
|
|
|
|
defer.returnValue(iterative_find_result)
|
|
|
|
else:
|
|
|
|
# The value wasn't found, but a list of contacts was returned
|
|
|
|
# Now, see if we have the value (it might seem wasteful to search on the network
|
|
|
|
# first, but it ensures that all values are properly propagated through the
|
|
|
|
# network
|
|
|
|
if self._dataStore.hasPeersForBlob(key):
|
|
|
|
# Ok, we have the value locally, so use that
|
|
|
|
# Send this value to the closest node without it
|
|
|
|
peers = self._dataStore.getPeersForBlob(key)
|
|
|
|
defer.returnValue({key: peers})
|
|
|
|
else:
|
|
|
|
# Ok, value does not exist in DHT at all
|
|
|
|
defer.returnValue(iterative_find_result)
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
def addContact(self, contact):
|
|
|
|
""" Add/update the given contact; simple wrapper for the same method
|
|
|
|
in this object's RoutingTable object
|
|
|
|
|
|
|
|
@param contact: The contact to add to this node's k-buckets
|
|
|
|
@type contact: kademlia.contact.Contact
|
|
|
|
"""
|
|
|
|
self._routingTable.addContact(contact)
|
|
|
|
|
|
|
|
def removeContact(self, contactID):
|
|
|
|
""" Remove the contact with the specified node ID from this node's
|
|
|
|
table of known nodes. This is a simple wrapper for the same method
|
|
|
|
in this object's RoutingTable object
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@param contactID: The node ID of the contact to remove
|
|
|
|
@type contactID: str
|
|
|
|
"""
|
|
|
|
self._routingTable.removeContact(contactID)
|
|
|
|
|
|
|
|
def findContact(self, contactID):
|
|
|
|
""" Find a entangled.kademlia.contact.Contact object for the specified
|
|
|
|
cotact ID
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@param contactID: The contact ID of the required Contact object
|
|
|
|
@type contactID: str
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@return: Contact object of remote node with the specified node ID,
|
|
|
|
or None if the contact was not found
|
|
|
|
@rtype: twisted.internet.defer.Deferred
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
contact = self._routingTable.getContact(contactID)
|
|
|
|
df = defer.Deferred()
|
|
|
|
df.callback(contact)
|
|
|
|
except ValueError:
|
|
|
|
def parseResults(nodes):
|
|
|
|
if contactID in nodes:
|
|
|
|
contact = nodes[nodes.index(contactID)]
|
|
|
|
return contact
|
|
|
|
else:
|
|
|
|
return None
|
2017-03-31 19:32:43 +02:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
df = self.iterativeFindNode(contactID)
|
|
|
|
df.addCallback(parseResults)
|
|
|
|
return df
|
|
|
|
|
|
|
|
@rpcmethod
|
|
|
|
def ping(self):
|
|
|
|
""" Used to verify contact between two Kademlia nodes
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@rtype: str
|
|
|
|
"""
|
|
|
|
return 'pong'
|
|
|
|
|
|
|
|
@rpcmethod
|
|
|
|
def store(self, key, value, originalPublisherID=None, self_store=False, **kwargs):
|
|
|
|
""" Store the received data in this node's local hash table
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@param key: The hashtable key of the data
|
|
|
|
@type key: str
|
|
|
|
@param value: The actual data (the value associated with C{key})
|
|
|
|
@type value: str
|
|
|
|
@param originalPublisherID: The node ID of the node that is the
|
|
|
|
B{original} publisher of the data
|
|
|
|
@type originalPublisherID: str
|
|
|
|
@param age: The relative age of the data (time in seconds since it was
|
|
|
|
originally published). Note that the original publish time
|
|
|
|
isn't actually given, to compensate for clock skew between
|
|
|
|
different nodes.
|
|
|
|
@type age: int
|
|
|
|
|
|
|
|
@rtype: str
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@todo: Since the data (value) may be large, passing it around as a buffer
|
|
|
|
(which is the case currently) might not be a good idea... will have
|
|
|
|
to fix this (perhaps use a stream from the Protocol class?)
|
|
|
|
"""
|
|
|
|
# Get the sender's ID (if any)
|
2017-04-25 20:21:13 +02:00
|
|
|
if originalPublisherID is None:
|
2015-08-20 17:27:15 +02:00
|
|
|
if '_rpcNodeID' in kwargs:
|
|
|
|
originalPublisherID = kwargs['_rpcNodeID']
|
|
|
|
else:
|
|
|
|
raise TypeError, 'No NodeID given. Therefore we can\'t store this node'
|
|
|
|
|
2016-01-26 22:07:33 +01:00
|
|
|
if self_store is True and self.externalIP:
|
2017-10-10 19:15:25 +02:00
|
|
|
contact = Contact(self.node_id, self.externalIP, self.port, None, None)
|
2015-08-20 17:27:15 +02:00
|
|
|
compact_ip = contact.compact_ip()
|
|
|
|
elif '_rpcNodeContact' in kwargs:
|
|
|
|
contact = kwargs['_rpcNodeContact']
|
|
|
|
compact_ip = contact.compact_ip()
|
|
|
|
else:
|
2017-10-10 19:29:54 +02:00
|
|
|
raise TypeError, 'No contact info available'
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2018-03-27 20:59:39 +02:00
|
|
|
if not self_store:
|
|
|
|
if 'token' not in value:
|
|
|
|
raise ValueError("Missing token")
|
|
|
|
if not self.verify_token(value['token'], compact_ip):
|
|
|
|
raise ValueError("Invalid token")
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
if 'port' in value:
|
|
|
|
port = int(value['port'])
|
|
|
|
if 0 <= port <= 65536:
|
|
|
|
compact_port = str(struct.pack('>H', port))
|
|
|
|
else:
|
2018-03-27 20:59:39 +02:00
|
|
|
raise TypeError('Invalid port')
|
2015-08-20 17:27:15 +02:00
|
|
|
else:
|
2018-03-27 20:59:39 +02:00
|
|
|
raise TypeError('No port available')
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
if 'lbryid' in value:
|
2017-10-12 05:23:07 +02:00
|
|
|
if len(value['lbryid']) != constants.key_bits / 8:
|
|
|
|
raise ValueError('Invalid lbryid (%i bytes): %s' % (len(value['lbryid']),
|
|
|
|
value['lbryid'].encode('hex')))
|
2015-08-20 17:27:15 +02:00
|
|
|
else:
|
|
|
|
compact_address = compact_ip + compact_port + value['lbryid']
|
|
|
|
else:
|
2018-03-27 20:59:39 +02:00
|
|
|
raise TypeError('No lbryid given')
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
now = int(time.time())
|
2017-03-31 19:32:43 +02:00
|
|
|
originallyPublished = now # - age
|
2017-05-25 20:01:39 +02:00
|
|
|
self._dataStore.addPeerToBlob(key, compact_address, now, originallyPublished,
|
|
|
|
originalPublisherID)
|
2015-08-20 17:27:15 +02:00
|
|
|
return 'OK'
|
|
|
|
|
|
|
|
@rpcmethod
|
|
|
|
def findNode(self, key, **kwargs):
|
|
|
|
""" Finds a number of known nodes closest to the node/value with the
|
|
|
|
specified key.
|
2016-11-04 21:09:40 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@param key: the n-bit key (i.e. the node or value ID) to search for
|
|
|
|
@type key: str
|
|
|
|
|
|
|
|
@return: A list of contact triples closest to the specified key.
|
|
|
|
This method will return C{k} (or C{count}, if specified)
|
|
|
|
contacts if at all possible; it will only return fewer if the
|
|
|
|
node is returning all of the contacts that it knows of.
|
|
|
|
@rtype: list
|
|
|
|
"""
|
2017-10-10 19:29:29 +02:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
# Get the sender's ID (if any)
|
|
|
|
if '_rpcNodeID' in kwargs:
|
2017-04-10 16:51:49 +02:00
|
|
|
rpc_sender_id = kwargs['_rpcNodeID']
|
2015-08-20 17:27:15 +02:00
|
|
|
else:
|
2017-04-10 16:51:49 +02:00
|
|
|
rpc_sender_id = None
|
|
|
|
contacts = self._routingTable.findCloseNodes(key, constants.k, rpc_sender_id)
|
|
|
|
contact_triples = []
|
2015-08-20 17:27:15 +02:00
|
|
|
for contact in contacts:
|
2017-04-10 16:51:49 +02:00
|
|
|
contact_triples.append((contact.id, contact.address, contact.port))
|
|
|
|
return contact_triples
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
@rpcmethod
|
|
|
|
def findValue(self, key, **kwargs):
|
|
|
|
""" Return the value associated with the specified key if present in
|
|
|
|
this node's data, otherwise execute FIND_NODE for the key
|
2016-11-04 21:09:40 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@param key: The hashtable key of the data to return
|
|
|
|
@type key: str
|
2016-11-04 21:09:40 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@return: A dictionary containing the requested key/value pair,
|
|
|
|
or a list of contact triples closest to the requested key.
|
|
|
|
@rtype: dict or list
|
|
|
|
"""
|
2017-05-25 20:01:39 +02:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
if self._dataStore.hasPeersForBlob(key):
|
|
|
|
rval = {key: self._dataStore.getPeersForBlob(key)}
|
|
|
|
else:
|
2017-04-10 16:51:49 +02:00
|
|
|
contact_triples = self.findNode(key, **kwargs)
|
|
|
|
rval = {'contacts': contact_triples}
|
2015-08-20 17:27:15 +02:00
|
|
|
if '_rpcNodeContact' in kwargs:
|
|
|
|
contact = kwargs['_rpcNodeContact']
|
|
|
|
compact_ip = contact.compact_ip()
|
|
|
|
rval['token'] = self.make_token(compact_ip)
|
2017-05-25 20:01:39 +02:00
|
|
|
self.hash_watcher.add_requested_hash(key, contact)
|
2015-08-20 17:27:15 +02:00
|
|
|
return rval
|
|
|
|
|
|
|
|
def _generateID(self):
|
|
|
|
""" Generates an n-bit pseudo-random identifier
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@return: A globally unique n-bit pseudo-random identifier
|
|
|
|
@rtype: str
|
|
|
|
"""
|
2017-04-10 16:51:49 +02:00
|
|
|
return generate_id()
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2017-05-25 20:01:39 +02:00
|
|
|
@defer.inlineCallbacks
|
2015-08-20 17:27:15 +02:00
|
|
|
def _iterativeFind(self, key, startupShortlist=None, rpc='findNode'):
|
|
|
|
""" The basic Kademlia iterative lookup operation (for nodes/values)
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
This builds a list of k "closest" contacts through iterative use of
|
|
|
|
the "FIND_NODE" RPC, or if C{findValue} is set to C{True}, using the
|
|
|
|
"FIND_VALUE" RPC, in which case the value (if found) may be returned
|
|
|
|
instead of a list of contacts
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@param key: the n-bit key (i.e. the node or value ID) to search for
|
|
|
|
@type key: str
|
|
|
|
@param startupShortlist: A list of contacts to use as the starting
|
|
|
|
shortlist for this search; this is normally
|
|
|
|
only used when the node joins the network
|
|
|
|
@type startupShortlist: list
|
|
|
|
@param rpc: The name of the RPC to issue to remote nodes during the
|
|
|
|
Kademlia lookup operation (e.g. this sets whether this
|
|
|
|
algorithm should search for a data value (if
|
|
|
|
rpc='findValue') or not. It can thus be used to perform
|
|
|
|
other operations that piggy-back on the basic Kademlia
|
|
|
|
lookup operation (Entangled's "delete" RPC, for instance).
|
|
|
|
@type rpc: str
|
2016-12-14 00:08:29 +01:00
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
@return: If C{findValue} is C{True}, the algorithm will stop as soon
|
|
|
|
as a data value for C{key} is found, and return a dictionary
|
|
|
|
containing the key and the found value. Otherwise, it will
|
|
|
|
return a list of the k closest nodes to the specified key
|
|
|
|
@rtype: twisted.internet.defer.Deferred
|
|
|
|
"""
|
2017-04-10 16:51:49 +02:00
|
|
|
findValue = rpc != 'findNode'
|
|
|
|
|
2017-04-25 20:21:13 +02:00
|
|
|
if startupShortlist is None:
|
2017-10-10 19:20:19 +02:00
|
|
|
shortlist = self._routingTable.findCloseNodes(key, constants.k)
|
2017-10-10 19:15:25 +02:00
|
|
|
if key != self.node_id:
|
2015-08-20 17:27:15 +02:00
|
|
|
# Update the "last accessed" timestamp for the appropriate k-bucket
|
|
|
|
self._routingTable.touchKBucket(key)
|
|
|
|
if len(shortlist) == 0:
|
2017-10-10 19:20:19 +02:00
|
|
|
log.warning("This node doesnt know any other nodes")
|
2015-08-20 17:27:15 +02:00
|
|
|
# This node doesn't know of any other nodes
|
|
|
|
fakeDf = defer.Deferred()
|
|
|
|
fakeDf.callback([])
|
2017-05-25 20:01:39 +02:00
|
|
|
result = yield fakeDf
|
|
|
|
defer.returnValue(result)
|
2015-08-20 17:27:15 +02:00
|
|
|
else:
|
|
|
|
# This is used during the bootstrap process; node ID's are most probably fake
|
|
|
|
shortlist = startupShortlist
|
|
|
|
|
|
|
|
outerDf = defer.Deferred()
|
2016-12-14 01:11:04 +01:00
|
|
|
|
|
|
|
helper = _IterativeFindHelper(self, outerDf, shortlist, key, findValue, rpc)
|
2015-08-20 17:27:15 +02:00
|
|
|
# Start the iterations
|
2016-12-14 01:11:04 +01:00
|
|
|
helper.searchIteration()
|
2017-05-25 20:01:39 +02:00
|
|
|
result = yield outerDf
|
|
|
|
defer.returnValue(result)
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2017-10-23 19:13:06 +02:00
|
|
|
@defer.inlineCallbacks
|
2015-08-20 17:27:15 +02:00
|
|
|
def _refreshNode(self):
|
|
|
|
""" Periodically called to perform k-bucket refreshes and data
|
|
|
|
replication/republishing as necessary """
|
2017-05-25 20:01:39 +02:00
|
|
|
|
2017-10-23 19:13:06 +02:00
|
|
|
yield self._refreshRoutingTable()
|
|
|
|
self._dataStore.removeExpiredPeers()
|
|
|
|
defer.returnValue(None)
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2017-10-23 21:36:50 +02:00
|
|
|
@defer.inlineCallbacks
|
2015-08-20 17:27:15 +02:00
|
|
|
def _refreshRoutingTable(self):
|
|
|
|
nodeIDs = self._routingTable.getRefreshList(0, False)
|
2017-10-23 21:36:50 +02:00
|
|
|
while nodeIDs:
|
|
|
|
searchID = nodeIDs.pop()
|
|
|
|
yield self.iterativeFindNode(searchID)
|
|
|
|
defer.returnValue(None)
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
|
2016-12-14 01:11:04 +01:00
|
|
|
# This was originally a set of nested methods in _iterativeFind
|
|
|
|
# but they have been moved into this helper class in-order to
|
|
|
|
# have better scoping and readability
|
|
|
|
class _IterativeFindHelper(object):
|
|
|
|
# TODO: use polymorphism to search for a value or node
|
|
|
|
# instead of using a find_value flag
|
|
|
|
def __init__(self, node, outer_d, shortlist, key, find_value, rpc):
|
|
|
|
self.node = node
|
|
|
|
self.outer_d = outer_d
|
|
|
|
self.shortlist = shortlist
|
|
|
|
self.key = key
|
|
|
|
self.find_value = find_value
|
|
|
|
self.rpc = rpc
|
2016-12-14 03:53:24 +01:00
|
|
|
# all distance operations in this class only care about the distance
|
|
|
|
# to self.key, so this makes it easier to calculate those
|
|
|
|
self.distance = Distance(key)
|
2016-12-14 01:11:04 +01:00
|
|
|
# List of active queries; len() indicates number of active probes
|
|
|
|
#
|
|
|
|
# n.b: using lists for these variables, because Python doesn't
|
|
|
|
# allow binding a new value to a name in an enclosing
|
|
|
|
# (non-global) scope
|
|
|
|
self.active_probes = []
|
|
|
|
# List of contact IDs that have already been queried
|
|
|
|
self.already_contacted = []
|
|
|
|
# Probes that were active during the previous iteration
|
|
|
|
# A list of found and known-to-be-active remote nodes
|
|
|
|
self.active_contacts = []
|
|
|
|
# This should only contain one entry; the next scheduled iteration call
|
|
|
|
self.pending_iteration_calls = []
|
|
|
|
self.prev_closest_node = [None]
|
|
|
|
self.find_value_result = {}
|
|
|
|
self.slow_node_count = [0]
|
|
|
|
|
|
|
|
def extendShortlist(self, responseTuple):
|
|
|
|
""" @type responseMsg: kademlia.msgtypes.ResponseMessage """
|
|
|
|
# The "raw response" tuple contains the response message,
|
|
|
|
# and the originating address info
|
|
|
|
responseMsg = responseTuple[0]
|
2017-03-31 19:32:43 +02:00
|
|
|
originAddress = responseTuple[1] # tuple: (ip adress, udp port)
|
2016-12-14 01:11:04 +01:00
|
|
|
# Make sure the responding node is valid, and abort the operation if it isn't
|
2017-10-10 19:15:25 +02:00
|
|
|
if responseMsg.nodeID in self.active_contacts or responseMsg.nodeID == self.node.node_id:
|
2016-12-14 01:11:04 +01:00
|
|
|
return responseMsg.nodeID
|
|
|
|
|
|
|
|
# Mark this node as active
|
|
|
|
aContact = self._getActiveContact(responseMsg, originAddress)
|
|
|
|
self.active_contacts.append(aContact)
|
|
|
|
|
|
|
|
# This makes sure "bootstrap"-nodes with "fake" IDs don't get queried twice
|
|
|
|
if responseMsg.nodeID not in self.already_contacted:
|
|
|
|
self.already_contacted.append(responseMsg.nodeID)
|
2017-05-25 20:01:39 +02:00
|
|
|
|
2016-12-14 01:11:04 +01:00
|
|
|
# Now grow extend the (unverified) shortlist with the returned contacts
|
|
|
|
result = responseMsg.response
|
2017-03-31 19:32:43 +02:00
|
|
|
# TODO: some validation on the result (for guarding against attacks)
|
2016-12-14 01:11:04 +01:00
|
|
|
# If we are looking for a value, first see if this result is the value
|
|
|
|
# we are looking for before treating it as a list of contact triples
|
|
|
|
if self.find_value is True and self.key in result and not 'contacts' in result:
|
|
|
|
# We have found the value
|
|
|
|
self.find_value_result[self.key] = result[self.key]
|
|
|
|
else:
|
|
|
|
if self.find_value is True:
|
|
|
|
self._setClosestNodeValue(responseMsg, aContact)
|
|
|
|
self._keepSearching(result)
|
|
|
|
return responseMsg.nodeID
|
|
|
|
|
|
|
|
def _getActiveContact(self, responseMsg, originAddress):
|
|
|
|
if responseMsg.nodeID in self.shortlist:
|
|
|
|
# Get the contact information from the shortlist...
|
|
|
|
return self.shortlist[self.shortlist.index(responseMsg.nodeID)]
|
|
|
|
else:
|
|
|
|
# If it's not in the shortlist; we probably used a fake ID to reach it
|
|
|
|
# - reconstruct the contact, using the real node ID this time
|
|
|
|
return Contact(
|
|
|
|
responseMsg.nodeID, originAddress[0], originAddress[1], self.node._protocol)
|
|
|
|
|
|
|
|
def _keepSearching(self, result):
|
|
|
|
contactTriples = self._getContactTriples(result)
|
|
|
|
for contactTriple in contactTriples:
|
|
|
|
self._addIfValid(contactTriple)
|
|
|
|
|
|
|
|
def _getContactTriples(self, result):
|
|
|
|
if self.find_value is True:
|
|
|
|
return result['contacts']
|
|
|
|
else:
|
|
|
|
return result
|
|
|
|
|
|
|
|
def _setClosestNodeValue(self, responseMsg, aContact):
|
|
|
|
# We are looking for a value, and the remote node didn't have it
|
|
|
|
# - mark it as the closest "empty" node, if it is
|
|
|
|
if 'closestNodeNoValue' in self.find_value_result:
|
|
|
|
if self._is_closer(responseMsg):
|
|
|
|
self.find_value_result['closestNodeNoValue'] = aContact
|
|
|
|
else:
|
|
|
|
self.find_value_result['closestNodeNoValue'] = aContact
|
|
|
|
|
|
|
|
def _is_closer(self, responseMsg):
|
2016-12-14 03:53:24 +01:00
|
|
|
return self.distance.is_closer(responseMsg.nodeID, self.active_contacts[0].id)
|
2016-12-14 01:11:04 +01:00
|
|
|
|
|
|
|
def _addIfValid(self, contactTriple):
|
|
|
|
if isinstance(contactTriple, (list, tuple)) and len(contactTriple) == 3:
|
|
|
|
testContact = Contact(
|
|
|
|
contactTriple[0], contactTriple[1], contactTriple[2], self.node._protocol)
|
|
|
|
if testContact not in self.shortlist:
|
|
|
|
self.shortlist.append(testContact)
|
|
|
|
|
2017-10-24 20:55:20 +02:00
|
|
|
def removeFromShortlist(self, failure, deadContactID):
|
2016-12-14 01:11:04 +01:00
|
|
|
""" @type failure: twisted.python.failure.Failure """
|
2018-02-20 19:40:02 +01:00
|
|
|
failure.trap(TimeoutError, defer.CancelledError, TypeError)
|
2017-10-24 20:55:20 +02:00
|
|
|
if len(deadContactID) != constants.key_bits / 8:
|
|
|
|
raise ValueError("invalid lbry id")
|
2016-12-14 01:11:04 +01:00
|
|
|
if deadContactID in self.shortlist:
|
|
|
|
self.shortlist.remove(deadContactID)
|
|
|
|
return deadContactID
|
|
|
|
|
|
|
|
def cancelActiveProbe(self, contactID):
|
|
|
|
self.active_probes.pop()
|
2017-03-31 19:32:43 +02:00
|
|
|
if len(self.active_probes) <= constants.alpha / 2 and len(self.pending_iteration_calls):
|
2016-12-14 01:11:04 +01:00
|
|
|
# Force the iteration
|
|
|
|
self.pending_iteration_calls[0].cancel()
|
|
|
|
del self.pending_iteration_calls[0]
|
|
|
|
self.searchIteration()
|
|
|
|
|
2016-12-14 03:53:24 +01:00
|
|
|
def sortByDistance(self, contact_list):
|
|
|
|
"""Sort the list of contacts in order by distance from key"""
|
|
|
|
ExpensiveSort(contact_list, self.distance.to_contact).sort()
|
|
|
|
|
2016-12-14 01:11:04 +01:00
|
|
|
# Send parallel, asynchronous FIND_NODE RPCs to the shortlist of contacts
|
|
|
|
def searchIteration(self):
|
|
|
|
self.slow_node_count[0] = len(self.active_probes)
|
|
|
|
# Sort the discovered active nodes from closest to furthest
|
2016-12-14 03:53:24 +01:00
|
|
|
self.sortByDistance(self.active_contacts)
|
2016-12-14 01:11:04 +01:00
|
|
|
# This makes sure a returning probe doesn't force calling this function by mistake
|
|
|
|
while len(self.pending_iteration_calls):
|
|
|
|
del self.pending_iteration_calls[0]
|
|
|
|
# See if should continue the search
|
|
|
|
if self.key in self.find_value_result:
|
|
|
|
self.outer_d.callback(self.find_value_result)
|
|
|
|
return
|
2017-04-25 20:21:13 +02:00
|
|
|
elif len(self.active_contacts) and self.find_value is False:
|
2016-12-14 01:11:04 +01:00
|
|
|
if self._is_all_done():
|
|
|
|
# TODO: Re-send the FIND_NODEs to all of the k closest nodes not already queried
|
|
|
|
#
|
|
|
|
# Ok, we're done; either we have accumulated k active
|
|
|
|
# contacts or no improvement in closestNode has been
|
|
|
|
# noted
|
|
|
|
self.outer_d.callback(self.active_contacts)
|
|
|
|
return
|
2017-05-25 20:01:39 +02:00
|
|
|
|
2016-12-14 01:11:04 +01:00
|
|
|
# The search continues...
|
|
|
|
if len(self.active_contacts):
|
|
|
|
self.prev_closest_node[0] = self.active_contacts[0]
|
|
|
|
contactedNow = 0
|
2016-12-14 03:53:24 +01:00
|
|
|
self.sortByDistance(self.shortlist)
|
2016-12-14 01:11:04 +01:00
|
|
|
# Store the current shortList length before contacting other nodes
|
|
|
|
prevShortlistLength = len(self.shortlist)
|
|
|
|
for contact in self.shortlist:
|
|
|
|
if contact.id not in self.already_contacted:
|
|
|
|
self._probeContact(contact)
|
|
|
|
contactedNow += 1
|
|
|
|
if contactedNow == constants.alpha:
|
|
|
|
break
|
|
|
|
if self._should_lookup_active_calls():
|
|
|
|
# Schedule the next iteration if there are any active
|
|
|
|
# calls (Kademlia uses loose parallelism)
|
2018-02-22 17:29:10 +01:00
|
|
|
call, _ = self.node.reactor_callLater(constants.iterativeLookupDelay, self.searchIteration)
|
2016-12-14 01:11:04 +01:00
|
|
|
self.pending_iteration_calls.append(call)
|
|
|
|
# Check for a quick contact response that made an update to the shortList
|
|
|
|
elif prevShortlistLength < len(self.shortlist):
|
|
|
|
# Ensure that the closest contacts are taken from the updated shortList
|
|
|
|
self.searchIteration()
|
|
|
|
else:
|
|
|
|
# If no probes were sent, there will not be any improvement, so we're done
|
|
|
|
self.outer_d.callback(self.active_contacts)
|
|
|
|
|
|
|
|
def _probeContact(self, contact):
|
|
|
|
self.active_probes.append(contact.id)
|
|
|
|
rpcMethod = getattr(contact, self.rpc)
|
|
|
|
df = rpcMethod(self.key, rawResponse=True)
|
|
|
|
df.addCallback(self.extendShortlist)
|
2017-10-24 20:55:20 +02:00
|
|
|
df.addErrback(self.removeFromShortlist, contact.id)
|
2016-12-14 01:11:04 +01:00
|
|
|
df.addCallback(self.cancelActiveProbe)
|
2017-05-25 20:01:39 +02:00
|
|
|
df.addErrback(lambda _: log.exception('Failed to contact %s', contact))
|
2016-12-14 01:11:04 +01:00
|
|
|
self.already_contacted.append(contact.id)
|
|
|
|
|
|
|
|
def _should_lookup_active_calls(self):
|
|
|
|
return (
|
|
|
|
len(self.active_probes) > self.slow_node_count[0] or
|
|
|
|
(
|
|
|
|
len(self.shortlist) < constants.k and
|
|
|
|
len(self.active_contacts) < len(self.shortlist) and
|
|
|
|
len(self.active_probes) > 0
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
|
|
|
def _is_all_done(self):
|
|
|
|
return (
|
|
|
|
len(self.active_contacts) >= constants.k or
|
|
|
|
(
|
|
|
|
self.active_contacts[0] == self.prev_closest_node[0] and
|
|
|
|
len(self.active_probes) == self.slow_node_count[0]
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2016-12-14 03:53:24 +01:00
|
|
|
class ExpensiveSort(object):
|
|
|
|
"""Sort a list in place.
|
|
|
|
|
|
|
|
The result of `key(item)` is cached for each item in the `to_sort`
|
|
|
|
list as an optimization. This can be useful when `key` is
|
|
|
|
expensive.
|
|
|
|
|
|
|
|
Attributes:
|
|
|
|
to_sort: a list of items to sort
|
|
|
|
key: callable, like `key` in normal python sort
|
|
|
|
attr: the attribute name used to cache the value on each item.
|
|
|
|
"""
|
2017-03-31 19:32:43 +02:00
|
|
|
|
2016-12-14 03:53:24 +01:00
|
|
|
def __init__(self, to_sort, key, attr='__value'):
|
|
|
|
self.to_sort = to_sort
|
|
|
|
self.key = key
|
|
|
|
self.attr = attr
|
|
|
|
|
|
|
|
def sort(self):
|
|
|
|
self._cacheValues()
|
|
|
|
self._sortByValue()
|
|
|
|
self._removeValue()
|
|
|
|
|
|
|
|
def _cacheValues(self):
|
|
|
|
for item in self.to_sort:
|
|
|
|
setattr(item, self.attr, self.key(item))
|
|
|
|
|
|
|
|
def _sortByValue(self):
|
|
|
|
self.to_sort.sort(key=operator.attrgetter(self.attr))
|
|
|
|
|
|
|
|
def _removeValue(self):
|
|
|
|
for item in self.to_sort:
|
|
|
|
delattr(item, self.attr)
|