2019-01-22 18:52:32 +01:00
|
|
|
import asyncio
|
|
|
|
import typing
|
|
|
|
import logging
|
|
|
|
from lbrynet.utils import drain_tasks
|
2019-01-30 20:57:09 +01:00
|
|
|
from lbrynet.blob_exchange.client import request_blob
|
2019-01-22 18:52:32 +01:00
|
|
|
if typing.TYPE_CHECKING:
|
2019-01-30 20:57:09 +01:00
|
|
|
from lbrynet.conf import Config
|
2019-01-22 18:52:32 +01:00
|
|
|
from lbrynet.dht.node import Node
|
|
|
|
from lbrynet.dht.peer import KademliaPeer
|
|
|
|
from lbrynet.blob.blob_manager import BlobFileManager
|
|
|
|
from lbrynet.blob.blob_file import BlobFile
|
|
|
|
|
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
|
|
|
def drain_into(a: list, b: list):
|
|
|
|
while a:
|
|
|
|
b.append(a.pop())
|
|
|
|
|
|
|
|
|
2019-01-30 20:57:09 +01:00
|
|
|
class BlobDownloader:
|
|
|
|
def __init__(self, loop: asyncio.BaseEventLoop, config: 'Config', blob_manager: 'BlobFileManager',
|
|
|
|
peer_queue: asyncio.Queue):
|
2019-01-22 18:52:32 +01:00
|
|
|
self.loop = loop
|
2019-01-30 20:57:09 +01:00
|
|
|
self.config = config
|
2019-01-22 18:52:32 +01:00
|
|
|
self.blob_manager = blob_manager
|
2019-01-30 20:57:09 +01:00
|
|
|
self.peer_queue = peer_queue
|
2019-01-31 19:05:00 +01:00
|
|
|
self.active_connections: typing.Dict['KademliaPeer', asyncio.Task] = {} # active request_blob calls
|
2019-01-30 20:57:09 +01:00
|
|
|
self.ignored: typing.Set['KademliaPeer'] = set()
|
2019-01-31 18:28:59 +01:00
|
|
|
self.scores: typing.Dict['KademliaPeer', int] = {}
|
2019-01-22 18:52:32 +01:00
|
|
|
|
2019-01-30 20:57:09 +01:00
|
|
|
def request_blob_from_peer(self, blob: 'BlobFile', peer: 'KademliaPeer'):
|
|
|
|
async def _request_blob():
|
|
|
|
if blob.get_is_verified():
|
|
|
|
return
|
2019-01-31 18:28:59 +01:00
|
|
|
success, keep_connection = await request_blob(
|
|
|
|
self.loop, blob, peer.address, peer.tcp_port, self.config.peer_connect_timeout,
|
|
|
|
self.config.blob_download_timeout
|
|
|
|
)
|
2019-01-30 20:57:09 +01:00
|
|
|
if not keep_connection and peer not in self.ignored:
|
|
|
|
self.ignored.add(peer)
|
|
|
|
log.debug("drop peer %s:%i", peer.address, peer.tcp_port)
|
|
|
|
elif keep_connection:
|
|
|
|
log.debug("keep peer %s:%i", peer.address, peer.tcp_port)
|
2019-01-31 18:28:59 +01:00
|
|
|
if success:
|
|
|
|
self.scores[peer] = self.scores.get(peer, 0) + 2
|
|
|
|
else:
|
|
|
|
self.scores[peer] = self.scores.get(peer, 0) - 1
|
2019-01-30 20:57:09 +01:00
|
|
|
return self.loop.create_task(_request_blob())
|
2019-01-22 18:52:32 +01:00
|
|
|
|
2019-01-30 20:57:09 +01:00
|
|
|
async def new_peer_or_finished(self, blob: 'BlobFile'):
|
|
|
|
async def get_and_re_add_peers():
|
|
|
|
new_peers = await self.peer_queue.get()
|
|
|
|
self.peer_queue.put_nowait(new_peers)
|
|
|
|
tasks = [self.loop.create_task(get_and_re_add_peers()), self.loop.create_task(blob.verified.wait())]
|
2019-01-22 18:52:32 +01:00
|
|
|
try:
|
2019-01-30 20:57:09 +01:00
|
|
|
await asyncio.wait(tasks, loop=self.loop, return_when='FIRST_COMPLETED')
|
2019-01-22 18:52:32 +01:00
|
|
|
except asyncio.CancelledError:
|
2019-01-30 20:57:09 +01:00
|
|
|
drain_tasks(tasks)
|
2019-01-22 18:52:32 +01:00
|
|
|
|
2019-01-30 20:57:09 +01:00
|
|
|
async def download_blob(self, blob_hash: str, length: typing.Optional[int] = None) -> 'BlobFile':
|
|
|
|
blob = self.blob_manager.get_blob(blob_hash, length)
|
|
|
|
if blob.get_is_verified():
|
|
|
|
return blob
|
2019-01-22 18:52:32 +01:00
|
|
|
try:
|
2019-01-30 20:57:09 +01:00
|
|
|
while not blob.get_is_verified():
|
|
|
|
batch: typing.List['KademliaPeer'] = []
|
|
|
|
while not self.peer_queue.empty():
|
|
|
|
batch.extend(await self.peer_queue.get())
|
|
|
|
for peer in batch:
|
2019-01-31 18:28:59 +01:00
|
|
|
if len(self.active_connections) >= self.config.max_connections_per_download:
|
|
|
|
break
|
2019-01-31 19:05:00 +01:00
|
|
|
if peer not in self.active_connections and peer not in self.ignored:
|
2019-01-31 18:28:59 +01:00
|
|
|
log.debug("request %s from %s:%i", blob_hash[:8], peer.address, peer.tcp_port)
|
2019-01-31 19:05:00 +01:00
|
|
|
self.active_connections[peer] = self.request_blob_from_peer(blob, peer)
|
2019-01-30 20:57:09 +01:00
|
|
|
await self.new_peer_or_finished(blob)
|
|
|
|
to_re_add = list(set(filter(lambda peer: peer not in self.ignored, batch)))
|
2019-01-31 18:28:59 +01:00
|
|
|
to_re_add.sort(key=lambda peer: self.scores.get(peer, 0), reverse=True)
|
2019-01-30 20:57:09 +01:00
|
|
|
if to_re_add:
|
|
|
|
self.peer_queue.put_nowait(to_re_add)
|
|
|
|
while self.active_connections:
|
|
|
|
peer, task = self.active_connections.popitem()
|
|
|
|
if task and not task.done():
|
|
|
|
task.cancel()
|
|
|
|
return blob
|
2019-01-22 18:52:32 +01:00
|
|
|
except asyncio.CancelledError:
|
2019-01-30 20:57:09 +01:00
|
|
|
while self.active_connections:
|
|
|
|
peer, task = self.active_connections.popitem()
|
|
|
|
if task and not task.done():
|
|
|
|
task.cancel()
|
2019-01-22 18:52:32 +01:00
|
|
|
raise
|
2019-01-30 20:57:09 +01:00
|
|
|
|
|
|
|
|
|
|
|
async def download_blob(loop, config: 'Config', blob_manager: 'BlobFileManager', node: 'Node',
|
|
|
|
blob_hash: str) -> 'BlobFile':
|
2019-01-31 18:28:59 +01:00
|
|
|
search_queue = asyncio.Queue(loop=loop, maxsize=config.max_connections_per_download)
|
2019-01-30 20:57:09 +01:00
|
|
|
search_queue.put_nowait(blob_hash)
|
|
|
|
peer_queue, accumulate_task = node.accumulate_peers(search_queue)
|
|
|
|
downloader = BlobDownloader(loop, config, blob_manager, peer_queue)
|
|
|
|
try:
|
|
|
|
return await downloader.download_blob(blob_hash)
|
|
|
|
finally:
|
|
|
|
if accumulate_task and not accumulate_task.done():
|
|
|
|
accumulate_task.cancel()
|