2019-01-22 18:54:17 +01:00
|
|
|
import os
|
|
|
|
import asyncio
|
2019-10-11 01:46:00 +02:00
|
|
|
import time
|
2019-01-22 18:54:17 +01:00
|
|
|
import typing
|
|
|
|
import logging
|
2020-01-15 16:18:38 +01:00
|
|
|
from typing import Optional
|
2019-05-24 04:40:59 +02:00
|
|
|
from aiohttp.web import Request, StreamResponse, HTTPRequestRangeNotSatisfiable
|
2019-11-19 19:57:14 +01:00
|
|
|
from lbry.error import DownloadSDTimeoutError
|
2019-06-21 02:55:47 +02:00
|
|
|
from lbry.schema.mime_types import guess_media_type
|
|
|
|
from lbry.stream.downloader import StreamDownloader
|
2019-10-09 19:32:52 +02:00
|
|
|
from lbry.stream.descriptor import StreamDescriptor, sanitize_file_name
|
2019-06-21 02:55:47 +02:00
|
|
|
from lbry.stream.reflector.client import StreamReflectorClient
|
2019-11-15 20:55:49 +01:00
|
|
|
from lbry.extras.daemon.storage import StoredContentClaim
|
2019-10-04 15:18:54 +02:00
|
|
|
from lbry.blob import MAX_BLOB_SIZE
|
2020-01-15 16:18:38 +01:00
|
|
|
from lbry.file.source import ManagedDownloadSource
|
2019-10-04 15:18:54 +02:00
|
|
|
|
2019-01-22 18:54:17 +01:00
|
|
|
if typing.TYPE_CHECKING:
|
2019-06-21 02:55:47 +02:00
|
|
|
from lbry.conf import Config
|
|
|
|
from lbry.schema.claim import Claim
|
|
|
|
from lbry.blob.blob_manager import BlobManager
|
|
|
|
from lbry.blob.blob_info import BlobInfo
|
|
|
|
from lbry.dht.node import Node
|
|
|
|
from lbry.extras.daemon.analytics import AnalyticsManager
|
|
|
|
from lbry.wallet.transaction import Transaction
|
2019-01-22 18:54:17 +01:00
|
|
|
|
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
2019-03-31 03:07:43 +02:00
|
|
|
def _get_next_available_file_name(download_directory: str, file_name: str) -> str:
|
|
|
|
base_name, ext = os.path.splitext(os.path.basename(file_name))
|
|
|
|
i = 0
|
|
|
|
while os.path.isfile(os.path.join(download_directory, file_name)):
|
|
|
|
i += 1
|
|
|
|
file_name = "%s_%i%s" % (base_name, i, ext)
|
|
|
|
|
|
|
|
return file_name
|
|
|
|
|
|
|
|
|
2019-08-02 19:14:41 +02:00
|
|
|
async def get_next_available_file_name(loop: asyncio.AbstractEventLoop, download_directory: str, file_name: str) -> str:
|
2019-03-31 03:07:43 +02:00
|
|
|
return await loop.run_in_executor(None, _get_next_available_file_name, download_directory, file_name)
|
|
|
|
|
|
|
|
|
2020-01-15 16:18:38 +01:00
|
|
|
class ManagedStream(ManagedDownloadSource):
|
2019-08-02 19:14:41 +02:00
|
|
|
def __init__(self, loop: asyncio.AbstractEventLoop, config: 'Config', blob_manager: 'BlobManager',
|
2020-01-15 16:18:38 +01:00
|
|
|
sd_hash: str, download_directory: Optional[str] = None, file_name: Optional[str] = None,
|
|
|
|
status: Optional[str] = ManagedDownloadSource.STATUS_STOPPED,
|
|
|
|
claim: Optional[StoredContentClaim] = None,
|
|
|
|
download_id: Optional[str] = None, rowid: Optional[int] = None,
|
|
|
|
descriptor: Optional[StreamDescriptor] = None,
|
|
|
|
content_fee: Optional['Transaction'] = None,
|
|
|
|
analytics_manager: Optional['AnalyticsManager'] = None,
|
|
|
|
added_on: Optional[int] = None):
|
|
|
|
super().__init__(loop, config, blob_manager.storage, sd_hash, file_name, download_directory, status, claim,
|
|
|
|
download_id, rowid, content_fee, analytics_manager, added_on)
|
2019-01-22 18:54:17 +01:00
|
|
|
self.blob_manager = blob_manager
|
2019-10-30 03:56:28 +01:00
|
|
|
self.purchase_receipt = None
|
2019-03-31 03:07:43 +02:00
|
|
|
self.downloader = StreamDownloader(self.loop, self.config, self.blob_manager, sd_hash, descriptor)
|
2019-03-31 19:42:27 +02:00
|
|
|
self.analytics_manager = analytics_manager
|
2019-05-01 23:09:50 +02:00
|
|
|
|
2020-04-10 16:56:45 +02:00
|
|
|
self.reflector_progress = 0
|
2020-04-20 17:57:09 +02:00
|
|
|
self.uploading_to_reflector = False
|
2019-03-31 03:07:43 +02:00
|
|
|
self.file_output_task: typing.Optional[asyncio.Task] = None
|
2019-05-01 23:09:50 +02:00
|
|
|
self.delayed_stop_task: typing.Optional[asyncio.Task] = None
|
2019-05-02 22:56:29 +02:00
|
|
|
self.streaming_responses: typing.List[typing.Tuple[Request, StreamResponse]] = []
|
2020-01-15 16:18:38 +01:00
|
|
|
self.fully_reflected = asyncio.Event(loop=self.loop)
|
2019-05-01 23:09:50 +02:00
|
|
|
self.streaming = asyncio.Event(loop=self.loop)
|
|
|
|
self._running = asyncio.Event(loop=self.loop)
|
2020-01-15 16:18:38 +01:00
|
|
|
|
|
|
|
@property
|
|
|
|
def sd_hash(self) -> str:
|
|
|
|
return self.identifier
|
2019-03-31 03:07:43 +02:00
|
|
|
|
2020-02-07 16:34:47 +01:00
|
|
|
@property
|
|
|
|
def is_fully_reflected(self) -> bool:
|
|
|
|
return self.fully_reflected.is_set()
|
|
|
|
|
2019-03-31 03:07:43 +02:00
|
|
|
@property
|
|
|
|
def descriptor(self) -> StreamDescriptor:
|
|
|
|
return self.downloader.descriptor
|
|
|
|
|
|
|
|
@property
|
|
|
|
def stream_hash(self) -> str:
|
|
|
|
return self.descriptor.stream_hash
|
2019-01-22 18:54:17 +01:00
|
|
|
|
2019-02-14 00:41:26 +01:00
|
|
|
@property
|
2020-01-15 16:18:38 +01:00
|
|
|
def file_name(self) -> Optional[str]:
|
2019-10-08 20:03:27 +02:00
|
|
|
return self._file_name or (self.descriptor.suggested_file_name if self.descriptor else None)
|
2019-02-14 00:41:26 +01:00
|
|
|
|
2019-07-15 09:45:22 +02:00
|
|
|
@property
|
|
|
|
def written_bytes(self) -> int:
|
|
|
|
return 0 if not self.output_file_exists else os.stat(self.full_path).st_size
|
|
|
|
|
2019-09-09 01:02:05 +02:00
|
|
|
@property
|
|
|
|
def completed(self):
|
|
|
|
return self.written_bytes >= self.descriptor.lower_bound_decrypted_length()
|
|
|
|
|
|
|
|
@property
|
|
|
|
def stream_url(self):
|
|
|
|
return f"http://{self.config.streaming_host}:{self.config.streaming_port}/stream/{self.sd_hash}"
|
|
|
|
|
2019-05-01 23:09:50 +02:00
|
|
|
async def update_status(self, status: str):
|
2019-01-22 18:54:17 +01:00
|
|
|
assert status in [self.STATUS_RUNNING, self.STATUS_STOPPED, self.STATUS_FINISHED]
|
|
|
|
self._status = status
|
2019-05-01 23:09:50 +02:00
|
|
|
await self.blob_manager.storage.change_file_status(self.stream_hash, status)
|
2019-01-22 18:54:17 +01:00
|
|
|
|
|
|
|
@property
|
|
|
|
def blobs_completed(self) -> int:
|
2019-08-12 00:11:13 +02:00
|
|
|
return sum([1 if b.blob_hash in self.blob_manager.completed_blob_hashes else 0
|
2019-01-22 18:54:17 +01:00
|
|
|
for b in self.descriptor.blobs[:-1]])
|
|
|
|
|
|
|
|
@property
|
|
|
|
def blobs_in_stream(self) -> int:
|
|
|
|
return len(self.descriptor.blobs) - 1
|
|
|
|
|
2019-02-01 22:17:10 +01:00
|
|
|
@property
|
|
|
|
def blobs_remaining(self) -> int:
|
|
|
|
return self.blobs_in_stream - self.blobs_completed
|
|
|
|
|
2019-03-31 03:07:43 +02:00
|
|
|
@property
|
|
|
|
def mime_type(self):
|
2019-04-05 05:10:18 +02:00
|
|
|
return guess_media_type(os.path.basename(self.descriptor.suggested_file_name))[0]
|
2019-03-31 03:07:43 +02:00
|
|
|
|
2020-05-17 13:12:31 +02:00
|
|
|
@property
|
|
|
|
def download_path(self):
|
|
|
|
return f"{self.download_directory}/{self._file_name}" if self.download_directory and self._file_name else None
|
|
|
|
|
2020-01-15 16:18:38 +01:00
|
|
|
# @classmethod
|
|
|
|
# async def create(cls, loop: asyncio.AbstractEventLoop, config: 'Config',
|
|
|
|
# file_path: str, key: Optional[bytes] = None,
|
|
|
|
# iv_generator: Optional[typing.Generator[bytes, None, None]] = None) -> 'ManagedDownloadSource':
|
|
|
|
# """
|
|
|
|
# Generate a stream from a file and save it to the db
|
|
|
|
# """
|
|
|
|
# descriptor = await StreamDescriptor.create_stream(
|
|
|
|
# loop, blob_manager.blob_dir, file_path, key=key, iv_generator=iv_generator,
|
|
|
|
# blob_completed_callback=blob_manager.blob_completed
|
|
|
|
# )
|
|
|
|
# await blob_manager.storage.store_stream(
|
|
|
|
# blob_manager.get_blob(descriptor.sd_hash), descriptor
|
|
|
|
# )
|
|
|
|
# row_id = await blob_manager.storage.save_published_file(descriptor.stream_hash, os.path.basename(file_path),
|
|
|
|
# os.path.dirname(file_path), 0)
|
|
|
|
# return cls(loop, config, blob_manager, descriptor.sd_hash, os.path.dirname(file_path),
|
|
|
|
# os.path.basename(file_path), status=cls.STATUS_FINISHED, rowid=row_id, descriptor=descriptor)
|
|
|
|
|
2020-01-29 01:24:05 +01:00
|
|
|
async def start(self, timeout: Optional[float] = None,
|
2019-05-01 23:09:50 +02:00
|
|
|
save_now: bool = False):
|
|
|
|
timeout = timeout or self.config.download_timeout
|
|
|
|
if self._running.is_set():
|
|
|
|
return
|
2019-05-06 02:22:10 +02:00
|
|
|
log.info("start downloader for stream (sd hash: %s)", self.sd_hash)
|
2019-05-01 23:09:50 +02:00
|
|
|
self._running.set()
|
|
|
|
try:
|
2020-01-29 01:24:05 +01:00
|
|
|
await asyncio.wait_for(self.downloader.start(), timeout, loop=self.loop)
|
2019-05-01 23:09:50 +02:00
|
|
|
except asyncio.TimeoutError:
|
|
|
|
self._running.clear()
|
2019-11-19 19:57:14 +01:00
|
|
|
raise DownloadSDTimeoutError(self.sd_hash)
|
2019-05-01 23:09:50 +02:00
|
|
|
|
|
|
|
if self.delayed_stop_task and not self.delayed_stop_task.done():
|
|
|
|
self.delayed_stop_task.cancel()
|
|
|
|
self.delayed_stop_task = self.loop.create_task(self._delayed_stop())
|
|
|
|
if not await self.blob_manager.storage.file_exists(self.sd_hash):
|
|
|
|
if save_now:
|
2020-01-29 01:24:05 +01:00
|
|
|
if not self._file_name:
|
|
|
|
self._file_name = await get_next_available_file_name(
|
|
|
|
self.loop, self.download_directory,
|
|
|
|
self._file_name or sanitize_file_name(self.descriptor.suggested_file_name)
|
|
|
|
)
|
2019-10-08 20:03:27 +02:00
|
|
|
file_name, download_dir = self._file_name, self.download_directory
|
2019-05-01 23:09:50 +02:00
|
|
|
else:
|
|
|
|
file_name, download_dir = None, None
|
2019-10-26 17:24:37 +02:00
|
|
|
self._added_on = int(time.time())
|
2019-05-01 23:09:50 +02:00
|
|
|
self.rowid = await self.blob_manager.storage.save_downloaded_file(
|
2019-10-26 17:24:37 +02:00
|
|
|
self.stream_hash, file_name, download_dir, 0.0, added_on=self._added_on
|
2019-05-01 23:09:50 +02:00
|
|
|
)
|
|
|
|
if self.status != self.STATUS_RUNNING:
|
|
|
|
await self.update_status(self.STATUS_RUNNING)
|
2019-03-31 03:07:43 +02:00
|
|
|
|
2019-05-01 23:09:50 +02:00
|
|
|
async def stop(self, finished: bool = False):
|
|
|
|
"""
|
|
|
|
Stop any running save/stream tasks as well as the downloader and update the status in the database
|
|
|
|
"""
|
2019-03-31 03:07:43 +02:00
|
|
|
|
2019-05-01 23:09:50 +02:00
|
|
|
self.stop_tasks()
|
|
|
|
if (finished and self.status != self.STATUS_FINISHED) or self.status == self.STATUS_RUNNING:
|
|
|
|
await self.update_status(self.STATUS_FINISHED if finished else self.STATUS_STOPPED)
|
2019-03-31 03:07:43 +02:00
|
|
|
|
2020-01-15 16:18:38 +01:00
|
|
|
async def _aiter_read_stream(self, start_blob_num: Optional[int] = 0, connection_id: int = 0)\
|
2019-05-01 23:09:50 +02:00
|
|
|
-> typing.AsyncIterator[typing.Tuple['BlobInfo', bytes]]:
|
2019-03-31 03:07:43 +02:00
|
|
|
if start_blob_num >= len(self.descriptor.blobs[:-1]):
|
|
|
|
raise IndexError(start_blob_num)
|
|
|
|
for i, blob_info in enumerate(self.descriptor.blobs[start_blob_num:-1]):
|
|
|
|
assert i + start_blob_num == blob_info.blob_num
|
2019-05-24 04:40:02 +02:00
|
|
|
if connection_id == self.STREAMING_ID:
|
|
|
|
decrypted = await self.downloader.cached_read_blob(blob_info)
|
|
|
|
else:
|
|
|
|
decrypted = await self.downloader.read_blob(blob_info, connection_id)
|
2019-05-01 23:09:50 +02:00
|
|
|
yield (blob_info, decrypted)
|
|
|
|
|
2020-05-12 06:32:36 +02:00
|
|
|
async def stream_file(self, request: Request) -> StreamResponse:
|
2019-05-02 22:55:53 +02:00
|
|
|
log.info("stream file to browser for lbry://%s#%s (sd hash %s...)", self.claim_name, self.claim_id,
|
|
|
|
self.sd_hash[:6])
|
2019-05-24 04:40:59 +02:00
|
|
|
headers, size, skip_blobs, first_blob_start_offset = self._prepare_range_response_headers(
|
|
|
|
request.headers.get('range', 'bytes=0-')
|
|
|
|
)
|
2020-05-12 06:32:36 +02:00
|
|
|
await self.start()
|
2019-05-01 23:09:50 +02:00
|
|
|
response = StreamResponse(
|
|
|
|
status=206,
|
|
|
|
headers=headers
|
|
|
|
)
|
|
|
|
await response.prepare(request)
|
2019-05-02 22:56:29 +02:00
|
|
|
self.streaming_responses.append((request, response))
|
2019-05-01 23:09:50 +02:00
|
|
|
self.streaming.set()
|
2019-09-26 23:07:05 +02:00
|
|
|
wrote = 0
|
2019-05-01 23:09:50 +02:00
|
|
|
try:
|
2019-05-24 04:40:59 +02:00
|
|
|
async for blob_info, decrypted in self._aiter_read_stream(skip_blobs, connection_id=self.STREAMING_ID):
|
|
|
|
if not wrote:
|
|
|
|
decrypted = decrypted[first_blob_start_offset:]
|
2019-05-01 23:09:50 +02:00
|
|
|
if (blob_info.blob_num == len(self.descriptor.blobs) - 2) or (len(decrypted) + wrote >= size):
|
2019-10-04 15:18:54 +02:00
|
|
|
decrypted += (b'\x00' * (size - len(decrypted) - wrote - (skip_blobs * (MAX_BLOB_SIZE - 1))))
|
2019-05-24 04:40:59 +02:00
|
|
|
log.debug("sending browser final blob (%i/%i)", blob_info.blob_num + 1,
|
|
|
|
len(self.descriptor.blobs) - 1)
|
2019-05-01 23:09:50 +02:00
|
|
|
await response.write_eof(decrypted)
|
|
|
|
else:
|
2019-05-24 04:40:59 +02:00
|
|
|
log.debug("sending browser blob (%i/%i)", blob_info.blob_num + 1, len(self.descriptor.blobs) - 1)
|
2019-05-01 23:09:50 +02:00
|
|
|
await response.write(decrypted)
|
|
|
|
wrote += len(decrypted)
|
2019-05-02 22:55:53 +02:00
|
|
|
log.info("sent browser %sblob %i/%i", "(final) " if response._eof_sent else "",
|
2019-05-01 23:09:50 +02:00
|
|
|
blob_info.blob_num + 1, len(self.descriptor.blobs) - 1)
|
|
|
|
if response._eof_sent:
|
|
|
|
break
|
|
|
|
return response
|
2019-09-26 23:07:05 +02:00
|
|
|
except ConnectionResetError:
|
|
|
|
log.warning("connection was reset after sending browser %i blob bytes", wrote)
|
|
|
|
raise asyncio.CancelledError("range request transport was reset")
|
2019-05-01 23:09:50 +02:00
|
|
|
finally:
|
|
|
|
response.force_close()
|
2019-05-02 22:56:29 +02:00
|
|
|
if (request, response) in self.streaming_responses:
|
|
|
|
self.streaming_responses.remove((request, response))
|
|
|
|
if not self.streaming_responses:
|
2019-05-01 23:09:50 +02:00
|
|
|
self.streaming.clear()
|
2019-03-31 03:07:43 +02:00
|
|
|
|
2019-05-08 22:07:32 +02:00
|
|
|
@staticmethod
|
2020-04-22 22:10:23 +02:00
|
|
|
def _write_decrypted_blob(output_path: str, data: bytes):
|
|
|
|
with open(output_path, 'ab') as handle:
|
|
|
|
handle.write(data)
|
|
|
|
handle.flush()
|
2019-05-08 22:07:32 +02:00
|
|
|
|
2021-10-15 08:23:45 +02:00
|
|
|
async def save_blobs(self):
|
|
|
|
async for _ in self._aiter_read_stream(0, connection_id=self.STREAMING_ID):
|
|
|
|
pass
|
|
|
|
|
2019-03-31 03:07:43 +02:00
|
|
|
async def _save_file(self, output_path: str):
|
2019-05-02 22:55:53 +02:00
|
|
|
log.info("save file for lbry://%s#%s (sd hash %s...) -> %s", self.claim_name, self.claim_id, self.sd_hash[:6],
|
|
|
|
output_path)
|
2019-03-31 03:07:43 +02:00
|
|
|
self.saving.set()
|
2019-05-10 20:50:01 +02:00
|
|
|
self.finished_write_attempt.clear()
|
2019-03-31 03:07:43 +02:00
|
|
|
self.finished_writing.clear()
|
2019-03-31 19:42:27 +02:00
|
|
|
self.started_writing.clear()
|
2019-03-31 03:07:43 +02:00
|
|
|
try:
|
2021-08-21 04:36:35 +02:00
|
|
|
open(output_path, 'wb').close() # pylint: disable=consider-using-with
|
2020-04-22 22:10:23 +02:00
|
|
|
async for blob_info, decrypted in self._aiter_read_stream(connection_id=self.SAVING_ID):
|
|
|
|
log.info("write blob %i/%i", blob_info.blob_num + 1, len(self.descriptor.blobs) - 1)
|
|
|
|
await self.loop.run_in_executor(None, self._write_decrypted_blob, output_path, decrypted)
|
|
|
|
if not self.started_writing.is_set():
|
|
|
|
self.started_writing.set()
|
2019-05-01 23:09:50 +02:00
|
|
|
await self.update_status(ManagedStream.STATUS_FINISHED)
|
2019-03-31 19:42:27 +02:00
|
|
|
if self.analytics_manager:
|
|
|
|
self.loop.create_task(self.analytics_manager.send_download_finished(
|
|
|
|
self.download_id, self.claim_name, self.sd_hash
|
|
|
|
))
|
2019-03-31 03:07:43 +02:00
|
|
|
self.finished_writing.set()
|
2019-05-03 20:53:23 +02:00
|
|
|
log.info("finished saving file for lbry://%s#%s (sd hash %s...) -> %s", self.claim_name, self.claim_id,
|
2019-05-03 22:36:24 +02:00
|
|
|
self.sd_hash[:6], self.full_path)
|
2019-05-07 20:30:35 +02:00
|
|
|
await self.blob_manager.storage.set_saved_file(self.stream_hash)
|
2019-03-31 03:07:43 +02:00
|
|
|
except Exception as err:
|
|
|
|
if os.path.isfile(output_path):
|
2019-05-02 22:55:53 +02:00
|
|
|
log.warning("removing incomplete download %s for %s", output_path, self.sd_hash)
|
2019-03-31 03:07:43 +02:00
|
|
|
os.remove(output_path)
|
2019-05-10 20:50:01 +02:00
|
|
|
if isinstance(err, asyncio.TimeoutError):
|
|
|
|
self.downloader.stop()
|
|
|
|
await self.blob_manager.storage.change_file_download_dir_and_file_name(
|
|
|
|
self.stream_hash, None, None
|
|
|
|
)
|
2019-10-08 20:03:27 +02:00
|
|
|
self._file_name, self.download_directory = None, None
|
2019-05-10 20:50:01 +02:00
|
|
|
await self.blob_manager.storage.clear_saved_file(self.stream_hash)
|
|
|
|
await self.update_status(self.STATUS_STOPPED)
|
|
|
|
return
|
|
|
|
elif not isinstance(err, asyncio.CancelledError):
|
2019-03-31 19:42:27 +02:00
|
|
|
log.exception("unexpected error encountered writing file for stream %s", self.sd_hash)
|
2019-03-31 03:07:43 +02:00
|
|
|
raise err
|
|
|
|
finally:
|
|
|
|
self.saving.clear()
|
2019-05-10 20:50:01 +02:00
|
|
|
self.finished_write_attempt.set()
|
2019-03-31 03:07:43 +02:00
|
|
|
|
2020-01-29 02:37:52 +01:00
|
|
|
async def save_file(self, file_name: Optional[str] = None, download_directory: Optional[str] = None):
|
2020-01-29 01:24:05 +01:00
|
|
|
await self.start()
|
2019-05-01 23:09:50 +02:00
|
|
|
if self.file_output_task and not self.file_output_task.done(): # cancel an already running save task
|
2019-03-31 03:07:43 +02:00
|
|
|
self.file_output_task.cancel()
|
2019-03-31 19:42:27 +02:00
|
|
|
self.download_directory = download_directory or self.download_directory or self.config.download_dir
|
2019-03-31 03:07:43 +02:00
|
|
|
if not self.download_directory:
|
|
|
|
raise ValueError("no directory to download to")
|
2019-10-08 20:03:27 +02:00
|
|
|
if not (file_name or self._file_name or self.descriptor.suggested_file_name):
|
2019-03-31 03:07:43 +02:00
|
|
|
raise ValueError("no file name to download to")
|
|
|
|
if not os.path.isdir(self.download_directory):
|
|
|
|
log.warning("download directory '%s' does not exist, attempting to make it", self.download_directory)
|
|
|
|
os.mkdir(self.download_directory)
|
2019-10-08 20:03:27 +02:00
|
|
|
self._file_name = await get_next_available_file_name(
|
2019-05-01 23:09:50 +02:00
|
|
|
self.loop, self.download_directory,
|
2019-10-09 19:32:52 +02:00
|
|
|
file_name or self._file_name or sanitize_file_name(self.descriptor.suggested_file_name)
|
2019-05-01 23:09:50 +02:00
|
|
|
)
|
|
|
|
await self.blob_manager.storage.change_file_download_dir_and_file_name(
|
|
|
|
self.stream_hash, self.download_directory, self.file_name
|
|
|
|
)
|
|
|
|
await self.update_status(ManagedStream.STATUS_RUNNING)
|
2019-03-31 03:07:43 +02:00
|
|
|
self.file_output_task = self.loop.create_task(self._save_file(self.full_path))
|
2019-12-02 05:15:31 +01:00
|
|
|
try:
|
|
|
|
await asyncio.wait_for(self.started_writing.wait(), self.config.download_timeout, loop=self.loop)
|
|
|
|
except asyncio.TimeoutError:
|
|
|
|
log.warning("timeout starting to write data for lbry://%s#%s", self.claim_name, self.claim_id)
|
|
|
|
self.stop_tasks()
|
|
|
|
await self.update_status(ManagedStream.STATUS_STOPPED)
|
2019-02-01 21:46:31 +01:00
|
|
|
|
2019-05-01 23:09:50 +02:00
|
|
|
def stop_tasks(self):
|
2019-03-31 03:07:43 +02:00
|
|
|
if self.file_output_task and not self.file_output_task.done():
|
|
|
|
self.file_output_task.cancel()
|
|
|
|
self.file_output_task = None
|
2019-05-01 23:09:50 +02:00
|
|
|
while self.streaming_responses:
|
2019-05-02 22:56:29 +02:00
|
|
|
req, response = self.streaming_responses.pop()
|
|
|
|
response.force_close()
|
|
|
|
req.transport.close()
|
2019-03-31 03:07:43 +02:00
|
|
|
self.downloader.stop()
|
2019-05-01 23:09:50 +02:00
|
|
|
self._running.clear()
|
2019-01-25 21:05:22 +01:00
|
|
|
|
|
|
|
async def upload_to_reflector(self, host: str, port: int) -> typing.List[str]:
|
|
|
|
sent = []
|
|
|
|
protocol = StreamReflectorClient(self.blob_manager, self.descriptor)
|
|
|
|
try:
|
2020-04-20 17:57:09 +02:00
|
|
|
self.uploading_to_reflector = True
|
2019-01-25 21:05:22 +01:00
|
|
|
await self.loop.create_connection(lambda: protocol, host, port)
|
|
|
|
await protocol.send_handshake()
|
|
|
|
sent_sd, needed = await protocol.send_descriptor()
|
2020-02-07 16:34:47 +01:00
|
|
|
if sent_sd: # reflector needed the sd blob
|
2019-01-25 21:05:22 +01:00
|
|
|
sent.append(self.sd_hash)
|
2020-02-07 16:34:47 +01:00
|
|
|
if not sent_sd and not needed: # reflector already has the stream
|
2019-01-30 20:58:26 +01:00
|
|
|
if not self.fully_reflected.is_set():
|
|
|
|
self.fully_reflected.set()
|
2019-01-31 18:30:31 +01:00
|
|
|
await self.blob_manager.storage.update_reflected_stream(self.sd_hash, f"{host}:{port}")
|
|
|
|
return []
|
2019-03-31 03:07:43 +02:00
|
|
|
we_have = [
|
|
|
|
blob_hash for blob_hash in needed if blob_hash in self.blob_manager.completed_blob_hashes
|
|
|
|
]
|
2019-10-29 17:46:29 +01:00
|
|
|
log.info("we have %i/%i needed blobs needed by reflector for lbry://%s#%s", len(we_have), len(needed),
|
|
|
|
self.claim_name, self.claim_id)
|
2020-04-10 16:56:45 +02:00
|
|
|
for i, blob_hash in enumerate(we_have):
|
2019-02-04 17:34:18 +01:00
|
|
|
await protocol.send_blob(blob_hash)
|
|
|
|
sent.append(blob_hash)
|
2020-04-10 16:56:45 +02:00
|
|
|
self.reflector_progress = int((i + 1) / len(we_have) * 100)
|
2019-02-22 03:07:45 +01:00
|
|
|
except (asyncio.TimeoutError, ValueError):
|
2019-02-04 17:34:18 +01:00
|
|
|
return sent
|
2021-01-08 15:53:12 +01:00
|
|
|
except ConnectionError:
|
2019-02-04 17:34:18 +01:00
|
|
|
return sent
|
2020-08-11 19:04:05 +02:00
|
|
|
except (OSError, Exception) as err:
|
|
|
|
if isinstance(err, asyncio.CancelledError):
|
|
|
|
log.warning("stopped uploading %s#%s to reflector", self.claim_name, self.claim_id)
|
2021-01-08 15:53:12 +01:00
|
|
|
elif isinstance(err, OSError):
|
2020-08-11 19:04:05 +02:00
|
|
|
log.warning(
|
|
|
|
"stopped uploading %s#%s to reflector because blobs were deleted or moved", self.claim_name,
|
|
|
|
self.claim_id
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
log.exception("unexpected error reflecting %s#%s", self.claim_name, self.claim_id)
|
2021-01-08 15:53:12 +01:00
|
|
|
return sent
|
2019-02-04 17:34:18 +01:00
|
|
|
finally:
|
2019-01-25 21:05:22 +01:00
|
|
|
if protocol.transport:
|
|
|
|
protocol.transport.close()
|
2020-04-20 17:57:09 +02:00
|
|
|
self.uploading_to_reflector = False
|
2020-08-11 19:04:05 +02:00
|
|
|
|
2019-01-25 21:05:22 +01:00
|
|
|
return sent
|
2019-01-31 18:32:52 +01:00
|
|
|
|
2020-01-15 16:18:38 +01:00
|
|
|
async def update_content_claim(self, claim_info: Optional[typing.Dict] = None):
|
2019-05-01 23:09:50 +02:00
|
|
|
if not claim_info:
|
|
|
|
claim_info = await self.blob_manager.storage.get_content_claim(self.stream_hash)
|
|
|
|
self.set_claim(claim_info, claim_info['value'])
|
|
|
|
|
|
|
|
async def _delayed_stop(self):
|
|
|
|
stalled_count = 0
|
|
|
|
while self._running.is_set():
|
|
|
|
if self.saving.is_set() or self.streaming.is_set():
|
|
|
|
stalled_count = 0
|
|
|
|
else:
|
|
|
|
stalled_count += 1
|
|
|
|
if stalled_count > 1:
|
2019-05-02 22:55:53 +02:00
|
|
|
log.info("stopping inactive download for lbry://%s#%s (%s...)", self.claim_name, self.claim_id,
|
|
|
|
self.sd_hash[:6])
|
2019-05-01 23:09:50 +02:00
|
|
|
await self.stop()
|
|
|
|
return
|
|
|
|
await asyncio.sleep(1, loop=self.loop)
|
|
|
|
|
2019-05-24 04:40:59 +02:00
|
|
|
def _prepare_range_response_headers(self, get_range: str) -> typing.Tuple[typing.Dict[str, str], int, int, int]:
|
2019-05-01 23:09:50 +02:00
|
|
|
if '=' in get_range:
|
|
|
|
get_range = get_range.split('=')[1]
|
|
|
|
start, end = get_range.split('-')
|
|
|
|
size = 0
|
2019-05-08 20:01:19 +02:00
|
|
|
|
2019-05-01 23:09:50 +02:00
|
|
|
for blob in self.descriptor.blobs[:-1]:
|
|
|
|
size += blob.length - 1
|
2019-05-08 20:01:19 +02:00
|
|
|
if self.stream_claim_info and self.stream_claim_info.claim.stream.source.size:
|
|
|
|
size_from_claim = int(self.stream_claim_info.claim.stream.source.size)
|
|
|
|
if not size_from_claim <= size <= size_from_claim + 16:
|
|
|
|
raise ValueError("claim contains implausible stream size")
|
|
|
|
log.debug("using stream size from claim")
|
|
|
|
size = size_from_claim
|
|
|
|
elif self.stream_claim_info:
|
|
|
|
log.debug("estimating stream size")
|
|
|
|
|
2019-05-01 23:09:50 +02:00
|
|
|
start = int(start)
|
2019-05-24 04:40:59 +02:00
|
|
|
if not 0 <= start < size:
|
|
|
|
raise HTTPRequestRangeNotSatisfiable()
|
|
|
|
|
2019-05-01 23:09:50 +02:00
|
|
|
end = int(end) if end else size - 1
|
2019-05-24 04:40:59 +02:00
|
|
|
|
|
|
|
if end >= size:
|
|
|
|
raise HTTPRequestRangeNotSatisfiable()
|
|
|
|
|
2019-10-04 15:18:54 +02:00
|
|
|
skip_blobs = start // (MAX_BLOB_SIZE - 2) # -2 because ... dont remember
|
|
|
|
skip = skip_blobs * (MAX_BLOB_SIZE - 1) # -1 because
|
2019-05-24 04:40:59 +02:00
|
|
|
skip_first_blob = start - skip
|
|
|
|
start = skip_first_blob + skip
|
2019-05-01 23:09:50 +02:00
|
|
|
final_size = end - start + 1
|
|
|
|
headers = {
|
|
|
|
'Accept-Ranges': 'bytes',
|
|
|
|
'Content-Range': f'bytes {start}-{end}/{size}',
|
|
|
|
'Content-Length': str(final_size),
|
|
|
|
'Content-Type': self.mime_type
|
|
|
|
}
|
2019-05-24 04:40:59 +02:00
|
|
|
return headers, size, skip_blobs, skip_first_blob
|