2019-02-15 00:19:01 +01:00
|
|
|
import os
|
2018-02-12 19:36:35 +01:00
|
|
|
import logging
|
2019-01-22 21:43:01 +01:00
|
|
|
import sqlite3
|
2018-11-30 22:11:23 +01:00
|
|
|
import typing
|
2019-01-22 21:43:01 +01:00
|
|
|
import asyncio
|
|
|
|
import binascii
|
2019-01-31 18:30:31 +01:00
|
|
|
import time
|
2020-01-03 06:52:28 +01:00
|
|
|
from typing import Optional
|
2020-01-03 04:18:49 +01:00
|
|
|
from lbry.wallet import SQLiteMixin
|
2019-06-21 02:55:47 +02:00
|
|
|
from lbry.conf import Config
|
|
|
|
from lbry.wallet.dewies import dewies_to_lbc, lbc_to_dewies
|
2020-01-27 06:10:55 +01:00
|
|
|
from lbry.wallet.transaction import Transaction, Output
|
2019-06-21 02:55:47 +02:00
|
|
|
from lbry.schema.claim import Claim
|
2020-01-03 04:57:28 +01:00
|
|
|
from lbry.dht.constants import DATA_EXPIRATION
|
2019-06-21 02:55:47 +02:00
|
|
|
from lbry.blob.blob_info import BlobInfo
|
2019-01-22 21:43:01 +01:00
|
|
|
|
|
|
|
if typing.TYPE_CHECKING:
|
2019-06-21 02:55:47 +02:00
|
|
|
from lbry.blob.blob_file import BlobFile
|
|
|
|
from lbry.stream.descriptor import StreamDescriptor
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
2018-11-30 22:11:23 +01:00
|
|
|
def calculate_effective_amount(amount: str, supports: typing.Optional[typing.List[typing.Dict]] = None) -> str:
|
|
|
|
return dewies_to_lbc(
|
|
|
|
lbc_to_dewies(amount) + sum([lbc_to_dewies(support['amount']) for support in supports])
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2019-11-15 20:55:49 +01:00
|
|
|
class StoredContentClaim:
|
2020-01-03 06:52:28 +01:00
|
|
|
def __init__(self, outpoint: Optional[str] = None, claim_id: Optional[str] = None, name: Optional[str] = None,
|
|
|
|
amount: Optional[int] = None, height: Optional[int] = None, serialized: Optional[str] = None,
|
|
|
|
channel_claim_id: Optional[str] = None, address: Optional[str] = None,
|
|
|
|
claim_sequence: Optional[int] = None, channel_name: Optional[str] = None):
|
2019-01-22 21:43:01 +01:00
|
|
|
self.claim_id = claim_id
|
|
|
|
self.outpoint = outpoint
|
|
|
|
self.claim_name = name
|
|
|
|
self.amount = amount
|
|
|
|
self.height = height
|
2019-03-20 06:46:23 +01:00
|
|
|
self.claim: typing.Optional[Claim] = None if not serialized else Claim.from_bytes(
|
|
|
|
binascii.unhexlify(serialized)
|
|
|
|
)
|
2019-01-22 21:43:01 +01:00
|
|
|
self.claim_address = address
|
|
|
|
self.claim_sequence = claim_sequence
|
|
|
|
self.channel_claim_id = channel_claim_id
|
|
|
|
self.channel_name = channel_name
|
|
|
|
|
|
|
|
@property
|
|
|
|
def txid(self) -> typing.Optional[str]:
|
|
|
|
return None if not self.outpoint else self.outpoint.split(":")[0]
|
|
|
|
|
|
|
|
@property
|
|
|
|
def nout(self) -> typing.Optional[int]:
|
|
|
|
return None if not self.outpoint else int(self.outpoint.split(":")[1])
|
|
|
|
|
|
|
|
def as_dict(self) -> typing.Dict:
|
|
|
|
return {
|
|
|
|
"name": self.claim_name,
|
|
|
|
"claim_id": self.claim_id,
|
|
|
|
"address": self.claim_address,
|
|
|
|
"claim_sequence": self.claim_sequence,
|
|
|
|
"value": self.claim,
|
|
|
|
"height": self.height,
|
|
|
|
"amount": dewies_to_lbc(self.amount),
|
|
|
|
"nout": self.nout,
|
|
|
|
"txid": self.txid,
|
|
|
|
"channel_claim_id": self.channel_claim_id,
|
|
|
|
"channel_name": self.channel_name
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2019-11-15 20:55:49 +01:00
|
|
|
def _get_content_claims(transaction: sqlite3.Connection, query: str,
|
|
|
|
source_hashes: typing.List[str]) -> typing.Dict[str, StoredContentClaim]:
|
|
|
|
claims = {}
|
|
|
|
for claim_info in _batched_select(transaction, query, source_hashes):
|
|
|
|
claims[claim_info[0]] = StoredContentClaim(*claim_info[1:])
|
|
|
|
return claims
|
|
|
|
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def get_claims_from_stream_hashes(transaction: sqlite3.Connection,
|
2019-11-15 20:55:49 +01:00
|
|
|
stream_hashes: typing.List[str]) -> typing.Dict[str, StoredContentClaim]:
|
2019-01-22 21:43:01 +01:00
|
|
|
query = (
|
|
|
|
"select content_claim.stream_hash, c.*, case when c.channel_claim_id is not null then "
|
|
|
|
" (select claim_name from claim where claim_id==c.channel_claim_id) "
|
|
|
|
" else null end as channel_name "
|
|
|
|
" from content_claim "
|
|
|
|
" inner join claim c on c.claim_outpoint=content_claim.claim_outpoint and content_claim.stream_hash in {}"
|
|
|
|
" order by c.rowid desc"
|
|
|
|
)
|
2019-11-15 20:55:49 +01:00
|
|
|
return _get_content_claims(transaction, query, stream_hashes)
|
|
|
|
|
|
|
|
|
|
|
|
def get_claims_from_torrent_info_hashes(transaction: sqlite3.Connection,
|
|
|
|
info_hashes: typing.List[str]) -> typing.Dict[str, StoredContentClaim]:
|
|
|
|
query = (
|
|
|
|
"select content_claim.bt_infohash, c.*, case when c.channel_claim_id is not null then "
|
|
|
|
" (select claim_name from claim where claim_id==c.channel_claim_id) "
|
|
|
|
" else null end as channel_name "
|
|
|
|
" from content_claim "
|
|
|
|
" inner join claim c on c.claim_outpoint=content_claim.claim_outpoint and content_claim.bt_infohash in {}"
|
|
|
|
" order by c.rowid desc"
|
|
|
|
)
|
|
|
|
return _get_content_claims(transaction, query, info_hashes)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
|
2019-02-09 01:51:13 +01:00
|
|
|
def _batched_select(transaction, query, parameters, batch_size=900):
|
|
|
|
for start_index in range(0, len(parameters), batch_size):
|
|
|
|
current_batch = parameters[start_index:start_index+batch_size]
|
2019-01-22 21:43:01 +01:00
|
|
|
bind = "({})".format(','.join(['?'] * len(current_batch)))
|
2019-10-02 18:38:56 +02:00
|
|
|
yield from transaction.execute(query.format(bind), current_batch)
|
2018-02-27 23:21:22 +01:00
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-11-14 00:55:11 +01:00
|
|
|
def _get_lbry_file_stream_dict(rowid, added_on, stream_hash, file_name, download_dir, data_rate, status,
|
|
|
|
sd_hash, stream_key, stream_name, suggested_file_name, claim, saved_file,
|
2020-02-07 16:34:47 +01:00
|
|
|
raw_content_fee, fully_reflected):
|
2019-11-14 00:55:11 +01:00
|
|
|
return {
|
|
|
|
"rowid": rowid,
|
|
|
|
"added_on": added_on,
|
|
|
|
"stream_hash": stream_hash,
|
|
|
|
"file_name": file_name, # hex
|
|
|
|
"download_directory": download_dir, # hex
|
|
|
|
"blob_data_rate": data_rate,
|
|
|
|
"status": status,
|
|
|
|
"sd_hash": sd_hash,
|
|
|
|
"key": stream_key,
|
|
|
|
"stream_name": stream_name, # hex
|
|
|
|
"suggested_file_name": suggested_file_name, # hex
|
|
|
|
"claim": claim,
|
|
|
|
"saved_file": bool(saved_file),
|
|
|
|
"content_fee": None if not raw_content_fee else Transaction(
|
|
|
|
binascii.unhexlify(raw_content_fee)
|
2020-02-07 16:34:47 +01:00
|
|
|
),
|
|
|
|
"fully_reflected": fully_reflected
|
2019-11-14 00:55:11 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2019-02-08 21:52:13 +01:00
|
|
|
def get_all_lbry_files(transaction: sqlite3.Connection) -> typing.List[typing.Dict]:
|
2019-02-09 01:51:13 +01:00
|
|
|
files = []
|
|
|
|
signed_claims = {}
|
2020-01-03 06:52:28 +01:00
|
|
|
for (rowid, stream_hash, _, file_name, download_dir, data_rate, status, saved_file, raw_content_fee,
|
2020-01-13 18:55:30 +01:00
|
|
|
added_on, _, sd_hash, stream_key, stream_name, suggested_file_name, *claim_args) in transaction.execute(
|
2020-02-07 16:34:47 +01:00
|
|
|
"select file.rowid, file.*, stream.*, c.*, "
|
|
|
|
" case when (SELECT 1 FROM reflected_stream r WHERE r.sd_hash=stream.sd_hash) "
|
|
|
|
" is null then 0 else 1 end as fully_reflected "
|
2020-01-03 06:52:28 +01:00
|
|
|
"from file inner join stream on file.stream_hash=stream.stream_hash "
|
|
|
|
"inner join content_claim cc on file.stream_hash=cc.stream_hash "
|
|
|
|
"inner join claim c on cc.claim_outpoint=c.claim_outpoint "
|
2020-01-13 18:55:30 +01:00
|
|
|
"order by c.rowid desc").fetchall():
|
2020-02-07 16:34:47 +01:00
|
|
|
claim_args, fully_reflected = tuple(claim_args[:-1]), claim_args[-1]
|
2019-11-15 20:55:49 +01:00
|
|
|
claim = StoredContentClaim(*claim_args)
|
2019-02-09 01:51:13 +01:00
|
|
|
if claim.channel_claim_id:
|
|
|
|
if claim.channel_claim_id not in signed_claims:
|
|
|
|
signed_claims[claim.channel_claim_id] = []
|
|
|
|
signed_claims[claim.channel_claim_id].append(claim)
|
|
|
|
files.append(
|
2019-11-14 00:55:11 +01:00
|
|
|
_get_lbry_file_stream_dict(
|
|
|
|
rowid, added_on, stream_hash, file_name, download_dir, data_rate, status,
|
|
|
|
sd_hash, stream_key, stream_name, suggested_file_name, claim, saved_file,
|
2020-02-07 16:34:47 +01:00
|
|
|
raw_content_fee, fully_reflected
|
2019-11-14 00:55:11 +01:00
|
|
|
)
|
2019-02-08 21:52:13 +01:00
|
|
|
)
|
2019-03-12 14:58:19 +01:00
|
|
|
for claim_name, claim_id in _batched_select(
|
|
|
|
transaction, "select c.claim_name, c.claim_id from claim c where c.claim_id in {}",
|
2019-09-17 16:26:38 +02:00
|
|
|
tuple(signed_claims.keys())):
|
2019-03-12 14:58:19 +01:00
|
|
|
for claim in signed_claims[claim_id]:
|
|
|
|
claim.channel_name = claim_name
|
2019-02-09 01:51:13 +01:00
|
|
|
return files
|
2019-02-08 21:52:13 +01:00
|
|
|
|
|
|
|
|
2019-02-15 00:17:50 +01:00
|
|
|
def store_stream(transaction: sqlite3.Connection, sd_blob: 'BlobFile', descriptor: 'StreamDescriptor'):
|
2019-05-06 21:49:52 +02:00
|
|
|
# add all blobs, except the last one, which is empty
|
|
|
|
transaction.executemany(
|
2021-09-08 16:55:21 +02:00
|
|
|
"insert or ignore into blob values (?, ?, ?, ?, ?, ?, ?, ?, ?)",
|
|
|
|
((blob.blob_hash, blob.length, 0, 0, "pending", 0, 0, blob.added_on, blob.is_mine)
|
2019-09-17 16:26:38 +02:00
|
|
|
for blob in (descriptor.blobs[:-1] if len(descriptor.blobs) > 1 else descriptor.blobs) + [sd_blob])
|
|
|
|
).fetchall()
|
2019-02-15 00:17:50 +01:00
|
|
|
# associate the blobs to the stream
|
|
|
|
transaction.execute("insert or ignore into stream values (?, ?, ?, ?, ?)",
|
|
|
|
(descriptor.stream_hash, sd_blob.blob_hash, descriptor.key,
|
|
|
|
binascii.hexlify(descriptor.stream_name.encode()).decode(),
|
2019-09-17 16:26:38 +02:00
|
|
|
binascii.hexlify(descriptor.suggested_file_name.encode()).decode())).fetchall()
|
2019-02-15 00:17:50 +01:00
|
|
|
# add the stream
|
|
|
|
transaction.executemany(
|
|
|
|
"insert or ignore into stream_blob values (?, ?, ?, ?)",
|
2019-09-17 16:26:38 +02:00
|
|
|
((descriptor.stream_hash, blob.blob_hash, blob.blob_num, blob.iv)
|
|
|
|
for blob in descriptor.blobs)
|
|
|
|
).fetchall()
|
2019-05-06 21:49:52 +02:00
|
|
|
# ensure should_announce is set regardless if insert was ignored
|
|
|
|
transaction.execute(
|
2022-02-09 02:10:42 +01:00
|
|
|
"update blob set should_announce=1 where blob_hash in (?)",
|
|
|
|
(sd_blob.blob_hash,)
|
2019-09-17 16:26:38 +02:00
|
|
|
).fetchall()
|
2019-02-15 00:17:50 +01:00
|
|
|
|
|
|
|
|
2019-02-14 18:33:48 +01:00
|
|
|
def delete_stream(transaction: sqlite3.Connection, descriptor: 'StreamDescriptor'):
|
2019-02-14 21:57:24 +01:00
|
|
|
blob_hashes = [(blob.blob_hash, ) for blob in descriptor.blobs[:-1]]
|
|
|
|
blob_hashes.append((descriptor.sd_hash, ))
|
2019-09-17 16:26:38 +02:00
|
|
|
transaction.execute("delete from content_claim where stream_hash=? ", (descriptor.stream_hash,)).fetchall()
|
|
|
|
transaction.execute("delete from file where stream_hash=? ", (descriptor.stream_hash,)).fetchall()
|
|
|
|
transaction.execute("delete from stream_blob where stream_hash=?", (descriptor.stream_hash,)).fetchall()
|
|
|
|
transaction.execute("delete from stream where stream_hash=? ", (descriptor.stream_hash,)).fetchall()
|
|
|
|
transaction.executemany("delete from blob where blob_hash=?", blob_hashes).fetchall()
|
2019-02-14 18:33:48 +01:00
|
|
|
|
|
|
|
|
2019-11-15 21:25:19 +01:00
|
|
|
def delete_torrent(transaction: sqlite3.Connection, bt_infohash: str):
|
|
|
|
transaction.execute("delete from content_claim where bt_infohash=?", (bt_infohash, )).fetchall()
|
|
|
|
transaction.execute("delete from torrent_tracker where bt_infohash=?", (bt_infohash,)).fetchall()
|
|
|
|
transaction.execute("delete from torrent_node where bt_infohash=?", (bt_infohash,)).fetchall()
|
|
|
|
transaction.execute("delete from torrent_http_seed where bt_infohash=?", (bt_infohash,)).fetchall()
|
|
|
|
transaction.execute("delete from file where bt_infohash=?", (bt_infohash,)).fetchall()
|
|
|
|
transaction.execute("delete from torrent where bt_infohash=?", (bt_infohash,)).fetchall()
|
|
|
|
|
|
|
|
|
2019-03-31 03:07:43 +02:00
|
|
|
def store_file(transaction: sqlite3.Connection, stream_hash: str, file_name: typing.Optional[str],
|
2019-05-07 20:30:35 +02:00
|
|
|
download_directory: typing.Optional[str], data_payment_rate: float, status: str,
|
2019-10-26 17:24:37 +02:00
|
|
|
content_fee: typing.Optional[Transaction], added_on: typing.Optional[int] = None) -> int:
|
2019-03-31 03:07:43 +02:00
|
|
|
if not file_name and not download_directory:
|
2019-05-07 20:30:35 +02:00
|
|
|
encoded_file_name, encoded_download_dir = None, None
|
2019-03-31 03:07:43 +02:00
|
|
|
else:
|
|
|
|
encoded_file_name = binascii.hexlify(file_name.encode()).decode()
|
|
|
|
encoded_download_dir = binascii.hexlify(download_directory.encode()).decode()
|
2019-10-26 17:24:37 +02:00
|
|
|
time_added = added_on or int(time.time())
|
2019-02-15 00:17:50 +01:00
|
|
|
transaction.execute(
|
2019-11-14 00:55:11 +01:00
|
|
|
"insert or replace into file values (?, NULL, ?, ?, ?, ?, ?, ?, ?)",
|
2019-05-07 20:30:35 +02:00
|
|
|
(stream_hash, encoded_file_name, encoded_download_dir, data_payment_rate, status,
|
|
|
|
1 if (file_name and download_directory and os.path.isfile(os.path.join(download_directory, file_name))) else 0,
|
2019-11-14 00:55:11 +01:00
|
|
|
None if not content_fee else binascii.hexlify(content_fee.raw).decode(), time_added)
|
2019-09-17 16:26:38 +02:00
|
|
|
).fetchall()
|
2019-05-07 20:30:35 +02:00
|
|
|
|
2019-02-15 22:44:31 +01:00
|
|
|
return transaction.execute("select rowid from file where stream_hash=?", (stream_hash, )).fetchone()[0]
|
2019-02-15 00:17:50 +01:00
|
|
|
|
|
|
|
|
2018-12-15 21:31:02 +01:00
|
|
|
class SQLiteStorage(SQLiteMixin):
|
2018-02-12 19:36:35 +01:00
|
|
|
CREATE_TABLES_QUERY = """
|
|
|
|
pragma foreign_keys=on;
|
|
|
|
pragma journal_mode=WAL;
|
2019-01-22 21:43:01 +01:00
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
create table if not exists blob (
|
|
|
|
blob_hash char(96) primary key not null,
|
|
|
|
blob_length integer not null,
|
|
|
|
next_announce_time integer not null,
|
|
|
|
should_announce integer not null default 0,
|
2018-03-27 20:58:29 +02:00
|
|
|
status text not null,
|
2018-03-29 00:47:37 +02:00
|
|
|
last_announced_time integer,
|
2021-09-08 16:55:21 +02:00
|
|
|
single_announce integer,
|
|
|
|
added_on integer not null,
|
|
|
|
is_mine integer not null default 0
|
2018-02-12 19:36:35 +01:00
|
|
|
);
|
2019-01-22 21:43:01 +01:00
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
create table if not exists stream (
|
|
|
|
stream_hash char(96) not null primary key,
|
2018-02-12 21:10:06 +01:00
|
|
|
sd_hash char(96) not null references blob,
|
2018-02-12 19:36:35 +01:00
|
|
|
stream_key text not null,
|
|
|
|
stream_name text not null,
|
2018-02-12 21:10:06 +01:00
|
|
|
suggested_filename text not null
|
2018-02-12 19:36:35 +01:00
|
|
|
);
|
2019-01-22 21:43:01 +01:00
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
create table if not exists stream_blob (
|
2018-02-12 21:10:06 +01:00
|
|
|
stream_hash char(96) not null references stream,
|
|
|
|
blob_hash char(96) references blob,
|
2018-02-12 19:36:35 +01:00
|
|
|
position integer not null,
|
|
|
|
iv char(32) not null,
|
2018-02-12 21:10:06 +01:00
|
|
|
primary key (stream_hash, blob_hash)
|
2018-02-12 19:36:35 +01:00
|
|
|
);
|
2019-01-22 21:43:01 +01:00
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
create table if not exists claim (
|
|
|
|
claim_outpoint text not null primary key,
|
|
|
|
claim_id char(40) not null,
|
|
|
|
claim_name text not null,
|
|
|
|
amount integer not null,
|
|
|
|
height integer not null,
|
|
|
|
serialized_metadata blob not null,
|
|
|
|
channel_claim_id text,
|
|
|
|
address text not null,
|
|
|
|
claim_sequence integer not null
|
|
|
|
);
|
|
|
|
|
2019-11-14 00:55:11 +01:00
|
|
|
create table if not exists torrent (
|
|
|
|
bt_infohash char(20) not null primary key,
|
|
|
|
tracker text,
|
|
|
|
length integer not null,
|
|
|
|
name text not null
|
|
|
|
);
|
|
|
|
|
|
|
|
create table if not exists torrent_node ( -- BEP-0005
|
|
|
|
bt_infohash char(20) not null references torrent,
|
|
|
|
host text not null,
|
|
|
|
port integer not null
|
|
|
|
);
|
|
|
|
|
|
|
|
create table if not exists torrent_tracker ( -- BEP-0012
|
|
|
|
bt_infohash char(20) not null references torrent,
|
|
|
|
tracker text not null
|
|
|
|
);
|
|
|
|
|
|
|
|
create table if not exists torrent_http_seed ( -- BEP-0017
|
|
|
|
bt_infohash char(20) not null references torrent,
|
|
|
|
http_seed text not null
|
|
|
|
);
|
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
create table if not exists file (
|
2019-11-14 00:55:11 +01:00
|
|
|
stream_hash char(96) references stream,
|
|
|
|
bt_infohash char(20) references torrent,
|
2019-05-07 20:30:35 +02:00
|
|
|
file_name text,
|
|
|
|
download_directory text,
|
2018-02-12 19:36:35 +01:00
|
|
|
blob_data_rate real not null,
|
2019-05-07 20:30:35 +02:00
|
|
|
status text not null,
|
|
|
|
saved_file integer not null,
|
2019-10-11 01:46:00 +02:00
|
|
|
content_fee text,
|
2019-10-26 17:24:37 +02:00
|
|
|
added_on integer not null
|
2018-02-12 19:36:35 +01:00
|
|
|
);
|
2019-01-22 21:43:01 +01:00
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
create table if not exists content_claim (
|
2019-11-14 00:55:11 +01:00
|
|
|
stream_hash char(96) references stream,
|
|
|
|
bt_infohash char(20) references torrent,
|
|
|
|
claim_outpoint text unique not null references claim
|
2018-02-12 19:36:35 +01:00
|
|
|
);
|
2019-01-22 21:43:01 +01:00
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
create table if not exists support (
|
|
|
|
support_outpoint text not null primary key,
|
|
|
|
claim_id text not null,
|
|
|
|
amount integer not null,
|
|
|
|
address text not null
|
|
|
|
);
|
2019-01-22 21:43:01 +01:00
|
|
|
|
2018-05-08 19:51:02 +02:00
|
|
|
create table if not exists reflected_stream (
|
|
|
|
sd_hash text not null,
|
|
|
|
reflector_address text not null,
|
|
|
|
timestamp integer,
|
|
|
|
primary key (sd_hash, reflector_address)
|
|
|
|
);
|
2019-10-31 13:57:30 +01:00
|
|
|
|
|
|
|
create table if not exists peer (
|
2019-11-11 15:32:24 +01:00
|
|
|
node_id char(96) not null primary key,
|
2019-10-31 13:57:30 +01:00
|
|
|
address text not null,
|
|
|
|
udp_port integer not null,
|
|
|
|
tcp_port integer,
|
2019-11-11 15:32:24 +01:00
|
|
|
unique (address, udp_port)
|
2019-10-31 13:57:30 +01:00
|
|
|
);
|
2021-11-03 19:13:37 +01:00
|
|
|
create index if not exists blob_data on blob(blob_hash, blob_length, is_mine);
|
2018-10-10 05:58:32 +02:00
|
|
|
"""
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-02-03 21:09:38 +01:00
|
|
|
def __init__(self, conf: Config, path, loop=None, time_getter: typing.Optional[typing.Callable[[], float]] = None):
|
2018-12-15 21:31:02 +01:00
|
|
|
super().__init__(path)
|
2019-01-21 21:55:50 +01:00
|
|
|
self.conf = conf
|
2018-12-15 21:31:02 +01:00
|
|
|
self.content_claim_callbacks = {}
|
2019-01-08 02:23:18 +01:00
|
|
|
self.loop = loop or asyncio.get_event_loop()
|
2019-02-03 21:09:38 +01:00
|
|
|
self.time_getter = time_getter or time.time
|
2018-12-15 21:31:02 +01:00
|
|
|
|
|
|
|
async def run_and_return_one_or_none(self, query, *args):
|
|
|
|
for row in await self.db.execute_fetchall(query, args):
|
2019-01-07 08:52:53 +01:00
|
|
|
if len(row) == 1:
|
|
|
|
return row[0]
|
2018-12-15 21:31:02 +01:00
|
|
|
return row
|
|
|
|
|
|
|
|
async def run_and_return_list(self, query, *args):
|
|
|
|
rows = list(await self.db.execute_fetchall(query, args))
|
|
|
|
return [col[0] for col in rows] if rows else []
|
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
# # # # # # # # # blob functions # # # # # # # # #
|
|
|
|
|
2021-09-08 16:55:21 +02:00
|
|
|
async def add_blobs(self, *blob_hashes_and_lengths: typing.Tuple[str, int, int, int], finished=False):
|
2019-04-15 22:14:19 +02:00
|
|
|
def _add_blobs(transaction: sqlite3.Connection):
|
|
|
|
transaction.executemany(
|
2021-09-08 16:55:21 +02:00
|
|
|
"insert or ignore into blob values (?, ?, ?, ?, ?, ?, ?, ?, ?)",
|
2019-09-17 16:26:38 +02:00
|
|
|
(
|
2021-09-08 16:55:21 +02:00
|
|
|
(blob_hash, length, 0, 0, "pending" if not finished else "finished", 0, 0, added_on, is_mine)
|
|
|
|
for blob_hash, length, added_on, is_mine in blob_hashes_and_lengths
|
2019-09-17 16:26:38 +02:00
|
|
|
)
|
|
|
|
).fetchall()
|
2019-04-15 22:14:19 +02:00
|
|
|
if finished:
|
|
|
|
transaction.executemany(
|
2019-09-17 16:26:38 +02:00
|
|
|
"update blob set status='finished' where blob.blob_hash=?", (
|
2021-09-08 16:55:21 +02:00
|
|
|
(blob_hash, ) for blob_hash, _, _, _ in blob_hashes_and_lengths
|
2019-09-17 16:26:38 +02:00
|
|
|
)
|
|
|
|
).fetchall()
|
2019-04-15 22:14:19 +02:00
|
|
|
return await self.db.run(_add_blobs)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def get_blob_status(self, blob_hash: str):
|
2018-02-12 19:36:35 +01:00
|
|
|
return self.run_and_return_one_or_none(
|
|
|
|
"select status from blob where blob_hash=?", blob_hash
|
|
|
|
)
|
|
|
|
|
2021-10-22 07:55:59 +02:00
|
|
|
def set_announce(self, *blob_hashes):
|
|
|
|
return self.db.execute_fetchall(
|
|
|
|
"update blob set should_announce=1 where blob_hash in (?, ?)", blob_hashes
|
|
|
|
)
|
|
|
|
|
2019-02-03 21:09:38 +01:00
|
|
|
def update_last_announced_blobs(self, blob_hashes: typing.List[str]):
|
2019-01-22 21:43:01 +01:00
|
|
|
def _update_last_announced_blobs(transaction: sqlite3.Connection):
|
2019-02-03 21:09:38 +01:00
|
|
|
last_announced = self.time_getter()
|
2019-01-22 21:43:01 +01:00
|
|
|
return transaction.executemany(
|
|
|
|
"update blob set next_announce_time=?, last_announced_time=?, single_announce=0 "
|
|
|
|
"where blob_hash=?",
|
2020-01-03 04:57:28 +01:00
|
|
|
((int(last_announced + (DATA_EXPIRATION / 2)), int(last_announced), blob_hash)
|
2019-09-17 16:26:38 +02:00
|
|
|
for blob_hash in blob_hashes)
|
|
|
|
).fetchall()
|
2019-01-22 21:43:01 +01:00
|
|
|
return self.db.run(_update_last_announced_blobs)
|
2018-03-27 21:12:44 +02:00
|
|
|
|
2018-03-29 00:47:37 +02:00
|
|
|
def should_single_announce_blobs(self, blob_hashes, immediate=False):
|
|
|
|
def set_single_announce(transaction):
|
2019-02-03 21:09:38 +01:00
|
|
|
now = int(self.time_getter())
|
2018-03-29 00:47:37 +02:00
|
|
|
for blob_hash in blob_hashes:
|
|
|
|
if immediate:
|
|
|
|
transaction.execute(
|
|
|
|
"update blob set single_announce=1, next_announce_time=? "
|
|
|
|
"where blob_hash=? and status='finished'", (int(now), blob_hash)
|
2019-09-17 16:26:38 +02:00
|
|
|
).fetchall()
|
2018-03-29 00:47:37 +02:00
|
|
|
else:
|
|
|
|
transaction.execute(
|
2019-01-22 21:43:01 +01:00
|
|
|
"update blob set single_announce=1 where blob_hash=? and status='finished'", (blob_hash,)
|
2019-09-17 16:26:38 +02:00
|
|
|
).fetchall()
|
2018-12-15 21:31:02 +01:00
|
|
|
return self.db.run(set_single_announce)
|
2018-03-29 00:47:37 +02:00
|
|
|
|
2018-03-27 21:12:44 +02:00
|
|
|
def get_blobs_to_announce(self):
|
2018-02-12 19:36:35 +01:00
|
|
|
def get_and_update(transaction):
|
2019-02-03 21:09:38 +01:00
|
|
|
timestamp = int(self.time_getter())
|
2019-01-28 15:51:02 +01:00
|
|
|
if self.conf.announce_head_and_sd_only:
|
2018-02-12 19:36:35 +01:00
|
|
|
r = transaction.execute(
|
|
|
|
"select blob_hash from blob "
|
2018-03-29 00:47:37 +02:00
|
|
|
"where blob_hash is not null and "
|
2019-02-08 22:29:55 +01:00
|
|
|
"(should_announce=1 or single_announce=1) and next_announce_time<? and status='finished' "
|
|
|
|
"order by next_announce_time asc limit ?",
|
|
|
|
(timestamp, int(self.conf.concurrent_blob_announcers * 10))
|
2019-09-17 16:26:38 +02:00
|
|
|
).fetchall()
|
2018-02-12 19:36:35 +01:00
|
|
|
else:
|
|
|
|
r = transaction.execute(
|
2018-03-08 22:46:39 +01:00
|
|
|
"select blob_hash from blob where blob_hash is not null "
|
2019-02-08 22:29:55 +01:00
|
|
|
"and next_announce_time<? and status='finished' "
|
|
|
|
"order by next_announce_time asc limit ?",
|
|
|
|
(timestamp, int(self.conf.concurrent_blob_announcers * 10))
|
2019-09-17 16:26:38 +02:00
|
|
|
).fetchall()
|
|
|
|
return [b[0] for b in r]
|
2018-12-15 21:31:02 +01:00
|
|
|
return self.db.run(get_and_update)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
def delete_blobs_from_db(self, blob_hashes):
|
|
|
|
def delete_blobs(transaction):
|
2019-01-22 21:43:01 +01:00
|
|
|
transaction.executemany(
|
2019-09-17 16:26:38 +02:00
|
|
|
"delete from blob where blob_hash=?;", ((blob_hash,) for blob_hash in blob_hashes)
|
|
|
|
).fetchall()
|
2019-02-14 21:42:12 +01:00
|
|
|
return self.db.run_with_foreign_keys_disabled(delete_blobs)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
def get_all_blob_hashes(self):
|
|
|
|
return self.run_and_return_list("select blob_hash from blob")
|
|
|
|
|
2021-10-24 23:19:06 +02:00
|
|
|
async def get_stored_blobs(self, is_mine: bool, is_network_blob=False):
|
2021-09-10 16:53:52 +02:00
|
|
|
is_mine = 1 if is_mine else 0
|
2021-10-24 23:19:06 +02:00
|
|
|
if is_network_blob:
|
2021-10-22 07:43:40 +02:00
|
|
|
return await self.db.execute_fetchall(
|
|
|
|
"select blob.blob_hash, blob.blob_length, blob.added_on "
|
|
|
|
"from blob left join stream_blob using (blob_hash) "
|
2022-06-08 16:08:42 +02:00
|
|
|
"where stream_blob.stream_hash is null and blob.is_mine=? and blob.status='finished'"
|
2022-03-21 08:26:27 +01:00
|
|
|
"order by blob.blob_length desc, blob.added_on asc",
|
2021-10-22 07:43:40 +02:00
|
|
|
(is_mine,)
|
|
|
|
)
|
|
|
|
|
2021-10-20 07:18:34 +02:00
|
|
|
sd_blobs = await self.db.execute_fetchall(
|
|
|
|
"select blob.blob_hash, blob.blob_length, blob.added_on "
|
|
|
|
"from blob join stream on blob.blob_hash=stream.sd_hash join file using (stream_hash) "
|
|
|
|
"where blob.is_mine=? order by blob.added_on asc",
|
2021-09-10 16:53:52 +02:00
|
|
|
(is_mine,)
|
2021-09-08 16:55:21 +02:00
|
|
|
)
|
2021-11-03 02:34:15 +01:00
|
|
|
content_blobs = await self.db.execute_fetchall(
|
2021-10-20 07:18:34 +02:00
|
|
|
"select blob.blob_hash, blob.blob_length, blob.added_on "
|
|
|
|
"from blob join stream_blob using (blob_hash) cross join stream using (stream_hash)"
|
2022-06-08 16:08:42 +02:00
|
|
|
"cross join file using (stream_hash)"
|
|
|
|
"where blob.is_mine=? and blob.status='finished' order by blob.added_on asc, blob.blob_length asc",
|
2021-10-20 07:18:34 +02:00
|
|
|
(is_mine,)
|
|
|
|
)
|
2021-11-03 02:34:15 +01:00
|
|
|
return content_blobs + sd_blobs
|
|
|
|
|
|
|
|
async def get_stored_blob_disk_usage(self):
|
|
|
|
total, network_size, content_size, private_size = await self.db.execute_fetchone("""
|
|
|
|
select coalesce(sum(blob_length), 0) as total,
|
|
|
|
coalesce(sum(case when
|
|
|
|
stream_blob.stream_hash is null
|
|
|
|
then blob_length else 0 end), 0) as network_storage,
|
|
|
|
coalesce(sum(case when
|
|
|
|
stream_blob.blob_hash is not null and is_mine=0
|
|
|
|
then blob_length else 0 end), 0) as content_storage,
|
|
|
|
coalesce(sum(case when
|
|
|
|
is_mine=1
|
|
|
|
then blob_length else 0 end), 0) as private_storage
|
2022-06-08 16:08:42 +02:00
|
|
|
from blob left join stream_blob using (blob_hash)
|
|
|
|
where blob_hash not in (select sd_hash from stream) and blob.status="finished"
|
2021-11-03 02:34:15 +01:00
|
|
|
""")
|
|
|
|
return {
|
|
|
|
'network_storage': network_size,
|
|
|
|
'content_storage': content_size,
|
|
|
|
'private_storage': private_size,
|
|
|
|
'total': total
|
|
|
|
}
|
2021-09-08 16:55:21 +02:00
|
|
|
|
2021-09-10 16:53:52 +02:00
|
|
|
async def update_blob_ownership(self, sd_hash, is_mine: bool):
|
|
|
|
is_mine = 1 if is_mine else 0
|
2021-09-08 16:55:21 +02:00
|
|
|
await self.db.execute_fetchall(
|
|
|
|
"update blob set is_mine = ? where blob_hash in ("
|
2021-09-10 16:53:52 +02:00
|
|
|
" select blob_hash from blob natural join stream_blob natural join stream where sd_hash = ?"
|
2021-09-15 16:37:08 +02:00
|
|
|
") OR blob_hash = ?", (is_mine, sd_hash, sd_hash)
|
2021-09-08 16:55:21 +02:00
|
|
|
)
|
|
|
|
|
2019-02-14 18:36:18 +01:00
|
|
|
def sync_missing_blobs(self, blob_files: typing.Set[str]) -> typing.Awaitable[typing.Set[str]]:
|
|
|
|
def _sync_blobs(transaction: sqlite3.Connection) -> typing.Set[str]:
|
2019-09-17 16:25:40 +02:00
|
|
|
finished_blob_hashes = tuple(
|
|
|
|
blob_hash for (blob_hash, ) in transaction.execute(
|
|
|
|
"select blob_hash from blob where status='finished'"
|
|
|
|
).fetchall()
|
|
|
|
)
|
|
|
|
finished_blobs_set = set(finished_blob_hashes)
|
|
|
|
to_update_set = finished_blobs_set.difference(blob_files)
|
2019-02-14 18:36:18 +01:00
|
|
|
transaction.executemany(
|
|
|
|
"update blob set status='pending' where blob_hash=?",
|
2019-09-17 16:25:40 +02:00
|
|
|
((blob_hash, ) for blob_hash in to_update_set)
|
|
|
|
).fetchall()
|
|
|
|
return blob_files.intersection(finished_blobs_set)
|
2019-02-14 18:36:18 +01:00
|
|
|
return self.db.run(_sync_blobs)
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
# # # # # # # # # stream functions # # # # # # # # #
|
2018-03-08 22:46:39 +01:00
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
async def stream_exists(self, sd_hash: str) -> bool:
|
|
|
|
streams = await self.run_and_return_one_or_none("select stream_hash from stream where sd_hash=?", sd_hash)
|
|
|
|
return streams is not None
|
2018-03-08 22:46:39 +01:00
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
async def file_exists(self, sd_hash: str) -> bool:
|
|
|
|
streams = await self.run_and_return_one_or_none("select f.stream_hash from file f "
|
|
|
|
"inner join stream s on "
|
|
|
|
"s.stream_hash=f.stream_hash and s.sd_hash=?", sd_hash)
|
|
|
|
return streams is not None
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def store_stream(self, sd_blob: 'BlobFile', descriptor: 'StreamDescriptor'):
|
2019-02-15 00:17:50 +01:00
|
|
|
return self.db.run(store_stream, sd_blob, descriptor)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-02-15 00:16:16 +01:00
|
|
|
def get_blobs_for_stream(self, stream_hash, only_completed=False) -> typing.Awaitable[typing.List[BlobInfo]]:
|
2019-01-22 23:44:25 +01:00
|
|
|
def _get_blobs_for_stream(transaction):
|
|
|
|
crypt_blob_infos = []
|
|
|
|
stream_blobs = transaction.execute(
|
2022-03-23 17:41:02 +01:00
|
|
|
"select s.blob_hash, s.position, s.iv, b.added_on "
|
|
|
|
"from stream_blob s left outer join blob b on b.blob_hash=s.blob_hash where stream_hash=? "
|
2019-02-15 00:16:16 +01:00
|
|
|
"order by position asc", (stream_hash, )
|
2019-01-22 23:44:25 +01:00
|
|
|
).fetchall()
|
|
|
|
if only_completed:
|
|
|
|
lengths = transaction.execute(
|
|
|
|
"select b.blob_hash, b.blob_length from blob b "
|
|
|
|
"inner join stream_blob s ON b.blob_hash=s.blob_hash and b.status='finished' and s.stream_hash=?",
|
|
|
|
(stream_hash, )
|
|
|
|
).fetchall()
|
|
|
|
else:
|
|
|
|
lengths = transaction.execute(
|
|
|
|
"select b.blob_hash, b.blob_length from blob b "
|
|
|
|
"inner join stream_blob s ON b.blob_hash=s.blob_hash and s.stream_hash=?",
|
|
|
|
(stream_hash, )
|
|
|
|
).fetchall()
|
|
|
|
|
|
|
|
blob_length_dict = {}
|
|
|
|
for blob_hash, length in lengths:
|
|
|
|
blob_length_dict[blob_hash] = length
|
|
|
|
|
2022-03-23 17:41:02 +01:00
|
|
|
current_time = time.time()
|
|
|
|
for blob_hash, position, iv, added_on in stream_blobs:
|
2019-01-22 23:44:25 +01:00
|
|
|
blob_length = blob_length_dict.get(blob_hash, 0)
|
2022-03-23 17:41:02 +01:00
|
|
|
crypt_blob_infos.append(BlobInfo(position, blob_length, iv, added_on or current_time, blob_hash))
|
2019-02-15 00:16:16 +01:00
|
|
|
if not blob_hash:
|
|
|
|
break
|
2019-01-22 23:44:25 +01:00
|
|
|
return crypt_blob_infos
|
|
|
|
return self.db.run(_get_blobs_for_stream)
|
|
|
|
|
|
|
|
def get_sd_blob_hash_for_stream(self, stream_hash):
|
|
|
|
return self.run_and_return_one_or_none(
|
|
|
|
"select sd_hash from stream where stream_hash=?", stream_hash
|
|
|
|
)
|
|
|
|
|
|
|
|
def get_stream_hash_for_sd_hash(self, sd_blob_hash):
|
|
|
|
return self.run_and_return_one_or_none(
|
|
|
|
"select stream_hash from stream where sd_hash = ?", sd_blob_hash
|
|
|
|
)
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def delete_stream(self, descriptor: 'StreamDescriptor'):
|
2019-02-14 18:33:48 +01:00
|
|
|
return self.db.run_with_foreign_keys_disabled(delete_stream, descriptor)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-11-15 21:25:19 +01:00
|
|
|
async def delete_torrent(self, bt_infohash: str):
|
|
|
|
return await self.db.run(delete_torrent, bt_infohash)
|
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
# # # # # # # # # file stuff # # # # # # # # #
|
|
|
|
|
2019-05-07 20:30:35 +02:00
|
|
|
def save_downloaded_file(self, stream_hash: str, file_name: typing.Optional[str],
|
|
|
|
download_directory: typing.Optional[str], data_payment_rate: float,
|
2019-10-11 01:46:00 +02:00
|
|
|
content_fee: typing.Optional[Transaction] = None,
|
2019-10-26 17:24:37 +02:00
|
|
|
added_on: typing.Optional[int] = None) -> typing.Awaitable[int]:
|
2019-01-22 21:43:01 +01:00
|
|
|
return self.save_published_file(
|
2019-05-07 20:30:35 +02:00
|
|
|
stream_hash, file_name, download_directory, data_payment_rate, status="running",
|
2019-10-26 17:24:37 +02:00
|
|
|
content_fee=content_fee, added_on=added_on
|
2018-02-12 19:36:35 +01:00
|
|
|
)
|
|
|
|
|
2019-03-31 03:07:43 +02:00
|
|
|
def save_published_file(self, stream_hash: str, file_name: typing.Optional[str],
|
|
|
|
download_directory: typing.Optional[str], data_payment_rate: float,
|
2019-05-07 20:30:35 +02:00
|
|
|
status: str = "finished",
|
2019-10-11 01:46:00 +02:00
|
|
|
content_fee: typing.Optional[Transaction] = None,
|
2019-10-26 17:24:37 +02:00
|
|
|
added_on: typing.Optional[int] = None) -> typing.Awaitable[int]:
|
2019-05-07 20:30:35 +02:00
|
|
|
return self.db.run(store_file, stream_hash, file_name, download_directory, data_payment_rate, status,
|
2019-10-26 17:24:37 +02:00
|
|
|
content_fee, added_on)
|
2019-05-07 20:30:35 +02:00
|
|
|
|
|
|
|
async def update_manually_removed_files_since_last_run(self):
|
|
|
|
"""
|
|
|
|
Update files that have been removed from the downloads directory since the last run
|
|
|
|
"""
|
|
|
|
def update_manually_removed_files(transaction: sqlite3.Connection):
|
2020-01-13 18:46:00 +01:00
|
|
|
files = {}
|
|
|
|
query = "select stream_hash, download_directory, file_name from file where saved_file=1 " \
|
2019-11-14 00:55:11 +01:00
|
|
|
"and stream_hash is not null"
|
2020-01-13 18:46:00 +01:00
|
|
|
for (stream_hash, download_directory, file_name) in transaction.execute(query).fetchall():
|
|
|
|
if download_directory and file_name:
|
|
|
|
files[stream_hash] = download_directory, file_name
|
|
|
|
return files
|
|
|
|
|
|
|
|
def detect_removed(files):
|
|
|
|
return [
|
|
|
|
stream_hash for stream_hash, (download_directory, file_name) in files.items()
|
|
|
|
if not os.path.isfile(os.path.join(binascii.unhexlify(download_directory).decode(),
|
|
|
|
binascii.unhexlify(file_name).decode()))
|
|
|
|
]
|
|
|
|
|
|
|
|
def update_db_removed(transaction: sqlite3.Connection, removed):
|
|
|
|
query = "update file set file_name=null, download_directory=null, saved_file=0 where stream_hash in {}"
|
|
|
|
for cur in _batched_select(transaction, query, removed):
|
|
|
|
cur.fetchall()
|
|
|
|
|
|
|
|
stream_and_file = await self.db.run(update_manually_removed_files)
|
|
|
|
removed = await self.loop.run_in_executor(None, detect_removed, stream_and_file)
|
|
|
|
if removed:
|
|
|
|
await self.db.run(update_db_removed, removed)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-02-09 01:51:13 +01:00
|
|
|
def get_all_lbry_files(self) -> typing.Awaitable[typing.List[typing.Dict]]:
|
2019-02-08 21:52:13 +01:00
|
|
|
return self.db.run(get_all_lbry_files)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def change_file_status(self, stream_hash: str, new_status: str):
|
2019-02-15 00:17:50 +01:00
|
|
|
log.debug("update file status %s -> %s", stream_hash, new_status)
|
2019-09-17 16:26:38 +02:00
|
|
|
return self.db.execute_fetchall("update file set status=? where stream_hash=?", (new_status, stream_hash))
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2021-09-15 16:37:08 +02:00
|
|
|
def stop_all_files(self):
|
|
|
|
log.debug("stopping all files")
|
|
|
|
return self.db.execute_fetchall("update file set status=?", ("stopped",))
|
|
|
|
|
2019-04-23 15:53:17 +02:00
|
|
|
async def change_file_download_dir_and_file_name(self, stream_hash: str, download_dir: typing.Optional[str],
|
|
|
|
file_name: typing.Optional[str]):
|
|
|
|
if not file_name or not download_dir:
|
2019-05-07 20:30:35 +02:00
|
|
|
encoded_file_name, encoded_download_dir = None, None
|
2019-04-23 15:53:17 +02:00
|
|
|
else:
|
|
|
|
encoded_file_name = binascii.hexlify(file_name.encode()).decode()
|
|
|
|
encoded_download_dir = binascii.hexlify(download_dir.encode()).decode()
|
2019-09-17 16:26:38 +02:00
|
|
|
return await self.db.execute_fetchall("update file set download_directory=?, file_name=? where stream_hash=?", (
|
2019-04-23 15:53:17 +02:00
|
|
|
encoded_download_dir, encoded_file_name, stream_hash,
|
2019-02-01 21:46:31 +01:00
|
|
|
))
|
|
|
|
|
2019-05-07 20:30:35 +02:00
|
|
|
async def save_content_fee(self, stream_hash: str, content_fee: Transaction):
|
2019-09-17 16:26:38 +02:00
|
|
|
return await self.db.execute_fetchall("update file set content_fee=? where stream_hash=?", (
|
2019-05-07 20:30:35 +02:00
|
|
|
binascii.hexlify(content_fee.raw), stream_hash,
|
|
|
|
))
|
|
|
|
|
|
|
|
async def set_saved_file(self, stream_hash: str):
|
2019-09-17 16:26:38 +02:00
|
|
|
return await self.db.execute_fetchall("update file set saved_file=1 where stream_hash=?", (
|
2019-05-07 20:30:35 +02:00
|
|
|
stream_hash,
|
|
|
|
))
|
|
|
|
|
|
|
|
async def clear_saved_file(self, stream_hash: str):
|
2019-09-17 16:26:38 +02:00
|
|
|
return await self.db.execute_fetchall("update file set saved_file=0 where stream_hash=?", (
|
2019-05-07 20:30:35 +02:00
|
|
|
stream_hash,
|
|
|
|
))
|
|
|
|
|
|
|
|
async def recover_streams(self, descriptors_and_sds: typing.List[typing.Tuple['StreamDescriptor', 'BlobFile',
|
|
|
|
typing.Optional[Transaction]]],
|
2019-02-15 00:19:01 +01:00
|
|
|
download_directory: str):
|
|
|
|
def _recover(transaction: sqlite3.Connection):
|
2019-05-07 20:30:35 +02:00
|
|
|
stream_hashes = [x[0].stream_hash for x in descriptors_and_sds]
|
|
|
|
for descriptor, sd_blob, content_fee in descriptors_and_sds:
|
2019-02-15 00:19:01 +01:00
|
|
|
content_claim = transaction.execute(
|
|
|
|
"select * from content_claim where stream_hash=?", (descriptor.stream_hash, )
|
|
|
|
).fetchone()
|
|
|
|
delete_stream(transaction, descriptor) # this will also delete the content claim
|
|
|
|
store_stream(transaction, sd_blob, descriptor)
|
|
|
|
store_file(transaction, descriptor.stream_hash, os.path.basename(descriptor.suggested_file_name),
|
2019-05-07 20:30:35 +02:00
|
|
|
download_directory, 0.0, 'stopped', content_fee=content_fee)
|
2019-02-15 00:19:01 +01:00
|
|
|
if content_claim:
|
2019-11-14 00:55:11 +01:00
|
|
|
transaction.execute("insert or ignore into content_claim values (?, ?, ?)", content_claim)
|
2019-02-15 00:19:01 +01:00
|
|
|
transaction.executemany(
|
|
|
|
"update file set status='stopped' where stream_hash=?",
|
2019-09-17 16:26:38 +02:00
|
|
|
((stream_hash, ) for stream_hash in stream_hashes)
|
|
|
|
).fetchall()
|
2019-02-15 00:19:01 +01:00
|
|
|
download_dir = binascii.hexlify(self.conf.download_dir.encode()).decode()
|
|
|
|
transaction.executemany(
|
2021-08-21 04:36:35 +02:00
|
|
|
"update file set download_directory=? where stream_hash=?",
|
2019-09-17 16:26:38 +02:00
|
|
|
((download_dir, stream_hash) for stream_hash in stream_hashes)
|
|
|
|
).fetchall()
|
2019-02-15 00:19:01 +01:00
|
|
|
await self.db.run_with_foreign_keys_disabled(_recover)
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def get_all_stream_hashes(self):
|
|
|
|
return self.run_and_return_list("select stream_hash from stream")
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
# # # # # # # # # support functions # # # # # # # # #
|
|
|
|
|
2019-04-20 01:42:35 +02:00
|
|
|
def save_supports(self, claim_id_to_supports: dict):
|
2018-02-12 19:36:35 +01:00
|
|
|
# TODO: add 'address' to support items returned for a claim from lbrycrdd and lbryum-server
|
|
|
|
def _save_support(transaction):
|
2019-04-20 01:42:35 +02:00
|
|
|
bind = "({})".format(','.join(['?'] * len(claim_id_to_supports)))
|
2019-09-17 16:26:38 +02:00
|
|
|
transaction.execute(
|
|
|
|
f"delete from support where claim_id in {bind}", tuple(claim_id_to_supports.keys())
|
|
|
|
).fetchall()
|
2019-04-20 01:42:35 +02:00
|
|
|
for claim_id, supports in claim_id_to_supports.items():
|
|
|
|
for support in supports:
|
|
|
|
transaction.execute(
|
|
|
|
"insert into support values (?, ?, ?, ?)",
|
|
|
|
("%s:%i" % (support['txid'], support['nout']), claim_id, lbc_to_dewies(support['amount']),
|
|
|
|
support.get('address', ""))
|
2019-09-17 16:26:38 +02:00
|
|
|
).fetchall()
|
2018-12-15 21:31:02 +01:00
|
|
|
return self.db.run(_save_support)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2018-05-09 00:46:29 +02:00
|
|
|
def get_supports(self, *claim_ids):
|
2018-02-12 19:36:35 +01:00
|
|
|
def _format_support(outpoint, supported_id, amount, address):
|
|
|
|
return {
|
|
|
|
"txid": outpoint.split(":")[0],
|
|
|
|
"nout": int(outpoint.split(":")[1]),
|
|
|
|
"claim_id": supported_id,
|
2018-11-30 22:11:23 +01:00
|
|
|
"amount": dewies_to_lbc(amount),
|
2018-02-12 19:36:35 +01:00
|
|
|
"address": address,
|
|
|
|
}
|
|
|
|
|
|
|
|
def _get_supports(transaction):
|
|
|
|
return [
|
|
|
|
_format_support(*support_info)
|
2018-11-07 14:25:46 +01:00
|
|
|
for support_info in _batched_select(
|
|
|
|
transaction,
|
|
|
|
"select * from support where claim_id in {}",
|
2019-09-17 16:26:38 +02:00
|
|
|
claim_ids
|
2018-11-07 14:25:46 +01:00
|
|
|
)
|
2018-02-12 19:36:35 +01:00
|
|
|
]
|
|
|
|
|
2018-12-15 21:31:02 +01:00
|
|
|
return self.db.run(_get_supports)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
# # # # # # # # # claim functions # # # # # # # # #
|
|
|
|
|
2018-12-15 21:31:02 +01:00
|
|
|
async def save_claims(self, claim_infos):
|
2019-04-20 01:42:35 +02:00
|
|
|
claim_id_to_supports = {}
|
2018-12-15 21:31:02 +01:00
|
|
|
update_file_callbacks = []
|
|
|
|
|
2018-06-20 17:41:43 +02:00
|
|
|
def _save_claims(transaction):
|
|
|
|
content_claims_to_update = []
|
|
|
|
for claim_info in claim_infos:
|
|
|
|
outpoint = "%s:%i" % (claim_info['txid'], claim_info['nout'])
|
|
|
|
claim_id = claim_info['claim_id']
|
|
|
|
name = claim_info['name']
|
2018-11-30 22:11:23 +01:00
|
|
|
amount = lbc_to_dewies(claim_info['amount'])
|
2018-06-20 17:41:43 +02:00
|
|
|
height = claim_info['height']
|
|
|
|
address = claim_info['address']
|
|
|
|
sequence = claim_info['claim_sequence']
|
2019-03-20 06:46:23 +01:00
|
|
|
certificate_id = claim_info['value'].signing_channel_id
|
2018-06-20 17:41:43 +02:00
|
|
|
try:
|
2019-04-20 07:12:43 +02:00
|
|
|
source_hash = claim_info['value'].stream.source.sd_hash
|
2019-03-20 06:46:23 +01:00
|
|
|
except (AttributeError, ValueError):
|
2018-06-20 17:41:43 +02:00
|
|
|
source_hash = None
|
2019-03-20 06:46:23 +01:00
|
|
|
serialized = binascii.hexlify(claim_info['value'].to_bytes())
|
2018-06-20 17:41:43 +02:00
|
|
|
transaction.execute(
|
|
|
|
"insert or replace into claim values (?, ?, ?, ?, ?, ?, ?, ?, ?)",
|
|
|
|
(outpoint, claim_id, name, amount, height, serialized, certificate_id, address, sequence)
|
2019-09-17 16:26:38 +02:00
|
|
|
).fetchall()
|
2019-01-22 21:43:01 +01:00
|
|
|
# if this response doesn't have support info don't overwrite the existing
|
|
|
|
# support info
|
|
|
|
if 'supports' in claim_info:
|
2019-04-20 01:42:35 +02:00
|
|
|
claim_id_to_supports[claim_id] = claim_info['supports']
|
2018-06-20 17:41:43 +02:00
|
|
|
if not source_hash:
|
|
|
|
continue
|
|
|
|
stream_hash = transaction.execute(
|
|
|
|
"select file.stream_hash from stream "
|
2019-01-22 21:43:01 +01:00
|
|
|
"inner join file on file.stream_hash=stream.stream_hash where sd_hash=?", (source_hash,)
|
2018-06-20 17:41:43 +02:00
|
|
|
).fetchone()
|
|
|
|
if not stream_hash:
|
|
|
|
continue
|
|
|
|
stream_hash = stream_hash[0]
|
|
|
|
known_outpoint = transaction.execute(
|
2019-01-22 21:43:01 +01:00
|
|
|
"select claim_outpoint from content_claim where stream_hash=?", (stream_hash,)
|
2019-03-20 06:46:23 +01:00
|
|
|
).fetchone()
|
2018-06-20 17:41:43 +02:00
|
|
|
known_claim_id = transaction.execute(
|
2018-02-28 20:20:33 +01:00
|
|
|
"select claim_id from claim "
|
|
|
|
"inner join content_claim c3 ON claim.claim_outpoint=c3.claim_outpoint "
|
2019-01-22 21:43:01 +01:00
|
|
|
"where c3.stream_hash=?", (stream_hash,)
|
2019-03-20 06:46:23 +01:00
|
|
|
).fetchone()
|
2018-06-20 17:41:43 +02:00
|
|
|
if not known_claim_id:
|
|
|
|
content_claims_to_update.append((stream_hash, outpoint))
|
|
|
|
elif known_outpoint != outpoint:
|
|
|
|
content_claims_to_update.append((stream_hash, outpoint))
|
|
|
|
for stream_hash, outpoint in content_claims_to_update:
|
|
|
|
self._save_content_claim(transaction, outpoint, stream_hash)
|
|
|
|
if stream_hash in self.content_claim_callbacks:
|
|
|
|
update_file_callbacks.append(self.content_claim_callbacks[stream_hash]())
|
|
|
|
|
2018-12-15 21:31:02 +01:00
|
|
|
await self.db.run(_save_claims)
|
|
|
|
if update_file_callbacks:
|
2022-07-29 21:07:13 +02:00
|
|
|
await asyncio.wait(map(asyncio.create_task, update_file_callbacks))
|
2019-04-20 01:42:35 +02:00
|
|
|
if claim_id_to_supports:
|
|
|
|
await self.save_supports(claim_id_to_supports)
|
2018-02-28 20:20:33 +01:00
|
|
|
|
2020-01-29 01:24:05 +01:00
|
|
|
def save_claim_from_output(self, ledger, *outputs: Output):
|
2020-01-27 06:10:55 +01:00
|
|
|
return self.save_claims([{
|
|
|
|
"claim_id": output.claim_id,
|
|
|
|
"name": output.claim_name,
|
|
|
|
"amount": dewies_to_lbc(output.amount),
|
|
|
|
"address": output.get_address(ledger),
|
|
|
|
"txid": output.tx_ref.id,
|
|
|
|
"nout": output.position,
|
|
|
|
"value": output.claim,
|
2020-01-29 01:24:05 +01:00
|
|
|
"height": output.tx_ref.height,
|
2020-01-27 06:10:55 +01:00
|
|
|
"claim_sequence": -1,
|
2020-01-29 01:24:05 +01:00
|
|
|
} for output in outputs])
|
2020-01-27 06:10:55 +01:00
|
|
|
|
2018-08-16 01:23:06 +02:00
|
|
|
def save_claims_for_resolve(self, claim_infos):
|
2019-04-20 01:52:58 +02:00
|
|
|
to_save = {}
|
2018-08-16 01:23:06 +02:00
|
|
|
for info in claim_infos:
|
|
|
|
if 'value' in info:
|
|
|
|
if info['value']:
|
2019-04-20 01:52:58 +02:00
|
|
|
to_save[info['claim_id']] = info
|
2018-08-16 01:23:06 +02:00
|
|
|
else:
|
2019-04-20 01:52:58 +02:00
|
|
|
for key in ('certificate', 'claim'):
|
|
|
|
if info.get(key, {}).get('value'):
|
|
|
|
to_save[info[key]['claim_id']] = info[key]
|
|
|
|
return self.save_claims(to_save.values())
|
2018-08-16 01:23:06 +02:00
|
|
|
|
2018-06-20 17:41:43 +02:00
|
|
|
@staticmethod
|
2020-02-11 03:15:18 +01:00
|
|
|
def _save_content_claim(transaction, claim_outpoint, stream_hash=None, bt_infohash=None):
|
|
|
|
assert stream_hash or bt_infohash
|
2018-06-20 17:41:43 +02:00
|
|
|
# get the claim id and serialized metadata
|
|
|
|
claim_info = transaction.execute(
|
|
|
|
"select claim_id, serialized_metadata from claim where claim_outpoint=?", (claim_outpoint,)
|
|
|
|
).fetchone()
|
|
|
|
if not claim_info:
|
|
|
|
raise Exception("claim not found")
|
2019-03-20 06:46:23 +01:00
|
|
|
new_claim_id, claim = claim_info[0], Claim.from_bytes(binascii.unhexlify(claim_info[1]))
|
2018-06-20 17:41:43 +02:00
|
|
|
|
|
|
|
# certificate claims should not be in the content_claim table
|
|
|
|
if not claim.is_stream:
|
|
|
|
raise Exception("claim does not contain a stream")
|
|
|
|
|
|
|
|
# get the known sd hash for this stream
|
|
|
|
known_sd_hash = transaction.execute(
|
|
|
|
"select sd_hash from stream where stream_hash=?", (stream_hash,)
|
|
|
|
).fetchone()
|
|
|
|
if not known_sd_hash:
|
|
|
|
raise Exception("stream not found")
|
|
|
|
# check the claim contains the same sd hash
|
2019-04-20 07:12:43 +02:00
|
|
|
if known_sd_hash[0] != claim.stream.source.sd_hash:
|
2018-06-20 17:41:43 +02:00
|
|
|
raise Exception("stream mismatch")
|
|
|
|
|
|
|
|
# if there is a current claim associated to the file, check that the new claim is an update to it
|
|
|
|
current_associated_content = transaction.execute(
|
|
|
|
"select claim_outpoint from content_claim where stream_hash=?", (stream_hash,)
|
|
|
|
).fetchone()
|
|
|
|
if current_associated_content:
|
|
|
|
current_associated_claim_id = transaction.execute(
|
|
|
|
"select claim_id from claim where claim_outpoint=?", current_associated_content
|
|
|
|
).fetchone()[0]
|
|
|
|
if current_associated_claim_id != new_claim_id:
|
2018-10-25 23:51:17 +02:00
|
|
|
raise Exception(
|
|
|
|
f"mismatching claim ids when updating stream {current_associated_claim_id} vs {new_claim_id}"
|
|
|
|
)
|
2018-06-20 17:41:43 +02:00
|
|
|
|
|
|
|
# update the claim associated to the file
|
2019-11-14 00:55:11 +01:00
|
|
|
transaction.execute("delete from content_claim where stream_hash=?", (stream_hash, )).fetchall()
|
2019-09-17 16:26:38 +02:00
|
|
|
transaction.execute(
|
2019-11-14 00:55:11 +01:00
|
|
|
"insert into content_claim values (?, NULL, ?)", (stream_hash, claim_outpoint)
|
2019-09-17 16:26:38 +02:00
|
|
|
).fetchall()
|
2018-06-20 17:41:43 +02:00
|
|
|
|
2018-12-15 21:31:02 +01:00
|
|
|
async def save_content_claim(self, stream_hash, claim_outpoint):
|
|
|
|
await self.db.run(self._save_content_claim, claim_outpoint, stream_hash)
|
2018-02-28 20:20:33 +01:00
|
|
|
# update corresponding ManagedEncryptedFileDownloader object
|
|
|
|
if stream_hash in self.content_claim_callbacks:
|
2018-12-15 21:31:02 +01:00
|
|
|
await self.content_claim_callbacks[stream_hash]()
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2020-02-11 03:15:18 +01:00
|
|
|
async def save_torrent_content_claim(self, bt_infohash, claim_outpoint, length, name):
|
|
|
|
def _save_torrent(transaction):
|
|
|
|
transaction.execute(
|
2020-02-25 22:18:48 +01:00
|
|
|
"insert or replace into torrent values (?, NULL, ?, ?)", (bt_infohash, length, name)
|
2020-02-11 03:15:18 +01:00
|
|
|
).fetchall()
|
|
|
|
transaction.execute(
|
2020-02-25 22:18:48 +01:00
|
|
|
"insert or replace into content_claim values (NULL, ?, ?)", (bt_infohash, claim_outpoint)
|
2020-02-11 03:15:18 +01:00
|
|
|
).fetchall()
|
|
|
|
await self.db.run(_save_torrent)
|
|
|
|
# update corresponding ManagedEncryptedFileDownloader object
|
|
|
|
if bt_infohash in self.content_claim_callbacks:
|
|
|
|
await self.content_claim_callbacks[bt_infohash]()
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
async def get_content_claim(self, stream_hash: str, include_supports: typing.Optional[bool] = True) -> typing.Dict:
|
|
|
|
claims = await self.db.run(get_claims_from_stream_hashes, [stream_hash])
|
|
|
|
claim = None
|
|
|
|
if claims:
|
|
|
|
claim = claims[stream_hash].as_dict()
|
|
|
|
if include_supports:
|
|
|
|
supports = await self.get_supports(claim['claim_id'])
|
2018-05-09 15:50:44 +02:00
|
|
|
claim['supports'] = supports
|
2018-11-30 22:11:23 +01:00
|
|
|
claim['effective_amount'] = calculate_effective_amount(claim['amount'], supports)
|
2019-01-22 21:43:01 +01:00
|
|
|
return claim
|
|
|
|
|
2020-02-11 03:15:18 +01:00
|
|
|
async def get_content_claim_for_torrent(self, bt_infohash):
|
|
|
|
claims = await self.db.run(get_claims_from_torrent_info_hashes, [bt_infohash])
|
|
|
|
return claims[bt_infohash].as_dict() if claims else None
|
|
|
|
|
2018-05-08 19:51:02 +02:00
|
|
|
# # # # # # # # # reflector functions # # # # # # # # #
|
|
|
|
|
|
|
|
def update_reflected_stream(self, sd_hash, reflector_address, success=True):
|
|
|
|
if success:
|
2019-09-17 16:26:38 +02:00
|
|
|
return self.db.execute_fetchall(
|
2018-05-08 19:51:02 +02:00
|
|
|
"insert or replace into reflected_stream values (?, ?, ?)",
|
2019-02-05 17:14:59 +01:00
|
|
|
(sd_hash, reflector_address, self.time_getter())
|
2018-05-08 19:51:02 +02:00
|
|
|
)
|
2019-09-17 16:26:38 +02:00
|
|
|
return self.db.execute_fetchall(
|
2018-05-08 19:51:02 +02:00
|
|
|
"delete from reflected_stream where sd_hash=? and reflector_address=?",
|
|
|
|
(sd_hash, reflector_address)
|
|
|
|
)
|
|
|
|
|
|
|
|
def get_streams_to_re_reflect(self):
|
|
|
|
return self.run_and_return_list(
|
|
|
|
"select s.sd_hash from stream s "
|
|
|
|
"left outer join reflected_stream r on s.sd_hash=r.sd_hash "
|
|
|
|
"where r.timestamp is null or r.timestamp < ?",
|
2019-02-03 21:09:38 +01:00
|
|
|
int(self.time_getter()) - 86400
|
2018-05-08 19:51:02 +02:00
|
|
|
)
|
2019-10-31 13:57:30 +01:00
|
|
|
|
|
|
|
# # # # # # # # # # dht functions # # # # # # # # # # #
|
2019-11-11 15:32:24 +01:00
|
|
|
async def get_persisted_kademlia_peers(self) -> typing.List[typing.Tuple[bytes, str, int, int]]:
|
2019-10-31 13:57:30 +01:00
|
|
|
query = 'select node_id, address, udp_port, tcp_port from peer'
|
|
|
|
return [(binascii.unhexlify(n), a, u, t) for n, a, u, t in await self.db.execute_fetchall(query)]
|
|
|
|
|
2019-11-11 15:32:24 +01:00
|
|
|
async def save_kademlia_peers(self, peers: typing.List['KademliaPeer']):
|
|
|
|
def _save_kademlia_peers(transaction: sqlite3.Connection):
|
2019-10-31 13:57:30 +01:00
|
|
|
transaction.execute('delete from peer').fetchall()
|
|
|
|
transaction.executemany(
|
2019-11-13 19:14:43 +01:00
|
|
|
'insert into peer(node_id, address, udp_port, tcp_port) values (?, ?, ?, ?)',
|
2021-08-21 04:36:35 +02:00
|
|
|
((binascii.hexlify(p.node_id), p.address, p.udp_port, p.tcp_port) for p in peers)
|
2019-10-31 13:57:30 +01:00
|
|
|
).fetchall()
|
2019-11-11 15:32:24 +01:00
|
|
|
return await self.db.run(_save_kademlia_peers)
|