2018-02-12 19:36:35 +01:00
|
|
|
import logging
|
2019-01-22 21:43:01 +01:00
|
|
|
import sqlite3
|
2018-11-30 22:11:23 +01:00
|
|
|
import typing
|
2019-01-22 21:43:01 +01:00
|
|
|
import asyncio
|
|
|
|
import binascii
|
2019-01-31 18:30:31 +01:00
|
|
|
import time
|
2019-01-22 21:43:01 +01:00
|
|
|
from torba.client.basedatabase import SQLiteMixin
|
2019-01-21 21:55:50 +01:00
|
|
|
from lbrynet.conf import Config
|
2019-01-22 21:43:01 +01:00
|
|
|
from lbrynet.extras.wallet.dewies import dewies_to_lbc, lbc_to_dewies
|
2018-09-17 22:31:44 +02:00
|
|
|
from lbrynet.schema.claim import ClaimDict
|
|
|
|
from lbrynet.schema.decode import smart_decode
|
2019-01-22 21:43:01 +01:00
|
|
|
from lbrynet.dht.constants import data_expiration
|
2019-01-22 23:44:25 +01:00
|
|
|
from lbrynet.blob.blob_info import BlobInfo
|
2019-01-22 21:43:01 +01:00
|
|
|
|
|
|
|
if typing.TYPE_CHECKING:
|
|
|
|
from lbrynet.blob.blob_file import BlobFile
|
|
|
|
from lbrynet.stream.descriptor import StreamDescriptor
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
log = logging.getLogger(__name__)
|
2019-01-22 21:43:01 +01:00
|
|
|
opt_str = typing.Optional[str]
|
|
|
|
opt_int = typing.Optional[int]
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
|
2018-11-30 22:11:23 +01:00
|
|
|
def calculate_effective_amount(amount: str, supports: typing.Optional[typing.List[typing.Dict]] = None) -> str:
|
|
|
|
return dewies_to_lbc(
|
|
|
|
lbc_to_dewies(amount) + sum([lbc_to_dewies(support['amount']) for support in supports])
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
class StoredStreamClaim:
|
|
|
|
def __init__(self, stream_hash: str, outpoint: opt_str = None, claim_id: opt_str = None, name: opt_str = None,
|
|
|
|
amount: opt_int = None, height: opt_int = None, serialized: opt_str = None,
|
|
|
|
channel_claim_id: opt_str = None, address: opt_str = None, claim_sequence: opt_int = None,
|
|
|
|
channel_name: opt_str = None):
|
|
|
|
self.stream_hash = stream_hash
|
|
|
|
self.claim_id = claim_id
|
|
|
|
self.outpoint = outpoint
|
|
|
|
self.claim_name = name
|
|
|
|
self.amount = amount
|
|
|
|
self.height = height
|
|
|
|
self.claim: typing.Optional[ClaimDict] = None if not serialized else smart_decode(serialized)
|
|
|
|
self.claim_address = address
|
|
|
|
self.claim_sequence = claim_sequence
|
|
|
|
self.channel_claim_id = channel_claim_id
|
|
|
|
self.channel_name = channel_name
|
|
|
|
|
|
|
|
@property
|
|
|
|
def txid(self) -> typing.Optional[str]:
|
|
|
|
return None if not self.outpoint else self.outpoint.split(":")[0]
|
|
|
|
|
|
|
|
@property
|
|
|
|
def nout(self) -> typing.Optional[int]:
|
|
|
|
return None if not self.outpoint else int(self.outpoint.split(":")[1])
|
|
|
|
|
|
|
|
@property
|
|
|
|
def metadata(self) -> typing.Optional[typing.Dict]:
|
|
|
|
return None if not self.claim else self.claim.claim_dict['stream']['metadata']
|
|
|
|
|
|
|
|
def as_dict(self) -> typing.Dict:
|
|
|
|
return {
|
|
|
|
"name": self.claim_name,
|
|
|
|
"claim_id": self.claim_id,
|
|
|
|
"address": self.claim_address,
|
|
|
|
"claim_sequence": self.claim_sequence,
|
|
|
|
"value": self.claim,
|
|
|
|
"height": self.height,
|
|
|
|
"amount": dewies_to_lbc(self.amount),
|
|
|
|
"nout": self.nout,
|
|
|
|
"txid": self.txid,
|
|
|
|
"channel_claim_id": self.channel_claim_id,
|
|
|
|
"channel_name": self.channel_name
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
def get_claims_from_stream_hashes(transaction: sqlite3.Connection,
|
|
|
|
stream_hashes: typing.List[str]) -> typing.Dict[str, StoredStreamClaim]:
|
|
|
|
query = (
|
|
|
|
"select content_claim.stream_hash, c.*, case when c.channel_claim_id is not null then "
|
|
|
|
" (select claim_name from claim where claim_id==c.channel_claim_id) "
|
|
|
|
" else null end as channel_name "
|
|
|
|
" from content_claim "
|
|
|
|
" inner join claim c on c.claim_outpoint=content_claim.claim_outpoint and content_claim.stream_hash in {}"
|
|
|
|
" order by c.rowid desc"
|
|
|
|
)
|
|
|
|
return {
|
|
|
|
claim_info.stream_hash: claim_info
|
|
|
|
for claim_info in [
|
|
|
|
None if not claim_info else StoredStreamClaim(*claim_info)
|
|
|
|
for claim_info in _batched_select(transaction, query, stream_hashes)
|
|
|
|
]
|
|
|
|
}
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def get_content_claim_from_outpoint(transaction: sqlite3.Connection,
|
|
|
|
outpoint: str) -> typing.Optional[StoredStreamClaim]:
|
|
|
|
query = (
|
|
|
|
"select content_claim.stream_hash, c.*, case when c.channel_claim_id is not null then "
|
|
|
|
" (select claim_name from claim where claim_id==c.channel_claim_id) "
|
|
|
|
" else null end as channel_name "
|
|
|
|
" from content_claim "
|
|
|
|
" inner join claim c on c.claim_outpoint=content_claim.claim_outpoint and content_claim.claim_outpoint=?"
|
2018-12-15 21:31:02 +01:00
|
|
|
)
|
2019-01-22 21:43:01 +01:00
|
|
|
claim_fields = transaction.execute(query, (outpoint, )).fetchone()
|
|
|
|
if claim_fields:
|
|
|
|
return StoredStreamClaim(*claim_fields)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def _batched_select(transaction, query, parameters):
|
|
|
|
for start_index in range(0, len(parameters), 900):
|
|
|
|
current_batch = parameters[start_index:start_index+900]
|
|
|
|
bind = "({})".format(','.join(['?'] * len(current_batch)))
|
|
|
|
for result in transaction.execute(query.format(bind), current_batch):
|
|
|
|
yield result
|
2018-02-27 23:21:22 +01:00
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2018-12-15 21:31:02 +01:00
|
|
|
class SQLiteStorage(SQLiteMixin):
|
2018-02-12 19:36:35 +01:00
|
|
|
CREATE_TABLES_QUERY = """
|
|
|
|
pragma foreign_keys=on;
|
|
|
|
pragma journal_mode=WAL;
|
2019-01-22 21:43:01 +01:00
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
create table if not exists blob (
|
|
|
|
blob_hash char(96) primary key not null,
|
|
|
|
blob_length integer not null,
|
|
|
|
next_announce_time integer not null,
|
|
|
|
should_announce integer not null default 0,
|
2018-03-27 20:58:29 +02:00
|
|
|
status text not null,
|
2018-03-29 00:47:37 +02:00
|
|
|
last_announced_time integer,
|
|
|
|
single_announce integer
|
2018-02-12 19:36:35 +01:00
|
|
|
);
|
2019-01-22 21:43:01 +01:00
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
create table if not exists stream (
|
|
|
|
stream_hash char(96) not null primary key,
|
2018-02-12 21:10:06 +01:00
|
|
|
sd_hash char(96) not null references blob,
|
2018-02-12 19:36:35 +01:00
|
|
|
stream_key text not null,
|
|
|
|
stream_name text not null,
|
2018-02-12 21:10:06 +01:00
|
|
|
suggested_filename text not null
|
2018-02-12 19:36:35 +01:00
|
|
|
);
|
2019-01-22 21:43:01 +01:00
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
create table if not exists stream_blob (
|
2018-02-12 21:10:06 +01:00
|
|
|
stream_hash char(96) not null references stream,
|
|
|
|
blob_hash char(96) references blob,
|
2018-02-12 19:36:35 +01:00
|
|
|
position integer not null,
|
|
|
|
iv char(32) not null,
|
2018-02-12 21:10:06 +01:00
|
|
|
primary key (stream_hash, blob_hash)
|
2018-02-12 19:36:35 +01:00
|
|
|
);
|
2019-01-22 21:43:01 +01:00
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
create table if not exists claim (
|
|
|
|
claim_outpoint text not null primary key,
|
|
|
|
claim_id char(40) not null,
|
|
|
|
claim_name text not null,
|
|
|
|
amount integer not null,
|
|
|
|
height integer not null,
|
|
|
|
serialized_metadata blob not null,
|
|
|
|
channel_claim_id text,
|
|
|
|
address text not null,
|
|
|
|
claim_sequence integer not null
|
|
|
|
);
|
|
|
|
|
|
|
|
create table if not exists file (
|
2018-02-12 21:10:06 +01:00
|
|
|
stream_hash text primary key not null references stream,
|
2018-02-12 19:36:35 +01:00
|
|
|
file_name text not null,
|
|
|
|
download_directory text not null,
|
|
|
|
blob_data_rate real not null,
|
2018-02-12 21:10:06 +01:00
|
|
|
status text not null
|
2018-02-12 19:36:35 +01:00
|
|
|
);
|
2019-01-22 21:43:01 +01:00
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
create table if not exists content_claim (
|
2018-02-12 21:10:06 +01:00
|
|
|
stream_hash text unique not null references file,
|
|
|
|
claim_outpoint text not null references claim,
|
|
|
|
primary key (stream_hash, claim_outpoint)
|
2018-02-12 19:36:35 +01:00
|
|
|
);
|
2019-01-22 21:43:01 +01:00
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
create table if not exists support (
|
|
|
|
support_outpoint text not null primary key,
|
|
|
|
claim_id text not null,
|
|
|
|
amount integer not null,
|
|
|
|
address text not null
|
|
|
|
);
|
2019-01-22 21:43:01 +01:00
|
|
|
|
2018-05-08 19:51:02 +02:00
|
|
|
create table if not exists reflected_stream (
|
|
|
|
sd_hash text not null,
|
|
|
|
reflector_address text not null,
|
|
|
|
timestamp integer,
|
|
|
|
primary key (sd_hash, reflector_address)
|
|
|
|
);
|
2018-10-10 05:58:32 +02:00
|
|
|
"""
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-01-21 21:55:50 +01:00
|
|
|
def __init__(self, conf: Config, path, loop=None):
|
2018-12-15 21:31:02 +01:00
|
|
|
super().__init__(path)
|
2019-01-21 21:55:50 +01:00
|
|
|
self.conf = conf
|
2018-12-15 21:31:02 +01:00
|
|
|
self.content_claim_callbacks = {}
|
2019-01-08 02:23:18 +01:00
|
|
|
self.loop = loop or asyncio.get_event_loop()
|
2018-12-15 21:31:02 +01:00
|
|
|
|
|
|
|
async def run_and_return_one_or_none(self, query, *args):
|
|
|
|
for row in await self.db.execute_fetchall(query, args):
|
2019-01-07 08:52:53 +01:00
|
|
|
if len(row) == 1:
|
|
|
|
return row[0]
|
2018-12-15 21:31:02 +01:00
|
|
|
return row
|
|
|
|
|
|
|
|
async def run_and_return_list(self, query, *args):
|
|
|
|
rows = list(await self.db.execute_fetchall(query, args))
|
|
|
|
return [col[0] for col in rows] if rows else []
|
|
|
|
|
2018-02-12 19:36:35 +01:00
|
|
|
# # # # # # # # # blob functions # # # # # # # # #
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def add_completed_blob(self, blob_hash: str):
|
|
|
|
log.debug("Adding a completed blob. blob_hash=%s", blob_hash)
|
|
|
|
return self.db.execute("update blob set status='finished' where blob.blob_hash=?", (blob_hash, ))
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def set_should_announce(self, blob_hash: str, next_announce_time: int, should_announce: int):
|
2018-12-15 21:31:02 +01:00
|
|
|
return self.db.execute(
|
2018-02-12 19:36:35 +01:00
|
|
|
"update blob set next_announce_time=?, should_announce=? where blob_hash=?",
|
2018-05-30 09:26:00 +02:00
|
|
|
(next_announce_time or 0, int(bool(should_announce)), blob_hash)
|
2018-02-12 19:36:35 +01:00
|
|
|
)
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def get_blob_status(self, blob_hash: str):
|
2018-02-12 19:36:35 +01:00
|
|
|
return self.run_and_return_one_or_none(
|
|
|
|
"select status from blob where blob_hash=?", blob_hash
|
|
|
|
)
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def add_known_blob(self, blob_hash: str, length: int):
|
2018-12-15 21:31:02 +01:00
|
|
|
return self.db.execute(
|
2018-06-14 20:49:53 +02:00
|
|
|
"insert or ignore into blob values (?, ?, ?, ?, ?, ?, ?)", (blob_hash, length, 0, 0, "pending", 0, 0)
|
|
|
|
)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def should_announce(self, blob_hash: str):
|
2018-02-12 19:36:35 +01:00
|
|
|
return self.run_and_return_one_or_none(
|
|
|
|
"select should_announce from blob where blob_hash=?", blob_hash
|
|
|
|
)
|
|
|
|
|
|
|
|
def count_should_announce_blobs(self):
|
|
|
|
return self.run_and_return_one_or_none(
|
2018-03-08 22:46:39 +01:00
|
|
|
"select count(*) from blob where should_announce=1 and status='finished'"
|
2018-02-12 19:36:35 +01:00
|
|
|
)
|
|
|
|
|
|
|
|
def get_all_should_announce_blobs(self):
|
|
|
|
return self.run_and_return_list(
|
2018-03-08 22:46:39 +01:00
|
|
|
"select blob_hash from blob where should_announce=1 and status='finished'"
|
2018-02-12 19:36:35 +01:00
|
|
|
)
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def get_all_finished_blobs(self):
|
|
|
|
return self.run_and_return_list(
|
2018-06-29 18:01:46 +02:00
|
|
|
"select blob_hash from blob where status='finished'"
|
|
|
|
)
|
|
|
|
|
2018-08-02 23:33:56 +02:00
|
|
|
def count_finished_blobs(self):
|
|
|
|
return self.run_and_return_one_or_none(
|
|
|
|
"select count(*) from blob where status='finished'"
|
|
|
|
)
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def update_last_announced_blobs(self, blob_hashes: typing.List[str], last_announced: float):
|
|
|
|
def _update_last_announced_blobs(transaction: sqlite3.Connection):
|
|
|
|
return transaction.executemany(
|
|
|
|
"update blob set next_announce_time=?, last_announced_time=?, single_announce=0 "
|
|
|
|
"where blob_hash=?",
|
|
|
|
[(int(last_announced + (data_expiration / 2)), int(last_announced), blob_hash)
|
|
|
|
for blob_hash in blob_hashes]
|
|
|
|
)
|
|
|
|
return self.db.run(_update_last_announced_blobs)
|
2018-03-27 21:12:44 +02:00
|
|
|
|
2018-03-29 00:47:37 +02:00
|
|
|
def should_single_announce_blobs(self, blob_hashes, immediate=False):
|
|
|
|
def set_single_announce(transaction):
|
2019-01-08 02:23:18 +01:00
|
|
|
now = self.loop.time()
|
2018-03-29 00:47:37 +02:00
|
|
|
for blob_hash in blob_hashes:
|
|
|
|
if immediate:
|
|
|
|
transaction.execute(
|
|
|
|
"update blob set single_announce=1, next_announce_time=? "
|
|
|
|
"where blob_hash=? and status='finished'", (int(now), blob_hash)
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
transaction.execute(
|
2019-01-22 21:43:01 +01:00
|
|
|
"update blob set single_announce=1 where blob_hash=? and status='finished'", (blob_hash,)
|
2018-03-29 00:47:37 +02:00
|
|
|
)
|
2018-12-15 21:31:02 +01:00
|
|
|
return self.db.run(set_single_announce)
|
2018-03-29 00:47:37 +02:00
|
|
|
|
2018-03-27 21:12:44 +02:00
|
|
|
def get_blobs_to_announce(self):
|
2018-02-12 19:36:35 +01:00
|
|
|
def get_and_update(transaction):
|
2019-01-31 05:09:57 +01:00
|
|
|
timestamp = int(self.loop.time())
|
2019-01-28 15:51:02 +01:00
|
|
|
if self.conf.announce_head_and_sd_only:
|
2018-02-12 19:36:35 +01:00
|
|
|
r = transaction.execute(
|
|
|
|
"select blob_hash from blob "
|
2018-03-29 00:47:37 +02:00
|
|
|
"where blob_hash is not null and "
|
|
|
|
"(should_announce=1 or single_announce=1) and next_announce_time<? and status='finished'",
|
2018-02-12 19:36:35 +01:00
|
|
|
(timestamp,)
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
r = transaction.execute(
|
2018-03-08 22:46:39 +01:00
|
|
|
"select blob_hash from blob where blob_hash is not null "
|
|
|
|
"and next_announce_time<? and status='finished'", (timestamp,)
|
2018-02-12 19:36:35 +01:00
|
|
|
)
|
2018-03-27 21:12:44 +02:00
|
|
|
blobs = [b[0] for b in r.fetchall()]
|
2018-02-12 19:36:35 +01:00
|
|
|
return blobs
|
2018-12-15 21:31:02 +01:00
|
|
|
return self.db.run(get_and_update)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
def delete_blobs_from_db(self, blob_hashes):
|
|
|
|
def delete_blobs(transaction):
|
2019-01-22 21:43:01 +01:00
|
|
|
transaction.executemany(
|
|
|
|
"delete from blob where blob_hash=?;", [(blob_hash,) for blob_hash in blob_hashes]
|
|
|
|
)
|
2018-12-15 21:31:02 +01:00
|
|
|
return self.db.run(delete_blobs)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
def get_all_blob_hashes(self):
|
|
|
|
return self.run_and_return_list("select blob_hash from blob")
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
# # # # # # # # # stream functions # # # # # # # # #
|
2018-03-08 22:46:39 +01:00
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
async def stream_exists(self, sd_hash: str) -> bool:
|
|
|
|
streams = await self.run_and_return_one_or_none("select stream_hash from stream where sd_hash=?", sd_hash)
|
|
|
|
return streams is not None
|
2018-03-08 22:46:39 +01:00
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
async def file_exists(self, sd_hash: str) -> bool:
|
|
|
|
streams = await self.run_and_return_one_or_none("select f.stream_hash from file f "
|
|
|
|
"inner join stream s on "
|
|
|
|
"s.stream_hash=f.stream_hash and s.sd_hash=?", sd_hash)
|
|
|
|
return streams is not None
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def store_stream(self, sd_blob: 'BlobFile', descriptor: 'StreamDescriptor'):
|
|
|
|
def _store_stream(transaction: sqlite3.Connection):
|
|
|
|
# add the head blob and set it to be announced
|
2018-12-15 21:31:02 +01:00
|
|
|
transaction.execute(
|
2019-01-22 21:43:01 +01:00
|
|
|
"insert or ignore into blob values (?, ?, ?, ?, ?, ?, ?), (?, ?, ?, ?, ?, ?, ?)",
|
|
|
|
(
|
|
|
|
sd_blob.blob_hash, sd_blob.length, 0, 1, "pending", 0, 0,
|
|
|
|
descriptor.blobs[0].blob_hash, descriptor.blobs[0].length, 0, 1, "pending", 0, 0
|
2018-12-15 21:31:02 +01:00
|
|
|
)
|
|
|
|
)
|
2019-01-22 21:43:01 +01:00
|
|
|
# add the rest of the blobs with announcement off
|
|
|
|
if len(descriptor.blobs) > 2:
|
|
|
|
transaction.executemany(
|
|
|
|
"insert or ignore into blob values (?, ?, ?, ?, ?, ?, ?)",
|
|
|
|
[(blob.blob_hash, blob.length, 0, 0, "pending", 0, 0)
|
|
|
|
for blob in descriptor.blobs[1:-1]]
|
2018-12-15 21:31:02 +01:00
|
|
|
)
|
2019-01-22 21:43:01 +01:00
|
|
|
# associate the blobs to the stream
|
|
|
|
transaction.execute("insert or ignore into stream values (?, ?, ?, ?, ?)",
|
|
|
|
(descriptor.stream_hash, sd_blob.blob_hash, descriptor.key,
|
|
|
|
binascii.hexlify(descriptor.stream_name.encode()).decode(),
|
|
|
|
binascii.hexlify(descriptor.suggested_file_name.encode()).decode()))
|
|
|
|
# add the stream
|
|
|
|
transaction.executemany(
|
|
|
|
"insert or ignore into stream_blob values (?, ?, ?, ?)",
|
|
|
|
[(descriptor.stream_hash, blob.blob_hash, blob.blob_num, blob.iv)
|
|
|
|
for blob in descriptor.blobs]
|
|
|
|
)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
return self.db.run(_store_stream)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-01-22 23:44:25 +01:00
|
|
|
def get_blobs_for_stream(self, stream_hash, only_completed=False):
|
|
|
|
def _get_blobs_for_stream(transaction):
|
|
|
|
crypt_blob_infos = []
|
|
|
|
stream_blobs = transaction.execute(
|
|
|
|
"select blob_hash, position, iv from stream_blob where stream_hash=?", (stream_hash, )
|
|
|
|
).fetchall()
|
|
|
|
if only_completed:
|
|
|
|
lengths = transaction.execute(
|
|
|
|
"select b.blob_hash, b.blob_length from blob b "
|
|
|
|
"inner join stream_blob s ON b.blob_hash=s.blob_hash and b.status='finished' and s.stream_hash=?",
|
|
|
|
(stream_hash, )
|
|
|
|
).fetchall()
|
|
|
|
else:
|
|
|
|
lengths = transaction.execute(
|
|
|
|
"select b.blob_hash, b.blob_length from blob b "
|
|
|
|
"inner join stream_blob s ON b.blob_hash=s.blob_hash and s.stream_hash=?",
|
|
|
|
(stream_hash, )
|
|
|
|
).fetchall()
|
|
|
|
|
|
|
|
blob_length_dict = {}
|
|
|
|
for blob_hash, length in lengths:
|
|
|
|
blob_length_dict[blob_hash] = length
|
|
|
|
|
|
|
|
for blob_hash, position, iv in stream_blobs:
|
|
|
|
blob_length = blob_length_dict.get(blob_hash, 0)
|
|
|
|
crypt_blob_infos.append(BlobInfo(position, blob_length, iv, blob_hash))
|
|
|
|
crypt_blob_infos = sorted(crypt_blob_infos, key=lambda info: info.blob_num)
|
|
|
|
return crypt_blob_infos
|
|
|
|
return self.db.run(_get_blobs_for_stream)
|
|
|
|
|
|
|
|
def get_sd_blob_hash_for_stream(self, stream_hash):
|
|
|
|
return self.run_and_return_one_or_none(
|
|
|
|
"select sd_hash from stream where stream_hash=?", stream_hash
|
|
|
|
)
|
|
|
|
|
|
|
|
def get_stream_hash_for_sd_hash(self, sd_blob_hash):
|
|
|
|
return self.run_and_return_one_or_none(
|
|
|
|
"select stream_hash from stream where sd_hash = ?", sd_blob_hash
|
|
|
|
)
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def delete_stream(self, descriptor: 'StreamDescriptor'):
|
|
|
|
def _delete_stream(transaction: sqlite3.Connection):
|
|
|
|
transaction.execute("delete from content_claim where stream_hash=? ", (descriptor.stream_hash,))
|
|
|
|
transaction.execute("delete from file where stream_hash=? ", (descriptor.stream_hash, ))
|
|
|
|
transaction.execute("delete from stream_blob where stream_hash=?", (descriptor.stream_hash, ))
|
|
|
|
transaction.execute("delete from stream where stream_hash=? ", (descriptor.stream_hash, ))
|
|
|
|
transaction.execute("delete from blob where blob_hash=?", (descriptor.sd_hash, ))
|
|
|
|
transaction.executemany("delete from blob where blob_hash=?",
|
|
|
|
[(blob.blob_hash, ) for blob in descriptor.blobs[:-1]])
|
|
|
|
return self.db.run(_delete_stream)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
# # # # # # # # # file stuff # # # # # # # # #
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def save_downloaded_file(self, stream_hash, file_name, download_directory, data_payment_rate):
|
|
|
|
return self.save_published_file(
|
2019-01-31 18:32:52 +01:00
|
|
|
stream_hash, file_name, download_directory, data_payment_rate, status="running"
|
2018-02-12 19:36:35 +01:00
|
|
|
)
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def save_published_file(self, stream_hash: str, file_name: str, download_directory: str, data_payment_rate: float,
|
|
|
|
status="finished"):
|
|
|
|
return self.db.execute(
|
2018-06-12 17:54:01 +02:00
|
|
|
"insert into file values (?, ?, ?, ?, ?)",
|
2019-01-31 18:32:52 +01:00
|
|
|
(stream_hash, binascii.hexlify(file_name.encode()).decode(),
|
|
|
|
binascii.hexlify(download_directory.encode()).decode(), data_payment_rate, status)
|
2018-06-12 17:54:01 +02:00
|
|
|
)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
async def get_all_lbry_files(self) -> typing.List[typing.Dict]:
|
2018-02-12 19:36:35 +01:00
|
|
|
def _lbry_file_dict(rowid, stream_hash, file_name, download_dir, data_rate, status, _, sd_hash, stream_key,
|
2019-01-22 21:43:01 +01:00
|
|
|
stream_name, suggested_file_name) -> typing.Dict:
|
2018-02-12 19:36:35 +01:00
|
|
|
return {
|
|
|
|
"row_id": rowid,
|
|
|
|
"stream_hash": stream_hash,
|
|
|
|
"file_name": file_name,
|
|
|
|
"download_directory": download_dir,
|
|
|
|
"blob_data_rate": data_rate,
|
|
|
|
"status": status,
|
|
|
|
"sd_hash": sd_hash,
|
|
|
|
"key": stream_key,
|
|
|
|
"stream_name": stream_name,
|
|
|
|
"suggested_file_name": suggested_file_name
|
|
|
|
}
|
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def _get_all_files(transaction: sqlite3.Connection) -> typing.List[typing.Dict]:
|
|
|
|
file_infos = list(map(lambda a: _lbry_file_dict(*a), transaction.execute(
|
|
|
|
"select file.rowid, file.*, stream.* "
|
|
|
|
"from file inner join stream on file.stream_hash=stream.stream_hash"
|
|
|
|
).fetchall()))
|
|
|
|
stream_hashes = [file_info['stream_hash'] for file_info in file_infos]
|
|
|
|
claim_infos = get_claims_from_stream_hashes(transaction, stream_hashes)
|
|
|
|
for index in range(len(file_infos)): # pylint: disable=consider-using-enumerate
|
|
|
|
file_infos[index]['claim'] = claim_infos.get(file_infos[index]['stream_hash'])
|
|
|
|
return file_infos
|
|
|
|
|
|
|
|
results = await self.db.run(_get_all_files)
|
|
|
|
if results:
|
|
|
|
return results
|
|
|
|
return []
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def change_file_status(self, stream_hash: str, new_status: str):
|
|
|
|
log.info("update file status %s -> %s", stream_hash, new_status)
|
|
|
|
return self.db.execute("update file set status=? where stream_hash=?", (new_status, stream_hash))
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
def get_all_stream_hashes(self):
|
|
|
|
return self.run_and_return_list("select stream_hash from stream")
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
# # # # # # # # # support functions # # # # # # # # #
|
|
|
|
|
|
|
|
def save_supports(self, claim_id, supports):
|
|
|
|
# TODO: add 'address' to support items returned for a claim from lbrycrdd and lbryum-server
|
|
|
|
def _save_support(transaction):
|
2019-01-22 21:43:01 +01:00
|
|
|
transaction.execute("delete from support where claim_id=?", (claim_id,))
|
2018-02-12 19:36:35 +01:00
|
|
|
for support in supports:
|
|
|
|
transaction.execute(
|
|
|
|
"insert into support values (?, ?, ?, ?)",
|
2018-11-30 22:11:23 +01:00
|
|
|
("%s:%i" % (support['txid'], support['nout']), claim_id, lbc_to_dewies(support['amount']),
|
2018-02-12 19:36:35 +01:00
|
|
|
support.get('address', ""))
|
|
|
|
)
|
2018-12-15 21:31:02 +01:00
|
|
|
return self.db.run(_save_support)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2018-05-09 00:46:29 +02:00
|
|
|
def get_supports(self, *claim_ids):
|
2018-02-12 19:36:35 +01:00
|
|
|
def _format_support(outpoint, supported_id, amount, address):
|
|
|
|
return {
|
|
|
|
"txid": outpoint.split(":")[0],
|
|
|
|
"nout": int(outpoint.split(":")[1]),
|
|
|
|
"claim_id": supported_id,
|
2018-11-30 22:11:23 +01:00
|
|
|
"amount": dewies_to_lbc(amount),
|
2018-02-12 19:36:35 +01:00
|
|
|
"address": address,
|
|
|
|
}
|
|
|
|
|
|
|
|
def _get_supports(transaction):
|
|
|
|
return [
|
|
|
|
_format_support(*support_info)
|
2018-11-07 14:25:46 +01:00
|
|
|
for support_info in _batched_select(
|
|
|
|
transaction,
|
|
|
|
"select * from support where claim_id in {}",
|
2018-05-09 00:46:29 +02:00
|
|
|
tuple(claim_ids)
|
2018-11-07 14:25:46 +01:00
|
|
|
)
|
2018-02-12 19:36:35 +01:00
|
|
|
]
|
|
|
|
|
2018-12-15 21:31:02 +01:00
|
|
|
return self.db.run(_get_supports)
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
# # # # # # # # # claim functions # # # # # # # # #
|
|
|
|
|
2018-12-15 21:31:02 +01:00
|
|
|
async def save_claims(self, claim_infos):
|
|
|
|
support_callbacks = []
|
|
|
|
update_file_callbacks = []
|
|
|
|
|
2018-06-20 17:41:43 +02:00
|
|
|
def _save_claims(transaction):
|
|
|
|
content_claims_to_update = []
|
|
|
|
for claim_info in claim_infos:
|
|
|
|
outpoint = "%s:%i" % (claim_info['txid'], claim_info['nout'])
|
|
|
|
claim_id = claim_info['claim_id']
|
|
|
|
name = claim_info['name']
|
2018-11-30 22:11:23 +01:00
|
|
|
amount = lbc_to_dewies(claim_info['amount'])
|
2018-06-20 17:41:43 +02:00
|
|
|
height = claim_info['height']
|
|
|
|
address = claim_info['address']
|
|
|
|
sequence = claim_info['claim_sequence']
|
|
|
|
try:
|
2019-01-08 22:43:36 +01:00
|
|
|
certificate_id = claim_info['value'].get('publisherSignature', {}).get('certificateId')
|
2018-06-20 17:41:43 +02:00
|
|
|
except AttributeError:
|
|
|
|
certificate_id = None
|
|
|
|
try:
|
|
|
|
if claim_info['value'].get('stream', {}).get('source', {}).get('sourceType') == "lbry_sd_hash":
|
|
|
|
source_hash = claim_info['value'].get('stream', {}).get('source', {}).get('source')
|
|
|
|
else:
|
|
|
|
source_hash = None
|
|
|
|
except AttributeError:
|
|
|
|
source_hash = None
|
2019-01-22 21:43:01 +01:00
|
|
|
serialized = claim_info.get('hex') or binascii.hexlify(
|
|
|
|
smart_decode(claim_info['value']).serialized).decode()
|
2018-06-20 17:41:43 +02:00
|
|
|
transaction.execute(
|
|
|
|
"insert or replace into claim values (?, ?, ?, ?, ?, ?, ?, ?, ?)",
|
|
|
|
(outpoint, claim_id, name, amount, height, serialized, certificate_id, address, sequence)
|
|
|
|
)
|
2019-01-22 21:43:01 +01:00
|
|
|
# if this response doesn't have support info don't overwrite the existing
|
|
|
|
# support info
|
|
|
|
if 'supports' in claim_info:
|
2018-12-15 21:31:02 +01:00
|
|
|
support_callbacks.append((claim_id, claim_info['supports']))
|
2018-06-20 17:41:43 +02:00
|
|
|
if not source_hash:
|
|
|
|
continue
|
|
|
|
stream_hash = transaction.execute(
|
|
|
|
"select file.stream_hash from stream "
|
2019-01-22 21:43:01 +01:00
|
|
|
"inner join file on file.stream_hash=stream.stream_hash where sd_hash=?", (source_hash,)
|
2018-06-20 17:41:43 +02:00
|
|
|
).fetchone()
|
|
|
|
if not stream_hash:
|
|
|
|
continue
|
|
|
|
stream_hash = stream_hash[0]
|
|
|
|
known_outpoint = transaction.execute(
|
2019-01-22 21:43:01 +01:00
|
|
|
"select claim_outpoint from content_claim where stream_hash=?", (stream_hash,)
|
2018-02-28 20:20:33 +01:00
|
|
|
)
|
2018-06-20 17:41:43 +02:00
|
|
|
known_claim_id = transaction.execute(
|
2018-02-28 20:20:33 +01:00
|
|
|
"select claim_id from claim "
|
|
|
|
"inner join content_claim c3 ON claim.claim_outpoint=c3.claim_outpoint "
|
2019-01-22 21:43:01 +01:00
|
|
|
"where c3.stream_hash=?", (stream_hash,)
|
2018-02-28 20:20:33 +01:00
|
|
|
)
|
2018-06-20 17:41:43 +02:00
|
|
|
if not known_claim_id:
|
|
|
|
content_claims_to_update.append((stream_hash, outpoint))
|
|
|
|
elif known_outpoint != outpoint:
|
|
|
|
content_claims_to_update.append((stream_hash, outpoint))
|
|
|
|
for stream_hash, outpoint in content_claims_to_update:
|
|
|
|
self._save_content_claim(transaction, outpoint, stream_hash)
|
|
|
|
if stream_hash in self.content_claim_callbacks:
|
|
|
|
update_file_callbacks.append(self.content_claim_callbacks[stream_hash]())
|
|
|
|
|
2018-12-15 21:31:02 +01:00
|
|
|
await self.db.run(_save_claims)
|
|
|
|
if update_file_callbacks:
|
|
|
|
await asyncio.wait(update_file_callbacks)
|
|
|
|
if support_callbacks:
|
|
|
|
await asyncio.wait([
|
|
|
|
self.save_supports(*args) for args in support_callbacks
|
|
|
|
])
|
2018-02-28 20:20:33 +01:00
|
|
|
|
2018-08-16 01:23:06 +02:00
|
|
|
def save_claims_for_resolve(self, claim_infos):
|
|
|
|
to_save = []
|
|
|
|
for info in claim_infos:
|
|
|
|
if 'value' in info:
|
|
|
|
if info['value']:
|
|
|
|
to_save.append(info)
|
|
|
|
else:
|
|
|
|
if 'certificate' in info and info['certificate']['value']:
|
|
|
|
to_save.append(info['certificate'])
|
|
|
|
if 'claim' in info and info['claim']['value']:
|
|
|
|
to_save.append(info['claim'])
|
|
|
|
return self.save_claims(to_save)
|
|
|
|
|
2018-06-20 17:41:43 +02:00
|
|
|
@staticmethod
|
|
|
|
def _save_content_claim(transaction, claim_outpoint, stream_hash):
|
|
|
|
# get the claim id and serialized metadata
|
|
|
|
claim_info = transaction.execute(
|
|
|
|
"select claim_id, serialized_metadata from claim where claim_outpoint=?", (claim_outpoint,)
|
|
|
|
).fetchone()
|
|
|
|
if not claim_info:
|
|
|
|
raise Exception("claim not found")
|
2019-01-22 21:43:01 +01:00
|
|
|
new_claim_id, claim = claim_info[0], ClaimDict.deserialize(binascii.unhexlify(claim_info[1]))
|
2018-06-20 17:41:43 +02:00
|
|
|
|
|
|
|
# certificate claims should not be in the content_claim table
|
|
|
|
if not claim.is_stream:
|
|
|
|
raise Exception("claim does not contain a stream")
|
|
|
|
|
|
|
|
# get the known sd hash for this stream
|
|
|
|
known_sd_hash = transaction.execute(
|
|
|
|
"select sd_hash from stream where stream_hash=?", (stream_hash,)
|
|
|
|
).fetchone()
|
|
|
|
if not known_sd_hash:
|
|
|
|
raise Exception("stream not found")
|
|
|
|
# check the claim contains the same sd hash
|
2018-07-31 19:20:25 +02:00
|
|
|
if known_sd_hash[0].encode() != claim.source_hash:
|
2018-06-20 17:41:43 +02:00
|
|
|
raise Exception("stream mismatch")
|
|
|
|
|
|
|
|
# if there is a current claim associated to the file, check that the new claim is an update to it
|
|
|
|
current_associated_content = transaction.execute(
|
|
|
|
"select claim_outpoint from content_claim where stream_hash=?", (stream_hash,)
|
|
|
|
).fetchone()
|
|
|
|
if current_associated_content:
|
|
|
|
current_associated_claim_id = transaction.execute(
|
|
|
|
"select claim_id from claim where claim_outpoint=?", current_associated_content
|
|
|
|
).fetchone()[0]
|
|
|
|
if current_associated_claim_id != new_claim_id:
|
2018-10-25 23:51:17 +02:00
|
|
|
raise Exception(
|
|
|
|
f"mismatching claim ids when updating stream {current_associated_claim_id} vs {new_claim_id}"
|
|
|
|
)
|
2018-06-20 17:41:43 +02:00
|
|
|
|
|
|
|
# update the claim associated to the file
|
|
|
|
transaction.execute("insert or replace into content_claim values (?, ?)", (stream_hash, claim_outpoint))
|
|
|
|
|
2018-12-15 21:31:02 +01:00
|
|
|
async def save_content_claim(self, stream_hash, claim_outpoint):
|
|
|
|
await self.db.run(self._save_content_claim, claim_outpoint, stream_hash)
|
2018-02-28 20:20:33 +01:00
|
|
|
# update corresponding ManagedEncryptedFileDownloader object
|
|
|
|
if stream_hash in self.content_claim_callbacks:
|
2018-12-15 21:31:02 +01:00
|
|
|
await self.content_claim_callbacks[stream_hash]()
|
2018-02-12 19:36:35 +01:00
|
|
|
|
2019-01-22 21:43:01 +01:00
|
|
|
async def get_content_claim(self, stream_hash: str, include_supports: typing.Optional[bool] = True) -> typing.Dict:
|
|
|
|
claims = await self.db.run(get_claims_from_stream_hashes, [stream_hash])
|
|
|
|
claim = None
|
|
|
|
if claims:
|
|
|
|
claim = claims[stream_hash].as_dict()
|
|
|
|
if include_supports:
|
|
|
|
supports = await self.get_supports(claim['claim_id'])
|
2018-05-09 15:50:44 +02:00
|
|
|
claim['supports'] = supports
|
2018-11-30 22:11:23 +01:00
|
|
|
claim['effective_amount'] = calculate_effective_amount(claim['amount'], supports)
|
2019-01-22 21:43:01 +01:00
|
|
|
return claim
|
|
|
|
|
|
|
|
async def get_claims_from_stream_hashes(self, stream_hashes: typing.List[str],
|
|
|
|
include_supports: typing.Optional[bool] = True):
|
|
|
|
claims = await self.db.run(get_claims_from_stream_hashes, stream_hashes)
|
|
|
|
return {stream_hash: claim_info.as_dict() for stream_hash, claim_info in claims.items()}
|
2018-05-09 15:50:44 +02:00
|
|
|
|
2018-12-15 21:31:02 +01:00
|
|
|
async def get_claim(self, claim_outpoint, include_supports=True):
|
2019-01-22 21:43:01 +01:00
|
|
|
claim_info = await self.db.run(get_content_claim_from_outpoint, claim_outpoint)
|
|
|
|
if not claim_info:
|
|
|
|
return
|
|
|
|
result = claim_info.as_dict()
|
2018-02-12 19:36:35 +01:00
|
|
|
if include_supports:
|
2019-01-22 21:43:01 +01:00
|
|
|
supports = await self.get_supports(result['claim_id'])
|
|
|
|
result['supports'] = supports
|
|
|
|
result['effective_amount'] = calculate_effective_amount(result['amount'], supports)
|
2018-12-15 21:31:02 +01:00
|
|
|
return result
|
2018-02-12 19:36:35 +01:00
|
|
|
|
|
|
|
def get_unknown_certificate_ids(self):
|
|
|
|
def _get_unknown_certificate_claim_ids(transaction):
|
|
|
|
return [
|
|
|
|
claim_id for (claim_id,) in transaction.execute(
|
|
|
|
"select distinct c1.channel_claim_id from claim as c1 "
|
|
|
|
"where c1.channel_claim_id!='' "
|
|
|
|
"and c1.channel_claim_id not in "
|
|
|
|
"(select c2.claim_id from claim as c2)"
|
|
|
|
).fetchall()
|
|
|
|
]
|
2018-12-15 21:31:02 +01:00
|
|
|
return self.db.run(_get_unknown_certificate_claim_ids)
|
2018-02-28 20:59:12 +01:00
|
|
|
|
2018-12-15 21:31:02 +01:00
|
|
|
async def get_pending_claim_outpoints(self):
|
|
|
|
claim_outpoints = await self.run_and_return_list("select claim_outpoint from claim where height=-1")
|
2018-02-28 20:59:12 +01:00
|
|
|
results = {} # {txid: [nout, ...]}
|
|
|
|
for outpoint_str in claim_outpoints:
|
|
|
|
txid, nout = outpoint_str.split(":")
|
|
|
|
outputs = results.get(txid, [])
|
|
|
|
outputs.append(int(nout))
|
|
|
|
results[txid] = outputs
|
|
|
|
if results:
|
|
|
|
log.debug("missing transaction heights for %i claims", len(results))
|
2018-12-15 21:31:02 +01:00
|
|
|
return results
|
2018-02-28 20:59:12 +01:00
|
|
|
|
|
|
|
def save_claim_tx_heights(self, claim_tx_heights):
|
|
|
|
def _save_claim_heights(transaction):
|
2018-07-21 23:11:44 +02:00
|
|
|
for outpoint, height in claim_tx_heights.items():
|
2018-02-28 20:59:12 +01:00
|
|
|
transaction.execute(
|
|
|
|
"update claim set height=? where claim_outpoint=? and height=-1",
|
|
|
|
(height, outpoint)
|
|
|
|
)
|
2018-12-15 21:31:02 +01:00
|
|
|
return self.db.run(_save_claim_heights)
|
2018-05-08 19:51:02 +02:00
|
|
|
|
|
|
|
# # # # # # # # # reflector functions # # # # # # # # #
|
|
|
|
|
|
|
|
def update_reflected_stream(self, sd_hash, reflector_address, success=True):
|
|
|
|
if success:
|
2018-12-15 21:31:02 +01:00
|
|
|
return self.db.execute(
|
2018-05-08 19:51:02 +02:00
|
|
|
"insert or replace into reflected_stream values (?, ?, ?)",
|
2019-01-31 18:30:31 +01:00
|
|
|
(sd_hash, reflector_address, time.time())
|
2018-05-08 19:51:02 +02:00
|
|
|
)
|
2018-12-15 21:31:02 +01:00
|
|
|
return self.db.execute(
|
2018-05-08 19:51:02 +02:00
|
|
|
"delete from reflected_stream where sd_hash=? and reflector_address=?",
|
|
|
|
(sd_hash, reflector_address)
|
|
|
|
)
|
|
|
|
|
|
|
|
def get_streams_to_re_reflect(self):
|
|
|
|
return self.run_and_return_list(
|
|
|
|
"select s.sd_hash from stream s "
|
|
|
|
"left outer join reflected_stream r on s.sd_hash=r.sd_hash "
|
|
|
|
"where r.timestamp is null or r.timestamp < ?",
|
2019-01-28 15:51:02 +01:00
|
|
|
self.loop.time() - 86400
|
2018-05-08 19:51:02 +02:00
|
|
|
)
|