2019-08-26 23:23:43 +02:00
|
|
|
import os
|
2019-12-08 00:13:13 +01:00
|
|
|
import apsw
|
2019-03-31 00:40:01 +01:00
|
|
|
from typing import Union, Tuple, Set, List
|
2019-05-26 05:06:22 +02:00
|
|
|
from itertools import chain
|
2019-06-23 04:25:22 +02:00
|
|
|
from decimal import Decimal
|
2019-12-08 00:13:13 +01:00
|
|
|
from collections import namedtuple
|
2020-01-10 16:47:57 +01:00
|
|
|
from multiprocessing import Manager
|
|
|
|
from binascii import unhexlify
|
2018-11-04 10:42:47 +01:00
|
|
|
|
2020-01-10 16:47:57 +01:00
|
|
|
from lbry.wallet.server.leveldb import LevelDB
|
2019-12-31 20:52:57 +01:00
|
|
|
from lbry.wallet.server.util import class_logger
|
2020-01-03 04:18:49 +01:00
|
|
|
from lbry.wallet.database import query, constraints_to_sql
|
2019-03-31 00:40:01 +01:00
|
|
|
|
2019-06-23 02:11:33 +02:00
|
|
|
from lbry.schema.tags import clean_tags
|
2019-06-21 02:55:47 +02:00
|
|
|
from lbry.schema.mime_types import guess_stream_type
|
2020-01-03 04:18:49 +01:00
|
|
|
from lbry.wallet import Ledger, RegTestLedger
|
2019-06-21 02:55:47 +02:00
|
|
|
from lbry.wallet.transaction import Transaction, Output
|
2019-07-11 19:29:26 +02:00
|
|
|
from lbry.wallet.server.db.canonical import register_canonical_functions
|
2019-11-25 19:09:44 +01:00
|
|
|
from lbry.wallet.server.db.full_text_search import update_full_text_search, CREATE_FULL_TEXT_SEARCH, first_sync_finished
|
2019-07-11 19:29:26 +02:00
|
|
|
from lbry.wallet.server.db.trending import (
|
2019-05-19 21:57:39 +02:00
|
|
|
CREATE_TREND_TABLE, calculate_trending, register_trending_functions
|
|
|
|
)
|
2019-03-31 00:40:01 +01:00
|
|
|
|
2019-07-20 07:14:06 +02:00
|
|
|
from .common import CLAIM_TYPES, STREAM_TYPES, COMMON_TAGS
|
2019-03-31 00:40:01 +01:00
|
|
|
|
2019-05-06 23:15:25 +02:00
|
|
|
|
2019-07-11 19:29:26 +02:00
|
|
|
ATTRIBUTE_ARRAY_MAX_LENGTH = 100
|
2019-05-06 23:15:25 +02:00
|
|
|
|
|
|
|
|
2019-03-31 00:40:01 +01:00
|
|
|
class SQLDB:
|
|
|
|
|
|
|
|
PRAGMAS = """
|
|
|
|
pragma journal_mode=WAL;
|
|
|
|
"""
|
|
|
|
|
|
|
|
CREATE_CLAIM_TABLE = """
|
|
|
|
create table if not exists claim (
|
|
|
|
claim_hash bytes primary key,
|
2019-05-25 04:40:39 +02:00
|
|
|
claim_id text not null,
|
2019-04-29 06:38:58 +02:00
|
|
|
claim_name text not null,
|
2019-05-24 05:55:57 +02:00
|
|
|
normalized text not null,
|
2019-03-31 00:40:01 +01:00
|
|
|
txo_hash bytes not null,
|
2019-04-29 06:38:58 +02:00
|
|
|
tx_position integer not null,
|
2019-05-25 04:40:39 +02:00
|
|
|
amount integer not null,
|
2019-05-26 05:06:22 +02:00
|
|
|
timestamp integer not null, -- last updated timestamp
|
|
|
|
creation_timestamp integer not null,
|
|
|
|
height integer not null, -- last updated height
|
2019-05-25 04:40:39 +02:00
|
|
|
creation_height integer not null,
|
2019-04-29 06:38:58 +02:00
|
|
|
activation_height integer,
|
2019-05-25 04:40:39 +02:00
|
|
|
expiration_height integer not null,
|
|
|
|
release_time integer not null,
|
|
|
|
|
2019-05-26 05:06:22 +02:00
|
|
|
short_url text not null, -- normalized#shortest-unique-claim_id
|
|
|
|
canonical_url text, -- channel's-short_url/normalized#shortest-unique-claim_id-within-channel
|
|
|
|
|
2019-11-14 20:31:49 +01:00
|
|
|
title text,
|
|
|
|
author text,
|
|
|
|
description text,
|
|
|
|
|
2019-06-04 06:10:59 +02:00
|
|
|
claim_type integer,
|
2019-11-18 21:48:52 +01:00
|
|
|
reposted integer default 0,
|
2019-06-04 06:10:59 +02:00
|
|
|
|
|
|
|
-- streams
|
|
|
|
stream_type text,
|
|
|
|
media_type text,
|
2019-06-23 04:25:22 +02:00
|
|
|
fee_amount integer default 0,
|
|
|
|
fee_currency text,
|
2019-06-04 06:10:59 +02:00
|
|
|
|
2019-10-15 07:21:05 +02:00
|
|
|
-- reposts
|
|
|
|
reposted_claim_hash bytes,
|
|
|
|
|
2019-05-26 05:06:22 +02:00
|
|
|
-- claims which are channels
|
|
|
|
public_key_bytes bytes,
|
2019-06-03 22:37:21 +02:00
|
|
|
public_key_hash bytes,
|
2019-05-26 05:06:22 +02:00
|
|
|
claims_in_channel integer,
|
|
|
|
|
|
|
|
-- claims which are inside channels
|
2019-05-25 04:40:39 +02:00
|
|
|
channel_hash bytes,
|
2019-05-26 05:06:22 +02:00
|
|
|
channel_join integer, -- height at which claim got valid signature / joined channel
|
|
|
|
signature bytes,
|
|
|
|
signature_digest bytes,
|
2019-06-04 20:16:11 +02:00
|
|
|
signature_valid bool,
|
2019-05-25 04:40:39 +02:00
|
|
|
|
2019-03-31 00:40:01 +01:00
|
|
|
effective_amount integer not null default 0,
|
2019-05-15 05:59:21 +02:00
|
|
|
support_amount integer not null default 0,
|
2019-05-19 01:54:13 +02:00
|
|
|
trending_group integer not null default 0,
|
|
|
|
trending_mixed integer not null default 0,
|
|
|
|
trending_local integer not null default 0,
|
|
|
|
trending_global integer not null default 0
|
2019-03-31 00:40:01 +01:00
|
|
|
);
|
2019-05-19 01:54:13 +02:00
|
|
|
|
2019-07-20 07:14:06 +02:00
|
|
|
create index if not exists claim_normalized_idx on claim (normalized, activation_height);
|
|
|
|
create index if not exists claim_channel_hash_idx on claim (channel_hash, signature, claim_hash);
|
2019-06-05 02:38:34 +02:00
|
|
|
create index if not exists claim_claims_in_channel_idx on claim (signature_valid, channel_hash, normalized);
|
2019-03-31 00:40:01 +01:00
|
|
|
create index if not exists claim_txo_hash_idx on claim (txo_hash);
|
2019-07-20 07:14:06 +02:00
|
|
|
create index if not exists claim_activation_height_idx on claim (activation_height, claim_hash);
|
2019-06-05 01:16:11 +02:00
|
|
|
create index if not exists claim_expiration_height_idx on claim (expiration_height);
|
2019-05-19 01:54:13 +02:00
|
|
|
"""
|
|
|
|
|
2019-03-31 00:40:01 +01:00
|
|
|
CREATE_SUPPORT_TABLE = """
|
|
|
|
create table if not exists support (
|
|
|
|
txo_hash bytes primary key,
|
2019-04-29 06:38:58 +02:00
|
|
|
tx_position integer not null,
|
2019-03-31 00:40:01 +01:00
|
|
|
height integer not null,
|
2019-04-29 06:38:58 +02:00
|
|
|
claim_hash bytes not null,
|
|
|
|
amount integer not null
|
2019-03-31 00:40:01 +01:00
|
|
|
);
|
|
|
|
create index if not exists support_claim_hash_idx on support (claim_hash, height);
|
|
|
|
"""
|
|
|
|
|
|
|
|
CREATE_TAG_TABLE = """
|
|
|
|
create table if not exists tag (
|
|
|
|
tag text not null,
|
2019-05-15 05:59:21 +02:00
|
|
|
claim_hash bytes not null,
|
2019-03-31 00:40:01 +01:00
|
|
|
height integer not null
|
|
|
|
);
|
2019-07-16 02:43:30 +02:00
|
|
|
create unique index if not exists tag_claim_hash_tag_idx on tag (claim_hash, tag);
|
2019-03-31 00:40:01 +01:00
|
|
|
"""
|
|
|
|
|
|
|
|
CREATE_CLAIMTRIE_TABLE = """
|
|
|
|
create table if not exists claimtrie (
|
2019-04-29 06:38:58 +02:00
|
|
|
normalized text primary key,
|
2019-03-31 00:40:01 +01:00
|
|
|
claim_hash bytes not null,
|
|
|
|
last_take_over_height integer not null
|
|
|
|
);
|
|
|
|
create index if not exists claimtrie_claim_hash_idx on claimtrie (claim_hash);
|
|
|
|
"""
|
|
|
|
|
2019-07-20 07:14:06 +02:00
|
|
|
SEARCH_INDEXES = """
|
|
|
|
-- used by any tag clouds
|
|
|
|
create index if not exists tag_tag_idx on tag (tag, claim_hash);
|
|
|
|
|
|
|
|
-- common ORDER BY
|
|
|
|
create unique index if not exists claim_effective_amount_idx on claim (effective_amount, claim_hash, release_time);
|
|
|
|
create unique index if not exists claim_release_time_idx on claim (release_time, claim_hash);
|
|
|
|
create unique index if not exists claim_trending_global_mixed_idx on claim (trending_global, trending_mixed, claim_hash);
|
2019-11-04 16:26:21 +01:00
|
|
|
create unique index if not exists claim_trending_group_mixed_idx on claim (trending_group, trending_mixed, claim_hash);
|
2019-07-24 20:27:55 +02:00
|
|
|
create unique index if not exists filter_fee_amount_order_release_time_idx on claim (fee_amount, release_time, claim_hash);
|
2019-07-20 07:14:06 +02:00
|
|
|
|
2019-07-29 20:11:12 +02:00
|
|
|
create unique index if not exists claim_type_trending_idx on claim (claim_type, trending_global, trending_mixed, claim_hash);
|
|
|
|
|
2019-07-20 07:14:06 +02:00
|
|
|
-- TODO: verify that all indexes below are used
|
|
|
|
create index if not exists claim_height_normalized_idx on claim (height, normalized asc);
|
|
|
|
|
|
|
|
create index if not exists claim_resolve_idx on claim (normalized, claim_id);
|
|
|
|
|
|
|
|
create index if not exists claim_id_idx on claim (claim_id, claim_hash);
|
|
|
|
create index if not exists claim_timestamp_idx on claim (timestamp);
|
|
|
|
create index if not exists claim_public_key_hash_idx on claim (public_key_hash);
|
|
|
|
|
|
|
|
create index if not exists claim_stream_type_idx on claim (stream_type);
|
|
|
|
create index if not exists claim_media_type_idx on claim (media_type);
|
|
|
|
|
|
|
|
create index if not exists claim_signature_valid_idx on claim (signature_valid);
|
2019-07-20 16:03:24 +02:00
|
|
|
"""
|
|
|
|
|
|
|
|
TAG_INDEXES = '\n'.join(
|
|
|
|
f"create unique index if not exists tag_{tag_key}_idx on tag (tag, claim_hash) WHERE tag='{tag_value}';"
|
2019-07-20 07:14:06 +02:00
|
|
|
for tag_value, tag_key in COMMON_TAGS.items()
|
2019-07-20 16:03:24 +02:00
|
|
|
)
|
2019-07-20 07:14:06 +02:00
|
|
|
|
2019-03-31 00:40:01 +01:00
|
|
|
CREATE_TABLES_QUERY = (
|
|
|
|
CREATE_CLAIM_TABLE +
|
2019-05-19 01:54:13 +02:00
|
|
|
CREATE_TREND_TABLE +
|
2019-11-14 20:31:49 +01:00
|
|
|
CREATE_FULL_TEXT_SEARCH +
|
2019-03-31 00:40:01 +01:00
|
|
|
CREATE_SUPPORT_TABLE +
|
|
|
|
CREATE_CLAIMTRIE_TABLE +
|
|
|
|
CREATE_TAG_TABLE
|
|
|
|
)
|
|
|
|
|
2020-01-10 16:47:57 +01:00
|
|
|
def __init__(self, main, path: str, filtering_channels: list):
|
2019-05-15 05:59:21 +02:00
|
|
|
self.main = main
|
2019-03-31 00:40:01 +01:00
|
|
|
self._db_path = path
|
|
|
|
self.db = None
|
2020-01-10 16:47:57 +01:00
|
|
|
self.state_manager = None
|
|
|
|
self.blocked_claims = None
|
2019-03-31 00:40:01 +01:00
|
|
|
self.logger = class_logger(__name__, self.__class__.__name__)
|
2020-01-10 16:47:57 +01:00
|
|
|
self.ledger = Ledger if main.coin.NET == 'mainnet' else RegTestLedger
|
2019-11-25 19:09:44 +01:00
|
|
|
self._fts_synced = False
|
2020-01-10 16:47:57 +01:00
|
|
|
self.filtering_channel_hashes = {
|
|
|
|
unhexlify(channel_id)[::-1] for channel_id in filtering_channels if channel_id
|
|
|
|
}
|
2019-03-31 00:40:01 +01:00
|
|
|
|
|
|
|
def open(self):
|
2019-12-08 00:13:13 +01:00
|
|
|
self.db = apsw.Connection(
|
|
|
|
self._db_path,
|
|
|
|
flags=(
|
|
|
|
apsw.SQLITE_OPEN_READWRITE |
|
|
|
|
apsw.SQLITE_OPEN_CREATE |
|
|
|
|
apsw.SQLITE_OPEN_URI
|
|
|
|
)
|
|
|
|
)
|
|
|
|
def exec_factory(cursor, statement, bindings):
|
|
|
|
tpl = namedtuple('row', (d[0] for d in cursor.getdescription()))
|
|
|
|
cursor.setrowtrace(lambda cursor, row: tpl(*row))
|
|
|
|
return True
|
|
|
|
self.db.setexectrace(exec_factory)
|
2019-12-21 09:44:39 +01:00
|
|
|
self.execute(self.PRAGMAS)
|
2019-12-08 00:13:13 +01:00
|
|
|
self.execute(self.CREATE_TABLES_QUERY)
|
2019-05-24 05:55:57 +02:00
|
|
|
register_canonical_functions(self.db)
|
2019-05-19 21:57:39 +02:00
|
|
|
register_trending_functions(self.db)
|
2020-01-10 16:47:57 +01:00
|
|
|
self.state_manager = Manager()
|
|
|
|
self.blocked_claims = self.state_manager.dict()
|
|
|
|
self.update_blocked_claims()
|
2019-03-31 00:40:01 +01:00
|
|
|
|
|
|
|
def close(self):
|
2019-12-08 00:13:13 +01:00
|
|
|
if self.db is not None:
|
|
|
|
self.db.close()
|
2020-01-10 16:47:57 +01:00
|
|
|
if self.state_manager is not None:
|
|
|
|
self.state_manager.shutdown()
|
|
|
|
|
|
|
|
def update_blocked_claims(self):
|
|
|
|
sql = query(
|
|
|
|
"SELECT channel_hash, reposted_claim_hash FROM claim",
|
|
|
|
reposted_claim_hash__is_not_null=1,
|
|
|
|
channel_hash__in=self.filtering_channel_hashes
|
|
|
|
)
|
|
|
|
blocked_claims = {}
|
|
|
|
for blocked_claim in self.execute(*sql):
|
|
|
|
blocked_claims[blocked_claim.reposted_claim_hash] = blocked_claim.channel_hash
|
|
|
|
self.blocked_claims.clear()
|
|
|
|
self.blocked_claims.update(blocked_claims)
|
2019-03-31 00:40:01 +01:00
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def _insert_sql(table: str, data: dict) -> Tuple[str, list]:
|
|
|
|
columns, values = [], []
|
|
|
|
for column, value in data.items():
|
|
|
|
columns.append(column)
|
|
|
|
values.append(value)
|
|
|
|
sql = (
|
|
|
|
f"INSERT INTO {table} ({', '.join(columns)}) "
|
|
|
|
f"VALUES ({', '.join(['?'] * len(values))})"
|
|
|
|
)
|
|
|
|
return sql, values
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def _update_sql(table: str, data: dict, where: str,
|
|
|
|
constraints: Union[list, tuple]) -> Tuple[str, list]:
|
|
|
|
columns, values = [], []
|
|
|
|
for column, value in data.items():
|
2019-10-08 18:19:01 +02:00
|
|
|
columns.append(f"{column} = ?")
|
2019-03-31 00:40:01 +01:00
|
|
|
values.append(value)
|
|
|
|
values.extend(constraints)
|
|
|
|
return f"UPDATE {table} SET {', '.join(columns)} WHERE {where}", values
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def _delete_sql(table: str, constraints: dict) -> Tuple[str, dict]:
|
|
|
|
where, values = constraints_to_sql(constraints)
|
|
|
|
return f"DELETE FROM {table} WHERE {where}", values
|
|
|
|
|
|
|
|
def execute(self, *args):
|
2019-12-08 00:13:13 +01:00
|
|
|
return self.db.cursor().execute(*args)
|
|
|
|
|
|
|
|
def executemany(self, *args):
|
|
|
|
return self.db.cursor().executemany(*args)
|
2019-03-31 00:40:01 +01:00
|
|
|
|
|
|
|
def begin(self):
|
|
|
|
self.execute('begin;')
|
|
|
|
|
|
|
|
def commit(self):
|
|
|
|
self.execute('commit;')
|
|
|
|
|
2019-05-26 05:06:22 +02:00
|
|
|
def _upsertable_claims(self, txos: List[Output], header, clear_first=False):
|
2019-12-08 00:13:13 +01:00
|
|
|
claim_hashes, claims, tags = set(), [], {}
|
2019-03-31 00:40:01 +01:00
|
|
|
for txo in txos:
|
|
|
|
tx = txo.tx_ref.tx
|
2019-04-29 06:38:58 +02:00
|
|
|
|
2019-03-31 00:40:01 +01:00
|
|
|
try:
|
|
|
|
assert txo.claim_name
|
2019-04-29 06:38:58 +02:00
|
|
|
assert txo.normalized_name
|
2019-05-14 04:46:52 +02:00
|
|
|
except:
|
|
|
|
#self.logger.exception(f"Could not decode claim name for {tx.id}:{txo.position}.")
|
2019-03-31 00:40:01 +01:00
|
|
|
continue
|
2019-04-29 06:38:58 +02:00
|
|
|
|
2019-12-08 00:13:13 +01:00
|
|
|
claim_hash = txo.claim_hash
|
|
|
|
claim_hashes.add(claim_hash)
|
2019-04-29 06:38:58 +02:00
|
|
|
claim_record = {
|
2019-05-15 05:59:21 +02:00
|
|
|
'claim_hash': claim_hash,
|
2019-05-25 04:40:39 +02:00
|
|
|
'claim_id': txo.claim_id,
|
2019-04-29 06:38:58 +02:00
|
|
|
'claim_name': txo.claim_name,
|
2019-05-25 04:40:39 +02:00
|
|
|
'normalized': txo.normalized_name,
|
2019-12-08 00:13:13 +01:00
|
|
|
'txo_hash': txo.ref.hash,
|
2019-04-29 06:38:58 +02:00
|
|
|
'tx_position': tx.position,
|
|
|
|
'amount': txo.amount,
|
2019-05-25 04:40:39 +02:00
|
|
|
'timestamp': header['timestamp'],
|
|
|
|
'height': tx.height,
|
2019-11-14 20:31:49 +01:00
|
|
|
'title': None,
|
|
|
|
'description': None,
|
|
|
|
'author': None,
|
2019-06-04 06:10:59 +02:00
|
|
|
'claim_type': None,
|
|
|
|
'stream_type': None,
|
|
|
|
'media_type': None,
|
2019-05-25 04:40:39 +02:00
|
|
|
'release_time': None,
|
2019-06-23 04:25:22 +02:00
|
|
|
'fee_currency': None,
|
2019-11-14 20:31:49 +01:00
|
|
|
'fee_amount': 0,
|
2019-10-15 07:21:05 +02:00
|
|
|
'reposted_claim_hash': None
|
2019-04-29 06:38:58 +02:00
|
|
|
}
|
|
|
|
claims.append(claim_record)
|
|
|
|
|
2019-03-31 00:40:01 +01:00
|
|
|
try:
|
|
|
|
claim = txo.claim
|
2019-05-14 04:46:52 +02:00
|
|
|
except:
|
|
|
|
#self.logger.exception(f"Could not parse claim protobuf for {tx.id}:{txo.position}.")
|
2019-03-31 00:40:01 +01:00
|
|
|
continue
|
2019-04-29 06:38:58 +02:00
|
|
|
|
2019-05-25 04:40:39 +02:00
|
|
|
if claim.is_stream:
|
2019-06-04 06:10:59 +02:00
|
|
|
claim_record['claim_type'] = CLAIM_TYPES['stream']
|
|
|
|
claim_record['media_type'] = claim.stream.source.media_type
|
|
|
|
claim_record['stream_type'] = STREAM_TYPES[guess_stream_type(claim_record['media_type'])]
|
2019-11-14 20:31:49 +01:00
|
|
|
claim_record['title'] = claim.stream.title
|
|
|
|
claim_record['description'] = claim.stream.description
|
|
|
|
claim_record['author'] = claim.stream.author
|
2019-05-25 04:40:39 +02:00
|
|
|
if claim.stream.release_time:
|
|
|
|
claim_record['release_time'] = claim.stream.release_time
|
2019-06-23 04:25:22 +02:00
|
|
|
if claim.stream.has_fee:
|
2019-06-26 16:44:25 +02:00
|
|
|
fee = claim.stream.fee
|
|
|
|
if isinstance(fee.currency, str):
|
|
|
|
claim_record['fee_currency'] = fee.currency.lower()
|
|
|
|
if isinstance(fee.amount, Decimal):
|
|
|
|
claim_record['fee_amount'] = int(fee.amount*1000)
|
2019-10-15 07:21:05 +02:00
|
|
|
elif claim.is_repost:
|
|
|
|
claim_record['reposted_claim_hash'] = claim.repost.reference.claim_hash
|
2019-05-25 04:40:39 +02:00
|
|
|
elif claim.is_channel:
|
2019-06-04 06:10:59 +02:00
|
|
|
claim_record['claim_type'] = CLAIM_TYPES['channel']
|
2019-05-25 04:40:39 +02:00
|
|
|
|
2019-06-23 02:11:33 +02:00
|
|
|
for tag in clean_tags(claim.message.tags):
|
2019-07-16 02:43:30 +02:00
|
|
|
tags[(tag, claim_hash)] = (tag, claim_hash, tx.height)
|
2019-05-15 05:59:21 +02:00
|
|
|
|
|
|
|
if clear_first:
|
|
|
|
self._clear_claim_metadata(claim_hashes)
|
2019-04-29 06:38:58 +02:00
|
|
|
|
2019-03-31 00:40:01 +01:00
|
|
|
if tags:
|
2019-12-08 00:13:13 +01:00
|
|
|
self.executemany(
|
2019-07-18 03:50:20 +02:00
|
|
|
"INSERT OR IGNORE INTO tag (tag, claim_hash, height) VALUES (?, ?, ?)", tags.values()
|
2019-03-31 00:40:01 +01:00
|
|
|
)
|
2019-04-29 06:38:58 +02:00
|
|
|
|
2019-03-31 00:40:01 +01:00
|
|
|
return claims
|
2018-11-04 10:42:47 +01:00
|
|
|
|
2019-05-26 05:06:22 +02:00
|
|
|
def insert_claims(self, txos: List[Output], header):
|
|
|
|
claims = self._upsertable_claims(txos, header)
|
2019-03-31 00:40:01 +01:00
|
|
|
if claims:
|
2019-12-08 00:13:13 +01:00
|
|
|
self.executemany("""
|
2019-06-19 00:52:33 +02:00
|
|
|
INSERT OR IGNORE INTO claim (
|
2019-05-26 05:06:22 +02:00
|
|
|
claim_hash, claim_id, claim_name, normalized, txo_hash, tx_position, amount,
|
2019-06-23 04:25:22 +02:00
|
|
|
claim_type, media_type, stream_type, timestamp, creation_timestamp,
|
2019-10-15 07:21:05 +02:00
|
|
|
fee_currency, fee_amount, title, description, author, height, reposted_claim_hash,
|
2019-06-04 06:10:59 +02:00
|
|
|
creation_height, release_time, activation_height, expiration_height, short_url)
|
2019-05-18 05:54:03 +02:00
|
|
|
VALUES (
|
2019-05-26 05:06:22 +02:00
|
|
|
:claim_hash, :claim_id, :claim_name, :normalized, :txo_hash, :tx_position, :amount,
|
2019-06-23 04:25:22 +02:00
|
|
|
:claim_type, :media_type, :stream_type, :timestamp, :timestamp,
|
2019-10-15 07:21:05 +02:00
|
|
|
:fee_currency, :fee_amount, :title, :description, :author, :height, :reposted_claim_hash, :height,
|
2019-05-25 04:40:39 +02:00
|
|
|
CASE WHEN :release_time IS NOT NULL THEN :release_time ELSE :timestamp END,
|
2019-05-24 05:55:57 +02:00
|
|
|
CASE WHEN :normalized NOT IN (SELECT normalized FROM claimtrie) THEN :height END,
|
2019-06-05 03:42:15 +02:00
|
|
|
CASE WHEN :height >= 137181 THEN :height+2102400 ELSE :height+262974 END,
|
2019-05-31 03:32:35 +02:00
|
|
|
:claim_name||COALESCE(
|
2019-05-25 04:40:39 +02:00
|
|
|
(SELECT shortest_id(claim_id, :claim_id) FROM claim WHERE normalized = :normalized),
|
|
|
|
'#'||substr(:claim_id, 1, 1)
|
2019-05-26 05:06:22 +02:00
|
|
|
)
|
2019-05-18 05:54:03 +02:00
|
|
|
)""", claims)
|
|
|
|
|
2019-05-26 05:06:22 +02:00
|
|
|
def update_claims(self, txos: List[Output], header):
|
|
|
|
claims = self._upsertable_claims(txos, header, clear_first=True)
|
2019-03-31 00:40:01 +01:00
|
|
|
if claims:
|
2019-12-08 00:13:13 +01:00
|
|
|
self.executemany("""
|
2019-05-26 05:06:22 +02:00
|
|
|
UPDATE claim SET
|
2019-06-04 06:10:59 +02:00
|
|
|
txo_hash=:txo_hash, tx_position=:tx_position, amount=:amount, height=:height,
|
|
|
|
claim_type=:claim_type, media_type=:media_type, stream_type=:stream_type,
|
2019-06-23 04:25:22 +02:00
|
|
|
timestamp=:timestamp, fee_amount=:fee_amount, fee_currency=:fee_currency,
|
2019-10-15 07:21:05 +02:00
|
|
|
title=:title, description=:description, author=:author, reposted_claim_hash=:reposted_claim_hash,
|
2019-05-26 05:06:22 +02:00
|
|
|
release_time=CASE WHEN :release_time IS NOT NULL THEN :release_time ELSE release_time END
|
2019-05-25 04:40:39 +02:00
|
|
|
WHERE claim_hash=:claim_hash;
|
|
|
|
""", claims)
|
2019-03-31 00:40:01 +01:00
|
|
|
|
2019-05-15 05:59:21 +02:00
|
|
|
def delete_claims(self, claim_hashes: Set[bytes]):
|
2019-03-31 00:40:01 +01:00
|
|
|
""" Deletes claim supports and from claimtrie in case of an abandon. """
|
|
|
|
if claim_hashes:
|
2019-06-23 01:25:32 +02:00
|
|
|
affected_channels = self.execute(*query(
|
2019-12-08 00:13:13 +01:00
|
|
|
"SELECT channel_hash FROM claim", channel_hash__is_not_null=1, claim_hash__in=claim_hashes
|
2019-06-23 01:25:32 +02:00
|
|
|
)).fetchall()
|
2019-03-31 00:40:01 +01:00
|
|
|
for table in ('claim', 'support', 'claimtrie'):
|
2019-12-08 00:13:13 +01:00
|
|
|
self.execute(*self._delete_sql(table, {'claim_hash__in': claim_hashes}))
|
|
|
|
self._clear_claim_metadata(claim_hashes)
|
|
|
|
return {r.channel_hash for r in affected_channels}
|
2019-06-23 01:25:32 +02:00
|
|
|
return set()
|
2019-03-31 00:40:01 +01:00
|
|
|
|
2019-12-08 00:13:13 +01:00
|
|
|
def _clear_claim_metadata(self, claim_hashes: Set[bytes]):
|
|
|
|
if claim_hashes:
|
2019-05-15 05:59:21 +02:00
|
|
|
for table in ('tag',): # 'language', 'location', etc
|
2019-12-08 00:13:13 +01:00
|
|
|
self.execute(*self._delete_sql(table, {'claim_hash__in': claim_hashes}))
|
2019-05-15 05:59:21 +02:00
|
|
|
|
|
|
|
def split_inputs_into_claims_supports_and_other(self, txis):
|
2019-05-19 01:54:13 +02:00
|
|
|
txo_hashes = {txi.txo_ref.hash for txi in txis}
|
|
|
|
claims = self.execute(*query(
|
2019-12-08 00:13:13 +01:00
|
|
|
"SELECT txo_hash, claim_hash, normalized FROM claim", txo_hash__in=txo_hashes
|
2019-05-19 01:54:13 +02:00
|
|
|
)).fetchall()
|
2019-12-08 00:13:13 +01:00
|
|
|
txo_hashes -= {r.txo_hash for r in claims}
|
2019-05-15 05:59:21 +02:00
|
|
|
supports = {}
|
|
|
|
if txo_hashes:
|
2019-05-19 01:54:13 +02:00
|
|
|
supports = self.execute(*query(
|
2019-12-08 00:13:13 +01:00
|
|
|
"SELECT txo_hash, claim_hash FROM support", txo_hash__in=txo_hashes
|
2019-05-19 01:54:13 +02:00
|
|
|
)).fetchall()
|
2019-12-08 00:13:13 +01:00
|
|
|
txo_hashes -= {r.txo_hash for r in supports}
|
2019-05-15 05:59:21 +02:00
|
|
|
return claims, supports, txo_hashes
|
2019-03-31 00:40:01 +01:00
|
|
|
|
2019-05-26 05:06:22 +02:00
|
|
|
def insert_supports(self, txos: List[Output]):
|
2019-03-31 00:40:01 +01:00
|
|
|
supports = []
|
|
|
|
for txo in txos:
|
|
|
|
tx = txo.tx_ref.tx
|
|
|
|
supports.append((
|
2019-12-08 00:13:13 +01:00
|
|
|
txo.ref.hash, tx.position, tx.height,
|
|
|
|
txo.claim_hash, txo.amount
|
2019-03-31 00:40:01 +01:00
|
|
|
))
|
|
|
|
if supports:
|
2019-12-08 00:13:13 +01:00
|
|
|
self.executemany(
|
2019-06-23 16:56:59 +02:00
|
|
|
"INSERT OR IGNORE INTO support ("
|
2019-04-29 06:38:58 +02:00
|
|
|
" txo_hash, tx_position, height, claim_hash, amount"
|
|
|
|
") "
|
|
|
|
"VALUES (?, ?, ?, ?, ?)", supports
|
2019-03-31 00:40:01 +01:00
|
|
|
)
|
|
|
|
|
2019-05-15 05:59:21 +02:00
|
|
|
def delete_supports(self, txo_hashes: Set[bytes]):
|
2019-03-31 00:40:01 +01:00
|
|
|
if txo_hashes:
|
2019-12-08 00:13:13 +01:00
|
|
|
self.execute(*self._delete_sql('support', {'txo_hash__in': txo_hashes}))
|
2019-03-31 00:40:01 +01:00
|
|
|
|
2019-11-18 22:23:03 +01:00
|
|
|
def calculate_reposts(self, txos: List[Output]):
|
2019-11-18 21:48:52 +01:00
|
|
|
targets = set()
|
2019-11-18 22:23:03 +01:00
|
|
|
for txo in txos:
|
|
|
|
try:
|
|
|
|
claim = txo.claim
|
|
|
|
except:
|
|
|
|
continue
|
|
|
|
if claim.is_repost:
|
|
|
|
targets.add((claim.repost.reference.claim_hash,))
|
2019-11-18 21:48:52 +01:00
|
|
|
if targets:
|
2019-12-08 00:13:13 +01:00
|
|
|
self.executemany(
|
2019-11-18 21:48:52 +01:00
|
|
|
"""
|
|
|
|
UPDATE claim SET reposted = (
|
|
|
|
SELECT count(*) FROM claim AS repost WHERE repost.reposted_claim_hash = claim.claim_hash
|
|
|
|
)
|
|
|
|
WHERE claim_hash = ?
|
|
|
|
""", targets
|
|
|
|
)
|
|
|
|
|
2019-06-23 01:25:32 +02:00
|
|
|
def validate_channel_signatures(self, height, new_claims, updated_claims, spent_claims, affected_channels, timer):
|
2019-05-28 04:20:21 +02:00
|
|
|
if not new_claims and not updated_claims and not spent_claims:
|
2019-05-26 05:06:22 +02:00
|
|
|
return
|
|
|
|
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer = timer.add_timer('segregate channels and signables')
|
|
|
|
sub_timer.start()
|
2019-05-26 05:06:22 +02:00
|
|
|
channels, new_channel_keys, signables = {}, {}, {}
|
|
|
|
for txo in chain(new_claims, updated_claims):
|
|
|
|
try:
|
|
|
|
claim = txo.claim
|
|
|
|
except:
|
|
|
|
continue
|
|
|
|
if claim.is_channel:
|
|
|
|
channels[txo.claim_hash] = txo
|
|
|
|
new_channel_keys[txo.claim_hash] = claim.channel.public_key_bytes
|
|
|
|
else:
|
|
|
|
signables[txo.claim_hash] = txo
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer.stop()
|
2019-05-26 05:06:22 +02:00
|
|
|
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer = timer.add_timer('make list of channels we need to lookup')
|
|
|
|
sub_timer.start()
|
2019-05-26 05:06:22 +02:00
|
|
|
missing_channel_keys = set()
|
|
|
|
for txo in signables.values():
|
|
|
|
claim = txo.claim
|
|
|
|
if claim.is_signed and claim.signing_channel_hash not in new_channel_keys:
|
|
|
|
missing_channel_keys.add(claim.signing_channel_hash)
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer.stop()
|
2019-05-26 05:06:22 +02:00
|
|
|
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer = timer.add_timer('lookup missing channels')
|
|
|
|
sub_timer.start()
|
2019-05-26 05:06:22 +02:00
|
|
|
all_channel_keys = {}
|
2019-06-23 01:25:32 +02:00
|
|
|
if new_channel_keys or missing_channel_keys or affected_channels:
|
2019-05-26 05:06:22 +02:00
|
|
|
all_channel_keys = dict(self.execute(*query(
|
|
|
|
"SELECT claim_hash, public_key_bytes FROM claim",
|
2019-12-08 00:13:13 +01:00
|
|
|
claim_hash__in=set(new_channel_keys) | missing_channel_keys | affected_channels
|
2019-05-26 05:06:22 +02:00
|
|
|
)))
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer.stop()
|
2019-05-26 05:06:22 +02:00
|
|
|
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer = timer.add_timer('prepare for updating claims')
|
|
|
|
sub_timer.start()
|
2019-05-26 05:06:22 +02:00
|
|
|
changed_channel_keys = {}
|
|
|
|
for claim_hash, new_key in new_channel_keys.items():
|
2019-05-26 05:50:37 +02:00
|
|
|
if claim_hash not in all_channel_keys or all_channel_keys[claim_hash] != new_key:
|
2019-05-26 05:06:22 +02:00
|
|
|
all_channel_keys[claim_hash] = new_key
|
|
|
|
changed_channel_keys[claim_hash] = new_key
|
|
|
|
|
|
|
|
claim_updates = []
|
|
|
|
|
|
|
|
for claim_hash, txo in signables.items():
|
|
|
|
claim = txo.claim
|
|
|
|
update = {
|
2019-12-08 00:13:13 +01:00
|
|
|
'claim_hash': claim_hash,
|
2019-05-26 05:06:22 +02:00
|
|
|
'channel_hash': None,
|
|
|
|
'signature': None,
|
|
|
|
'signature_digest': None,
|
2019-06-04 20:16:11 +02:00
|
|
|
'signature_valid': None
|
2019-05-26 05:06:22 +02:00
|
|
|
}
|
|
|
|
if claim.is_signed:
|
|
|
|
update.update({
|
2019-12-08 00:13:13 +01:00
|
|
|
'channel_hash': claim.signing_channel_hash,
|
|
|
|
'signature': txo.get_encoded_signature(),
|
|
|
|
'signature_digest': txo.get_signature_digest(self.ledger),
|
2019-06-04 20:16:11 +02:00
|
|
|
'signature_valid': 0
|
2019-05-26 05:06:22 +02:00
|
|
|
})
|
|
|
|
claim_updates.append(update)
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer.stop()
|
2019-05-26 05:06:22 +02:00
|
|
|
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer = timer.add_timer('find claims affected by a change in channel key')
|
|
|
|
sub_timer.start()
|
2019-05-26 05:06:22 +02:00
|
|
|
if changed_channel_keys:
|
|
|
|
sql = f"""
|
|
|
|
SELECT * FROM claim WHERE
|
|
|
|
channel_hash IN ({','.join('?' for _ in changed_channel_keys)}) AND
|
|
|
|
signature IS NOT NULL
|
|
|
|
"""
|
2019-12-08 00:25:13 +01:00
|
|
|
for affected_claim in self.execute(sql, changed_channel_keys.keys()):
|
2019-12-08 00:13:13 +01:00
|
|
|
if affected_claim.claim_hash not in signables:
|
2019-05-26 05:06:22 +02:00
|
|
|
claim_updates.append({
|
2019-12-08 00:13:13 +01:00
|
|
|
'claim_hash': affected_claim.claim_hash,
|
|
|
|
'channel_hash': affected_claim.channel_hash,
|
|
|
|
'signature': affected_claim.signature,
|
|
|
|
'signature_digest': affected_claim.signature_digest,
|
2019-06-04 20:16:11 +02:00
|
|
|
'signature_valid': 0
|
2019-05-26 05:06:22 +02:00
|
|
|
})
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer.stop()
|
2019-05-26 05:06:22 +02:00
|
|
|
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer = timer.add_timer('verify signatures')
|
|
|
|
sub_timer.start()
|
2019-05-26 05:06:22 +02:00
|
|
|
for update in claim_updates:
|
|
|
|
channel_pub_key = all_channel_keys.get(update['channel_hash'])
|
|
|
|
if channel_pub_key and update['signature']:
|
2019-06-04 20:16:11 +02:00
|
|
|
update['signature_valid'] = Output.is_signature_valid(
|
2019-05-26 05:06:22 +02:00
|
|
|
bytes(update['signature']), bytes(update['signature_digest']), channel_pub_key
|
|
|
|
)
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer.stop()
|
2019-05-26 05:06:22 +02:00
|
|
|
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer = timer.add_timer('update claims')
|
|
|
|
sub_timer.start()
|
2019-05-26 05:06:22 +02:00
|
|
|
if claim_updates:
|
2019-12-08 00:13:13 +01:00
|
|
|
self.executemany(f"""
|
2019-05-26 05:06:22 +02:00
|
|
|
UPDATE claim SET
|
|
|
|
channel_hash=:channel_hash, signature=:signature, signature_digest=:signature_digest,
|
2019-06-04 20:16:11 +02:00
|
|
|
signature_valid=:signature_valid,
|
2019-05-26 05:06:22 +02:00
|
|
|
channel_join=CASE
|
2019-10-23 19:34:10 +02:00
|
|
|
WHEN signature_valid=1 AND :signature_valid=1 AND channel_hash=:channel_hash THEN channel_join
|
2019-06-04 20:16:11 +02:00
|
|
|
WHEN :signature_valid=1 THEN {height}
|
2019-05-26 05:06:22 +02:00
|
|
|
END,
|
|
|
|
canonical_url=CASE
|
2019-10-23 19:34:10 +02:00
|
|
|
WHEN signature_valid=1 AND :signature_valid=1 AND channel_hash=:channel_hash THEN canonical_url
|
2019-06-04 20:16:11 +02:00
|
|
|
WHEN :signature_valid=1 THEN
|
2019-05-26 05:06:22 +02:00
|
|
|
(SELECT short_url FROM claim WHERE claim_hash=:channel_hash)||'/'||
|
2019-05-31 03:32:35 +02:00
|
|
|
claim_name||COALESCE(
|
2019-05-26 05:06:22 +02:00
|
|
|
(SELECT shortest_id(other_claim.claim_id, claim.claim_id) FROM claim AS other_claim
|
2019-06-05 02:38:34 +02:00
|
|
|
WHERE other_claim.signature_valid = 1 AND
|
2019-05-26 05:06:22 +02:00
|
|
|
other_claim.channel_hash = :channel_hash AND
|
2019-06-05 02:38:34 +02:00
|
|
|
other_claim.normalized = claim.normalized),
|
2019-05-26 05:06:22 +02:00
|
|
|
'#'||substr(claim_id, 1, 1)
|
|
|
|
)
|
|
|
|
END
|
|
|
|
WHERE claim_hash=:claim_hash;
|
|
|
|
""", claim_updates)
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer.stop()
|
2019-05-26 05:06:22 +02:00
|
|
|
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer = timer.add_timer('update claims affected by spent channels')
|
|
|
|
sub_timer.start()
|
2019-05-28 04:20:21 +02:00
|
|
|
if spent_claims:
|
|
|
|
self.execute(
|
|
|
|
f"""
|
2019-06-04 20:16:11 +02:00
|
|
|
UPDATE claim SET
|
|
|
|
signature_valid=CASE WHEN signature IS NOT NULL THEN 0 END,
|
|
|
|
channel_join=NULL, canonical_url=NULL
|
2019-05-28 04:20:21 +02:00
|
|
|
WHERE channel_hash IN ({','.join('?' for _ in spent_claims)})
|
2019-12-08 00:13:13 +01:00
|
|
|
""", spent_claims
|
2019-05-28 04:20:21 +02:00
|
|
|
)
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer.stop()
|
2019-05-28 04:20:21 +02:00
|
|
|
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer = timer.add_timer('update channels')
|
|
|
|
sub_timer.start()
|
2019-05-26 05:06:22 +02:00
|
|
|
if channels:
|
2019-12-08 00:13:13 +01:00
|
|
|
self.executemany(
|
2019-06-03 22:37:21 +02:00
|
|
|
"""
|
|
|
|
UPDATE claim SET
|
|
|
|
public_key_bytes=:public_key_bytes,
|
|
|
|
public_key_hash=:public_key_hash
|
|
|
|
WHERE claim_hash=:claim_hash""", [{
|
2019-12-08 00:13:13 +01:00
|
|
|
'claim_hash': claim_hash,
|
|
|
|
'public_key_bytes': txo.claim.channel.public_key_bytes,
|
|
|
|
'public_key_hash': self.ledger.address_to_hash160(
|
|
|
|
self.ledger.public_key_to_address(txo.claim.channel.public_key_bytes)
|
|
|
|
)
|
2019-05-26 05:06:22 +02:00
|
|
|
} for claim_hash, txo in channels.items()]
|
|
|
|
)
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer.stop()
|
2019-05-26 05:06:22 +02:00
|
|
|
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer = timer.add_timer('update claims_in_channel counts')
|
|
|
|
sub_timer.start()
|
2019-05-26 05:06:22 +02:00
|
|
|
if all_channel_keys:
|
2019-12-08 00:13:13 +01:00
|
|
|
self.executemany(f"""
|
2019-05-26 05:06:22 +02:00
|
|
|
UPDATE claim SET
|
|
|
|
claims_in_channel=(
|
|
|
|
SELECT COUNT(*) FROM claim AS claim_in_channel
|
2019-06-05 02:18:04 +02:00
|
|
|
WHERE claim_in_channel.signature_valid=1 AND
|
|
|
|
claim_in_channel.channel_hash=claim.claim_hash
|
2019-05-26 05:06:22 +02:00
|
|
|
)
|
|
|
|
WHERE claim_hash = ?
|
2019-12-08 00:13:13 +01:00
|
|
|
""", [(channel_hash,) for channel_hash in all_channel_keys.keys()])
|
2019-06-05 01:27:32 +02:00
|
|
|
sub_timer.stop()
|
2019-05-26 05:06:22 +02:00
|
|
|
|
2020-01-10 16:47:57 +01:00
|
|
|
sub_timer = timer.add_timer('update blocked claims list')
|
|
|
|
sub_timer.start()
|
|
|
|
if self.filtering_channel_hashes.intersection(all_channel_keys):
|
|
|
|
self.update_blocked_claims()
|
|
|
|
sub_timer.stop()
|
|
|
|
|
2019-05-15 05:59:21 +02:00
|
|
|
def _update_support_amount(self, claim_hashes):
|
|
|
|
if claim_hashes:
|
|
|
|
self.execute(f"""
|
|
|
|
UPDATE claim SET
|
|
|
|
support_amount = COALESCE(
|
|
|
|
(SELECT SUM(amount) FROM support WHERE support.claim_hash=claim.claim_hash), 0
|
|
|
|
)
|
|
|
|
WHERE claim_hash IN ({','.join('?' for _ in claim_hashes)})
|
2019-05-16 07:34:18 +02:00
|
|
|
""", claim_hashes)
|
2019-05-15 05:59:21 +02:00
|
|
|
|
2019-05-18 05:54:03 +02:00
|
|
|
def _update_effective_amount(self, height, claim_hashes=None):
|
|
|
|
self.execute(
|
|
|
|
f"UPDATE claim SET effective_amount = amount + support_amount "
|
|
|
|
f"WHERE activation_height = {height}"
|
|
|
|
)
|
|
|
|
if claim_hashes:
|
|
|
|
self.execute(
|
|
|
|
f"UPDATE claim SET effective_amount = amount + support_amount "
|
|
|
|
f"WHERE activation_height < {height} "
|
|
|
|
f" AND claim_hash IN ({','.join('?' for _ in claim_hashes)})",
|
|
|
|
claim_hashes
|
|
|
|
)
|
2019-05-16 07:34:18 +02:00
|
|
|
|
2019-05-18 05:54:03 +02:00
|
|
|
def _calculate_activation_height(self, height):
|
|
|
|
last_take_over_height = f"""COALESCE(
|
|
|
|
(SELECT last_take_over_height FROM claimtrie
|
|
|
|
WHERE claimtrie.normalized=claim.normalized),
|
|
|
|
{height}
|
|
|
|
)
|
|
|
|
"""
|
2019-05-16 07:34:18 +02:00
|
|
|
self.execute(f"""
|
|
|
|
UPDATE claim SET activation_height =
|
2019-05-18 05:54:03 +02:00
|
|
|
{height} + min(4032, cast(({height} - {last_take_over_height}) / 32 AS INT))
|
2019-05-16 07:34:18 +02:00
|
|
|
WHERE activation_height IS NULL
|
|
|
|
""")
|
2019-03-31 00:40:01 +01:00
|
|
|
|
2019-05-19 01:54:13 +02:00
|
|
|
def _perform_overtake(self, height, changed_claim_hashes, deleted_names):
|
|
|
|
deleted_names_sql = claim_hashes_sql = ""
|
|
|
|
if changed_claim_hashes:
|
|
|
|
claim_hashes_sql = f"OR claim_hash IN ({','.join('?' for _ in changed_claim_hashes)})"
|
|
|
|
if deleted_names:
|
|
|
|
deleted_names_sql = f"OR normalized IN ({','.join('?' for _ in deleted_names)})"
|
2019-05-16 07:34:18 +02:00
|
|
|
overtakes = self.execute(f"""
|
2019-05-20 20:11:53 +02:00
|
|
|
SELECT winner.normalized, winner.claim_hash,
|
|
|
|
claimtrie.claim_hash AS current_winner,
|
2019-12-08 00:13:13 +01:00
|
|
|
MAX(winner.effective_amount) AS max_winner_effective_amount
|
2019-05-20 20:11:53 +02:00
|
|
|
FROM (
|
|
|
|
SELECT normalized, claim_hash, effective_amount FROM claim
|
2019-05-19 01:54:13 +02:00
|
|
|
WHERE normalized IN (
|
2019-05-20 20:11:53 +02:00
|
|
|
SELECT normalized FROM claim WHERE activation_height={height} {claim_hashes_sql}
|
|
|
|
) {deleted_names_sql}
|
|
|
|
ORDER BY effective_amount DESC, height ASC, tx_position ASC
|
2019-05-18 05:54:03 +02:00
|
|
|
) AS winner LEFT JOIN claimtrie USING (normalized)
|
|
|
|
GROUP BY winner.normalized
|
2019-05-19 01:54:13 +02:00
|
|
|
HAVING current_winner IS NULL OR current_winner <> winner.claim_hash
|
2019-12-08 00:25:13 +01:00
|
|
|
""", list(changed_claim_hashes)+deleted_names)
|
2019-05-16 07:34:18 +02:00
|
|
|
for overtake in overtakes:
|
2019-12-08 00:13:13 +01:00
|
|
|
if overtake.current_winner:
|
2019-05-18 05:54:03 +02:00
|
|
|
self.execute(
|
|
|
|
f"UPDATE claimtrie SET claim_hash = ?, last_take_over_height = {height} "
|
|
|
|
f"WHERE normalized = ?",
|
2019-12-08 00:13:13 +01:00
|
|
|
(overtake.claim_hash, overtake.normalized)
|
2019-05-18 05:54:03 +02:00
|
|
|
)
|
|
|
|
else:
|
|
|
|
self.execute(
|
|
|
|
f"INSERT INTO claimtrie (claim_hash, normalized, last_take_over_height) "
|
|
|
|
f"VALUES (?, ?, {height})",
|
2019-12-08 00:13:13 +01:00
|
|
|
(overtake.claim_hash, overtake.normalized)
|
2019-05-18 05:54:03 +02:00
|
|
|
)
|
2019-03-31 00:40:01 +01:00
|
|
|
self.execute(
|
2019-05-06 01:16:17 +02:00
|
|
|
f"UPDATE claim SET activation_height = {height} WHERE normalized = ? "
|
|
|
|
f"AND (activation_height IS NULL OR activation_height > {height})",
|
2019-12-08 00:13:13 +01:00
|
|
|
(overtake.normalized,)
|
2019-03-31 00:40:01 +01:00
|
|
|
)
|
|
|
|
|
2019-05-15 17:37:53 +02:00
|
|
|
def _copy(self, height):
|
|
|
|
if height > 50:
|
|
|
|
self.execute(f"DROP TABLE claimtrie{height-50}")
|
|
|
|
self.execute(f"CREATE TABLE claimtrie{height} AS SELECT * FROM claimtrie")
|
|
|
|
|
2019-05-19 01:54:13 +02:00
|
|
|
def update_claimtrie(self, height, changed_claim_hashes, deleted_names, timer):
|
2019-05-19 21:57:39 +02:00
|
|
|
r = timer.run
|
|
|
|
|
2019-05-18 05:54:03 +02:00
|
|
|
r(self._calculate_activation_height, height)
|
2019-12-08 00:13:13 +01:00
|
|
|
r(self._update_support_amount, changed_claim_hashes)
|
2019-05-16 07:34:18 +02:00
|
|
|
|
2019-12-08 00:13:13 +01:00
|
|
|
r(self._update_effective_amount, height, changed_claim_hashes)
|
|
|
|
r(self._perform_overtake, height, changed_claim_hashes, list(deleted_names))
|
2019-05-16 07:34:18 +02:00
|
|
|
|
2019-05-18 05:54:03 +02:00
|
|
|
r(self._update_effective_amount, height)
|
2019-05-19 01:54:13 +02:00
|
|
|
r(self._perform_overtake, height, [], [])
|
2019-03-31 00:40:01 +01:00
|
|
|
|
2019-06-05 00:59:25 +02:00
|
|
|
def get_expiring(self, height):
|
|
|
|
return self.execute(
|
|
|
|
f"SELECT claim_hash, normalized FROM claim WHERE expiration_height = {height}"
|
|
|
|
)
|
|
|
|
|
2019-05-20 02:22:25 +02:00
|
|
|
def advance_txs(self, height, all_txs, header, daemon_height, timer):
|
2019-05-26 05:06:22 +02:00
|
|
|
insert_claims = []
|
|
|
|
update_claims = []
|
2019-12-14 23:30:42 +01:00
|
|
|
update_claim_hashes = set()
|
2019-05-18 05:54:03 +02:00
|
|
|
delete_claim_hashes = set()
|
2019-05-26 05:06:22 +02:00
|
|
|
insert_supports = []
|
2019-05-19 01:54:13 +02:00
|
|
|
delete_support_txo_hashes = set()
|
|
|
|
recalculate_claim_hashes = set() # added/deleted supports, added/updated claim
|
|
|
|
deleted_claim_names = set()
|
2019-05-26 05:06:22 +02:00
|
|
|
delete_others = set()
|
2019-05-16 07:34:18 +02:00
|
|
|
body_timer = timer.add_timer('body')
|
|
|
|
for position, (etx, txid) in enumerate(all_txs):
|
|
|
|
tx = timer.run(
|
2019-11-25 04:37:11 +01:00
|
|
|
Transaction, etx.raw, height=height, position=position
|
2019-05-16 07:34:18 +02:00
|
|
|
)
|
2019-05-19 01:54:13 +02:00
|
|
|
# Inputs
|
2019-05-26 05:06:22 +02:00
|
|
|
spent_claims, spent_supports, spent_others = timer.run(
|
2019-05-16 07:34:18 +02:00
|
|
|
self.split_inputs_into_claims_supports_and_other, tx.inputs
|
|
|
|
)
|
|
|
|
body_timer.start()
|
2019-12-08 00:13:13 +01:00
|
|
|
delete_claim_hashes.update({r.claim_hash for r in spent_claims})
|
|
|
|
deleted_claim_names.update({r.normalized for r in spent_claims})
|
|
|
|
delete_support_txo_hashes.update({r.txo_hash for r in spent_supports})
|
|
|
|
recalculate_claim_hashes.update({r.claim_hash for r in spent_supports})
|
2019-05-26 05:06:22 +02:00
|
|
|
delete_others.update(spent_others)
|
2019-05-19 01:54:13 +02:00
|
|
|
# Outputs
|
2019-05-16 07:34:18 +02:00
|
|
|
for output in tx.outputs:
|
|
|
|
if output.is_support:
|
2019-05-26 05:06:22 +02:00
|
|
|
insert_supports.append(output)
|
2019-05-19 01:54:13 +02:00
|
|
|
recalculate_claim_hashes.add(output.claim_hash)
|
2019-05-16 07:34:18 +02:00
|
|
|
elif output.script.is_claim_name:
|
2019-05-26 05:06:22 +02:00
|
|
|
insert_claims.append(output)
|
2019-05-19 01:54:13 +02:00
|
|
|
recalculate_claim_hashes.add(output.claim_hash)
|
2019-05-16 07:34:18 +02:00
|
|
|
elif output.script.is_update_claim:
|
|
|
|
claim_hash = output.claim_hash
|
2019-05-26 05:06:22 +02:00
|
|
|
update_claims.append(output)
|
|
|
|
recalculate_claim_hashes.add(claim_hash)
|
2019-05-16 07:34:18 +02:00
|
|
|
body_timer.stop()
|
2019-06-05 00:59:25 +02:00
|
|
|
|
2019-12-14 23:30:42 +01:00
|
|
|
skip_update_claim_timer = timer.add_timer('skip update of abandoned claims')
|
|
|
|
skip_update_claim_timer.start()
|
|
|
|
for updated_claim in list(update_claims):
|
|
|
|
if updated_claim.ref.hash in delete_others:
|
|
|
|
update_claims.remove(updated_claim)
|
2019-12-18 04:32:48 +01:00
|
|
|
for updated_claim in update_claims:
|
|
|
|
claim_hash = updated_claim.claim_hash
|
|
|
|
delete_claim_hashes.discard(claim_hash)
|
|
|
|
update_claim_hashes.add(claim_hash)
|
2019-12-14 23:30:42 +01:00
|
|
|
skip_update_claim_timer.stop()
|
|
|
|
|
|
|
|
skip_insert_claim_timer = timer.add_timer('skip insertion of abandoned claims')
|
|
|
|
skip_insert_claim_timer.start()
|
2019-05-26 05:06:22 +02:00
|
|
|
for new_claim in list(insert_claims):
|
|
|
|
if new_claim.ref.hash in delete_others:
|
2019-12-14 23:30:42 +01:00
|
|
|
if new_claim.claim_hash not in update_claim_hashes:
|
|
|
|
insert_claims.remove(new_claim)
|
|
|
|
skip_insert_claim_timer.stop()
|
2019-06-05 00:59:25 +02:00
|
|
|
|
2019-12-14 23:49:39 +01:00
|
|
|
skip_insert_support_timer = timer.add_timer('skip insertion of abandoned supports')
|
|
|
|
skip_insert_support_timer.start()
|
|
|
|
for new_support in list(insert_supports):
|
|
|
|
if new_support.ref.hash in delete_others:
|
|
|
|
insert_supports.remove(new_support)
|
|
|
|
skip_insert_support_timer.stop()
|
|
|
|
|
2019-06-05 00:59:25 +02:00
|
|
|
expire_timer = timer.add_timer('recording expired claims')
|
|
|
|
expire_timer.start()
|
|
|
|
for expired in self.get_expiring(height):
|
2019-12-08 00:13:13 +01:00
|
|
|
delete_claim_hashes.add(expired.claim_hash)
|
|
|
|
deleted_claim_names.add(expired.normalized)
|
2019-06-05 00:59:25 +02:00
|
|
|
expire_timer.stop()
|
|
|
|
|
2019-05-16 07:34:18 +02:00
|
|
|
r = timer.run
|
2019-11-14 20:31:49 +01:00
|
|
|
r(update_full_text_search, 'before-delete',
|
2019-12-08 00:13:13 +01:00
|
|
|
delete_claim_hashes, self.db.cursor(), self.main.first_sync)
|
2019-06-23 01:25:32 +02:00
|
|
|
affected_channels = r(self.delete_claims, delete_claim_hashes)
|
2019-05-19 01:54:13 +02:00
|
|
|
r(self.delete_supports, delete_support_txo_hashes)
|
2019-05-26 05:06:22 +02:00
|
|
|
r(self.insert_claims, insert_claims, header)
|
2019-11-18 21:48:52 +01:00
|
|
|
r(self.calculate_reposts, insert_claims)
|
2019-11-14 20:31:49 +01:00
|
|
|
r(update_full_text_search, 'after-insert',
|
2019-12-08 00:13:13 +01:00
|
|
|
[txo.claim_hash for txo in insert_claims], self.db.cursor(), self.main.first_sync)
|
2019-11-14 20:31:49 +01:00
|
|
|
r(update_full_text_search, 'before-update',
|
2019-12-08 00:13:13 +01:00
|
|
|
[txo.claim_hash for txo in update_claims], self.db.cursor(), self.main.first_sync)
|
2019-05-26 05:06:22 +02:00
|
|
|
r(self.update_claims, update_claims, header)
|
2019-11-14 20:31:49 +01:00
|
|
|
r(update_full_text_search, 'after-update',
|
2019-12-08 00:13:13 +01:00
|
|
|
[txo.claim_hash for txo in update_claims], self.db.cursor(), self.main.first_sync)
|
2019-06-05 01:27:32 +02:00
|
|
|
r(self.validate_channel_signatures, height, insert_claims,
|
2019-06-23 01:25:32 +02:00
|
|
|
update_claims, delete_claim_hashes, affected_channels, forward_timer=True)
|
2019-05-16 07:34:18 +02:00
|
|
|
r(self.insert_supports, insert_supports)
|
2019-05-19 01:54:13 +02:00
|
|
|
r(self.update_claimtrie, height, recalculate_claim_hashes, deleted_claim_names, forward_timer=True)
|
2019-12-08 00:13:13 +01:00
|
|
|
r(calculate_trending, self.db.cursor(), height, daemon_height)
|
2019-11-25 19:09:44 +01:00
|
|
|
if not self._fts_synced and self.main.first_sync and height == daemon_height:
|
2019-12-08 00:13:13 +01:00
|
|
|
r(first_sync_finished, self.db.cursor())
|
2019-11-25 19:09:44 +01:00
|
|
|
self._fts_synced = True
|
2019-05-16 07:34:18 +02:00
|
|
|
|
2018-11-04 10:42:47 +01:00
|
|
|
|
2020-01-10 16:47:57 +01:00
|
|
|
class LBRYLevelDB(LevelDB):
|
2018-11-04 10:42:47 +01:00
|
|
|
|
|
|
|
def __init__(self, *args, **kwargs):
|
|
|
|
super().__init__(*args, **kwargs)
|
2019-08-26 23:23:43 +02:00
|
|
|
path = os.path.join(self.env.db_dir, 'claims.db')
|
2020-01-10 16:47:57 +01:00
|
|
|
# space separated list of channel URIs used for filtering bad content
|
|
|
|
self.sql = SQLDB(self, path, self.env.default('FILTERING_CHANNELS_IDS', '').split(' '))
|
2018-11-04 10:42:47 +01:00
|
|
|
|
2018-12-15 21:31:02 +01:00
|
|
|
def close(self):
|
|
|
|
super().close()
|
2019-03-31 00:40:01 +01:00
|
|
|
self.sql.close()
|
2018-11-04 10:42:47 +01:00
|
|
|
|
2019-04-27 20:27:18 +02:00
|
|
|
async def _open_dbs(self, *args, **kwargs):
|
|
|
|
await super()._open_dbs(*args, **kwargs)
|
2019-03-31 00:40:01 +01:00
|
|
|
self.sql.open()
|