lbry-sdk/lbry/wallet/database.py

1306 lines
52 KiB
Python
Raw Normal View History

2020-02-21 04:11:25 +01:00
import os
2020-01-03 04:18:49 +01:00
import logging
import asyncio
import sqlite3
2020-03-19 18:24:25 +01:00
import platform
2020-01-03 04:18:49 +01:00
from binascii import hexlify
2020-06-04 16:18:04 +02:00
from collections import defaultdict
2020-02-21 04:11:25 +01:00
from dataclasses import dataclass
from contextvars import ContextVar
2020-01-03 04:18:49 +01:00
from typing import Tuple, List, Union, Callable, Any, Awaitable, Iterable, Dict, Optional
from datetime import date
2020-05-03 03:23:17 +02:00
from prometheus_client import Gauge, Counter, Histogram
from lbry.utils import LockWithMetrics
2018-06-12 17:53:29 +02:00
2020-01-03 04:18:49 +01:00
from .bip32 import PubKey
2020-06-04 16:18:04 +02:00
from .transaction import Transaction, Output, OutputScript, TXRefImmutable, Input
2020-01-03 04:18:49 +01:00
from .constants import TXO_TYPES, CLAIM_TYPES
from .util import date_to_julian_day
2019-03-24 21:55:04 +01:00
2020-09-29 13:52:28 +02:00
from concurrent.futures.thread import ThreadPoolExecutor # pylint: disable=wrong-import-order
if platform.system() == 'Windows' or 'ANDROID_ARGUMENT' or 'KIVY_BUILD' in os.environ:
from concurrent.futures.thread import ThreadPoolExecutor as ReaderExecutorClass # pylint: disable=reimported
else:
from concurrent.futures.process import ProcessPoolExecutor as ReaderExecutorClass
2018-06-12 17:53:29 +02:00
2020-01-03 04:18:49 +01:00
log = logging.getLogger(__name__)
sqlite3.enable_callback_tracebacks(True)
2018-06-12 17:53:29 +02:00
2020-05-03 04:30:25 +02:00
HISTOGRAM_BUCKETS = (
.005, .01, .025, .05, .075, .1, .25, .5, .75, 1.0, 2.5, 5.0, 7.5, 10.0, 15.0, 20.0, 30.0, 60.0, float('inf')
)
2020-01-03 04:18:49 +01:00
2020-02-21 04:11:25 +01:00
@dataclass
class ReaderProcessState:
cursor: sqlite3.Cursor
reader_context: Optional[ContextVar[ReaderProcessState]] = ContextVar('reader_context')
def initializer(path):
2020-02-21 04:33:24 +01:00
db = sqlite3.connect(path)
db.row_factory = dict_row_factory
2020-02-21 04:33:24 +01:00
db.executescript("pragma journal_mode=WAL;")
reader = ReaderProcessState(db.cursor())
2020-02-21 04:11:25 +01:00
reader_context.set(reader)
def run_read_only_fetchall(sql, params):
cursor = reader_context.get().cursor
try:
return cursor.execute(sql, params).fetchall()
except (Exception, OSError) as e:
log.exception('Error running transaction:', exc_info=e)
raise
def run_read_only_fetchone(sql, params):
cursor = reader_context.get().cursor
try:
return cursor.execute(sql, params).fetchone()
except (Exception, OSError) as e:
log.exception('Error running transaction:', exc_info=e)
raise
2020-01-03 04:18:49 +01:00
class AIOSQLite:
2020-03-19 18:24:25 +01:00
reader_executor: ReaderExecutorClass
2020-01-03 04:18:49 +01:00
2020-04-24 03:17:44 +02:00
waiting_writes_metric = Gauge(
"waiting_writes_count", "Number of waiting db writes", namespace="daemon_database"
)
waiting_reads_metric = Gauge(
"waiting_reads_count", "Number of waiting db writes", namespace="daemon_database"
)
2020-05-03 03:23:17 +02:00
write_count_metric = Counter(
"write_count", "Number of database writes", namespace="daemon_database"
)
read_count_metric = Counter(
"read_count", "Number of database reads", namespace="daemon_database"
)
acquire_write_lock_metric = Histogram(
2020-05-03 04:30:25 +02:00
f'write_lock_acquired', 'Time to acquire the write lock', namespace="daemon_database", buckets=HISTOGRAM_BUCKETS
2020-05-03 03:23:17 +02:00
)
held_write_lock_metric = Histogram(
2020-05-04 18:09:09 +02:00
f'write_lock_held', 'Length of time the write lock is held for', namespace="daemon_database",
buckets=HISTOGRAM_BUCKETS
2020-05-03 03:23:17 +02:00
)
2020-04-24 03:17:44 +02:00
2020-01-03 04:18:49 +01:00
def __init__(self):
# has to be single threaded as there is no mapping of thread:connection
2020-02-18 00:12:52 +01:00
self.writer_executor = ThreadPoolExecutor(max_workers=1)
self.writer_connection: Optional[sqlite3.Connection] = None
2020-01-03 04:18:49 +01:00
self._closing = False
self.query_count = 0
2020-05-03 03:23:17 +02:00
self.write_lock = LockWithMetrics(self.acquire_write_lock_metric, self.held_write_lock_metric)
2020-02-25 20:15:27 +01:00
self.writers = 0
self.read_ready = asyncio.Event()
self.urgent_read_done = asyncio.Event()
2020-01-03 04:18:49 +01:00
@classmethod
async def connect(cls, path: Union[bytes, str], *args, **kwargs):
sqlite3.enable_callback_tracebacks(True)
db = cls()
2020-02-18 00:12:52 +01:00
def _connect_writer():
db.writer_connection = sqlite3.connect(path, *args, **kwargs)
2020-02-21 04:11:25 +01:00
readers = max(os.cpu_count() - 2, 2)
2020-03-19 18:24:25 +01:00
db.reader_executor = ReaderExecutorClass(
2020-02-21 04:11:25 +01:00
max_workers=readers, initializer=initializer, initargs=(path, )
)
2020-02-18 00:12:52 +01:00
await asyncio.get_event_loop().run_in_executor(db.writer_executor, _connect_writer)
2020-02-25 20:15:27 +01:00
db.read_ready.set()
db.urgent_read_done.set()
2020-01-03 04:18:49 +01:00
return db
async def close(self):
if self._closing:
return
self._closing = True
2020-11-16 14:53:41 +01:00
def __checkpoint_and_close(conn: sqlite3.Connection):
conn.execute("PRAGMA WAL_CHECKPOINT(FULL);")
log.info("DB checkpoint finished.")
2020-11-16 14:53:41 +01:00
conn.close()
await asyncio.get_event_loop().run_in_executor(
self.writer_executor, __checkpoint_and_close, self.writer_connection)
2020-02-18 00:12:52 +01:00
self.writer_executor.shutdown(wait=True)
2020-02-21 04:11:25 +01:00
self.reader_executor.shutdown(wait=True)
2020-02-25 20:15:27 +01:00
self.read_ready.clear()
2020-02-18 00:12:52 +01:00
self.writer_connection = None
2020-01-03 04:18:49 +01:00
def executemany(self, sql: str, params: Iterable):
params = params if params is not None else []
# this fetchall is needed to prevent SQLITE_MISUSE
return self.run(lambda conn: conn.executemany(sql, params).fetchall())
def executescript(self, script: str) -> Awaitable:
return self.run(lambda conn: conn.executescript(script))
2020-02-25 20:15:27 +01:00
async def _execute_fetch(self, sql: str, parameters: Iterable = None,
read_only=False, fetch_all: bool = False) -> List[dict]:
2020-02-25 20:15:27 +01:00
read_only_fn = run_read_only_fetchall if fetch_all else run_read_only_fetchone
2020-01-03 04:18:49 +01:00
parameters = parameters if parameters is not None else []
still_waiting = False
urgent_read = False
2020-02-21 04:11:25 +01:00
if read_only:
2020-04-24 03:17:44 +02:00
self.waiting_reads_metric.inc()
2020-05-03 03:23:17 +02:00
self.read_count_metric.inc()
try:
2020-10-20 22:07:18 +02:00
while self.writers and not self._closing: # more writes can come in while we are waiting for the first
if not urgent_read and still_waiting and self.urgent_read_done.is_set():
# throttle the writes if they pile up
self.urgent_read_done.clear()
urgent_read = True
# wait until the running writes have finished
await self.read_ready.wait()
still_waiting = True
2020-10-20 22:07:18 +02:00
if self._closing:
2020-12-04 00:25:38 +01:00
raise asyncio.CancelledError()
return await asyncio.get_event_loop().run_in_executor(
self.reader_executor, read_only_fn, sql, parameters
)
finally:
if urgent_read:
# unthrottle the writers if they had to be throttled
self.urgent_read_done.set()
2020-04-24 03:17:44 +02:00
self.waiting_reads_metric.dec()
2020-02-25 20:15:27 +01:00
if fetch_all:
return await self.run(lambda conn: conn.execute(sql, parameters).fetchall())
return await self.run(lambda conn: conn.execute(sql, parameters).fetchone())
2020-01-03 04:18:49 +01:00
2020-02-25 20:15:27 +01:00
async def execute_fetchall(self, sql: str, parameters: Iterable = None,
read_only=False) -> List[dict]:
2020-02-25 20:15:27 +01:00
return await self._execute_fetch(sql, parameters, read_only, fetch_all=True)
async def execute_fetchone(self, sql: str, parameters: Iterable = None,
read_only=False) -> List[dict]:
2020-02-25 20:15:27 +01:00
return await self._execute_fetch(sql, parameters, read_only, fetch_all=False)
2020-01-03 04:18:49 +01:00
def execute(self, sql: str, parameters: Iterable = None) -> Awaitable[sqlite3.Cursor]:
parameters = parameters if parameters is not None else []
return self.run(lambda conn: conn.execute(sql, parameters))
2020-02-25 20:15:27 +01:00
async def run(self, fun, *args, **kwargs):
2020-05-03 03:23:17 +02:00
self.write_count_metric.inc()
2020-04-24 03:17:44 +02:00
self.waiting_writes_metric.inc()
2020-05-04 19:47:37 +02:00
# it's possible many writes are coming in one after the other, these can
# block reader calls for a long time
# if the reader waits for the writers to finish and then has to wait for
# yet more, it will clear the urgent_read_done event to block more writers
# piling on
2020-04-24 03:17:44 +02:00
try:
await self.urgent_read_done.wait()
except Exception as e:
self.waiting_writes_metric.dec()
raise e
2020-02-25 20:15:27 +01:00
self.writers += 1
2020-05-04 19:47:37 +02:00
# block readers
2020-02-25 20:15:27 +01:00
self.read_ready.clear()
2020-05-01 17:40:52 +02:00
try:
async with self.write_lock:
2020-10-20 22:07:18 +02:00
if self._closing:
2020-12-04 00:25:38 +01:00
raise asyncio.CancelledError()
2020-02-25 20:15:27 +01:00
return await asyncio.get_event_loop().run_in_executor(
self.writer_executor, lambda: self.__run_transaction(fun, *args, **kwargs)
)
2020-05-01 17:40:52 +02:00
finally:
self.writers -= 1
2020-04-24 03:17:44 +02:00
self.waiting_writes_metric.dec()
2020-05-01 17:40:52 +02:00
if not self.writers:
2020-05-04 19:47:37 +02:00
# unblock the readers once the last enqueued writer finishes
2020-05-01 17:40:52 +02:00
self.read_ready.set()
2020-01-03 04:18:49 +01:00
def __run_transaction(self, fun: Callable[[sqlite3.Connection, Any, Any], Any], *args, **kwargs):
2020-02-18 00:12:52 +01:00
self.writer_connection.execute('begin')
2020-01-03 04:18:49 +01:00
try:
self.query_count += 1
2020-02-18 00:12:52 +01:00
result = fun(self.writer_connection, *args, **kwargs) # type: ignore
self.writer_connection.commit()
2020-01-03 04:18:49 +01:00
return result
except (Exception, OSError) as e:
log.exception('Error running transaction:', exc_info=e)
2020-02-18 00:12:52 +01:00
self.writer_connection.rollback()
2020-01-03 04:18:49 +01:00
log.warning("rolled back")
raise
2020-05-03 03:23:17 +02:00
async def run_with_foreign_keys_disabled(self, fun, *args, **kwargs):
self.write_count_metric.inc()
self.waiting_writes_metric.inc()
try:
await self.urgent_read_done.wait()
except Exception as e:
self.waiting_writes_metric.dec()
raise e
self.writers += 1
self.read_ready.clear()
try:
async with self.write_lock:
2020-10-20 22:07:18 +02:00
if self._closing:
2020-12-04 00:25:38 +01:00
raise asyncio.CancelledError()
2020-05-03 03:23:17 +02:00
return await asyncio.get_event_loop().run_in_executor(
self.writer_executor, self.__run_transaction_with_foreign_keys_disabled, fun, args, kwargs
)
finally:
self.writers -= 1
self.waiting_writes_metric.dec()
if not self.writers:
self.read_ready.set()
2020-01-03 04:18:49 +01:00
def __run_transaction_with_foreign_keys_disabled(self,
fun: Callable[[sqlite3.Connection, Any, Any], Any],
args, kwargs):
2020-02-18 00:12:52 +01:00
foreign_keys_enabled, = self.writer_connection.execute("pragma foreign_keys").fetchone()
2020-01-03 04:18:49 +01:00
if not foreign_keys_enabled:
raise sqlite3.IntegrityError("foreign keys are disabled, use `AIOSQLite.run` instead")
try:
2020-02-18 00:12:52 +01:00
self.writer_connection.execute('pragma foreign_keys=off').fetchone()
2020-01-03 04:18:49 +01:00
return self.__run_transaction(fun, *args, **kwargs)
finally:
2020-02-18 00:12:52 +01:00
self.writer_connection.execute('pragma foreign_keys=on').fetchone()
2020-01-03 04:18:49 +01:00
def constraints_to_sql(constraints, joiner=' AND ', prepend_key=''):
sql, values = [], {}
for key, constraint in constraints.items():
tag = '0'
if '#' in key:
key, tag = key[:key.index('#')], key[key.index('#')+1:]
col, op, key = key, '=', key.replace('.', '_')
if not key:
sql.append(constraint)
continue
2020-03-07 06:34:47 +01:00
if key.startswith('$$'):
col, key = col[2:], key[1:]
elif key.startswith('$'):
2020-01-03 04:18:49 +01:00
values[key] = constraint
continue
if key.endswith('__not'):
col, op = col[:-len('__not')], '!='
elif key.endswith('__is_null'):
col = col[:-len('__is_null')]
sql.append(f'{col} IS NULL')
continue
if key.endswith('__is_not_null'):
col = col[:-len('__is_not_null')]
sql.append(f'{col} IS NOT NULL')
continue
if key.endswith('__lt'):
col, op = col[:-len('__lt')], '<'
elif key.endswith('__lte'):
col, op = col[:-len('__lte')], '<='
elif key.endswith('__gt'):
col, op = col[:-len('__gt')], '>'
elif key.endswith('__gte'):
col, op = col[:-len('__gte')], '>='
elif key.endswith('__like'):
col, op = col[:-len('__like')], 'LIKE'
elif key.endswith('__not_like'):
col, op = col[:-len('__not_like')], 'NOT LIKE'
elif key.endswith('__in') or key.endswith('__not_in'):
if key.endswith('__in'):
col, op, one_val_op = col[:-len('__in')], 'IN', '='
2020-01-03 04:18:49 +01:00
else:
col, op, one_val_op = col[:-len('__not_in')], 'NOT IN', '!='
2020-01-03 04:18:49 +01:00
if constraint:
if isinstance(constraint, (list, set, tuple)):
if len(constraint) == 1:
2020-02-12 17:41:32 +01:00
values[f'{key}{tag}'] = next(iter(constraint))
sql.append(f'{col} {one_val_op} :{key}{tag}')
else:
keys = []
for i, val in enumerate(constraint):
keys.append(f':{key}{tag}_{i}')
values[f'{key}{tag}_{i}'] = val
sql.append(f'{col} {op} ({", ".join(keys)})')
2020-01-03 04:18:49 +01:00
elif isinstance(constraint, str):
sql.append(f'{col} {op} ({constraint})')
else:
raise ValueError(f"{col} requires a list, set or string as constraint value.")
continue
elif key.endswith('__any') or key.endswith('__or'):
where, subvalues = constraints_to_sql(constraint, ' OR ', key+tag+'_')
sql.append(f'({where})')
values.update(subvalues)
continue
if key.endswith('__and'):
where, subvalues = constraints_to_sql(constraint, ' AND ', key+tag+'_')
sql.append(f'({where})')
values.update(subvalues)
continue
sql.append(f'{col} {op} :{prepend_key}{key}{tag}')
values[prepend_key+key+tag] = constraint
return joiner.join(sql) if sql else '', values
def query(select, **constraints) -> Tuple[str, Dict[str, Any]]:
sql = [select]
limit = constraints.pop('limit', None)
offset = constraints.pop('offset', None)
order_by = constraints.pop('order_by', None)
group_by = constraints.pop('group_by', None)
2020-01-03 04:18:49 +01:00
accounts = constraints.pop('accounts', [])
if accounts:
constraints['account__in'] = [a.public_key.address for a in accounts]
where, values = constraints_to_sql(constraints)
if where:
sql.append('WHERE')
sql.append(where)
if group_by is not None:
sql.append(f'GROUP BY {group_by}')
2020-01-03 04:18:49 +01:00
if order_by:
sql.append('ORDER BY')
if isinstance(order_by, str):
sql.append(order_by)
elif isinstance(order_by, list):
sql.append(', '.join(order_by))
else:
raise ValueError("order_by must be string or list")
if limit is not None:
sql.append(f'LIMIT {limit}')
if offset is not None:
sql.append(f'OFFSET {offset}')
return ' '.join(sql), values
def interpolate(sql, values):
for k in sorted(values.keys(), reverse=True):
value = values[k]
if isinstance(value, bytes):
value = f"X'{hexlify(value).decode()}'"
elif isinstance(value, str):
value = f"'{value}'"
else:
value = str(value)
sql = sql.replace(f":{k}", value)
return sql
def constrain_single_or_list(constraints, column, value, convert=lambda x: x, negate=False):
2020-03-07 06:34:47 +01:00
if value is not None:
if isinstance(value, list):
value = [convert(v) for v in value]
if len(value) == 1:
if negate:
constraints[f"{column}__or"] = {
f"{column}__is_null": True,
f"{column}__not": value[0]
}
else:
constraints[column] = value[0]
2020-03-07 06:34:47 +01:00
elif len(value) > 1:
if negate:
constraints[f"{column}__or"] = {
f"{column}__is_null": True,
f"{column}__not_in": value
}
else:
constraints[f"{column}__in"] = value
elif negate:
constraints[f"{column}__or"] = {
f"{column}__is_null": True,
f"{column}__not": convert(value)
}
2020-03-07 06:34:47 +01:00
else:
constraints[column] = convert(value)
return constraints
2020-01-03 04:18:49 +01:00
class SQLiteMixin:
SCHEMA_VERSION: Optional[str] = None
CREATE_TABLES_QUERY: str
MAX_QUERY_VARIABLES = 900
CREATE_VERSION_TABLE = """
create table if not exists version (
version text
);
"""
def __init__(self, path):
self._db_path = path
self.db: AIOSQLite = None
self.ledger = None
async def open(self):
log.info("connecting to database: %s", self._db_path)
self.db = await AIOSQLite.connect(self._db_path, isolation_level=None)
if self.SCHEMA_VERSION:
tables = [t[0] for t in await self.db.execute_fetchall(
"SELECT name FROM sqlite_master WHERE type='table';"
)]
if tables:
if 'version' in tables:
version = await self.db.execute_fetchone("SELECT version FROM version LIMIT 1;")
if version == (self.SCHEMA_VERSION,):
return
await self.db.executescript('\n'.join(
f"DROP TABLE {table};" for table in tables
2020-12-23 02:08:26 +01:00
) + '\n' + 'PRAGMA WAL_CHECKPOINT(FULL);' + '\n' + 'VACUUM;')
2020-01-03 04:18:49 +01:00
await self.db.execute(self.CREATE_VERSION_TABLE)
await self.db.execute("INSERT INTO version VALUES (?)", (self.SCHEMA_VERSION,))
await self.db.executescript(self.CREATE_TABLES_QUERY)
async def close(self):
await self.db.close()
@staticmethod
def _insert_sql(table: str, data: dict, ignore_duplicate: bool = False,
replace: bool = False) -> Tuple[str, List]:
columns, values = [], []
for column, value in data.items():
columns.append(column)
values.append(value)
policy = ""
if ignore_duplicate:
policy = " OR IGNORE"
if replace:
policy = " OR REPLACE"
sql = "INSERT{} INTO {} ({}) VALUES ({})".format(
policy, table, ', '.join(columns), ', '.join(['?'] * len(values))
)
return sql, values
@staticmethod
def _update_sql(table: str, data: dict, where: str,
constraints: Union[list, tuple]) -> Tuple[str, list]:
columns, values = [], []
for column, value in data.items():
columns.append(f"{column} = ?")
values.append(value)
values.extend(constraints)
sql = "UPDATE {} SET {} WHERE {}".format(
table, ', '.join(columns), where
)
return sql, values
def dict_row_factory(cursor, row):
d = {}
for idx, col in enumerate(cursor.description):
d[col[0]] = row[idx]
return d
2020-06-04 16:18:04 +02:00
SQLITE_MAX_INTEGER = 9223372036854775807
def _get_spendable_utxos(transaction: sqlite3.Connection, accounts: List, decoded_transactions: Dict[str, Transaction],
result: Dict[Tuple[bytes, int, bool], List[int]], reserved: List[Transaction],
amount_to_reserve: int, reserved_amount: int, floor: int, ceiling: int,
fee_per_byte: int) -> int:
accounts_fmt = ",".join(["?"] * len(accounts))
txo_query = f"""
SELECT tx.txid, txo.txoid, tx.raw, tx.height, txo.position as nout, tx.is_verified, txo.amount FROM txo
INNER JOIN account_address USING (address)
LEFT JOIN txi USING (txoid)
INNER JOIN tx USING (txid)
WHERE txo.txo_type=0 AND txi.txoid IS NULL AND tx.txid IS NOT NULL AND NOT txo.is_reserved
AND txo.amount >= ? AND txo.amount < ?
"""
if accounts:
txo_query += f"""
AND account_address.account {'= ?' if len(accounts_fmt) == 1 else 'IN (' + accounts_fmt + ')'}
"""
txo_query += """
ORDER BY txo.amount ASC, tx.height DESC
"""
2020-06-04 16:18:04 +02:00
# prefer confirmed, but save unconfirmed utxos from this selection in case they are needed
unconfirmed = []
for row in transaction.execute(txo_query, (floor, ceiling, *accounts)):
(txid, txoid, raw, height, nout, verified, amount) = row.values()
# verified or non verified transactions were found- reset the gap count
# multiple txos can come from the same tx, only decode it once and cache
if txid not in decoded_transactions:
# cache the decoded transaction
decoded_transactions[txid] = Transaction(raw)
decoded_tx = decoded_transactions[txid]
# save the unconfirmed txo for possible use later, if still needed
if verified:
# add the txo to the reservation, minus the fee for including it
reserved_amount += amount
reserved_amount -= Input.spend(decoded_tx.outputs[nout]).size * fee_per_byte
# mark it as reserved
result[(raw, height, verified)].append(nout)
reserved.append(txoid)
# if we've reserved enough, return
if reserved_amount >= amount_to_reserve:
return reserved_amount
else:
unconfirmed.append((txid, txoid, raw, height, nout, verified, amount))
# we're popping the items, so to get them in the order they were seen they are reversed
unconfirmed.reverse()
# add available unconfirmed txos if any were previously found
while unconfirmed and reserved_amount < amount_to_reserve:
(txid, txoid, raw, height, nout, verified, amount) = unconfirmed.pop()
# it's already decoded
decoded_tx = decoded_transactions[txid]
# add to the reserved amount
reserved_amount += amount
reserved_amount -= Input.spend(decoded_tx.outputs[nout]).size * fee_per_byte
result[(raw, height, verified)].append(nout)
reserved.append(txoid)
return reserved_amount
def get_and_reserve_spendable_utxos(transaction: sqlite3.Connection, accounts: List, amount_to_reserve: int, floor: int,
2020-06-15 16:20:55 +02:00
fee_per_byte: int, set_reserved: bool, return_insufficient_funds: bool,
base_multiplier: int = 100):
2020-06-04 16:18:04 +02:00
txs = defaultdict(list)
decoded_transactions = {}
reserved = []
reserved_dewies = 0
2020-06-15 16:20:55 +02:00
multiplier = base_multiplier
2020-06-04 16:18:04 +02:00
gap_count = 0
while reserved_dewies < amount_to_reserve and gap_count < 5 and floor * multiplier < SQLITE_MAX_INTEGER:
previous_reserved_dewies = reserved_dewies
reserved_dewies = _get_spendable_utxos(
transaction, accounts, decoded_transactions, txs, reserved, amount_to_reserve, reserved_dewies,
floor, floor * multiplier, fee_per_byte
)
floor *= multiplier
if previous_reserved_dewies == reserved_dewies:
gap_count += 1
multiplier **= 2
else:
gap_count = 0
2020-06-15 16:20:55 +02:00
multiplier = base_multiplier
2020-06-04 16:18:04 +02:00
# reserve the accumulated txos if enough were found
if reserved_dewies >= amount_to_reserve:
if set_reserved:
transaction.executemany("UPDATE txo SET is_reserved = ? WHERE txoid = ?",
[(True, txoid) for txoid in reserved]).fetchall()
return txs
# return_insufficient_funds and set_reserved are used for testing
return txs if return_insufficient_funds else {}
2020-06-04 16:18:04 +02:00
2020-01-03 04:18:49 +01:00
class Database(SQLiteMixin):
2020-12-23 02:08:26 +01:00
SCHEMA_VERSION = "1.5"
2020-01-03 04:18:49 +01:00
PRAGMAS = """
pragma journal_mode=WAL;
"""
CREATE_ACCOUNT_TABLE = """
create table if not exists account_address (
account text not null,
address text not null,
chain integer not null,
pubkey blob not null,
chain_code blob not null,
n integer not null,
depth integer not null,
primary key (account, address)
);
create index if not exists address_account_idx on account_address (address, account);
"""
CREATE_PUBKEY_ADDRESS_TABLE = """
create table if not exists pubkey_address (
address text primary key,
history text,
used_times integer not null default 0
);
"""
CREATE_TX_TABLE = """
create table if not exists tx (
txid text primary key,
raw blob not null,
height integer not null,
position integer not null,
is_verified boolean not null default 0,
purchased_claim_id text,
day integer
);
create index if not exists tx_purchased_claim_id_idx on tx (purchased_claim_id);
"""
CREATE_TXO_TABLE = """
create table if not exists txo (
2018-07-15 05:02:19 +02:00
txid text references tx,
txoid text primary key,
address text references pubkey_address,
position integer not null,
amount integer not null,
script blob not null,
is_reserved boolean not null default 0,
2018-07-12 05:18:59 +02:00
txo_type integer not null default 0,
2018-07-15 05:02:19 +02:00
claim_id text,
2020-03-21 01:22:57 +01:00
claim_name text,
channel_id text,
2020-03-21 01:22:57 +01:00
reposted_claim_id text
);
create index if not exists txo_txid_idx on txo (txid);
create index if not exists txo_address_idx on txo (address);
create index if not exists txo_claim_id_idx on txo (claim_id, txo_type);
2020-03-07 06:34:47 +01:00
create index if not exists txo_claim_name_idx on txo (claim_name);
create index if not exists txo_txo_type_idx on txo (txo_type);
create index if not exists txo_channel_id_idx on txo (channel_id);
2020-03-21 01:22:57 +01:00
create index if not exists txo_reposted_claim_idx on txo (reposted_claim_id);
"""
2020-01-03 04:18:49 +01:00
CREATE_TXI_TABLE = """
create table if not exists txi (
txid text references tx,
txoid text references txo primary key,
address text references pubkey_address,
position integer not null
2020-01-03 04:18:49 +01:00
);
create index if not exists txi_address_idx on txi (address);
create index if not exists first_input_idx on txi (txid, address) where position=0;
2020-01-03 04:18:49 +01:00
"""
2018-06-12 17:53:29 +02:00
CREATE_TABLES_QUERY = (
2020-01-03 04:50:27 +01:00
PRAGMAS +
CREATE_ACCOUNT_TABLE +
CREATE_PUBKEY_ADDRESS_TABLE +
CREATE_TX_TABLE +
CREATE_TXO_TABLE +
CREATE_TXI_TABLE
2018-06-12 17:53:29 +02:00
)
async def open(self):
await super().open()
self.db.writer_connection.row_factory = dict_row_factory
def txo_to_row(self, tx, txo):
2020-01-03 04:18:49 +01:00
row = {
'txid': tx.id,
'txoid': txo.id,
'address': txo.get_address(self.ledger),
2020-01-03 04:18:49 +01:00
'position': txo.position,
'amount': txo.amount,
'script': sqlite3.Binary(txo.script.source)
}
if txo.is_claim:
2019-09-13 15:32:34 +02:00
if txo.can_decode_claim:
2020-03-21 01:22:57 +01:00
claim = txo.claim
row['txo_type'] = TXO_TYPES.get(claim.claim_type, TXO_TYPES['stream'])
if claim.is_repost:
row['reposted_claim_id'] = claim.repost.reference.claim_id
if claim.is_signed:
row['channel_id'] = claim.signing_channel_id
2019-09-13 15:32:34 +02:00
else:
row['txo_type'] = TXO_TYPES['stream']
elif txo.is_support:
row['txo_type'] = TXO_TYPES['support']
2021-01-25 16:14:34 +01:00
support = txo.can_decode_support
if support and support.is_signed:
row['channel_id'] = support.signing_channel_id
elif txo.purchase is not None:
row['txo_type'] = TXO_TYPES['purchase']
row['claim_id'] = txo.purchased_claim_id
if txo.script.is_claim_involved:
row['claim_id'] = txo.claim_id
row['claim_name'] = txo.claim_name
return row
2018-07-12 05:18:59 +02:00
def tx_to_row(self, tx):
2020-01-03 04:18:49 +01:00
row = {
'txid': tx.id,
'raw': sqlite3.Binary(tx.raw),
'height': tx.height,
'position': tx.position,
'is_verified': tx.is_verified,
'day': tx.get_julian_day(self.ledger),
2020-01-03 04:18:49 +01:00
}
txos = tx.outputs
if len(txos) >= 2 and txos[1].can_decode_purchase_data:
txos[0].purchase = txos[1]
row['purchased_claim_id'] = txos[1].purchase_data.claim_id
return row
async def insert_transaction(self, tx):
await self.db.execute_fetchall(*self._insert_sql('tx', self.tx_to_row(tx)))
async def update_transaction(self, tx):
await self.db.execute_fetchall(*self._update_sql("tx", {
'height': tx.height, 'position': tx.position, 'is_verified': tx.is_verified
}, 'txid = ?', (tx.id,)))
def _transaction_io(self, conn: sqlite3.Connection, tx: Transaction, address, txhash):
2020-01-12 22:11:57 +01:00
conn.execute(*self._insert_sql('tx', self.tx_to_row(tx), replace=True)).fetchall()
2020-01-03 04:18:49 +01:00
is_my_input = False
2020-01-03 04:18:49 +01:00
for txi in tx.inputs:
if txi.txo_ref.txo is not None:
txo = txi.txo_ref.txo
if txo.has_address and txo.get_address(self.ledger) == address:
is_my_input = True
2020-01-03 04:18:49 +01:00
conn.execute(*self._insert_sql("txi", {
'txid': tx.id,
'txoid': txo.id,
'address': address,
'position': txi.position
2020-01-03 04:18:49 +01:00
}, ignore_duplicate=True)).fetchall()
for txo in tx.outputs:
if txo.script.is_pay_pubkey_hash and (txo.pubkey_hash == txhash or is_my_input):
conn.execute(*self._insert_sql(
"txo", self.txo_to_row(tx, txo), ignore_duplicate=True
)).fetchall()
elif txo.script.is_pay_script_hash:
# TODO: implement script hash payments
log.warning('Database.save_transaction_io: pay script hash is not implemented!')
2020-01-03 04:18:49 +01:00
def save_transaction_io(self, tx: Transaction, address, txhash, history):
return self.save_transaction_io_batch([tx], address, txhash, history)
2020-01-03 04:18:49 +01:00
def save_transaction_io_batch(self, txs: Iterable[Transaction], address, txhash, history):
history_count = history.count(':') // 2
2020-01-03 04:18:49 +01:00
def __many(conn):
for tx in txs:
self._transaction_io(conn, tx, address, txhash)
conn.execute(
"UPDATE pubkey_address SET history = ?, used_times = ? WHERE address = ?",
(history, history_count, address)
).fetchall()
2020-01-03 04:18:49 +01:00
return self.db.run(__many)
async def reserve_outputs(self, txos, is_reserved=True):
2020-05-03 04:30:25 +02:00
txoids = [(is_reserved, txo.id) for txo in txos]
2020-01-03 04:18:49 +01:00
await self.db.executemany("UPDATE txo SET is_reserved = ? WHERE txoid = ?", txoids)
async def release_outputs(self, txos):
await self.reserve_outputs(txos, is_reserved=False)
async def rewind_blockchain(self, above_height): # pylint: disable=no-self-use
# TODO:
# 1. delete transactions above_height
# 2. update address histories removing deleted TXs
return True
2020-06-15 16:20:55 +02:00
async def get_spendable_utxos(self, ledger, reserve_amount, accounts: Optional[Iterable], min_amount: int = 1,
fee_per_byte: int = 50, set_reserved: bool = True,
return_insufficient_funds: bool = False) -> List:
2020-06-04 16:18:04 +02:00
to_spend = await self.db.run(
get_and_reserve_spendable_utxos, tuple(account.id for account in accounts), reserve_amount, min_amount,
fee_per_byte, set_reserved, return_insufficient_funds
2020-06-04 16:18:04 +02:00
)
txos = []
for (raw, height, verified), positions in to_spend.items():
tx = Transaction(raw, height=height, is_verified=verified)
for nout in positions:
txos.append(tx.outputs[nout].get_estimator(ledger))
return txos
2020-03-17 15:29:26 +01:00
async def select_transactions(self, cols, accounts=None, read_only=False, **constraints):
2020-01-03 04:18:49 +01:00
if not {'txid', 'txid__in'}.intersection(constraints):
assert accounts, "'accounts' argument required when no 'txid' constraint is present"
2020-03-07 06:34:47 +01:00
where, values = constraints_to_sql({
'$$account_address.account__in': [a.public_key.address for a in accounts]
2020-01-03 04:18:49 +01:00
})
constraints['txid__in'] = f"""
2020-03-07 06:34:47 +01:00
SELECT txo.txid FROM txo JOIN account_address USING (address) WHERE {where}
2020-01-03 04:18:49 +01:00
UNION
2020-03-07 06:34:47 +01:00
SELECT txi.txid FROM txi JOIN account_address USING (address) WHERE {where}
2020-01-03 04:18:49 +01:00
"""
2020-03-07 06:34:47 +01:00
constraints.update(values)
2020-01-03 04:18:49 +01:00
return await self.db.execute_fetchall(
2020-02-21 04:11:25 +01:00
*query(f"SELECT {cols} FROM tx", **constraints), read_only=read_only
2020-01-03 04:18:49 +01:00
)
TXO_NOT_MINE = Output(None, None, is_my_output=False)
2020-01-03 04:18:49 +01:00
async def get_transactions(self, wallet=None, **constraints):
include_is_spent = constraints.pop('include_is_spent', False)
include_is_my_input = constraints.pop('include_is_my_input', False)
include_is_my_output = constraints.pop('include_is_my_output', False)
2020-01-03 04:18:49 +01:00
tx_rows = await self.select_transactions(
'txid, raw, height, position, is_verified',
order_by=constraints.pop('order_by', ["height=0 DESC", "height DESC", "position DESC"]),
**constraints
)
if not tx_rows:
return []
txids, txs, txi_txoids = [], [], []
for row in tx_rows:
txids.append(row['txid'])
2020-01-03 04:18:49 +01:00
txs.append(Transaction(
raw=row['raw'], height=row['height'], position=row['position'],
is_verified=bool(row['is_verified'])
2020-01-03 04:18:49 +01:00
))
for txi in txs[-1].inputs:
txi_txoids.append(txi.txo_ref.id)
step = self.MAX_QUERY_VARIABLES
annotated_txos = {}
for offset in range(0, len(txids), step):
annotated_txos.update({
txo.id: txo for txo in
(await self.get_txos(
wallet=wallet,
txid__in=txids[offset:offset+step], order_by='txo.txid',
include_is_spent=include_is_spent,
include_is_my_input=include_is_my_input,
include_is_my_output=include_is_my_output,
2020-01-03 04:18:49 +01:00
))
})
referenced_txos = {}
for offset in range(0, len(txi_txoids), step):
referenced_txos.update({
txo.id: txo for txo in
(await self.get_txos(
wallet=wallet,
txoid__in=txi_txoids[offset:offset+step], order_by='txo.txoid',
include_is_my_output=include_is_my_output,
2020-01-03 04:18:49 +01:00
))
})
for tx in txs:
for txi in tx.inputs:
txo = referenced_txos.get(txi.txo_ref.id)
if txo:
txi.txo_ref = txo.ref
for txo in tx.outputs:
_txo = annotated_txos.get(txo.id)
if _txo:
txo.update_annotations(_txo)
else:
txo.update_annotations(self.TXO_NOT_MINE)
2020-01-03 04:18:49 +01:00
for tx in txs:
txos = tx.outputs
if len(txos) >= 2 and txos[1].can_decode_purchase_data:
txos[0].purchase = txos[1]
2020-01-03 04:18:49 +01:00
return txs
2020-01-03 04:18:49 +01:00
async def get_transaction_count(self, **constraints):
constraints.pop('wallet', None)
constraints.pop('offset', None)
constraints.pop('limit', None)
constraints.pop('order_by', None)
count = await self.select_transactions('COUNT(*) as total', **constraints)
return count[0]['total'] or 0
2020-01-03 04:18:49 +01:00
async def get_transaction(self, **constraints):
txs = await self.get_transactions(limit=1, **constraints)
if txs:
return txs[0]
async def select_txos(
self, cols, accounts=None, is_my_input=None, is_my_output=True,
is_my_input_or_output=None, exclude_internal_transfers=False,
include_is_spent=False, include_is_my_input=False,
2020-04-02 02:44:34 +02:00
is_spent=None, read_only=False, **constraints):
for rename_col in ('txid', 'txoid'):
for rename_constraint in (rename_col, rename_col+'__in', rename_col+'__not_in'):
if rename_constraint in constraints:
constraints['txo.'+rename_constraint] = constraints.pop(rename_constraint)
if accounts:
account_in_sql, values = constraints_to_sql({
'$$account__in': [a.public_key.address for a in accounts]
2020-03-09 04:11:03 +01:00
})
my_addresses = f"SELECT address FROM account_address WHERE {account_in_sql}"
2020-03-09 04:11:03 +01:00
constraints.update(values)
if is_my_input_or_output:
include_is_my_input = True
constraints['received_or_sent__or'] = {
'txo.address__in': my_addresses,
'sent__and': {
'txi.address__is_not_null': True,
'txi.address__in': my_addresses
}
}
else:
if is_my_output:
constraints['txo.address__in'] = my_addresses
elif is_my_output is False:
constraints['txo.address__not_in'] = my_addresses
if is_my_input:
include_is_my_input = True
constraints['txi.address__is_not_null'] = True
constraints['txi.address__in'] = my_addresses
elif is_my_input is False:
include_is_my_input = True
constraints['is_my_input_false__or'] = {
'txi.address__is_null': True,
'txi.address__not_in': my_addresses
}
if exclude_internal_transfers:
include_is_my_input = True
constraints['exclude_internal_payments__or'] = {
'txo.txo_type__not': TXO_TYPES['other'],
'txo.address__not_in': my_addresses,
'txi.address__is_null': True,
'txi.address__not_in': my_addresses,
}
sql = [f"SELECT {cols} FROM txo JOIN tx ON (tx.txid=txo.txid)"]
if is_spent:
constraints['spent.txoid__is_not_null'] = True
2020-04-02 02:44:34 +02:00
elif is_spent is False:
constraints['is_reserved'] = False
constraints['spent.txoid__is_null'] = True
if include_is_spent or is_spent is not None:
sql.append("LEFT JOIN txi AS spent ON (spent.txoid=txo.txoid)")
if include_is_my_input:
sql.append("LEFT JOIN txi ON (txi.position=0 AND txi.txid=txo.txid)")
return await self.db.execute_fetchall(*query(' '.join(sql), **constraints), read_only=read_only)
2020-04-02 02:44:34 +02:00
async def get_txos(self, wallet=None, no_tx=False, read_only=False, **constraints):
include_is_spent = constraints.get('include_is_spent', False)
include_is_my_input = constraints.get('include_is_my_input', False)
include_is_my_output = constraints.pop('include_is_my_output', False)
include_received_tips = constraints.pop('include_received_tips', False)
select_columns = [
"tx.txid, raw, tx.height, tx.position as tx_position, tx.is_verified, "
"txo_type, txo.position as txo_position, amount, script"
]
2020-01-03 04:18:49 +01:00
my_accounts = {a.public_key.address for a in wallet.accounts} if wallet else set()
my_accounts_sql = ""
if include_is_my_output or include_is_my_input:
my_accounts_sql, values = constraints_to_sql({'$$account__in#_wallet': my_accounts})
constraints.update(values)
if include_is_my_output and my_accounts:
if constraints.get('is_my_output', None) in (True, False):
select_columns.append(f"{1 if constraints['is_my_output'] else 0} AS is_my_output")
else:
select_columns.append(f"""(
txo.address IN (SELECT address FROM account_address WHERE {my_accounts_sql})
) AS is_my_output""")
if include_is_my_input and my_accounts:
if constraints.get('is_my_input', None) in (True, False):
select_columns.append(f"{1 if constraints['is_my_input'] else 0} AS is_my_input")
else:
select_columns.append(f"""(
txi.address IS NOT NULL AND
txi.address IN (SELECT address FROM account_address WHERE {my_accounts_sql})
) AS is_my_input""")
if include_is_spent:
select_columns.append("spent.txoid IS NOT NULL AS is_spent")
if include_received_tips:
select_columns.append(f"""(
SELECT COALESCE(SUM(support.amount), 0) FROM txo AS support WHERE
support.claim_id = txo.claim_id AND
support.txo_type = {TXO_TYPES['support']} AND
support.address IN (SELECT address FROM account_address WHERE {my_accounts_sql}) AND
support.txoid NOT IN (SELECT txoid FROM txi)
) AS received_tips""")
if 'order_by' not in constraints or constraints['order_by'] == 'height':
2020-01-03 04:18:49 +01:00
constraints['order_by'] = [
"tx.height=0 DESC", "tx.height DESC", "tx.position DESC", "txo.position"
]
2020-03-21 23:16:25 +01:00
elif constraints.get('order_by', None) == 'none':
del constraints['order_by']
rows = await self.select_txos(', '.join(select_columns), read_only=read_only, **constraints)
2020-01-03 04:18:49 +01:00
txos = []
txs = {}
for row in rows:
if no_tx:
txo = Output(
amount=row['amount'],
script=OutputScript(row['script']),
tx_ref=TXRefImmutable.from_id(row['txid'], row['height']),
position=row['txo_position']
2020-01-03 04:18:49 +01:00
)
else:
if row['txid'] not in txs:
txs[row['txid']] = Transaction(
row['raw'], height=row['height'], position=row['tx_position'],
is_verified=bool(row['is_verified'])
2020-01-03 04:18:49 +01:00
)
txo = txs[row['txid']].outputs[row['txo_position']]
if include_is_spent:
txo.is_spent = bool(row['is_spent'])
if include_is_my_input:
txo.is_my_input = bool(row['is_my_input'])
if include_is_my_output:
txo.is_my_output = bool(row['is_my_output'])
if include_is_my_input and include_is_my_output:
if txo.is_my_input and txo.is_my_output and row['txo_type'] == TXO_TYPES['other']:
txo.is_internal_transfer = True
else:
txo.is_internal_transfer = False
if include_received_tips:
txo.received_tips = row['received_tips']
2020-01-03 04:18:49 +01:00
txos.append(txo)
2018-10-05 15:02:02 +02:00
channel_ids = set()
2018-10-05 15:02:02 +02:00
for txo in txos:
if txo.is_claim and txo.can_decode_claim:
2019-03-20 06:46:23 +01:00
if txo.claim.is_signed:
channel_ids.add(txo.claim.signing_channel_id)
if txo.claim.is_channel and wallet:
for account in wallet.accounts:
private_key = await account.get_channel_private_key(
txo.claim.channel.public_key_bytes
)
if private_key:
txo.private_key = private_key
break
2018-10-05 15:02:02 +02:00
if channel_ids:
2018-10-05 15:02:02 +02:00
channels = {
txo.claim_id: txo for txo in
(await self.get_channels(
wallet=wallet,
2020-02-25 20:18:09 +01:00
claim_id__in=channel_ids,
read_only=read_only
2018-10-05 15:02:02 +02:00
))
}
for txo in txos:
if txo.is_claim and txo.can_decode_claim:
txo.channel = channels.get(txo.claim.signing_channel_id, None)
2018-10-05 15:02:02 +02:00
return txos
2020-04-02 02:53:09 +02:00
@staticmethod
def _clean_txo_constraints_for_aggregation(constraints):
constraints.pop('include_is_spent', None)
constraints.pop('include_is_my_input', None)
constraints.pop('include_is_my_output', None)
constraints.pop('include_received_tips', None)
constraints.pop('wallet', None)
constraints.pop('resolve', None)
2020-01-03 04:18:49 +01:00
constraints.pop('offset', None)
constraints.pop('limit', None)
constraints.pop('order_by', None)
2020-03-20 23:24:24 +01:00
2020-04-02 02:44:34 +02:00
async def get_txo_count(self, **constraints):
self._clean_txo_constraints_for_aggregation(constraints)
count = await self.select_txos('COUNT(*) AS total', **constraints)
return count[0]['total'] or 0
2020-01-03 04:18:49 +01:00
2020-04-02 02:44:34 +02:00
async def get_txo_sum(self, **constraints):
self._clean_txo_constraints_for_aggregation(constraints)
result = await self.select_txos('SUM(amount) AS total', **constraints)
return result[0]['total'] or 0
2020-03-20 23:24:24 +01:00
2020-04-02 02:44:34 +02:00
async def get_txo_plot(self, start_day=None, days_back=0, end_day=None, days_after=None, **constraints):
self._clean_txo_constraints_for_aggregation(constraints)
if start_day is None:
constraints['day__gte'] = self.ledger.headers.estimated_julian_day(
self.ledger.headers.height
) - days_back
else:
constraints['day__gte'] = date_to_julian_day(
date.fromisoformat(start_day)
)
if end_day is not None:
constraints['day__lte'] = date_to_julian_day(
date.fromisoformat(end_day)
)
elif days_after is not None:
constraints['day__lte'] = constraints['day__gte'] + days_after
return await self.select_txos(
"DATE(day) AS day, SUM(amount) AS total",
group_by='day', order_by='day', **constraints
)
2020-03-17 15:29:26 +01:00
def get_utxos(self, read_only=False, **constraints):
2020-04-02 02:44:34 +02:00
return self.get_txos(is_spent=False, read_only=read_only, **constraints)
2020-01-03 04:18:49 +01:00
def get_utxo_count(self, **constraints):
2020-04-02 02:44:34 +02:00
return self.get_txo_count(is_spent=False, **constraints)
2020-01-03 04:18:49 +01:00
2020-03-17 15:29:26 +01:00
async def get_balance(self, wallet=None, accounts=None, read_only=False, **constraints):
2020-01-03 04:18:49 +01:00
assert wallet or accounts, \
"'wallet' or 'accounts' constraints required to calculate balance"
constraints['accounts'] = accounts or wallet.accounts
2020-04-02 02:44:34 +02:00
balance = await self.select_txos(
'SUM(amount) as total', is_spent=False, read_only=read_only, **constraints
)
return balance[0]['total'] or 0
2020-01-03 04:18:49 +01:00
2020-03-17 15:29:26 +01:00
async def select_addresses(self, cols, read_only=False, **constraints):
2020-01-03 04:18:49 +01:00
return await self.db.execute_fetchall(*query(
f"SELECT {cols} FROM pubkey_address JOIN account_address USING (address)",
**constraints
2020-02-21 04:11:25 +01:00
), read_only=read_only)
2020-01-03 04:18:49 +01:00
2020-03-17 15:29:26 +01:00
async def get_addresses(self, cols=None, read_only=False, **constraints):
2020-01-03 04:18:49 +01:00
cols = cols or (
'address', 'account', 'chain', 'history', 'used_times',
'pubkey', 'chain_code', 'n', 'depth'
)
addresses = await self.select_addresses(', '.join(cols), read_only=read_only, **constraints)
2020-01-03 04:18:49 +01:00
if 'pubkey' in cols:
for address in addresses:
address['pubkey'] = PubKey(
self.ledger, address.pop('pubkey'), address.pop('chain_code'),
address.pop('n'), address.pop('depth')
)
return addresses
2020-03-17 15:29:26 +01:00
async def get_address_count(self, cols=None, read_only=False, **constraints):
count = await self.select_addresses('COUNT(*) as total', read_only=read_only, **constraints)
return count[0]['total'] or 0
2020-01-03 04:18:49 +01:00
2020-03-17 15:29:26 +01:00
async def get_address(self, read_only=False, **constraints):
2020-02-25 20:18:09 +01:00
addresses = await self.get_addresses(read_only=read_only, limit=1, **constraints)
2020-01-03 04:18:49 +01:00
if addresses:
return addresses[0]
async def add_keys(self, account, chain, pubkeys):
await self.db.executemany(
"insert or ignore into account_address "
"(account, address, chain, pubkey, chain_code, n, depth) values "
"(?, ?, ?, ?, ?, ?, ?)", ((
account.id, k.address, chain,
sqlite3.Binary(k.pubkey_bytes),
sqlite3.Binary(k.chain_code),
k.n, k.depth
) for k in pubkeys)
)
await self.db.executemany(
"insert or ignore into pubkey_address (address) values (?)",
((pubkey.address,) for pubkey in pubkeys)
)
async def _set_address_history(self, address, history):
await self.db.execute_fetchall(
"UPDATE pubkey_address SET history = ?, used_times = ? WHERE address = ?",
(history, history.count(':')//2, address)
)
async def set_address_history(self, address, history):
await self._set_address_history(address, history)
@staticmethod
def constrain_purchases(constraints):
accounts = constraints.pop('accounts', None)
assert accounts, "'accounts' argument required to find purchases"
if not {'purchased_claim_id', 'purchased_claim_id__in'}.intersection(constraints):
constraints['purchased_claim_id__is_not_null'] = True
constraints.update({
f'$account{i}': a.public_key.address for i, a in enumerate(accounts)
})
account_values = ', '.join([f':$account{i}' for i in range(len(accounts))])
constraints['txid__in'] = f"""
SELECT txid FROM txi JOIN account_address USING (address)
WHERE account_address.account IN ({account_values})
"""
async def get_purchases(self, **constraints):
self.constrain_purchases(constraints)
return [tx.outputs[0] for tx in await self.get_transactions(**constraints)]
def get_purchase_count(self, **constraints):
self.constrain_purchases(constraints)
return self.get_transaction_count(**constraints)
@staticmethod
def constrain_claims(constraints):
2020-03-07 06:34:47 +01:00
if {'txo_type', 'txo_type__in'}.intersection(constraints):
return
claim_types = constraints.pop('claim_type', None)
2020-03-07 06:34:47 +01:00
if claim_types:
constrain_single_or_list(
constraints, 'txo_type', claim_types, lambda x: TXO_TYPES[x]
)
else:
constraints['txo_type__in'] = CLAIM_TYPES
2019-11-12 18:17:35 +01:00
2020-03-17 15:29:26 +01:00
async def get_claims(self, read_only=False, **constraints) -> List[Output]:
self.constrain_claims(constraints)
2020-02-25 20:18:09 +01:00
return await self.get_utxos(read_only=read_only, **constraints)
def get_claim_count(self, **constraints):
self.constrain_claims(constraints)
return self.get_utxo_count(**constraints)
@staticmethod
def constrain_streams(constraints):
constraints['txo_type'] = TXO_TYPES['stream']
2020-03-17 15:29:26 +01:00
def get_streams(self, read_only=False, **constraints):
self.constrain_streams(constraints)
2020-02-25 20:18:09 +01:00
return self.get_claims(read_only=read_only, **constraints)
def get_stream_count(self, **constraints):
self.constrain_streams(constraints)
return self.get_claim_count(**constraints)
@staticmethod
def constrain_channels(constraints):
constraints['txo_type'] = TXO_TYPES['channel']
def get_channels(self, **constraints):
self.constrain_channels(constraints)
return self.get_claims(**constraints)
2018-07-12 05:18:59 +02:00
def get_channel_count(self, **constraints):
self.constrain_channels(constraints)
return self.get_claim_count(**constraints)
2019-03-24 21:55:04 +01:00
@staticmethod
def constrain_supports(constraints):
constraints['txo_type'] = TXO_TYPES['support']
2019-03-24 21:55:04 +01:00
def get_supports(self, **constraints):
self.constrain_supports(constraints)
return self.get_utxos(**constraints)
def get_support_count(self, **constraints):
self.constrain_supports(constraints)
2019-11-12 18:17:35 +01:00
return self.get_utxo_count(**constraints)
@staticmethod
def constrain_collections(constraints):
constraints['txo_type'] = TXO_TYPES['collection']
def get_collections(self, **constraints):
2019-11-13 23:50:35 +01:00
self.constrain_collections(constraints)
2019-11-12 18:17:35 +01:00
return self.get_utxos(**constraints)
def get_collection_count(self, **constraints):
2019-11-13 23:50:35 +01:00
self.constrain_collections(constraints)
2019-03-24 21:55:04 +01:00
return self.get_utxo_count(**constraints)
2019-01-04 08:49:29 +01:00
2020-08-04 21:02:43 +02:00
async def release_all_outputs(self, account=None):
if account is None:
await self.db.execute_fetchall("UPDATE txo SET is_reserved = 0 WHERE is_reserved = 1")
else:
await self.db.execute_fetchall(
"UPDATE txo SET is_reserved = 0 WHERE"
" is_reserved = 1 AND txo.address IN ("
" SELECT address from account_address WHERE account = ?"
" )", (account.public_key.address, )
)
2020-03-20 23:24:24 +01:00
def get_supports_summary(self, read_only=False, **constraints):
return self.get_txos(
txo_type=TXO_TYPES['support'],
2020-04-02 02:44:34 +02:00
is_spent=False, is_my_output=True,
2020-03-20 23:24:24 +01:00
include_is_my_input=True,
no_tx=True, read_only=read_only,
**constraints
)