2018-11-03 23:50:34 +01:00
|
|
|
# Copyright (c) 2016-2018, Neil Booth
|
|
|
|
#
|
|
|
|
# All rights reserved.
|
|
|
|
#
|
|
|
|
# See the file "LICENCE" for information about the copyright
|
|
|
|
# and warranty status of this software.
|
|
|
|
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Mempool handling."""
|
2020-03-01 20:40:50 +01:00
|
|
|
import os
|
2018-12-06 20:27:38 +01:00
|
|
|
import asyncio
|
2018-11-03 23:50:34 +01:00
|
|
|
import itertools
|
|
|
|
import time
|
|
|
|
from abc import ABC, abstractmethod
|
|
|
|
from collections import defaultdict
|
2020-02-25 20:17:33 +01:00
|
|
|
from concurrent.futures.thread import ThreadPoolExecutor
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
import attr
|
|
|
|
|
2019-12-31 20:52:57 +01:00
|
|
|
from lbry.wallet.server.hash import hash_to_hex_str, hex_str_to_hash
|
|
|
|
from lbry.wallet.server.util import class_logger, chunks
|
|
|
|
from lbry.wallet.server.leveldb import UTXO
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
|
|
|
|
@attr.s(slots=True)
|
2018-11-26 02:51:41 +01:00
|
|
|
class MemPoolTx:
|
2018-11-03 23:50:34 +01:00
|
|
|
prevouts = attr.ib()
|
|
|
|
# A pair is a (hashX, value) tuple
|
|
|
|
in_pairs = attr.ib()
|
|
|
|
out_pairs = attr.ib()
|
|
|
|
fee = attr.ib()
|
|
|
|
size = attr.ib()
|
|
|
|
|
|
|
|
|
|
|
|
@attr.s(slots=True)
|
2018-11-26 02:51:41 +01:00
|
|
|
class MemPoolTxSummary:
|
2018-11-03 23:50:34 +01:00
|
|
|
hash = attr.ib()
|
|
|
|
fee = attr.ib()
|
|
|
|
has_unconfirmed_inputs = attr.ib()
|
|
|
|
|
|
|
|
|
|
|
|
class MemPoolAPI(ABC):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""A concrete instance of this class is passed to the MemPool object
|
|
|
|
and used by it to query DB and blockchain state."""
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
async def height(self):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Query bitcoind for its height."""
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
def cached_height(self):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Return the height of bitcoind the last time it was queried,
|
2018-11-03 23:50:34 +01:00
|
|
|
for any reason, without actually querying it.
|
2019-04-16 09:50:35 +02:00
|
|
|
"""
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
async def mempool_hashes(self):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Query bitcoind for the hashes of all transactions in its
|
|
|
|
mempool, returned as a list."""
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
async def raw_transactions(self, hex_hashes):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Query bitcoind for the serialized raw transactions with the given
|
2018-11-03 23:50:34 +01:00
|
|
|
hashes. Missing transactions are returned as None.
|
|
|
|
|
2019-04-16 09:50:35 +02:00
|
|
|
hex_hashes is an iterable of hexadecimal hash strings."""
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
async def lookup_utxos(self, prevouts):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Return a list of (hashX, value) pairs each prevout if unspent,
|
2018-11-03 23:50:34 +01:00
|
|
|
otherwise return None if spent or not found.
|
|
|
|
|
|
|
|
prevouts - an iterable of (hash, index) pairs
|
2019-04-16 09:50:35 +02:00
|
|
|
"""
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
@abstractmethod
|
|
|
|
async def on_mempool(self, touched, height):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Called each time the mempool is synchronized. touched is a set of
|
2018-11-03 23:50:34 +01:00
|
|
|
hashXs touched since the previous call. height is the
|
2019-04-16 09:50:35 +02:00
|
|
|
daemon's height at the time the mempool was obtained."""
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
|
2018-11-26 02:51:41 +01:00
|
|
|
class MemPool:
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Representation of the daemon's mempool.
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
coin - a coin class from coins.py
|
|
|
|
api - an object implementing MemPoolAPI
|
|
|
|
|
|
|
|
Updated regularly in caught-up state. Goal is to enable efficient
|
|
|
|
response to the calls in the external interface. To that end we
|
|
|
|
maintain the following maps:
|
|
|
|
|
|
|
|
tx: tx_hash -> MemPoolTx
|
|
|
|
hashXs: hashX -> set of all hashes of txs touching the hashX
|
2019-04-16 09:50:35 +02:00
|
|
|
"""
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
def __init__(self, coin, api, refresh_secs=5.0, log_status_secs=120.0):
|
|
|
|
assert isinstance(api, MemPoolAPI)
|
|
|
|
self.coin = coin
|
|
|
|
self.api = api
|
|
|
|
self.logger = class_logger(__name__, self.__class__.__name__)
|
|
|
|
self.txs = {}
|
|
|
|
self.hashXs = defaultdict(set) # None can be a key
|
|
|
|
self.cached_compact_histogram = []
|
|
|
|
self.refresh_secs = refresh_secs
|
|
|
|
self.log_status_secs = log_status_secs
|
|
|
|
# Prevents mempool refreshes during fee histogram calculation
|
2019-12-18 03:00:28 +01:00
|
|
|
self.lock = asyncio.Lock()
|
|
|
|
self.wakeup = asyncio.Event()
|
2020-03-01 20:40:50 +01:00
|
|
|
self.executor = ThreadPoolExecutor(max(os.cpu_count() - 1, 1))
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
async def _logging(self, synchronized_event):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Print regular logs of mempool stats."""
|
2018-11-03 23:50:34 +01:00
|
|
|
self.logger.info('beginning processing of daemon mempool. '
|
|
|
|
'This can take some time...')
|
|
|
|
start = time.time()
|
|
|
|
await synchronized_event.wait()
|
|
|
|
elapsed = time.time() - start
|
|
|
|
self.logger.info(f'synced in {elapsed:.2f}s')
|
|
|
|
while True:
|
|
|
|
self.logger.info(f'{len(self.txs):,d} txs '
|
|
|
|
f'touching {len(self.hashXs):,d} addresses')
|
2019-12-18 03:00:28 +01:00
|
|
|
await asyncio.sleep(self.log_status_secs)
|
2018-11-03 23:50:34 +01:00
|
|
|
await synchronized_event.wait()
|
|
|
|
|
|
|
|
async def _refresh_histogram(self, synchronized_event):
|
2020-02-25 20:17:33 +01:00
|
|
|
try:
|
|
|
|
while True:
|
|
|
|
await synchronized_event.wait()
|
|
|
|
async with self.lock:
|
|
|
|
# Threaded as can be expensive
|
|
|
|
await asyncio.get_event_loop().run_in_executor(self.executor, self._update_histogram, 100_000)
|
|
|
|
await asyncio.sleep(self.coin.MEMPOOL_HISTOGRAM_REFRESH_SECS)
|
|
|
|
finally:
|
|
|
|
self.executor.shutdown(wait=True)
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
def _update_histogram(self, bin_size):
|
|
|
|
# Build a histogram by fee rate
|
|
|
|
histogram = defaultdict(int)
|
|
|
|
for tx in self.txs.values():
|
|
|
|
histogram[tx.fee // tx.size] += tx.size
|
|
|
|
|
|
|
|
# Now compact it. For efficiency, get_fees returns a
|
|
|
|
# compact histogram with variable bin size. The compact
|
|
|
|
# histogram is an array of (fee_rate, vsize) values.
|
|
|
|
# vsize_n is the cumulative virtual size of mempool
|
|
|
|
# transactions with a fee rate in the interval
|
|
|
|
# [rate_(n-1), rate_n)], and rate_(n-1) > rate_n.
|
|
|
|
# Intervals are chosen to create tranches containing at
|
|
|
|
# least 100kb of transactions
|
|
|
|
compact = []
|
|
|
|
cum_size = 0
|
|
|
|
r = 0 # ?
|
|
|
|
for fee_rate, size in sorted(histogram.items(), reverse=True):
|
|
|
|
cum_size += size
|
|
|
|
if cum_size + r > bin_size:
|
|
|
|
compact.append((fee_rate, cum_size))
|
|
|
|
r += cum_size - bin_size
|
|
|
|
cum_size = 0
|
|
|
|
bin_size *= 1.1
|
|
|
|
self.logger.info(f'compact fee histogram: {compact}')
|
|
|
|
self.cached_compact_histogram = compact
|
|
|
|
|
|
|
|
def _accept_transactions(self, tx_map, utxo_map, touched):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Accept transactions in tx_map to the mempool if all their inputs
|
2018-11-03 23:50:34 +01:00
|
|
|
can be found in the existing mempool or a utxo_map from the
|
|
|
|
DB.
|
|
|
|
|
|
|
|
Returns an (unprocessed tx_map, unspent utxo_map) pair.
|
2019-04-16 09:50:35 +02:00
|
|
|
"""
|
2018-11-03 23:50:34 +01:00
|
|
|
hashXs = self.hashXs
|
|
|
|
txs = self.txs
|
|
|
|
|
|
|
|
deferred = {}
|
|
|
|
unspent = set(utxo_map)
|
|
|
|
# Try to find all prevouts so we can accept the TX
|
|
|
|
for hash, tx in tx_map.items():
|
|
|
|
in_pairs = []
|
|
|
|
try:
|
|
|
|
for prevout in tx.prevouts:
|
|
|
|
utxo = utxo_map.get(prevout)
|
|
|
|
if not utxo:
|
|
|
|
prev_hash, prev_index = prevout
|
|
|
|
# Raises KeyError if prev_hash is not in txs
|
|
|
|
utxo = txs[prev_hash].out_pairs[prev_index]
|
|
|
|
in_pairs.append(utxo)
|
|
|
|
except KeyError:
|
|
|
|
deferred[hash] = tx
|
|
|
|
continue
|
|
|
|
|
|
|
|
# Spend the prevouts
|
|
|
|
unspent.difference_update(tx.prevouts)
|
|
|
|
|
|
|
|
# Save the in_pairs, compute the fee and accept the TX
|
|
|
|
tx.in_pairs = tuple(in_pairs)
|
|
|
|
# Avoid negative fees if dealing with generation-like transactions
|
|
|
|
# because some in_parts would be missing
|
|
|
|
tx.fee = max(0, (sum(v for _, v in tx.in_pairs) -
|
|
|
|
sum(v for _, v in tx.out_pairs)))
|
|
|
|
txs[hash] = tx
|
|
|
|
|
|
|
|
for hashX, value in itertools.chain(tx.in_pairs, tx.out_pairs):
|
|
|
|
touched.add(hashX)
|
|
|
|
hashXs[hashX].add(hash)
|
|
|
|
|
|
|
|
return deferred, {prevout: utxo_map[prevout] for prevout in unspent}
|
|
|
|
|
|
|
|
async def _refresh_hashes(self, synchronized_event):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Refresh our view of the daemon's mempool."""
|
2018-11-03 23:50:34 +01:00
|
|
|
while True:
|
|
|
|
height = self.api.cached_height()
|
|
|
|
hex_hashes = await self.api.mempool_hashes()
|
|
|
|
if height != await self.api.height():
|
|
|
|
continue
|
2019-10-02 18:38:56 +02:00
|
|
|
hashes = {hex_str_to_hash(hh) for hh in hex_hashes}
|
2018-11-03 23:50:34 +01:00
|
|
|
async with self.lock:
|
|
|
|
touched = await self._process_mempool(hashes)
|
|
|
|
synchronized_event.set()
|
|
|
|
synchronized_event.clear()
|
|
|
|
await self.api.on_mempool(touched, height)
|
2019-12-18 03:00:28 +01:00
|
|
|
try:
|
2019-12-19 20:20:12 +01:00
|
|
|
# we wait up to `refresh_secs` but go early if a broadcast happens (which triggers wakeup event)
|
2019-12-18 03:00:28 +01:00
|
|
|
await asyncio.wait_for(self.wakeup.wait(), timeout=self.refresh_secs)
|
|
|
|
except asyncio.TimeoutError:
|
|
|
|
pass
|
|
|
|
finally:
|
|
|
|
self.wakeup.clear()
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
async def _process_mempool(self, all_hashes):
|
|
|
|
# Re-sync with the new set of hashes
|
|
|
|
txs = self.txs
|
|
|
|
hashXs = self.hashXs
|
|
|
|
touched = set()
|
|
|
|
|
|
|
|
# First handle txs that have disappeared
|
|
|
|
for tx_hash in set(txs).difference(all_hashes):
|
|
|
|
tx = txs.pop(tx_hash)
|
2019-10-02 18:38:56 +02:00
|
|
|
tx_hashXs = {hashX for hashX, value in tx.in_pairs}
|
2018-11-03 23:50:34 +01:00
|
|
|
tx_hashXs.update(hashX for hashX, value in tx.out_pairs)
|
|
|
|
for hashX in tx_hashXs:
|
|
|
|
hashXs[hashX].remove(tx_hash)
|
|
|
|
if not hashXs[hashX]:
|
|
|
|
del hashXs[hashX]
|
|
|
|
touched.update(tx_hashXs)
|
|
|
|
|
|
|
|
# Process new transactions
|
|
|
|
new_hashes = list(all_hashes.difference(txs))
|
|
|
|
if new_hashes:
|
2018-12-07 02:03:22 +01:00
|
|
|
fetches = []
|
2018-11-03 23:50:34 +01:00
|
|
|
for hashes in chunks(new_hashes, 200):
|
2018-12-07 02:03:22 +01:00
|
|
|
fetches.append(self._fetch_and_accept(hashes, all_hashes, touched))
|
2018-11-03 23:50:34 +01:00
|
|
|
tx_map = {}
|
|
|
|
utxo_map = {}
|
2018-12-07 02:03:22 +01:00
|
|
|
for fetch in asyncio.as_completed(fetches):
|
|
|
|
deferred, unspent = await fetch
|
2018-11-03 23:50:34 +01:00
|
|
|
tx_map.update(deferred)
|
|
|
|
utxo_map.update(unspent)
|
|
|
|
|
|
|
|
prior_count = 0
|
|
|
|
# FIXME: this is not particularly efficient
|
|
|
|
while tx_map and len(tx_map) != prior_count:
|
|
|
|
prior_count = len(tx_map)
|
|
|
|
tx_map, utxo_map = self._accept_transactions(tx_map, utxo_map,
|
|
|
|
touched)
|
|
|
|
if tx_map:
|
|
|
|
self.logger.info(f'{len(tx_map)} txs dropped')
|
|
|
|
|
|
|
|
return touched
|
|
|
|
|
|
|
|
async def _fetch_and_accept(self, hashes, all_hashes, touched):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Fetch a list of mempool transactions."""
|
2018-11-03 23:50:34 +01:00
|
|
|
hex_hashes_iter = (hash_to_hex_str(hash) for hash in hashes)
|
|
|
|
raw_txs = await self.api.raw_transactions(hex_hashes_iter)
|
|
|
|
|
|
|
|
def deserialize_txs(): # This function is pure
|
|
|
|
to_hashX = self.coin.hashX_from_script
|
|
|
|
deserializer = self.coin.DESERIALIZER
|
|
|
|
|
|
|
|
txs = {}
|
|
|
|
for hash, raw_tx in zip(hashes, raw_txs):
|
|
|
|
# The daemon may have evicted the tx from its
|
|
|
|
# mempool or it may have gotten in a block
|
|
|
|
if not raw_tx:
|
|
|
|
continue
|
|
|
|
tx, tx_size = deserializer(raw_tx).read_tx_and_vsize()
|
|
|
|
# Convert the inputs and outputs into (hashX, value) pairs
|
|
|
|
# Drop generation-like inputs from MemPoolTx.prevouts
|
|
|
|
txin_pairs = tuple((txin.prev_hash, txin.prev_idx)
|
|
|
|
for txin in tx.inputs
|
|
|
|
if not txin.is_generation())
|
|
|
|
txout_pairs = tuple((to_hashX(txout.pk_script), txout.value)
|
|
|
|
for txout in tx.outputs)
|
|
|
|
txs[hash] = MemPoolTx(txin_pairs, None, txout_pairs,
|
|
|
|
0, tx_size)
|
|
|
|
return txs
|
|
|
|
|
|
|
|
# Thread this potentially slow operation so as not to block
|
2020-02-25 20:17:33 +01:00
|
|
|
tx_map = await asyncio.get_event_loop().run_in_executor(self.executor, deserialize_txs)
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
# Determine all prevouts not in the mempool, and fetch the
|
|
|
|
# UTXO information from the database. Failed prevout lookups
|
|
|
|
# return None - concurrent database updates happen - which is
|
|
|
|
# relied upon by _accept_transactions. Ignore prevouts that are
|
|
|
|
# generation-like.
|
|
|
|
prevouts = tuple(prevout for tx in tx_map.values()
|
|
|
|
for prevout in tx.prevouts
|
|
|
|
if prevout[0] not in all_hashes)
|
|
|
|
utxos = await self.api.lookup_utxos(prevouts)
|
|
|
|
utxo_map = {prevout: utxo for prevout, utxo in zip(prevouts, utxos)}
|
|
|
|
|
|
|
|
return self._accept_transactions(tx_map, utxo_map, touched)
|
|
|
|
|
|
|
|
#
|
|
|
|
# External interface
|
|
|
|
#
|
|
|
|
|
|
|
|
async def keep_synchronized(self, synchronized_event):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Keep the mempool synchronized with the daemon."""
|
2018-12-07 02:03:22 +01:00
|
|
|
await asyncio.wait([
|
|
|
|
self._refresh_hashes(synchronized_event),
|
|
|
|
self._refresh_histogram(synchronized_event),
|
|
|
|
self._logging(synchronized_event)
|
|
|
|
])
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
async def balance_delta(self, hashX):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Return the unconfirmed amount in the mempool for hashX.
|
2018-11-03 23:50:34 +01:00
|
|
|
|
|
|
|
Can be positive or negative.
|
2019-04-16 09:50:35 +02:00
|
|
|
"""
|
2018-11-03 23:50:34 +01:00
|
|
|
value = 0
|
|
|
|
if hashX in self.hashXs:
|
|
|
|
for hash in self.hashXs[hashX]:
|
|
|
|
tx = self.txs[hash]
|
|
|
|
value -= sum(v for h168, v in tx.in_pairs if h168 == hashX)
|
|
|
|
value += sum(v for h168, v in tx.out_pairs if h168 == hashX)
|
|
|
|
return value
|
|
|
|
|
|
|
|
async def compact_fee_histogram(self):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Return a compact fee histogram of the current mempool."""
|
2018-11-03 23:50:34 +01:00
|
|
|
return self.cached_compact_histogram
|
|
|
|
|
|
|
|
async def potential_spends(self, hashX):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Return a set of (prev_hash, prev_idx) pairs from mempool
|
2018-11-03 23:50:34 +01:00
|
|
|
transactions that touch hashX.
|
|
|
|
|
|
|
|
None, some or all of these may be spends of the hashX, but all
|
|
|
|
actual spends of it (in the DB or mempool) will be included.
|
2019-04-16 09:50:35 +02:00
|
|
|
"""
|
2018-11-03 23:50:34 +01:00
|
|
|
result = set()
|
|
|
|
for tx_hash in self.hashXs.get(hashX, ()):
|
|
|
|
tx = self.txs[tx_hash]
|
|
|
|
result.update(tx.prevouts)
|
|
|
|
return result
|
|
|
|
|
|
|
|
async def transaction_summaries(self, hashX):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Return a list of MemPoolTxSummary objects for the hashX."""
|
2018-11-03 23:50:34 +01:00
|
|
|
result = []
|
|
|
|
for tx_hash in self.hashXs.get(hashX, ()):
|
|
|
|
tx = self.txs[tx_hash]
|
|
|
|
has_ui = any(hash in self.txs for hash, idx in tx.prevouts)
|
|
|
|
result.append(MemPoolTxSummary(tx_hash, tx.fee, has_ui))
|
|
|
|
return result
|
|
|
|
|
|
|
|
async def unordered_UTXOs(self, hashX):
|
2019-04-16 09:50:35 +02:00
|
|
|
"""Return an unordered list of UTXO named tuples from mempool
|
2018-11-03 23:50:34 +01:00
|
|
|
transactions that pay to hashX.
|
|
|
|
|
|
|
|
This does not consider if any other mempool transactions spend
|
|
|
|
the outputs.
|
2019-04-16 09:50:35 +02:00
|
|
|
"""
|
2018-11-03 23:50:34 +01:00
|
|
|
utxos = []
|
|
|
|
for tx_hash in self.hashXs.get(hashX, ()):
|
|
|
|
tx = self.txs.get(tx_hash)
|
|
|
|
for pos, (hX, value) in enumerate(tx.out_pairs):
|
|
|
|
if hX == hashX:
|
|
|
|
utxos.append(UTXO(-1, pos, tx_hash, 0, value))
|
|
|
|
return utxos
|