2020-06-26 16:39:58 +02:00
|
|
|
import os
|
|
|
|
import asyncio
|
|
|
|
import logging
|
2020-07-12 00:18:33 +02:00
|
|
|
from typing import Optional, Tuple, Set, List, Coroutine
|
2020-09-16 16:37:49 +02:00
|
|
|
from concurrent.futures import ThreadPoolExecutor
|
2020-06-26 16:39:58 +02:00
|
|
|
|
|
|
|
from lbry.db import Database
|
2020-07-12 00:18:33 +02:00
|
|
|
from lbry.db import queries as q
|
2020-07-13 15:30:32 +02:00
|
|
|
from lbry.db.constants import TXO_TYPES, CLAIM_TYPE_CODES
|
2020-07-12 00:18:33 +02:00
|
|
|
from lbry.db.query_context import Event, Progress
|
2020-09-16 16:37:49 +02:00
|
|
|
from lbry.event import BroadcastSubscription, EventController
|
2020-06-26 16:39:58 +02:00
|
|
|
from lbry.service.base import Sync, BlockEvent
|
|
|
|
from lbry.blockchain.lbrycrd import Lbrycrd
|
2020-09-14 16:53:13 +02:00
|
|
|
from lbry.error import LbrycrdEventSubscriptionError
|
2020-06-26 16:39:58 +02:00
|
|
|
|
2020-07-12 00:18:33 +02:00
|
|
|
from . import blocks as block_phase, claims as claim_phase, supports as support_phase
|
2020-09-16 16:37:49 +02:00
|
|
|
from .context import uninitialize
|
2020-09-21 16:26:19 +02:00
|
|
|
from .filter_builder import split_range_into_10k_batches
|
2020-06-26 16:39:58 +02:00
|
|
|
|
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
|
2020-07-13 06:55:30 +02:00
|
|
|
BLOCKS_INIT_EVENT = Event.add("blockchain.sync.blocks.init", "steps")
|
|
|
|
BLOCKS_MAIN_EVENT = Event.add("blockchain.sync.blocks.main", "blocks", "txs")
|
2020-09-21 16:26:19 +02:00
|
|
|
FILTER_INIT_EVENT = Event.add("blockchain.sync.filters.init", "steps")
|
|
|
|
FILTER_MAIN_EVENT = Event.add("blockchain.sync.filters.main", "blocks")
|
2020-07-13 06:55:30 +02:00
|
|
|
CLAIMS_INIT_EVENT = Event.add("blockchain.sync.claims.init", "steps")
|
|
|
|
CLAIMS_MAIN_EVENT = Event.add("blockchain.sync.claims.main", "claims")
|
|
|
|
TRENDS_INIT_EVENT = Event.add("blockchain.sync.trends.init", "steps")
|
|
|
|
TRENDS_MAIN_EVENT = Event.add("blockchain.sync.trends.main", "blocks")
|
|
|
|
SUPPORTS_INIT_EVENT = Event.add("blockchain.sync.supports.init", "steps")
|
|
|
|
SUPPORTS_MAIN_EVENT = Event.add("blockchain.sync.supports.main", "supports")
|
2020-07-12 00:18:33 +02:00
|
|
|
|
2020-06-26 16:39:58 +02:00
|
|
|
|
|
|
|
class BlockchainSync(Sync):
|
|
|
|
|
2020-08-03 18:24:13 +02:00
|
|
|
TX_FLUSH_SIZE = 25_000 # flush to db after processing this many TXs and update progress
|
|
|
|
CLAIM_FLUSH_SIZE = 25_000 # flush to db after processing this many claims and update progress
|
|
|
|
SUPPORT_FLUSH_SIZE = 25_000 # flush to db after processing this many supports and update progress
|
2020-07-12 00:18:33 +02:00
|
|
|
FILTER_FLUSH_SIZE = 10_000 # flush to db after processing this many filters and update progress
|
|
|
|
|
2020-06-26 16:39:58 +02:00
|
|
|
def __init__(self, chain: Lbrycrd, db: Database):
|
|
|
|
super().__init__(chain.ledger, db)
|
|
|
|
self.chain = chain
|
2020-07-12 00:18:33 +02:00
|
|
|
self.pid = os.getpid()
|
2020-09-16 16:37:49 +02:00
|
|
|
self.on_block_hash_subscription: Optional[BroadcastSubscription] = None
|
|
|
|
self.on_tx_hash_subscription: Optional[BroadcastSubscription] = None
|
2020-06-26 16:39:58 +02:00
|
|
|
self.advance_loop_task: Optional[asyncio.Task] = None
|
2020-09-16 16:37:49 +02:00
|
|
|
self.block_hash_event = asyncio.Event()
|
|
|
|
self.tx_hash_event = asyncio.Event()
|
|
|
|
self._on_mempool_controller = EventController()
|
|
|
|
self.on_mempool = self._on_mempool_controller.stream
|
2020-06-26 16:39:58 +02:00
|
|
|
|
2020-09-07 22:17:51 +02:00
|
|
|
async def wait_for_chain_ready(self):
|
|
|
|
while True:
|
|
|
|
try:
|
|
|
|
return await self.chain.ensure_subscribable()
|
|
|
|
except asyncio.CancelledError:
|
|
|
|
raise
|
2020-09-14 16:53:13 +02:00
|
|
|
except LbrycrdEventSubscriptionError as e:
|
|
|
|
log.warning(
|
|
|
|
"Lbrycrd is misconfigured. Please double check if"
|
|
|
|
" zmqpubhashblock is properly set on lbrycrd.conf"
|
|
|
|
)
|
2020-09-11 19:03:58 +02:00
|
|
|
raise
|
2020-09-07 22:17:51 +02:00
|
|
|
except Exception as e:
|
|
|
|
log.warning("Blockchain not ready, waiting for it: %s", str(e))
|
|
|
|
await asyncio.sleep(1)
|
|
|
|
|
2020-06-26 16:39:58 +02:00
|
|
|
async def start(self):
|
2020-07-12 18:02:58 +02:00
|
|
|
self.db.stop_event.clear()
|
2020-09-07 22:17:51 +02:00
|
|
|
await self.wait_for_chain_ready()
|
2020-07-12 00:18:33 +02:00
|
|
|
self.advance_loop_task = asyncio.create_task(self.advance())
|
|
|
|
await self.advance_loop_task
|
2020-07-27 16:58:57 +02:00
|
|
|
await self.chain.subscribe()
|
2020-06-26 16:39:58 +02:00
|
|
|
self.advance_loop_task = asyncio.create_task(self.advance_loop())
|
2020-09-16 16:37:49 +02:00
|
|
|
self.on_block_hash_subscription = self.chain.on_block_hash.listen(
|
|
|
|
lambda e: self.block_hash_event.set()
|
|
|
|
)
|
|
|
|
self.on_tx_hash_subscription = self.chain.on_tx_hash.listen(
|
|
|
|
lambda e: self.tx_hash_event.set()
|
2020-06-26 16:39:58 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
async def stop(self):
|
|
|
|
self.chain.unsubscribe()
|
|
|
|
self.db.stop_event.set()
|
2020-09-16 16:37:49 +02:00
|
|
|
for subscription in (
|
|
|
|
self.on_block_hash_subscription,
|
|
|
|
self.on_tx_hash_subscription,
|
|
|
|
self.advance_loop_task
|
|
|
|
):
|
|
|
|
if subscription is not None:
|
|
|
|
subscription.cancel()
|
|
|
|
if isinstance(self.db.executor, ThreadPoolExecutor):
|
|
|
|
await self.db.run(uninitialize)
|
2020-06-26 16:39:58 +02:00
|
|
|
|
2020-07-12 00:18:33 +02:00
|
|
|
async def run_tasks(self, tasks: List[Coroutine]) -> Optional[Set[asyncio.Future]]:
|
2020-06-26 16:39:58 +02:00
|
|
|
done, pending = await asyncio.wait(
|
|
|
|
tasks, return_when=asyncio.FIRST_EXCEPTION
|
|
|
|
)
|
|
|
|
if pending:
|
|
|
|
self.db.stop_event.set()
|
|
|
|
for future in pending:
|
|
|
|
future.cancel()
|
|
|
|
for future in done:
|
|
|
|
future.result()
|
|
|
|
return
|
2020-07-12 00:18:33 +02:00
|
|
|
return done
|
2020-06-26 16:39:58 +02:00
|
|
|
|
2020-07-12 00:18:33 +02:00
|
|
|
async def get_best_block_height_for_file(self, file_number) -> int:
|
|
|
|
return await self.db.run(
|
|
|
|
block_phase.get_best_block_height_for_file, file_number
|
2020-07-06 05:03:45 +02:00
|
|
|
)
|
2020-07-12 00:18:33 +02:00
|
|
|
|
|
|
|
async def sync_blocks(self) -> Optional[Tuple[int, int]]:
|
|
|
|
tasks = []
|
|
|
|
starting_height = None
|
|
|
|
tx_count = block_count = 0
|
2020-07-13 06:55:30 +02:00
|
|
|
with Progress(self.db.message_queue, BLOCKS_INIT_EVENT) as p:
|
2020-07-12 00:18:33 +02:00
|
|
|
ending_height = await self.chain.db.get_best_height()
|
|
|
|
for chain_file in p.iter(await self.chain.db.get_block_files()):
|
|
|
|
# block files may be read and saved out of order, need to check
|
|
|
|
# each file individually to see if we have missing blocks
|
|
|
|
our_best_file_height = await self.get_best_block_height_for_file(
|
|
|
|
chain_file['file_number']
|
|
|
|
)
|
|
|
|
if our_best_file_height == chain_file['best_height']:
|
|
|
|
# we have all blocks in this file, skipping
|
|
|
|
continue
|
|
|
|
if -1 < our_best_file_height < chain_file['best_height']:
|
|
|
|
# we have some blocks, need to figure out what we're missing
|
|
|
|
# call get_block_files again limited to this file and current_height
|
|
|
|
chain_file = (await self.chain.db.get_block_files(
|
|
|
|
file_number=chain_file['file_number'], start_height=our_best_file_height+1,
|
|
|
|
))[0]
|
|
|
|
tx_count += chain_file['txs']
|
|
|
|
block_count += chain_file['blocks']
|
2020-08-03 22:53:40 +02:00
|
|
|
file_start_height = chain_file['start_height']
|
2020-07-12 00:18:33 +02:00
|
|
|
starting_height = min(
|
2020-08-03 22:53:40 +02:00
|
|
|
file_start_height if starting_height is None else starting_height,
|
|
|
|
file_start_height
|
2020-07-12 00:18:33 +02:00
|
|
|
)
|
|
|
|
tasks.append(self.db.run(
|
2020-08-03 22:53:40 +02:00
|
|
|
block_phase.sync_block_file, chain_file['file_number'], file_start_height,
|
2020-07-12 00:18:33 +02:00
|
|
|
chain_file['txs'], self.TX_FLUSH_SIZE
|
|
|
|
))
|
2020-07-13 06:55:30 +02:00
|
|
|
with Progress(self.db.message_queue, BLOCKS_MAIN_EVENT) as p:
|
2020-07-12 00:18:33 +02:00
|
|
|
p.start(block_count, tx_count, extra={
|
|
|
|
"starting_height": starting_height,
|
|
|
|
"ending_height": ending_height,
|
|
|
|
"files": len(tasks),
|
|
|
|
"claims": await self.chain.db.get_claim_metadata_count(starting_height, ending_height),
|
|
|
|
"supports": await self.chain.db.get_support_metadata_count(starting_height, ending_height),
|
|
|
|
})
|
|
|
|
completed = await self.run_tasks(tasks)
|
|
|
|
if completed:
|
2020-09-21 16:26:19 +02:00
|
|
|
if starting_height == 0:
|
|
|
|
await self.db.run(block_phase.blocks_constraints_and_indexes)
|
|
|
|
else:
|
|
|
|
await self.db.run(block_phase.blocks_vacuum)
|
2020-07-12 00:18:33 +02:00
|
|
|
best_height_processed = max(f.result() for f in completed)
|
|
|
|
return starting_height, best_height_processed
|
|
|
|
|
|
|
|
async def sync_filters(self):
|
2020-09-21 16:26:19 +02:00
|
|
|
if not self.conf.spv_address_filters:
|
|
|
|
return
|
|
|
|
with Progress(self.db.message_queue, FILTER_INIT_EVENT) as p:
|
|
|
|
p.start(2)
|
|
|
|
initial_sync = not await self.db.has_filters()
|
|
|
|
p.step()
|
|
|
|
if initial_sync:
|
|
|
|
blocks = [0, await self.db.get_best_block_height()]
|
|
|
|
else:
|
|
|
|
blocks = await self.db.run(block_phase.get_block_range_without_filters)
|
|
|
|
if blocks != (-1, -1):
|
|
|
|
batches = split_range_into_10k_batches(*blocks)
|
|
|
|
p.step()
|
|
|
|
else:
|
|
|
|
p.step()
|
|
|
|
return
|
|
|
|
with Progress(self.db.message_queue, FILTER_MAIN_EVENT) as p:
|
|
|
|
p.start((blocks[1]-blocks[0])+1)
|
|
|
|
await self.run_tasks([
|
|
|
|
self.db.run(block_phase.sync_filters, *batch) for batch in batches
|
|
|
|
])
|
|
|
|
if initial_sync:
|
|
|
|
await self.db.run(block_phase.filters_constraints_and_indexes)
|
|
|
|
else:
|
|
|
|
await self.db.run(block_phase.filters_vacuum)
|
2020-07-12 00:18:33 +02:00
|
|
|
|
2020-07-13 06:55:30 +02:00
|
|
|
async def sync_spends(self, blocks_added):
|
2020-07-06 05:03:45 +02:00
|
|
|
if blocks_added:
|
2020-07-13 06:55:30 +02:00
|
|
|
await self.db.run(block_phase.sync_spends, blocks_added[0] == 0)
|
2020-07-12 00:18:33 +02:00
|
|
|
|
|
|
|
async def count_unspent_txos(
|
|
|
|
self,
|
|
|
|
txo_types: Tuple[int, ...],
|
|
|
|
blocks: Tuple[int, int] = None,
|
|
|
|
missing_in_supports_table: bool = False,
|
|
|
|
missing_in_claims_table: bool = False,
|
|
|
|
missing_or_stale_in_claims_table: bool = False,
|
|
|
|
) -> int:
|
|
|
|
return await self.db.run(
|
|
|
|
q.count_unspent_txos, txo_types, blocks,
|
|
|
|
missing_in_supports_table,
|
|
|
|
missing_in_claims_table,
|
|
|
|
missing_or_stale_in_claims_table,
|
|
|
|
)
|
|
|
|
|
|
|
|
async def distribute_unspent_txos(
|
|
|
|
self,
|
|
|
|
txo_types: Tuple[int, ...],
|
|
|
|
blocks: Tuple[int, int] = None,
|
|
|
|
missing_in_supports_table: bool = False,
|
|
|
|
missing_in_claims_table: bool = False,
|
|
|
|
missing_or_stale_in_claims_table: bool = False,
|
|
|
|
) -> int:
|
|
|
|
return await self.db.run(
|
|
|
|
q.distribute_unspent_txos, txo_types, blocks,
|
|
|
|
missing_in_supports_table,
|
|
|
|
missing_in_claims_table,
|
|
|
|
missing_or_stale_in_claims_table,
|
2020-08-03 18:24:13 +02:00
|
|
|
self.db.workers
|
2020-07-12 00:18:33 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
async def count_abandoned_supports(self) -> int:
|
|
|
|
return await self.db.run(q.count_abandoned_supports)
|
|
|
|
|
|
|
|
async def count_abandoned_claims(self) -> int:
|
|
|
|
return await self.db.run(q.count_abandoned_claims)
|
|
|
|
|
|
|
|
async def count_claims_with_changed_supports(self, blocks) -> int:
|
|
|
|
return await self.db.run(q.count_claims_with_changed_supports, blocks)
|
|
|
|
|
2020-09-24 21:19:30 +02:00
|
|
|
async def count_claims_with_changed_reposts(self, blocks) -> int:
|
|
|
|
return await self.db.run(q.count_claims_with_changed_reposts, blocks)
|
|
|
|
|
2020-07-12 00:18:33 +02:00
|
|
|
async def count_channels_with_changed_content(self, blocks) -> int:
|
|
|
|
return await self.db.run(q.count_channels_with_changed_content, blocks)
|
|
|
|
|
|
|
|
async def count_takeovers(self, blocks) -> int:
|
|
|
|
return await self.chain.db.get_takeover_count(
|
|
|
|
start_height=blocks[0], end_height=blocks[-1]
|
|
|
|
)
|
|
|
|
|
2020-07-13 15:30:32 +02:00
|
|
|
async def sync_claims(self, blocks) -> bool:
|
2020-09-24 21:19:30 +02:00
|
|
|
delete_claims = takeovers = claims_with_changed_supports = claims_with_changed_reposts = 0
|
2020-07-12 00:18:33 +02:00
|
|
|
initial_sync = not await self.db.has_claims()
|
2020-07-13 06:55:30 +02:00
|
|
|
with Progress(self.db.message_queue, CLAIMS_INIT_EVENT) as p:
|
2020-07-12 00:18:33 +02:00
|
|
|
if initial_sync:
|
2020-07-13 15:30:32 +02:00
|
|
|
total, batches = await self.distribute_unspent_txos(CLAIM_TYPE_CODES)
|
2020-07-12 00:18:33 +02:00
|
|
|
elif blocks:
|
2020-09-24 21:19:30 +02:00
|
|
|
p.start(5)
|
2020-07-13 15:30:32 +02:00
|
|
|
# 1. content claims to be inserted or updated
|
|
|
|
total = await self.count_unspent_txos(
|
|
|
|
CLAIM_TYPE_CODES, blocks, missing_or_stale_in_claims_table=True
|
2020-07-12 00:18:33 +02:00
|
|
|
)
|
2020-07-13 15:30:32 +02:00
|
|
|
batches = [blocks] if total else []
|
2020-07-12 00:18:33 +02:00
|
|
|
p.step()
|
2020-07-13 15:30:32 +02:00
|
|
|
# 2. claims to be deleted
|
2020-07-12 00:18:33 +02:00
|
|
|
delete_claims = await self.count_abandoned_claims()
|
|
|
|
total += delete_claims
|
|
|
|
p.step()
|
2020-07-13 15:30:32 +02:00
|
|
|
# 3. claims to be updated with new support totals
|
2020-07-12 00:18:33 +02:00
|
|
|
claims_with_changed_supports = await self.count_claims_with_changed_supports(blocks)
|
|
|
|
total += claims_with_changed_supports
|
|
|
|
p.step()
|
2020-09-24 21:19:30 +02:00
|
|
|
# 4. claims to be updated with new repost totals
|
|
|
|
claims_with_changed_reposts = await self.count_claims_with_changed_reposts(blocks)
|
|
|
|
total += claims_with_changed_reposts
|
|
|
|
p.step()
|
2020-07-13 15:30:32 +02:00
|
|
|
# 5. claims to be updated due to name takeovers
|
2020-07-12 00:18:33 +02:00
|
|
|
takeovers = await self.count_takeovers(blocks)
|
|
|
|
total += takeovers
|
|
|
|
p.step()
|
|
|
|
else:
|
2020-07-13 15:30:32 +02:00
|
|
|
return initial_sync
|
2020-07-13 06:55:30 +02:00
|
|
|
with Progress(self.db.message_queue, CLAIMS_MAIN_EVENT) as p:
|
2020-07-12 00:18:33 +02:00
|
|
|
p.start(total)
|
2020-07-13 15:30:32 +02:00
|
|
|
if batches:
|
|
|
|
await self.run_tasks([
|
2020-08-03 18:24:13 +02:00
|
|
|
self.db.run(claim_phase.claims_insert, batch, not initial_sync, self.CLAIM_FLUSH_SIZE)
|
|
|
|
for batch in batches
|
2020-07-13 15:30:32 +02:00
|
|
|
])
|
|
|
|
if not initial_sync:
|
2020-07-12 00:18:33 +02:00
|
|
|
await self.run_tasks([
|
2020-07-13 15:30:32 +02:00
|
|
|
self.db.run(claim_phase.claims_update, batch) for batch in batches
|
2020-07-12 00:18:33 +02:00
|
|
|
])
|
|
|
|
if delete_claims:
|
|
|
|
await self.db.run(claim_phase.claims_delete, delete_claims)
|
|
|
|
if takeovers:
|
|
|
|
await self.db.run(claim_phase.update_takeovers, blocks, takeovers)
|
|
|
|
if claims_with_changed_supports:
|
|
|
|
await self.db.run(claim_phase.update_stakes, blocks, claims_with_changed_supports)
|
2020-09-24 21:19:30 +02:00
|
|
|
if claims_with_changed_reposts:
|
|
|
|
await self.db.run(claim_phase.update_reposts, blocks, claims_with_changed_reposts)
|
2020-07-13 06:55:30 +02:00
|
|
|
if initial_sync:
|
|
|
|
await self.db.run(claim_phase.claims_constraints_and_indexes)
|
2020-07-14 19:26:32 +02:00
|
|
|
else:
|
|
|
|
await self.db.run(claim_phase.claims_vacuum)
|
2020-07-13 15:30:32 +02:00
|
|
|
return initial_sync
|
2020-07-12 00:18:33 +02:00
|
|
|
|
|
|
|
async def sync_supports(self, blocks):
|
|
|
|
delete_supports = 0
|
|
|
|
initial_sync = not await self.db.has_supports()
|
2020-07-13 06:55:30 +02:00
|
|
|
with Progress(self.db.message_queue, SUPPORTS_INIT_EVENT) as p:
|
2020-07-12 00:18:33 +02:00
|
|
|
if initial_sync:
|
|
|
|
total, support_batches = await self.distribute_unspent_txos(TXO_TYPES['support'])
|
|
|
|
elif blocks:
|
|
|
|
p.start(2)
|
|
|
|
# 1. supports to be inserted
|
|
|
|
total = await self.count_unspent_txos(
|
|
|
|
TXO_TYPES['support'], blocks, missing_in_supports_table=True
|
|
|
|
)
|
|
|
|
support_batches = [blocks] if total else []
|
|
|
|
p.step()
|
|
|
|
# 2. supports to be deleted
|
|
|
|
delete_supports = await self.count_abandoned_supports()
|
|
|
|
total += delete_supports
|
|
|
|
p.step()
|
|
|
|
else:
|
|
|
|
return
|
2020-07-13 06:55:30 +02:00
|
|
|
with Progress(self.db.message_queue, SUPPORTS_MAIN_EVENT) as p:
|
2020-07-12 00:18:33 +02:00
|
|
|
p.start(total)
|
|
|
|
if support_batches:
|
|
|
|
await self.run_tasks([
|
|
|
|
self.db.run(
|
2020-08-03 18:24:13 +02:00
|
|
|
support_phase.supports_insert, batch, not initial_sync, self.SUPPORT_FLUSH_SIZE
|
2020-07-12 00:18:33 +02:00
|
|
|
) for batch in support_batches
|
|
|
|
])
|
|
|
|
if delete_supports:
|
|
|
|
await self.db.run(support_phase.supports_delete, delete_supports)
|
2020-07-13 06:55:30 +02:00
|
|
|
if initial_sync:
|
|
|
|
await self.db.run(support_phase.supports_constraints_and_indexes)
|
2020-07-14 19:26:32 +02:00
|
|
|
else:
|
|
|
|
await self.db.run(support_phase.supports_vacuum)
|
2020-07-12 00:18:33 +02:00
|
|
|
|
2020-07-13 15:30:32 +02:00
|
|
|
async def sync_channel_stats(self, blocks, initial_sync):
|
|
|
|
await self.db.run(claim_phase.update_channel_stats, blocks, initial_sync)
|
2020-07-12 00:18:33 +02:00
|
|
|
|
|
|
|
async def sync_trends(self):
|
|
|
|
pass
|
|
|
|
|
|
|
|
async def advance(self):
|
|
|
|
blocks_added = await self.sync_blocks()
|
2020-07-13 06:55:30 +02:00
|
|
|
await self.sync_spends(blocks_added)
|
2020-09-17 17:25:50 +02:00
|
|
|
await self.sync_filters()
|
2020-07-13 15:30:32 +02:00
|
|
|
initial_claim_sync = await self.sync_claims(blocks_added)
|
2020-07-12 00:18:33 +02:00
|
|
|
await self.sync_supports(blocks_added)
|
2020-07-13 15:30:32 +02:00
|
|
|
await self.sync_channel_stats(blocks_added, initial_claim_sync)
|
2020-09-17 17:25:50 +02:00
|
|
|
await self.sync_trends()
|
2020-07-06 05:03:45 +02:00
|
|
|
if blocks_added:
|
|
|
|
await self._on_block_controller.add(BlockEvent(blocks_added[-1]))
|
2020-06-26 16:39:58 +02:00
|
|
|
|
2020-09-16 16:37:49 +02:00
|
|
|
async def sync_mempool(self):
|
|
|
|
await self.db.run(block_phase.sync_mempool)
|
|
|
|
await self.sync_spends([-1])
|
|
|
|
await self.db.run(claim_phase.claims_insert, [-2, 0], True, self.CLAIM_FLUSH_SIZE)
|
|
|
|
await self.db.run(claim_phase.claims_vacuum)
|
|
|
|
|
2020-06-26 16:39:58 +02:00
|
|
|
async def advance_loop(self):
|
|
|
|
while True:
|
|
|
|
try:
|
2020-09-16 16:37:49 +02:00
|
|
|
await asyncio.wait([
|
|
|
|
self.tx_hash_event.wait(),
|
|
|
|
self.block_hash_event.wait(),
|
|
|
|
], return_when=asyncio.FIRST_COMPLETED)
|
|
|
|
if self.block_hash_event.is_set():
|
|
|
|
self.block_hash_event.clear()
|
|
|
|
await self.db.run(block_phase.clear_mempool)
|
|
|
|
await self.advance()
|
|
|
|
self.tx_hash_event.clear()
|
|
|
|
await self.sync_mempool()
|
2020-06-26 16:39:58 +02:00
|
|
|
except asyncio.CancelledError:
|
|
|
|
return
|
|
|
|
except Exception as e:
|
|
|
|
log.exception(e)
|
|
|
|
await self.stop()
|
2020-08-03 22:53:40 +02:00
|
|
|
|
|
|
|
async def rewind(self, height):
|
|
|
|
await self.db.run(block_phase.rewind, height)
|