# Copyright (c) 2016-2017, Neil Booth # # All rights reserved. # # See the file "LICENCE" for information about the copyright # and warranty status of this software. """Class for handling asynchronous connections to a blockchain daemon.""" import asyncio import itertools import json import time from calendar import timegm from struct import pack from time import strptime import aiohttp from torba.server.util import hex_to_bytes, class_logger,\ unpack_le_uint16_from, pack_varint from torba.server.hash import hex_str_to_hash, hash_to_hex_str from torba.server.tx import DeserializerDecred from torba.rpc import JSONRPC class DaemonError(Exception): """Raised when the daemon returns an error in its results.""" class WarmingUpError(Exception): """Internal - when the daemon is warming up.""" class WorkQueueFullError(Exception): """Internal - when the daemon's work queue is full.""" class Daemon: """Handles connections to a daemon at the given URL.""" WARMING_UP = -28 id_counter = itertools.count() def __init__(self, coin, url, max_workqueue=10, init_retry=0.25, max_retry=4.0): self.coin = coin self.logger = class_logger(__name__, self.__class__.__name__) self.set_url(url) # Limit concurrent RPC calls to this number. # See DEFAULT_HTTP_WORKQUEUE in bitcoind, which is typically 16 self.workqueue_semaphore = asyncio.Semaphore(value=max_workqueue) self.init_retry = init_retry self.max_retry = max_retry self._height = None self.available_rpcs = {} def set_url(self, url): """Set the URLS to the given list, and switch to the first one.""" urls = url.split(',') urls = [self.coin.sanitize_url(url) for url in urls] for n, url in enumerate(urls): status = '' if n else ' (current)' logged_url = self.logged_url(url) self.logger.info(f'daemon #{n + 1} at {logged_url}{status}') self.url_index = 0 self.urls = urls def current_url(self): """Returns the current daemon URL.""" return self.urls[self.url_index] def logged_url(self, url=None): """The host and port part, for logging.""" url = url or self.current_url() return url[url.rindex('@') + 1:] def failover(self): """Call to fail-over to the next daemon URL. Returns False if there is only one, otherwise True. """ if len(self.urls) > 1: self.url_index = (self.url_index + 1) % len(self.urls) self.logger.info(f'failing over to {self.logged_url()}') return True return False def client_session(self): """An aiohttp client session.""" return aiohttp.ClientSession() async def _send_data(self, data): async with self.workqueue_semaphore: async with self.client_session() as session: async with session.post(self.current_url(), data=data) as resp: kind = resp.headers.get('Content-Type', None) if kind == 'application/json': return await resp.json() # bitcoind's HTTP protocol "handling" is a bad joke text = await resp.text() if 'Work queue depth exceeded' in text: raise WorkQueueFullError text = text.strip() or resp.reason self.logger.error(text) raise DaemonError(text) async def _send(self, payload, processor): """Send a payload to be converted to JSON. Handles temporary connection issues. Daemon reponse errors are raise through DaemonError. """ def log_error(error): nonlocal last_error_log, retry now = time.time() if now - last_error_log > 60: last_error_log = now self.logger.error(f'{error} Retrying occasionally...') if retry == self.max_retry and self.failover(): retry = 0 on_good_message = None last_error_log = 0 data = json.dumps(payload) retry = self.init_retry while True: try: result = await self._send_data(data) result = processor(result) if on_good_message: self.logger.info(on_good_message) return result except asyncio.TimeoutError: log_error('timeout error.') except aiohttp.ServerDisconnectedError: log_error('disconnected.') on_good_message = 'connection restored' except aiohttp.ClientConnectionError: log_error('connection problem - is your daemon running?') on_good_message = 'connection restored' except aiohttp.ClientError as e: log_error(f'daemon error: {e}') on_good_message = 'running normally' except WarmingUpError: log_error('starting up checking blocks.') on_good_message = 'running normally' except WorkQueueFullError: log_error('work queue full.') on_good_message = 'running normally' await asyncio.sleep(retry) retry = max(min(self.max_retry, retry * 2), self.init_retry) async def _send_single(self, method, params=None): """Send a single request to the daemon.""" def processor(result): err = result['error'] if not err: return result['result'] if err.get('code') == self.WARMING_UP: raise WarmingUpError raise DaemonError(err) payload = {'method': method, 'id': next(self.id_counter)} if params: payload['params'] = params return await self._send(payload, processor) async def _send_vector(self, method, params_iterable, replace_errs=False): """Send several requests of the same method. The result will be an array of the same length as params_iterable. If replace_errs is true, any item with an error is returned as None, otherwise an exception is raised.""" def processor(result): errs = [item['error'] for item in result if item['error']] if any(err.get('code') == self.WARMING_UP for err in errs): raise WarmingUpError if not errs or replace_errs: return [item['result'] for item in result] raise DaemonError(errs) payload = [{'method': method, 'params': p, 'id': next(self.id_counter)} for p in params_iterable] if payload: return await self._send(payload, processor) return [] async def _is_rpc_available(self, method): """Return whether given RPC method is available in the daemon. Results are cached and the daemon will generally not be queried with the same method more than once.""" available = self.available_rpcs.get(method) if available is None: available = True try: await self._send_single(method) except DaemonError as e: err = e.args[0] error_code = err.get("code") available = error_code != JSONRPC.METHOD_NOT_FOUND self.available_rpcs[method] = available return available async def block_hex_hashes(self, first, count): """Return the hex hashes of count block starting at height first.""" params_iterable = ((h, ) for h in range(first, first + count)) return await self._send_vector('getblockhash', params_iterable) async def deserialised_block(self, hex_hash): """Return the deserialised block with the given hex hash.""" return await self._send_single('getblock', (hex_hash, True)) async def raw_blocks(self, hex_hashes): """Return the raw binary blocks with the given hex hashes.""" params_iterable = ((h, False) for h in hex_hashes) blocks = await self._send_vector('getblock', params_iterable) # Convert hex string to bytes return [hex_to_bytes(block) for block in blocks] async def mempool_hashes(self): """Update our record of the daemon's mempool hashes.""" return await self._send_single('getrawmempool') async def estimatefee(self, block_count): """Return the fee estimate for the block count. Units are whole currency units per KB, e.g. 0.00000995, or -1 if no estimate is available. """ args = (block_count, ) if await self._is_rpc_available('estimatesmartfee'): estimate = await self._send_single('estimatesmartfee', args) return estimate.get('feerate', -1) return await self._send_single('estimatefee', args) async def getnetworkinfo(self): """Return the result of the 'getnetworkinfo' RPC call.""" return await self._send_single('getnetworkinfo') async def relayfee(self): """The minimum fee a low-priority tx must pay in order to be accepted to the daemon's memory pool.""" network_info = await self.getnetworkinfo() return network_info['relayfee'] async def getrawtransaction(self, hex_hash, verbose=False): """Return the serialized raw transaction with the given hash.""" # Cast to int because some coin daemons are old and require it return await self._send_single('getrawtransaction', (hex_hash, int(verbose))) async def getrawtransactions(self, hex_hashes, replace_errs=True): """Return the serialized raw transactions with the given hashes. Replaces errors with None by default.""" params_iterable = ((hex_hash, 0) for hex_hash in hex_hashes) txs = await self._send_vector('getrawtransaction', params_iterable, replace_errs=replace_errs) # Convert hex strings to bytes return [hex_to_bytes(tx) if tx else None for tx in txs] async def broadcast_transaction(self, raw_tx): """Broadcast a transaction to the network.""" return await self._send_single('sendrawtransaction', (raw_tx, )) async def height(self): """Query the daemon for its current height.""" self._height = await self._send_single('getblockcount') return self._height def cached_height(self): """Return the cached daemon height. If the daemon has not been queried yet this returns None.""" return self._height class DashDaemon(Daemon): async def masternode_broadcast(self, params): """Broadcast a transaction to the network.""" return await self._send_single('masternodebroadcast', params) async def masternode_list(self, params): """Return the masternode status.""" return await self._send_single('masternodelist', params) class FakeEstimateFeeDaemon(Daemon): """Daemon that simulates estimatefee and relayfee RPC calls. Coin that wants to use this daemon must define ESTIMATE_FEE & RELAY_FEE""" async def estimatefee(self, block_count): """Return the fee estimate for the given parameters.""" return self.coin.ESTIMATE_FEE async def relayfee(self): """The minimum fee a low-priority tx must pay in order to be accepted to the daemon's memory pool.""" return self.coin.RELAY_FEE class LegacyRPCDaemon(Daemon): """Handles connections to a daemon at the given URL. This class is useful for daemons that don't have the new 'getblock' RPC call that returns the block in hex, the workaround is to manually recreate the block bytes. The recreated block bytes may not be the exact as in the underlying blockchain but it is good enough for our indexing purposes.""" async def raw_blocks(self, hex_hashes): """Return the raw binary blocks with the given hex hashes.""" params_iterable = ((h, ) for h in hex_hashes) block_info = await self._send_vector('getblock', params_iterable) blocks = [] for i in block_info: raw_block = await self.make_raw_block(i) blocks.append(raw_block) # Convert hex string to bytes return blocks async def make_raw_header(self, b): pbh = b.get('previousblockhash') if pbh is None: pbh = '0' * 64 return b''.join([ pack(' 0: transactions = await self.getrawtransactions(b.get('tx'), False) raw_block = header num_txs = len(transactions) if num_txs > 0: raw_block += pack_varint(num_txs) raw_block += b''.join(transactions) else: raw_block += b'\x00' return raw_block def timestamp_safe(self, t): if isinstance(t, int): return t return timegm(strptime(t, "%Y-%m-%d %H:%M:%S %Z")) class DecredDaemon(Daemon): async def raw_blocks(self, hex_hashes): """Return the raw binary blocks with the given hex hashes.""" params_iterable = ((h, False) for h in hex_hashes) blocks = await self._send_vector('getblock', params_iterable) raw_blocks = [] valid_tx_tree = {} for block in blocks: # Convert to bytes from hex raw_block = hex_to_bytes(block) raw_blocks.append(raw_block) # Check if previous block is valid prev = self.prev_hex_hash(raw_block) votebits = unpack_le_uint16_from(raw_block[100:102])[0] valid_tx_tree[prev] = self.is_valid_tx_tree(votebits) processed_raw_blocks = [] for hash, raw_block in zip(hex_hashes, raw_blocks): if hash in valid_tx_tree: is_valid = valid_tx_tree[hash] else: # Do something complicated to figure out if this block is valid header = await self._send_single('getblockheader', (hash, )) if 'nextblockhash' not in header: raise DaemonError(f'Could not find next block for {hash}') next_hash = header['nextblockhash'] next_header = await self._send_single('getblockheader', (next_hash, )) is_valid = self.is_valid_tx_tree(next_header['votebits']) if is_valid: processed_raw_blocks.append(raw_block) else: # If this block is invalid remove the normal transactions self.logger.info(f'block {hash} is invalidated') processed_raw_blocks.append(self.strip_tx_tree(raw_block)) return processed_raw_blocks @staticmethod def prev_hex_hash(raw_block): return hash_to_hex_str(raw_block[4:36]) @staticmethod def is_valid_tx_tree(votebits): # Check if previous block was invalidated. return bool(votebits & (1 << 0) != 0) def strip_tx_tree(self, raw_block): c = self.coin assert issubclass(c.DESERIALIZER, DeserializerDecred) d = c.DESERIALIZER(raw_block, start=c.BASIC_HEADER_SIZE) d.read_tx_tree() # Skip normal transactions # Create a fake block without any normal transactions return raw_block[:c.BASIC_HEADER_SIZE] + b'\x00' + raw_block[d.cursor:] async def height(self): height = await super().height() if height > 0: # Lie about the daemon height as the current tip can be invalidated height -= 1 self._height = height return height async def mempool_hashes(self): mempool = await super().mempool_hashes() # Add current tip transactions to the 'fake' mempool. real_height = await self._send_single('getblockcount') tip_hash = await self._send_single('getblockhash', (real_height,)) tip = await self.deserialised_block(tip_hash) # Add normal transactions except coinbase mempool += tip['tx'][1:] # Add stake transactions if applicable mempool += tip.get('stx', []) return mempool def client_session(self): # FIXME allow self signed certificates connector = aiohttp.TCPConnector(verify_ssl=False) return aiohttp.ClientSession(connector=connector) class PreLegacyRPCDaemon(LegacyRPCDaemon): """Handles connections to a daemon at the given URL. This class is useful for daemons that don't have the new 'getblock' RPC call that returns the block in hex, and need the False parameter for the getblock""" async def deserialised_block(self, hex_hash): """Return the deserialised block with the given hex hash.""" return await self._send_single('getblock', (hex_hash, False))