return whole batches
This commit is contained in:
parent
b2027cfd66
commit
270da80d64
1 changed files with 16 additions and 14 deletions
|
@ -605,6 +605,7 @@ class Ledger(metaclass=LedgerRegistry):
|
||||||
header = await self.headers.get(remote_height)
|
header = await self.headers.get(remote_height)
|
||||||
tx.position = merkle['pos']
|
tx.position = merkle['pos']
|
||||||
tx.is_verified = merkle_root == header['merkle_root']
|
tx.is_verified = merkle_root == header['merkle_root']
|
||||||
|
return tx
|
||||||
|
|
||||||
async def request_transactions(self, to_request: Tuple[Tuple[str, int], ...], cached=False):
|
async def request_transactions(self, to_request: Tuple[Tuple[str, int], ...], cached=False):
|
||||||
batches = [[]]
|
batches = [[]]
|
||||||
|
@ -614,7 +615,8 @@ class Ledger(metaclass=LedgerRegistry):
|
||||||
if cached:
|
if cached:
|
||||||
if txid in self._tx_cache:
|
if txid in self._tx_cache:
|
||||||
if self._tx_cache[txid].tx is not None and self._tx_cache[txid].tx.is_verified:
|
if self._tx_cache[txid].tx is not None and self._tx_cache[txid].tx.is_verified:
|
||||||
yield self._tx_cache[txid].tx
|
tx = self._tx_cache[txid].tx
|
||||||
|
yield {tx.id: self._tx_cache[txid].tx}
|
||||||
continue
|
continue
|
||||||
else:
|
else:
|
||||||
self._tx_cache[txid] = TransactionCacheItem()
|
self._tx_cache[txid] = TransactionCacheItem()
|
||||||
|
@ -626,29 +628,29 @@ class Ledger(metaclass=LedgerRegistry):
|
||||||
batches.pop()
|
batches.pop()
|
||||||
|
|
||||||
for batch in batches:
|
for batch in batches:
|
||||||
async for tx in self._single_batch(batch, remote_heights):
|
txs = await self._single_batch(batch, remote_heights)
|
||||||
if cached:
|
if cached:
|
||||||
self._tx_cache[tx.id].tx = tx
|
for txid, tx in txs.items():
|
||||||
yield tx
|
self._tx_cache[txid].tx = tx
|
||||||
|
yield txs
|
||||||
|
|
||||||
async def request_synced_transactions(self, to_request, remote_history, address):
|
async def request_synced_transactions(self, to_request, remote_history, address):
|
||||||
pending_sync = {}
|
async for txs in self.request_transactions(((txid, height) for txid, height in to_request.values())):
|
||||||
async for tx in self.request_transactions(((txid, height) for txid, height in to_request.values())):
|
for tx in txs.values():
|
||||||
pending_sync[tx.id] = tx
|
yield tx
|
||||||
yield tx
|
await self._sync_and_save_batch(address, remote_history, txs)
|
||||||
if len(pending_sync) > 10:
|
|
||||||
await self._sync_and_save_batch(address, remote_history, pending_sync)
|
|
||||||
await self._sync_and_save_batch(address, remote_history, pending_sync)
|
|
||||||
|
|
||||||
async def _single_batch(self, batch, remote_heights):
|
async def _single_batch(self, batch, remote_heights):
|
||||||
heights = {remote_heights[txid] for txid in batch}
|
heights = {remote_heights[txid] for txid in batch}
|
||||||
unrestriced = 0 < min(heights) < max(heights) < max(self.headers.checkpoints or [0])
|
unrestriced = 0 < min(heights) < max(heights) < max(self.headers.checkpoints or [0])
|
||||||
batch_result = await self.network.retriable_call(self.network.get_transaction_batch, batch, not unrestriced)
|
batch_result = await self.network.retriable_call(self.network.get_transaction_batch, batch, not unrestriced)
|
||||||
|
txs = {}
|
||||||
for txid, (raw, merkle) in batch_result.items():
|
for txid, (raw, merkle) in batch_result.items():
|
||||||
remote_height = remote_heights[txid]
|
remote_height = remote_heights[txid]
|
||||||
tx = Transaction(unhexlify(raw), height=remote_height)
|
tx = Transaction(unhexlify(raw), height=remote_height)
|
||||||
|
txs[tx.id] = tx
|
||||||
await self.maybe_verify_transaction(tx, remote_height, merkle)
|
await self.maybe_verify_transaction(tx, remote_height, merkle)
|
||||||
yield tx
|
return txs
|
||||||
|
|
||||||
async def _sync_and_save_batch(self, address, remote_history, pending_txs):
|
async def _sync_and_save_batch(self, address, remote_history, pending_txs):
|
||||||
await asyncio.gather(*(self._sync(tx, remote_history, pending_txs) for tx in pending_txs.values()))
|
await asyncio.gather(*(self._sync(tx, remote_history, pending_txs) for tx in pending_txs.values()))
|
||||||
|
@ -757,7 +759,7 @@ class Ledger(metaclass=LedgerRegistry):
|
||||||
txs: List[Transaction] = []
|
txs: List[Transaction] = []
|
||||||
if len(outputs.txs) > 0:
|
if len(outputs.txs) > 0:
|
||||||
async for tx in self.request_transactions(tuple(outputs.txs), cached=True):
|
async for tx in self.request_transactions(tuple(outputs.txs), cached=True):
|
||||||
txs.append(tx)
|
txs.extend(tx.values())
|
||||||
|
|
||||||
_txos, blocked = outputs.inflate(txs)
|
_txos, blocked = outputs.inflate(txs)
|
||||||
|
|
||||||
|
|
Loading…
Add table
Reference in a new issue