consolidate leveldb block advance/reorg
-move methods from History to LevelDB
This commit is contained in:
parent
3ed748f2fd
commit
bc7fe680c0
3 changed files with 403 additions and 298 deletions
|
@ -384,7 +384,7 @@ class BlockProcessor:
|
||||||
one_MB = 1000*1000
|
one_MB = 1000*1000
|
||||||
utxo_cache_size = len(self.utxo_cache) * 205
|
utxo_cache_size = len(self.utxo_cache) * 205
|
||||||
db_deletes_size = len(self.db_deletes) * 57
|
db_deletes_size = len(self.db_deletes) * 57
|
||||||
hist_cache_size = self.db.history.unflushed_memsize()
|
hist_cache_size = len(self.db.history.unflushed) * 180 + self.db.history.unflushed_count * 4
|
||||||
# Roughly ntxs * 32 + nblocks * 42
|
# Roughly ntxs * 32 + nblocks * 42
|
||||||
tx_hash_size = ((self.tx_count - self.db.fs_tx_count) * 32
|
tx_hash_size = ((self.tx_count - self.db.fs_tx_count) * 32
|
||||||
+ (self.height - self.db.fs_height) * 42)
|
+ (self.height - self.db.fs_height) * 42)
|
||||||
|
@ -475,7 +475,16 @@ class BlockProcessor:
|
||||||
self.db.total_transactions.append(tx_hash)
|
self.db.total_transactions.append(tx_hash)
|
||||||
tx_num += 1
|
tx_num += 1
|
||||||
|
|
||||||
self.db.history.add_unflushed(hashXs_by_tx, self.tx_count)
|
# self.db.add_unflushed(hashXs_by_tx, self.tx_count)
|
||||||
|
first_tx_num = self.tx_count
|
||||||
|
_unflushed = self.db.history.unflushed
|
||||||
|
_count = 0
|
||||||
|
for _tx_num, _hashXs in enumerate(hashXs_by_tx, start=first_tx_num):
|
||||||
|
for _hashX in set(_hashXs):
|
||||||
|
_unflushed[_hashX].append(_tx_num)
|
||||||
|
_count += len(_hashXs)
|
||||||
|
self.db.history.unflushed_count += _count
|
||||||
|
|
||||||
self.tx_count = tx_num
|
self.tx_count = tx_num
|
||||||
self.db.tx_counts.append(tx_num)
|
self.db.tx_counts.append(tx_num)
|
||||||
|
|
||||||
|
|
|
@ -34,150 +34,122 @@ class History:
|
||||||
self.max_hist_row_entries = 12500
|
self.max_hist_row_entries = 12500
|
||||||
self.unflushed = defaultdict(partial(array.array, 'I'))
|
self.unflushed = defaultdict(partial(array.array, 'I'))
|
||||||
self.unflushed_count = 0
|
self.unflushed_count = 0
|
||||||
self.db = None
|
self.flush_count = 0
|
||||||
|
self.comp_flush_count = -1
|
||||||
|
self.comp_cursor = -1
|
||||||
|
# self.db = None
|
||||||
|
|
||||||
def open_db(self, db, for_sync, utxo_flush_count, compacting):
|
# def close_db(self):
|
||||||
self.db = db #db_class('hist', for_sync)
|
# if self.db:
|
||||||
self.read_state()
|
# # self.db.close()
|
||||||
self.clear_excess(utxo_flush_count)
|
# self.db = None
|
||||||
# An incomplete compaction needs to be cancelled otherwise
|
|
||||||
# restarting it will corrupt the history
|
|
||||||
if not compacting:
|
|
||||||
self._cancel_compaction()
|
|
||||||
return self.flush_count
|
|
||||||
|
|
||||||
def close_db(self):
|
# def read_state(self):
|
||||||
if self.db:
|
# state = self.db.get(HIST_STATE)
|
||||||
# self.db.close()
|
# if state:
|
||||||
self.db = None
|
# state = ast.literal_eval(state.decode())
|
||||||
|
# if not isinstance(state, dict):
|
||||||
|
# raise RuntimeError('failed reading state from history DB')
|
||||||
|
# self.flush_count = state['flush_count']
|
||||||
|
# self.comp_flush_count = state.get('comp_flush_count', -1)
|
||||||
|
# self.comp_cursor = state.get('comp_cursor', -1)
|
||||||
|
# self.db_version = state.get('db_version', 0)
|
||||||
|
# else:
|
||||||
|
# self.flush_count = 0
|
||||||
|
# self.comp_flush_count = -1
|
||||||
|
# self.comp_cursor = -1
|
||||||
|
# self.db_version = max(self.DB_VERSIONS)
|
||||||
|
#
|
||||||
|
# self.logger.info(f'history DB version: {self.db_version}')
|
||||||
|
# if self.db_version not in self.DB_VERSIONS:
|
||||||
|
# msg = f'this software only handles DB versions {self.DB_VERSIONS}'
|
||||||
|
# self.logger.error(msg)
|
||||||
|
# raise RuntimeError(msg)
|
||||||
|
# self.logger.info(f'flush count: {self.flush_count:,d}')
|
||||||
|
|
||||||
def read_state(self):
|
# def clear_excess(self, utxo_flush_count):
|
||||||
state = self.db.get(HIST_STATE)
|
# # < might happen at end of compaction as both DBs cannot be
|
||||||
if state:
|
# # updated atomically
|
||||||
state = ast.literal_eval(state.decode())
|
# if self.flush_count <= utxo_flush_count:
|
||||||
if not isinstance(state, dict):
|
# return
|
||||||
raise RuntimeError('failed reading state from history DB')
|
#
|
||||||
self.flush_count = state['flush_count']
|
# self.logger.info('DB shut down uncleanly. Scanning for '
|
||||||
self.comp_flush_count = state.get('comp_flush_count', -1)
|
# 'excess history flushes...')
|
||||||
self.comp_cursor = state.get('comp_cursor', -1)
|
#
|
||||||
self.db_version = state.get('db_version', 0)
|
# keys = []
|
||||||
else:
|
# for key, hist in self.db.iterator(prefix=HASHX_HISTORY_PREFIX):
|
||||||
self.flush_count = 0
|
# k = key[1:]
|
||||||
self.comp_flush_count = -1
|
# flush_id, = unpack_be_uint16_from(k[-2:])
|
||||||
self.comp_cursor = -1
|
# if flush_id > utxo_flush_count:
|
||||||
self.db_version = max(self.DB_VERSIONS)
|
# keys.append(k)
|
||||||
|
#
|
||||||
|
# self.logger.info(f'deleting {len(keys):,d} history entries')
|
||||||
|
#
|
||||||
|
# self.flush_count = utxo_flush_count
|
||||||
|
# with self.db.write_batch() as batch:
|
||||||
|
# for key in keys:
|
||||||
|
# batch.delete(HASHX_HISTORY_PREFIX + key)
|
||||||
|
# self.write_state(batch)
|
||||||
|
#
|
||||||
|
# self.logger.info('deleted excess history entries')
|
||||||
|
#
|
||||||
|
# def write_state(self, batch):
|
||||||
|
# """Write state to the history DB."""
|
||||||
|
# state = {
|
||||||
|
# 'flush_count': self.flush_count,
|
||||||
|
# 'comp_flush_count': self.comp_flush_count,
|
||||||
|
# 'comp_cursor': self.comp_cursor,
|
||||||
|
# 'db_version': self.db_version,
|
||||||
|
# }
|
||||||
|
# # History entries are not prefixed; the suffix \0\0 ensures we
|
||||||
|
# # look similar to other entries and aren't interfered with
|
||||||
|
# batch.put(HIST_STATE, repr(state).encode())
|
||||||
|
|
||||||
self.logger.info(f'history DB version: {self.db_version}')
|
# def add_unflushed(self, hashXs_by_tx, first_tx_num):
|
||||||
if self.db_version not in self.DB_VERSIONS:
|
# unflushed = self.unflushed
|
||||||
msg = f'this software only handles DB versions {self.DB_VERSIONS}'
|
# count = 0
|
||||||
self.logger.error(msg)
|
# for tx_num, hashXs in enumerate(hashXs_by_tx, start=first_tx_num):
|
||||||
raise RuntimeError(msg)
|
# hashXs = set(hashXs)
|
||||||
self.logger.info(f'flush count: {self.flush_count:,d}')
|
# for hashX in hashXs:
|
||||||
|
# unflushed[hashX].append(tx_num)
|
||||||
|
# count += len(hashXs)
|
||||||
|
# self.unflushed_count += count
|
||||||
|
|
||||||
def clear_excess(self, utxo_flush_count):
|
# def unflushed_memsize(self):
|
||||||
# < might happen at end of compaction as both DBs cannot be
|
# return len(self.unflushed) * 180 + self.unflushed_count * 4
|
||||||
# updated atomically
|
|
||||||
if self.flush_count <= utxo_flush_count:
|
|
||||||
return
|
|
||||||
|
|
||||||
self.logger.info('DB shut down uncleanly. Scanning for '
|
# def assert_flushed(self):
|
||||||
'excess history flushes...')
|
# assert not self.unflushed
|
||||||
|
|
||||||
keys = []
|
# def backup(self, hashXs, tx_count):
|
||||||
for key, hist in self.db.iterator(prefix=HASHX_HISTORY_PREFIX):
|
# # Not certain this is needed, but it doesn't hurt
|
||||||
k = key[1:]
|
# self.flush_count += 1
|
||||||
flush_id, = unpack_be_uint16_from(k[-2:])
|
# nremoves = 0
|
||||||
if flush_id > utxo_flush_count:
|
# bisect_left = bisect.bisect_left
|
||||||
keys.append(k)
|
#
|
||||||
|
# with self.db.write_batch() as batch:
|
||||||
self.logger.info(f'deleting {len(keys):,d} history entries')
|
# for hashX in sorted(hashXs):
|
||||||
|
# deletes = []
|
||||||
self.flush_count = utxo_flush_count
|
# puts = {}
|
||||||
with self.db.write_batch() as batch:
|
# for key, hist in self.db.iterator(prefix=HASHX_HISTORY_PREFIX + hashX, reverse=True):
|
||||||
for key in keys:
|
# k = key[1:]
|
||||||
batch.delete(HASHX_HISTORY_PREFIX + key)
|
# a = array.array('I')
|
||||||
self.write_state(batch)
|
# a.frombytes(hist)
|
||||||
|
# # Remove all history entries >= tx_count
|
||||||
self.logger.info('deleted excess history entries')
|
# idx = bisect_left(a, tx_count)
|
||||||
|
# nremoves += len(a) - idx
|
||||||
def write_state(self, batch):
|
# if idx > 0:
|
||||||
"""Write state to the history DB."""
|
# puts[k] = a[:idx].tobytes()
|
||||||
state = {
|
# break
|
||||||
'flush_count': self.flush_count,
|
# deletes.append(k)
|
||||||
'comp_flush_count': self.comp_flush_count,
|
#
|
||||||
'comp_cursor': self.comp_cursor,
|
# for key in deletes:
|
||||||
'db_version': self.db_version,
|
# batch.delete(key)
|
||||||
}
|
# for key, value in puts.items():
|
||||||
# History entries are not prefixed; the suffix \0\0 ensures we
|
# batch.put(key, value)
|
||||||
# look similar to other entries and aren't interfered with
|
# self.write_state(batch)
|
||||||
batch.put(HIST_STATE, repr(state).encode())
|
#
|
||||||
|
# self.logger.info(f'backing up removed {nremoves:,d} history entries')
|
||||||
def add_unflushed(self, hashXs_by_tx, first_tx_num):
|
|
||||||
unflushed = self.unflushed
|
|
||||||
count = 0
|
|
||||||
for tx_num, hashXs in enumerate(hashXs_by_tx, start=first_tx_num):
|
|
||||||
hashXs = set(hashXs)
|
|
||||||
for hashX in hashXs:
|
|
||||||
unflushed[hashX].append(tx_num)
|
|
||||||
count += len(hashXs)
|
|
||||||
self.unflushed_count += count
|
|
||||||
|
|
||||||
def unflushed_memsize(self):
|
|
||||||
return len(self.unflushed) * 180 + self.unflushed_count * 4
|
|
||||||
|
|
||||||
def assert_flushed(self):
|
|
||||||
assert not self.unflushed
|
|
||||||
|
|
||||||
def flush(self):
|
|
||||||
start_time = time.time()
|
|
||||||
self.flush_count += 1
|
|
||||||
flush_id = pack_be_uint16(self.flush_count)
|
|
||||||
unflushed = self.unflushed
|
|
||||||
|
|
||||||
with self.db.write_batch() as batch:
|
|
||||||
for hashX in sorted(unflushed):
|
|
||||||
key = hashX + flush_id
|
|
||||||
batch.put(HASHX_HISTORY_PREFIX + key, unflushed[hashX].tobytes())
|
|
||||||
self.write_state(batch)
|
|
||||||
|
|
||||||
count = len(unflushed)
|
|
||||||
unflushed.clear()
|
|
||||||
self.unflushed_count = 0
|
|
||||||
|
|
||||||
if self.db.for_sync:
|
|
||||||
elapsed = time.time() - start_time
|
|
||||||
self.logger.info(f'flushed history in {elapsed:.1f}s '
|
|
||||||
f'for {count:,d} addrs')
|
|
||||||
|
|
||||||
def backup(self, hashXs, tx_count):
|
|
||||||
# Not certain this is needed, but it doesn't hurt
|
|
||||||
self.flush_count += 1
|
|
||||||
nremoves = 0
|
|
||||||
bisect_left = bisect.bisect_left
|
|
||||||
|
|
||||||
with self.db.write_batch() as batch:
|
|
||||||
for hashX in sorted(hashXs):
|
|
||||||
deletes = []
|
|
||||||
puts = {}
|
|
||||||
for key, hist in self.db.iterator(prefix=HASHX_HISTORY_PREFIX + hashX, reverse=True):
|
|
||||||
k = key[1:]
|
|
||||||
a = array.array('I')
|
|
||||||
a.frombytes(hist)
|
|
||||||
# Remove all history entries >= tx_count
|
|
||||||
idx = bisect_left(a, tx_count)
|
|
||||||
nremoves += len(a) - idx
|
|
||||||
if idx > 0:
|
|
||||||
puts[k] = a[:idx].tobytes()
|
|
||||||
break
|
|
||||||
deletes.append(k)
|
|
||||||
|
|
||||||
for key in deletes:
|
|
||||||
batch.delete(key)
|
|
||||||
for key, value in puts.items():
|
|
||||||
batch.put(key, value)
|
|
||||||
self.write_state(batch)
|
|
||||||
|
|
||||||
self.logger.info(f'backing up removed {nremoves:,d} history entries')
|
|
||||||
|
|
||||||
# def get_txnums(self, hashX, limit=1000):
|
# def get_txnums(self, hashX, limit=1000):
|
||||||
# """Generator that returns an unpruned, sorted list of tx_nums in the
|
# """Generator that returns an unpruned, sorted list of tx_nums in the
|
||||||
|
@ -213,119 +185,120 @@ class History:
|
||||||
# When compaction is complete and the final flush takes place,
|
# When compaction is complete and the final flush takes place,
|
||||||
# flush_count is reset to comp_flush_count, and comp_flush_count to -1
|
# flush_count is reset to comp_flush_count, and comp_flush_count to -1
|
||||||
|
|
||||||
def _flush_compaction(self, cursor, write_items, keys_to_delete):
|
# def _flush_compaction(self, cursor, write_items, keys_to_delete):
|
||||||
"""Flush a single compaction pass as a batch."""
|
# """Flush a single compaction pass as a batch."""
|
||||||
# Update compaction state
|
# # Update compaction state
|
||||||
if cursor == 65536:
|
# if cursor == 65536:
|
||||||
self.flush_count = self.comp_flush_count
|
# self.flush_count = self.comp_flush_count
|
||||||
self.comp_cursor = -1
|
# self.comp_cursor = -1
|
||||||
self.comp_flush_count = -1
|
# self.comp_flush_count = -1
|
||||||
else:
|
# else:
|
||||||
self.comp_cursor = cursor
|
# self.comp_cursor = cursor
|
||||||
|
#
|
||||||
|
# # History DB. Flush compacted history and updated state
|
||||||
|
# with self.db.write_batch() as batch:
|
||||||
|
# # Important: delete first! The keyspace may overlap.
|
||||||
|
# for key in keys_to_delete:
|
||||||
|
# batch.delete(HASHX_HISTORY_PREFIX + key)
|
||||||
|
# for key, value in write_items:
|
||||||
|
# batch.put(HASHX_HISTORY_PREFIX + key, value)
|
||||||
|
# self.write_state(batch)
|
||||||
|
|
||||||
# History DB. Flush compacted history and updated state
|
# def _compact_hashX(self, hashX, hist_map, hist_list,
|
||||||
with self.db.write_batch() as batch:
|
# write_items, keys_to_delete):
|
||||||
# Important: delete first! The keyspace may overlap.
|
# """Compress history for a hashX. hist_list is an ordered list of
|
||||||
for key in keys_to_delete:
|
# the histories to be compressed."""
|
||||||
batch.delete(HASHX_HISTORY_PREFIX + key)
|
# # History entries (tx numbers) are 4 bytes each. Distribute
|
||||||
for key, value in write_items:
|
# # over rows of up to 50KB in size. A fixed row size means
|
||||||
batch.put(HASHX_HISTORY_PREFIX + key, value)
|
# # future compactions will not need to update the first N - 1
|
||||||
self.write_state(batch)
|
# # rows.
|
||||||
|
# max_row_size = self.max_hist_row_entries * 4
|
||||||
|
# full_hist = b''.join(hist_list)
|
||||||
|
# nrows = (len(full_hist) + max_row_size - 1) // max_row_size
|
||||||
|
# if nrows > 4:
|
||||||
|
# self.logger.info('hashX {} is large: {:,d} entries across '
|
||||||
|
# '{:,d} rows'
|
||||||
|
# .format(hash_to_hex_str(hashX),
|
||||||
|
# len(full_hist) // 4, nrows))
|
||||||
|
#
|
||||||
|
# # Find what history needs to be written, and what keys need to
|
||||||
|
# # be deleted. Start by assuming all keys are to be deleted,
|
||||||
|
# # and then remove those that are the same on-disk as when
|
||||||
|
# # compacted.
|
||||||
|
# write_size = 0
|
||||||
|
# keys_to_delete.update(hist_map)
|
||||||
|
# for n, chunk in enumerate(util.chunks(full_hist, max_row_size)):
|
||||||
|
# key = hashX + pack_be_uint16(n)
|
||||||
|
# if hist_map.get(key) == chunk:
|
||||||
|
# keys_to_delete.remove(key)
|
||||||
|
# else:
|
||||||
|
# write_items.append((key, chunk))
|
||||||
|
# write_size += len(chunk)
|
||||||
|
#
|
||||||
|
# assert n + 1 == nrows
|
||||||
|
# self.comp_flush_count = max(self.comp_flush_count, n)
|
||||||
|
#
|
||||||
|
# return write_size
|
||||||
|
|
||||||
def _compact_hashX(self, hashX, hist_map, hist_list,
|
# def _compact_prefix(self, prefix, write_items, keys_to_delete):
|
||||||
write_items, keys_to_delete):
|
# """Compact all history entries for hashXs beginning with the
|
||||||
"""Compress history for a hashX. hist_list is an ordered list of
|
# given prefix. Update keys_to_delete and write."""
|
||||||
the histories to be compressed."""
|
# prior_hashX = None
|
||||||
# History entries (tx numbers) are 4 bytes each. Distribute
|
# hist_map = {}
|
||||||
# over rows of up to 50KB in size. A fixed row size means
|
# hist_list = []
|
||||||
# future compactions will not need to update the first N - 1
|
#
|
||||||
# rows.
|
# key_len = HASHX_LEN + 2
|
||||||
max_row_size = self.max_hist_row_entries * 4
|
# write_size = 0
|
||||||
full_hist = b''.join(hist_list)
|
# for key, hist in self.db.iterator(prefix=HASHX_HISTORY_PREFIX + prefix):
|
||||||
nrows = (len(full_hist) + max_row_size - 1) // max_row_size
|
# k = key[1:]
|
||||||
if nrows > 4:
|
# # Ignore non-history entries
|
||||||
self.logger.info('hashX {} is large: {:,d} entries across '
|
# if len(k) != key_len:
|
||||||
'{:,d} rows'
|
# continue
|
||||||
.format(hash_to_hex_str(hashX),
|
# hashX = k[:-2]
|
||||||
len(full_hist) // 4, nrows))
|
# if hashX != prior_hashX and prior_hashX:
|
||||||
|
# write_size += self._compact_hashX(prior_hashX, hist_map,
|
||||||
|
# hist_list, write_items,
|
||||||
|
# keys_to_delete)
|
||||||
|
# hist_map.clear()
|
||||||
|
# hist_list.clear()
|
||||||
|
# prior_hashX = hashX
|
||||||
|
# hist_map[k] = hist
|
||||||
|
# hist_list.append(hist)
|
||||||
|
#
|
||||||
|
# if prior_hashX:
|
||||||
|
# write_size += self._compact_hashX(prior_hashX, hist_map, hist_list,
|
||||||
|
# write_items, keys_to_delete)
|
||||||
|
# return write_size
|
||||||
|
|
||||||
# Find what history needs to be written, and what keys need to
|
# def _compact_history(self, limit):
|
||||||
# be deleted. Start by assuming all keys are to be deleted,
|
# """Inner loop of history compaction. Loops until limit bytes have
|
||||||
# and then remove those that are the same on-disk as when
|
# been processed.
|
||||||
# compacted.
|
# """
|
||||||
write_size = 0
|
# fnord
|
||||||
keys_to_delete.update(hist_map)
|
# keys_to_delete = set()
|
||||||
for n, chunk in enumerate(util.chunks(full_hist, max_row_size)):
|
# write_items = [] # A list of (key, value) pairs
|
||||||
key = hashX + pack_be_uint16(n)
|
# write_size = 0
|
||||||
if hist_map.get(key) == chunk:
|
#
|
||||||
keys_to_delete.remove(key)
|
# # Loop over 2-byte prefixes
|
||||||
else:
|
# cursor = self.comp_cursor
|
||||||
write_items.append((key, chunk))
|
# while write_size < limit and cursor < 65536:
|
||||||
write_size += len(chunk)
|
# prefix = pack_be_uint16(cursor)
|
||||||
|
# write_size += self._compact_prefix(prefix, write_items,
|
||||||
|
# keys_to_delete)
|
||||||
|
# cursor += 1
|
||||||
|
#
|
||||||
|
# max_rows = self.comp_flush_count + 1
|
||||||
|
# self._flush_compaction(cursor, write_items, keys_to_delete)
|
||||||
|
#
|
||||||
|
# self.logger.info('history compaction: wrote {:,d} rows ({:.1f} MB), '
|
||||||
|
# 'removed {:,d} rows, largest: {:,d}, {:.1f}% complete'
|
||||||
|
# .format(len(write_items), write_size / 1000000,
|
||||||
|
# len(keys_to_delete), max_rows,
|
||||||
|
# 100 * cursor / 65536))
|
||||||
|
# return write_size
|
||||||
|
|
||||||
assert n + 1 == nrows
|
# def _cancel_compaction(self):
|
||||||
self.comp_flush_count = max(self.comp_flush_count, n)
|
# if self.comp_cursor != -1:
|
||||||
|
# self.logger.warning('cancelling in-progress history compaction')
|
||||||
return write_size
|
# self.comp_flush_count = -1
|
||||||
|
# self.comp_cursor = -1
|
||||||
def _compact_prefix(self, prefix, write_items, keys_to_delete):
|
|
||||||
"""Compact all history entries for hashXs beginning with the
|
|
||||||
given prefix. Update keys_to_delete and write."""
|
|
||||||
prior_hashX = None
|
|
||||||
hist_map = {}
|
|
||||||
hist_list = []
|
|
||||||
|
|
||||||
key_len = HASHX_LEN + 2
|
|
||||||
write_size = 0
|
|
||||||
for key, hist in self.db.iterator(prefix=HASHX_HISTORY_PREFIX + prefix):
|
|
||||||
k = key[1:]
|
|
||||||
# Ignore non-history entries
|
|
||||||
if len(k) != key_len:
|
|
||||||
continue
|
|
||||||
hashX = k[:-2]
|
|
||||||
if hashX != prior_hashX and prior_hashX:
|
|
||||||
write_size += self._compact_hashX(prior_hashX, hist_map,
|
|
||||||
hist_list, write_items,
|
|
||||||
keys_to_delete)
|
|
||||||
hist_map.clear()
|
|
||||||
hist_list.clear()
|
|
||||||
prior_hashX = hashX
|
|
||||||
hist_map[k] = hist
|
|
||||||
hist_list.append(hist)
|
|
||||||
|
|
||||||
if prior_hashX:
|
|
||||||
write_size += self._compact_hashX(prior_hashX, hist_map, hist_list,
|
|
||||||
write_items, keys_to_delete)
|
|
||||||
return write_size
|
|
||||||
|
|
||||||
def _compact_history(self, limit):
|
|
||||||
"""Inner loop of history compaction. Loops until limit bytes have
|
|
||||||
been processed.
|
|
||||||
"""
|
|
||||||
keys_to_delete = set()
|
|
||||||
write_items = [] # A list of (key, value) pairs
|
|
||||||
write_size = 0
|
|
||||||
|
|
||||||
# Loop over 2-byte prefixes
|
|
||||||
cursor = self.comp_cursor
|
|
||||||
while write_size < limit and cursor < 65536:
|
|
||||||
prefix = pack_be_uint16(cursor)
|
|
||||||
write_size += self._compact_prefix(prefix, write_items,
|
|
||||||
keys_to_delete)
|
|
||||||
cursor += 1
|
|
||||||
|
|
||||||
max_rows = self.comp_flush_count + 1
|
|
||||||
self._flush_compaction(cursor, write_items, keys_to_delete)
|
|
||||||
|
|
||||||
self.logger.info('history compaction: wrote {:,d} rows ({:.1f} MB), '
|
|
||||||
'removed {:,d} rows, largest: {:,d}, {:.1f}% complete'
|
|
||||||
.format(len(write_items), write_size / 1000000,
|
|
||||||
len(keys_to_delete), max_rows,
|
|
||||||
100 * cursor / 65536))
|
|
||||||
return write_size
|
|
||||||
|
|
||||||
def _cancel_compaction(self):
|
|
||||||
if self.comp_cursor != -1:
|
|
||||||
self.logger.warning('cancelling in-progress history compaction')
|
|
||||||
self.comp_flush_count = -1
|
|
||||||
self.comp_cursor = -1
|
|
||||||
|
|
|
@ -19,7 +19,7 @@ import zlib
|
||||||
import typing
|
import typing
|
||||||
from typing import Optional, List, Tuple, Iterable
|
from typing import Optional, List, Tuple, Iterable
|
||||||
from asyncio import sleep
|
from asyncio import sleep
|
||||||
from bisect import bisect_right
|
from bisect import bisect_right, bisect_left
|
||||||
from collections import namedtuple
|
from collections import namedtuple
|
||||||
from glob import glob
|
from glob import glob
|
||||||
from struct import pack, unpack
|
from struct import pack, unpack
|
||||||
|
@ -29,7 +29,7 @@ from lbry.utils import LRUCacheWithMetrics
|
||||||
from lbry.wallet.server import util
|
from lbry.wallet.server import util
|
||||||
from lbry.wallet.server.hash import hash_to_hex_str, HASHX_LEN
|
from lbry.wallet.server.hash import hash_to_hex_str, HASHX_LEN
|
||||||
from lbry.wallet.server.merkle import Merkle, MerkleCache
|
from lbry.wallet.server.merkle import Merkle, MerkleCache
|
||||||
from lbry.wallet.server.util import formatted_time, pack_be_uint16
|
from lbry.wallet.server.util import formatted_time, pack_be_uint16, unpack_be_uint16_from
|
||||||
from lbry.wallet.server.storage import db_class
|
from lbry.wallet.server.storage import db_class
|
||||||
from lbry.wallet.server.history import History
|
from lbry.wallet.server.history import History
|
||||||
|
|
||||||
|
@ -41,6 +41,7 @@ BLOCK_HASH_PREFIX = b'C'
|
||||||
HEADER_PREFIX = b'H'
|
HEADER_PREFIX = b'H'
|
||||||
TX_NUM_PREFIX = b'N'
|
TX_NUM_PREFIX = b'N'
|
||||||
TX_COUNT_PREFIX = b'T'
|
TX_COUNT_PREFIX = b'T'
|
||||||
|
UNDO_PREFIX = b'U'
|
||||||
TX_HASH_PREFIX = b'X'
|
TX_HASH_PREFIX = b'X'
|
||||||
|
|
||||||
HASHX_UTXO_PREFIX = b'h'
|
HASHX_UTXO_PREFIX = b'h'
|
||||||
|
@ -50,9 +51,6 @@ UTXO_PREFIX = b'u'
|
||||||
HASHX_HISTORY_PREFIX = b'x'
|
HASHX_HISTORY_PREFIX = b'x'
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
@attr.s(slots=True)
|
@attr.s(slots=True)
|
||||||
class FlushData:
|
class FlushData:
|
||||||
height = attr.ib()
|
height = attr.ib()
|
||||||
|
@ -107,6 +105,19 @@ class LevelDB:
|
||||||
self._tx_and_merkle_cache = LRUCacheWithMetrics(2 ** 17, metric_name='tx_and_merkle', namespace="wallet_server")
|
self._tx_and_merkle_cache = LRUCacheWithMetrics(2 ** 17, metric_name='tx_and_merkle', namespace="wallet_server")
|
||||||
self.total_transactions = None
|
self.total_transactions = None
|
||||||
|
|
||||||
|
# def add_unflushed(self, hashXs_by_tx, first_tx_num):
|
||||||
|
# unflushed = self.history.unflushed
|
||||||
|
# count = 0
|
||||||
|
# for tx_num, hashXs in enumerate(hashXs_by_tx, start=first_tx_num):
|
||||||
|
# hashXs = set(hashXs)
|
||||||
|
# for hashX in hashXs:
|
||||||
|
# unflushed[hashX].append(tx_num)
|
||||||
|
# count += len(hashXs)
|
||||||
|
# self.history.unflushed_count += count
|
||||||
|
|
||||||
|
# def unflushed_memsize(self):
|
||||||
|
# return len(self.history.unflushed) * 180 + self.history.unflushed_count * 4
|
||||||
|
|
||||||
async def _read_tx_counts(self):
|
async def _read_tx_counts(self):
|
||||||
if self.tx_counts is not None:
|
if self.tx_counts is not None:
|
||||||
return
|
return
|
||||||
|
@ -172,10 +183,88 @@ class LevelDB:
|
||||||
self.read_utxo_state()
|
self.read_utxo_state()
|
||||||
|
|
||||||
# Then history DB
|
# Then history DB
|
||||||
self.utxo_flush_count = self.history.open_db(
|
state = self.db.get(HIST_STATE)
|
||||||
self.db, for_sync, self.utxo_flush_count, compacting
|
if state:
|
||||||
)
|
state = ast.literal_eval(state.decode())
|
||||||
self.clear_excess_undo_info()
|
if not isinstance(state, dict):
|
||||||
|
raise RuntimeError('failed reading state from history DB')
|
||||||
|
self.history.flush_count = state['flush_count']
|
||||||
|
self.history.comp_flush_count = state.get('comp_flush_count', -1)
|
||||||
|
self.history.comp_cursor = state.get('comp_cursor', -1)
|
||||||
|
self.history.db_version = state.get('db_version', 0)
|
||||||
|
else:
|
||||||
|
self.history.flush_count = 0
|
||||||
|
self.history.comp_flush_count = -1
|
||||||
|
self.history.comp_cursor = -1
|
||||||
|
self.history.db_version = max(self.DB_VERSIONS)
|
||||||
|
|
||||||
|
self.logger.info(f'history DB version: {self.history.db_version}')
|
||||||
|
if self.history.db_version not in self.DB_VERSIONS:
|
||||||
|
msg = f'this software only handles DB versions {self.DB_VERSIONS}'
|
||||||
|
self.logger.error(msg)
|
||||||
|
raise RuntimeError(msg)
|
||||||
|
self.logger.info(f'flush count: {self.history.flush_count:,d}')
|
||||||
|
|
||||||
|
# self.history.clear_excess(self.utxo_flush_count)
|
||||||
|
# < might happen at end of compaction as both DBs cannot be
|
||||||
|
# updated atomically
|
||||||
|
if self.history.flush_count > self.utxo_flush_count:
|
||||||
|
self.logger.info('DB shut down uncleanly. Scanning for '
|
||||||
|
'excess history flushes...')
|
||||||
|
|
||||||
|
keys = []
|
||||||
|
for key, hist in self.db.iterator(prefix=HASHX_HISTORY_PREFIX):
|
||||||
|
k = key[1:]
|
||||||
|
flush_id, = unpack_be_uint16_from(k[-2:])
|
||||||
|
if flush_id > self.utxo_flush_count:
|
||||||
|
keys.append(k)
|
||||||
|
|
||||||
|
self.logger.info(f'deleting {len(keys):,d} history entries')
|
||||||
|
|
||||||
|
self.history.flush_count = self.utxo_flush_count
|
||||||
|
with self.db.write_batch() as batch:
|
||||||
|
for key in keys:
|
||||||
|
batch.delete(HASHX_HISTORY_PREFIX + key)
|
||||||
|
state = {
|
||||||
|
'flush_count': self.history.flush_count,
|
||||||
|
'comp_flush_count': self.history.comp_flush_count,
|
||||||
|
'comp_cursor': self.history.comp_cursor,
|
||||||
|
'db_version': self.history.db_version,
|
||||||
|
}
|
||||||
|
# History entries are not prefixed; the suffix \0\0 ensures we
|
||||||
|
# look similar to other entries and aren't interfered with
|
||||||
|
batch.put(HIST_STATE, repr(state).encode())
|
||||||
|
|
||||||
|
self.logger.info('deleted excess history entries')
|
||||||
|
|
||||||
|
self.utxo_flush_count = self.history.flush_count
|
||||||
|
|
||||||
|
min_height = self.min_undo_height(self.db_height)
|
||||||
|
keys = []
|
||||||
|
for key, hist in self.db.iterator(prefix=UNDO_PREFIX):
|
||||||
|
height, = unpack('>I', key[-4:])
|
||||||
|
if height >= min_height:
|
||||||
|
break
|
||||||
|
keys.append(key)
|
||||||
|
|
||||||
|
if keys:
|
||||||
|
with self.db.write_batch() as batch:
|
||||||
|
for key in keys:
|
||||||
|
batch.delete(key)
|
||||||
|
self.logger.info(f'deleted {len(keys):,d} stale undo entries')
|
||||||
|
|
||||||
|
# delete old block files
|
||||||
|
prefix = self.raw_block_prefix()
|
||||||
|
paths = [path for path in glob(f'{prefix}[0-9]*')
|
||||||
|
if len(path) > len(prefix)
|
||||||
|
and int(path[len(prefix):]) < min_height]
|
||||||
|
if paths:
|
||||||
|
for path in paths:
|
||||||
|
try:
|
||||||
|
os.remove(path)
|
||||||
|
except FileNotFoundError:
|
||||||
|
pass
|
||||||
|
self.logger.info(f'deleted {len(paths):,d} stale block files')
|
||||||
|
|
||||||
# Read TX counts (requires meta directory)
|
# Read TX counts (requires meta directory)
|
||||||
await self._read_tx_counts()
|
await self._read_tx_counts()
|
||||||
|
@ -185,7 +274,6 @@ class LevelDB:
|
||||||
|
|
||||||
def close(self):
|
def close(self):
|
||||||
self.db.close()
|
self.db.close()
|
||||||
self.history.close_db()
|
|
||||||
self.executor.shutdown(wait=True)
|
self.executor.shutdown(wait=True)
|
||||||
self.executor = None
|
self.executor = None
|
||||||
|
|
||||||
|
@ -240,7 +328,7 @@ class LevelDB:
|
||||||
assert not flush_data.adds
|
assert not flush_data.adds
|
||||||
assert not flush_data.deletes
|
assert not flush_data.deletes
|
||||||
assert not flush_data.undo_infos
|
assert not flush_data.undo_infos
|
||||||
self.history.assert_flushed()
|
assert not self.history.unflushed
|
||||||
|
|
||||||
def flush_utxo_db(self, batch, flush_data):
|
def flush_utxo_db(self, batch, flush_data):
|
||||||
"""Flush the cached DB writes and UTXO set to the batch."""
|
"""Flush the cached DB writes and UTXO set to the batch."""
|
||||||
|
@ -263,12 +351,13 @@ class LevelDB:
|
||||||
# suffix = tx_idx + tx_num
|
# suffix = tx_idx + tx_num
|
||||||
hashX = value[:-12]
|
hashX = value[:-12]
|
||||||
suffix = key[-2:] + value[-12:-8]
|
suffix = key[-2:] + value[-12:-8]
|
||||||
batch_put(b'h' + key[:4] + suffix, hashX)
|
batch_put(HASHX_UTXO_PREFIX + key[:4] + suffix, hashX)
|
||||||
batch_put(b'u' + hashX + suffix, value[-8:])
|
batch_put(UTXO_PREFIX + hashX + suffix, value[-8:])
|
||||||
flush_data.adds.clear()
|
flush_data.adds.clear()
|
||||||
|
|
||||||
# New undo information
|
# New undo information
|
||||||
self.flush_undo_infos(batch_put, flush_data.undo_infos)
|
for undo_info, height in flush_data.undo_infos:
|
||||||
|
batch_put(self.undo_key(height), b''.join(undo_info))
|
||||||
flush_data.undo_infos.clear()
|
flush_data.undo_infos.clear()
|
||||||
|
|
||||||
if self.db.for_sync:
|
if self.db.for_sync:
|
||||||
|
@ -285,6 +374,17 @@ class LevelDB:
|
||||||
self.db_tx_count = flush_data.tx_count
|
self.db_tx_count = flush_data.tx_count
|
||||||
self.db_tip = flush_data.tip
|
self.db_tip = flush_data.tip
|
||||||
|
|
||||||
|
def write_history_state(self, batch):
|
||||||
|
state = {
|
||||||
|
'flush_count': self.history.flush_count,
|
||||||
|
'comp_flush_count': self.history.comp_flush_count,
|
||||||
|
'comp_cursor': self.history.comp_cursor,
|
||||||
|
'db_version': self.db_version,
|
||||||
|
}
|
||||||
|
# History entries are not prefixed; the suffix \0\0 ensures we
|
||||||
|
# look similar to other entries and aren't interfered with
|
||||||
|
batch.put(HIST_STATE, repr(state).encode())
|
||||||
|
|
||||||
def flush_dbs(self, flush_data, flush_utxos, estimate_txs_remaining):
|
def flush_dbs(self, flush_data, flush_utxos, estimate_txs_remaining):
|
||||||
"""Flush out cached state. History is always flushed; UTXOs are
|
"""Flush out cached state. History is always flushed; UTXOs are
|
||||||
flushed if flush_utxos."""
|
flushed if flush_utxos."""
|
||||||
|
@ -351,7 +451,7 @@ class LevelDB:
|
||||||
for hashX in sorted(unflushed):
|
for hashX in sorted(unflushed):
|
||||||
key = hashX + flush_id
|
key = hashX + flush_id
|
||||||
batch_put(HASHX_HISTORY_PREFIX + key, unflushed[hashX].tobytes())
|
batch_put(HASHX_HISTORY_PREFIX + key, unflushed[hashX].tobytes())
|
||||||
self.history.write_state(batch)
|
self.write_history_state(batch)
|
||||||
|
|
||||||
unflushed.clear()
|
unflushed.clear()
|
||||||
self.history.unflushed_count = 0
|
self.history.unflushed_count = 0
|
||||||
|
@ -396,45 +496,74 @@ class LevelDB:
|
||||||
self.logger.info(f'sync time: {formatted_time(self.wall_time)} '
|
self.logger.info(f'sync time: {formatted_time(self.wall_time)} '
|
||||||
f'ETA: {formatted_time(eta)}')
|
f'ETA: {formatted_time(eta)}')
|
||||||
|
|
||||||
def flush_state(self, batch):
|
# def flush_state(self, batch):
|
||||||
"""Flush chain state to the batch."""
|
# """Flush chain state to the batch."""
|
||||||
now = time.time()
|
# now = time.time()
|
||||||
self.wall_time += now - self.last_flush
|
# self.wall_time += now - self.last_flush
|
||||||
self.last_flush = now
|
# self.last_flush = now
|
||||||
self.last_flush_tx_count = self.fs_tx_count
|
# self.last_flush_tx_count = self.fs_tx_count
|
||||||
self.write_utxo_state(batch)
|
# self.write_utxo_state(batch)
|
||||||
|
|
||||||
def flush_backup(self, flush_data, touched):
|
def flush_backup(self, flush_data, touched):
|
||||||
"""Like flush_dbs() but when backing up. All UTXOs are flushed."""
|
"""Like flush_dbs() but when backing up. All UTXOs are flushed."""
|
||||||
assert not flush_data.headers
|
assert not flush_data.headers
|
||||||
assert not flush_data.block_txs
|
assert not flush_data.block_txs
|
||||||
assert flush_data.height < self.db_height
|
assert flush_data.height < self.db_height
|
||||||
self.history.assert_flushed()
|
assert not self.history.unflushed
|
||||||
|
|
||||||
start_time = time.time()
|
start_time = time.time()
|
||||||
tx_delta = flush_data.tx_count - self.last_flush_tx_count
|
tx_delta = flush_data.tx_count - self.last_flush_tx_count
|
||||||
|
###
|
||||||
|
while self.fs_height > flush_data.height:
|
||||||
|
self.fs_height -= 1
|
||||||
|
self.headers.pop()
|
||||||
|
self.fs_tx_count = flush_data.tx_count
|
||||||
|
# Truncate header_mc: header count is 1 more than the height.
|
||||||
|
self.header_mc.truncate(flush_data.height + 1)
|
||||||
|
|
||||||
|
###
|
||||||
|
# Not certain this is needed, but it doesn't hurt
|
||||||
|
self.history.flush_count += 1
|
||||||
|
nremoves = 0
|
||||||
|
|
||||||
self.backup_fs(flush_data.height, flush_data.tx_count)
|
|
||||||
self.history.backup(touched, flush_data.tx_count)
|
|
||||||
with self.db.write_batch() as batch:
|
with self.db.write_batch() as batch:
|
||||||
|
tx_count = flush_data.tx_count
|
||||||
|
for hashX in sorted(touched):
|
||||||
|
deletes = []
|
||||||
|
puts = {}
|
||||||
|
for key, hist in self.db.iterator(prefix=HASHX_HISTORY_PREFIX + hashX, reverse=True):
|
||||||
|
k = key[1:]
|
||||||
|
a = array.array('I')
|
||||||
|
a.frombytes(hist)
|
||||||
|
# Remove all history entries >= tx_count
|
||||||
|
idx = bisect_left(a, tx_count)
|
||||||
|
nremoves += len(a) - idx
|
||||||
|
if idx > 0:
|
||||||
|
puts[k] = a[:idx].tobytes()
|
||||||
|
break
|
||||||
|
deletes.append(k)
|
||||||
|
|
||||||
|
for key in deletes:
|
||||||
|
batch.delete(key)
|
||||||
|
for key, value in puts.items():
|
||||||
|
batch.put(key, value)
|
||||||
|
self.write_history_state(batch)
|
||||||
|
|
||||||
self.flush_utxo_db(batch, flush_data)
|
self.flush_utxo_db(batch, flush_data)
|
||||||
# Flush state last as it reads the wall time.
|
# Flush state last as it reads the wall time.
|
||||||
self.flush_state(batch)
|
now = time.time()
|
||||||
|
self.wall_time += now - self.last_flush
|
||||||
|
self.last_flush = now
|
||||||
|
self.last_flush_tx_count = self.fs_tx_count
|
||||||
|
self.write_utxo_state(batch)
|
||||||
|
|
||||||
|
|
||||||
|
self.logger.info(f'backing up removed {nremoves:,d} history entries')
|
||||||
elapsed = self.last_flush - start_time
|
elapsed = self.last_flush - start_time
|
||||||
self.logger.info(f'backup flush #{self.history.flush_count:,d} took '
|
self.logger.info(f'backup flush #{self.history.flush_count:,d} took '
|
||||||
f'{elapsed:.1f}s. Height {flush_data.height:,d} '
|
f'{elapsed:.1f}s. Height {flush_data.height:,d} '
|
||||||
f'txs: {flush_data.tx_count:,d} ({tx_delta:+,d})')
|
f'txs: {flush_data.tx_count:,d} ({tx_delta:+,d})')
|
||||||
|
|
||||||
def backup_fs(self, height, tx_count):
|
|
||||||
"""Back up during a reorg. This just updates our pointers."""
|
|
||||||
while self.fs_height > height:
|
|
||||||
self.fs_height -= 1
|
|
||||||
self.headers.pop()
|
|
||||||
self.fs_tx_count = tx_count
|
|
||||||
# Truncate header_mc: header count is 1 more than the height.
|
|
||||||
self.header_mc.truncate(height + 1)
|
|
||||||
|
|
||||||
def raw_header(self, height):
|
def raw_header(self, height):
|
||||||
"""Return the binary header at the given height."""
|
"""Return the binary header at the given height."""
|
||||||
header, n = self.read_headers(height, 1)
|
header, n = self.read_headers(height, 1)
|
||||||
|
@ -555,7 +684,7 @@ class LevelDB:
|
||||||
cnt = 0
|
cnt = 0
|
||||||
txs = []
|
txs = []
|
||||||
|
|
||||||
for hist in self.history.db.iterator(prefix=HASHX_HISTORY_PREFIX + hashX, include_key=False):
|
for hist in self.db.iterator(prefix=HASHX_HISTORY_PREFIX + hashX, include_key=False):
|
||||||
a = array.array('I')
|
a = array.array('I')
|
||||||
a.frombytes(hist)
|
a.frombytes(hist)
|
||||||
for tx_num in a:
|
for tx_num in a:
|
||||||
|
@ -586,17 +715,12 @@ class LevelDB:
|
||||||
|
|
||||||
def undo_key(self, height):
|
def undo_key(self, height):
|
||||||
"""DB key for undo information at the given height."""
|
"""DB key for undo information at the given height."""
|
||||||
return b'U' + pack('>I', height)
|
return UNDO_PREFIX + pack('>I', height)
|
||||||
|
|
||||||
def read_undo_info(self, height):
|
def read_undo_info(self, height):
|
||||||
"""Read undo information from a file for the current height."""
|
"""Read undo information from a file for the current height."""
|
||||||
return self.db.get(self.undo_key(height))
|
return self.db.get(self.undo_key(height))
|
||||||
|
|
||||||
def flush_undo_infos(self, batch_put, undo_infos):
|
|
||||||
"""undo_infos is a list of (undo_info, height) pairs."""
|
|
||||||
for undo_info, height in undo_infos:
|
|
||||||
batch_put(self.undo_key(height), b''.join(undo_info))
|
|
||||||
|
|
||||||
def raw_block_prefix(self):
|
def raw_block_prefix(self):
|
||||||
return 'block'
|
return 'block'
|
||||||
|
|
||||||
|
@ -626,10 +750,9 @@ class LevelDB:
|
||||||
|
|
||||||
def clear_excess_undo_info(self):
|
def clear_excess_undo_info(self):
|
||||||
"""Clear excess undo info. Only most recent N are kept."""
|
"""Clear excess undo info. Only most recent N are kept."""
|
||||||
prefix = b'U'
|
|
||||||
min_height = self.min_undo_height(self.db_height)
|
min_height = self.min_undo_height(self.db_height)
|
||||||
keys = []
|
keys = []
|
||||||
for key, hist in self.db.iterator(prefix=prefix):
|
for key, hist in self.db.iterator(prefix=UNDO_PREFIX):
|
||||||
height, = unpack('>I', key[-4:])
|
height, = unpack('>I', key[-4:])
|
||||||
if height >= min_height:
|
if height >= min_height:
|
||||||
break
|
break
|
||||||
|
@ -733,7 +856,7 @@ class LevelDB:
|
||||||
fs_tx_hash = self.fs_tx_hash
|
fs_tx_hash = self.fs_tx_hash
|
||||||
# Key: b'u' + address_hashX + tx_idx + tx_num
|
# Key: b'u' + address_hashX + tx_idx + tx_num
|
||||||
# Value: the UTXO value as a 64-bit unsigned integer
|
# Value: the UTXO value as a 64-bit unsigned integer
|
||||||
prefix = b'u' + hashX
|
prefix = UTXO_PREFIX + hashX
|
||||||
for db_key, db_value in self.db.iterator(prefix=prefix):
|
for db_key, db_value in self.db.iterator(prefix=prefix):
|
||||||
tx_pos, tx_num = s_unpack('<HI', db_key[-6:])
|
tx_pos, tx_num = s_unpack('<HI', db_key[-6:])
|
||||||
value, = unpack('<Q', db_value)
|
value, = unpack('<Q', db_value)
|
||||||
|
@ -764,7 +887,7 @@ class LevelDB:
|
||||||
|
|
||||||
# Key: b'h' + compressed_tx_hash + tx_idx + tx_num
|
# Key: b'h' + compressed_tx_hash + tx_idx + tx_num
|
||||||
# Value: hashX
|
# Value: hashX
|
||||||
prefix = b'h' + tx_hash[:4] + idx_packed
|
prefix = HASHX_UTXO_PREFIX + tx_hash[:4] + idx_packed
|
||||||
|
|
||||||
# Find which entry, if any, the TX_HASH matches.
|
# Find which entry, if any, the TX_HASH matches.
|
||||||
for db_key, hashX in self.db.iterator(prefix=prefix):
|
for db_key, hashX in self.db.iterator(prefix=prefix):
|
||||||
|
@ -785,7 +908,7 @@ class LevelDB:
|
||||||
return None
|
return None
|
||||||
# Key: b'u' + address_hashX + tx_idx + tx_num
|
# Key: b'u' + address_hashX + tx_idx + tx_num
|
||||||
# Value: the UTXO value as a 64-bit unsigned integer
|
# Value: the UTXO value as a 64-bit unsigned integer
|
||||||
key = b'u' + hashX + suffix
|
key = UTXO_PREFIX + hashX + suffix
|
||||||
db_value = self.db.get(key)
|
db_value = self.db.get(key)
|
||||||
if not db_value:
|
if not db_value:
|
||||||
# This can happen if the DB was updated between
|
# This can happen if the DB was updated between
|
||||||
|
|
Loading…
Reference in a new issue