5ffd552214
This modifies the block nodes used in the blockchain package for keeping track of the block index to use int64 for the timestamps instead of time.Time. This is being done because a time.Time takes 24 bytes while an int64 only takes 8 and the plan is to eventually move the entire block index into memory instead of the current dynamically-loaded version, so cutting the number of bytes used for the timestamp by a third is highly desirable. Also, the consensus code requires working with unix-style timestamps anyways, so switching over to them in the block node does not seem unreasonable. Finally, this does not go so far as to change all of the time.Time references, particularly those that are in the public API, so it is purely an internal change.
1969 lines
69 KiB
Go
1969 lines
69 KiB
Go
// Copyright (c) 2015-2017 The btcsuite developers
|
|
// Use of this source code is governed by an ISC
|
|
// license that can be found in the LICENSE file.
|
|
|
|
package blockchain
|
|
|
|
import (
|
|
"bytes"
|
|
"encoding/binary"
|
|
"fmt"
|
|
"math/big"
|
|
"sort"
|
|
"time"
|
|
|
|
"github.com/btcsuite/btcd/chaincfg"
|
|
"github.com/btcsuite/btcd/chaincfg/chainhash"
|
|
"github.com/btcsuite/btcd/database"
|
|
"github.com/btcsuite/btcd/wire"
|
|
"github.com/btcsuite/btcutil"
|
|
)
|
|
|
|
var (
|
|
// hashIndexBucketName is the name of the db bucket used to house to the
|
|
// block hash -> block height index.
|
|
hashIndexBucketName = []byte("hashidx")
|
|
|
|
// heightIndexBucketName is the name of the db bucket used to house to
|
|
// the block height -> block hash index.
|
|
heightIndexBucketName = []byte("heightidx")
|
|
|
|
// chainStateKeyName is the name of the db key used to store the best
|
|
// chain state.
|
|
chainStateKeyName = []byte("chainstate")
|
|
|
|
// spendJournalBucketName is the name of the db bucket used to house
|
|
// transactions outputs that are spent in each block.
|
|
spendJournalBucketName = []byte("spendjournal")
|
|
|
|
// utxoSetBucketName is the name of the db bucket used to house the
|
|
// unspent transaction output set.
|
|
utxoSetBucketName = []byte("utxoset")
|
|
|
|
// thresholdBucketName is the name of the db bucket used to house cached
|
|
// threshold states.
|
|
thresholdBucketName = []byte("thresholdstate")
|
|
|
|
// numDeploymentsKeyName is the name of the db key used to store the
|
|
// number of saved deployment caches.
|
|
numDeploymentsKeyName = []byte("numdeployments")
|
|
|
|
// deploymentBucketName is the name of the db bucket used to house the
|
|
// cached threshold states for the actively defined rule deployments.
|
|
deploymentBucketName = []byte("deploymentcache")
|
|
|
|
// deploymentStateKeyName is the name of the db key used to store the
|
|
// deployment state associated with the threshold cache for a given rule
|
|
// deployment.
|
|
deploymentStateKeyName = []byte("deploymentstate")
|
|
|
|
// warningBucketName is the name of the db bucket used to house the
|
|
// cached threshold states for unknown rule deployments.
|
|
warningBucketName = []byte("warningcache")
|
|
|
|
// byteOrder is the preferred byte order used for serializing numeric
|
|
// fields for storage in the database.
|
|
byteOrder = binary.LittleEndian
|
|
)
|
|
|
|
// errNotInMainChain signifies that a block hash or height that is not in the
|
|
// main chain was requested.
|
|
type errNotInMainChain string
|
|
|
|
// Error implements the error interface.
|
|
func (e errNotInMainChain) Error() string {
|
|
return string(e)
|
|
}
|
|
|
|
// isNotInMainChainErr returns whether or not the passed error is an
|
|
// errNotInMainChain error.
|
|
func isNotInMainChainErr(err error) bool {
|
|
_, ok := err.(errNotInMainChain)
|
|
return ok
|
|
}
|
|
|
|
// errDeserialize signifies that a problem was encountered when deserializing
|
|
// data.
|
|
type errDeserialize string
|
|
|
|
// Error implements the error interface.
|
|
func (e errDeserialize) Error() string {
|
|
return string(e)
|
|
}
|
|
|
|
// isDeserializeErr returns whether or not the passed error is an errDeserialize
|
|
// error.
|
|
func isDeserializeErr(err error) bool {
|
|
_, ok := err.(errDeserialize)
|
|
return ok
|
|
}
|
|
|
|
// isDbBucketNotFoundErr returns whether or not the passed error is a
|
|
// database.Error with an error code of database.ErrBucketNotFound.
|
|
func isDbBucketNotFoundErr(err error) bool {
|
|
dbErr, ok := err.(database.Error)
|
|
return ok && dbErr.ErrorCode == database.ErrBucketNotFound
|
|
}
|
|
|
|
// -----------------------------------------------------------------------------
|
|
// The transaction spend journal consists of an entry for each block connected
|
|
// to the main chain which contains the transaction outputs the block spends
|
|
// serialized such that the order is the reverse of the order they were spent.
|
|
//
|
|
// This is required because reorganizing the chain necessarily entails
|
|
// disconnecting blocks to get back to the point of the fork which implies
|
|
// unspending all of the transaction outputs that each block previously spent.
|
|
// Since the utxo set, by definition, only contains unspent transaction outputs,
|
|
// the spent transaction outputs must be resurrected from somewhere. There is
|
|
// more than one way this could be done, however this is the most straight
|
|
// forward method that does not require having a transaction index and unpruned
|
|
// blockchain.
|
|
//
|
|
// NOTE: This format is NOT self describing. The additional details such as
|
|
// the number of entries (transaction inputs) are expected to come from the
|
|
// block itself and the utxo set. The rationale in doing this is to save a
|
|
// significant amount of space. This is also the reason the spent outputs are
|
|
// serialized in the reverse order they are spent because later transactions
|
|
// are allowed to spend outputs from earlier ones in the same block.
|
|
//
|
|
// The serialized format is:
|
|
//
|
|
// [<header code><version><compressed txout>],...
|
|
//
|
|
// Field Type Size
|
|
// header code VLQ variable
|
|
// version VLQ variable
|
|
// compressed txout
|
|
// compressed amount VLQ variable
|
|
// compressed script []byte variable
|
|
//
|
|
// The serialized header code format is:
|
|
// bit 0 - containing transaction is a coinbase
|
|
// bits 1-x - height of the block that contains the spent txout
|
|
//
|
|
// NOTE: The header code and version are only encoded when the spent txout was
|
|
// the final unspent output of the containing transaction. Otherwise, the
|
|
// header code will be 0 and the version is not serialized at all. This is
|
|
// done because that information is only needed when the utxo set no longer
|
|
// has it.
|
|
//
|
|
// Example 1:
|
|
// From block 170 in main blockchain.
|
|
//
|
|
// 1301320511db93e1dcdb8a016b49840f8c53bc1eb68a382e97b1482ecad7b148a6909a5c
|
|
// <><><------------------------------------------------------------------>
|
|
// | | |
|
|
// | version compressed txout
|
|
// header code
|
|
//
|
|
// - header code: 0x13 (coinbase, height 9)
|
|
// - transaction version: 1
|
|
// - compressed txout 0:
|
|
// - 0x32: VLQ-encoded compressed amount for 5000000000 (50 BTC)
|
|
// - 0x05: special script type pay-to-pubkey
|
|
// - 0x11...5c: x-coordinate of the pubkey
|
|
//
|
|
// Example 2:
|
|
// Adapted from block 100025 in main blockchain.
|
|
//
|
|
// 0091f20f006edbc6c4d31bae9f1ccc38538a114bf42de65e868b99700186c64700b2fb57eadf61e106a100a7445a8c3f67898841ec
|
|
// <><----------------------------------------------><----><><---------------------------------------------->
|
|
// | | | | |
|
|
// | compressed txout | version compressed txout
|
|
// header code header code
|
|
//
|
|
// - Last spent output:
|
|
// - header code: 0x00 (was not the final unspent output for containing tx)
|
|
// - transaction version: Nothing since header code is 0
|
|
// - compressed txout:
|
|
// - 0x91f20f: VLQ-encoded compressed amount for 34405000000 (344.05 BTC)
|
|
// - 0x00: special script type pay-to-pubkey-hash
|
|
// - 0x6e...86: pubkey hash
|
|
// - Second to last spent output:
|
|
// - header code: 0x8b9970 (not coinbase, height 100024)
|
|
// - transaction version: 1
|
|
// - compressed txout:
|
|
// - 0x86c647: VLQ-encoded compressed amount for 13761000000 (137.61 BTC)
|
|
// - 0x00: special script type pay-to-pubkey-hash
|
|
// - 0xb2...ec: pubkey hash
|
|
// -----------------------------------------------------------------------------
|
|
|
|
// spentTxOut contains a spent transaction output and potentially additional
|
|
// contextual information such as whether or not it was contained in a coinbase
|
|
// transaction, the version of the transaction it was contained in, and which
|
|
// block height the containing transaction was included in. As described in
|
|
// the comments above, the additional contextual information will only be valid
|
|
// when this spent txout is spending the last unspent output of the containing
|
|
// transaction.
|
|
type spentTxOut struct {
|
|
compressed bool // The amount and public key script are compressed.
|
|
version int32 // The version of creating tx.
|
|
amount int64 // The amount of the output.
|
|
pkScript []byte // The public key script for the output.
|
|
|
|
// These fields are only set when this is spending the final output of
|
|
// the creating tx.
|
|
height int32 // Height of the the block containing the creating tx.
|
|
isCoinBase bool // Whether creating tx is a coinbase.
|
|
}
|
|
|
|
// spentTxOutHeaderCode returns the calculated header code to be used when
|
|
// serializing the provided stxo entry.
|
|
func spentTxOutHeaderCode(stxo *spentTxOut) uint64 {
|
|
// The header code is 0 when there is no height set for the stxo.
|
|
if stxo.height == 0 {
|
|
return 0
|
|
}
|
|
|
|
// As described in the serialization format comments, the header code
|
|
// encodes the height shifted over one bit and the coinbase flag in the
|
|
// lowest bit.
|
|
headerCode := uint64(stxo.height) << 1
|
|
if stxo.isCoinBase {
|
|
headerCode |= 0x01
|
|
}
|
|
|
|
return headerCode
|
|
}
|
|
|
|
// spentTxOutSerializeSize returns the number of bytes it would take to
|
|
// serialize the passed stxo according to the format described above.
|
|
func spentTxOutSerializeSize(stxo *spentTxOut) int {
|
|
headerCode := spentTxOutHeaderCode(stxo)
|
|
size := serializeSizeVLQ(headerCode)
|
|
if headerCode != 0 {
|
|
size += serializeSizeVLQ(uint64(stxo.version))
|
|
}
|
|
return size + compressedTxOutSize(uint64(stxo.amount), stxo.pkScript,
|
|
stxo.version, stxo.compressed)
|
|
}
|
|
|
|
// putSpentTxOut serializes the passed stxo according to the format described
|
|
// above directly into the passed target byte slice. The target byte slice must
|
|
// be at least large enough to handle the number of bytes returned by the
|
|
// spentTxOutSerializeSize function or it will panic.
|
|
func putSpentTxOut(target []byte, stxo *spentTxOut) int {
|
|
headerCode := spentTxOutHeaderCode(stxo)
|
|
offset := putVLQ(target, headerCode)
|
|
if headerCode != 0 {
|
|
offset += putVLQ(target[offset:], uint64(stxo.version))
|
|
}
|
|
return offset + putCompressedTxOut(target[offset:], uint64(stxo.amount),
|
|
stxo.pkScript, stxo.version, stxo.compressed)
|
|
}
|
|
|
|
// decodeSpentTxOut decodes the passed serialized stxo entry, possibly followed
|
|
// by other data, into the passed stxo struct. It returns the number of bytes
|
|
// read.
|
|
//
|
|
// Since the serialized stxo entry does not contain the height, version, or
|
|
// coinbase flag of the containing transaction when it still has utxos, the
|
|
// caller is responsible for passing in the containing transaction version in
|
|
// that case. The provided version is ignore when it is serialized as a part of
|
|
// the stxo.
|
|
//
|
|
// An error will be returned if the version is not serialized as a part of the
|
|
// stxo and is also not provided to the function.
|
|
func decodeSpentTxOut(serialized []byte, stxo *spentTxOut, txVersion int32) (int, error) {
|
|
// Ensure there are bytes to decode.
|
|
if len(serialized) == 0 {
|
|
return 0, errDeserialize("no serialized bytes")
|
|
}
|
|
|
|
// Deserialize the header code.
|
|
code, offset := deserializeVLQ(serialized)
|
|
if offset >= len(serialized) {
|
|
return offset, errDeserialize("unexpected end of data after " +
|
|
"header code")
|
|
}
|
|
|
|
// Decode the header code and deserialize the containing transaction
|
|
// version if needed.
|
|
//
|
|
// Bit 0 indicates containing transaction is a coinbase.
|
|
// Bits 1-x encode height of containing transaction.
|
|
if code != 0 {
|
|
version, bytesRead := deserializeVLQ(serialized[offset:])
|
|
offset += bytesRead
|
|
if offset >= len(serialized) {
|
|
return offset, errDeserialize("unexpected end of data " +
|
|
"after version")
|
|
}
|
|
|
|
stxo.isCoinBase = code&0x01 != 0
|
|
stxo.height = int32(code >> 1)
|
|
stxo.version = int32(version)
|
|
} else {
|
|
// Ensure a tx version was specified if the stxo did not encode
|
|
// it. This should never happen unless there is database
|
|
// corruption or this function is being called without the
|
|
// proper state.
|
|
if txVersion == 0 {
|
|
return offset, AssertError("decodeSpentTxOut called " +
|
|
"without a containing tx version when the " +
|
|
"serialized stxo that does not encode the " +
|
|
"version")
|
|
}
|
|
stxo.version = txVersion
|
|
}
|
|
|
|
// Decode the compressed txout.
|
|
compAmount, compScript, bytesRead, err := decodeCompressedTxOut(
|
|
serialized[offset:], stxo.version)
|
|
offset += bytesRead
|
|
if err != nil {
|
|
return offset, errDeserialize(fmt.Sprintf("unable to decode "+
|
|
"txout: %v", err))
|
|
}
|
|
stxo.amount = int64(compAmount)
|
|
stxo.pkScript = compScript
|
|
stxo.compressed = true
|
|
return offset, nil
|
|
}
|
|
|
|
// deserializeSpendJournalEntry decodes the passed serialized byte slice into a
|
|
// slice of spent txouts according to the format described in detail above.
|
|
//
|
|
// Since the serialization format is not self describing, as noted in the
|
|
// format comments, this function also requires the transactions that spend the
|
|
// txouts and a utxo view that contains any remaining existing utxos in the
|
|
// transactions referenced by the inputs to the passed transasctions.
|
|
func deserializeSpendJournalEntry(serialized []byte, txns []*wire.MsgTx, view *UtxoViewpoint) ([]spentTxOut, error) {
|
|
// Calculate the total number of stxos.
|
|
var numStxos int
|
|
for _, tx := range txns {
|
|
numStxos += len(tx.TxIn)
|
|
}
|
|
|
|
// When a block has no spent txouts there is nothing to serialize.
|
|
if len(serialized) == 0 {
|
|
// Ensure the block actually has no stxos. This should never
|
|
// happen unless there is database corruption or an empty entry
|
|
// erroneously made its way into the database.
|
|
if numStxos != 0 {
|
|
return nil, AssertError(fmt.Sprintf("mismatched spend "+
|
|
"journal serialization - no serialization for "+
|
|
"expected %d stxos", numStxos))
|
|
}
|
|
|
|
return nil, nil
|
|
}
|
|
|
|
// Loop backwards through all transactions so everything is read in
|
|
// reverse order to match the serialization order.
|
|
stxoIdx := numStxos - 1
|
|
stxoInFlight := make(map[chainhash.Hash]int)
|
|
offset := 0
|
|
stxos := make([]spentTxOut, numStxos)
|
|
for txIdx := len(txns) - 1; txIdx > -1; txIdx-- {
|
|
tx := txns[txIdx]
|
|
|
|
// Loop backwards through all of the transaction inputs and read
|
|
// the associated stxo.
|
|
for txInIdx := len(tx.TxIn) - 1; txInIdx > -1; txInIdx-- {
|
|
txIn := tx.TxIn[txInIdx]
|
|
stxo := &stxos[stxoIdx]
|
|
stxoIdx--
|
|
|
|
// Get the transaction version for the stxo based on
|
|
// whether or not it should be serialized as a part of
|
|
// the stxo. Recall that it is only serialized when the
|
|
// stxo spends the final utxo of a transaction. Since
|
|
// they are deserialized in reverse order, this means
|
|
// the first time an entry for a given containing tx is
|
|
// encountered that is not already in the utxo view it
|
|
// must have been the final spend and thus the extra
|
|
// data will be serialized with the stxo. Otherwise,
|
|
// the version must be pulled from the utxo entry.
|
|
//
|
|
// Since the view is not actually modified as the stxos
|
|
// are read here and it's possible later entries
|
|
// reference earlier ones, an inflight map is maintained
|
|
// to detect this case and pull the tx version from the
|
|
// entry that contains the version information as just
|
|
// described.
|
|
var txVersion int32
|
|
originHash := &txIn.PreviousOutPoint.Hash
|
|
entry := view.LookupEntry(originHash)
|
|
if entry != nil {
|
|
txVersion = entry.Version()
|
|
} else if idx, ok := stxoInFlight[*originHash]; ok {
|
|
txVersion = stxos[idx].version
|
|
} else {
|
|
stxoInFlight[*originHash] = stxoIdx + 1
|
|
}
|
|
|
|
n, err := decodeSpentTxOut(serialized[offset:], stxo,
|
|
txVersion)
|
|
offset += n
|
|
if err != nil {
|
|
return nil, errDeserialize(fmt.Sprintf("unable "+
|
|
"to decode stxo for %v: %v",
|
|
txIn.PreviousOutPoint, err))
|
|
}
|
|
}
|
|
}
|
|
|
|
return stxos, nil
|
|
}
|
|
|
|
// serializeSpendJournalEntry serializes all of the passed spent txouts into a
|
|
// single byte slice according to the format described in detail above.
|
|
func serializeSpendJournalEntry(stxos []spentTxOut) []byte {
|
|
if len(stxos) == 0 {
|
|
return nil
|
|
}
|
|
|
|
// Calculate the size needed to serialize the entire journal entry.
|
|
var size int
|
|
for i := range stxos {
|
|
size += spentTxOutSerializeSize(&stxos[i])
|
|
}
|
|
serialized := make([]byte, size)
|
|
|
|
// Serialize each individual stxo directly into the slice in reverse
|
|
// order one after the other.
|
|
var offset int
|
|
for i := len(stxos) - 1; i > -1; i-- {
|
|
offset += putSpentTxOut(serialized[offset:], &stxos[i])
|
|
}
|
|
|
|
return serialized
|
|
}
|
|
|
|
// dbFetchSpendJournalEntry fetches the spend journal entry for the passed
|
|
// block and deserializes it into a slice of spent txout entries. The provided
|
|
// view MUST have the utxos referenced by all of the transactions available for
|
|
// the passed block since that information is required to reconstruct the spent
|
|
// txouts.
|
|
func dbFetchSpendJournalEntry(dbTx database.Tx, block *btcutil.Block, view *UtxoViewpoint) ([]spentTxOut, error) {
|
|
// Exclude the coinbase transaction since it can't spend anything.
|
|
spendBucket := dbTx.Metadata().Bucket(spendJournalBucketName)
|
|
serialized := spendBucket.Get(block.Hash()[:])
|
|
blockTxns := block.MsgBlock().Transactions[1:]
|
|
stxos, err := deserializeSpendJournalEntry(serialized, blockTxns, view)
|
|
if err != nil {
|
|
// Ensure any deserialization errors are returned as database
|
|
// corruption errors.
|
|
if isDeserializeErr(err) {
|
|
return nil, database.Error{
|
|
ErrorCode: database.ErrCorruption,
|
|
Description: fmt.Sprintf("corrupt spend "+
|
|
"information for %v: %v", block.Hash(),
|
|
err),
|
|
}
|
|
}
|
|
|
|
return nil, err
|
|
}
|
|
|
|
return stxos, nil
|
|
}
|
|
|
|
// dbPutSpendJournalEntry uses an existing database transaction to update the
|
|
// spend journal entry for the given block hash using the provided slice of
|
|
// spent txouts. The spent txouts slice must contain an entry for every txout
|
|
// the transactions in the block spend in the order they are spent.
|
|
func dbPutSpendJournalEntry(dbTx database.Tx, blockHash *chainhash.Hash, stxos []spentTxOut) error {
|
|
spendBucket := dbTx.Metadata().Bucket(spendJournalBucketName)
|
|
serialized := serializeSpendJournalEntry(stxos)
|
|
return spendBucket.Put(blockHash[:], serialized)
|
|
}
|
|
|
|
// dbRemoveSpendJournalEntry uses an existing database transaction to remove the
|
|
// spend journal entry for the passed block hash.
|
|
func dbRemoveSpendJournalEntry(dbTx database.Tx, blockHash *chainhash.Hash) error {
|
|
spendBucket := dbTx.Metadata().Bucket(spendJournalBucketName)
|
|
return spendBucket.Delete(blockHash[:])
|
|
}
|
|
|
|
// -----------------------------------------------------------------------------
|
|
// The unspent transaction output (utxo) set consists of an entry for each
|
|
// transaction which contains a utxo serialized using a format that is highly
|
|
// optimized to reduce space using domain specific compression algorithms. This
|
|
// format is a slightly modified version of the format used in Bitcoin Core.
|
|
//
|
|
// The serialized format is:
|
|
//
|
|
// <version><height><header code><unspentness bitmap>[<compressed txouts>,...]
|
|
//
|
|
// Field Type Size
|
|
// version VLQ variable
|
|
// block height VLQ variable
|
|
// header code VLQ variable
|
|
// unspentness bitmap []byte variable
|
|
// compressed txouts
|
|
// compressed amount VLQ variable
|
|
// compressed script []byte variable
|
|
//
|
|
// The serialized header code format is:
|
|
// bit 0 - containing transaction is a coinbase
|
|
// bit 1 - output zero is unspent
|
|
// bit 2 - output one is unspent
|
|
// bits 3-x - number of bytes in unspentness bitmap. When both bits 1 and 2
|
|
// are unset, it encodes N-1 since there must be at least one unspent
|
|
// output.
|
|
//
|
|
// The rationale for the header code scheme is as follows:
|
|
// - Transactions which only pay to a single output and a change output are
|
|
// extremely common, thus an extra byte for the unspentness bitmap can be
|
|
// avoided for them by encoding those two outputs in the low order bits.
|
|
// - Given it is encoded as a VLQ which can encode values up to 127 with a
|
|
// single byte, that leaves 4 bits to represent the number of bytes in the
|
|
// unspentness bitmap while still only consuming a single byte for the
|
|
// header code. In other words, an unspentness bitmap with up to 120
|
|
// transaction outputs can be encoded with a single-byte header code.
|
|
// This covers the vast majority of transactions.
|
|
// - Encoding N-1 bytes when both bits 1 and 2 are unset allows an additional
|
|
// 8 outpoints to be encoded before causing the header code to require an
|
|
// additional byte.
|
|
//
|
|
// Example 1:
|
|
// From tx in main blockchain:
|
|
// Blk 1, 0e3e2357e806b6cdb1f70b54c3a3a17b6714ee1f0e68bebb44a74b1efd512098
|
|
//
|
|
// 010103320496b538e853519c726a2c91e61ec11600ae1390813a627c66fb8be7947be63c52
|
|
// <><><><------------------------------------------------------------------>
|
|
// | | \--------\ |
|
|
// | height | compressed txout 0
|
|
// version header code
|
|
//
|
|
// - version: 1
|
|
// - height: 1
|
|
// - header code: 0x03 (coinbase, output zero unspent, 0 bytes of unspentness)
|
|
// - unspentness: Nothing since it is zero bytes
|
|
// - compressed txout 0:
|
|
// - 0x32: VLQ-encoded compressed amount for 5000000000 (50 BTC)
|
|
// - 0x04: special script type pay-to-pubkey
|
|
// - 0x96...52: x-coordinate of the pubkey
|
|
//
|
|
// Example 2:
|
|
// From tx in main blockchain:
|
|
// Blk 113931, 4a16969aa4764dd7507fc1de7f0baa4850a246de90c45e59a3207f9a26b5036f
|
|
//
|
|
// 0185f90b0a011200e2ccd6ec7c6e2e581349c77e067385fa8236bf8a800900b8025be1b3efc63b0ad48e7f9f10e87544528d58
|
|
// <><----><><><------------------------------------------><-------------------------------------------->
|
|
// | | | \-------------------\ | |
|
|
// version | \--------\ unspentness | compressed txout 2
|
|
// height header code compressed txout 0
|
|
//
|
|
// - version: 1
|
|
// - height: 113931
|
|
// - header code: 0x0a (output zero unspent, 1 byte in unspentness bitmap)
|
|
// - unspentness: [0x01] (bit 0 is set, so output 0+2 = 2 is unspent)
|
|
// NOTE: It's +2 since the first two outputs are encoded in the header code
|
|
// - compressed txout 0:
|
|
// - 0x12: VLQ-encoded compressed amount for 20000000 (0.2 BTC)
|
|
// - 0x00: special script type pay-to-pubkey-hash
|
|
// - 0xe2...8a: pubkey hash
|
|
// - compressed txout 2:
|
|
// - 0x8009: VLQ-encoded compressed amount for 15000000 (0.15 BTC)
|
|
// - 0x00: special script type pay-to-pubkey-hash
|
|
// - 0xb8...58: pubkey hash
|
|
//
|
|
// Example 3:
|
|
// From tx in main blockchain:
|
|
// Blk 338156, 1b02d1c8cfef60a189017b9a420c682cf4a0028175f2f563209e4ff61c8c3620
|
|
//
|
|
// 0193d06c100000108ba5b9e763011dd46a006572d820e448e12d2bbb38640bc718e6
|
|
// <><----><><----><-------------------------------------------------->
|
|
// | | | \-----------------\ |
|
|
// version | \--------\ unspentness |
|
|
// height header code compressed txout 22
|
|
//
|
|
// - version: 1
|
|
// - height: 338156
|
|
// - header code: 0x10 (2+1 = 3 bytes in unspentness bitmap)
|
|
// NOTE: It's +1 since neither bit 1 nor 2 are set, so N-1 is encoded.
|
|
// - unspentness: [0x00 0x00 0x10] (bit 20 is set, so output 20+2 = 22 is unspent)
|
|
// NOTE: It's +2 since the first two outputs are encoded in the header code
|
|
// - compressed txout 22:
|
|
// - 0x8ba5b9e763: VLQ-encoded compressed amount for 366875659 (3.66875659 BTC)
|
|
// - 0x01: special script type pay-to-script-hash
|
|
// - 0x1d...e6: script hash
|
|
// -----------------------------------------------------------------------------
|
|
|
|
// utxoEntryHeaderCode returns the calculated header code to be used when
|
|
// serializing the provided utxo entry and the number of bytes needed to encode
|
|
// the unspentness bitmap.
|
|
func utxoEntryHeaderCode(entry *UtxoEntry, highestOutputIndex uint32) (uint64, int, error) {
|
|
// The first two outputs are encoded separately, so offset the index
|
|
// accordingly to calculate the correct number of bytes needed to encode
|
|
// up to the highest unspent output index.
|
|
numBitmapBytes := int((highestOutputIndex + 6) / 8)
|
|
|
|
// As previously described, one less than the number of bytes is encoded
|
|
// when both output 0 and 1 are spent because there must be at least one
|
|
// unspent output. Adjust the number of bytes to encode accordingly and
|
|
// encode the value by shifting it over 3 bits.
|
|
output0Unspent := !entry.IsOutputSpent(0)
|
|
output1Unspent := !entry.IsOutputSpent(1)
|
|
var numBitmapBytesAdjustment int
|
|
if !output0Unspent && !output1Unspent {
|
|
if numBitmapBytes == 0 {
|
|
return 0, 0, AssertError("attempt to serialize utxo " +
|
|
"header for fully spent transaction")
|
|
}
|
|
numBitmapBytesAdjustment = 1
|
|
}
|
|
headerCode := uint64(numBitmapBytes-numBitmapBytesAdjustment) << 3
|
|
|
|
// Set the coinbase, output 0, and output 1 bits in the header code
|
|
// accordingly.
|
|
if entry.isCoinBase {
|
|
headerCode |= 0x01 // bit 0
|
|
}
|
|
if output0Unspent {
|
|
headerCode |= 0x02 // bit 1
|
|
}
|
|
if output1Unspent {
|
|
headerCode |= 0x04 // bit 2
|
|
}
|
|
|
|
return headerCode, numBitmapBytes, nil
|
|
}
|
|
|
|
// serializeUtxoEntry returns the entry serialized to a format that is suitable
|
|
// for long-term storage. The format is described in detail above.
|
|
func serializeUtxoEntry(entry *UtxoEntry) ([]byte, error) {
|
|
// Fully spent entries have no serialization.
|
|
if entry.IsFullySpent() {
|
|
return nil, nil
|
|
}
|
|
|
|
// Determine the output order by sorting the sparse output index keys.
|
|
outputOrder := make([]int, 0, len(entry.sparseOutputs))
|
|
for outputIndex := range entry.sparseOutputs {
|
|
outputOrder = append(outputOrder, int(outputIndex))
|
|
}
|
|
sort.Ints(outputOrder)
|
|
|
|
// Encode the header code and determine the number of bytes the
|
|
// unspentness bitmap needs.
|
|
highIndex := uint32(outputOrder[len(outputOrder)-1])
|
|
headerCode, numBitmapBytes, err := utxoEntryHeaderCode(entry, highIndex)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Calculate the size needed to serialize the entry.
|
|
size := serializeSizeVLQ(uint64(entry.version)) +
|
|
serializeSizeVLQ(uint64(entry.blockHeight)) +
|
|
serializeSizeVLQ(headerCode) + numBitmapBytes
|
|
for _, outputIndex := range outputOrder {
|
|
out := entry.sparseOutputs[uint32(outputIndex)]
|
|
if out.spent {
|
|
continue
|
|
}
|
|
size += compressedTxOutSize(uint64(out.amount), out.pkScript,
|
|
entry.version, out.compressed)
|
|
}
|
|
|
|
// Serialize the version, block height of the containing transaction,
|
|
// and header code.
|
|
serialized := make([]byte, size)
|
|
offset := putVLQ(serialized, uint64(entry.version))
|
|
offset += putVLQ(serialized[offset:], uint64(entry.blockHeight))
|
|
offset += putVLQ(serialized[offset:], headerCode)
|
|
|
|
// Serialize the unspentness bitmap.
|
|
for i := uint32(0); i < uint32(numBitmapBytes); i++ {
|
|
unspentBits := byte(0)
|
|
for j := uint32(0); j < 8; j++ {
|
|
// The first 2 outputs are encoded via the header code,
|
|
// so adjust the output index accordingly.
|
|
if !entry.IsOutputSpent(2 + i*8 + j) {
|
|
unspentBits |= 1 << uint8(j)
|
|
}
|
|
}
|
|
serialized[offset] = unspentBits
|
|
offset++
|
|
}
|
|
|
|
// Serialize the compressed unspent transaction outputs. Outputs that
|
|
// are already compressed are serialized without modifications.
|
|
for _, outputIndex := range outputOrder {
|
|
out := entry.sparseOutputs[uint32(outputIndex)]
|
|
if out.spent {
|
|
continue
|
|
}
|
|
|
|
offset += putCompressedTxOut(serialized[offset:],
|
|
uint64(out.amount), out.pkScript, entry.version,
|
|
out.compressed)
|
|
}
|
|
|
|
return serialized, nil
|
|
}
|
|
|
|
// deserializeUtxoEntry decodes a utxo entry from the passed serialized byte
|
|
// slice into a new UtxoEntry using a format that is suitable for long-term
|
|
// storage. The format is described in detail above.
|
|
func deserializeUtxoEntry(serialized []byte) (*UtxoEntry, error) {
|
|
// Deserialize the version.
|
|
version, bytesRead := deserializeVLQ(serialized)
|
|
offset := bytesRead
|
|
if offset >= len(serialized) {
|
|
return nil, errDeserialize("unexpected end of data after version")
|
|
}
|
|
|
|
// Deserialize the block height.
|
|
blockHeight, bytesRead := deserializeVLQ(serialized[offset:])
|
|
offset += bytesRead
|
|
if offset >= len(serialized) {
|
|
return nil, errDeserialize("unexpected end of data after height")
|
|
}
|
|
|
|
// Deserialize the header code.
|
|
code, bytesRead := deserializeVLQ(serialized[offset:])
|
|
offset += bytesRead
|
|
if offset >= len(serialized) {
|
|
return nil, errDeserialize("unexpected end of data after header")
|
|
}
|
|
|
|
// Decode the header code.
|
|
//
|
|
// Bit 0 indicates whether the containing transaction is a coinbase.
|
|
// Bit 1 indicates output 0 is unspent.
|
|
// Bit 2 indicates output 1 is unspent.
|
|
// Bits 3-x encodes the number of non-zero unspentness bitmap bytes that
|
|
// follow. When both output 0 and 1 are spent, it encodes N-1.
|
|
isCoinBase := code&0x01 != 0
|
|
output0Unspent := code&0x02 != 0
|
|
output1Unspent := code&0x04 != 0
|
|
numBitmapBytes := code >> 3
|
|
if !output0Unspent && !output1Unspent {
|
|
numBitmapBytes++
|
|
}
|
|
|
|
// Ensure there are enough bytes left to deserialize the unspentness
|
|
// bitmap.
|
|
if uint64(len(serialized[offset:])) < numBitmapBytes {
|
|
return nil, errDeserialize("unexpected end of data for " +
|
|
"unspentness bitmap")
|
|
}
|
|
|
|
// Create a new utxo entry with the details deserialized above to house
|
|
// all of the utxos.
|
|
entry := newUtxoEntry(int32(version), isCoinBase, int32(blockHeight))
|
|
|
|
// Add sparse output for unspent outputs 0 and 1 as needed based on the
|
|
// details provided by the header code.
|
|
var outputIndexes []uint32
|
|
if output0Unspent {
|
|
outputIndexes = append(outputIndexes, 0)
|
|
}
|
|
if output1Unspent {
|
|
outputIndexes = append(outputIndexes, 1)
|
|
}
|
|
|
|
// Decode the unspentness bitmap adding a sparse output for each unspent
|
|
// output.
|
|
for i := uint32(0); i < uint32(numBitmapBytes); i++ {
|
|
unspentBits := serialized[offset]
|
|
for j := uint32(0); j < 8; j++ {
|
|
if unspentBits&0x01 != 0 {
|
|
// The first 2 outputs are encoded via the
|
|
// header code, so adjust the output number
|
|
// accordingly.
|
|
outputNum := 2 + i*8 + j
|
|
outputIndexes = append(outputIndexes, outputNum)
|
|
}
|
|
unspentBits >>= 1
|
|
}
|
|
offset++
|
|
}
|
|
|
|
// Decode and add all of the utxos.
|
|
for i, outputIndex := range outputIndexes {
|
|
// Decode the next utxo. The script and amount fields of the
|
|
// utxo output are left compressed so decompression can be
|
|
// avoided on those that are not accessed. This is done since
|
|
// it is quite common for a redeeming transaction to only
|
|
// reference a single utxo from a referenced transaction.
|
|
compAmount, compScript, bytesRead, err := decodeCompressedTxOut(
|
|
serialized[offset:], int32(version))
|
|
if err != nil {
|
|
return nil, errDeserialize(fmt.Sprintf("unable to "+
|
|
"decode utxo at index %d: %v", i, err))
|
|
}
|
|
offset += bytesRead
|
|
|
|
entry.sparseOutputs[outputIndex] = &utxoOutput{
|
|
spent: false,
|
|
compressed: true,
|
|
pkScript: compScript,
|
|
amount: int64(compAmount),
|
|
}
|
|
}
|
|
|
|
return entry, nil
|
|
}
|
|
|
|
// dbFetchUtxoEntry uses an existing database transaction to fetch all unspent
|
|
// outputs for the provided Bitcoin transaction hash from the utxo set.
|
|
//
|
|
// When there is no entry for the provided hash, nil will be returned for the
|
|
// both the entry and the error.
|
|
func dbFetchUtxoEntry(dbTx database.Tx, hash *chainhash.Hash) (*UtxoEntry, error) {
|
|
// Fetch the unspent transaction output information for the passed
|
|
// transaction hash. Return now when there is no entry.
|
|
utxoBucket := dbTx.Metadata().Bucket(utxoSetBucketName)
|
|
serializedUtxo := utxoBucket.Get(hash[:])
|
|
if serializedUtxo == nil {
|
|
return nil, nil
|
|
}
|
|
|
|
// A non-nil zero-length entry means there is an entry in the database
|
|
// for a fully spent transaction which should never be the case.
|
|
if len(serializedUtxo) == 0 {
|
|
return nil, AssertError(fmt.Sprintf("database contains entry "+
|
|
"for fully spent tx %v", hash))
|
|
}
|
|
|
|
// Deserialize the utxo entry and return it.
|
|
entry, err := deserializeUtxoEntry(serializedUtxo)
|
|
if err != nil {
|
|
// Ensure any deserialization errors are returned as database
|
|
// corruption errors.
|
|
if isDeserializeErr(err) {
|
|
return nil, database.Error{
|
|
ErrorCode: database.ErrCorruption,
|
|
Description: fmt.Sprintf("corrupt utxo entry "+
|
|
"for %v: %v", hash, err),
|
|
}
|
|
}
|
|
|
|
return nil, err
|
|
}
|
|
|
|
return entry, nil
|
|
}
|
|
|
|
// dbPutUtxoView uses an existing database transaction to update the utxo set
|
|
// in the database based on the provided utxo view contents and state. In
|
|
// particular, only the entries that have been marked as modified are written
|
|
// to the database.
|
|
func dbPutUtxoView(dbTx database.Tx, view *UtxoViewpoint) error {
|
|
utxoBucket := dbTx.Metadata().Bucket(utxoSetBucketName)
|
|
for txHashIter, entry := range view.entries {
|
|
// No need to update the database if the entry was not modified.
|
|
if entry == nil || !entry.modified {
|
|
continue
|
|
}
|
|
|
|
// Serialize the utxo entry without any entries that have been
|
|
// spent.
|
|
serialized, err := serializeUtxoEntry(entry)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Make a copy of the hash because the iterator changes on each
|
|
// loop iteration and thus slicing it directly would cause the
|
|
// data to change out from under the put/delete funcs below.
|
|
txHash := txHashIter
|
|
|
|
// Remove the utxo entry if it is now fully spent.
|
|
if serialized == nil {
|
|
if err := utxoBucket.Delete(txHash[:]); err != nil {
|
|
return err
|
|
}
|
|
|
|
continue
|
|
}
|
|
|
|
// At this point the utxo entry is not fully spent, so store its
|
|
// serialization in the database.
|
|
err = utxoBucket.Put(txHash[:], serialized)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// -----------------------------------------------------------------------------
|
|
// The block index consists of two buckets with an entry for every block in the
|
|
// main chain. One bucket is for the hash to height mapping and the other is
|
|
// for the height to hash mapping.
|
|
//
|
|
// The serialized format for values in the hash to height bucket is:
|
|
// <height>
|
|
//
|
|
// Field Type Size
|
|
// height uint32 4 bytes
|
|
//
|
|
// The serialized format for values in the height to hash bucket is:
|
|
// <hash>
|
|
//
|
|
// Field Type Size
|
|
// hash chainhash.Hash chainhash.HashSize
|
|
// -----------------------------------------------------------------------------
|
|
|
|
// dbPutBlockIndex uses an existing database transaction to update or add the
|
|
// block index entries for the hash to height and height to hash mappings for
|
|
// the provided values.
|
|
func dbPutBlockIndex(dbTx database.Tx, hash *chainhash.Hash, height int32) error {
|
|
// Serialize the height for use in the index entries.
|
|
var serializedHeight [4]byte
|
|
byteOrder.PutUint32(serializedHeight[:], uint32(height))
|
|
|
|
// Add the block hash to height mapping to the index.
|
|
meta := dbTx.Metadata()
|
|
hashIndex := meta.Bucket(hashIndexBucketName)
|
|
if err := hashIndex.Put(hash[:], serializedHeight[:]); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Add the block height to hash mapping to the index.
|
|
heightIndex := meta.Bucket(heightIndexBucketName)
|
|
return heightIndex.Put(serializedHeight[:], hash[:])
|
|
}
|
|
|
|
// dbRemoveBlockIndex uses an existing database transaction remove block index
|
|
// entries from the hash to height and height to hash mappings for the provided
|
|
// values.
|
|
func dbRemoveBlockIndex(dbTx database.Tx, hash *chainhash.Hash, height int32) error {
|
|
// Remove the block hash to height mapping.
|
|
meta := dbTx.Metadata()
|
|
hashIndex := meta.Bucket(hashIndexBucketName)
|
|
if err := hashIndex.Delete(hash[:]); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Remove the block height to hash mapping.
|
|
var serializedHeight [4]byte
|
|
byteOrder.PutUint32(serializedHeight[:], uint32(height))
|
|
heightIndex := meta.Bucket(heightIndexBucketName)
|
|
return heightIndex.Delete(serializedHeight[:])
|
|
}
|
|
|
|
// dbFetchHeightByHash uses an existing database transaction to retrieve the
|
|
// height for the provided hash from the index.
|
|
func dbFetchHeightByHash(dbTx database.Tx, hash *chainhash.Hash) (int32, error) {
|
|
meta := dbTx.Metadata()
|
|
hashIndex := meta.Bucket(hashIndexBucketName)
|
|
serializedHeight := hashIndex.Get(hash[:])
|
|
if serializedHeight == nil {
|
|
str := fmt.Sprintf("block %s is not in the main chain", hash)
|
|
return 0, errNotInMainChain(str)
|
|
}
|
|
|
|
return int32(byteOrder.Uint32(serializedHeight)), nil
|
|
}
|
|
|
|
// dbFetchHashByHeight uses an existing database transaction to retrieve the
|
|
// hash for the provided height from the index.
|
|
func dbFetchHashByHeight(dbTx database.Tx, height int32) (*chainhash.Hash, error) {
|
|
var serializedHeight [4]byte
|
|
byteOrder.PutUint32(serializedHeight[:], uint32(height))
|
|
|
|
meta := dbTx.Metadata()
|
|
heightIndex := meta.Bucket(heightIndexBucketName)
|
|
hashBytes := heightIndex.Get(serializedHeight[:])
|
|
if hashBytes == nil {
|
|
str := fmt.Sprintf("no block at height %d exists", height)
|
|
return nil, errNotInMainChain(str)
|
|
}
|
|
|
|
var hash chainhash.Hash
|
|
copy(hash[:], hashBytes)
|
|
return &hash, nil
|
|
}
|
|
|
|
// -----------------------------------------------------------------------------
|
|
// The best chain state consists of the best block hash and height, the total
|
|
// number of transactions up to and including those in the best block, and the
|
|
// accumulated work sum up to and including the best block.
|
|
//
|
|
// The serialized format is:
|
|
//
|
|
// <block hash><block height><total txns><work sum length><work sum>
|
|
//
|
|
// Field Type Size
|
|
// block hash chainhash.Hash chainhash.HashSize
|
|
// block height uint32 4 bytes
|
|
// total txns uint64 8 bytes
|
|
// work sum length uint32 4 bytes
|
|
// work sum big.Int work sum length
|
|
// -----------------------------------------------------------------------------
|
|
|
|
// bestChainState represents the data to be stored the database for the current
|
|
// best chain state.
|
|
type bestChainState struct {
|
|
hash chainhash.Hash
|
|
height uint32
|
|
totalTxns uint64
|
|
workSum *big.Int
|
|
}
|
|
|
|
// serializeBestChainState returns the serialization of the passed block best
|
|
// chain state. This is data to be stored in the chain state bucket.
|
|
func serializeBestChainState(state bestChainState) []byte {
|
|
// Calculate the full size needed to serialize the chain state.
|
|
workSumBytes := state.workSum.Bytes()
|
|
workSumBytesLen := uint32(len(workSumBytes))
|
|
serializedLen := chainhash.HashSize + 4 + 8 + 4 + workSumBytesLen
|
|
|
|
// Serialize the chain state.
|
|
serializedData := make([]byte, serializedLen)
|
|
copy(serializedData[0:chainhash.HashSize], state.hash[:])
|
|
offset := uint32(chainhash.HashSize)
|
|
byteOrder.PutUint32(serializedData[offset:], state.height)
|
|
offset += 4
|
|
byteOrder.PutUint64(serializedData[offset:], state.totalTxns)
|
|
offset += 8
|
|
byteOrder.PutUint32(serializedData[offset:], workSumBytesLen)
|
|
offset += 4
|
|
copy(serializedData[offset:], workSumBytes)
|
|
return serializedData[:]
|
|
}
|
|
|
|
// deserializeBestChainState deserializes the passed serialized best chain
|
|
// state. This is data stored in the chain state bucket and is updated after
|
|
// every block is connected or disconnected form the main chain.
|
|
// block.
|
|
func deserializeBestChainState(serializedData []byte) (bestChainState, error) {
|
|
// Ensure the serialized data has enough bytes to properly deserialize
|
|
// the hash, height, total transactions, and work sum length.
|
|
if len(serializedData) < chainhash.HashSize+16 {
|
|
return bestChainState{}, database.Error{
|
|
ErrorCode: database.ErrCorruption,
|
|
Description: "corrupt best chain state",
|
|
}
|
|
}
|
|
|
|
state := bestChainState{}
|
|
copy(state.hash[:], serializedData[0:chainhash.HashSize])
|
|
offset := uint32(chainhash.HashSize)
|
|
state.height = byteOrder.Uint32(serializedData[offset : offset+4])
|
|
offset += 4
|
|
state.totalTxns = byteOrder.Uint64(serializedData[offset : offset+8])
|
|
offset += 8
|
|
workSumBytesLen := byteOrder.Uint32(serializedData[offset : offset+4])
|
|
offset += 4
|
|
|
|
// Ensure the serialized data has enough bytes to deserialize the work
|
|
// sum.
|
|
if uint32(len(serializedData[offset:])) < workSumBytesLen {
|
|
return bestChainState{}, database.Error{
|
|
ErrorCode: database.ErrCorruption,
|
|
Description: "corrupt best chain state",
|
|
}
|
|
}
|
|
workSumBytes := serializedData[offset : offset+workSumBytesLen]
|
|
state.workSum = new(big.Int).SetBytes(workSumBytes)
|
|
|
|
return state, nil
|
|
}
|
|
|
|
// dbPutBestState uses an existing database transaction to update the best chain
|
|
// state with the given parameters.
|
|
func dbPutBestState(dbTx database.Tx, snapshot *BestState, workSum *big.Int) error {
|
|
// Serialize the current best chain state.
|
|
serializedData := serializeBestChainState(bestChainState{
|
|
hash: *snapshot.Hash,
|
|
height: uint32(snapshot.Height),
|
|
totalTxns: snapshot.TotalTxns,
|
|
workSum: workSum,
|
|
})
|
|
|
|
// Store the current best chain state into the database.
|
|
return dbTx.Metadata().Put(chainStateKeyName, serializedData)
|
|
}
|
|
|
|
// createChainState initializes both the database and the chain state to the
|
|
// genesis block. This includes creating the necessary buckets and inserting
|
|
// the genesis block, so it must only be called on an uninitialized database.
|
|
func (b *BlockChain) createChainState() error {
|
|
// Create a new node from the genesis block and set it as the best node.
|
|
genesisBlock := btcutil.NewBlock(b.chainParams.GenesisBlock)
|
|
header := &genesisBlock.MsgBlock().Header
|
|
node := newBlockNode(header, genesisBlock.Hash(), 0)
|
|
node.inMainChain = true
|
|
b.bestNode = node
|
|
|
|
// Add the new node to the index which is used for faster lookups.
|
|
b.index[*node.hash] = node
|
|
|
|
// Initialize the state related to the best block. Since it is the
|
|
// genesis block, use its timestamp for the median time.
|
|
numTxns := uint64(len(genesisBlock.MsgBlock().Transactions))
|
|
blockSize := uint64(genesisBlock.MsgBlock().SerializeSize())
|
|
b.stateSnapshot = newBestState(b.bestNode, blockSize, numTxns, numTxns,
|
|
time.Unix(b.bestNode.timestamp, 0))
|
|
|
|
// Create the initial the database chain state including creating the
|
|
// necessary index buckets and inserting the genesis block.
|
|
err := b.db.Update(func(dbTx database.Tx) error {
|
|
// Create the bucket that houses the chain block hash to height
|
|
// index.
|
|
meta := dbTx.Metadata()
|
|
_, err := meta.CreateBucket(hashIndexBucketName)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Create the bucket that houses the chain block height to hash
|
|
// index.
|
|
_, err = meta.CreateBucket(heightIndexBucketName)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Create the bucket that houses the spend journal data.
|
|
_, err = meta.CreateBucket(spendJournalBucketName)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Create the bucket that houses the utxo set. Note that the
|
|
// genesis block coinbase transaction is intentionally not
|
|
// inserted here since it is not spendable by consensus rules.
|
|
_, err = meta.CreateBucket(utxoSetBucketName)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Add the genesis block hash to height and height to hash
|
|
// mappings to the index.
|
|
err = dbPutBlockIndex(dbTx, b.bestNode.hash, b.bestNode.height)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Store the current best chain state into the database.
|
|
err = dbPutBestState(dbTx, b.stateSnapshot, b.bestNode.workSum)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Store the genesis block into the database.
|
|
return dbTx.StoreBlock(genesisBlock)
|
|
})
|
|
return err
|
|
}
|
|
|
|
// initChainState attempts to load and initialize the chain state from the
|
|
// database. When the db does not yet contain any chain state, both it and the
|
|
// chain state are initialized to the genesis block.
|
|
func (b *BlockChain) initChainState() error {
|
|
// Attempt to load the chain state from the database.
|
|
var isStateInitialized bool
|
|
err := b.db.View(func(dbTx database.Tx) error {
|
|
// Fetch the stored chain state from the database metadata.
|
|
// When it doesn't exist, it means the database hasn't been
|
|
// initialized for use with chain yet, so break out now to allow
|
|
// that to happen under a writable database transaction.
|
|
serializedData := dbTx.Metadata().Get(chainStateKeyName)
|
|
if serializedData == nil {
|
|
return nil
|
|
}
|
|
log.Tracef("Serialized chain state: %x", serializedData)
|
|
state, err := deserializeBestChainState(serializedData)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Load the raw block bytes for the best block.
|
|
blockBytes, err := dbTx.FetchBlock(&state.hash)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
var block wire.MsgBlock
|
|
err = block.Deserialize(bytes.NewReader(blockBytes))
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Create a new node and set it as the best node. The preceding
|
|
// nodes will be loaded on demand as needed.
|
|
header := &block.Header
|
|
node := newBlockNode(header, &state.hash, int32(state.height))
|
|
node.inMainChain = true
|
|
node.workSum = state.workSum
|
|
b.bestNode = node
|
|
|
|
// Add the new node to the indices for faster lookups.
|
|
prevHash := node.parentHash
|
|
b.index[*node.hash] = node
|
|
b.depNodes[*prevHash] = append(b.depNodes[*prevHash], node)
|
|
|
|
// Calculate the median time for the block.
|
|
medianTime, err := b.calcPastMedianTime(node)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Initialize the state related to the best block.
|
|
blockSize := uint64(len(blockBytes))
|
|
numTxns := uint64(len(block.Transactions))
|
|
b.stateSnapshot = newBestState(b.bestNode, blockSize, numTxns,
|
|
state.totalTxns, medianTime)
|
|
|
|
isStateInitialized = true
|
|
return nil
|
|
})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// There is nothing more to do if the chain state was initialized.
|
|
if isStateInitialized {
|
|
return nil
|
|
}
|
|
|
|
// At this point the database has not already been initialized, so
|
|
// initialize both it and the chain state to the genesis block.
|
|
return b.createChainState()
|
|
}
|
|
|
|
// dbFetchHeaderByHash uses an existing database transaction to retrieve the
|
|
// block header for the provided hash.
|
|
func dbFetchHeaderByHash(dbTx database.Tx, hash *chainhash.Hash) (*wire.BlockHeader, error) {
|
|
headerBytes, err := dbTx.FetchBlockHeader(hash)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
var header wire.BlockHeader
|
|
err = header.Deserialize(bytes.NewReader(headerBytes))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return &header, nil
|
|
}
|
|
|
|
// dbFetchHeaderByHeight uses an existing database transaction to retrieve the
|
|
// block header for the provided height.
|
|
func dbFetchHeaderByHeight(dbTx database.Tx, height int32) (*wire.BlockHeader, error) {
|
|
hash, err := dbFetchHashByHeight(dbTx, height)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return dbFetchHeaderByHash(dbTx, hash)
|
|
}
|
|
|
|
// dbFetchBlockByHash uses an existing database transaction to retrieve the raw
|
|
// block for the provided hash, deserialize it, retrieve the appropriate height
|
|
// from the index, and return a btcutil.Block with the height set.
|
|
func dbFetchBlockByHash(dbTx database.Tx, hash *chainhash.Hash) (*btcutil.Block, error) {
|
|
// First find the height associated with the provided hash in the index.
|
|
blockHeight, err := dbFetchHeightByHash(dbTx, hash)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Load the raw block bytes from the database.
|
|
blockBytes, err := dbTx.FetchBlock(hash)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Create the encapsulated block and set the height appropriately.
|
|
block, err := btcutil.NewBlockFromBytes(blockBytes)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
block.SetHeight(blockHeight)
|
|
|
|
return block, nil
|
|
}
|
|
|
|
// dbFetchBlockByHeight uses an existing database transaction to retrieve the
|
|
// raw block for the provided height, deserialize it, and return a btcutil.Block
|
|
// with the height set.
|
|
func dbFetchBlockByHeight(dbTx database.Tx, height int32) (*btcutil.Block, error) {
|
|
// First find the hash associated with the provided height in the index.
|
|
hash, err := dbFetchHashByHeight(dbTx, height)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Load the raw block bytes from the database.
|
|
blockBytes, err := dbTx.FetchBlock(hash)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Create the encapsulated block and set the height appropriately.
|
|
block, err := btcutil.NewBlockFromBytes(blockBytes)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
block.SetHeight(height)
|
|
|
|
return block, nil
|
|
}
|
|
|
|
// dbMainChainHasBlock uses an existing database transaction to return whether
|
|
// or not the main chain contains the block identified by the provided hash.
|
|
func dbMainChainHasBlock(dbTx database.Tx, hash *chainhash.Hash) bool {
|
|
hashIndex := dbTx.Metadata().Bucket(hashIndexBucketName)
|
|
return hashIndex.Get(hash[:]) != nil
|
|
}
|
|
|
|
// MainChainHasBlock returns whether or not the block with the given hash is in
|
|
// the main chain.
|
|
//
|
|
// This function is safe for concurrent access.
|
|
func (b *BlockChain) MainChainHasBlock(hash *chainhash.Hash) (bool, error) {
|
|
var exists bool
|
|
err := b.db.View(func(dbTx database.Tx) error {
|
|
exists = dbMainChainHasBlock(dbTx, hash)
|
|
return nil
|
|
})
|
|
return exists, err
|
|
}
|
|
|
|
// BlockHeightByHash returns the height of the block with the given hash in the
|
|
// main chain.
|
|
//
|
|
// This function is safe for concurrent access.
|
|
func (b *BlockChain) BlockHeightByHash(hash *chainhash.Hash) (int32, error) {
|
|
var height int32
|
|
err := b.db.View(func(dbTx database.Tx) error {
|
|
var err error
|
|
height, err = dbFetchHeightByHash(dbTx, hash)
|
|
return err
|
|
})
|
|
return height, err
|
|
}
|
|
|
|
// BlockHashByHeight returns the hash of the block at the given height in the
|
|
// main chain.
|
|
//
|
|
// This function is safe for concurrent access.
|
|
func (b *BlockChain) BlockHashByHeight(blockHeight int32) (*chainhash.Hash, error) {
|
|
var hash *chainhash.Hash
|
|
err := b.db.View(func(dbTx database.Tx) error {
|
|
var err error
|
|
hash, err = dbFetchHashByHeight(dbTx, blockHeight)
|
|
return err
|
|
})
|
|
return hash, err
|
|
}
|
|
|
|
// BlockByHeight returns the block at the given height in the main chain.
|
|
//
|
|
// This function is safe for concurrent access.
|
|
func (b *BlockChain) BlockByHeight(blockHeight int32) (*btcutil.Block, error) {
|
|
var block *btcutil.Block
|
|
err := b.db.View(func(dbTx database.Tx) error {
|
|
var err error
|
|
block, err = dbFetchBlockByHeight(dbTx, blockHeight)
|
|
return err
|
|
})
|
|
return block, err
|
|
}
|
|
|
|
// BlockByHash returns the block from the main chain with the given hash with
|
|
// the appropriate chain height set.
|
|
//
|
|
// This function is safe for concurrent access.
|
|
func (b *BlockChain) BlockByHash(hash *chainhash.Hash) (*btcutil.Block, error) {
|
|
var block *btcutil.Block
|
|
err := b.db.View(func(dbTx database.Tx) error {
|
|
var err error
|
|
block, err = dbFetchBlockByHash(dbTx, hash)
|
|
return err
|
|
})
|
|
return block, err
|
|
}
|
|
|
|
// HeightRange returns a range of block hashes for the given start and end
|
|
// heights. It is inclusive of the start height and exclusive of the end
|
|
// height. The end height will be limited to the current main chain height.
|
|
//
|
|
// This function is safe for concurrent access.
|
|
func (b *BlockChain) HeightRange(startHeight, endHeight int32) ([]chainhash.Hash, error) {
|
|
// Ensure requested heights are sane.
|
|
if startHeight < 0 {
|
|
return nil, fmt.Errorf("start height of fetch range must not "+
|
|
"be less than zero - got %d", startHeight)
|
|
}
|
|
if endHeight < startHeight {
|
|
return nil, fmt.Errorf("end height of fetch range must not "+
|
|
"be less than the start height - got start %d, end %d",
|
|
startHeight, endHeight)
|
|
}
|
|
|
|
// There is nothing to do when the start and end heights are the same,
|
|
// so return now to avoid the chain lock and a database transaction.
|
|
if startHeight == endHeight {
|
|
return nil, nil
|
|
}
|
|
|
|
// Grab a lock on the chain to prevent it from changing due to a reorg
|
|
// while building the hashes.
|
|
b.chainLock.RLock()
|
|
defer b.chainLock.RUnlock()
|
|
|
|
// When the requested start height is after the most recent best chain
|
|
// height, there is nothing to do.
|
|
latestHeight := b.bestNode.height
|
|
if startHeight > latestHeight {
|
|
return nil, nil
|
|
}
|
|
|
|
// Limit the ending height to the latest height of the chain.
|
|
if endHeight > latestHeight+1 {
|
|
endHeight = latestHeight + 1
|
|
}
|
|
|
|
// Fetch as many as are available within the specified range.
|
|
var hashList []chainhash.Hash
|
|
err := b.db.View(func(dbTx database.Tx) error {
|
|
hashes := make([]chainhash.Hash, 0, endHeight-startHeight)
|
|
for i := startHeight; i < endHeight; i++ {
|
|
hash, err := dbFetchHashByHeight(dbTx, i)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
hashes = append(hashes, *hash)
|
|
}
|
|
|
|
// Set the list to be returned to the constructed list.
|
|
hashList = hashes
|
|
return nil
|
|
})
|
|
return hashList, err
|
|
}
|
|
|
|
// -----------------------------------------------------------------------------
|
|
// The threshold state consists of individual threshold cache buckets for each
|
|
// cache id under one main threshold state bucket. Each threshold cache bucket
|
|
// contains entries keyed by the block hash for the final block in each window
|
|
// and their associated threshold states as well as the associated deployment
|
|
// parameters.
|
|
//
|
|
// The serialized value format is for each cache entry keyed by hash is:
|
|
//
|
|
// <thresholdstate>
|
|
//
|
|
// Field Type Size
|
|
// threshold state uint8 1 byte
|
|
//
|
|
//
|
|
// In addition, the threshold cache buckets for deployments contain the specific
|
|
// deployment parameters they were created with. This allows the cache
|
|
// invalidation when there any changes to their definitions.
|
|
//
|
|
// The serialized value format for the deployment parameters is:
|
|
//
|
|
// <bit number><start time><expire time>
|
|
//
|
|
// Field Type Size
|
|
// bit number uint8 1 byte
|
|
// start time uint64 8 bytes
|
|
// expire time uint64 8 bytes
|
|
//
|
|
//
|
|
// Finally, the main threshold bucket also contains the number of stored
|
|
// deployment buckets as described above.
|
|
//
|
|
// The serialized value format for the number of stored deployment buckets is:
|
|
//
|
|
// <num deployments>
|
|
//
|
|
// Field Type Size
|
|
// num deployments uint32 4 bytes
|
|
// -----------------------------------------------------------------------------
|
|
|
|
// serializeDeploymentCacheParams serializes the parameters for the passed
|
|
// deployment into a single byte slice according to the format described in
|
|
// detail above.
|
|
func serializeDeploymentCacheParams(deployment *chaincfg.ConsensusDeployment) []byte {
|
|
serialized := make([]byte, 1+8+8)
|
|
serialized[0] = deployment.BitNumber
|
|
byteOrder.PutUint64(serialized[1:], deployment.StartTime)
|
|
byteOrder.PutUint64(serialized[9:], deployment.ExpireTime)
|
|
return serialized
|
|
}
|
|
|
|
// deserializeDeploymentCacheParams deserializes the passed serialized
|
|
// deployment cache parameters into a deployment struct.
|
|
func deserializeDeploymentCacheParams(serialized []byte) (chaincfg.ConsensusDeployment, error) {
|
|
// Ensure the serialized data has enough bytes to properly deserialize
|
|
// the bit number, start time, and expire time.
|
|
if len(serialized) != 1+8+8 {
|
|
return chaincfg.ConsensusDeployment{}, database.Error{
|
|
ErrorCode: database.ErrCorruption,
|
|
Description: "corrupt deployment cache state",
|
|
}
|
|
}
|
|
|
|
var deployment chaincfg.ConsensusDeployment
|
|
deployment.BitNumber = serialized[0]
|
|
deployment.StartTime = byteOrder.Uint64(serialized[1:])
|
|
deployment.ExpireTime = byteOrder.Uint64(serialized[9:])
|
|
return deployment, nil
|
|
}
|
|
|
|
// dbPutDeploymentCacheParams uses an existing database transaction to update
|
|
// the deployment cache params with the given values.
|
|
func dbPutDeploymentCacheParams(bucket database.Bucket, deployment *chaincfg.ConsensusDeployment) error {
|
|
serialized := serializeDeploymentCacheParams(deployment)
|
|
return bucket.Put(deploymentStateKeyName, serialized)
|
|
}
|
|
|
|
// dbFetchDeploymentCacheParams uses an existing database transaction to
|
|
// retrieve the deployment parameters from the given bucket, deserialize them,
|
|
// and returns the resulting deployment struct.
|
|
func dbFetchDeploymentCacheParams(bucket database.Bucket) (chaincfg.ConsensusDeployment, error) {
|
|
serialized := bucket.Get(deploymentStateKeyName)
|
|
return deserializeDeploymentCacheParams(serialized)
|
|
}
|
|
|
|
// serializeNumDeployments serializes the parameters for the passed number of
|
|
// deployments into a single byte slice according to the format described in
|
|
// detail above.
|
|
func serializeNumDeployments(numDeployments uint32) []byte {
|
|
serialized := make([]byte, 4)
|
|
byteOrder.PutUint32(serialized, numDeployments)
|
|
return serialized
|
|
}
|
|
|
|
// deserializeDeploymentCacheParams deserializes the passed serialized
|
|
// number of deployments.
|
|
func deserializeNumDeployments(serialized []byte) (uint32, error) {
|
|
if len(serialized) != 4 {
|
|
return 0, database.Error{
|
|
ErrorCode: database.ErrCorruption,
|
|
Description: "corrupt stored number of deployments",
|
|
}
|
|
}
|
|
return byteOrder.Uint32(serialized), nil
|
|
}
|
|
|
|
// dbPutNumDeployments uses an existing database transaction to update the
|
|
// number of deployments to the given value.
|
|
func dbPutNumDeployments(bucket database.Bucket, numDeployments uint32) error {
|
|
serialized := serializeNumDeployments(numDeployments)
|
|
return bucket.Put(numDeploymentsKeyName, serialized)
|
|
}
|
|
|
|
// dbFetchNumDeployments uses an existing database transaction to retrieve the
|
|
// number of deployments, deserialize it, and returns the result.
|
|
func dbFetchNumDeployments(bucket database.Bucket) (uint32, error) {
|
|
// Ensure the serialized data has enough bytes to properly deserialize
|
|
// the number of stored deployments.
|
|
serialized := bucket.Get(numDeploymentsKeyName)
|
|
return deserializeNumDeployments(serialized)
|
|
}
|
|
|
|
// thresholdCacheBucket returns the serialized bucket name to use for a
|
|
// threshold cache given a prefix and an ID.
|
|
func thresholdCacheBucket(prefix []byte, id uint32) []byte {
|
|
bucketName := make([]byte, len(prefix)+4)
|
|
copy(bucketName, prefix)
|
|
byteOrder.PutUint32(bucketName[len(bucketName)-4:], id)
|
|
return bucketName
|
|
}
|
|
|
|
// dbPutThresholdState uses an existing database transaction to update or add
|
|
// the rule change threshold state for the provided block hash.
|
|
func dbPutThresholdState(bucket database.Bucket, hash chainhash.Hash, state ThresholdState) error {
|
|
// Add the block hash to threshold state mapping.
|
|
var serializedState [1]byte
|
|
serializedState[0] = byte(state)
|
|
return bucket.Put(hash[:], serializedState[:])
|
|
}
|
|
|
|
// dbPutThresholdCaches uses an existing database transaction to update the
|
|
// provided threshold state caches using the given bucket prefix.
|
|
func dbPutThresholdCaches(dbTx database.Tx, caches []thresholdStateCache, bucketPrefix []byte) error {
|
|
// Loop through each of the defined cache IDs in the provided cache and
|
|
// populate the associated bucket with all of the block hash to
|
|
// threshold state mappings for it.
|
|
cachesBucket := dbTx.Metadata().Bucket(thresholdBucketName)
|
|
for i := uint32(0); i < uint32(len(caches)); i++ {
|
|
cache := &caches[i]
|
|
if len(cache.dbUpdates) == 0 {
|
|
continue
|
|
}
|
|
|
|
cacheIDBucketName := thresholdCacheBucket(bucketPrefix, i)
|
|
bucket := cachesBucket.Bucket(cacheIDBucketName)
|
|
for blockHash, state := range cache.dbUpdates {
|
|
err := dbPutThresholdState(bucket, blockHash, state)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// putThresholdCaches uses an existing database transaction to update the
|
|
// threshold state caches.
|
|
func (b *BlockChain) putThresholdCaches(dbTx database.Tx) error {
|
|
err := dbPutThresholdCaches(dbTx, b.deploymentCaches,
|
|
deploymentBucketName)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
return dbPutThresholdCaches(dbTx, b.warningCaches, warningBucketName)
|
|
}
|
|
|
|
// markThresholdCachesFlushed clears any pending updates to be written from
|
|
// threshold state caches. Callers are intended to call this after the pending
|
|
// updates have been successfully written to the database via the
|
|
// putThresholdCaches function and its associated database transation is closed.
|
|
// This approach is taken to ensure the memory state is not updated until after
|
|
// the atomic database update was successful.
|
|
func (b *BlockChain) markThresholdCachesFlushed() {
|
|
for i := 0; i < len(b.deploymentCaches); i++ {
|
|
b.deploymentCaches[i].MarkFlushed()
|
|
}
|
|
for i := 0; i < len(b.warningCaches); i++ {
|
|
b.warningCaches[i].MarkFlushed()
|
|
}
|
|
}
|
|
|
|
// dbFetchThresholdCaches uses an existing database transaction to retrieve
|
|
// the threshold state caches from the provided bucket prefix into the given
|
|
// cache parameter. When the db does not contain any information for a specific
|
|
// id within that cache, that entry will simply be empty.
|
|
func dbFetchThresholdCaches(dbTx database.Tx, caches []thresholdStateCache, bucketPrefix []byte) error {
|
|
// Nothing to load if the main threshold state caches bucket
|
|
// doesn't exist.
|
|
cachesBucket := dbTx.Metadata().Bucket(thresholdBucketName)
|
|
if cachesBucket == nil {
|
|
return nil
|
|
}
|
|
|
|
// Loop through each of the cache IDs and load any saved threshold
|
|
// states.
|
|
for i := 0; i < len(caches); i++ {
|
|
// Nothing to do for this cache ID if there is no bucket for it.
|
|
cacheIDBucketName := thresholdCacheBucket(bucketPrefix, uint32(i))
|
|
cacheIDBucket := cachesBucket.Bucket(cacheIDBucketName[:])
|
|
if cacheIDBucket == nil {
|
|
continue
|
|
}
|
|
|
|
// Load all of the cached block hash to threshold state mappings
|
|
// from the bucket.
|
|
err := cacheIDBucket.ForEach(func(k, v []byte) error {
|
|
// Skip non-hash entries.
|
|
if len(k) != chainhash.HashSize {
|
|
return nil
|
|
}
|
|
|
|
var hash chainhash.Hash
|
|
copy(hash[:], k)
|
|
caches[i].entries[hash] = ThresholdState(v[0])
|
|
return nil
|
|
})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// invalidateThresholdCaches removes any threshold state caches that are no
|
|
// longer valid. This can happen if a deployment ID is changed such as when it
|
|
// is reused, or if it is reordered in the parameter definitions. It is also
|
|
// necessary for specific bits in the warning cache when deployment definitions
|
|
// are added and removed since it could change the expected block versions and
|
|
// hence potentially change the result of the warning states for that bit.
|
|
func (b *BlockChain) invalidateThresholdCaches(cachesBucket database.Bucket) error {
|
|
deployments := b.chainParams.Deployments[:]
|
|
|
|
// Remove any stored deployments that are no longer defined along with
|
|
// the warning cache associated with their bits.
|
|
numStoredDeployments, err := dbFetchNumDeployments(cachesBucket)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
definedDeployments := uint32(len(deployments))
|
|
for i := definedDeployments; i < numStoredDeployments; i++ {
|
|
// Nothing to do when nothing is stored for the deployment.
|
|
deployBucketKey := thresholdCacheBucket(deploymentBucketName, i)
|
|
deployBucket := cachesBucket.Bucket(deployBucketKey)
|
|
if deployBucket == nil {
|
|
continue
|
|
}
|
|
|
|
// Load the deployment details the cache was created for from
|
|
// the database.
|
|
stored, err := dbFetchDeploymentCacheParams(deployBucket)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Remove the warning cache for the bit associated with the old
|
|
// deployment definition.
|
|
oldBit := uint32(stored.BitNumber)
|
|
bn := thresholdCacheBucket(warningBucketName, oldBit)
|
|
err = cachesBucket.DeleteBucket(bn)
|
|
if err != nil && !isDbBucketNotFoundErr(err) {
|
|
return err
|
|
}
|
|
|
|
// Remove deployment state and cache.
|
|
err = cachesBucket.DeleteBucket(deployBucketKey)
|
|
if err != nil && !isDbBucketNotFoundErr(err) {
|
|
return err
|
|
}
|
|
log.Debugf("Removed threshold state caches for deployment %d "+
|
|
"and warning bit %d", i, oldBit)
|
|
}
|
|
|
|
// Remove any deployment caches that no longer match the associated
|
|
// deployment definition.
|
|
for i := uint32(0); i < uint32(len(deployments)); i++ {
|
|
// Remove the warning cache for the bit associated with the new
|
|
// deployment definition if nothing is already stored for the
|
|
// deployment.
|
|
deployBucketKey := thresholdCacheBucket(deploymentBucketName, i)
|
|
deployBucket := cachesBucket.Bucket(deployBucketKey)
|
|
if deployBucket == nil {
|
|
// Remove the warning cache for the bit associated with
|
|
// the new deployment definition.
|
|
newBit := uint32(deployments[i].BitNumber)
|
|
bn := thresholdCacheBucket(warningBucketName, newBit)
|
|
err = cachesBucket.DeleteBucket(bn)
|
|
if err != nil && !isDbBucketNotFoundErr(err) {
|
|
return err
|
|
}
|
|
log.Debugf("Removed threshold state cache for warning "+
|
|
"bit %d ", newBit)
|
|
continue
|
|
}
|
|
|
|
// Load the deployment details the cache was created for from
|
|
// the database, compare them against the currently defined
|
|
// deployment, and invalidate the relevant caches if they don't
|
|
// match.
|
|
stored, err := dbFetchDeploymentCacheParams(deployBucket)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if stored != deployments[i] {
|
|
// Remove deployment state and cache.
|
|
err := cachesBucket.DeleteBucket(deployBucketKey)
|
|
if err != nil && !isDbBucketNotFoundErr(err) {
|
|
return err
|
|
}
|
|
|
|
// Remove the warning cache for the bit associated with
|
|
// the new deployment definition.
|
|
newBit := uint32(deployments[i].BitNumber)
|
|
bn := thresholdCacheBucket(warningBucketName, newBit)
|
|
err = cachesBucket.DeleteBucket(bn)
|
|
if err != nil && !isDbBucketNotFoundErr(err) {
|
|
return err
|
|
}
|
|
|
|
// Remove the warning cache for the bit associated with
|
|
// the old deployment definition if it is different than
|
|
// the new one.
|
|
oldBit := uint32(stored.BitNumber)
|
|
if oldBit == newBit {
|
|
log.Debugf("Removed threshold state caches for "+
|
|
"deployment %d and warning bit %d", i,
|
|
newBit)
|
|
continue
|
|
}
|
|
bn = thresholdCacheBucket(warningBucketName, oldBit)
|
|
err = cachesBucket.DeleteBucket(bn)
|
|
if err != nil && !isDbBucketNotFoundErr(err) {
|
|
return err
|
|
}
|
|
log.Debugf("Removed threshold state caches for "+
|
|
"deployment %d and warning bits %d and %d", i,
|
|
oldBit, newBit)
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// initThresholdCacheBuckets creates any missing buckets needed for the defined
|
|
// threshold caches and populates them with state-related details so they can
|
|
// be invalidated as needed.
|
|
func (b *BlockChain) initThresholdCacheBuckets(meta database.Bucket) error {
|
|
// Create overall bucket that houses all of the threshold caches and
|
|
// their related state as needed.
|
|
cachesBucket, err := meta.CreateBucketIfNotExists(thresholdBucketName)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Update the number of stored deployment as needed.
|
|
definedDeployments := uint32(len(b.deploymentCaches))
|
|
storedDeployments, err := dbFetchNumDeployments(cachesBucket)
|
|
if err != nil || storedDeployments != definedDeployments {
|
|
err := dbPutNumDeployments(cachesBucket, definedDeployments)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
// Create buckets for each of the deployment caches as needed, and
|
|
// populate the created buckets with the specific deployment details so
|
|
// that the cache(s) can be invalidated properly with future updates.
|
|
for i := uint32(0); i < definedDeployments; i++ {
|
|
name := thresholdCacheBucket(deploymentBucketName, i)
|
|
if bucket := cachesBucket.Bucket(name); bucket != nil {
|
|
continue
|
|
}
|
|
|
|
deployBucket, err := cachesBucket.CreateBucket(name)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
deployment := &b.chainParams.Deployments[i]
|
|
err = dbPutDeploymentCacheParams(deployBucket, deployment)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
// Create buckets for each of the warning caches as needed.
|
|
for i := uint32(0); i < uint32(len(b.warningCaches)); i++ {
|
|
name := thresholdCacheBucket(warningBucketName, i)
|
|
_, err := cachesBucket.CreateBucketIfNotExists(name)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// initThresholdCaches initializes the threshold state caches from the database.
|
|
// When the db does not yet contain any information for a specific threshold
|
|
// cache or a given id within that cache, it will simply be empty which will
|
|
// lead to it being calculated as needed.
|
|
func (b *BlockChain) initThresholdCaches() error {
|
|
// Create and initialize missing threshold state cache buckets and
|
|
// remove any that are no longer valid.
|
|
err := b.db.Update(func(dbTx database.Tx) error {
|
|
meta := dbTx.Metadata()
|
|
cachesBucket := meta.Bucket(thresholdBucketName)
|
|
if cachesBucket != nil {
|
|
err := b.invalidateThresholdCaches(cachesBucket)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
// Create all cache buckets as needed.
|
|
return b.initThresholdCacheBuckets(meta)
|
|
})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Load the deployment caches.
|
|
err = b.db.View(func(dbTx database.Tx) error {
|
|
// Load the deployment threshold states.
|
|
err := dbFetchThresholdCaches(dbTx, b.deploymentCaches,
|
|
deploymentBucketName)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Load the warning threshold states.
|
|
return dbFetchThresholdCaches(dbTx, b.warningCaches,
|
|
warningBucketName)
|
|
})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Inform the user the states might take a while to recalculate if any
|
|
// of the threshold state caches aren't populated.
|
|
var showMsg bool
|
|
for i := 0; i < len(b.warningCaches); i++ {
|
|
if len(b.warningCaches[i].entries) == 0 {
|
|
showMsg = true
|
|
break
|
|
}
|
|
}
|
|
if !showMsg {
|
|
for i := 0; i < len(b.deploymentCaches); i++ {
|
|
if len(b.deploymentCaches[i].entries) == 0 {
|
|
showMsg = true
|
|
break
|
|
}
|
|
}
|
|
}
|
|
if showMsg {
|
|
log.Info("Recalculating threshold states due to definition " +
|
|
"change. This might take a while...")
|
|
}
|
|
|
|
// Get the previous block node. This function is used over simply
|
|
// accessing b.bestNode.parent directly as it will dynamically create
|
|
// previous block nodes as needed. This helps allow only the pieces of
|
|
// the chain that are needed to remain in memory.
|
|
prevNode, err := b.getPrevNodeFromNode(b.bestNode)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Initialize the warning and deployment caches by calculating the
|
|
// threshold state for each of them. This will ensure the caches are
|
|
// populated and any states that needed to be recalculated due to
|
|
// definition changes is done now.
|
|
for bit := uint32(0); bit < vbNumBits; bit++ {
|
|
checker := bitConditionChecker{bit: bit, chain: b}
|
|
cache := &b.warningCaches[bit]
|
|
_, err := b.thresholdState(prevNode, checker, cache)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
for id := 0; id < len(b.chainParams.Deployments); id++ {
|
|
deployment := &b.chainParams.Deployments[id]
|
|
cache := &b.deploymentCaches[id]
|
|
checker := deploymentChecker{deployment: deployment, chain: b}
|
|
_, err := b.thresholdState(prevNode, checker, cache)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
// No warnings about unknown rules or versions until the chain is
|
|
// current.
|
|
if b.isCurrent() {
|
|
// Warn if a high enough percentage of the last blocks have
|
|
// unexpected versions.
|
|
if err := b.warnUnknownVersions(b.bestNode); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Warn if any unknown new rules are either about to activate or
|
|
// have already been activated.
|
|
if err := b.warnUnknownRuleActivations(b.bestNode); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
// Update the cached threshold states in the database as needed.
|
|
err = b.db.Update(func(dbTx database.Tx) error {
|
|
return b.putThresholdCaches(dbTx)
|
|
})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Mark all modified entries in the threshold caches as flushed now that
|
|
// they have been committed to the database.
|
|
b.markThresholdCachesFlushed()
|
|
|
|
return nil
|
|
}
|