2016-02-03 18:42:04 +01:00
|
|
|
// Copyright (c) 2015-2016 The btcsuite developers
|
|
|
|
// Use of this source code is governed by an ISC
|
|
|
|
// license that can be found in the LICENSE file.
|
|
|
|
|
|
|
|
// This file is part of the ffldb package rather than the ffldb_test package as
|
|
|
|
// it provides whitebox testing.
|
|
|
|
|
|
|
|
package ffldb
|
|
|
|
|
|
|
|
import (
|
|
|
|
"compress/bzip2"
|
|
|
|
"encoding/binary"
|
|
|
|
"fmt"
|
|
|
|
"hash/crc32"
|
|
|
|
"io"
|
|
|
|
"os"
|
|
|
|
"path/filepath"
|
|
|
|
"testing"
|
|
|
|
|
|
|
|
"github.com/btcsuite/goleveldb/leveldb"
|
|
|
|
ldberrors "github.com/btcsuite/goleveldb/leveldb/errors"
|
2021-10-15 07:45:32 +02:00
|
|
|
"github.com/lbryio/lbcd/database"
|
|
|
|
"github.com/lbryio/lbcd/wire"
|
|
|
|
btcutil "github.com/lbryio/lbcutil"
|
2016-02-03 18:42:04 +01:00
|
|
|
)
|
|
|
|
|
|
|
|
var (
|
|
|
|
// blockDataNet is the expected network in the test block data.
|
|
|
|
blockDataNet = wire.MainNet
|
|
|
|
|
|
|
|
// blockDataFile is the path to a file containing the first 256 blocks
|
|
|
|
// of the block chain.
|
|
|
|
blockDataFile = filepath.Join("..", "testdata", "blocks1-256.bz2")
|
|
|
|
|
|
|
|
// errSubTestFail is used to signal that a sub test returned false.
|
|
|
|
errSubTestFail = fmt.Errorf("sub test failure")
|
|
|
|
)
|
|
|
|
|
|
|
|
// loadBlocks loads the blocks contained in the testdata directory and returns
|
|
|
|
// a slice of them.
|
|
|
|
func loadBlocks(t *testing.T, dataFile string, network wire.BitcoinNet) ([]*btcutil.Block, error) {
|
|
|
|
// Open the file that contains the blocks for reading.
|
|
|
|
fi, err := os.Open(dataFile)
|
|
|
|
if err != nil {
|
|
|
|
t.Errorf("failed to open file %v, err %v", dataFile, err)
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
defer func() {
|
|
|
|
if err := fi.Close(); err != nil {
|
|
|
|
t.Errorf("failed to close file %v %v", dataFile,
|
|
|
|
err)
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
dr := bzip2.NewReader(fi)
|
|
|
|
|
|
|
|
// Set the first block as the genesis block.
|
|
|
|
blocks := make([]*btcutil.Block, 0, 256)
|
|
|
|
genesis := btcutil.NewBlock(chaincfg.MainNetParams.GenesisBlock)
|
|
|
|
blocks = append(blocks, genesis)
|
|
|
|
|
|
|
|
// Load the remaining blocks.
|
|
|
|
for height := 1; ; height++ {
|
|
|
|
var net uint32
|
|
|
|
err := binary.Read(dr, binary.LittleEndian, &net)
|
|
|
|
if err == io.EOF {
|
|
|
|
// Hit end of file at the expected offset. No error.
|
|
|
|
break
|
|
|
|
}
|
|
|
|
if err != nil {
|
|
|
|
t.Errorf("Failed to load network type for block %d: %v",
|
|
|
|
height, err)
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
if net != uint32(network) {
|
|
|
|
t.Errorf("Block doesn't match network: %v expects %v",
|
|
|
|
net, network)
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
var blockLen uint32
|
|
|
|
err = binary.Read(dr, binary.LittleEndian, &blockLen)
|
|
|
|
if err != nil {
|
|
|
|
t.Errorf("Failed to load block size for block %d: %v",
|
|
|
|
height, err)
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Read the block.
|
|
|
|
blockBytes := make([]byte, blockLen)
|
|
|
|
_, err = io.ReadFull(dr, blockBytes)
|
|
|
|
if err != nil {
|
|
|
|
t.Errorf("Failed to load block %d: %v", height, err)
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Deserialize and store the block.
|
|
|
|
block, err := btcutil.NewBlockFromBytes(blockBytes)
|
|
|
|
if err != nil {
|
|
|
|
t.Errorf("Failed to parse block %v: %v", height, err)
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
blocks = append(blocks, block)
|
|
|
|
}
|
|
|
|
|
|
|
|
return blocks, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// checkDbError ensures the passed error is a database.Error with an error code
|
|
|
|
// that matches the passed error code.
|
|
|
|
func checkDbError(t *testing.T, testName string, gotErr error, wantErrCode database.ErrorCode) bool {
|
|
|
|
dbErr, ok := gotErr.(database.Error)
|
|
|
|
if !ok {
|
|
|
|
t.Errorf("%s: unexpected error type - got %T, want %T",
|
|
|
|
testName, gotErr, database.Error{})
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if dbErr.ErrorCode != wantErrCode {
|
|
|
|
t.Errorf("%s: unexpected error code - got %s (%s), want %s",
|
|
|
|
testName, dbErr.ErrorCode, dbErr.Description,
|
|
|
|
wantErrCode)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testContext is used to store context information about a running test which
|
|
|
|
// is passed into helper functions.
|
|
|
|
type testContext struct {
|
|
|
|
t *testing.T
|
|
|
|
db database.DB
|
|
|
|
files map[uint32]*lockableFile
|
|
|
|
maxFileSizes map[uint32]int64
|
|
|
|
blocks []*btcutil.Block
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestConvertErr ensures the leveldb error to database error conversion works
|
|
|
|
// as expected.
|
|
|
|
func TestConvertErr(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
tests := []struct {
|
|
|
|
err error
|
|
|
|
wantErrCode database.ErrorCode
|
|
|
|
}{
|
|
|
|
{&ldberrors.ErrCorrupted{}, database.ErrCorruption},
|
|
|
|
{leveldb.ErrClosed, database.ErrDbNotOpen},
|
|
|
|
{leveldb.ErrSnapshotReleased, database.ErrTxClosed},
|
|
|
|
{leveldb.ErrIterReleased, database.ErrTxClosed},
|
|
|
|
}
|
|
|
|
|
|
|
|
for i, test := range tests {
|
|
|
|
gotErr := convertErr("test", test.err)
|
|
|
|
if gotErr.ErrorCode != test.wantErrCode {
|
|
|
|
t.Errorf("convertErr #%d unexpected error - got %v, "+
|
|
|
|
"want %v", i, gotErr.ErrorCode, test.wantErrCode)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestCornerCases ensures several corner cases which can happen when opening
|
|
|
|
// a database and/or block files work as expected.
|
|
|
|
func TestCornerCases(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
// Create a file at the datapase path to force the open below to fail.
|
|
|
|
dbPath := filepath.Join(os.TempDir(), "ffldb-errors")
|
|
|
|
_ = os.RemoveAll(dbPath)
|
|
|
|
fi, err := os.Create(dbPath)
|
|
|
|
if err != nil {
|
|
|
|
t.Errorf("os.Create: unexpected error: %v", err)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
fi.Close()
|
|
|
|
|
|
|
|
// Ensure creating a new database fails when a file exists where a
|
|
|
|
// directory is needed.
|
|
|
|
testName := "openDB: fail due to file at target location"
|
|
|
|
wantErrCode := database.ErrDriverSpecific
|
|
|
|
idb, err := openDB(dbPath, blockDataNet, true)
|
|
|
|
if !checkDbError(t, testName, err, wantErrCode) {
|
|
|
|
if err == nil {
|
|
|
|
idb.Close()
|
|
|
|
}
|
|
|
|
_ = os.RemoveAll(dbPath)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// Remove the file and create the database to run tests against. It
|
|
|
|
// should be successful this time.
|
|
|
|
_ = os.RemoveAll(dbPath)
|
|
|
|
idb, err = openDB(dbPath, blockDataNet, true)
|
|
|
|
if err != nil {
|
|
|
|
t.Errorf("openDB: unexpected error: %v", err)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
defer os.RemoveAll(dbPath)
|
|
|
|
defer idb.Close()
|
|
|
|
|
|
|
|
// Ensure attempting to write to a file that can't be created returns
|
|
|
|
// the expected error.
|
|
|
|
testName = "writeBlock: open file failure"
|
|
|
|
filePath := blockFilePath(dbPath, 0)
|
|
|
|
if err := os.Mkdir(filePath, 0755); err != nil {
|
|
|
|
t.Errorf("os.Mkdir: unexpected error: %v", err)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
store := idb.(*db).store
|
|
|
|
_, err = store.writeBlock([]byte{0x00})
|
|
|
|
if !checkDbError(t, testName, err, database.ErrDriverSpecific) {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
_ = os.RemoveAll(filePath)
|
|
|
|
|
2016-02-03 18:41:46 +01:00
|
|
|
// Close the underlying leveldb database out from under the database.
|
|
|
|
ldb := idb.(*db).cache.ldb
|
2016-02-03 18:42:04 +01:00
|
|
|
ldb.Close()
|
|
|
|
|
|
|
|
// Ensure initilization errors in the underlying database work as
|
|
|
|
// expected.
|
|
|
|
testName = "initDB: reinitialization"
|
|
|
|
wantErrCode = database.ErrDbNotOpen
|
|
|
|
err = initDB(ldb)
|
|
|
|
if !checkDbError(t, testName, err, wantErrCode) {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the View handles errors in the underlying leveldb database
|
|
|
|
// properly.
|
|
|
|
testName = "View: underlying leveldb error"
|
|
|
|
wantErrCode = database.ErrDbNotOpen
|
|
|
|
err = idb.View(func(tx database.Tx) error {
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if !checkDbError(t, testName, err, wantErrCode) {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the Update handles errors in the underlying leveldb database
|
|
|
|
// properly.
|
|
|
|
testName = "Update: underlying leveldb error"
|
|
|
|
err = idb.Update(func(tx database.Tx) error {
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if !checkDbError(t, testName, err, wantErrCode) {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// resetDatabase removes everything from the opened database associated with the
|
|
|
|
// test context including all metadata and the mock files.
|
|
|
|
func resetDatabase(tc *testContext) bool {
|
|
|
|
// Reset the metadata.
|
|
|
|
err := tc.db.Update(func(tx database.Tx) error {
|
|
|
|
// Remove all the keys using a cursor while also generating a
|
|
|
|
// list of buckets. It's not safe to remove keys during ForEach
|
|
|
|
// iteration nor is it safe to remove buckets during cursor
|
|
|
|
// iteration, so this dual approach is needed.
|
|
|
|
var bucketNames [][]byte
|
|
|
|
cursor := tx.Metadata().Cursor()
|
|
|
|
for ok := cursor.First(); ok; ok = cursor.Next() {
|
|
|
|
if cursor.Value() != nil {
|
|
|
|
if err := cursor.Delete(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
bucketNames = append(bucketNames, cursor.Key())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Remove the buckets.
|
|
|
|
for _, k := range bucketNames {
|
|
|
|
if err := tx.Metadata().DeleteBucket(k); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
_, err := tx.Metadata().CreateBucket(blockIdxBucketName)
|
|
|
|
return err
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("Update: unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Reset the mock files.
|
|
|
|
store := tc.db.(*db).store
|
|
|
|
wc := store.writeCursor
|
|
|
|
wc.curFile.Lock()
|
|
|
|
if wc.curFile.file != nil {
|
|
|
|
wc.curFile.file.Close()
|
|
|
|
wc.curFile.file = nil
|
|
|
|
}
|
|
|
|
wc.curFile.Unlock()
|
|
|
|
wc.Lock()
|
|
|
|
wc.curFileNum = 0
|
|
|
|
wc.curOffset = 0
|
|
|
|
wc.Unlock()
|
|
|
|
tc.files = make(map[uint32]*lockableFile)
|
|
|
|
tc.maxFileSizes = make(map[uint32]int64)
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testWriteFailures tests various failures paths when writing to the block
|
|
|
|
// files.
|
|
|
|
func testWriteFailures(tc *testContext) bool {
|
|
|
|
if !resetDatabase(tc) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
2016-02-03 18:41:46 +01:00
|
|
|
// Ensure file sync errors during flush return the expected error.
|
2016-02-03 18:42:04 +01:00
|
|
|
store := tc.db.(*db).store
|
2016-02-03 18:41:46 +01:00
|
|
|
testName := "flush: file sync failure"
|
2016-02-03 18:42:04 +01:00
|
|
|
store.writeCursor.Lock()
|
|
|
|
oldFile := store.writeCursor.curFile
|
|
|
|
store.writeCursor.curFile = &lockableFile{
|
|
|
|
file: &mockFile{forceSyncErr: true, maxSize: -1},
|
|
|
|
}
|
|
|
|
store.writeCursor.Unlock()
|
2016-02-03 18:41:46 +01:00
|
|
|
err := tc.db.(*db).cache.flush()
|
2016-02-03 18:42:04 +01:00
|
|
|
if !checkDbError(tc.t, testName, err, database.ErrDriverSpecific) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
store.writeCursor.Lock()
|
|
|
|
store.writeCursor.curFile = oldFile
|
|
|
|
store.writeCursor.Unlock()
|
|
|
|
|
|
|
|
// Force errors in the various error paths when writing data by using
|
|
|
|
// mock files with a limited max size.
|
|
|
|
block0Bytes, _ := tc.blocks[0].Bytes()
|
|
|
|
tests := []struct {
|
|
|
|
fileNum uint32
|
|
|
|
maxSize int64
|
|
|
|
}{
|
|
|
|
// Force an error when writing the network bytes.
|
|
|
|
{fileNum: 0, maxSize: 2},
|
|
|
|
|
|
|
|
// Force an error when writing the block size.
|
|
|
|
{fileNum: 0, maxSize: 6},
|
|
|
|
|
|
|
|
// Force an error when writing the block.
|
|
|
|
{fileNum: 0, maxSize: 17},
|
|
|
|
|
|
|
|
// Force an error when writing the checksum.
|
|
|
|
{fileNum: 0, maxSize: int64(len(block0Bytes)) + 10},
|
|
|
|
|
|
|
|
// Force an error after writing enough blocks for force multiple
|
|
|
|
// files.
|
|
|
|
{fileNum: 15, maxSize: 1},
|
|
|
|
}
|
|
|
|
|
|
|
|
for i, test := range tests {
|
|
|
|
if !resetDatabase(tc) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure storing the specified number of blocks using a mock
|
|
|
|
// file that fails the write fails when the transaction is
|
|
|
|
// committed, not when the block is stored.
|
|
|
|
tc.maxFileSizes = map[uint32]int64{test.fileNum: test.maxSize}
|
|
|
|
err := tc.db.Update(func(tx database.Tx) error {
|
|
|
|
for i, block := range tc.blocks {
|
|
|
|
err := tx.StoreBlock(block)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("StoreBlock (%d): unexpected "+
|
|
|
|
"error: %v", i, err)
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
testName := fmt.Sprintf("Force update commit failure - test "+
|
|
|
|
"%d, fileNum %d, maxsize %d", i, test.fileNum,
|
|
|
|
test.maxSize)
|
|
|
|
if !checkDbError(tc.t, testName, err, database.ErrDriverSpecific) {
|
|
|
|
tc.t.Errorf("%v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the commit rollback removed all extra files and data.
|
|
|
|
if len(tc.files) != 1 {
|
|
|
|
tc.t.Errorf("Update rollback: new not removed - want "+
|
|
|
|
"1 file, got %d", len(tc.files))
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if _, ok := tc.files[0]; !ok {
|
|
|
|
tc.t.Error("Update rollback: file 0 does not exist")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
file := tc.files[0].file.(*mockFile)
|
|
|
|
if len(file.data) != 0 {
|
|
|
|
tc.t.Errorf("Update rollback: file did not truncate - "+
|
|
|
|
"want len 0, got len %d", len(file.data))
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testBlockFileErrors ensures the database returns expected errors with various
|
|
|
|
// file-related issues such as closed and missing files.
|
|
|
|
func testBlockFileErrors(tc *testContext) bool {
|
|
|
|
if !resetDatabase(tc) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure errors in blockFile and openFile when requesting invalid file
|
|
|
|
// numbers.
|
|
|
|
store := tc.db.(*db).store
|
|
|
|
testName := "blockFile invalid file open"
|
|
|
|
_, err := store.blockFile(^uint32(0))
|
|
|
|
if !checkDbError(tc.t, testName, err, database.ErrDriverSpecific) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
testName = "openFile invalid file open"
|
|
|
|
_, err = store.openFile(^uint32(0))
|
|
|
|
if !checkDbError(tc.t, testName, err, database.ErrDriverSpecific) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Insert the first block into the mock file.
|
|
|
|
err = tc.db.Update(func(tx database.Tx) error {
|
|
|
|
err := tx.StoreBlock(tc.blocks[0])
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("StoreBlock: unexpected error: %v", err)
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
if err != errSubTestFail {
|
|
|
|
tc.t.Errorf("Update: unexpected error: %v", err)
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure errors in readBlock and readBlockRegion when requesting a file
|
|
|
|
// number that doesn't exist.
|
2016-08-08 21:04:33 +02:00
|
|
|
block0Hash := tc.blocks[0].Hash()
|
2016-02-03 18:42:04 +01:00
|
|
|
testName = "readBlock invalid file number"
|
|
|
|
invalidLoc := blockLocation{
|
|
|
|
blockFileNum: ^uint32(0),
|
|
|
|
blockLen: 80,
|
|
|
|
}
|
|
|
|
_, err = store.readBlock(block0Hash, invalidLoc)
|
|
|
|
if !checkDbError(tc.t, testName, err, database.ErrDriverSpecific) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
testName = "readBlockRegion invalid file number"
|
|
|
|
_, err = store.readBlockRegion(invalidLoc, 0, 80)
|
|
|
|
if !checkDbError(tc.t, testName, err, database.ErrDriverSpecific) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Close the block file out from under the database.
|
|
|
|
store.writeCursor.curFile.Lock()
|
|
|
|
store.writeCursor.curFile.file.Close()
|
|
|
|
store.writeCursor.curFile.Unlock()
|
|
|
|
|
|
|
|
// Ensure failures in FetchBlock and FetchBlockRegion(s) since the
|
|
|
|
// underlying file they need to read from has been closed.
|
|
|
|
err = tc.db.View(func(tx database.Tx) error {
|
|
|
|
testName = "FetchBlock closed file"
|
|
|
|
wantErrCode := database.ErrDriverSpecific
|
|
|
|
_, err := tx.FetchBlock(block0Hash)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
testName = "FetchBlockRegion closed file"
|
|
|
|
regions := []database.BlockRegion{
|
|
|
|
{
|
|
|
|
Hash: block0Hash,
|
|
|
|
Len: 80,
|
|
|
|
Offset: 0,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
_, err = tx.FetchBlockRegion(®ions[0])
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
testName = "FetchBlockRegions closed file"
|
|
|
|
_, err = tx.FetchBlockRegions(regions)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
if err != errSubTestFail {
|
|
|
|
tc.t.Errorf("View: unexpected error: %v", err)
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testCorruption ensures the database returns expected errors under various
|
|
|
|
// corruption scenarios.
|
|
|
|
func testCorruption(tc *testContext) bool {
|
|
|
|
if !resetDatabase(tc) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Insert the first block into the mock file.
|
|
|
|
err := tc.db.Update(func(tx database.Tx) error {
|
|
|
|
err := tx.StoreBlock(tc.blocks[0])
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("StoreBlock: unexpected error: %v", err)
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
if err != errSubTestFail {
|
|
|
|
tc.t.Errorf("Update: unexpected error: %v", err)
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure corruption is detected by intentionally modifying the bytes
|
|
|
|
// stored to the mock file and reading the block.
|
|
|
|
block0Bytes, _ := tc.blocks[0].Bytes()
|
2016-08-08 21:04:33 +02:00
|
|
|
block0Hash := tc.blocks[0].Hash()
|
2016-02-03 18:42:04 +01:00
|
|
|
tests := []struct {
|
|
|
|
offset uint32
|
|
|
|
fixChecksum bool
|
|
|
|
wantErrCode database.ErrorCode
|
|
|
|
}{
|
|
|
|
// One of the network bytes. The checksum needs to be fixed so
|
|
|
|
// the invalid network is detected.
|
|
|
|
{2, true, database.ErrDriverSpecific},
|
|
|
|
|
|
|
|
// The same network byte, but this time don't fix the checksum
|
|
|
|
// to ensure the corruption is detected.
|
|
|
|
{2, false, database.ErrCorruption},
|
|
|
|
|
|
|
|
// One of the block length bytes.
|
|
|
|
{6, false, database.ErrCorruption},
|
|
|
|
|
|
|
|
// Random header byte.
|
|
|
|
{17, false, database.ErrCorruption},
|
|
|
|
|
|
|
|
// Random transaction byte.
|
|
|
|
{90, false, database.ErrCorruption},
|
|
|
|
|
|
|
|
// Random checksum byte.
|
|
|
|
{uint32(len(block0Bytes)) + 10, false, database.ErrCorruption},
|
|
|
|
}
|
|
|
|
err = tc.db.View(func(tx database.Tx) error {
|
|
|
|
data := tc.files[0].file.(*mockFile).data
|
|
|
|
for i, test := range tests {
|
|
|
|
// Corrupt the byte at the offset by a single bit.
|
|
|
|
data[test.offset] ^= 0x10
|
|
|
|
|
|
|
|
// Fix the checksum if requested to force other errors.
|
|
|
|
fileLen := len(data)
|
|
|
|
var oldChecksumBytes [4]byte
|
|
|
|
copy(oldChecksumBytes[:], data[fileLen-4:])
|
|
|
|
if test.fixChecksum {
|
|
|
|
toSum := data[:fileLen-4]
|
|
|
|
cksum := crc32.Checksum(toSum, castagnoli)
|
|
|
|
binary.BigEndian.PutUint32(data[fileLen-4:], cksum)
|
|
|
|
}
|
|
|
|
|
|
|
|
testName := fmt.Sprintf("FetchBlock (test #%d): "+
|
|
|
|
"corruption", i)
|
|
|
|
_, err := tx.FetchBlock(block0Hash)
|
|
|
|
if !checkDbError(tc.t, testName, err, test.wantErrCode) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
// Reset the corrupted data back to the original.
|
|
|
|
data[test.offset] ^= 0x10
|
|
|
|
if test.fixChecksum {
|
|
|
|
copy(data[fileLen-4:], oldChecksumBytes[:])
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
if err != errSubTestFail {
|
|
|
|
tc.t.Errorf("View: unexpected error: %v", err)
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestFailureScenarios ensures several failure scenarios such as database
|
|
|
|
// corruption, block file write failures, and rollback failures are handled
|
|
|
|
// correctly.
|
|
|
|
func TestFailureScenarios(t *testing.T) {
|
|
|
|
// Create a new database to run tests against.
|
|
|
|
dbPath := filepath.Join(os.TempDir(), "ffldb-failurescenarios")
|
|
|
|
_ = os.RemoveAll(dbPath)
|
|
|
|
idb, err := database.Create(dbType, dbPath, blockDataNet)
|
|
|
|
if err != nil {
|
|
|
|
t.Errorf("Failed to create test database (%s) %v", dbType, err)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
defer os.RemoveAll(dbPath)
|
|
|
|
defer idb.Close()
|
|
|
|
|
|
|
|
// Create a test context to pass around.
|
|
|
|
tc := &testContext{
|
|
|
|
t: t,
|
|
|
|
db: idb,
|
|
|
|
files: make(map[uint32]*lockableFile),
|
|
|
|
maxFileSizes: make(map[uint32]int64),
|
|
|
|
}
|
|
|
|
|
|
|
|
// Change the maximum file size to a small value to force multiple flat
|
|
|
|
// files with the test data set and replace the file-related functions
|
|
|
|
// to make use of mock files in memory. This allows injection of
|
|
|
|
// various file-related errors.
|
|
|
|
store := idb.(*db).store
|
|
|
|
store.maxBlockFileSize = 1024 // 1KiB
|
|
|
|
store.openWriteFileFunc = func(fileNum uint32) (filer, error) {
|
|
|
|
if file, ok := tc.files[fileNum]; ok {
|
|
|
|
// "Reopen" the file.
|
|
|
|
file.Lock()
|
|
|
|
mock := file.file.(*mockFile)
|
|
|
|
mock.Lock()
|
|
|
|
mock.closed = false
|
|
|
|
mock.Unlock()
|
|
|
|
file.Unlock()
|
|
|
|
return mock, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Limit the max size of the mock file as specified in the test
|
|
|
|
// context.
|
|
|
|
maxSize := int64(-1)
|
|
|
|
if maxFileSize, ok := tc.maxFileSizes[fileNum]; ok {
|
2020-05-13 14:58:39 +02:00
|
|
|
maxSize = maxFileSize
|
2016-02-03 18:42:04 +01:00
|
|
|
}
|
2020-05-13 14:58:39 +02:00
|
|
|
file := &mockFile{maxSize: maxSize}
|
2016-02-03 18:42:04 +01:00
|
|
|
tc.files[fileNum] = &lockableFile{file: file}
|
|
|
|
return file, nil
|
|
|
|
}
|
|
|
|
store.openFileFunc = func(fileNum uint32) (*lockableFile, error) {
|
|
|
|
// Force error when trying to open max file num.
|
|
|
|
if fileNum == ^uint32(0) {
|
|
|
|
return nil, makeDbErr(database.ErrDriverSpecific,
|
|
|
|
"test", nil)
|
|
|
|
}
|
|
|
|
if file, ok := tc.files[fileNum]; ok {
|
|
|
|
// "Reopen" the file.
|
|
|
|
file.Lock()
|
|
|
|
mock := file.file.(*mockFile)
|
|
|
|
mock.Lock()
|
|
|
|
mock.closed = false
|
|
|
|
mock.Unlock()
|
|
|
|
file.Unlock()
|
|
|
|
return file, nil
|
|
|
|
}
|
|
|
|
file := &lockableFile{file: &mockFile{}}
|
|
|
|
tc.files[fileNum] = file
|
|
|
|
return file, nil
|
|
|
|
}
|
|
|
|
store.deleteFileFunc = func(fileNum uint32) error {
|
|
|
|
if file, ok := tc.files[fileNum]; ok {
|
|
|
|
file.Lock()
|
|
|
|
file.file.Close()
|
|
|
|
file.Unlock()
|
|
|
|
delete(tc.files, fileNum)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
str := fmt.Sprintf("file %d does not exist", fileNum)
|
|
|
|
return makeDbErr(database.ErrDriverSpecific, str, nil)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load the test blocks and save in the test context for use throughout
|
|
|
|
// the tests.
|
|
|
|
blocks, err := loadBlocks(t, blockDataFile, blockDataNet)
|
|
|
|
if err != nil {
|
|
|
|
t.Errorf("loadBlocks: Unexpected error: %v", err)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
tc.blocks = blocks
|
|
|
|
|
|
|
|
// Test various failures paths when writing to the block files.
|
|
|
|
if !testWriteFailures(tc) {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// Test various file-related issues such as closed and missing files.
|
|
|
|
if !testBlockFileErrors(tc) {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// Test various corruption scenarios.
|
|
|
|
testCorruption(tc)
|
|
|
|
}
|