2016-02-03 18:42:04 +01:00
|
|
|
// Copyright (c) 2015-2016 The btcsuite developers
|
|
|
|
// Use of this source code is governed by an ISC
|
|
|
|
// license that can be found in the LICENSE file.
|
|
|
|
|
|
|
|
// This file intended to be copied into each backend driver directory. Each
|
|
|
|
// driver should have their own driver_test.go file which creates a database and
|
|
|
|
// invokes the testInterface function in this file to ensure the driver properly
|
|
|
|
// implements the interface.
|
|
|
|
//
|
|
|
|
// NOTE: When copying this file into the backend driver folder, the package name
|
|
|
|
// will need to be changed accordingly.
|
|
|
|
|
|
|
|
package ffldb_test
|
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
|
|
|
"compress/bzip2"
|
|
|
|
"encoding/binary"
|
|
|
|
"fmt"
|
|
|
|
"io"
|
|
|
|
"os"
|
|
|
|
"path/filepath"
|
|
|
|
"reflect"
|
|
|
|
"sync/atomic"
|
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
2021-10-15 07:45:32 +02:00
|
|
|
"github.com/lbryio/lbcd/chaincfg/chainhash"
|
|
|
|
"github.com/lbryio/lbcd/database"
|
|
|
|
"github.com/lbryio/lbcd/wire"
|
|
|
|
btcutil "github.com/lbryio/lbcutil"
|
2016-02-03 18:42:04 +01:00
|
|
|
)
|
|
|
|
|
|
|
|
var (
|
|
|
|
// blockDataNet is the expected network in the test block data.
|
|
|
|
blockDataNet = wire.MainNet
|
|
|
|
|
|
|
|
// blockDataFile is the path to a file containing the first 256 blocks
|
|
|
|
// of the block chain.
|
|
|
|
blockDataFile = filepath.Join("..", "testdata", "blocks1-256.bz2")
|
|
|
|
|
|
|
|
// errSubTestFail is used to signal that a sub test returned false.
|
|
|
|
errSubTestFail = fmt.Errorf("sub test failure")
|
|
|
|
)
|
|
|
|
|
|
|
|
// loadBlocks loads the blocks contained in the testdata directory and returns
|
|
|
|
// a slice of them.
|
|
|
|
func loadBlocks(t *testing.T, dataFile string, network wire.BitcoinNet) ([]*btcutil.Block, error) {
|
|
|
|
// Open the file that contains the blocks for reading.
|
|
|
|
fi, err := os.Open(dataFile)
|
|
|
|
if err != nil {
|
|
|
|
t.Errorf("failed to open file %v, err %v", dataFile, err)
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
defer func() {
|
|
|
|
if err := fi.Close(); err != nil {
|
|
|
|
t.Errorf("failed to close file %v %v", dataFile,
|
|
|
|
err)
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
dr := bzip2.NewReader(fi)
|
|
|
|
|
|
|
|
// Set the first block as the genesis block.
|
2021-08-19 20:41:48 +02:00
|
|
|
blocks := make([]*btcutil.Block, 0, 257)
|
2016-02-03 18:42:04 +01:00
|
|
|
|
|
|
|
// Load the remaining blocks.
|
2021-08-19 20:41:48 +02:00
|
|
|
for {
|
2016-02-03 18:42:04 +01:00
|
|
|
var net uint32
|
|
|
|
err := binary.Read(dr, binary.LittleEndian, &net)
|
|
|
|
if err == io.EOF {
|
|
|
|
// Hit end of file at the expected offset. No error.
|
|
|
|
break
|
|
|
|
}
|
|
|
|
if err != nil {
|
|
|
|
t.Errorf("Failed to load network type for block %d: %v",
|
2021-08-19 20:41:48 +02:00
|
|
|
len(blocks), err)
|
2016-02-03 18:42:04 +01:00
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
if net != uint32(network) {
|
2021-08-19 20:41:48 +02:00
|
|
|
continue
|
2016-02-03 18:42:04 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
var blockLen uint32
|
|
|
|
err = binary.Read(dr, binary.LittleEndian, &blockLen)
|
|
|
|
if err != nil {
|
|
|
|
t.Errorf("Failed to load block size for block %d: %v",
|
2021-08-19 20:41:48 +02:00
|
|
|
len(blocks), err)
|
2016-02-03 18:42:04 +01:00
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Read the block.
|
|
|
|
blockBytes := make([]byte, blockLen)
|
|
|
|
_, err = io.ReadFull(dr, blockBytes)
|
|
|
|
if err != nil {
|
2021-08-19 20:41:48 +02:00
|
|
|
t.Errorf("Failed to load block %d: %v", len(blocks), err)
|
2016-02-03 18:42:04 +01:00
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Deserialize and store the block.
|
|
|
|
block, err := btcutil.NewBlockFromBytes(blockBytes)
|
|
|
|
if err != nil {
|
2021-08-19 20:41:48 +02:00
|
|
|
t.Errorf("Failed to parse block %v: %v", len(blocks), err)
|
2016-02-03 18:42:04 +01:00
|
|
|
return nil, err
|
|
|
|
}
|
2021-08-19 20:41:48 +02:00
|
|
|
// NOTE: there's a bug here in that it doesn't read the checksum;
|
|
|
|
// we account for that by checking the network above; it probably skips every other block
|
2016-02-03 18:42:04 +01:00
|
|
|
blocks = append(blocks, block)
|
2021-08-19 20:41:48 +02:00
|
|
|
if len(blocks) == 257 {
|
|
|
|
break
|
|
|
|
}
|
2016-02-03 18:42:04 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
return blocks, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// checkDbError ensures the passed error is a database.Error with an error code
|
|
|
|
// that matches the passed error code.
|
|
|
|
func checkDbError(t *testing.T, testName string, gotErr error, wantErrCode database.ErrorCode) bool {
|
|
|
|
dbErr, ok := gotErr.(database.Error)
|
|
|
|
if !ok {
|
|
|
|
t.Errorf("%s: unexpected error type - got %T, want %T",
|
|
|
|
testName, gotErr, database.Error{})
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if dbErr.ErrorCode != wantErrCode {
|
|
|
|
t.Errorf("%s: unexpected error code - got %s (%s), want %s",
|
|
|
|
testName, dbErr.ErrorCode, dbErr.Description,
|
|
|
|
wantErrCode)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testContext is used to store context information about a running test which
|
|
|
|
// is passed into helper functions.
|
|
|
|
type testContext struct {
|
|
|
|
t *testing.T
|
|
|
|
db database.DB
|
|
|
|
bucketDepth int
|
|
|
|
isWritable bool
|
|
|
|
blocks []*btcutil.Block
|
|
|
|
}
|
|
|
|
|
|
|
|
// keyPair houses a key/value pair. It is used over maps so ordering can be
|
|
|
|
// maintained.
|
|
|
|
type keyPair struct {
|
|
|
|
key []byte
|
|
|
|
value []byte
|
|
|
|
}
|
|
|
|
|
|
|
|
// lookupKey is a convenience method to lookup the requested key from the
|
|
|
|
// provided keypair slice along with whether or not the key was found.
|
|
|
|
func lookupKey(key []byte, values []keyPair) ([]byte, bool) {
|
|
|
|
for _, item := range values {
|
|
|
|
if bytes.Equal(item.key, key) {
|
|
|
|
return item.value, true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil, false
|
|
|
|
}
|
|
|
|
|
|
|
|
// toGetValues returns a copy of the provided keypairs with all of the nil
|
|
|
|
// values set to an empty byte slice. This is used to ensure that keys set to
|
|
|
|
// nil values result in empty byte slices when retrieved instead of nil.
|
|
|
|
func toGetValues(values []keyPair) []keyPair {
|
|
|
|
ret := make([]keyPair, len(values))
|
|
|
|
copy(ret, values)
|
|
|
|
for i := range ret {
|
|
|
|
if ret[i].value == nil {
|
|
|
|
ret[i].value = make([]byte, 0)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return ret
|
|
|
|
}
|
|
|
|
|
|
|
|
// rollbackValues returns a copy of the provided keypairs with all values set to
|
|
|
|
// nil. This is used to test that values are properly rolled back.
|
|
|
|
func rollbackValues(values []keyPair) []keyPair {
|
|
|
|
ret := make([]keyPair, len(values))
|
|
|
|
copy(ret, values)
|
|
|
|
for i := range ret {
|
|
|
|
ret[i].value = nil
|
|
|
|
}
|
|
|
|
return ret
|
|
|
|
}
|
|
|
|
|
|
|
|
// testCursorKeyPair checks that the provide key and value match the expected
|
|
|
|
// keypair at the provided index. It also ensures the index is in range for the
|
|
|
|
// provided slice of expected keypairs.
|
|
|
|
func testCursorKeyPair(tc *testContext, k, v []byte, index int, values []keyPair) bool {
|
|
|
|
if index >= len(values) || index < 0 {
|
|
|
|
tc.t.Errorf("Cursor: exceeded the expected range of values - "+
|
|
|
|
"index %d, num values %d", index, len(values))
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
pair := &values[index]
|
|
|
|
if !bytes.Equal(k, pair.key) {
|
|
|
|
tc.t.Errorf("Mismatched cursor key: index %d does not match "+
|
|
|
|
"the expected key - got %q, want %q", index, k,
|
|
|
|
pair.key)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if !bytes.Equal(v, pair.value) {
|
|
|
|
tc.t.Errorf("Mismatched cursor value: index %d does not match "+
|
|
|
|
"the expected value - got %q, want %q", index, v,
|
|
|
|
pair.value)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testGetValues checks that all of the provided key/value pairs can be
|
|
|
|
// retrieved from the database and the retrieved values match the provided
|
|
|
|
// values.
|
|
|
|
func testGetValues(tc *testContext, bucket database.Bucket, values []keyPair) bool {
|
|
|
|
for _, item := range values {
|
|
|
|
gotValue := bucket.Get(item.key)
|
|
|
|
if !reflect.DeepEqual(gotValue, item.value) {
|
|
|
|
tc.t.Errorf("Get: unexpected value for %q - got %q, "+
|
|
|
|
"want %q", item.key, gotValue, item.value)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testPutValues stores all of the provided key/value pairs in the provided
|
|
|
|
// bucket while checking for errors.
|
|
|
|
func testPutValues(tc *testContext, bucket database.Bucket, values []keyPair) bool {
|
|
|
|
for _, item := range values {
|
|
|
|
if err := bucket.Put(item.key, item.value); err != nil {
|
|
|
|
tc.t.Errorf("Put: unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testDeleteValues removes all of the provided key/value pairs from the
|
|
|
|
// provided bucket.
|
|
|
|
func testDeleteValues(tc *testContext, bucket database.Bucket, values []keyPair) bool {
|
|
|
|
for _, item := range values {
|
|
|
|
if err := bucket.Delete(item.key); err != nil {
|
|
|
|
tc.t.Errorf("Delete: unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testCursorInterface ensures the cursor itnerface is working properly by
|
|
|
|
// exercising all of its functions on the passed bucket.
|
|
|
|
func testCursorInterface(tc *testContext, bucket database.Bucket) bool {
|
|
|
|
// Ensure a cursor can be obtained for the bucket.
|
|
|
|
cursor := bucket.Cursor()
|
|
|
|
if cursor == nil {
|
|
|
|
tc.t.Error("Bucket.Cursor: unexpected nil cursor returned")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the cursor returns the same bucket it was created for.
|
|
|
|
if cursor.Bucket() != bucket {
|
|
|
|
tc.t.Error("Cursor.Bucket: does not match the bucket it was " +
|
|
|
|
"created for")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
if tc.isWritable {
|
|
|
|
unsortedValues := []keyPair{
|
|
|
|
{[]byte("cursor"), []byte("val1")},
|
|
|
|
{[]byte("abcd"), []byte("val2")},
|
|
|
|
{[]byte("bcd"), []byte("val3")},
|
|
|
|
{[]byte("defg"), nil},
|
|
|
|
}
|
|
|
|
sortedValues := []keyPair{
|
|
|
|
{[]byte("abcd"), []byte("val2")},
|
|
|
|
{[]byte("bcd"), []byte("val3")},
|
|
|
|
{[]byte("cursor"), []byte("val1")},
|
|
|
|
{[]byte("defg"), nil},
|
|
|
|
}
|
|
|
|
|
|
|
|
// Store the values to be used in the cursor tests in unsorted
|
|
|
|
// order and ensure they were actually stored.
|
|
|
|
if !testPutValues(tc, bucket, unsortedValues) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if !testGetValues(tc, bucket, toGetValues(unsortedValues)) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the cursor returns all items in byte-sorted order when
|
|
|
|
// iterating forward.
|
|
|
|
curIdx := 0
|
|
|
|
for ok := cursor.First(); ok; ok = cursor.Next() {
|
|
|
|
k, v := cursor.Key(), cursor.Value()
|
|
|
|
if !testCursorKeyPair(tc, k, v, curIdx, sortedValues) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
curIdx++
|
|
|
|
}
|
|
|
|
if curIdx != len(unsortedValues) {
|
|
|
|
tc.t.Errorf("Cursor: expected to iterate %d values, "+
|
|
|
|
"but only iterated %d", len(unsortedValues),
|
|
|
|
curIdx)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the cursor returns all items in reverse byte-sorted
|
|
|
|
// order when iterating in reverse.
|
|
|
|
curIdx = len(sortedValues) - 1
|
|
|
|
for ok := cursor.Last(); ok; ok = cursor.Prev() {
|
|
|
|
k, v := cursor.Key(), cursor.Value()
|
|
|
|
if !testCursorKeyPair(tc, k, v, curIdx, sortedValues) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
curIdx--
|
|
|
|
}
|
|
|
|
if curIdx > -1 {
|
|
|
|
tc.t.Errorf("Reverse cursor: expected to iterate %d "+
|
|
|
|
"values, but only iterated %d",
|
|
|
|
len(sortedValues), len(sortedValues)-(curIdx+1))
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
2016-02-25 18:17:12 +01:00
|
|
|
// Ensure forward iteration works as expected after seeking.
|
2016-02-03 18:42:04 +01:00
|
|
|
middleIdx := (len(sortedValues) - 1) / 2
|
|
|
|
seekKey := sortedValues[middleIdx].key
|
|
|
|
curIdx = middleIdx
|
|
|
|
for ok := cursor.Seek(seekKey); ok; ok = cursor.Next() {
|
|
|
|
k, v := cursor.Key(), cursor.Value()
|
|
|
|
if !testCursorKeyPair(tc, k, v, curIdx, sortedValues) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
curIdx++
|
|
|
|
}
|
|
|
|
if curIdx != len(sortedValues) {
|
|
|
|
tc.t.Errorf("Cursor after seek: expected to iterate "+
|
|
|
|
"%d values, but only iterated %d",
|
|
|
|
len(sortedValues)-middleIdx, curIdx-middleIdx)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure reverse iteration works as expected after seeking.
|
|
|
|
curIdx = middleIdx
|
|
|
|
for ok := cursor.Seek(seekKey); ok; ok = cursor.Prev() {
|
|
|
|
k, v := cursor.Key(), cursor.Value()
|
|
|
|
if !testCursorKeyPair(tc, k, v, curIdx, sortedValues) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
curIdx--
|
|
|
|
}
|
|
|
|
if curIdx > -1 {
|
|
|
|
tc.t.Errorf("Reverse cursor after seek: expected to "+
|
|
|
|
"iterate %d values, but only iterated %d",
|
|
|
|
len(sortedValues)-middleIdx, middleIdx-curIdx)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the cursor deletes items properly.
|
|
|
|
if !cursor.First() {
|
|
|
|
tc.t.Errorf("Cursor.First: no value")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
k := cursor.Key()
|
|
|
|
if err := cursor.Delete(); err != nil {
|
|
|
|
tc.t.Errorf("Cursor.Delete: unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if val := bucket.Get(k); val != nil {
|
|
|
|
tc.t.Errorf("Cursor.Delete: value for key %q was not "+
|
|
|
|
"deleted", k)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testNestedBucket reruns the testBucketInterface against a nested bucket along
|
|
|
|
// with a counter to only test a couple of level deep.
|
|
|
|
func testNestedBucket(tc *testContext, testBucket database.Bucket) bool {
|
|
|
|
// Don't go more than 2 nested levels deep.
|
|
|
|
if tc.bucketDepth > 1 {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
tc.bucketDepth++
|
|
|
|
defer func() {
|
|
|
|
tc.bucketDepth--
|
|
|
|
}()
|
2016-11-03 05:02:04 +01:00
|
|
|
return testBucketInterface(tc, testBucket)
|
2016-02-03 18:42:04 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// testBucketInterface ensures the bucket interface is working properly by
|
|
|
|
// exercising all of its functions. This includes the cursor interface for the
|
|
|
|
// cursor returned from the bucket.
|
|
|
|
func testBucketInterface(tc *testContext, bucket database.Bucket) bool {
|
|
|
|
if bucket.Writable() != tc.isWritable {
|
|
|
|
tc.t.Errorf("Bucket writable state does not match.")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
if tc.isWritable {
|
|
|
|
// keyValues holds the keys and values to use when putting
|
|
|
|
// values into the bucket.
|
|
|
|
keyValues := []keyPair{
|
|
|
|
{[]byte("bucketkey1"), []byte("foo1")},
|
|
|
|
{[]byte("bucketkey2"), []byte("foo2")},
|
|
|
|
{[]byte("bucketkey3"), []byte("foo3")},
|
|
|
|
{[]byte("bucketkey4"), nil},
|
|
|
|
}
|
|
|
|
expectedKeyValues := toGetValues(keyValues)
|
|
|
|
if !testPutValues(tc, bucket, keyValues) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
if !testGetValues(tc, bucket, expectedKeyValues) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure errors returned from the user-supplied ForEach
|
|
|
|
// function are returned.
|
|
|
|
forEachError := fmt.Errorf("example foreach error")
|
|
|
|
err := bucket.ForEach(func(k, v []byte) error {
|
|
|
|
return forEachError
|
|
|
|
})
|
|
|
|
if err != forEachError {
|
|
|
|
tc.t.Errorf("ForEach: inner function error not "+
|
|
|
|
"returned - got %v, want %v", err, forEachError)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Iterate all of the keys using ForEach while making sure the
|
|
|
|
// stored values are the expected values.
|
|
|
|
keysFound := make(map[string]struct{}, len(keyValues))
|
|
|
|
err = bucket.ForEach(func(k, v []byte) error {
|
|
|
|
wantV, found := lookupKey(k, expectedKeyValues)
|
|
|
|
if !found {
|
|
|
|
return fmt.Errorf("ForEach: key '%s' should "+
|
|
|
|
"exist", k)
|
|
|
|
}
|
|
|
|
|
|
|
|
if !reflect.DeepEqual(v, wantV) {
|
|
|
|
return fmt.Errorf("ForEach: value for key '%s' "+
|
|
|
|
"does not match - got %s, want %s", k,
|
|
|
|
v, wantV)
|
|
|
|
}
|
|
|
|
|
|
|
|
keysFound[string(k)] = struct{}{}
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("%v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure all keys were iterated.
|
|
|
|
for _, item := range keyValues {
|
|
|
|
if _, ok := keysFound[string(item.key)]; !ok {
|
|
|
|
tc.t.Errorf("ForEach: key '%s' was not iterated "+
|
|
|
|
"when it should have been", item.key)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Delete the keys and ensure they were deleted.
|
|
|
|
if !testDeleteValues(tc, bucket, keyValues) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if !testGetValues(tc, bucket, rollbackValues(keyValues)) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure creating a new bucket works as expected.
|
|
|
|
testBucketName := []byte("testbucket")
|
|
|
|
testBucket, err := bucket.CreateBucket(testBucketName)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("CreateBucket: unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if !testNestedBucket(tc, testBucket) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure errors returned from the user-supplied ForEachBucket
|
|
|
|
// function are returned.
|
|
|
|
err = bucket.ForEachBucket(func(k []byte) error {
|
|
|
|
return forEachError
|
|
|
|
})
|
|
|
|
if err != forEachError {
|
|
|
|
tc.t.Errorf("ForEachBucket: inner function error not "+
|
|
|
|
"returned - got %v, want %v", err, forEachError)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure creating a bucket that already exists fails with the
|
|
|
|
// expected error.
|
|
|
|
wantErrCode := database.ErrBucketExists
|
|
|
|
_, err = bucket.CreateBucket(testBucketName)
|
|
|
|
if !checkDbError(tc.t, "CreateBucket", err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure CreateBucketIfNotExists returns an existing bucket.
|
|
|
|
testBucket, err = bucket.CreateBucketIfNotExists(testBucketName)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("CreateBucketIfNotExists: unexpected "+
|
|
|
|
"error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if !testNestedBucket(tc, testBucket) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure retrieving an existing bucket works as expected.
|
|
|
|
testBucket = bucket.Bucket(testBucketName)
|
|
|
|
if !testNestedBucket(tc, testBucket) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure deleting a bucket works as intended.
|
|
|
|
if err := bucket.DeleteBucket(testBucketName); err != nil {
|
|
|
|
tc.t.Errorf("DeleteBucket: unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if b := bucket.Bucket(testBucketName); b != nil {
|
|
|
|
tc.t.Errorf("DeleteBucket: bucket '%s' still exists",
|
|
|
|
testBucketName)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure deleting a bucket that doesn't exist returns the
|
|
|
|
// expected error.
|
|
|
|
wantErrCode = database.ErrBucketNotFound
|
|
|
|
err = bucket.DeleteBucket(testBucketName)
|
|
|
|
if !checkDbError(tc.t, "DeleteBucket", err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure CreateBucketIfNotExists creates a new bucket when
|
|
|
|
// it doesn't already exist.
|
|
|
|
testBucket, err = bucket.CreateBucketIfNotExists(testBucketName)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("CreateBucketIfNotExists: unexpected "+
|
|
|
|
"error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if !testNestedBucket(tc, testBucket) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the cursor interface works as expected.
|
|
|
|
if !testCursorInterface(tc, testBucket) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Delete the test bucket to avoid leaving it around for future
|
|
|
|
// calls.
|
|
|
|
if err := bucket.DeleteBucket(testBucketName); err != nil {
|
|
|
|
tc.t.Errorf("DeleteBucket: unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if b := bucket.Bucket(testBucketName); b != nil {
|
|
|
|
tc.t.Errorf("DeleteBucket: bucket '%s' still exists",
|
|
|
|
testBucketName)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// Put should fail with bucket that is not writable.
|
|
|
|
testName := "unwritable tx put"
|
|
|
|
wantErrCode := database.ErrTxNotWritable
|
|
|
|
failBytes := []byte("fail")
|
|
|
|
err := bucket.Put(failBytes, failBytes)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Delete should fail with bucket that is not writable.
|
|
|
|
testName = "unwritable tx delete"
|
|
|
|
err = bucket.Delete(failBytes)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// CreateBucket should fail with bucket that is not writable.
|
|
|
|
testName = "unwritable tx create bucket"
|
|
|
|
_, err = bucket.CreateBucket(failBytes)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// CreateBucketIfNotExists should fail with bucket that is not
|
|
|
|
// writable.
|
|
|
|
testName = "unwritable tx create bucket if not exists"
|
|
|
|
_, err = bucket.CreateBucketIfNotExists(failBytes)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// DeleteBucket should fail with bucket that is not writable.
|
|
|
|
testName = "unwritable tx delete bucket"
|
|
|
|
err = bucket.DeleteBucket(failBytes)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the cursor interface works as expected with read-only
|
|
|
|
// buckets.
|
|
|
|
if !testCursorInterface(tc, bucket) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// rollbackOnPanic rolls the passed transaction back if the code in the calling
|
|
|
|
// function panics. This is useful in case the tests unexpectedly panic which
|
|
|
|
// would leave any manually created transactions with the database mutex locked
|
|
|
|
// thereby leading to a deadlock and masking the real reason for the panic. It
|
|
|
|
// also logs a test error and repanics so the original panic can be traced.
|
|
|
|
func rollbackOnPanic(t *testing.T, tx database.Tx) {
|
|
|
|
if err := recover(); err != nil {
|
|
|
|
t.Errorf("Unexpected panic: %v", err)
|
|
|
|
_ = tx.Rollback()
|
|
|
|
panic(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// testMetadataManualTxInterface ensures that the manual transactions metadata
|
|
|
|
// interface works as expected.
|
|
|
|
func testMetadataManualTxInterface(tc *testContext) bool {
|
|
|
|
// populateValues tests that populating values works as expected.
|
|
|
|
//
|
|
|
|
// When the writable flag is false, a read-only tranasction is created,
|
|
|
|
// standard bucket tests for read-only transactions are performed, and
|
|
|
|
// the Commit function is checked to ensure it fails as expected.
|
|
|
|
//
|
|
|
|
// Otherwise, a read-write transaction is created, the values are
|
|
|
|
// written, standard bucket tests for read-write transactions are
|
2016-02-25 18:17:12 +01:00
|
|
|
// performed, and then the transaction is either committed or rolled
|
2016-02-03 18:42:04 +01:00
|
|
|
// back depending on the flag.
|
|
|
|
bucket1Name := []byte("bucket1")
|
|
|
|
populateValues := func(writable, rollback bool, putValues []keyPair) bool {
|
|
|
|
tx, err := tc.db.Begin(writable)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("Begin: unexpected error %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
defer rollbackOnPanic(tc.t, tx)
|
|
|
|
|
|
|
|
metadataBucket := tx.Metadata()
|
|
|
|
if metadataBucket == nil {
|
|
|
|
tc.t.Errorf("Metadata: unexpected nil bucket")
|
|
|
|
_ = tx.Rollback()
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
bucket1 := metadataBucket.Bucket(bucket1Name)
|
|
|
|
if bucket1 == nil {
|
|
|
|
tc.t.Errorf("Bucket1: unexpected nil bucket")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
tc.isWritable = writable
|
|
|
|
if !testBucketInterface(tc, bucket1) {
|
|
|
|
_ = tx.Rollback()
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
if !writable {
|
|
|
|
// The transaction is not writable, so it should fail
|
|
|
|
// the commit.
|
|
|
|
testName := "unwritable tx commit"
|
|
|
|
wantErrCode := database.ErrTxNotWritable
|
|
|
|
err := tx.Commit()
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
_ = tx.Rollback()
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
if !testPutValues(tc, bucket1, putValues) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
if rollback {
|
|
|
|
// Rollback the transaction.
|
|
|
|
if err := tx.Rollback(); err != nil {
|
|
|
|
tc.t.Errorf("Rollback: unexpected "+
|
|
|
|
"error %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// The commit should succeed.
|
|
|
|
if err := tx.Commit(); err != nil {
|
|
|
|
tc.t.Errorf("Commit: unexpected error "+
|
|
|
|
"%v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// checkValues starts a read-only transaction and checks that all of
|
|
|
|
// the key/value pairs specified in the expectedValues parameter match
|
|
|
|
// what's in the database.
|
|
|
|
checkValues := func(expectedValues []keyPair) bool {
|
|
|
|
tx, err := tc.db.Begin(false)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("Begin: unexpected error %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
defer rollbackOnPanic(tc.t, tx)
|
|
|
|
|
|
|
|
metadataBucket := tx.Metadata()
|
|
|
|
if metadataBucket == nil {
|
|
|
|
tc.t.Errorf("Metadata: unexpected nil bucket")
|
|
|
|
_ = tx.Rollback()
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
bucket1 := metadataBucket.Bucket(bucket1Name)
|
|
|
|
if bucket1 == nil {
|
|
|
|
tc.t.Errorf("Bucket1: unexpected nil bucket")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
if !testGetValues(tc, bucket1, expectedValues) {
|
|
|
|
_ = tx.Rollback()
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Rollback the read-only transaction.
|
|
|
|
if err := tx.Rollback(); err != nil {
|
|
|
|
tc.t.Errorf("Commit: unexpected error %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// deleteValues starts a read-write transaction and deletes the keys
|
|
|
|
// in the passed key/value pairs.
|
|
|
|
deleteValues := func(values []keyPair) bool {
|
|
|
|
tx, err := tc.db.Begin(true)
|
|
|
|
if err != nil {
|
|
|
|
|
|
|
|
}
|
|
|
|
defer rollbackOnPanic(tc.t, tx)
|
|
|
|
|
|
|
|
metadataBucket := tx.Metadata()
|
|
|
|
if metadataBucket == nil {
|
|
|
|
tc.t.Errorf("Metadata: unexpected nil bucket")
|
|
|
|
_ = tx.Rollback()
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
bucket1 := metadataBucket.Bucket(bucket1Name)
|
|
|
|
if bucket1 == nil {
|
|
|
|
tc.t.Errorf("Bucket1: unexpected nil bucket")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Delete the keys and ensure they were deleted.
|
|
|
|
if !testDeleteValues(tc, bucket1, values) {
|
|
|
|
_ = tx.Rollback()
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if !testGetValues(tc, bucket1, rollbackValues(values)) {
|
|
|
|
_ = tx.Rollback()
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Commit the changes and ensure it was successful.
|
|
|
|
if err := tx.Commit(); err != nil {
|
|
|
|
tc.t.Errorf("Commit: unexpected error %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// keyValues holds the keys and values to use when putting values into a
|
|
|
|
// bucket.
|
|
|
|
var keyValues = []keyPair{
|
|
|
|
{[]byte("umtxkey1"), []byte("foo1")},
|
|
|
|
{[]byte("umtxkey2"), []byte("foo2")},
|
|
|
|
{[]byte("umtxkey3"), []byte("foo3")},
|
|
|
|
{[]byte("umtxkey4"), nil},
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure that attempting populating the values using a read-only
|
|
|
|
// transaction fails as expected.
|
|
|
|
if !populateValues(false, true, keyValues) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if !checkValues(rollbackValues(keyValues)) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure that attempting populating the values using a read-write
|
|
|
|
// transaction and then rolling it back yields the expected values.
|
|
|
|
if !populateValues(true, true, keyValues) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if !checkValues(rollbackValues(keyValues)) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure that attempting populating the values using a read-write
|
|
|
|
// transaction and then committing it stores the expected values.
|
|
|
|
if !populateValues(true, false, keyValues) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if !checkValues(toGetValues(keyValues)) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Clean up the keys.
|
|
|
|
if !deleteValues(keyValues) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testManagedTxPanics ensures calling Rollback of Commit inside a managed
|
|
|
|
// transaction panics.
|
|
|
|
func testManagedTxPanics(tc *testContext) bool {
|
|
|
|
testPanic := func(fn func()) (paniced bool) {
|
|
|
|
// Setup a defer to catch the expected panic and update the
|
|
|
|
// return variable.
|
|
|
|
defer func() {
|
|
|
|
if err := recover(); err != nil {
|
|
|
|
paniced = true
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
fn()
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure calling Commit on a managed read-only transaction panics.
|
|
|
|
paniced := testPanic(func() {
|
|
|
|
tc.db.View(func(tx database.Tx) error {
|
|
|
|
tx.Commit()
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
})
|
|
|
|
if !paniced {
|
|
|
|
tc.t.Error("Commit called inside View did not panic")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure calling Rollback on a managed read-only transaction panics.
|
|
|
|
paniced = testPanic(func() {
|
|
|
|
tc.db.View(func(tx database.Tx) error {
|
|
|
|
tx.Rollback()
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
})
|
|
|
|
if !paniced {
|
|
|
|
tc.t.Error("Rollback called inside View did not panic")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure calling Commit on a managed read-write transaction panics.
|
|
|
|
paniced = testPanic(func() {
|
|
|
|
tc.db.Update(func(tx database.Tx) error {
|
|
|
|
tx.Commit()
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
})
|
|
|
|
if !paniced {
|
|
|
|
tc.t.Error("Commit called inside Update did not panic")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure calling Rollback on a managed read-write transaction panics.
|
|
|
|
paniced = testPanic(func() {
|
|
|
|
tc.db.Update(func(tx database.Tx) error {
|
|
|
|
tx.Rollback()
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
})
|
|
|
|
if !paniced {
|
|
|
|
tc.t.Error("Rollback called inside Update did not panic")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testMetadataTxInterface tests all facets of the managed read/write and
|
|
|
|
// manual transaction metadata interfaces as well as the bucket interfaces under
|
|
|
|
// them.
|
|
|
|
func testMetadataTxInterface(tc *testContext) bool {
|
|
|
|
if !testManagedTxPanics(tc) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
bucket1Name := []byte("bucket1")
|
|
|
|
err := tc.db.Update(func(tx database.Tx) error {
|
|
|
|
_, err := tx.Metadata().CreateBucket(bucket1Name)
|
|
|
|
return err
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("Update: unexpected error creating bucket: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
if !testMetadataManualTxInterface(tc) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// keyValues holds the keys and values to use when putting values
|
|
|
|
// into a bucket.
|
|
|
|
keyValues := []keyPair{
|
|
|
|
{[]byte("mtxkey1"), []byte("foo1")},
|
|
|
|
{[]byte("mtxkey2"), []byte("foo2")},
|
|
|
|
{[]byte("mtxkey3"), []byte("foo3")},
|
|
|
|
{[]byte("mtxkey4"), nil},
|
|
|
|
}
|
|
|
|
|
|
|
|
// Test the bucket interface via a managed read-only transaction.
|
|
|
|
err = tc.db.View(func(tx database.Tx) error {
|
|
|
|
metadataBucket := tx.Metadata()
|
|
|
|
if metadataBucket == nil {
|
|
|
|
return fmt.Errorf("Metadata: unexpected nil bucket")
|
|
|
|
}
|
|
|
|
|
|
|
|
bucket1 := metadataBucket.Bucket(bucket1Name)
|
|
|
|
if bucket1 == nil {
|
|
|
|
return fmt.Errorf("Bucket1: unexpected nil bucket")
|
|
|
|
}
|
|
|
|
|
|
|
|
tc.isWritable = false
|
|
|
|
if !testBucketInterface(tc, bucket1) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
if err != errSubTestFail {
|
|
|
|
tc.t.Errorf("%v", err)
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure errors returned from the user-supplied View function are
|
|
|
|
// returned.
|
|
|
|
viewError := fmt.Errorf("example view error")
|
|
|
|
err = tc.db.View(func(tx database.Tx) error {
|
|
|
|
return viewError
|
|
|
|
})
|
|
|
|
if err != viewError {
|
|
|
|
tc.t.Errorf("View: inner function error not returned - got "+
|
|
|
|
"%v, want %v", err, viewError)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Test the bucket interface via a managed read-write transaction.
|
|
|
|
// Also, put a series of values and force a rollback so the following
|
|
|
|
// code can ensure the values were not stored.
|
|
|
|
forceRollbackError := fmt.Errorf("force rollback")
|
|
|
|
err = tc.db.Update(func(tx database.Tx) error {
|
|
|
|
metadataBucket := tx.Metadata()
|
|
|
|
if metadataBucket == nil {
|
|
|
|
return fmt.Errorf("Metadata: unexpected nil bucket")
|
|
|
|
}
|
|
|
|
|
|
|
|
bucket1 := metadataBucket.Bucket(bucket1Name)
|
|
|
|
if bucket1 == nil {
|
|
|
|
return fmt.Errorf("Bucket1: unexpected nil bucket")
|
|
|
|
}
|
|
|
|
|
|
|
|
tc.isWritable = true
|
|
|
|
if !testBucketInterface(tc, bucket1) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
if !testPutValues(tc, bucket1, keyValues) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
// Return an error to force a rollback.
|
|
|
|
return forceRollbackError
|
|
|
|
})
|
|
|
|
if err != forceRollbackError {
|
|
|
|
if err == errSubTestFail {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
tc.t.Errorf("Update: inner function error not returned - got "+
|
|
|
|
"%v, want %v", err, forceRollbackError)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the values that should not have been stored due to the forced
|
|
|
|
// rollback above were not actually stored.
|
|
|
|
err = tc.db.View(func(tx database.Tx) error {
|
|
|
|
metadataBucket := tx.Metadata()
|
|
|
|
if metadataBucket == nil {
|
|
|
|
return fmt.Errorf("Metadata: unexpected nil bucket")
|
|
|
|
}
|
|
|
|
|
|
|
|
if !testGetValues(tc, metadataBucket, rollbackValues(keyValues)) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
if err != errSubTestFail {
|
|
|
|
tc.t.Errorf("%v", err)
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Store a series of values via a managed read-write transaction.
|
|
|
|
err = tc.db.Update(func(tx database.Tx) error {
|
|
|
|
metadataBucket := tx.Metadata()
|
|
|
|
if metadataBucket == nil {
|
|
|
|
return fmt.Errorf("Metadata: unexpected nil bucket")
|
|
|
|
}
|
|
|
|
|
|
|
|
bucket1 := metadataBucket.Bucket(bucket1Name)
|
|
|
|
if bucket1 == nil {
|
|
|
|
return fmt.Errorf("Bucket1: unexpected nil bucket")
|
|
|
|
}
|
|
|
|
|
|
|
|
if !testPutValues(tc, bucket1, keyValues) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
if err != errSubTestFail {
|
|
|
|
tc.t.Errorf("%v", err)
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the values stored above were committed as expected.
|
|
|
|
err = tc.db.View(func(tx database.Tx) error {
|
|
|
|
metadataBucket := tx.Metadata()
|
|
|
|
if metadataBucket == nil {
|
|
|
|
return fmt.Errorf("Metadata: unexpected nil bucket")
|
|
|
|
}
|
|
|
|
|
|
|
|
bucket1 := metadataBucket.Bucket(bucket1Name)
|
|
|
|
if bucket1 == nil {
|
|
|
|
return fmt.Errorf("Bucket1: unexpected nil bucket")
|
|
|
|
}
|
|
|
|
|
|
|
|
if !testGetValues(tc, bucket1, toGetValues(keyValues)) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
if err != errSubTestFail {
|
|
|
|
tc.t.Errorf("%v", err)
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Clean up the values stored above in a managed read-write transaction.
|
|
|
|
err = tc.db.Update(func(tx database.Tx) error {
|
|
|
|
metadataBucket := tx.Metadata()
|
|
|
|
if metadataBucket == nil {
|
|
|
|
return fmt.Errorf("Metadata: unexpected nil bucket")
|
|
|
|
}
|
|
|
|
|
|
|
|
bucket1 := metadataBucket.Bucket(bucket1Name)
|
|
|
|
if bucket1 == nil {
|
|
|
|
return fmt.Errorf("Bucket1: unexpected nil bucket")
|
|
|
|
}
|
|
|
|
|
|
|
|
if !testDeleteValues(tc, bucket1, keyValues) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
if err != errSubTestFail {
|
|
|
|
tc.t.Errorf("%v", err)
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testFetchBlockIOMissing ensures that all of the block retrieval API functions
|
|
|
|
// work as expected when requesting blocks that don't exist.
|
|
|
|
func testFetchBlockIOMissing(tc *testContext, tx database.Tx) bool {
|
|
|
|
wantErrCode := database.ErrBlockNotFound
|
|
|
|
|
|
|
|
// ---------------------
|
|
|
|
// Non-bulk Block IO API
|
|
|
|
// ---------------------
|
|
|
|
|
|
|
|
// Test the individual block APIs one block at a time to ensure they
|
|
|
|
// return the expected error. Also, build the data needed to test the
|
|
|
|
// bulk APIs below while looping.
|
2016-08-08 21:04:33 +02:00
|
|
|
allBlockHashes := make([]chainhash.Hash, len(tc.blocks))
|
2016-02-03 18:42:04 +01:00
|
|
|
allBlockRegions := make([]database.BlockRegion, len(tc.blocks))
|
|
|
|
for i, block := range tc.blocks {
|
2016-08-08 21:04:33 +02:00
|
|
|
blockHash := block.Hash()
|
2016-02-03 18:42:04 +01:00
|
|
|
allBlockHashes[i] = *blockHash
|
|
|
|
|
|
|
|
txLocs, err := block.TxLoc()
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("block.TxLoc(%d): unexpected error: %v", i,
|
|
|
|
err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure FetchBlock returns expected error.
|
|
|
|
testName := fmt.Sprintf("FetchBlock #%d on missing block", i)
|
|
|
|
_, err = tx.FetchBlock(blockHash)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure FetchBlockHeader returns expected error.
|
|
|
|
testName = fmt.Sprintf("FetchBlockHeader #%d on missing block",
|
|
|
|
i)
|
|
|
|
_, err = tx.FetchBlockHeader(blockHash)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the first transaction fetched as a block region from
|
|
|
|
// the database returns the expected error.
|
|
|
|
region := database.BlockRegion{
|
|
|
|
Hash: blockHash,
|
|
|
|
Offset: uint32(txLocs[0].TxStart),
|
|
|
|
Len: uint32(txLocs[0].TxLen),
|
|
|
|
}
|
|
|
|
allBlockRegions[i] = region
|
|
|
|
_, err = tx.FetchBlockRegion(®ion)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure HasBlock returns false.
|
|
|
|
hasBlock, err := tx.HasBlock(blockHash)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("HasBlock #%d: unexpected err: %v", i, err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if hasBlock {
|
|
|
|
tc.t.Errorf("HasBlock #%d: should not have block", i)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----------------
|
|
|
|
// Bulk Block IO API
|
|
|
|
// -----------------
|
|
|
|
|
|
|
|
// Ensure FetchBlocks returns expected error.
|
|
|
|
testName := "FetchBlocks on missing blocks"
|
|
|
|
_, err := tx.FetchBlocks(allBlockHashes)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure FetchBlockHeaders returns expected error.
|
|
|
|
testName = "FetchBlockHeaders on missing blocks"
|
|
|
|
_, err = tx.FetchBlockHeaders(allBlockHashes)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure FetchBlockRegions returns expected error.
|
|
|
|
testName = "FetchBlockRegions on missing blocks"
|
|
|
|
_, err = tx.FetchBlockRegions(allBlockRegions)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure HasBlocks returns false for all blocks.
|
|
|
|
hasBlocks, err := tx.HasBlocks(allBlockHashes)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("HasBlocks: unexpected err: %v", err)
|
|
|
|
}
|
|
|
|
for i, hasBlock := range hasBlocks {
|
|
|
|
if hasBlock {
|
|
|
|
tc.t.Errorf("HasBlocks #%d: should not have block", i)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testFetchBlockIO ensures all of the block retrieval API functions work as
|
|
|
|
// expected for the provide set of blocks. The blocks must already be stored in
|
|
|
|
// the database, or at least stored into the the passed transaction. It also
|
|
|
|
// tests several error conditions such as ensuring the expected errors are
|
|
|
|
// returned when fetching blocks, headers, and regions that don't exist.
|
|
|
|
func testFetchBlockIO(tc *testContext, tx database.Tx) bool {
|
|
|
|
// ---------------------
|
|
|
|
// Non-bulk Block IO API
|
|
|
|
// ---------------------
|
|
|
|
|
|
|
|
// Test the individual block APIs one block at a time. Also, build the
|
|
|
|
// data needed to test the bulk APIs below while looping.
|
2016-08-08 21:04:33 +02:00
|
|
|
allBlockHashes := make([]chainhash.Hash, len(tc.blocks))
|
2016-02-03 18:42:04 +01:00
|
|
|
allBlockBytes := make([][]byte, len(tc.blocks))
|
|
|
|
allBlockTxLocs := make([][]wire.TxLoc, len(tc.blocks))
|
|
|
|
allBlockRegions := make([]database.BlockRegion, len(tc.blocks))
|
|
|
|
for i, block := range tc.blocks {
|
2016-08-08 21:04:33 +02:00
|
|
|
blockHash := block.Hash()
|
2016-02-03 18:42:04 +01:00
|
|
|
allBlockHashes[i] = *blockHash
|
|
|
|
|
|
|
|
blockBytes, err := block.Bytes()
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("block.Bytes(%d): unexpected error: %v", i,
|
|
|
|
err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
allBlockBytes[i] = blockBytes
|
|
|
|
|
|
|
|
txLocs, err := block.TxLoc()
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("block.TxLoc(%d): unexpected error: %v", i,
|
|
|
|
err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
allBlockTxLocs[i] = txLocs
|
|
|
|
|
|
|
|
// Ensure the block data fetched from the database matches the
|
|
|
|
// expected bytes.
|
|
|
|
gotBlockBytes, err := tx.FetchBlock(blockHash)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("FetchBlock(%s): unexpected error: %v",
|
|
|
|
blockHash, err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if !bytes.Equal(gotBlockBytes, blockBytes) {
|
|
|
|
tc.t.Errorf("FetchBlock(%s): bytes mismatch: got %x, "+
|
|
|
|
"want %x", blockHash, gotBlockBytes, blockBytes)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the block header fetched from the database matches the
|
|
|
|
// expected bytes.
|
|
|
|
wantHeaderBytes := blockBytes[0:wire.MaxBlockHeaderPayload]
|
|
|
|
gotHeaderBytes, err := tx.FetchBlockHeader(blockHash)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("FetchBlockHeader(%s): unexpected error: %v",
|
|
|
|
blockHash, err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if !bytes.Equal(gotHeaderBytes, wantHeaderBytes) {
|
|
|
|
tc.t.Errorf("FetchBlockHeader(%s): bytes mismatch: "+
|
|
|
|
"got %x, want %x", blockHash, gotHeaderBytes,
|
|
|
|
wantHeaderBytes)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the first transaction fetched as a block region from
|
|
|
|
// the database matches the expected bytes.
|
|
|
|
region := database.BlockRegion{
|
|
|
|
Hash: blockHash,
|
|
|
|
Offset: uint32(txLocs[0].TxStart),
|
|
|
|
Len: uint32(txLocs[0].TxLen),
|
|
|
|
}
|
|
|
|
allBlockRegions[i] = region
|
|
|
|
endRegionOffset := region.Offset + region.Len
|
|
|
|
wantRegionBytes := blockBytes[region.Offset:endRegionOffset]
|
|
|
|
gotRegionBytes, err := tx.FetchBlockRegion(®ion)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("FetchBlockRegion(%s): unexpected error: %v",
|
|
|
|
blockHash, err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if !bytes.Equal(gotRegionBytes, wantRegionBytes) {
|
|
|
|
tc.t.Errorf("FetchBlockRegion(%s): bytes mismatch: "+
|
|
|
|
"got %x, want %x", blockHash, gotRegionBytes,
|
|
|
|
wantRegionBytes)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
2019-03-03 02:42:32 +01:00
|
|
|
// Ensure block hash exists as expected.
|
2016-02-03 18:42:04 +01:00
|
|
|
hasBlock, err := tx.HasBlock(blockHash)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("HasBlock(%s): unexpected error: %v",
|
|
|
|
blockHash, err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if !hasBlock {
|
|
|
|
tc.t.Errorf("HasBlock(%s): database claims it doesn't "+
|
|
|
|
"have the block when it should", blockHash)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----------------------
|
|
|
|
// Invalid blocks/regions.
|
|
|
|
// -----------------------
|
|
|
|
|
|
|
|
// Ensure fetching a block that doesn't exist returns the
|
|
|
|
// expected error.
|
2016-08-08 21:04:33 +02:00
|
|
|
badBlockHash := &chainhash.Hash{}
|
2016-02-03 18:42:04 +01:00
|
|
|
testName := fmt.Sprintf("FetchBlock(%s) invalid block",
|
|
|
|
badBlockHash)
|
|
|
|
wantErrCode := database.ErrBlockNotFound
|
|
|
|
_, err = tx.FetchBlock(badBlockHash)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure fetching a block header that doesn't exist returns
|
|
|
|
// the expected error.
|
|
|
|
testName = fmt.Sprintf("FetchBlockHeader(%s) invalid block",
|
|
|
|
badBlockHash)
|
|
|
|
_, err = tx.FetchBlockHeader(badBlockHash)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure fetching a block region in a block that doesn't exist
|
|
|
|
// return the expected error.
|
|
|
|
testName = fmt.Sprintf("FetchBlockRegion(%s) invalid hash",
|
|
|
|
badBlockHash)
|
|
|
|
wantErrCode = database.ErrBlockNotFound
|
|
|
|
region.Hash = badBlockHash
|
|
|
|
region.Offset = ^uint32(0)
|
|
|
|
_, err = tx.FetchBlockRegion(®ion)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure fetching a block region that is out of bounds returns
|
|
|
|
// the expected error.
|
|
|
|
testName = fmt.Sprintf("FetchBlockRegion(%s) invalid region",
|
|
|
|
blockHash)
|
|
|
|
wantErrCode = database.ErrBlockRegionInvalid
|
|
|
|
region.Hash = blockHash
|
|
|
|
region.Offset = ^uint32(0)
|
|
|
|
_, err = tx.FetchBlockRegion(®ion)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----------------
|
|
|
|
// Bulk Block IO API
|
|
|
|
// -----------------
|
|
|
|
|
|
|
|
// Ensure the bulk block data fetched from the database matches the
|
|
|
|
// expected bytes.
|
|
|
|
blockData, err := tx.FetchBlocks(allBlockHashes)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("FetchBlocks: unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if len(blockData) != len(allBlockBytes) {
|
|
|
|
tc.t.Errorf("FetchBlocks: unexpected number of results - got "+
|
|
|
|
"%d, want %d", len(blockData), len(allBlockBytes))
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
for i := 0; i < len(blockData); i++ {
|
|
|
|
blockHash := allBlockHashes[i]
|
|
|
|
wantBlockBytes := allBlockBytes[i]
|
|
|
|
gotBlockBytes := blockData[i]
|
|
|
|
if !bytes.Equal(gotBlockBytes, wantBlockBytes) {
|
|
|
|
tc.t.Errorf("FetchBlocks(%s): bytes mismatch: got %x, "+
|
|
|
|
"want %x", blockHash, gotBlockBytes,
|
|
|
|
wantBlockBytes)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the bulk block headers fetched from the database match the
|
|
|
|
// expected bytes.
|
|
|
|
blockHeaderData, err := tx.FetchBlockHeaders(allBlockHashes)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("FetchBlockHeaders: unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if len(blockHeaderData) != len(allBlockBytes) {
|
|
|
|
tc.t.Errorf("FetchBlockHeaders: unexpected number of results "+
|
|
|
|
"- got %d, want %d", len(blockHeaderData),
|
|
|
|
len(allBlockBytes))
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
for i := 0; i < len(blockHeaderData); i++ {
|
|
|
|
blockHash := allBlockHashes[i]
|
|
|
|
wantHeaderBytes := allBlockBytes[i][0:wire.MaxBlockHeaderPayload]
|
|
|
|
gotHeaderBytes := blockHeaderData[i]
|
|
|
|
if !bytes.Equal(gotHeaderBytes, wantHeaderBytes) {
|
|
|
|
tc.t.Errorf("FetchBlockHeaders(%s): bytes mismatch: "+
|
|
|
|
"got %x, want %x", blockHash, gotHeaderBytes,
|
|
|
|
wantHeaderBytes)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the first transaction of every block fetched in bulk block
|
|
|
|
// regions from the database matches the expected bytes.
|
|
|
|
allRegionBytes, err := tx.FetchBlockRegions(allBlockRegions)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("FetchBlockRegions: unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
|
|
|
|
}
|
|
|
|
if len(allRegionBytes) != len(allBlockRegions) {
|
|
|
|
tc.t.Errorf("FetchBlockRegions: unexpected number of results "+
|
|
|
|
"- got %d, want %d", len(allRegionBytes),
|
|
|
|
len(allBlockRegions))
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
for i, gotRegionBytes := range allRegionBytes {
|
|
|
|
region := &allBlockRegions[i]
|
|
|
|
endRegionOffset := region.Offset + region.Len
|
|
|
|
wantRegionBytes := blockData[i][region.Offset:endRegionOffset]
|
|
|
|
if !bytes.Equal(gotRegionBytes, wantRegionBytes) {
|
|
|
|
tc.t.Errorf("FetchBlockRegions(%d): bytes mismatch: "+
|
|
|
|
"got %x, want %x", i, gotRegionBytes,
|
|
|
|
wantRegionBytes)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the bulk determination of whether a set of block hashes are in
|
|
|
|
// the database returns true for all loaded blocks.
|
|
|
|
hasBlocks, err := tx.HasBlocks(allBlockHashes)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("HasBlocks: unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
for i, hasBlock := range hasBlocks {
|
|
|
|
if !hasBlock {
|
|
|
|
tc.t.Errorf("HasBlocks(%d): should have block", i)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----------------------
|
|
|
|
// Invalid blocks/regions.
|
|
|
|
// -----------------------
|
|
|
|
|
|
|
|
// Ensure fetching blocks for which one doesn't exist returns the
|
|
|
|
// expected error.
|
|
|
|
testName := "FetchBlocks invalid hash"
|
2016-08-08 21:04:33 +02:00
|
|
|
badBlockHashes := make([]chainhash.Hash, len(allBlockHashes)+1)
|
2016-02-03 18:42:04 +01:00
|
|
|
copy(badBlockHashes, allBlockHashes)
|
2016-08-08 21:04:33 +02:00
|
|
|
badBlockHashes[len(badBlockHashes)-1] = chainhash.Hash{}
|
2016-02-03 18:42:04 +01:00
|
|
|
wantErrCode := database.ErrBlockNotFound
|
|
|
|
_, err = tx.FetchBlocks(badBlockHashes)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure fetching block headers for which one doesn't exist returns the
|
|
|
|
// expected error.
|
|
|
|
testName = "FetchBlockHeaders invalid hash"
|
|
|
|
_, err = tx.FetchBlockHeaders(badBlockHashes)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure fetching block regions for which one of blocks doesn't exist
|
|
|
|
// returns expected error.
|
|
|
|
testName = "FetchBlockRegions invalid hash"
|
|
|
|
badBlockRegions := make([]database.BlockRegion, len(allBlockRegions)+1)
|
|
|
|
copy(badBlockRegions, allBlockRegions)
|
2016-08-08 21:04:33 +02:00
|
|
|
badBlockRegions[len(badBlockRegions)-1].Hash = &chainhash.Hash{}
|
2016-02-03 18:42:04 +01:00
|
|
|
wantErrCode = database.ErrBlockNotFound
|
|
|
|
_, err = tx.FetchBlockRegions(badBlockRegions)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure fetching block regions that are out of bounds returns the
|
|
|
|
// expected error.
|
|
|
|
testName = "FetchBlockRegions invalid regions"
|
|
|
|
badBlockRegions = badBlockRegions[:len(badBlockRegions)-1]
|
|
|
|
for i := range badBlockRegions {
|
|
|
|
badBlockRegions[i].Offset = ^uint32(0)
|
|
|
|
}
|
|
|
|
wantErrCode = database.ErrBlockRegionInvalid
|
|
|
|
_, err = tx.FetchBlockRegions(badBlockRegions)
|
2016-11-03 05:02:04 +01:00
|
|
|
return checkDbError(tc.t, testName, err, wantErrCode)
|
2016-02-03 18:42:04 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// testBlockIOTxInterface ensures that the block IO interface works as expected
|
|
|
|
// for both managed read/write and manual transactions. This function leaves
|
|
|
|
// all of the stored blocks in the database.
|
|
|
|
func testBlockIOTxInterface(tc *testContext) bool {
|
|
|
|
// Ensure attempting to store a block with a read-only transaction fails
|
|
|
|
// with the expected error.
|
|
|
|
err := tc.db.View(func(tx database.Tx) error {
|
|
|
|
wantErrCode := database.ErrTxNotWritable
|
|
|
|
for i, block := range tc.blocks {
|
|
|
|
testName := fmt.Sprintf("StoreBlock(%d) on ro tx", i)
|
|
|
|
err := tx.StoreBlock(block)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
if err != errSubTestFail {
|
|
|
|
tc.t.Errorf("%v", err)
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Populate the database with loaded blocks and ensure all of the data
|
|
|
|
// fetching APIs work properly on them within the transaction before a
|
|
|
|
// commit or rollback. Then, force a rollback so the code below can
|
|
|
|
// ensure none of the data actually gets stored.
|
|
|
|
forceRollbackError := fmt.Errorf("force rollback")
|
|
|
|
err = tc.db.Update(func(tx database.Tx) error {
|
|
|
|
// Store all blocks in the same transaction.
|
|
|
|
for i, block := range tc.blocks {
|
|
|
|
err := tx.StoreBlock(block)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("StoreBlock #%d: unexpected error: "+
|
|
|
|
"%v", i, err)
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure attempting to store the same block again, before the
|
|
|
|
// transaction has been committed, returns the expected error.
|
|
|
|
wantErrCode := database.ErrBlockExists
|
|
|
|
for i, block := range tc.blocks {
|
|
|
|
testName := fmt.Sprintf("duplicate block entry #%d "+
|
|
|
|
"(before commit)", i)
|
|
|
|
err := tx.StoreBlock(block)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure that all data fetches from the stored blocks before
|
|
|
|
// the transaction has been committed work as expected.
|
|
|
|
if !testFetchBlockIO(tc, tx) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
return forceRollbackError
|
|
|
|
})
|
|
|
|
if err != forceRollbackError {
|
|
|
|
if err == errSubTestFail {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
tc.t.Errorf("Update: inner function error not returned - got "+
|
|
|
|
"%v, want %v", err, forceRollbackError)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure rollback was successful
|
|
|
|
err = tc.db.View(func(tx database.Tx) error {
|
|
|
|
if !testFetchBlockIOMissing(tc, tx) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
if err != errSubTestFail {
|
|
|
|
tc.t.Errorf("%v", err)
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Populate the database with loaded blocks and ensure all of the data
|
|
|
|
// fetching APIs work properly.
|
|
|
|
err = tc.db.Update(func(tx database.Tx) error {
|
|
|
|
// Store a bunch of blocks in the same transaction.
|
|
|
|
for i, block := range tc.blocks {
|
|
|
|
err := tx.StoreBlock(block)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("StoreBlock #%d: unexpected error: "+
|
|
|
|
"%v", i, err)
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure attempting to store the same block again while in the
|
|
|
|
// same transaction, but before it has been committed, returns
|
|
|
|
// the expected error.
|
|
|
|
for i, block := range tc.blocks {
|
|
|
|
testName := fmt.Sprintf("duplicate block entry #%d "+
|
|
|
|
"(before commit)", i)
|
|
|
|
wantErrCode := database.ErrBlockExists
|
|
|
|
err := tx.StoreBlock(block)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure that all data fetches from the stored blocks before
|
|
|
|
// the transaction has been committed work as expected.
|
|
|
|
if !testFetchBlockIO(tc, tx) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
if err != errSubTestFail {
|
|
|
|
tc.t.Errorf("%v", err)
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure all data fetch tests work as expected using a managed
|
|
|
|
// read-only transaction after the data was successfully committed
|
|
|
|
// above.
|
|
|
|
err = tc.db.View(func(tx database.Tx) error {
|
|
|
|
if !testFetchBlockIO(tc, tx) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
if err != errSubTestFail {
|
|
|
|
tc.t.Errorf("%v", err)
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure all data fetch tests work as expected using a managed
|
|
|
|
// read-write transaction after the data was successfully committed
|
|
|
|
// above.
|
|
|
|
err = tc.db.Update(func(tx database.Tx) error {
|
|
|
|
if !testFetchBlockIO(tc, tx) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure attempting to store existing blocks again returns the
|
|
|
|
// expected error. Note that this is different from the
|
|
|
|
// previous version since this is a new transaction after the
|
|
|
|
// blocks have been committed.
|
|
|
|
wantErrCode := database.ErrBlockExists
|
|
|
|
for i, block := range tc.blocks {
|
|
|
|
testName := fmt.Sprintf("duplicate block entry #%d "+
|
|
|
|
"(before commit)", i)
|
|
|
|
err := tx.StoreBlock(block)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return errSubTestFail
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
if err != errSubTestFail {
|
|
|
|
tc.t.Errorf("%v", err)
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testClosedTxInterface ensures that both the metadata and block IO API
|
|
|
|
// functions behave as expected when attempted against a closed transaction.
|
|
|
|
func testClosedTxInterface(tc *testContext, tx database.Tx) bool {
|
|
|
|
wantErrCode := database.ErrTxClosed
|
|
|
|
bucket := tx.Metadata()
|
|
|
|
cursor := tx.Metadata().Cursor()
|
|
|
|
bucketName := []byte("closedtxbucket")
|
|
|
|
keyName := []byte("closedtxkey")
|
|
|
|
|
|
|
|
// ------------
|
|
|
|
// Metadata API
|
|
|
|
// ------------
|
|
|
|
|
|
|
|
// Ensure that attempting to get an existing bucket returns nil when the
|
|
|
|
// transaction is closed.
|
|
|
|
if b := bucket.Bucket(bucketName); b != nil {
|
|
|
|
tc.t.Errorf("Bucket: did not return nil on closed tx")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure CreateBucket returns expected error.
|
|
|
|
testName := "CreateBucket on closed tx"
|
|
|
|
_, err := bucket.CreateBucket(bucketName)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure CreateBucketIfNotExists returns expected error.
|
|
|
|
testName = "CreateBucketIfNotExists on closed tx"
|
|
|
|
_, err = bucket.CreateBucketIfNotExists(bucketName)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure Delete returns expected error.
|
|
|
|
testName = "Delete on closed tx"
|
|
|
|
err = bucket.Delete(keyName)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure DeleteBucket returns expected error.
|
|
|
|
testName = "DeleteBucket on closed tx"
|
|
|
|
err = bucket.DeleteBucket(bucketName)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure ForEach returns expected error.
|
|
|
|
testName = "ForEach on closed tx"
|
|
|
|
err = bucket.ForEach(nil)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure ForEachBucket returns expected error.
|
|
|
|
testName = "ForEachBucket on closed tx"
|
|
|
|
err = bucket.ForEachBucket(nil)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure Get returns expected error.
|
|
|
|
testName = "Get on closed tx"
|
|
|
|
if k := bucket.Get(keyName); k != nil {
|
|
|
|
tc.t.Errorf("Get: did not return nil on closed tx")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure Put returns expected error.
|
|
|
|
testName = "Put on closed tx"
|
|
|
|
err = bucket.Put(keyName, []byte("test"))
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// -------------------
|
|
|
|
// Metadata Cursor API
|
|
|
|
// -------------------
|
|
|
|
|
|
|
|
// Ensure attempting to get a bucket from a cursor on a closed tx gives
|
|
|
|
// back nil.
|
|
|
|
if b := cursor.Bucket(); b != nil {
|
|
|
|
tc.t.Error("Cursor.Bucket: returned non-nil on closed tx")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure Cursor.Delete returns expected error.
|
|
|
|
testName = "Cursor.Delete on closed tx"
|
|
|
|
err = cursor.Delete()
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure Cursor.First on a closed tx returns false and nil key/value.
|
|
|
|
if cursor.First() {
|
|
|
|
tc.t.Error("Cursor.First: claims ok on closed tx")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if cursor.Key() != nil || cursor.Value() != nil {
|
|
|
|
tc.t.Error("Cursor.First: key and/or value are not nil on " +
|
|
|
|
"closed tx")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure Cursor.Last on a closed tx returns false and nil key/value.
|
|
|
|
if cursor.Last() {
|
|
|
|
tc.t.Error("Cursor.Last: claims ok on closed tx")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if cursor.Key() != nil || cursor.Value() != nil {
|
|
|
|
tc.t.Error("Cursor.Last: key and/or value are not nil on " +
|
|
|
|
"closed tx")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure Cursor.Next on a closed tx returns false and nil key/value.
|
|
|
|
if cursor.Next() {
|
|
|
|
tc.t.Error("Cursor.Next: claims ok on closed tx")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if cursor.Key() != nil || cursor.Value() != nil {
|
|
|
|
tc.t.Error("Cursor.Next: key and/or value are not nil on " +
|
|
|
|
"closed tx")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure Cursor.Prev on a closed tx returns false and nil key/value.
|
|
|
|
if cursor.Prev() {
|
|
|
|
tc.t.Error("Cursor.Prev: claims ok on closed tx")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if cursor.Key() != nil || cursor.Value() != nil {
|
|
|
|
tc.t.Error("Cursor.Prev: key and/or value are not nil on " +
|
|
|
|
"closed tx")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure Cursor.Seek on a closed tx returns false and nil key/value.
|
|
|
|
if cursor.Seek([]byte{}) {
|
|
|
|
tc.t.Error("Cursor.Seek: claims ok on closed tx")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if cursor.Key() != nil || cursor.Value() != nil {
|
|
|
|
tc.t.Error("Cursor.Seek: key and/or value are not nil on " +
|
|
|
|
"closed tx")
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// ---------------------
|
|
|
|
// Non-bulk Block IO API
|
|
|
|
// ---------------------
|
|
|
|
|
|
|
|
// Test the individual block APIs one block at a time to ensure they
|
|
|
|
// return the expected error. Also, build the data needed to test the
|
|
|
|
// bulk APIs below while looping.
|
2016-08-08 21:04:33 +02:00
|
|
|
allBlockHashes := make([]chainhash.Hash, len(tc.blocks))
|
2016-02-03 18:42:04 +01:00
|
|
|
allBlockRegions := make([]database.BlockRegion, len(tc.blocks))
|
|
|
|
for i, block := range tc.blocks {
|
2016-08-08 21:04:33 +02:00
|
|
|
blockHash := block.Hash()
|
2016-02-03 18:42:04 +01:00
|
|
|
allBlockHashes[i] = *blockHash
|
|
|
|
|
|
|
|
txLocs, err := block.TxLoc()
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("block.TxLoc(%d): unexpected error: %v", i,
|
|
|
|
err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure StoreBlock returns expected error.
|
|
|
|
testName = "StoreBlock on closed tx"
|
|
|
|
err = tx.StoreBlock(block)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure FetchBlock returns expected error.
|
|
|
|
testName = fmt.Sprintf("FetchBlock #%d on closed tx", i)
|
|
|
|
_, err = tx.FetchBlock(blockHash)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure FetchBlockHeader returns expected error.
|
|
|
|
testName = fmt.Sprintf("FetchBlockHeader #%d on closed tx", i)
|
|
|
|
_, err = tx.FetchBlockHeader(blockHash)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure the first transaction fetched as a block region from
|
|
|
|
// the database returns the expected error.
|
|
|
|
region := database.BlockRegion{
|
|
|
|
Hash: blockHash,
|
|
|
|
Offset: uint32(txLocs[0].TxStart),
|
|
|
|
Len: uint32(txLocs[0].TxLen),
|
|
|
|
}
|
|
|
|
allBlockRegions[i] = region
|
|
|
|
_, err = tx.FetchBlockRegion(®ion)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure HasBlock returns expected error.
|
|
|
|
testName = fmt.Sprintf("HasBlock #%d on closed tx", i)
|
|
|
|
_, err = tx.HasBlock(blockHash)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// -----------------
|
|
|
|
// Bulk Block IO API
|
|
|
|
// -----------------
|
|
|
|
|
|
|
|
// Ensure FetchBlocks returns expected error.
|
|
|
|
testName = "FetchBlocks on closed tx"
|
|
|
|
_, err = tx.FetchBlocks(allBlockHashes)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure FetchBlockHeaders returns expected error.
|
|
|
|
testName = "FetchBlockHeaders on closed tx"
|
|
|
|
_, err = tx.FetchBlockHeaders(allBlockHashes)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure FetchBlockRegions returns expected error.
|
|
|
|
testName = "FetchBlockRegions on closed tx"
|
|
|
|
_, err = tx.FetchBlockRegions(allBlockRegions)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure HasBlocks returns expected error.
|
|
|
|
testName = "HasBlocks on closed tx"
|
|
|
|
_, err = tx.HasBlocks(allBlockHashes)
|
|
|
|
if !checkDbError(tc.t, testName, err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// ---------------
|
|
|
|
// Commit/Rollback
|
|
|
|
// ---------------
|
|
|
|
|
|
|
|
// Ensure that attempting to rollback or commit a transaction that is
|
|
|
|
// already closed returns the expected error.
|
|
|
|
err = tx.Rollback()
|
|
|
|
if !checkDbError(tc.t, "closed tx rollback", err, wantErrCode) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
err = tx.Commit()
|
2016-11-03 05:02:04 +01:00
|
|
|
return checkDbError(tc.t, "closed tx commit", err, wantErrCode)
|
2016-02-03 18:42:04 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// testTxClosed ensures that both the metadata and block IO API functions behave
|
|
|
|
// as expected when attempted against both read-only and read-write
|
|
|
|
// transactions.
|
|
|
|
func testTxClosed(tc *testContext) bool {
|
|
|
|
bucketName := []byte("closedtxbucket")
|
|
|
|
keyName := []byte("closedtxkey")
|
|
|
|
|
|
|
|
// Start a transaction, create a bucket and key used for testing, and
|
|
|
|
// immediately perform a commit on it so it is closed.
|
|
|
|
tx, err := tc.db.Begin(true)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("Begin(true): unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
defer rollbackOnPanic(tc.t, tx)
|
|
|
|
if _, err := tx.Metadata().CreateBucket(bucketName); err != nil {
|
|
|
|
tc.t.Errorf("CreateBucket: unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if err := tx.Metadata().Put(keyName, []byte("test")); err != nil {
|
|
|
|
tc.t.Errorf("Put: unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if err := tx.Commit(); err != nil {
|
|
|
|
tc.t.Errorf("Commit: unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure invoking all of the functions on the closed read-write
|
|
|
|
// transaction behave as expected.
|
|
|
|
if !testClosedTxInterface(tc, tx) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Repeat the tests with a rolled-back read-only transaction.
|
|
|
|
tx, err = tc.db.Begin(false)
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("Begin(false): unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
defer rollbackOnPanic(tc.t, tx)
|
|
|
|
if err := tx.Rollback(); err != nil {
|
|
|
|
tc.t.Errorf("Rollback: unexpected error: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure invoking all of the functions on the closed read-only
|
|
|
|
// transaction behave as expected.
|
|
|
|
return testClosedTxInterface(tc, tx)
|
|
|
|
}
|
|
|
|
|
|
|
|
// testConcurrecy ensure the database properly supports concurrent readers and
|
|
|
|
// only a single writer. It also ensures views act as snapshots at the time
|
|
|
|
// they are acquired.
|
|
|
|
func testConcurrecy(tc *testContext) bool {
|
|
|
|
// sleepTime is how long each of the concurrent readers should sleep to
|
|
|
|
// aid in detection of whether or not the data is actually being read
|
|
|
|
// concurrently. It starts with a sane lower bound.
|
|
|
|
var sleepTime = time.Millisecond * 250
|
|
|
|
|
|
|
|
// Determine about how long it takes for a single block read. When it's
|
|
|
|
// longer than the default minimum sleep time, adjust the sleep time to
|
|
|
|
// help prevent durations that are too short which would cause erroneous
|
|
|
|
// test failures on slower systems.
|
|
|
|
startTime := time.Now()
|
|
|
|
err := tc.db.View(func(tx database.Tx) error {
|
2016-08-08 21:04:33 +02:00
|
|
|
_, err := tx.FetchBlock(tc.blocks[0].Hash())
|
2016-11-03 05:02:04 +01:00
|
|
|
return err
|
2016-02-03 18:42:04 +01:00
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("Unexpected error in view: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
2016-11-03 05:02:04 +01:00
|
|
|
elapsed := time.Since(startTime)
|
2016-02-03 18:42:04 +01:00
|
|
|
if sleepTime < elapsed {
|
|
|
|
sleepTime = elapsed
|
|
|
|
}
|
|
|
|
tc.t.Logf("Time to load block 0: %v, using sleep time: %v", elapsed,
|
|
|
|
sleepTime)
|
|
|
|
|
|
|
|
// reader takes a block number to load and channel to return the result
|
|
|
|
// of the operation on. It is used below to launch multiple concurrent
|
|
|
|
// readers.
|
|
|
|
numReaders := len(tc.blocks)
|
|
|
|
resultChan := make(chan bool, numReaders)
|
|
|
|
reader := func(blockNum int) {
|
|
|
|
err := tc.db.View(func(tx database.Tx) error {
|
|
|
|
time.Sleep(sleepTime)
|
2016-08-08 21:04:33 +02:00
|
|
|
_, err := tx.FetchBlock(tc.blocks[blockNum].Hash())
|
2016-11-03 05:02:04 +01:00
|
|
|
return err
|
2016-02-03 18:42:04 +01:00
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("Unexpected error in concurrent view: %v",
|
|
|
|
err)
|
|
|
|
resultChan <- false
|
|
|
|
}
|
|
|
|
resultChan <- true
|
|
|
|
}
|
|
|
|
|
|
|
|
// Start up several concurrent readers for the same block and wait for
|
|
|
|
// the results.
|
|
|
|
startTime = time.Now()
|
|
|
|
for i := 0; i < numReaders; i++ {
|
|
|
|
go reader(0)
|
|
|
|
}
|
|
|
|
for i := 0; i < numReaders; i++ {
|
|
|
|
if result := <-resultChan; !result {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
2016-11-03 05:02:04 +01:00
|
|
|
elapsed = time.Since(startTime)
|
2016-02-03 18:42:04 +01:00
|
|
|
tc.t.Logf("%d concurrent reads of same block elapsed: %v", numReaders,
|
|
|
|
elapsed)
|
|
|
|
|
|
|
|
// Consider it a failure if it took longer than half the time it would
|
|
|
|
// take with no concurrency.
|
|
|
|
if elapsed > sleepTime*time.Duration(numReaders/2) {
|
|
|
|
tc.t.Errorf("Concurrent views for same block did not appear to "+
|
|
|
|
"run simultaneously: elapsed %v", elapsed)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Start up several concurrent readers for different blocks and wait for
|
|
|
|
// the results.
|
|
|
|
startTime = time.Now()
|
|
|
|
for i := 0; i < numReaders; i++ {
|
|
|
|
go reader(i)
|
|
|
|
}
|
|
|
|
for i := 0; i < numReaders; i++ {
|
|
|
|
if result := <-resultChan; !result {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
2016-11-03 05:02:04 +01:00
|
|
|
elapsed = time.Since(startTime)
|
2016-02-03 18:42:04 +01:00
|
|
|
tc.t.Logf("%d concurrent reads of different blocks elapsed: %v",
|
|
|
|
numReaders, elapsed)
|
|
|
|
|
|
|
|
// Consider it a failure if it took longer than half the time it would
|
|
|
|
// take with no concurrency.
|
|
|
|
if elapsed > sleepTime*time.Duration(numReaders/2) {
|
|
|
|
tc.t.Errorf("Concurrent views for different blocks did not "+
|
|
|
|
"appear to run simultaneously: elapsed %v", elapsed)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Start up a few readers and wait for them to acquire views. Each
|
|
|
|
// reader waits for a signal from the writer to be finished to ensure
|
|
|
|
// that the data written by the writer is not seen by the view since it
|
|
|
|
// was started before the data was set.
|
|
|
|
concurrentKey := []byte("notthere")
|
|
|
|
concurrentVal := []byte("someval")
|
|
|
|
started := make(chan struct{})
|
|
|
|
writeComplete := make(chan struct{})
|
|
|
|
reader = func(blockNum int) {
|
|
|
|
err := tc.db.View(func(tx database.Tx) error {
|
|
|
|
started <- struct{}{}
|
|
|
|
|
|
|
|
// Wait for the writer to complete.
|
|
|
|
<-writeComplete
|
|
|
|
|
|
|
|
// Since this reader was created before the write took
|
|
|
|
// place, the data it added should not be visible.
|
|
|
|
val := tx.Metadata().Get(concurrentKey)
|
|
|
|
if val != nil {
|
|
|
|
return fmt.Errorf("%s should not be visible",
|
|
|
|
concurrentKey)
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("Unexpected error in concurrent view: %v",
|
|
|
|
err)
|
|
|
|
resultChan <- false
|
|
|
|
}
|
|
|
|
resultChan <- true
|
|
|
|
}
|
|
|
|
for i := 0; i < numReaders; i++ {
|
|
|
|
go reader(0)
|
|
|
|
}
|
|
|
|
for i := 0; i < numReaders; i++ {
|
|
|
|
<-started
|
|
|
|
}
|
|
|
|
|
|
|
|
// All readers are started and waiting for completion of the writer.
|
|
|
|
// Set some data the readers are expecting to not find and signal the
|
|
|
|
// readers the write is done by closing the writeComplete channel.
|
|
|
|
err = tc.db.Update(func(tx database.Tx) error {
|
2016-11-03 05:02:04 +01:00
|
|
|
return tx.Metadata().Put(concurrentKey, concurrentVal)
|
2016-02-03 18:42:04 +01:00
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("Unexpected error in update: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
close(writeComplete)
|
|
|
|
|
|
|
|
// Wait for reader results.
|
|
|
|
for i := 0; i < numReaders; i++ {
|
|
|
|
if result := <-resultChan; !result {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Start a few writers and ensure the total time is at least the
|
|
|
|
// writeSleepTime * numWriters. This ensures only one write transaction
|
|
|
|
// can be active at a time.
|
|
|
|
writeSleepTime := time.Millisecond * 250
|
|
|
|
writer := func() {
|
|
|
|
err := tc.db.Update(func(tx database.Tx) error {
|
|
|
|
time.Sleep(writeSleepTime)
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("Unexpected error in concurrent view: %v",
|
|
|
|
err)
|
|
|
|
resultChan <- false
|
|
|
|
}
|
|
|
|
resultChan <- true
|
|
|
|
}
|
|
|
|
numWriters := 3
|
|
|
|
startTime = time.Now()
|
|
|
|
for i := 0; i < numWriters; i++ {
|
|
|
|
go writer()
|
|
|
|
}
|
|
|
|
for i := 0; i < numWriters; i++ {
|
|
|
|
if result := <-resultChan; !result {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
2016-11-03 05:02:04 +01:00
|
|
|
elapsed = time.Since(startTime)
|
2016-02-03 18:42:04 +01:00
|
|
|
tc.t.Logf("%d concurrent writers elapsed using sleep time %v: %v",
|
|
|
|
numWriters, writeSleepTime, elapsed)
|
|
|
|
|
|
|
|
// The total time must have been at least the sum of all sleeps if the
|
|
|
|
// writes blocked properly.
|
|
|
|
if elapsed < writeSleepTime*time.Duration(numWriters) {
|
|
|
|
tc.t.Errorf("Concurrent writes appeared to run simultaneously: "+
|
|
|
|
"elapsed %v", elapsed)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testConcurrentClose ensures that closing the database with open transactions
|
|
|
|
// blocks until the transactions are finished.
|
|
|
|
//
|
|
|
|
// The database will be closed upon returning from this function.
|
|
|
|
func testConcurrentClose(tc *testContext) bool {
|
|
|
|
// Start up a few readers and wait for them to acquire views. Each
|
|
|
|
// reader waits for a signal to complete to ensure the transactions stay
|
|
|
|
// open until they are explicitly signalled to be closed.
|
|
|
|
var activeReaders int32
|
|
|
|
numReaders := 3
|
|
|
|
started := make(chan struct{})
|
|
|
|
finishReaders := make(chan struct{})
|
|
|
|
resultChan := make(chan bool, numReaders+1)
|
|
|
|
reader := func() {
|
|
|
|
err := tc.db.View(func(tx database.Tx) error {
|
|
|
|
atomic.AddInt32(&activeReaders, 1)
|
|
|
|
started <- struct{}{}
|
|
|
|
<-finishReaders
|
|
|
|
atomic.AddInt32(&activeReaders, -1)
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("Unexpected error in concurrent view: %v",
|
|
|
|
err)
|
|
|
|
resultChan <- false
|
|
|
|
}
|
|
|
|
resultChan <- true
|
|
|
|
}
|
|
|
|
for i := 0; i < numReaders; i++ {
|
|
|
|
go reader()
|
|
|
|
}
|
|
|
|
for i := 0; i < numReaders; i++ {
|
|
|
|
<-started
|
|
|
|
}
|
|
|
|
|
|
|
|
// Close the database in a separate goroutine. This should block until
|
|
|
|
// the transactions are finished. Once the close has taken place, the
|
|
|
|
// dbClosed channel is closed to signal the main goroutine below.
|
|
|
|
dbClosed := make(chan struct{})
|
|
|
|
go func() {
|
|
|
|
started <- struct{}{}
|
|
|
|
err := tc.db.Close()
|
|
|
|
if err != nil {
|
|
|
|
tc.t.Errorf("Unexpected error in concurrent view: %v",
|
|
|
|
err)
|
|
|
|
resultChan <- false
|
|
|
|
}
|
|
|
|
close(dbClosed)
|
|
|
|
resultChan <- true
|
|
|
|
}()
|
|
|
|
<-started
|
|
|
|
|
|
|
|
// Wait a short period and then signal the reader transactions to
|
|
|
|
// finish. When the db closed channel is received, ensure there are no
|
|
|
|
// active readers open.
|
|
|
|
time.AfterFunc(time.Millisecond*250, func() { close(finishReaders) })
|
|
|
|
<-dbClosed
|
|
|
|
if nr := atomic.LoadInt32(&activeReaders); nr != 0 {
|
|
|
|
tc.t.Errorf("Close did not appear to block with active "+
|
|
|
|
"readers: %d active", nr)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wait for all results.
|
|
|
|
for i := 0; i < numReaders+1; i++ {
|
|
|
|
if result := <-resultChan; !result {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// testInterface tests performs tests for the various interfaces of the database
|
|
|
|
// package which require state in the database for the given database type.
|
|
|
|
func testInterface(t *testing.T, db database.DB) {
|
|
|
|
// Create a test context to pass around.
|
|
|
|
context := testContext{t: t, db: db}
|
|
|
|
|
|
|
|
// Load the test blocks and store in the test context for use throughout
|
|
|
|
// the tests.
|
|
|
|
blocks, err := loadBlocks(t, blockDataFile, blockDataNet)
|
|
|
|
if err != nil {
|
|
|
|
t.Errorf("loadBlocks: Unexpected error: %v", err)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
context.blocks = blocks
|
|
|
|
|
|
|
|
// Test the transaction metadata interface including managed and manual
|
|
|
|
// transactions as well as buckets.
|
|
|
|
if !testMetadataTxInterface(&context) {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// Test the transaction block IO interface using managed and manual
|
|
|
|
// transactions. This function leaves all of the stored blocks in the
|
|
|
|
// database since they're used later.
|
|
|
|
if !testBlockIOTxInterface(&context) {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// Test all of the transaction interface functions against a closed
|
|
|
|
// transaction work as expected.
|
|
|
|
if !testTxClosed(&context) {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// Test the database properly supports concurrency.
|
|
|
|
if !testConcurrecy(&context) {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// Test that closing the database with open transactions blocks until
|
|
|
|
// the transactions are finished.
|
|
|
|
//
|
|
|
|
// The database will be closed upon returning from this function, so it
|
|
|
|
// must be the last thing called.
|
|
|
|
testConcurrentClose(&context)
|
|
|
|
}
|