35f98ce162
-Added travis support -updated travis to analyze code beneath the root. -refactored upload.go to fix travis errors. -gocyclo should ignore test files. $GOFILES needed to be adjusted. -fix rows.Close() ignoring error. Created func to handle so defer can be used when needed also. -fixed ignored errors. -fixed unit test that was not passing correctly to anonymous function. -fixed govet error for passing param inside go func. -removed returned error, in favor of logging instead. -added error logging for ignored error. -fixed potential race conditions. -removed unused append -fixed time usage to align with go standards. -removed unused variables -made changes for code review. -code comments for exported functions. -Documented bitmap.go and insert into contact list. -Documented dht, message, bootstrap -Fixed comment typos -Documented message,node, routing_table, testing in DHT package. -Documented server, client, prism, server and shared in peer and reflector packages. -Documented the stores in Store package. -made defer adjustments inline and deleted the separate function. -adjusted method in upload to take the only parameter it requires.
324 lines
8.2 KiB
Go
324 lines
8.2 KiB
Go
package db
|
|
|
|
import (
|
|
"database/sql"
|
|
|
|
"github.com/lbryio/lbry.go/errors"
|
|
"github.com/lbryio/lbry.go/querytools"
|
|
"github.com/lbryio/reflector.go/types"
|
|
// blank import for db driver
|
|
_ "github.com/go-sql-driver/mysql"
|
|
log "github.com/sirupsen/logrus"
|
|
)
|
|
|
|
// DB interface communicates to a backend database with a simple set of methods that supports tracking blobs that are
|
|
// used together with a BlobStore. The DB tracks pointers and the BlobStore stores the data.
|
|
type DB interface {
|
|
Connect(string) error
|
|
HasBlob(string) (bool, error)
|
|
AddBlob(string, int, bool) error
|
|
AddSDBlob(string, int, types.SdBlob) error
|
|
}
|
|
|
|
// SQL is the container for the supporting MySQL database connection.
|
|
type SQL struct {
|
|
conn *sql.DB
|
|
}
|
|
|
|
func logQuery(query string, args ...interface{}) {
|
|
s, err := querytools.InterpolateParams(query, args...)
|
|
if err != nil {
|
|
log.Errorln(err)
|
|
} else {
|
|
log.Debugln(s)
|
|
}
|
|
}
|
|
|
|
// Connect will create a connection to the database
|
|
func (s *SQL) Connect(dsn string) error {
|
|
var err error
|
|
dsn += "?parseTime=1&collation=utf8mb4_unicode_ci"
|
|
s.conn, err = sql.Open("mysql", dsn)
|
|
if err != nil {
|
|
return errors.Err(err)
|
|
}
|
|
|
|
return errors.Err(s.conn.Ping())
|
|
}
|
|
|
|
// AddBlob adds a blobs information to the database.
|
|
func (s *SQL) AddBlob(hash string, length int, stored bool) error {
|
|
if s.conn == nil {
|
|
return errors.Err("not connected")
|
|
}
|
|
|
|
return withTx(s.conn, func(tx *sql.Tx) error {
|
|
return addBlob(tx, hash, length, stored)
|
|
})
|
|
}
|
|
|
|
func addBlob(tx *sql.Tx, hash string, length int, stored bool) error {
|
|
if length <= 0 {
|
|
return errors.Err("length must be positive")
|
|
}
|
|
|
|
query := "INSERT INTO blob_ (hash, stored, length) VALUES (?,?,?) ON DUPLICATE KEY UPDATE stored = (stored or VALUES(stored))"
|
|
args := []interface{}{hash, stored, length}
|
|
|
|
logQuery(query, args...)
|
|
|
|
stmt, err := tx.Prepare(query)
|
|
if err != nil {
|
|
return errors.Err(err)
|
|
}
|
|
|
|
_, err = stmt.Exec(args...)
|
|
if err != nil {
|
|
return errors.Err(err)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// HasBlob checks if the database contains the blob information.
|
|
func (s *SQL) HasBlob(hash string) (bool, error) {
|
|
if s.conn == nil {
|
|
return false, errors.Err("not connected")
|
|
}
|
|
|
|
query := "SELECT EXISTS(SELECT 1 FROM blob_ WHERE hash = ? AND stored = ?)"
|
|
args := []interface{}{hash, true}
|
|
|
|
logQuery(query, args...)
|
|
|
|
row := s.conn.QueryRow(query, args...)
|
|
|
|
exists := false
|
|
err := row.Scan(&exists)
|
|
|
|
return exists, errors.Err(err)
|
|
}
|
|
|
|
// HasBlobs checks if the database contains the set of blobs and returns a bool map.
|
|
func (s *SQL) HasBlobs(hashes []string) (map[string]bool, error) {
|
|
if s.conn == nil {
|
|
return nil, errors.Err("not connected")
|
|
}
|
|
|
|
var hash string
|
|
exists := make(map[string]bool)
|
|
maxBatchSize := 100
|
|
doneIndex := 0
|
|
|
|
for len(hashes) > doneIndex {
|
|
sliceEnd := doneIndex + maxBatchSize
|
|
if sliceEnd > len(hashes) {
|
|
sliceEnd = len(hashes)
|
|
}
|
|
log.Debugf("getting hashes[%d:%d] of %d", doneIndex, sliceEnd, len(hashes))
|
|
batch := hashes[doneIndex:sliceEnd]
|
|
|
|
query := "SELECT hash FROM blob_ WHERE stored = ? && hash IN (" + querytools.Qs(len(batch)) + ")"
|
|
args := make([]interface{}, len(batch)+1)
|
|
args[0] = true
|
|
for i := range batch {
|
|
args[i+1] = batch[i]
|
|
}
|
|
|
|
logQuery(query, args...)
|
|
|
|
rows, err := s.conn.Query(query, args...)
|
|
if err != nil {
|
|
closeRows(rows)
|
|
return exists, err
|
|
}
|
|
|
|
for rows.Next() {
|
|
err := rows.Scan(&hash)
|
|
if err != nil {
|
|
closeRows(rows)
|
|
return exists, err
|
|
}
|
|
exists[hash] = true
|
|
}
|
|
|
|
err = rows.Err()
|
|
if err != nil {
|
|
closeRows(rows)
|
|
return exists, err
|
|
}
|
|
|
|
closeRows(rows)
|
|
doneIndex += len(batch)
|
|
}
|
|
|
|
return exists, nil
|
|
}
|
|
|
|
// AddSDBlob takes the SD Hash number of blobs and the set of blobs. In a single db tx it inserts the sdblob information
|
|
// into a stream, and inserts the associated blobs' information in the database. If a blob fails the transaction is
|
|
// rolled back and error(s) are returned.
|
|
func (s *SQL) AddSDBlob(sdHash string, sdBlobLength int, sdBlob types.SdBlob) error {
|
|
if s.conn == nil {
|
|
return errors.Err("not connected")
|
|
}
|
|
|
|
return withTx(s.conn, func(tx *sql.Tx) error {
|
|
// insert sd blob
|
|
err := addBlob(tx, sdHash, sdBlobLength, true)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// insert stream
|
|
query := "INSERT IGNORE INTO stream (hash, sd_hash) VALUES (?,?)"
|
|
args := []interface{}{sdBlob.StreamHash, sdHash}
|
|
|
|
logQuery(query, args...)
|
|
|
|
stmt, err := tx.Prepare(query)
|
|
if err != nil {
|
|
return errors.Err(err)
|
|
}
|
|
|
|
_, err = stmt.Exec(args...)
|
|
if err != nil {
|
|
return errors.Err(err)
|
|
}
|
|
|
|
// insert content blobs and connect them to stream
|
|
for _, contentBlob := range sdBlob.Blobs {
|
|
if contentBlob.BlobHash == "" {
|
|
// null terminator blob
|
|
continue
|
|
}
|
|
|
|
err := addBlob(tx, contentBlob.BlobHash, contentBlob.Length, false)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
query := "INSERT IGNORE INTO stream_blob (stream_hash, blob_hash, num) VALUES (?,?,?)"
|
|
args := []interface{}{sdBlob.StreamHash, contentBlob.BlobHash, contentBlob.BlobNum}
|
|
|
|
logQuery(query, args...)
|
|
|
|
stmt, err := tx.Prepare(query)
|
|
if err != nil {
|
|
return errors.Err(err)
|
|
}
|
|
|
|
_, err = stmt.Exec(args...)
|
|
if err != nil {
|
|
return errors.Err(err)
|
|
}
|
|
}
|
|
return nil
|
|
})
|
|
}
|
|
|
|
// txFunc is a function that can be wrapped in a transaction
|
|
type txFunc func(tx *sql.Tx) error
|
|
|
|
// withTx wraps a function in an sql transaction. the transaction is committed if there's no error, or rolled back if there is one.
|
|
// if dbOrTx is an sql.DB, a new transaction is started
|
|
func withTx(dbOrTx interface{}, f txFunc) (err error) {
|
|
var tx *sql.Tx
|
|
|
|
switch t := dbOrTx.(type) {
|
|
case *sql.Tx:
|
|
tx = t
|
|
case *sql.DB:
|
|
tx, err = t.Begin()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer func() {
|
|
if p := recover(); p != nil {
|
|
if rollBackError := tx.Rollback(); rollBackError != nil {
|
|
log.Error("failed to rollback tx on panic - ", rollBackError)
|
|
}
|
|
panic(p)
|
|
} else if err != nil {
|
|
if rollBackError := tx.Rollback(); rollBackError != nil {
|
|
log.Error("failed to rollback tx on panic - ", rollBackError)
|
|
}
|
|
} else {
|
|
err = errors.Err(tx.Commit())
|
|
}
|
|
}()
|
|
default:
|
|
return errors.Err("db or tx required")
|
|
}
|
|
|
|
return f(tx)
|
|
}
|
|
|
|
func closeRows(rows *sql.Rows) {
|
|
if err := rows.Close(); err != nil {
|
|
log.Error("error closing rows: ", err)
|
|
}
|
|
}
|
|
|
|
/*// func to generate schema. SQL below that.
|
|
func schema() {
|
|
_ = `
|
|
CREATE TABLE blob_ (
|
|
hash char(96) NOT NULL,
|
|
stored TINYINT(1) NOT NULL DEFAULT 0,
|
|
length bigint(20) unsigned DEFAULT NULL,
|
|
last_announced_at datetime DEFAULT NULL,
|
|
PRIMARY KEY (hash),
|
|
KEY last_announced_at_idx (last_announced_at)
|
|
) CHARACTER SET utf8mb4 COLLATE utf8mb4_unicode_ci;
|
|
|
|
CREATE TABLE stream (
|
|
hash char(96) NOT NULL,
|
|
sd_hash char(96) NOT NULL,
|
|
PRIMARY KEY (hash),
|
|
KEY sd_hash_idx (sd_hash),
|
|
FOREIGN KEY (sd_hash) REFERENCES blob_ (hash) ON DELETE RESTRICT ON UPDATE CASCADE
|
|
) CHARACTER SET utf8mb4 COLLATE utf8mb4_unicode_ci;
|
|
|
|
CREATE TABLE stream_blob (
|
|
stream_hash char(96) NOT NULL,
|
|
blob_hash char(96) NOT NULL,
|
|
num int NOT NULL,
|
|
PRIMARY KEY (stream_hash, blob_hash),
|
|
FOREIGN KEY (stream_hash) REFERENCES stream (hash) ON DELETE CASCADE ON UPDATE CASCADE,
|
|
FOREIGN KEY (blob_hash) REFERENCES blob_ (hash) ON DELETE CASCADE ON UPDATE CASCADE
|
|
) CHARACTER SET utf8mb4 COLLATE utf8mb4_unicode_ci;
|
|
|
|
`
|
|
}*/
|
|
|
|
/* SQL script to create schema
|
|
CREATE TABLE `reflector`.`blob_`
|
|
(
|
|
`hash` char(96) NOT NULL,
|
|
`stored` TINYINT(1) NOT NULL DEFAULT 0,
|
|
`length` bigint(20) unsigned DEFAULT NULL,
|
|
`last_announced_at` datetime DEFAULT NULL,
|
|
PRIMARY KEY (`hash`),
|
|
KEY `last_announced_at_idx` (`last_announced_at`)
|
|
) CHARACTER SET utf8mb4 COLLATE utf8mb4_unicode_ci;
|
|
|
|
CREATE TABLE `reflector`.`stream`
|
|
(
|
|
`hash` char(96) NOT NULL,
|
|
`sd_hash` char(96) NOT NULL,
|
|
PRIMARY KEY (hash),
|
|
KEY `sd_hash_idx` (`sd_hash`),
|
|
FOREIGN KEY (`sd_hash`) REFERENCES `blob_` (`hash`) ON DELETE RESTRICT ON UPDATE CASCADE
|
|
) CHARACTER SET utf8mb4 COLLATE utf8mb4_unicode_ci;
|
|
|
|
CREATE TABLE `reflector`.`stream_blob`
|
|
(
|
|
`stream_hash` char(96) NOT NULL,
|
|
`blob_hash` char(96) NOT NULL,
|
|
`num` int NOT NULL,
|
|
PRIMARY KEY (`stream_hash`, `blob_hash`),
|
|
FOREIGN KEY (`stream_hash`) REFERENCES `stream` (`hash`) ON DELETE CASCADE ON UPDATE CASCADE,
|
|
FOREIGN KEY (`blob_hash`) REFERENCES `blob_` (`hash`) ON DELETE CASCADE ON UPDATE CASCADE
|
|
) CHARACTER SET utf8mb4 COLLATE utf8mb4_unicode_ci;
|
|
*/
|