2015-08-20 17:27:15 +02:00
|
|
|
import logging
|
|
|
|
import os
|
|
|
|
import time
|
2015-09-04 22:22:02 +02:00
|
|
|
import sqlite3
|
2016-09-27 19:52:44 +02:00
|
|
|
|
2015-09-08 21:42:56 +02:00
|
|
|
from twisted.internet import threads, defer
|
2015-08-20 17:27:15 +02:00
|
|
|
from twisted.python.failure import Failure
|
2015-09-04 22:22:02 +02:00
|
|
|
from twisted.enterprise import adbapi
|
2015-08-20 17:27:15 +02:00
|
|
|
from lbrynet.core.HashBlob import BlobFile, TempBlob, BlobFileCreator, TempBlobCreator
|
|
|
|
from lbrynet.core.server.DHTHashAnnouncer import DHTHashSupplier
|
|
|
|
from lbrynet.core.utils import is_valid_blobhash
|
|
|
|
from lbrynet.core.cryptoutils import get_lbry_hash_obj
|
2015-09-04 22:22:02 +02:00
|
|
|
from lbrynet.core.Error import NoSuchBlobError
|
|
|
|
from lbrynet.core.sqlite_helpers import rerun_if_locked
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2015-09-08 21:42:56 +02:00
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
class BlobManager(DHTHashSupplier):
|
|
|
|
"""This class is subclassed by classes which keep track of which blobs are available
|
|
|
|
and which give access to new/existing blobs"""
|
|
|
|
def __init__(self, hash_announcer):
|
|
|
|
DHTHashSupplier.__init__(self, hash_announcer)
|
|
|
|
|
|
|
|
def setup(self):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def get_blob(self, blob_hash, upload_allowed, length):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def get_blob_creator(self):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def _make_new_blob(self, blob_hash, upload_allowed, length):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def blob_completed(self, blob, next_announce_time=None):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def completed_blobs(self, blobs_to_check):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def hashes_to_announce(self):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def creator_finished(self, blob_creator):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def delete_blob(self, blob_hash):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def get_blob_length(self, blob_hash):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def check_consistency(self):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def blob_requested(self, blob_hash):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def blob_downloaded(self, blob_hash):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def blob_searched_on(self, blob_hash):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def blob_paid_for(self, blob_hash, amount):
|
|
|
|
pass
|
|
|
|
|
2016-08-26 06:32:33 +02:00
|
|
|
def get_all_verified_blobs(self):
|
|
|
|
pass
|
|
|
|
|
2016-09-27 19:52:44 +02:00
|
|
|
def add_blob_to_download_history(self, blob_hash, host, rate):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def add_blob_to_upload_history(self, blob_hash, host, rate):
|
|
|
|
pass
|
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
class DiskBlobManager(BlobManager):
|
|
|
|
"""This class stores blobs on the hard disk"""
|
|
|
|
def __init__(self, hash_announcer, blob_dir, db_dir):
|
|
|
|
BlobManager.__init__(self, hash_announcer)
|
|
|
|
self.blob_dir = blob_dir
|
2015-09-04 22:22:02 +02:00
|
|
|
self.db_file = os.path.join(db_dir, "blobs.db")
|
|
|
|
self.db_conn = None
|
2015-08-20 17:27:15 +02:00
|
|
|
self.blob_type = BlobFile
|
|
|
|
self.blob_creator_type = BlobFileCreator
|
|
|
|
self.blobs = {}
|
2016-08-26 06:32:33 +02:00
|
|
|
self.blob_hashes_to_delete = {} # {blob_hash: being_deleted (True/False)}
|
2015-08-20 17:27:15 +02:00
|
|
|
self._next_manage_call = None
|
|
|
|
|
|
|
|
def setup(self):
|
2015-09-15 06:29:18 +02:00
|
|
|
log.info("Setting up the DiskBlobManager. blob_dir: %s, db_file: %s", str(self.blob_dir),
|
|
|
|
str(self.db_file))
|
2015-09-04 22:22:02 +02:00
|
|
|
d = self._open_db()
|
2015-08-20 17:27:15 +02:00
|
|
|
d.addCallback(lambda _: self._manage())
|
|
|
|
return d
|
|
|
|
|
|
|
|
def stop(self):
|
2015-09-15 06:29:18 +02:00
|
|
|
log.info("Stopping the DiskBlobManager")
|
2015-08-20 17:27:15 +02:00
|
|
|
if self._next_manage_call is not None and self._next_manage_call.active():
|
|
|
|
self._next_manage_call.cancel()
|
|
|
|
self._next_manage_call = None
|
2015-09-04 22:22:02 +02:00
|
|
|
self.db_conn = None
|
2015-08-20 17:27:15 +02:00
|
|
|
return defer.succeed(True)
|
|
|
|
|
|
|
|
def get_blob(self, blob_hash, upload_allowed, length=None):
|
2016-11-30 21:20:45 +01:00
|
|
|
"""Return a blob identified by blob_hash, which may be a new blob or a
|
|
|
|
blob that is already on the hard disk
|
|
|
|
"""
|
|
|
|
# TODO: if blob.upload_allowed and upload_allowed is False,
|
|
|
|
# change upload_allowed in blob and on disk
|
2015-08-20 17:27:15 +02:00
|
|
|
if blob_hash in self.blobs:
|
|
|
|
return defer.succeed(self.blobs[blob_hash])
|
|
|
|
return self._make_new_blob(blob_hash, upload_allowed, length)
|
|
|
|
|
|
|
|
def get_blob_creator(self):
|
|
|
|
return self.blob_creator_type(self, self.blob_dir)
|
|
|
|
|
|
|
|
def _make_new_blob(self, blob_hash, upload_allowed, length=None):
|
2016-12-11 00:02:13 +01:00
|
|
|
log.debug('Making a new blob for %s', blob_hash)
|
2015-08-20 17:27:15 +02:00
|
|
|
blob = self.blob_type(self.blob_dir, blob_hash, upload_allowed, length)
|
|
|
|
self.blobs[blob_hash] = blob
|
2015-09-04 22:22:02 +02:00
|
|
|
d = self._completed_blobs([blob_hash])
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
def check_completed(completed_blobs):
|
|
|
|
|
|
|
|
def set_length(length):
|
|
|
|
blob.length = length
|
|
|
|
|
|
|
|
if len(completed_blobs) == 1 and completed_blobs[0] == blob_hash:
|
|
|
|
blob.verified = True
|
2015-09-04 22:22:02 +02:00
|
|
|
inner_d = self._get_blob_length(blob_hash)
|
2015-08-20 17:27:15 +02:00
|
|
|
inner_d.addCallback(set_length)
|
|
|
|
inner_d.addCallback(lambda _: blob)
|
|
|
|
else:
|
|
|
|
inner_d = defer.succeed(blob)
|
|
|
|
return inner_d
|
|
|
|
|
|
|
|
d.addCallback(check_completed)
|
|
|
|
return d
|
|
|
|
|
|
|
|
def blob_completed(self, blob, next_announce_time=None):
|
|
|
|
if next_announce_time is None:
|
2016-12-11 00:02:13 +01:00
|
|
|
next_announce_time = time.time() + self.hash_reannounce_time
|
|
|
|
d = self._add_completed_blob(blob.blob_hash, blob.length,
|
|
|
|
time.time(), next_announce_time)
|
|
|
|
d.addCallback(lambda _: self.hash_announcer.immediate_announce([blob.blob_hash]))
|
|
|
|
return d
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
def completed_blobs(self, blobs_to_check):
|
2015-09-04 22:22:02 +02:00
|
|
|
return self._completed_blobs(blobs_to_check)
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
def hashes_to_announce(self):
|
|
|
|
next_announce_time = time.time() + self.hash_reannounce_time
|
2015-09-04 22:22:02 +02:00
|
|
|
return self._get_blobs_to_announce(next_announce_time)
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
def creator_finished(self, blob_creator):
|
2015-09-08 21:42:56 +02:00
|
|
|
log.debug("blob_creator.blob_hash: %s", blob_creator.blob_hash)
|
2015-08-20 17:27:15 +02:00
|
|
|
assert blob_creator.blob_hash is not None
|
|
|
|
assert blob_creator.blob_hash not in self.blobs
|
|
|
|
assert blob_creator.length is not None
|
|
|
|
new_blob = self.blob_type(self.blob_dir, blob_creator.blob_hash, True, blob_creator.length)
|
|
|
|
new_blob.verified = True
|
|
|
|
self.blobs[blob_creator.blob_hash] = new_blob
|
|
|
|
if self.hash_announcer is not None:
|
|
|
|
self.hash_announcer.immediate_announce([blob_creator.blob_hash])
|
|
|
|
next_announce_time = time.time() + self.hash_reannounce_time
|
|
|
|
d = self.blob_completed(new_blob, next_announce_time)
|
|
|
|
else:
|
|
|
|
d = self.blob_completed(new_blob)
|
|
|
|
return d
|
|
|
|
|
|
|
|
def delete_blobs(self, blob_hashes):
|
|
|
|
for blob_hash in blob_hashes:
|
|
|
|
if not blob_hash in self.blob_hashes_to_delete:
|
|
|
|
self.blob_hashes_to_delete[blob_hash] = False
|
|
|
|
|
|
|
|
def update_all_last_verified_dates(self, timestamp):
|
2015-09-04 22:22:02 +02:00
|
|
|
return self._update_all_last_verified_dates(timestamp)
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
def immediate_announce_all_blobs(self):
|
2015-09-04 22:22:02 +02:00
|
|
|
d = self._get_all_verified_blob_hashes()
|
2015-08-20 17:27:15 +02:00
|
|
|
d.addCallback(self.hash_announcer.immediate_announce)
|
|
|
|
return d
|
|
|
|
|
|
|
|
def get_blob_length(self, blob_hash):
|
2015-09-04 22:22:02 +02:00
|
|
|
return self._get_blob_length(blob_hash)
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
def check_consistency(self):
|
2015-09-04 22:22:02 +02:00
|
|
|
return self._check_consistency()
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2016-08-26 06:32:33 +02:00
|
|
|
def get_all_verified_blobs(self):
|
|
|
|
d = self._get_all_verified_blob_hashes()
|
|
|
|
d.addCallback(self.completed_blobs)
|
|
|
|
return d
|
|
|
|
|
2016-09-27 19:52:44 +02:00
|
|
|
def add_blob_to_download_history(self, blob_hash, host, rate):
|
|
|
|
d = self._add_blob_to_download_history(blob_hash, host, rate)
|
|
|
|
return d
|
|
|
|
|
|
|
|
def add_blob_to_upload_history(self, blob_hash, host, rate):
|
|
|
|
d = self._add_blob_to_upload_history(blob_hash, host, rate)
|
|
|
|
return d
|
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
def _manage(self):
|
|
|
|
from twisted.internet import reactor
|
|
|
|
|
|
|
|
d = self._delete_blobs_marked_for_deletion()
|
|
|
|
|
|
|
|
def set_next_manage_call():
|
|
|
|
self._next_manage_call = reactor.callLater(1, self._manage)
|
|
|
|
|
|
|
|
d.addCallback(lambda _: set_next_manage_call())
|
|
|
|
|
|
|
|
def _delete_blobs_marked_for_deletion(self):
|
|
|
|
|
|
|
|
def remove_from_list(b_h):
|
|
|
|
del self.blob_hashes_to_delete[b_h]
|
|
|
|
return b_h
|
|
|
|
|
|
|
|
def set_not_deleting(err, b_h):
|
2015-09-08 21:42:56 +02:00
|
|
|
log.warning("Failed to delete blob %s. Reason: %s", str(b_h), err.getErrorMessage())
|
2015-08-20 17:27:15 +02:00
|
|
|
self.blob_hashes_to_delete[b_h] = False
|
|
|
|
return err
|
|
|
|
|
|
|
|
def delete_from_db(result):
|
|
|
|
b_hs = [r[1] for r in result if r[0] is True]
|
|
|
|
if b_hs:
|
2015-09-04 22:22:02 +02:00
|
|
|
d = self._delete_blobs_from_db(b_hs)
|
2015-08-20 17:27:15 +02:00
|
|
|
else:
|
|
|
|
d = defer.succeed(True)
|
|
|
|
|
|
|
|
def log_error(err):
|
2016-11-30 21:20:45 +01:00
|
|
|
log.warning(
|
|
|
|
"Failed to delete completed blobs from the db: %s", err.getErrorMessage())
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
d.addErrback(log_error)
|
|
|
|
return d
|
|
|
|
|
|
|
|
def delete(blob, b_h):
|
|
|
|
d = blob.delete()
|
|
|
|
d.addCallbacks(lambda _: remove_from_list(b_h), set_not_deleting, errbackArgs=(b_h,))
|
|
|
|
return d
|
|
|
|
|
|
|
|
ds = []
|
|
|
|
for blob_hash, being_deleted in self.blob_hashes_to_delete.items():
|
|
|
|
if being_deleted is False:
|
|
|
|
self.blob_hashes_to_delete[blob_hash] = True
|
|
|
|
d = self.get_blob(blob_hash, True)
|
2016-11-30 21:20:45 +01:00
|
|
|
d.addCallbacks(
|
|
|
|
delete, set_not_deleting,
|
|
|
|
callbackArgs=(blob_hash,), errbackArgs=(blob_hash,))
|
2015-08-20 17:27:15 +02:00
|
|
|
ds.append(d)
|
|
|
|
dl = defer.DeferredList(ds, consumeErrors=True)
|
|
|
|
dl.addCallback(delete_from_db)
|
|
|
|
return defer.DeferredList(ds)
|
|
|
|
|
|
|
|
######### database calls #########
|
|
|
|
|
|
|
|
def _open_db(self):
|
2015-09-04 22:22:02 +02:00
|
|
|
# check_same_thread=False is solely to quiet a spurious error that appears to be due
|
|
|
|
# to a bug in twisted, where the connection is closed by a different thread than the
|
|
|
|
# one that opened it. The individual connections in the pool are not used in multiple
|
|
|
|
# threads.
|
|
|
|
self.db_conn = adbapi.ConnectionPool('sqlite3', self.db_file, check_same_thread=False)
|
2016-09-27 19:52:44 +02:00
|
|
|
|
|
|
|
def create_tables(transaction):
|
|
|
|
transaction.execute("create table if not exists blobs (" +
|
|
|
|
" blob_hash text primary key, " +
|
|
|
|
" blob_length integer, " +
|
|
|
|
" last_verified_time real, " +
|
|
|
|
" next_announce_time real)")
|
|
|
|
|
|
|
|
transaction.execute("create table if not exists download (" +
|
|
|
|
" id integer primary key autoincrement, " +
|
|
|
|
" blob text, " +
|
|
|
|
" host text, " +
|
|
|
|
" rate float, " +
|
|
|
|
" ts integer)")
|
|
|
|
|
|
|
|
transaction.execute("create table if not exists upload (" +
|
|
|
|
" id integer primary key autoincrement, " +
|
|
|
|
" blob text, " +
|
|
|
|
" host text, " +
|
|
|
|
" rate float, " +
|
|
|
|
" ts integer)")
|
|
|
|
|
|
|
|
return self.db_conn.runInteraction(create_tables)
|
2015-09-04 22:22:02 +02:00
|
|
|
|
|
|
|
@rerun_if_locked
|
2015-08-20 17:27:15 +02:00
|
|
|
def _add_completed_blob(self, blob_hash, length, timestamp, next_announce_time=None):
|
2015-09-08 21:42:56 +02:00
|
|
|
log.debug("Adding a completed blob. blob_hash=%s, length=%s", blob_hash, str(length))
|
2015-08-20 17:27:15 +02:00
|
|
|
if next_announce_time is None:
|
|
|
|
next_announce_time = timestamp
|
2015-09-04 22:22:02 +02:00
|
|
|
d = self.db_conn.runQuery("insert into blobs values (?, ?, ?, ?)",
|
|
|
|
(blob_hash, length, timestamp, next_announce_time))
|
|
|
|
d.addErrback(lambda err: err.trap(sqlite3.IntegrityError))
|
|
|
|
return d
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2015-09-04 22:22:02 +02:00
|
|
|
@rerun_if_locked
|
2015-08-20 17:27:15 +02:00
|
|
|
def _completed_blobs(self, blobs_to_check):
|
2015-09-04 22:22:02 +02:00
|
|
|
blobs_to_check = filter(is_valid_blobhash, blobs_to_check)
|
|
|
|
|
|
|
|
def get_blobs_in_db(db_transaction):
|
|
|
|
blobs_in_db = [] # [(blob_hash, last_verified_time)]
|
|
|
|
for b in blobs_to_check:
|
2016-11-30 21:20:45 +01:00
|
|
|
result = db_transaction.execute(
|
|
|
|
"select last_verified_time from blobs where blob_hash = ?",
|
|
|
|
(b,))
|
2015-09-04 22:22:02 +02:00
|
|
|
row = result.fetchone()
|
|
|
|
if row is not None:
|
|
|
|
blobs_in_db.append((b, row[0]))
|
|
|
|
return blobs_in_db
|
|
|
|
|
|
|
|
def get_valid_blobs(blobs_in_db):
|
|
|
|
|
|
|
|
def check_blob_verified_date(b, verified_time):
|
2015-08-20 17:27:15 +02:00
|
|
|
file_path = os.path.join(self.blob_dir, b)
|
|
|
|
if os.path.isfile(file_path):
|
|
|
|
if verified_time > os.path.getctime(file_path):
|
2015-09-04 22:22:02 +02:00
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
|
|
def return_valid_blobs(results):
|
|
|
|
valid_blobs = []
|
|
|
|
for (b, verified_date), (success, result) in zip(blobs_in_db, results):
|
|
|
|
if success is True and result is True:
|
|
|
|
valid_blobs.append(b)
|
|
|
|
return valid_blobs
|
|
|
|
|
|
|
|
ds = []
|
|
|
|
for b, verified_date in blobs_in_db:
|
|
|
|
ds.append(threads.deferToThread(check_blob_verified_date, b, verified_date))
|
|
|
|
dl = defer.DeferredList(ds)
|
|
|
|
dl.addCallback(return_valid_blobs)
|
|
|
|
return dl
|
|
|
|
|
|
|
|
d = self.db_conn.runInteraction(get_blobs_in_db)
|
|
|
|
d.addCallback(get_valid_blobs)
|
|
|
|
return d
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2015-09-04 22:22:02 +02:00
|
|
|
@rerun_if_locked
|
2015-08-20 17:27:15 +02:00
|
|
|
def _get_blob_length(self, blob):
|
2015-09-04 22:22:02 +02:00
|
|
|
d = self.db_conn.runQuery("select blob_length from blobs where blob_hash = ?", (blob,))
|
2015-09-11 03:43:46 +02:00
|
|
|
d.addCallback(lambda r: r[0][0] if len(r) else Failure(NoSuchBlobError(blob)))
|
2015-09-04 22:22:02 +02:00
|
|
|
return d
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2015-09-04 22:22:02 +02:00
|
|
|
|
|
|
|
@rerun_if_locked
|
2015-08-20 17:27:15 +02:00
|
|
|
def _update_blob_verified_timestamp(self, blob, timestamp):
|
2015-09-04 22:22:02 +02:00
|
|
|
return self.db_conn.runQuery("update blobs set last_verified_time = ? where blob_hash = ?",
|
|
|
|
(blob, timestamp))
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2015-09-04 22:22:02 +02:00
|
|
|
@rerun_if_locked
|
2015-08-20 17:27:15 +02:00
|
|
|
def _get_blobs_to_announce(self, next_announce_time):
|
|
|
|
|
2015-09-04 22:22:02 +02:00
|
|
|
def get_and_update(transaction):
|
|
|
|
timestamp = time.time()
|
|
|
|
r = transaction.execute("select blob_hash from blobs " +
|
|
|
|
"where next_announce_time < ? and blob_hash is not null",
|
|
|
|
(timestamp,))
|
|
|
|
blobs = [b for b, in r.fetchall()]
|
2016-11-30 21:20:45 +01:00
|
|
|
transaction.execute(
|
|
|
|
"update blobs set next_announce_time = ? where next_announce_time < ?",
|
|
|
|
(next_announce_time, timestamp))
|
2015-09-04 22:22:02 +02:00
|
|
|
return blobs
|
|
|
|
|
|
|
|
return self.db_conn.runInteraction(get_and_update)
|
|
|
|
|
|
|
|
@rerun_if_locked
|
2015-08-20 17:27:15 +02:00
|
|
|
def _update_all_last_verified_dates(self, timestamp):
|
2015-09-04 22:22:02 +02:00
|
|
|
return self.db_conn.runQuery("update blobs set last_verified_date = ?", (timestamp,))
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2015-09-04 22:22:02 +02:00
|
|
|
@rerun_if_locked
|
2015-08-20 17:27:15 +02:00
|
|
|
def _delete_blobs_from_db(self, blob_hashes):
|
|
|
|
|
2015-09-04 22:22:02 +02:00
|
|
|
def delete_blobs(transaction):
|
|
|
|
for b in blob_hashes:
|
|
|
|
transaction.execute("delete from blobs where blob_hash = ?", (b,))
|
|
|
|
|
|
|
|
return self.db_conn.runInteraction(delete_blobs)
|
|
|
|
|
|
|
|
@rerun_if_locked
|
2015-08-20 17:27:15 +02:00
|
|
|
def _check_consistency(self):
|
2015-09-04 22:22:02 +02:00
|
|
|
|
|
|
|
ALREADY_VERIFIED = 1
|
|
|
|
NEWLY_VERIFIED = 2
|
|
|
|
INVALID = 3
|
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
current_time = time.time()
|
2015-09-04 22:22:02 +02:00
|
|
|
d = self.db_conn.runQuery("select blob_hash, blob_length, last_verified_time from blobs")
|
|
|
|
|
|
|
|
def check_blob(blob_hash, blob_length, verified_time):
|
2015-08-20 17:27:15 +02:00
|
|
|
file_path = os.path.join(self.blob_dir, blob_hash)
|
|
|
|
if os.path.isfile(file_path):
|
2015-09-04 22:22:02 +02:00
|
|
|
if verified_time >= os.path.getctime(file_path):
|
|
|
|
return ALREADY_VERIFIED
|
|
|
|
else:
|
2015-08-20 17:27:15 +02:00
|
|
|
h = get_lbry_hash_obj()
|
|
|
|
len_so_far = 0
|
|
|
|
f = open(file_path)
|
|
|
|
while True:
|
|
|
|
data = f.read(2**12)
|
|
|
|
if not data:
|
|
|
|
break
|
|
|
|
h.update(data)
|
|
|
|
len_so_far += len(data)
|
2015-09-04 22:22:02 +02:00
|
|
|
if len_so_far == blob_length and h.hexdigest() == blob_hash:
|
|
|
|
return NEWLY_VERIFIED
|
|
|
|
return INVALID
|
|
|
|
|
|
|
|
def do_check(blobs):
|
|
|
|
already_verified = []
|
|
|
|
newly_verified = []
|
|
|
|
invalid = []
|
|
|
|
for blob_hash, blob_length, verified_time in blobs:
|
|
|
|
status = check_blob(blob_hash, blob_length, verified_time)
|
|
|
|
if status == ALREADY_VERIFIED:
|
|
|
|
already_verified.append(blob_hash)
|
|
|
|
elif status == NEWLY_VERIFIED:
|
|
|
|
newly_verified.append(blob_hash)
|
|
|
|
else:
|
|
|
|
invalid.append(blob_hash)
|
|
|
|
return already_verified, newly_verified, invalid
|
|
|
|
|
|
|
|
def update_newly_verified(transaction, blobs):
|
|
|
|
for b in blobs:
|
|
|
|
transaction.execute("update blobs set last_verified_time = ? where blob_hash = ?",
|
|
|
|
(current_time, b))
|
|
|
|
|
|
|
|
def check_blobs(blobs):
|
|
|
|
|
|
|
|
@rerun_if_locked
|
|
|
|
def update_and_return(status_lists):
|
|
|
|
|
|
|
|
already_verified, newly_verified, invalid = status_lists
|
|
|
|
|
|
|
|
d = self.db_conn.runInteraction(update_newly_verified, newly_verified)
|
|
|
|
d.addCallback(lambda _: status_lists)
|
|
|
|
return d
|
|
|
|
|
|
|
|
d = threads.deferToThread(do_check, blobs)
|
|
|
|
|
|
|
|
d.addCallback(update_and_return)
|
|
|
|
return d
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2015-09-04 22:22:02 +02:00
|
|
|
d.addCallback(check_blobs)
|
|
|
|
return d
|
|
|
|
|
|
|
|
@rerun_if_locked
|
2015-08-20 17:27:15 +02:00
|
|
|
def _get_all_verified_blob_hashes(self):
|
2015-09-04 22:22:02 +02:00
|
|
|
d = self.db_conn.runQuery("select blob_hash, last_verified_time from blobs")
|
|
|
|
|
|
|
|
def get_verified_blobs(blobs):
|
|
|
|
verified_blobs = []
|
|
|
|
for blob_hash, verified_time in blobs:
|
|
|
|
file_path = os.path.join(self.blob_dir, blob_hash)
|
|
|
|
if os.path.isfile(file_path):
|
|
|
|
if verified_time > os.path.getctime(file_path):
|
|
|
|
verified_blobs.append(blob_hash)
|
|
|
|
return verified_blobs
|
|
|
|
|
|
|
|
d.addCallback(lambda blobs: threads.deferToThread(get_verified_blobs, blobs))
|
|
|
|
return d
|
2015-08-20 17:27:15 +02:00
|
|
|
|
2016-09-27 19:52:44 +02:00
|
|
|
@rerun_if_locked
|
|
|
|
def _add_blob_to_download_history(self, blob_hash, host, rate):
|
|
|
|
ts = int(time.time())
|
2016-11-30 21:20:45 +01:00
|
|
|
d = self.db_conn.runQuery(
|
|
|
|
"insert into download values (null, ?, ?, ?, ?) ",
|
|
|
|
(blob_hash, str(host), float(rate), ts))
|
2016-09-27 19:52:44 +02:00
|
|
|
return d
|
|
|
|
|
|
|
|
@rerun_if_locked
|
|
|
|
def _add_blob_to_upload_history(self, blob_hash, host, rate):
|
|
|
|
ts = int(time.time())
|
2016-11-30 21:20:45 +01:00
|
|
|
d = self.db_conn.runQuery(
|
|
|
|
"insert into upload values (null, ?, ?, ?, ?) ",
|
|
|
|
(blob_hash, str(host), float(rate), ts))
|
2016-09-27 19:52:44 +02:00
|
|
|
return d
|
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
class TempBlobManager(BlobManager):
|
|
|
|
"""This class stores blobs in memory"""
|
|
|
|
def __init__(self, hash_announcer):
|
|
|
|
BlobManager.__init__(self, hash_announcer)
|
|
|
|
self.blob_type = TempBlob
|
|
|
|
self.blob_creator_type = TempBlobCreator
|
|
|
|
self.blobs = {}
|
|
|
|
self.blob_next_announces = {}
|
|
|
|
self.blob_hashes_to_delete = {} # {blob_hash: being_deleted (True/False)}
|
|
|
|
self._next_manage_call = None
|
|
|
|
|
|
|
|
def setup(self):
|
|
|
|
self._manage()
|
|
|
|
return defer.succeed(True)
|
|
|
|
|
|
|
|
def stop(self):
|
|
|
|
if self._next_manage_call is not None and self._next_manage_call.active():
|
|
|
|
self._next_manage_call.cancel()
|
|
|
|
self._next_manage_call = None
|
|
|
|
|
|
|
|
def get_blob(self, blob_hash, upload_allowed, length=None):
|
|
|
|
if blob_hash in self.blobs:
|
|
|
|
return defer.succeed(self.blobs[blob_hash])
|
|
|
|
return self._make_new_blob(blob_hash, upload_allowed, length)
|
|
|
|
|
|
|
|
def get_blob_creator(self):
|
|
|
|
return self.blob_creator_type(self)
|
|
|
|
|
|
|
|
def _make_new_blob(self, blob_hash, upload_allowed, length=None):
|
|
|
|
blob = self.blob_type(blob_hash, upload_allowed, length)
|
|
|
|
self.blobs[blob_hash] = blob
|
|
|
|
return defer.succeed(blob)
|
|
|
|
|
|
|
|
def blob_completed(self, blob, next_announce_time=None):
|
|
|
|
if next_announce_time is None:
|
|
|
|
next_announce_time = time.time()
|
|
|
|
self.blob_next_announces[blob.blob_hash] = next_announce_time
|
|
|
|
return defer.succeed(True)
|
|
|
|
|
|
|
|
def completed_blobs(self, blobs_to_check):
|
2016-11-30 21:20:45 +01:00
|
|
|
blobs = [
|
|
|
|
b.blob_hash for b in self.blobs.itervalues()
|
|
|
|
if b.blob_hash in blobs_to_check and b.is_validated()
|
|
|
|
]
|
2015-08-20 17:27:15 +02:00
|
|
|
return defer.succeed(blobs)
|
|
|
|
|
2016-08-26 06:32:33 +02:00
|
|
|
def get_all_verified_blobs(self):
|
|
|
|
d = self.completed_blobs(self.blobs)
|
|
|
|
return d
|
|
|
|
|
2015-08-20 17:27:15 +02:00
|
|
|
def hashes_to_announce(self):
|
|
|
|
now = time.time()
|
2016-11-30 21:20:45 +01:00
|
|
|
blobs = [
|
|
|
|
blob_hash for blob_hash, announce_time in self.blob_next_announces.iteritems()
|
|
|
|
if announce_time < now
|
|
|
|
]
|
2015-08-20 17:27:15 +02:00
|
|
|
next_announce_time = now + self.hash_reannounce_time
|
|
|
|
for b in blobs:
|
|
|
|
self.blob_next_announces[b] = next_announce_time
|
|
|
|
return defer.succeed(blobs)
|
|
|
|
|
|
|
|
def creator_finished(self, blob_creator):
|
|
|
|
assert blob_creator.blob_hash is not None
|
|
|
|
assert blob_creator.blob_hash not in self.blobs
|
|
|
|
assert blob_creator.length is not None
|
|
|
|
new_blob = self.blob_type(blob_creator.blob_hash, True, blob_creator.length)
|
|
|
|
new_blob.verified = True
|
|
|
|
new_blob.data_buffer = blob_creator.data_buffer
|
|
|
|
new_blob.length = blob_creator.length
|
|
|
|
self.blobs[blob_creator.blob_hash] = new_blob
|
|
|
|
if self.hash_announcer is not None:
|
|
|
|
self.hash_announcer.immediate_announce([blob_creator.blob_hash])
|
|
|
|
next_announce_time = time.time() + self.hash_reannounce_time
|
|
|
|
d = self.blob_completed(new_blob, next_announce_time)
|
|
|
|
else:
|
|
|
|
d = self.blob_completed(new_blob)
|
|
|
|
d.addCallback(lambda _: new_blob)
|
|
|
|
return d
|
|
|
|
|
|
|
|
def delete_blobs(self, blob_hashes):
|
|
|
|
for blob_hash in blob_hashes:
|
|
|
|
if not blob_hash in self.blob_hashes_to_delete:
|
|
|
|
self.blob_hashes_to_delete[blob_hash] = False
|
|
|
|
|
|
|
|
def get_blob_length(self, blob_hash):
|
|
|
|
if blob_hash in self.blobs:
|
|
|
|
if self.blobs[blob_hash].length is not None:
|
|
|
|
return defer.succeed(self.blobs[blob_hash].length)
|
2015-09-04 22:22:02 +02:00
|
|
|
return defer.fail(NoSuchBlobError(blob_hash))
|
2015-08-20 17:27:15 +02:00
|
|
|
|
|
|
|
def immediate_announce_all_blobs(self):
|
|
|
|
return self.hash_announcer.immediate_announce(self.blobs.iterkeys())
|
|
|
|
|
|
|
|
def _manage(self):
|
|
|
|
from twisted.internet import reactor
|
|
|
|
|
|
|
|
d = self._delete_blobs_marked_for_deletion()
|
|
|
|
|
|
|
|
def set_next_manage_call():
|
2015-09-08 21:42:56 +02:00
|
|
|
log.info("Setting the next manage call in %s", str(self))
|
2015-08-20 17:27:15 +02:00
|
|
|
self._next_manage_call = reactor.callLater(1, self._manage)
|
|
|
|
|
|
|
|
d.addCallback(lambda _: set_next_manage_call())
|
|
|
|
|
|
|
|
def _delete_blobs_marked_for_deletion(self):
|
|
|
|
def remove_from_list(b_h):
|
|
|
|
del self.blob_hashes_to_delete[b_h]
|
2015-09-08 21:42:56 +02:00
|
|
|
log.info("Deleted blob %s", blob_hash)
|
2015-08-20 17:27:15 +02:00
|
|
|
return b_h
|
|
|
|
|
|
|
|
def set_not_deleting(err, b_h):
|
2015-09-08 21:42:56 +02:00
|
|
|
log.warning("Failed to delete blob %s. Reason: %s", str(b_h), err.getErrorMessage())
|
2015-08-20 17:27:15 +02:00
|
|
|
self.blob_hashes_to_delete[b_h] = False
|
|
|
|
return b_h
|
|
|
|
|
|
|
|
ds = []
|
|
|
|
for blob_hash, being_deleted in self.blob_hashes_to_delete.items():
|
|
|
|
if being_deleted is False:
|
|
|
|
if blob_hash in self.blobs:
|
|
|
|
self.blob_hashes_to_delete[blob_hash] = True
|
2015-09-08 21:42:56 +02:00
|
|
|
log.info("Found a blob marked for deletion: %s", blob_hash)
|
2015-08-20 17:27:15 +02:00
|
|
|
blob = self.blobs[blob_hash]
|
|
|
|
d = blob.delete()
|
|
|
|
|
|
|
|
d.addCallbacks(lambda _: remove_from_list(blob_hash), set_not_deleting,
|
|
|
|
errbackArgs=(blob_hash,))
|
|
|
|
|
|
|
|
ds.append(d)
|
|
|
|
else:
|
|
|
|
remove_from_list(blob_hash)
|
2015-09-04 22:22:02 +02:00
|
|
|
d = defer.fail(Failure(NoSuchBlobError(blob_hash)))
|
2015-09-08 21:42:56 +02:00
|
|
|
log.warning("Blob %s cannot be deleted because it is unknown")
|
2015-08-20 17:27:15 +02:00
|
|
|
ds.append(d)
|
2016-09-27 19:52:44 +02:00
|
|
|
return defer.DeferredList(ds)
|