2015-09-04 22:22:02 +02:00
|
|
|
import sqlite3
|
|
|
|
import unqlite
|
|
|
|
import leveldb
|
|
|
|
import shutil
|
|
|
|
import os
|
|
|
|
import logging
|
|
|
|
import json
|
|
|
|
|
|
|
|
|
2015-09-08 21:42:56 +02:00
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
2015-09-04 22:22:02 +02:00
|
|
|
known_dbs = ['lbryfile_desc.db', 'lbryfiles.db', 'valuable_blobs.db', 'blobs.db',
|
|
|
|
'lbryfile_blob.db', 'lbryfile_info.db', 'settings.db', 'blind_settings.db',
|
|
|
|
'blind_peers.db', 'blind_info.db', 'lbryfile_info.db', 'lbryfile_manager.db',
|
|
|
|
'live_stream.db', 'stream_info.db', 'stream_blob.db', 'stream_desc.db']
|
|
|
|
|
|
|
|
|
|
|
|
def do_move(from_dir, to_dir):
|
|
|
|
for known_db in known_dbs:
|
|
|
|
known_db_path = os.path.join(from_dir, known_db)
|
|
|
|
if os.path.exists(known_db_path):
|
2015-09-08 21:42:56 +02:00
|
|
|
log.debug("Moving %s to %s",
|
|
|
|
os.path.abspath(known_db_path),
|
|
|
|
os.path.abspath(os.path.join(to_dir, known_db)))
|
2015-09-04 22:22:02 +02:00
|
|
|
shutil.move(known_db_path, os.path.join(to_dir, known_db))
|
|
|
|
else:
|
2015-09-08 21:42:56 +02:00
|
|
|
log.debug("Did not find %s", os.path.abspath(known_db_path))
|
2015-09-04 22:22:02 +02:00
|
|
|
|
|
|
|
|
|
|
|
def do_migration(db_dir):
|
|
|
|
old_dir = os.path.join(db_dir, "_0_to_1_old")
|
|
|
|
new_dir = os.path.join(db_dir, "_0_to_1_new")
|
|
|
|
try:
|
2015-09-08 21:42:56 +02:00
|
|
|
log.info("Moving dbs from the real directory to %s", os.path.abspath(old_dir))
|
2015-09-04 22:22:02 +02:00
|
|
|
os.makedirs(old_dir)
|
|
|
|
do_move(db_dir, old_dir)
|
|
|
|
except:
|
2015-09-08 21:42:56 +02:00
|
|
|
log.error("An error occurred moving the old db files.")
|
2015-09-04 22:22:02 +02:00
|
|
|
raise
|
|
|
|
try:
|
2015-09-08 21:42:56 +02:00
|
|
|
log.info("Creating the new directory in %s", os.path.abspath(new_dir))
|
2015-09-04 22:22:02 +02:00
|
|
|
os.makedirs(new_dir)
|
|
|
|
|
|
|
|
except:
|
2015-09-08 21:42:56 +02:00
|
|
|
log.error("An error occurred creating the new directory.")
|
2015-09-04 22:22:02 +02:00
|
|
|
raise
|
|
|
|
try:
|
2015-09-08 21:42:56 +02:00
|
|
|
log.info("Doing the migration")
|
2015-09-04 22:22:02 +02:00
|
|
|
migrate_blob_db(old_dir, new_dir)
|
|
|
|
migrate_lbryfile_db(old_dir, new_dir)
|
|
|
|
migrate_livestream_db(old_dir, new_dir)
|
|
|
|
migrate_ptc_db(old_dir, new_dir)
|
|
|
|
migrate_lbryfile_manager_db(old_dir, new_dir)
|
|
|
|
migrate_settings_db(old_dir, new_dir)
|
|
|
|
migrate_repeater_db(old_dir, new_dir)
|
2015-09-08 21:42:56 +02:00
|
|
|
log.info("Migration succeeded")
|
2015-09-04 22:22:02 +02:00
|
|
|
except:
|
2015-09-08 21:42:56 +02:00
|
|
|
log.error("An error occurred during the migration. Restoring.")
|
2015-09-04 22:22:02 +02:00
|
|
|
do_move(old_dir, db_dir)
|
|
|
|
raise
|
|
|
|
try:
|
2015-09-08 21:42:56 +02:00
|
|
|
log.info("Moving dbs in the new directory to the real directory")
|
2015-09-04 22:22:02 +02:00
|
|
|
do_move(new_dir, db_dir)
|
|
|
|
db_revision = open(os.path.join(db_dir, 'db_revision'), mode='w+')
|
|
|
|
db_revision.write("1")
|
|
|
|
db_revision.close()
|
|
|
|
os.rmdir(new_dir)
|
|
|
|
except:
|
2015-09-08 21:42:56 +02:00
|
|
|
log.error("An error occurred moving the new db files.")
|
2015-09-04 22:22:02 +02:00
|
|
|
raise
|
|
|
|
return old_dir
|
|
|
|
|
|
|
|
|
|
|
|
def migrate_blob_db(old_db_dir, new_db_dir):
|
|
|
|
old_blob_db_path = os.path.join(old_db_dir, "blobs.db")
|
|
|
|
if not os.path.exists(old_blob_db_path):
|
|
|
|
return True
|
|
|
|
|
|
|
|
old_db = leveldb.LevelDB(old_blob_db_path)
|
|
|
|
new_db_conn = sqlite3.connect(os.path.join(new_db_dir, "blobs.db"))
|
|
|
|
c = new_db_conn.cursor()
|
|
|
|
c.execute("create table if not exists blobs (" +
|
|
|
|
" blob_hash text primary key, " +
|
|
|
|
" blob_length integer, " +
|
|
|
|
" last_verified_time real, " +
|
|
|
|
" next_announce_time real"
|
|
|
|
")")
|
|
|
|
new_db_conn.commit()
|
|
|
|
c = new_db_conn.cursor()
|
|
|
|
for blob_hash, blob_info in old_db.RangeIter():
|
|
|
|
blob_length, verified_time, announce_time = json.loads(blob_info)
|
|
|
|
c.execute("insert into blobs values (?, ?, ?, ?)",
|
|
|
|
(blob_hash, blob_length, verified_time, announce_time))
|
|
|
|
new_db_conn.commit()
|
|
|
|
new_db_conn.close()
|
|
|
|
|
|
|
|
|
|
|
|
def migrate_lbryfile_db(old_db_dir, new_db_dir):
|
|
|
|
old_lbryfile_db_path = os.path.join(old_db_dir, "lbryfiles.db")
|
|
|
|
if not os.path.exists(old_lbryfile_db_path):
|
|
|
|
return True
|
|
|
|
|
|
|
|
stream_info_db = leveldb.LevelDB(os.path.join(old_db_dir, "lbryfile_info.db"))
|
|
|
|
stream_blob_db = leveldb.LevelDB(os.path.join(old_db_dir, "lbryfile_blob.db"))
|
|
|
|
stream_desc_db = leveldb.LevelDB(os.path.join(old_db_dir, "lbryfile_desc.db"))
|
|
|
|
|
|
|
|
db_conn = sqlite3.connect(os.path.join(new_db_dir, "lbryfile_info.db"))
|
|
|
|
c = db_conn.cursor()
|
|
|
|
c.execute("create table if not exists lbry_files (" +
|
|
|
|
" stream_hash text primary key, " +
|
|
|
|
" key text, " +
|
|
|
|
" stream_name text, " +
|
|
|
|
" suggested_file_name text" +
|
|
|
|
")")
|
|
|
|
c.execute("create table if not exists lbry_file_blobs (" +
|
|
|
|
" blob_hash text, " +
|
|
|
|
" stream_hash text, " +
|
|
|
|
" position integer, " +
|
|
|
|
" iv text, " +
|
|
|
|
" length integer, " +
|
|
|
|
" foreign key(stream_hash) references lbry_files(stream_hash)" +
|
|
|
|
")")
|
|
|
|
c.execute("create table if not exists lbry_file_descriptors (" +
|
|
|
|
" sd_blob_hash TEXT PRIMARY KEY, " +
|
|
|
|
" stream_hash TEXT, " +
|
|
|
|
" foreign key(stream_hash) references lbry_files(stream_hash)" +
|
|
|
|
")")
|
|
|
|
db_conn.commit()
|
|
|
|
c = db_conn.cursor()
|
|
|
|
for stream_hash, stream_info in stream_info_db.RangeIter():
|
|
|
|
key, name, suggested_file_name = json.loads(stream_info)
|
|
|
|
c.execute("insert into lbry_files values (?, ?, ?, ?)",
|
|
|
|
(stream_hash, key, name, suggested_file_name))
|
|
|
|
db_conn.commit()
|
|
|
|
c = db_conn.cursor()
|
|
|
|
for blob_hash_stream_hash, blob_info in stream_blob_db.RangeIter():
|
|
|
|
b_h, s_h = json.loads(blob_hash_stream_hash)
|
|
|
|
position, iv, length = json.loads(blob_info)
|
|
|
|
c.execute("insert into lbry_file_blobs values (?, ?, ?, ?, ?)",
|
|
|
|
(b_h, s_h, position, iv, length))
|
|
|
|
db_conn.commit()
|
|
|
|
c = db_conn.cursor()
|
|
|
|
for sd_blob_hash, stream_hash in stream_desc_db.RangeIter():
|
|
|
|
c.execute("insert into lbry_file_descriptors values (?, ?)",
|
|
|
|
(sd_blob_hash, stream_hash))
|
|
|
|
db_conn.commit()
|
|
|
|
db_conn.close()
|
|
|
|
|
|
|
|
|
|
|
|
def migrate_livestream_db(old_db_dir, new_db_dir):
|
|
|
|
old_db_path = os.path.join(old_db_dir, "stream_info.db")
|
|
|
|
if not os.path.exists(old_db_path):
|
|
|
|
return True
|
|
|
|
stream_info_db = leveldb.LevelDB(os.path.join(old_db_dir, "stream_info.db"))
|
|
|
|
stream_blob_db = leveldb.LevelDB(os.path.join(old_db_dir, "stream_blob.db"))
|
|
|
|
stream_desc_db = leveldb.LevelDB(os.path.join(old_db_dir, "stream_desc.db"))
|
|
|
|
|
|
|
|
db_conn = sqlite3.connect(os.path.join(new_db_dir, "live_stream.db"))
|
|
|
|
|
|
|
|
c = db_conn.cursor()
|
|
|
|
|
|
|
|
c.execute("create table if not exists live_streams (" +
|
|
|
|
" stream_hash text primary key, " +
|
|
|
|
" public_key text, " +
|
|
|
|
" key text, " +
|
|
|
|
" stream_name text, " +
|
|
|
|
" next_announce_time real" +
|
|
|
|
")")
|
|
|
|
c.execute("create table if not exists live_stream_blobs (" +
|
|
|
|
" blob_hash text, " +
|
|
|
|
" stream_hash text, " +
|
|
|
|
" position integer, " +
|
|
|
|
" revision integer, " +
|
|
|
|
" iv text, " +
|
|
|
|
" length integer, " +
|
|
|
|
" signature text, " +
|
|
|
|
" foreign key(stream_hash) references live_streams(stream_hash)" +
|
|
|
|
")")
|
|
|
|
c.execute("create table if not exists live_stream_descriptors (" +
|
|
|
|
" sd_blob_hash TEXT PRIMARY KEY, " +
|
|
|
|
" stream_hash TEXT, " +
|
|
|
|
" foreign key(stream_hash) references live_streams(stream_hash)" +
|
|
|
|
")")
|
|
|
|
|
|
|
|
db_conn.commit()
|
|
|
|
|
|
|
|
c = db_conn.cursor()
|
|
|
|
for stream_hash, stream_info in stream_info_db.RangeIter():
|
|
|
|
public_key, key, name, next_announce_time = json.loads(stream_info)
|
|
|
|
c.execute("insert into live_streams values (?, ?, ?, ?, ?)",
|
|
|
|
(stream_hash, public_key, key, name, next_announce_time))
|
|
|
|
db_conn.commit()
|
|
|
|
c = db_conn.cursor()
|
|
|
|
for blob_hash_stream_hash, blob_info in stream_blob_db.RangeIter():
|
|
|
|
b_h, s_h = json.loads(blob_hash_stream_hash)
|
|
|
|
position, revision, iv, length, signature = json.loads(blob_info)
|
|
|
|
c.execute("insert into live_stream_blobs values (?, ?, ?, ?, ?, ?, ?)",
|
|
|
|
(b_h, s_h, position, revision, iv, length, signature))
|
|
|
|
db_conn.commit()
|
|
|
|
c = db_conn.cursor()
|
|
|
|
for sd_blob_hash, stream_hash in stream_desc_db.RangeIter():
|
|
|
|
c.execute("insert into live_stream_descriptors values (?, ?)",
|
|
|
|
(sd_blob_hash, stream_hash))
|
|
|
|
db_conn.commit()
|
|
|
|
db_conn.close()
|
|
|
|
|
|
|
|
|
|
|
|
def migrate_ptc_db(old_db_dir, new_db_dir):
|
|
|
|
old_db_path = os.path.join(old_db_dir, "ptcwallet.db")
|
|
|
|
if not os.path.exists(old_db_path):
|
|
|
|
return True
|
|
|
|
old_db = leveldb.LevelDB(old_db_path)
|
|
|
|
try:
|
|
|
|
p_key = old_db.Get("private_key")
|
|
|
|
new_db = unqlite.UnQLite(os.path.join(new_db_dir, "ptcwallet.db"))
|
|
|
|
new_db['private_key'] = p_key
|
|
|
|
except KeyError:
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
def migrate_lbryfile_manager_db(old_db_dir, new_db_dir):
|
|
|
|
old_db_path = os.path.join(old_db_dir, "lbryfiles.db")
|
|
|
|
if not os.path.exists(old_db_path):
|
|
|
|
return True
|
|
|
|
old_db = leveldb.LevelDB(old_db_path)
|
|
|
|
new_db = sqlite3.connect(os.path.join(new_db_dir, "lbryfile_info.db"))
|
|
|
|
c = new_db.cursor()
|
|
|
|
c.execute("create table if not exists lbry_file_options (" +
|
|
|
|
" blob_data_rate real, " +
|
|
|
|
" status text," +
|
|
|
|
" stream_hash text,"
|
|
|
|
" foreign key(stream_hash) references lbry_files(stream_hash)" +
|
|
|
|
")")
|
|
|
|
new_db.commit()
|
|
|
|
LBRYFILE_STATUS = "t"
|
|
|
|
LBRYFILE_OPTIONS = "o"
|
|
|
|
c = new_db.cursor()
|
|
|
|
for k, v in old_db.RangeIter():
|
|
|
|
key_type, stream_hash = json.loads(k)
|
|
|
|
if key_type == LBRYFILE_STATUS:
|
|
|
|
try:
|
|
|
|
rate = json.loads(old_db.Get(json.dumps((LBRYFILE_OPTIONS, stream_hash))))[0]
|
|
|
|
except KeyError:
|
|
|
|
rate = None
|
|
|
|
c.execute("insert into lbry_file_options values (?, ?, ?)",
|
|
|
|
(rate, v, stream_hash))
|
|
|
|
new_db.commit()
|
|
|
|
new_db.close()
|
|
|
|
|
|
|
|
|
|
|
|
def migrate_settings_db(old_db_dir, new_db_dir):
|
|
|
|
old_settings_db_path = os.path.join(old_db_dir, "settings.db")
|
|
|
|
if not os.path.exists(old_settings_db_path):
|
|
|
|
return True
|
|
|
|
old_db = leveldb.LevelDB(old_settings_db_path)
|
|
|
|
new_db = unqlite.UnQLite(os.path.join(new_db_dir, "settings.db"))
|
|
|
|
for k, v in old_db.RangeIter():
|
|
|
|
new_db[k] = v
|
|
|
|
|
|
|
|
|
|
|
|
def migrate_repeater_db(old_db_dir, new_db_dir):
|
|
|
|
old_repeater_db_path = os.path.join(old_db_dir, "valuable_blobs.db")
|
|
|
|
if not os.path.exists(old_repeater_db_path):
|
|
|
|
return True
|
|
|
|
old_db = leveldb.LevelDB(old_repeater_db_path)
|
|
|
|
info_db = sqlite3.connect(os.path.join(new_db_dir, "blind_info.db"))
|
|
|
|
peer_db = sqlite3.connect(os.path.join(new_db_dir, "blind_peers.db"))
|
|
|
|
unql_db = unqlite.UnQLite(os.path.join(new_db_dir, "blind_settings.db"))
|
|
|
|
BLOB_INFO_TYPE = 'b'
|
|
|
|
SETTING_TYPE = 's'
|
|
|
|
PEER_TYPE = 'p'
|
|
|
|
info_c = info_db.cursor()
|
|
|
|
info_c.execute("create table if not exists valuable_blobs (" +
|
|
|
|
" blob_hash text primary key, " +
|
|
|
|
" blob_length integer, " +
|
|
|
|
" reference text, " +
|
|
|
|
" peer_host text, " +
|
|
|
|
" peer_port integer, " +
|
|
|
|
" peer_score text" +
|
|
|
|
")")
|
|
|
|
info_db.commit()
|
|
|
|
peer_c = peer_db.cursor()
|
|
|
|
peer_c.execute("create table if not exists approved_peers (" +
|
|
|
|
" ip_address text, " +
|
|
|
|
" port integer" +
|
|
|
|
")")
|
|
|
|
peer_db.commit()
|
|
|
|
info_c = info_db.cursor()
|
|
|
|
peer_c = peer_db.cursor()
|
|
|
|
for k, v in old_db.RangeIter():
|
|
|
|
key_type, key_rest = json.loads(k)
|
|
|
|
if key_type == PEER_TYPE:
|
|
|
|
host, port = key_rest
|
|
|
|
peer_c.execute("insert into approved_peers values (?, ?)",
|
|
|
|
(host, port))
|
|
|
|
elif key_type == SETTING_TYPE:
|
|
|
|
unql_db[key_rest] = v
|
|
|
|
elif key_type == BLOB_INFO_TYPE:
|
|
|
|
blob_hash = key_rest
|
|
|
|
length, reference, peer_host, peer_port, peer_score = json.loads(v)
|
|
|
|
info_c.execute("insert into valuable_blobs values (?, ?, ?, ?, ?, ?)",
|
|
|
|
(blob_hash, length, reference, peer_host, peer_port, peer_score))
|
|
|
|
info_db.commit()
|
|
|
|
peer_db.commit()
|
|
|
|
info_db.close()
|
|
|
|
peer_db.close()
|