2016-07-25 19:19:19 +02:00
import binascii
2016-07-25 19:25:10 +02:00
import distutils . version
2016-03-03 00:32:48 +01:00
import locale
2016-07-25 19:19:19 +02:00
import logging . handlers
2016-06-03 09:45:46 +02:00
import mimetypes
2016-02-29 19:25:47 +01:00
import os
2016-07-25 19:19:19 +02:00
import platform
import random
2016-07-20 17:44:06 +02:00
import re
2016-07-25 19:19:19 +02:00
import socket
import string
2016-05-30 21:49:25 +02:00
import subprocess
2016-02-29 19:25:47 +01:00
import sys
2016-04-07 09:12:09 +02:00
import base58
2016-07-25 19:19:19 +02:00
import requests
import simplejson as json
2016-07-27 09:54:10 +02:00
import pkg_resources
2016-03-24 03:27:48 +01:00
2016-07-27 09:54:10 +02:00
from urllib2 import urlopen
from appdirs import user_data_dir
from datetime import datetime
from decimal import Decimal
2016-04-21 04:02:52 +02:00
from twisted . web import server
2016-03-24 03:27:48 +01:00
from twisted . internet import defer , threads , error , reactor
2016-04-18 19:21:53 +02:00
from twisted . internet . task import LoopingCall
2016-03-24 03:27:48 +01:00
from txjsonrpc import jsonrpclib
from txjsonrpc . web import jsonrpc
2016-07-29 03:25:37 +02:00
from txjsonrpc . web . jsonrpc import Handler , Proxy
2016-02-29 19:25:47 +01:00
2016-04-12 04:28:46 +02:00
from lbrynet import __version__ as lbrynet_version
2016-06-28 08:06:04 +02:00
from lbryum . version import LBRYUM_VERSION as lbryum_version
2016-01-26 02:28:05 +01:00
from lbrynet . core . PaymentRateManager import PaymentRateManager
from lbrynet . core . server . BlobAvailabilityHandler import BlobAvailabilityHandlerFactory
from lbrynet . core . server . BlobRequestHandler import BlobRequestHandlerFactory
from lbrynet . core . server . ServerProtocol import ServerProtocolFactory
2016-07-29 00:48:29 +02:00
from lbrynet . core . Error import UnknownNameError , InsufficientFundsError , InvalidNameError
2015-12-06 23:32:17 +01:00
from lbrynet . lbryfile . StreamDescriptor import LBRYFileStreamType
from lbrynet . lbryfile . client . LBRYFileDownloader import LBRYFileSaverFactory , LBRYFileOpenerFactory
from lbrynet . lbryfile . client . LBRYFileOptions import add_lbry_file_to_sd_identifier
2016-05-30 21:49:25 +02:00
from lbrynet . lbrynet_daemon . LBRYUIManager import LBRYUIManager
2016-05-04 11:20:38 +02:00
from lbrynet . lbrynet_daemon . LBRYDownloader import GetStream
2016-01-21 04:00:28 +01:00
from lbrynet . lbrynet_daemon . LBRYPublisher import Publisher
2016-07-28 11:30:13 +02:00
from lbrynet . lbrynet_daemon . LBRYExchangeRateManager import ExchangeRateManager
2016-08-03 00:34:48 +02:00
from lbrynet . lbrynet_daemon . Lighthouse import LighthouseClient
2016-08-05 02:43:39 +02:00
from lbrynet . core . LBRYMetadata import Metadata
2016-07-25 23:09:13 +02:00
from lbrynet . core import utils
2016-07-29 00:48:29 +02:00
from lbrynet . core . LBRYMetadata import verify_name_characters
2015-12-06 23:32:17 +01:00
from lbrynet . core . utils import generate_id
from lbrynet . lbrynet_console . LBRYSettings import LBRYSettings
2016-04-21 04:02:52 +02:00
from lbrynet . conf import MIN_BLOB_DATA_PAYMENT_RATE , DEFAULT_MAX_SEARCH_RESULTS , KNOWN_DHT_NODES , DEFAULT_MAX_KEY_FEE , \
2016-07-21 22:11:14 +02:00
DEFAULT_WALLET , DEFAULT_SEARCH_TIMEOUT , DEFAULT_CACHE_TIME , DEFAULT_UI_BRANCH , LOG_POST_URL , LOG_FILE_NAME , SOURCE_TYPES
2016-08-03 23:57:26 +02:00
from lbrynet . conf import DEFAULT_SD_DOWNLOAD_TIMEOUT
2016-05-30 21:49:25 +02:00
from lbrynet . conf import DEFAULT_TIMEOUT , WALLET_TYPES
2016-08-02 04:31:49 +02:00
from lbrynet . core . StreamDescriptor import StreamDescriptorIdentifier , download_sd_blob , BlobStreamDescriptorReader
2015-12-06 23:32:17 +01:00
from lbrynet . core . Session import LBRYSession
from lbrynet . core . PTCWallet import PTCWallet
2016-07-04 10:57:30 +02:00
from lbrynet . core . LBRYWallet import LBRYcrdWallet , LBRYumWallet
2015-12-06 23:32:17 +01:00
from lbrynet . lbryfilemanager . LBRYFileManager import LBRYFileManager
from lbrynet . lbryfile . LBRYFileMetadataManager import DBLBRYFileMetadataManager , TempLBRYFileMetadataManager
2016-06-28 20:28:59 +02:00
# from lbryum import LOG_PATH as lbryum_log
2016-02-23 04:32:07 +01:00
2016-06-07 10:19:51 +02:00
2016-07-25 20:04:30 +02:00
# TODO: this code snippet is everywhere. Make it go away
2016-06-07 10:19:51 +02:00
if sys . platform != " darwin " :
log_dir = os . path . join ( os . path . expanduser ( " ~ " ) , " .lbrynet " )
else :
log_dir = user_data_dir ( " LBRY " )
if not os . path . isdir ( log_dir ) :
os . mkdir ( log_dir )
lbrynet_log = os . path . join ( log_dir , LOG_FILE_NAME )
2016-07-25 07:40:26 +02:00
log = logging . getLogger ( __name__ )
2016-06-28 20:28:59 +02:00
2016-06-04 23:05:14 +02:00
if os . path . isfile ( lbrynet_log ) :
2016-07-25 20:04:30 +02:00
with open ( lbrynet_log , ' r ' ) as f :
PREVIOUS_LBRYNET_LOG = len ( f . read ( ) )
2016-04-18 01:54:04 +02:00
else :
2016-06-04 23:05:14 +02:00
PREVIOUS_LBRYNET_LOG = 0
2016-04-07 09:12:09 +02:00
2016-04-18 19:21:53 +02:00
INITIALIZING_CODE = ' initializing '
LOADING_DB_CODE = ' loading_db '
LOADING_WALLET_CODE = ' loading_wallet '
LOADING_FILE_MANAGER_CODE = ' loading_file_manager '
LOADING_SERVER_CODE = ' loading_server '
STARTED_CODE = ' started '
2016-05-11 08:47:33 +02:00
WAITING_FOR_FIRST_RUN_CREDITS = ' waiting_for_credits '
2016-04-09 08:09:30 +02:00
STARTUP_STAGES = [
2016-04-18 19:21:53 +02:00
( INITIALIZING_CODE , ' Initializing... ' ) ,
( LOADING_DB_CODE , ' Loading databases... ' ) ,
( LOADING_WALLET_CODE , ' Catching up with the blockchain... %s ' ) ,
( LOADING_FILE_MANAGER_CODE , ' Setting up file manager ' ) ,
( LOADING_SERVER_CODE , ' Starting lbrynet ' ) ,
2016-05-11 08:47:33 +02:00
( STARTED_CODE , ' Started lbrynet ' ) ,
( WAITING_FOR_FIRST_RUN_CREDITS , ' Waiting for first run credits... ' )
2016-04-09 08:09:30 +02:00
]
2016-04-26 04:35:21 +02:00
DOWNLOAD_METADATA_CODE = ' downloading_metadata '
DOWNLOAD_TIMEOUT_CODE = ' timeout '
DOWNLOAD_RUNNING_CODE = ' running '
DOWNLOAD_STOPPED_CODE = ' stopped '
STREAM_STAGES = [
( INITIALIZING_CODE , ' Initializing... ' ) ,
( DOWNLOAD_METADATA_CODE , ' Downloading metadata ' ) ,
2016-05-06 04:26:25 +02:00
( DOWNLOAD_RUNNING_CODE , ' Started %s , got %s / %s blobs, stream status: %s ' ) ,
2016-04-26 04:35:21 +02:00
( DOWNLOAD_STOPPED_CODE , ' Paused stream ' ) ,
( DOWNLOAD_TIMEOUT_CODE , ' Stream timed out ' )
]
2016-04-18 19:21:53 +02:00
CONNECT_CODE_VERSION_CHECK = ' version_check '
CONNECT_CODE_NETWORK = ' network_connection '
CONNECT_CODE_WALLET = ' wallet_catchup_lag '
2016-04-26 04:35:21 +02:00
CONNECTION_PROBLEM_CODES = [
( CONNECT_CODE_VERSION_CHECK , " There was a problem checking for updates on github " ) ,
( CONNECT_CODE_NETWORK , " Your internet connection appears to have been interrupted " ) ,
( CONNECT_CODE_WALLET , " Synchronization with the blockchain is lagging... if this continues try restarting LBRY " )
]
2016-04-18 19:21:53 +02:00
2016-04-13 20:47:34 +02:00
ALLOWED_DURING_STARTUP = [ ' is_running ' , ' is_first_run ' ,
' get_time_behind_blockchain ' , ' stop ' ,
' daemon_status ' , ' get_start_notice ' ,
2016-05-30 21:49:25 +02:00
' version ' ]
2016-04-09 08:09:30 +02:00
2016-03-24 03:27:48 +01:00
BAD_REQUEST = 400
NOT_FOUND = 404
OK_CODE = 200
2016-04-22 08:45:05 +02:00
2016-02-16 19:39:08 +01:00
# TODO add login credentials in a conf file
# TODO alert if your copy of a lbry file is out of date with the name record
2016-01-12 00:01:46 +01:00
2015-12-06 23:32:17 +01:00
2016-04-18 19:21:53 +02:00
REMOTE_SERVER = " www.google.com "
2016-07-21 17:44:59 +02:00
class Parameters ( object ) :
def __init__ ( self , * * kwargs ) :
self . __dict__ . update ( kwargs )
2016-03-24 03:27:48 +01:00
class LBRYDaemon ( jsonrpc . JSONRPC ) :
2015-12-06 23:32:17 +01:00
"""
2016-03-24 03:27:48 +01:00
LBRYnet daemon , a jsonrpc interface to lbry functions
2015-12-06 23:32:17 +01:00
"""
2016-04-12 04:28:46 +02:00
2016-03-24 03:27:48 +01:00
isLeaf = True
2016-07-09 19:31:07 +02:00
def __init__ ( self , root , wallet_type = None ) :
2016-04-12 04:28:46 +02:00
jsonrpc . JSONRPC . __init__ ( self )
reactor . addSystemEventTrigger ( ' before ' , ' shutdown ' , self . _shutdown )
2016-04-12 08:03:57 +02:00
self . startup_status = STARTUP_STAGES [ 0 ]
self . startup_message = None
self . announced_startup = False
2016-04-18 19:21:53 +02:00
self . connected_to_internet = True
self . connection_problem = None
2016-04-12 08:03:57 +02:00
self . query_handlers = { }
2016-07-30 04:36:30 +02:00
self . git_lbrynet_version = None
self . git_lbryum_version = None
2016-05-30 21:49:25 +02:00
self . ui_version = None
self . ip = None
2016-07-17 20:43:35 +02:00
# TODO: this is confusing to set here, and then to be reset below.
2016-04-12 08:03:57 +02:00
self . wallet_type = wallet_type
self . first_run = None
2016-06-07 10:19:51 +02:00
self . log_file = lbrynet_log
2016-04-12 04:28:46 +02:00
self . current_db_revision = 1
self . run_server = True
self . session = None
2016-07-28 11:30:13 +02:00
self . exchange_rate_manager = ExchangeRateManager ( )
2016-08-03 00:34:48 +02:00
self . lighthouse_client = LighthouseClient ( )
2016-04-23 00:48:44 +02:00
self . waiting_on = { }
2016-05-04 11:20:38 +02:00
self . streams = { }
2016-07-27 09:00:44 +02:00
self . pending_claims = { }
2016-04-12 04:28:46 +02:00
self . known_dht_nodes = KNOWN_DHT_NODES
2016-05-30 21:49:25 +02:00
self . first_run_after_update = False
2016-07-20 08:36:55 +02:00
self . uploaded_temp_files = [ ]
2016-04-12 08:03:57 +02:00
2016-04-12 04:28:46 +02:00
if os . name == " nt " :
from lbrynet . winhelpers . knownpaths import get_path , FOLDERID , UserHandle
2016-04-22 08:45:05 +02:00
default_download_directory = get_path ( FOLDERID . Downloads , UserHandle . current )
2016-04-12 08:03:57 +02:00
self . db_dir = os . path . join ( get_path ( FOLDERID . RoamingAppData , UserHandle . current ) , " lbrynet " )
2016-04-12 04:28:46 +02:00
elif sys . platform == " darwin " :
2016-04-22 08:45:05 +02:00
default_download_directory = os . path . join ( os . path . expanduser ( " ~ " ) , ' Downloads ' )
2016-04-12 08:03:57 +02:00
self . db_dir = user_data_dir ( " LBRY " )
2016-04-12 04:28:46 +02:00
else :
2016-07-04 23:46:26 +02:00
default_download_directory = os . path . join ( os . path . expanduser ( " ~ " ) , ' Downloads ' )
2016-04-12 08:03:57 +02:00
self . db_dir = os . path . join ( os . path . expanduser ( " ~ " ) , " .lbrynet " )
2016-07-05 02:02:38 +02:00
try :
if not os . path . isdir ( default_download_directory ) :
os . mkdir ( default_download_directory )
except :
log . info ( " Couldn ' t make download directory, using home " )
default_download_directory = os . path . expanduser ( " ~ " )
2016-04-12 08:03:57 +02:00
2016-04-12 04:28:46 +02:00
self . daemon_conf = os . path . join ( self . db_dir , ' daemon_settings.json ' )
2016-04-12 08:03:57 +02:00
2016-04-12 04:28:46 +02:00
self . default_settings = {
' run_on_startup ' : False ,
' data_rate ' : MIN_BLOB_DATA_PAYMENT_RATE ,
2016-04-12 08:03:57 +02:00
' max_key_fee ' : DEFAULT_MAX_KEY_FEE ,
2016-04-22 08:45:05 +02:00
' download_directory ' : default_download_directory ,
2016-04-12 04:28:46 +02:00
' max_upload ' : 0.0 ,
' max_download ' : 0.0 ,
2016-04-12 08:03:57 +02:00
' upload_log ' : True ,
2016-04-22 08:45:05 +02:00
' search_timeout ' : DEFAULT_SEARCH_TIMEOUT ,
' download_timeout ' : DEFAULT_TIMEOUT ,
2016-04-12 08:03:57 +02:00
' max_search_results ' : DEFAULT_MAX_SEARCH_RESULTS ,
2016-05-14 23:36:30 +02:00
' wallet_type ' : DEFAULT_WALLET ,
2016-04-12 08:03:57 +02:00
' delete_blobs_on_remove ' : True ,
' peer_port ' : 3333 ,
' dht_node_port ' : 4444 ,
' use_upnp ' : True ,
' start_lbrycrdd ' : True ,
2016-04-26 04:35:21 +02:00
' requested_first_run_credits ' : False ,
2016-05-30 21:49:25 +02:00
' cache_time ' : DEFAULT_CACHE_TIME ,
' startup_scripts ' : [ ] ,
' last_version ' : { ' lbrynet ' : lbrynet_version , ' lbryum ' : lbryum_version }
2016-04-12 04:28:46 +02:00
}
2016-04-22 08:45:05 +02:00
if os . path . isfile ( self . daemon_conf ) :
f = open ( self . daemon_conf , " r " )
loaded_settings = json . loads ( f . read ( ) )
f . close ( )
missing_settings = { }
removed_settings = { }
for k in self . default_settings . keys ( ) :
if k not in loaded_settings . keys ( ) :
missing_settings [ k ] = self . default_settings [ k ]
for k in loaded_settings . keys ( ) :
if not k in self . default_settings . keys ( ) :
log . info ( " Removing unused setting: " + k + " with value: " + str ( loaded_settings [ k ] ) )
removed_settings [ k ] = loaded_settings [ k ]
del loaded_settings [ k ]
for k in missing_settings . keys ( ) :
log . info ( " Adding missing setting: " + k + " with default value: " + str ( missing_settings [ k ] ) )
loaded_settings [ k ] = missing_settings [ k ]
2016-06-04 20:18:58 +02:00
if loaded_settings [ ' wallet_type ' ] != self . wallet_type and self . wallet_type :
loaded_settings [ ' wallet_type ' ] = self . wallet_type
2016-04-22 08:45:05 +02:00
if missing_settings or removed_settings :
2016-06-04 20:18:58 +02:00
log . info ( " Updated and loaded lbrynet-daemon configuration " )
2016-04-22 08:45:05 +02:00
else :
log . info ( " Loaded lbrynet-daemon configuration " )
2016-06-04 20:18:58 +02:00
self . session_settings = loaded_settings
2016-04-22 08:45:05 +02:00
else :
2016-06-02 03:05:38 +02:00
missing_settings = self . default_settings
log . info ( " Writing default settings : " + json . dumps ( self . default_settings ) + " --> " + str ( self . daemon_conf ) )
2016-06-04 20:18:58 +02:00
self . session_settings = self . default_settings
2016-04-22 08:45:05 +02:00
2016-05-30 21:49:25 +02:00
if ' last_version ' in missing_settings . keys ( ) :
self . session_settings [ ' last_version ' ] = None
if self . session_settings [ ' last_version ' ] != self . default_settings [ ' last_version ' ] :
self . session_settings [ ' last_version ' ] = self . default_settings [ ' last_version ' ]
self . first_run_after_update = True
log . info ( " First run after update " )
2016-07-21 22:11:14 +02:00
log . info ( " lbrynet %s --> %s " % ( self . session_settings [ ' last_version ' ] [ ' lbrynet ' ] , self . default_settings [ ' last_version ' ] [ ' lbrynet ' ] ) )
log . info ( " lbryum %s --> %s " % ( self . session_settings [ ' last_version ' ] [ ' lbryum ' ] , self . default_settings [ ' last_version ' ] [ ' lbryum ' ] ) )
2016-05-30 21:49:25 +02:00
2016-06-04 20:18:58 +02:00
f = open ( self . daemon_conf , " w " )
f . write ( json . dumps ( self . session_settings ) )
f . close ( )
2016-04-22 08:45:05 +02:00
self . run_on_startup = self . session_settings [ ' run_on_startup ' ]
self . data_rate = self . session_settings [ ' data_rate ' ]
self . max_key_fee = self . session_settings [ ' max_key_fee ' ]
self . download_directory = self . session_settings [ ' download_directory ' ]
self . max_upload = self . session_settings [ ' max_upload ' ]
self . max_download = self . session_settings [ ' max_download ' ]
self . upload_log = self . session_settings [ ' upload_log ' ]
self . search_timeout = self . session_settings [ ' search_timeout ' ]
self . download_timeout = self . session_settings [ ' download_timeout ' ]
self . max_search_results = self . session_settings [ ' max_search_results ' ]
2016-07-17 20:43:35 +02:00
####
#
# Ignore the saved wallet type. Some users will have their wallet type
2016-07-18 18:30:21 +02:00
# saved as lbrycrd and we want wallets to be lbryum unless explicitly
# set on the command line to be lbrycrd.
2016-07-17 20:43:35 +02:00
#
# if self.session_settings['wallet_type'] in WALLET_TYPES and not wallet_type:
# self.wallet_type = self.session_settings['wallet_type']
# log.info("Using wallet type %s from config" % self.wallet_type)
# else:
# self.wallet_type = wallet_type
# self.session_settings['wallet_type'] = wallet_type
# log.info("Using wallet type %s specified from command line" % self.wallet_type)
#
# Instead, if wallet is not set on the command line, default to the default wallet
#
if wallet_type :
log . info ( " Using wallet type %s specified from command line " , wallet_type )
2016-06-02 02:52:15 +02:00
self . wallet_type = wallet_type
2016-07-17 20:43:35 +02:00
else :
log . info ( " Using the default wallet type %s " , DEFAULT_WALLET )
self . wallet_type = DEFAULT_WALLET
#
####
2016-04-22 08:45:05 +02:00
self . delete_blobs_on_remove = self . session_settings [ ' delete_blobs_on_remove ' ]
self . peer_port = self . session_settings [ ' peer_port ' ]
self . dht_node_port = self . session_settings [ ' dht_node_port ' ]
self . use_upnp = self . session_settings [ ' use_upnp ' ]
self . start_lbrycrdd = self . session_settings [ ' start_lbrycrdd ' ]
self . requested_first_run_credits = self . session_settings [ ' requested_first_run_credits ' ]
2016-04-26 04:35:21 +02:00
self . cache_time = self . session_settings [ ' cache_time ' ]
2016-05-30 21:49:25 +02:00
self . startup_scripts = self . session_settings [ ' startup_scripts ' ]
2016-04-26 04:35:21 +02:00
if os . path . isfile ( os . path . join ( self . db_dir , " stream_info_cache.json " ) ) :
f = open ( os . path . join ( self . db_dir , " stream_info_cache.json " ) , " r " )
self . name_cache = json . loads ( f . read ( ) )
f . close ( )
log . info ( " Loaded claim info cache " )
else :
self . name_cache = { }
2016-04-22 08:45:05 +02:00
2016-05-14 23:36:30 +02:00
if os . name == " nt " :
from lbrynet . winhelpers . knownpaths import get_path , FOLDERID , UserHandle
self . lbrycrdd_path = " lbrycrdd.exe "
if self . wallet_type == " lbrycrd " :
self . wallet_dir = os . path . join ( get_path ( FOLDERID . RoamingAppData , UserHandle . current ) , " lbrycrd " )
else :
self . wallet_dir = os . path . join ( get_path ( FOLDERID . RoamingAppData , UserHandle . current ) , " lbryum " )
elif sys . platform == " darwin " :
2016-07-04 22:18:08 +02:00
# use the path from the bundle if its available.
try :
import Foundation
bundle = Foundation . NSBundle . mainBundle ( )
self . lbrycrdd_path = bundle . pathForResource_ofType_ ( ' lbrycrdd ' , None )
except Exception :
log . exception ( ' Failed to get path from bundle, falling back to default ' )
self . lbrycrdd_path = " ./lbrycrdd "
2016-05-14 23:36:30 +02:00
if self . wallet_type == " lbrycrd " :
self . wallet_dir = user_data_dir ( " lbrycrd " )
else :
self . wallet_dir = user_data_dir ( " LBRY " )
else :
2016-07-07 04:43:27 +02:00
self . lbrycrdd_path = " lbrycrdd "
2016-05-14 23:36:30 +02:00
if self . wallet_type == " lbrycrd " :
self . wallet_dir = os . path . join ( os . path . expanduser ( " ~ " ) , " .lbrycrd " )
else :
self . wallet_dir = os . path . join ( os . path . expanduser ( " ~ " ) , " .lbryum " )
2016-07-05 01:34:51 +02:00
if os . name != ' nt ' :
lbrycrdd_path_conf = os . path . join ( os . path . expanduser ( " ~ " ) , " .lbrycrddpath.conf " )
if not os . path . isfile ( lbrycrdd_path_conf ) :
2016-07-05 02:19:04 +02:00
f = open ( lbrycrdd_path_conf , " w " )
f . write ( str ( self . lbrycrdd_path ) )
2016-07-05 01:34:51 +02:00
f . close ( )
2016-05-14 23:36:30 +02:00
self . created_data_dir = False
if not os . path . exists ( self . db_dir ) :
os . mkdir ( self . db_dir )
self . created_data_dir = True
self . blobfile_dir = os . path . join ( self . db_dir , " blobfiles " )
self . lbrycrd_conf = os . path . join ( self . wallet_dir , " lbrycrd.conf " )
self . autofetcher_conf = os . path . join ( self . wallet_dir , " autofetcher.conf " )
self . wallet_conf = os . path . join ( self . wallet_dir , " lbrycrd.conf " )
self . wallet_user = None
self . wallet_password = None
self . internet_connection_checker = LoopingCall ( self . _check_network_connection )
self . version_checker = LoopingCall ( self . _check_remote_versions )
self . connection_problem_checker = LoopingCall ( self . _check_connection_problems )
2016-07-27 09:00:44 +02:00
self . pending_claim_checker = LoopingCall ( self . _check_pending_claims )
2016-05-14 23:36:30 +02:00
# self.lbrynet_connection_checker = LoopingCall(self._check_lbrynet_connection)
self . sd_identifier = StreamDescriptorIdentifier ( )
self . stream_info_manager = TempLBRYFileMetadataManager ( )
self . settings = LBRYSettings ( self . db_dir )
2016-05-30 21:49:25 +02:00
self . lbry_ui_manager = LBRYUIManager ( root )
2016-05-14 23:36:30 +02:00
self . blob_request_payment_rate_manager = None
self . lbry_file_metadata_manager = None
self . lbry_file_manager = None
2016-07-04 22:16:34 +02:00
if self . wallet_type == " lbrycrd " :
if os . path . isfile ( self . lbrycrd_conf ) :
log . info ( " Using lbrycrd.conf found at " + self . lbrycrd_conf )
else :
log . info ( " No lbrycrd.conf found at " + self . lbrycrd_conf + " . Generating now... " )
password = " " . join ( random . SystemRandom ( ) . choice ( string . ascii_letters + string . digits + " _ " ) for i in range ( 20 ) )
with open ( self . lbrycrd_conf , ' w ' ) as f :
f . write ( " rpcuser=rpcuser \n " )
f . write ( " rpcpassword= " + password )
log . info ( " Done writing lbrycrd.conf " )
2016-08-03 18:46:46 +02:00
def _responseFailed ( self , err , call ) :
2016-08-05 05:08:54 +02:00
log . error ( err . getTraceback ( ) )
2016-08-03 18:46:46 +02:00
call . cancel ( )
2016-03-24 03:27:48 +01:00
def render ( self , request ) :
request . content . seek ( 0 , 0 )
# Unmarshal the JSON-RPC data.
content = request . content . read ( )
parsed = jsonrpclib . loads ( content )
functionPath = parsed . get ( " method " )
args = parsed . get ( ' params ' )
2016-04-12 04:28:46 +02:00
#TODO convert args to correct types if possible
2016-03-24 03:27:48 +01:00
id = parsed . get ( ' id ' )
version = parsed . get ( ' jsonrpc ' )
if version :
version = int ( float ( version ) )
elif id and not version :
version = jsonrpclib . VERSION_1
else :
version = jsonrpclib . VERSION_PRE1
# XXX this all needs to be re-worked to support logic for multiple
# versions...
2016-03-29 22:42:47 +02:00
if not self . announced_startup :
2016-04-13 20:47:34 +02:00
if functionPath not in ALLOWED_DURING_STARTUP :
2016-04-02 04:11:11 +02:00
return server . failure
2016-03-29 22:42:47 +02:00
2016-05-14 23:57:34 +02:00
if self . wallet_type == " lbryum " and functionPath in [ ' set_miner ' , ' get_miner_status ' ] :
return server . failure
2016-03-24 03:27:48 +01:00
try :
function = self . _getFunction ( functionPath )
except jsonrpclib . Fault , f :
self . _cbRender ( f , request , id , version )
else :
2016-03-28 22:49:51 +02:00
request . setHeader ( " Access-Control-Allow-Origin " , " * " )
2016-03-24 03:27:48 +01:00
request . setHeader ( " content-type " , " text/json " )
2016-04-14 05:10:17 +02:00
if args == [ { } ] :
2016-03-24 03:27:48 +01:00
d = defer . maybeDeferred ( function )
else :
d = defer . maybeDeferred ( function , * args )
2016-08-03 23:49:13 +02:00
# cancel the response if the connection is broken
2016-08-03 18:46:46 +02:00
request . notifyFinish ( ) . addErrback ( self . _responseFailed , d )
2016-08-03 23:49:13 +02:00
2016-03-24 03:27:48 +01:00
d . addErrback ( self . _ebRender , id )
d . addCallback ( self . _cbRender , request , id , version )
return server . NOT_DONE_YET
2015-12-06 23:32:17 +01:00
2016-03-29 22:42:47 +02:00
def _cbRender ( self , result , request , id , version ) :
2016-07-29 00:12:54 +02:00
def default_decimal ( obj ) :
if isinstance ( obj , Decimal ) :
return float ( obj )
2016-03-29 22:42:47 +02:00
if isinstance ( result , Handler ) :
result = result . result
if isinstance ( result , dict ) :
result = result [ ' result ' ]
if version == jsonrpclib . VERSION_PRE1 :
if not isinstance ( result , jsonrpclib . Fault ) :
result = ( result , )
# Convert the result (python) to JSON-RPC
try :
2016-07-29 00:12:54 +02:00
s = jsonrpclib . dumps ( result , version = version , default = default_decimal )
2016-03-29 22:42:47 +02:00
except :
f = jsonrpclib . Fault ( self . FAILURE , " can ' t serialize output " )
s = jsonrpclib . dumps ( f , version = version )
2016-08-03 18:46:46 +02:00
2016-03-29 22:42:47 +02:00
request . setHeader ( " content-length " , str ( len ( s ) ) )
request . write ( s )
request . finish ( )
def _ebRender ( self , failure , id ) :
if isinstance ( failure . value , jsonrpclib . Fault ) :
return failure . value
log . error ( failure )
return jsonrpclib . Fault ( self . FAILURE , " error " )
2016-07-06 07:27:25 +02:00
def setup ( self , branch = DEFAULT_UI_BRANCH , user_specified = False , branch_specified = False , host_ui = True ) :
2016-04-07 09:12:09 +02:00
def _log_starting_vals ( ) :
2016-06-29 05:20:28 +02:00
log . info ( " Starting balance: " + str ( self . session . wallet . wallet_balance ) )
return defer . succeed ( None )
2016-03-29 22:42:47 +02:00
2016-04-07 09:12:09 +02:00
def _announce_startup ( ) :
2016-05-11 08:47:33 +02:00
def _wait_for_credits ( ) :
if float ( self . session . wallet . wallet_balance ) == 0.0 :
self . startup_status = STARTUP_STAGES [ 6 ]
return reactor . callLater ( 1 , _wait_for_credits )
else :
return _announce ( )
2016-04-18 01:54:04 +02:00
def _announce ( ) :
self . announced_startup = True
self . startup_status = STARTUP_STAGES [ 5 ]
2016-07-16 08:15:58 +02:00
log . info ( " Started lbrynet-daemon " )
2016-05-30 21:49:25 +02:00
if len ( self . startup_scripts ) :
log . info ( " Scheduling scripts " )
reactor . callLater ( 3 , self . _run_scripts )
2016-04-22 08:45:05 +02:00
# self.lbrynet_connection_checker.start(3600)
2016-04-18 01:54:04 +02:00
if self . first_run :
2016-05-21 01:59:49 +02:00
d = self . _upload_log ( log_type = " first_run " )
2016-05-31 23:38:00 +02:00
elif self . upload_log :
2016-05-21 01:59:49 +02:00
d = self . _upload_log ( exclude_previous = True , log_type = " start " )
2016-05-31 23:38:00 +02:00
else :
d = defer . succeed ( None )
2016-04-18 01:54:04 +02:00
2016-07-01 08:55:20 +02:00
# if float(self.session.wallet.wallet_balance) == 0.0:
# d.addCallback(lambda _: self._check_first_run())
# d.addCallback(self._show_first_run_result)
2016-04-18 01:54:04 +02:00
2016-07-04 10:57:30 +02:00
# d.addCallback(lambda _: _wait_for_credits() if self.requested_first_run_credits else _announce())
d . addCallback ( lambda _ : _announce ( ) )
2016-04-18 01:54:04 +02:00
return d
2016-01-11 21:52:38 +01:00
2016-07-16 08:15:58 +02:00
log . info ( " Starting lbrynet-daemon " )
2016-02-25 23:17:07 +01:00
2016-04-27 01:48:43 +02:00
self . internet_connection_checker . start ( 3600 )
self . version_checker . start ( 3600 * 12 )
2016-04-18 19:21:53 +02:00
self . connection_problem_checker . start ( 1 )
2016-07-28 11:30:13 +02:00
self . exchange_rate_manager . start ( )
2016-07-06 07:27:25 +02:00
if host_ui :
self . lbry_ui_manager . update_checker . start ( 1800 , now = False )
2016-04-18 19:21:53 +02:00
2015-12-06 23:32:17 +01:00
d = defer . Deferred ( )
2016-07-06 07:27:25 +02:00
if host_ui :
d . addCallback ( lambda _ : self . lbry_ui_manager . setup ( branch = branch ,
user_specified = user_specified ,
branch_specified = branch_specified ) )
2016-04-07 09:12:09 +02:00
d . addCallback ( lambda _ : self . _initial_setup ( ) )
2015-12-14 16:22:52 +01:00
d . addCallback ( lambda _ : threads . deferToThread ( self . _setup_data_directory ) )
d . addCallback ( lambda _ : self . _check_db_migration ( ) )
2015-12-06 23:32:17 +01:00
d . addCallback ( lambda _ : self . _get_settings ( ) )
d . addCallback ( lambda _ : self . _get_session ( ) )
d . addCallback ( lambda _ : add_lbry_file_to_sd_identifier ( self . sd_identifier ) )
d . addCallback ( lambda _ : self . _setup_stream_identifier ( ) )
d . addCallback ( lambda _ : self . _setup_lbry_file_manager ( ) )
d . addCallback ( lambda _ : self . _setup_lbry_file_opener ( ) )
2016-01-26 02:28:05 +01:00
d . addCallback ( lambda _ : self . _setup_query_handlers ( ) )
d . addCallback ( lambda _ : self . _setup_server ( ) )
2016-04-07 09:12:09 +02:00
d . addCallback ( lambda _ : _log_starting_vals ( ) )
2016-03-29 22:42:47 +02:00
d . addCallback ( lambda _ : _announce_startup ( ) )
2015-12-06 23:32:17 +01:00
d . callback ( None )
return defer . succeed ( None )
2016-05-30 21:49:25 +02:00
def _get_platform ( self ) :
r = {
" processor " : platform . processor ( ) ,
" python_version: " : platform . python_version ( ) ,
" platform " : platform . platform ( ) ,
" os_release " : platform . release ( ) ,
" os_system " : platform . system ( ) ,
" lbrynet_version: " : lbrynet_version ,
" lbryum_version: " : lbryum_version ,
" ui_version " : self . lbry_ui_manager . loaded_git_version ,
}
if not self . ip :
try :
r [ ' ip ' ] = json . load ( urlopen ( ' http://jsonip.com ' ) ) [ ' ip ' ]
self . ip = r [ ' ip ' ]
except :
r [ ' ip ' ] = " Could not determine "
return r
2016-03-24 03:27:48 +01:00
def _initial_setup ( self ) :
2016-04-07 09:12:09 +02:00
def _log_platform ( ) :
2016-05-30 21:49:25 +02:00
log . info ( " Platform: " + json . dumps ( self . _get_platform ( ) ) )
2016-04-07 09:12:09 +02:00
return defer . succeed ( None )
d = _log_platform ( )
return d
2016-04-18 19:21:53 +02:00
def _check_network_connection ( self ) :
try :
host = socket . gethostbyname ( REMOTE_SERVER )
s = socket . create_connection ( ( host , 80 ) , 2 )
self . connected_to_internet = True
except :
2016-07-16 08:15:58 +02:00
log . info ( " Internet connection not working " )
2016-04-18 19:21:53 +02:00
self . connected_to_internet = False
2016-01-26 02:28:05 +01:00
2016-04-22 08:45:05 +02:00
def _check_lbrynet_connection ( self ) :
def _log_success ( ) :
2016-07-16 08:15:58 +02:00
log . info ( " lbrynet connectivity test passed " )
2016-04-22 08:45:05 +02:00
def _log_failure ( ) :
2016-07-16 08:15:58 +02:00
log . info ( " lbrynet connectivity test failed " )
2016-04-22 08:45:05 +02:00
wonderfullife_sh = " 6f3af0fa3924be98a54766aa2715d22c6c1509c3f7fa32566df4899a41f3530a9f97b2ecb817fa1dcbf1b30553aefaa7 "
d = download_sd_blob ( self . session , wonderfullife_sh , self . session . base_payment_rate_manager )
d . addCallbacks ( lambda _ : _log_success , lambda _ : _log_failure )
2016-04-18 19:21:53 +02:00
def _check_remote_versions ( self ) :
def _get_lbryum_version ( ) :
try :
2016-07-30 04:36:30 +02:00
r = urlopen ( " https://raw.githubusercontent.com/lbryio/lbryum/master/lib/version.py " ) . read ( ) . split ( ' \n ' )
version = next ( line . split ( " = " ) [ 1 ] . split ( " # " ) [ 0 ] . replace ( " " , " " )
for line in r if " LBRYUM_VERSION " in line )
version = version . replace ( " ' " , " " )
log . info (
" remote lbryum %s > local lbryum %s = %s " ,
version , lbryum_version ,
utils . version_is_greater_than ( version , lbryum_version )
)
self . git_lbryum_version = version
2016-04-18 19:21:53 +02:00
return defer . succeed ( None )
except :
2016-07-16 08:15:58 +02:00
log . info ( " Failed to get lbryum version from git " )
2016-07-30 04:36:30 +02:00
self . git_lbryum_version = None
2016-04-18 19:21:53 +02:00
return defer . fail ( None )
def _get_lbrynet_version ( ) :
try :
2016-07-30 04:36:30 +02:00
version = get_lbrynet_version_from_github ( )
log . info (
" remote lbrynet %s > local lbrynet %s = %s " ,
version , lbrynet_version ,
utils . version_is_greater_than ( version , lbrynet_version )
)
self . git_lbrynet_version = version
2016-04-18 19:21:53 +02:00
return defer . succeed ( None )
except :
2016-07-16 08:15:58 +02:00
log . info ( " Failed to get lbrynet version from git " )
2016-07-30 04:36:30 +02:00
self . git_lbrynet_version = None
2016-04-18 19:21:53 +02:00
return defer . fail ( None )
d = _get_lbrynet_version ( )
d . addCallback ( lambda _ : _get_lbryum_version ( ) )
def _check_connection_problems ( self ) :
2016-07-30 04:36:30 +02:00
if not self . git_lbrynet_version or not self . git_lbryum_version :
2016-04-18 19:21:53 +02:00
self . connection_problem = CONNECTION_PROBLEM_CODES [ 0 ]
elif self . startup_status [ 0 ] == ' loading_wallet ' :
if self . session . wallet . is_lagging :
self . connection_problem = CONNECTION_PROBLEM_CODES [ 2 ]
else :
self . connection_problem = None
if not self . connected_to_internet :
self . connection_problem = CONNECTION_PROBLEM_CODES [ 1 ]
2016-07-27 09:00:44 +02:00
def _add_to_pending_claims ( self , name , txid ) :
log . info ( " Adding lbry:// %s to pending claims, txid %s " % ( name , txid ) )
self . pending_claims [ name ] = txid
return txid
def _check_pending_claims ( self ) :
2016-07-28 18:43:47 +02:00
# TODO: this was blatantly copied from jsonrpc_start_lbry_file. Be DRY.
def _start_file ( f ) :
d = self . lbry_file_manager . toggle_lbry_file_running ( f )
2016-08-03 00:34:48 +02:00
d . addCallback ( lambda _ : self . lighthouse_client . announce_sd ( f . sd_hash ) )
2016-07-28 18:43:47 +02:00
return defer . succeed ( " Started LBRY file " )
def _get_and_start_file ( name ) :
2016-07-27 09:00:44 +02:00
d = defer . succeed ( self . pending_claims . pop ( name ) )
d . addCallback ( lambda _ : self . _get_lbry_file ( " name " , name , return_json = False ) )
d . addCallback ( lambda l : _start_file ( l ) if l . stopped else " LBRY file was already running " )
2016-07-28 19:16:33 +02:00
def re_add_to_pending_claims ( name ) :
txid = self . pending_claims . pop ( name )
self . _add_to_pending_claims ( name , txid )
2016-07-27 20:12:20 +02:00
def _process_lbry_file ( name , lbry_file ) :
2016-07-27 20:59:26 +02:00
# lbry_file is an instance of ManagedLBRYFileDownloader or None
2016-07-28 18:43:47 +02:00
# TODO: check for sd_hash in addition to txid
2016-07-27 20:12:20 +02:00
ready_to_start = (
2016-07-27 20:59:26 +02:00
lbry_file and
2016-07-28 18:43:47 +02:00
self . pending_claims [ name ] == lbry_file . txid
2016-07-27 20:12:20 +02:00
)
if ready_to_start :
2016-07-28 19:16:33 +02:00
_get_and_start_file ( name )
2016-07-27 20:12:20 +02:00
else :
re_add_to_pending_claims ( name )
2016-07-27 09:00:44 +02:00
for name in self . pending_claims :
log . info ( " Checking if new claim for lbry:// %s is confirmed " % name )
d = self . _resolve_name ( name , force_refresh = True )
d . addCallback ( lambda _ : self . _get_lbry_file_by_uri ( name ) )
2016-07-27 20:12:20 +02:00
d . addCallbacks (
lambda lbry_file : _process_lbry_file ( name , lbry_file ) ,
lambda _ : re_add_to_pending_claims ( name )
)
2016-07-27 09:00:44 +02:00
2016-04-18 19:21:53 +02:00
def _start_server ( self ) :
2016-01-26 02:28:05 +01:00
if self . peer_port is not None :
server_factory = ServerProtocolFactory ( self . session . rate_limiter ,
self . query_handlers ,
self . session . peer_manager )
try :
self . lbry_server_port = reactor . listenTCP ( self . peer_port , server_factory )
except error . CannotListenError as e :
import traceback
log . error ( " Couldn ' t bind to port %d . %s " , self . peer_port , traceback . format_exc ( ) )
raise ValueError ( " %s lbrynet may already be running on your computer. " , str ( e ) )
return defer . succeed ( True )
def _stop_server ( self ) :
2016-05-30 21:49:25 +02:00
try :
if self . lbry_server_port is not None :
self . lbry_server_port , p = None , self . lbry_server_port
return defer . maybeDeferred ( p . stopListening )
else :
return defer . succeed ( True )
except AttributeError :
2016-01-26 02:28:05 +01:00
return defer . succeed ( True )
def _setup_server ( self ) :
def restore_running_status ( running ) :
if running is True :
return self . _start_server ( )
return defer . succeed ( True )
2016-04-12 04:28:46 +02:00
self . startup_status = STARTUP_STAGES [ 4 ]
2016-04-09 07:33:27 +02:00
2016-01-26 02:28:05 +01:00
dl = self . settings . get_server_running_status ( )
dl . addCallback ( restore_running_status )
return dl
def _setup_query_handlers ( self ) :
handlers = [
2016-02-16 19:39:08 +01:00
# CryptBlobInfoQueryHandlerFactory(self.lbry_file_metadata_manager, self.session.wallet,
2016-01-26 02:28:05 +01:00
# self._server_payment_rate_manager),
BlobAvailabilityHandlerFactory ( self . session . blob_manager ) ,
2016-02-16 19:39:08 +01:00
# BlobRequestHandlerFactory(self.session.blob_manager, self.session.wallet,
2016-01-26 02:28:05 +01:00
# self._server_payment_rate_manager),
self . session . wallet . get_wallet_info_query_handler_factory ( ) ,
]
def get_blob_request_handler_factory ( rate ) :
self . blob_request_payment_rate_manager = PaymentRateManager (
self . session . base_payment_rate_manager , rate
)
handlers . append ( BlobRequestHandlerFactory ( self . session . blob_manager , self . session . wallet ,
self . blob_request_payment_rate_manager ) )
d1 = self . settings . get_server_data_payment_rate ( )
d1 . addCallback ( get_blob_request_handler_factory )
dl = defer . DeferredList ( [ d1 ] )
dl . addCallback ( lambda _ : self . _add_query_handlers ( handlers ) )
return dl
def _add_query_handlers ( self , query_handlers ) :
def _set_query_handlers ( statuses ) :
from future_builtins import zip
for handler , ( success , status ) in zip ( query_handlers , statuses ) :
if success is True :
self . query_handlers [ handler ] = status
ds = [ ]
for handler in query_handlers :
ds . append ( self . settings . get_query_handler_status ( handler . get_primary_query_identifier ( ) ) )
dl = defer . DeferredList ( ds )
dl . addCallback ( _set_query_handlers )
return dl
2016-05-21 01:59:49 +02:00
def _upload_log ( self , log_type = None , exclude_previous = False , force = False ) :
2016-04-22 08:45:05 +02:00
if self . upload_log or force :
2016-06-28 20:28:59 +02:00
for lm , lp in [ ( ' lbrynet ' , lbrynet_log ) ] : #, ('lbryum', lbryum_log)]:
2016-06-04 23:05:14 +02:00
if os . path . isfile ( lp ) :
if exclude_previous :
f = open ( lp , " r " )
2016-06-28 20:28:59 +02:00
f . seek ( PREVIOUS_LBRYNET_LOG ) # if lm == 'lbrynet' else PREVIOUS_LBRYUM_LOG)
2016-06-04 23:05:14 +02:00
log_contents = f . read ( )
f . close ( )
else :
f = open ( lp , " r " )
log_contents = f . read ( )
f . close ( )
params = {
' date ' : datetime . utcnow ( ) . strftime ( ' % Y % m %d - % H % M % S ' ) ,
' hash ' : base58 . b58encode ( self . lbryid ) [ : 20 ] ,
' sys ' : platform . system ( ) ,
' type ' : " %s - %s " % ( lm , log_type ) if log_type else lm ,
' log ' : log_contents
}
requests . post ( LOG_POST_URL , params )
2016-04-20 20:44:26 +02:00
return defer . succeed ( None )
else :
return defer . succeed ( None )
2016-04-07 09:12:09 +02:00
2016-07-20 08:36:55 +02:00
def _clean_up_temp_files ( self ) :
for path in self . uploaded_temp_files :
try :
os . remove ( path )
except OSError :
pass
2016-01-11 21:52:38 +01:00
def _shutdown ( self ) :
2016-03-24 03:27:48 +01:00
log . info ( " Closing lbrynet session " )
2016-04-18 01:54:04 +02:00
log . info ( " Status at time of shutdown: " + self . startup_status [ 0 ] )
2016-05-30 21:49:25 +02:00
if self . internet_connection_checker . running :
self . internet_connection_checker . stop ( )
if self . version_checker . running :
self . version_checker . stop ( )
if self . connection_problem_checker . running :
self . connection_problem_checker . stop ( )
2016-07-06 07:27:25 +02:00
if self . lbry_ui_manager . update_checker . running :
self . lbry_ui_manager . update_checker . stop ( )
2016-07-27 09:00:44 +02:00
if self . pending_claim_checker . running :
self . pending_claim_checker . stop ( )
2016-04-18 01:54:04 +02:00
2016-07-20 08:36:55 +02:00
self . _clean_up_temp_files ( )
2016-05-21 01:59:49 +02:00
d = self . _upload_log ( log_type = " close " , exclude_previous = False if self . first_run else True )
2016-04-07 09:12:09 +02:00
d . addCallback ( lambda _ : self . _stop_server ( ) )
2016-05-30 21:49:25 +02:00
d . addErrback ( lambda err : True )
2016-05-11 02:47:35 +02:00
d . addCallback ( lambda _ : self . lbry_file_manager . stop ( ) )
2016-05-30 21:49:25 +02:00
d . addErrback ( lambda err : True )
2016-01-11 21:52:38 +01:00
if self . session is not None :
2016-01-26 02:28:05 +01:00
d . addCallback ( lambda _ : self . session . shut_down ( ) )
2016-05-30 21:49:25 +02:00
d . addErrback ( lambda err : True )
2016-01-11 21:52:38 +01:00
return d
2016-03-24 03:27:48 +01:00
def _update_settings ( self , settings ) :
2016-04-07 09:12:09 +02:00
for k in settings . keys ( ) :
if k == ' run_on_startup ' :
if type ( settings [ ' run_on_startup ' ] ) is bool :
self . session_settings [ ' run_on_startup ' ] = settings [ ' run_on_startup ' ]
else :
return defer . fail ( )
elif k == ' data_rate ' :
if type ( settings [ ' data_rate ' ] ) is float :
self . session_settings [ ' data_rate ' ] = settings [ ' data_rate ' ]
2016-04-09 05:38:57 +02:00
elif type ( settings [ ' data_rate ' ] ) is int :
self . session_settings [ ' data_rate ' ] = float ( settings [ ' data_rate ' ] )
2016-04-07 09:12:09 +02:00
else :
return defer . fail ( )
elif k == ' max_key_fee ' :
if type ( settings [ ' max_key_fee ' ] ) is float :
self . session_settings [ ' max_key_fee ' ] = settings [ ' max_key_fee ' ]
2016-04-09 05:38:57 +02:00
elif type ( settings [ ' max_key_fee ' ] ) is int :
self . session_settings [ ' max_key_fee ' ] = float ( settings [ ' max_key_fee ' ] )
2016-04-07 09:12:09 +02:00
else :
return defer . fail ( )
2016-04-22 08:45:05 +02:00
elif k == ' download_directory ' :
if type ( settings [ ' download_directory ' ] ) is unicode :
if os . path . isdir ( settings [ ' download_directory ' ] ) :
self . session_settings [ ' download_directory ' ] = settings [ ' download_directory ' ]
2016-04-07 09:12:09 +02:00
else :
pass
else :
return defer . fail ( )
elif k == ' max_upload ' :
if type ( settings [ ' max_upload ' ] ) is float :
self . session_settings [ ' max_upload ' ] = settings [ ' max_upload ' ]
2016-04-09 05:38:57 +02:00
elif type ( settings [ ' max_upload ' ] ) is int :
self . session_settings [ ' max_upload ' ] = float ( settings [ ' max_upload ' ] )
2016-04-07 09:12:09 +02:00
else :
return defer . fail ( )
elif k == ' max_download ' :
if type ( settings [ ' max_download ' ] ) is float :
self . session_settings [ ' max_download ' ] = settings [ ' max_download ' ]
2016-04-09 05:38:57 +02:00
if type ( settings [ ' max_download ' ] ) is int :
self . session_settings [ ' max_download ' ] = float ( settings [ ' max_download ' ] )
2016-04-07 09:12:09 +02:00
else :
return defer . fail ( )
elif k == ' upload_log ' :
if type ( settings [ ' upload_log ' ] ) is bool :
self . session_settings [ ' upload_log ' ] = settings [ ' upload_log ' ]
else :
return defer . fail ( )
2016-04-23 00:18:17 +02:00
elif k == ' download_timeout ' :
if type ( settings [ ' download_timeout ' ] ) is int :
self . session_settings [ ' download_timeout ' ] = settings [ ' download_timeout ' ]
2016-05-31 23:38:00 +02:00
elif type ( settings [ ' download_timeout ' ] ) is float :
self . session_settings [ ' download_timeout ' ] = int ( settings [ ' download_timeout ' ] )
2016-04-23 00:18:17 +02:00
else :
return defer . fail ( )
2016-04-26 04:35:21 +02:00
elif k == ' search_timeout ' :
if type ( settings [ ' search_timeout ' ] ) is float :
self . session_settings [ ' search_timeout ' ] = settings [ ' search_timeout ' ]
2016-05-31 23:38:00 +02:00
elif type ( settings [ ' search_timeout ' ] ) is int :
self . session_settings [ ' search_timeout ' ] = float ( settings [ ' search_timeout ' ] )
2016-04-26 04:35:21 +02:00
else :
return defer . fail ( )
elif k == ' cache_time ' :
if type ( settings [ ' cache_time ' ] ) is int :
self . session_settings [ ' cache_time ' ] = settings [ ' cache_time ' ]
2016-05-31 23:38:00 +02:00
elif type ( settings [ ' cache_time ' ] ) is float :
self . session_settings [ ' cache_time ' ] = int ( settings [ ' cache_time ' ] )
2016-04-26 04:35:21 +02:00
else :
return defer . fail ( )
2016-04-22 08:45:05 +02:00
self . run_on_startup = self . session_settings [ ' run_on_startup ' ]
self . data_rate = self . session_settings [ ' data_rate ' ]
self . max_key_fee = self . session_settings [ ' max_key_fee ' ]
self . download_directory = self . session_settings [ ' download_directory ' ]
self . max_upload = self . session_settings [ ' max_upload ' ]
self . max_download = self . session_settings [ ' max_download ' ]
self . upload_log = self . session_settings [ ' upload_log ' ]
2016-04-23 00:18:17 +02:00
self . download_timeout = self . session_settings [ ' download_timeout ' ]
2016-04-26 04:35:21 +02:00
self . search_timeout = self . session_settings [ ' search_timeout ' ]
self . cache_time = self . session_settings [ ' cache_time ' ]
2016-04-22 08:45:05 +02:00
2016-04-07 09:12:09 +02:00
f = open ( self . daemon_conf , " w " )
f . write ( json . dumps ( self . session_settings ) )
f . close ( )
2016-03-24 03:27:48 +01:00
return defer . succeed ( True )
2016-01-07 16:24:55 +01:00
2015-12-14 16:22:52 +01:00
def _setup_data_directory ( self ) :
2016-04-09 08:09:30 +02:00
self . startup_status = STARTUP_STAGES [ 1 ]
2016-03-24 03:27:48 +01:00
log . info ( " Loading databases... " )
2015-12-14 16:22:52 +01:00
if self . created_data_dir :
db_revision = open ( os . path . join ( self . db_dir , " db_revision " ) , mode = ' w ' )
db_revision . write ( str ( self . current_db_revision ) )
db_revision . close ( )
log . debug ( " Created the db revision file: %s " , str ( os . path . join ( self . db_dir , " db_revision " ) ) )
if not os . path . exists ( self . blobfile_dir ) :
os . mkdir ( self . blobfile_dir )
log . debug ( " Created the blobfile directory: %s " , str ( self . blobfile_dir ) )
def _check_db_migration ( self ) :
old_revision = 0
db_revision_file = os . path . join ( self . db_dir , " db_revision " )
if os . path . exists ( db_revision_file ) :
old_revision = int ( open ( db_revision_file ) . read ( ) . strip ( ) )
if old_revision < self . current_db_revision :
from lbrynet . db_migrator import dbmigrator
2016-03-24 03:27:48 +01:00
log . info ( " Upgrading your databases... " )
2015-12-14 16:22:52 +01:00
d = threads . deferToThread ( dbmigrator . migrate_db , self . db_dir , old_revision , self . current_db_revision )
def print_success ( old_dirs ) :
success_string = " Finished upgrading the databases. It is now safe to delete the "
success_string + = " following directories, if you feel like it. It won ' t make any "
success_string + = " difference. \n Anyway here they are: "
for i , old_dir in enumerate ( old_dirs ) :
success_string + = old_dir
if i + 1 < len ( old_dir ) :
success_string + = " , "
2016-03-24 03:27:48 +01:00
log . info ( success_string )
2015-12-14 16:22:52 +01:00
d . addCallback ( print_success )
return d
return defer . succeed ( True )
2015-12-06 23:32:17 +01:00
def _get_settings ( self ) :
d = self . settings . start ( )
d . addCallback ( lambda _ : self . settings . get_lbryid ( ) )
2016-01-26 02:28:05 +01:00
d . addCallback ( self . _set_lbryid )
2015-12-06 23:32:17 +01:00
return d
2016-01-26 02:28:05 +01:00
def _set_lbryid ( self , lbryid ) :
2015-12-06 23:32:17 +01:00
if lbryid is None :
return self . _make_lbryid ( )
else :
2016-04-07 09:12:09 +02:00
log . info ( " LBRY ID: " + base58 . b58encode ( lbryid ) )
2015-12-06 23:32:17 +01:00
self . lbryid = lbryid
def _make_lbryid ( self ) :
self . lbryid = generate_id ( )
2016-04-07 09:12:09 +02:00
log . info ( " Generated new LBRY ID: " + base58 . b58encode ( self . lbryid ) )
2015-12-06 23:32:17 +01:00
d = self . settings . save_lbryid ( self . lbryid )
return d
def _setup_lbry_file_manager ( self ) :
2016-04-12 04:28:46 +02:00
self . startup_status = STARTUP_STAGES [ 3 ]
2015-12-06 23:32:17 +01:00
self . lbry_file_metadata_manager = DBLBRYFileMetadataManager ( self . db_dir )
d = self . lbry_file_metadata_manager . setup ( )
def set_lbry_file_manager ( ) :
2016-03-24 03:27:48 +01:00
self . lbry_file_manager = LBRYFileManager ( self . session ,
self . lbry_file_metadata_manager ,
self . sd_identifier ,
2016-04-24 23:51:24 +02:00
download_directory = self . download_directory )
2015-12-06 23:32:17 +01:00
return self . lbry_file_manager . setup ( )
d . addCallback ( lambda _ : set_lbry_file_manager ( ) )
return d
def _get_session ( self ) :
def get_default_data_rate ( ) :
d = self . settings . get_default_data_payment_rate ( )
2016-02-16 19:39:08 +01:00
d . addCallback ( lambda rate : { " default_data_payment_rate " : rate if rate is not None else
2016-03-24 03:27:48 +01:00
MIN_BLOB_DATA_PAYMENT_RATE } )
2015-12-06 23:32:17 +01:00
return d
def get_wallet ( ) :
2016-07-17 20:43:35 +02:00
if self . wallet_type == " lbrycrd " :
2016-02-25 23:17:07 +01:00
log . info ( " Using lbrycrd wallet " )
2016-01-07 16:24:55 +01:00
d = defer . succeed ( LBRYcrdWallet ( self . db_dir , wallet_dir = self . wallet_dir , wallet_conf = self . lbrycrd_conf ,
2016-07-06 09:37:31 +02:00
lbrycrdd_path = self . lbrycrdd_path ) )
2016-07-09 19:31:07 +02:00
elif self . wallet_type == " lbryum " :
log . info ( " Using lbryum wallet " )
d = defer . succeed ( LBRYumWallet ( self . db_dir ) )
2016-02-23 04:32:07 +01:00
elif self . wallet_type == " ptc " :
2016-02-25 23:17:07 +01:00
log . info ( " Using PTC wallet " )
2015-12-06 23:32:17 +01:00
d = defer . succeed ( PTCWallet ( self . db_dir ) )
2016-02-23 04:32:07 +01:00
else :
2016-07-17 20:43:35 +02:00
# TODO: should fail here. Can't switch to lbrycrd because the wallet_dir, conf and path won't be set
log . info ( " Requested unknown wallet ' %s ' , using default lbryum " , self . wallet_type )
2016-05-30 21:49:25 +02:00
d = defer . succeed ( LBRYumWallet ( self . db_dir ) )
2016-02-23 04:32:07 +01:00
2015-12-06 23:32:17 +01:00
d . addCallback ( lambda wallet : { " wallet " : wallet } )
return d
d1 = get_default_data_rate ( )
d2 = get_wallet ( )
def combine_results ( results ) :
r = { }
for success , result in results :
if success is True :
r . update ( result )
return r
def create_session ( results ) :
self . session = LBRYSession ( results [ ' default_data_payment_rate ' ] , db_dir = self . db_dir , lbryid = self . lbryid ,
blob_dir = self . blobfile_dir , dht_node_port = self . dht_node_port ,
known_dht_nodes = self . known_dht_nodes , peer_port = self . peer_port ,
use_upnp = self . use_upnp , wallet = results [ ' wallet ' ] )
2016-04-12 04:28:46 +02:00
self . startup_status = STARTUP_STAGES [ 2 ]
2015-12-06 23:32:17 +01:00
dl = defer . DeferredList ( [ d1 , d2 ] , fireOnOneErrback = True )
dl . addCallback ( combine_results )
dl . addCallback ( create_session )
dl . addCallback ( lambda _ : self . session . setup ( ) )
2016-04-18 01:54:04 +02:00
2015-12-06 23:32:17 +01:00
return dl
2016-07-09 19:31:07 +02:00
# def _check_first_run(self):
# def _set_first_run_false():
# log.info("Not first run")
# self.first_run = False
# self.session_settings['requested_first_run_credits'] = True
# f = open(self.daemon_conf, "w")
# f.write(json.dumps(self.session_settings))
# f.close()
# return 0.0
#
# if self.wallet_type == 'lbryum':
# d = self.session.wallet.is_first_run()
# d.addCallback(lambda is_first_run: self._do_first_run() if is_first_run or not self.requested_first_run_credits
# else _set_first_run_false())
# else:
# d = defer.succeed(None)
# d.addCallback(lambda _: _set_first_run_false())
# return d
#
# def _do_first_run(self):
# def send_request(url, data):
# log.info("Requesting first run credits")
# r = requests.post(url, json=data)
# if r.status_code == 200:
# self.requested_first_run_credits = True
# self.session_settings['requested_first_run_credits'] = True
# f = open(self.daemon_conf, "w")
# f.write(json.dumps(self.session_settings))
# f.close()
# return r.json()['credits_sent']
# return 0.0
#
# def log_error(err):
# log.warning("unable to request free credits. %s", err.getErrorMessage())
# return 0.0
#
# def request_credits(address):
# url = "http://credreq.lbry.io/requestcredits"
# data = {"address": address}
# d = threads.deferToThread(send_request, url, data)
# d.addErrback(log_error)
# return d
#
# self.first_run = True
# d = self.session.wallet.get_new_address()
# d.addCallback(request_credits)
#
# return d
#
# def _show_first_run_result(self, credits_received):
# if credits_received != 0.0:
# points_string = locale.format_string("%.2f LBC", (round(credits_received, 2),), grouping=True)
# self.startup_message = "Thank you for testing the alpha version of LBRY! You have been given %s for free because we love you. Please hang on for a few minutes for the next block to be mined. When you refresh this page and see your credits you're ready to go!." % points_string
# else:
# self.startup_message = None
2016-03-03 00:32:48 +01:00
2015-12-06 23:32:17 +01:00
def _setup_stream_identifier ( self ) :
file_saver_factory = LBRYFileSaverFactory ( self . session . peer_finder , self . session . rate_limiter ,
self . session . blob_manager , self . stream_info_manager ,
2016-04-22 08:45:05 +02:00
self . session . wallet , self . download_directory )
2015-12-06 23:32:17 +01:00
self . sd_identifier . add_stream_downloader_factory ( LBRYFileStreamType , file_saver_factory )
file_opener_factory = LBRYFileOpenerFactory ( self . session . peer_finder , self . session . rate_limiter ,
self . session . blob_manager , self . stream_info_manager ,
self . session . wallet )
self . sd_identifier . add_stream_downloader_factory ( LBRYFileStreamType , file_opener_factory )
return defer . succeed ( None )
def _setup_lbry_file_opener ( self ) :
downloader_factory = LBRYFileOpenerFactory ( self . session . peer_finder , self . session . rate_limiter ,
self . session . blob_manager , self . stream_info_manager ,
self . session . wallet )
self . sd_identifier . add_stream_downloader_factory ( LBRYFileStreamType , downloader_factory )
return defer . succeed ( True )
2016-08-03 23:57:26 +02:00
def _download_sd_blob ( self , sd_hash , timeout = DEFAULT_SD_DOWNLOAD_TIMEOUT ) :
2016-08-02 08:54:04 +02:00
def cb ( result ) :
2016-08-02 22:03:47 +02:00
if not r . called :
r . callback ( result )
2016-08-02 08:54:04 +02:00
def eb ( ) :
if not r . called :
r . errback ( Exception ( " sd timeout " ) )
r = defer . Deferred ( None )
2016-08-03 23:57:26 +02:00
reactor . callLater ( timeout , eb )
2016-08-02 05:15:32 +02:00
d = download_sd_blob ( self . session , sd_hash , PaymentRateManager ( self . session . base_payment_rate_manager ) )
d . addCallback ( BlobStreamDescriptorReader )
d . addCallback ( lambda blob : blob . get_info ( ) )
2016-08-02 08:54:04 +02:00
d . addCallback ( cb )
return r
2016-08-02 05:15:32 +02:00
2016-05-14 01:35:38 +02:00
def _download_name ( self , name , timeout = DEFAULT_TIMEOUT , download_directory = None ,
2016-07-21 17:44:59 +02:00
file_name = None , stream_info = None , wait_for_write = True ) :
2016-04-24 23:51:24 +02:00
"""
2016-05-04 11:20:38 +02:00
Add a lbry file to the file manager , start the download , and return the new lbry file .
If it already exists in the file manager , return the existing lbry file
2016-04-24 23:51:24 +02:00
"""
2016-07-21 17:44:59 +02:00
helper = _DownloadNameHelper (
self , name , timeout , download_directory , file_name , wait_for_write )
2016-01-12 00:01:46 +01:00
2016-05-05 03:25:46 +02:00
if not stream_info :
self . waiting_on [ name ] = True
d = self . _resolve_name ( name )
else :
d = defer . succeed ( stream_info )
2016-07-21 17:44:59 +02:00
d . addCallback ( helper . _setup_stream )
d . addCallback ( helper . wait_or_get_stream )
2016-05-05 03:25:46 +02:00
if not stream_info :
2016-07-21 17:44:59 +02:00
d . addCallback ( helper . _remove_from_wait )
return d
def add_stream ( self , name , timeout , download_directory , file_name , stream_info ) :
""" Makes, adds and starts a stream """
self . streams [ name ] = GetStream ( self . sd_identifier ,
self . session ,
self . session . wallet ,
self . lbry_file_manager ,
2016-08-05 12:00:35 +02:00
self . exchange_rate_manager ,
2016-07-21 17:44:59 +02:00
max_key_fee = self . max_key_fee ,
data_rate = self . data_rate ,
timeout = timeout ,
download_directory = download_directory ,
file_name = file_name )
d = self . streams [ name ] . start ( stream_info , name )
2015-12-15 03:49:01 +01:00
return d
2016-04-26 04:35:21 +02:00
def _get_long_count_timestamp ( self ) :
return int ( ( datetime . utcnow ( ) - ( datetime ( year = 2012 , month = 12 , day = 21 ) ) ) . total_seconds ( ) )
2016-01-20 03:07:16 +01:00
2016-04-26 04:35:21 +02:00
def _update_claim_cache ( self ) :
f = open ( os . path . join ( self . db_dir , " stream_info_cache.json " ) , " w " )
f . write ( json . dumps ( self . name_cache ) )
f . close ( )
return defer . succeed ( True )
2015-12-15 03:49:01 +01:00
2016-05-30 21:49:25 +02:00
def _resolve_name ( self , name , force_refresh = False ) :
2016-07-21 17:44:59 +02:00
""" Resolves a name. Checks the cache first before going out to the blockchain.
2016-01-20 03:07:16 +01:00
2016-07-21 17:44:59 +02:00
Args :
name : the lbry : / / < name > to resolve
force_refresh : if True , always go out to the blockchain to resolve .
"""
if name . startswith ( ' lbry:// ' ) :
raise ValueError ( ' name %s should not start with lbry:// ' )
helper = _ResolveNameHelper ( self , name , force_refresh )
return helper . get_deferred ( )
2016-01-12 00:01:46 +01:00
2016-05-30 21:49:25 +02:00
def _delete_lbry_file ( self , lbry_file , delete_file = True ) :
2016-01-20 03:07:16 +01:00
d = self . lbry_file_manager . delete_lbry_file ( lbry_file )
def finish_deletion ( lbry_file ) :
d = lbry_file . delete_data ( )
d . addCallback ( lambda _ : _delete_stream_data ( lbry_file ) )
return d
2016-01-12 18:42:32 +01:00
2016-01-20 03:07:16 +01:00
def _delete_stream_data ( lbry_file ) :
s_h = lbry_file . stream_hash
d = self . lbry_file_manager . get_count_for_stream_hash ( s_h )
# TODO: could possibly be a timing issue here
d . addCallback ( lambda c : self . stream_info_manager . delete_stream ( s_h ) if c == 0 else True )
2016-05-30 21:49:25 +02:00
if delete_file :
d . addCallback ( lambda _ : os . remove ( os . path . join ( self . download_directory , lbry_file . file_name ) ) if
2016-04-22 08:45:05 +02:00
os . path . isfile ( os . path . join ( self . download_directory , lbry_file . file_name ) ) else defer . succeed ( None ) )
2016-01-20 03:07:16 +01:00
return d
d . addCallback ( lambda _ : finish_deletion ( lbry_file ) )
2016-07-16 08:15:58 +02:00
d . addCallback ( lambda _ : log . info ( " Delete lbry file " ) )
2016-01-20 03:07:16 +01:00
return d
2016-01-24 06:10:22 +01:00
def _get_est_cost ( self , name ) :
def _check_est ( d , name ) :
2016-08-08 07:43:28 +02:00
try :
if d . result :
log . info ( " Cost est for lbry:// " + name + " : " + str ( d . result ) + " LBC " )
return defer . succeed ( None )
except AttributeError :
pass
log . info ( " Timeout estimating cost for lbry:// " + name + " , using key fee " )
d . cancel ( )
2016-01-24 06:10:22 +01:00
return defer . succeed ( None )
def _add_key_fee ( data_cost ) :
2016-04-26 04:35:21 +02:00
d = self . _resolve_name ( name )
2016-07-28 22:12:20 +02:00
d . addCallback ( lambda info : self . exchange_rate_manager . to_lbc ( info . get ( ' fee ' , None ) ) )
d . addCallback ( lambda fee : data_cost if fee is None else data_cost + fee . amount )
2016-01-24 06:10:22 +01:00
return d
2016-04-26 04:35:21 +02:00
d = self . _resolve_name ( name )
2016-06-27 23:07:59 +02:00
d . addCallback ( lambda info : info [ ' sources ' ] [ ' lbry_sd_hash ' ] )
2016-04-02 04:11:11 +02:00
d . addCallback ( lambda sd_hash : download_sd_blob ( self . session , sd_hash ,
2016-01-26 05:59:57 +01:00
self . blob_request_payment_rate_manager ) )
2016-01-24 06:10:22 +01:00
d . addCallback ( self . sd_identifier . get_metadata_for_sd_blob )
d . addCallback ( lambda metadata : metadata . validator . info_to_show ( ) )
2016-02-19 02:41:23 +01:00
d . addCallback ( lambda info : int ( dict ( info ) [ ' stream_size ' ] ) / 1000000 * self . data_rate )
2016-07-28 22:12:20 +02:00
d . addCallbacks ( _add_key_fee , lambda _ : _add_key_fee ( 0.0 ) )
2016-02-15 16:12:49 +01:00
reactor . callLater ( self . search_timeout , _check_est , d , name )
2016-01-24 06:10:22 +01:00
return d
2016-04-24 10:42:42 +02:00
def _get_lbry_file_by_uri ( self , name ) :
2016-04-24 23:51:24 +02:00
def _get_file ( stream_info ) :
2016-06-27 23:07:59 +02:00
sd = stream_info [ ' sources ' ] [ ' lbry_sd_hash ' ]
2016-04-24 23:51:24 +02:00
for l in self . lbry_file_manager . lbry_files :
if l . sd_hash == sd :
return defer . succeed ( l )
return defer . succeed ( None )
2016-04-26 04:35:21 +02:00
d = self . _resolve_name ( name )
2016-04-24 23:51:24 +02:00
d . addCallback ( _get_file )
2016-04-24 10:42:42 +02:00
return d
def _get_lbry_file_by_sd_hash ( self , sd_hash ) :
2016-04-24 23:51:24 +02:00
for l in self . lbry_file_manager . lbry_files :
if l . sd_hash == sd_hash :
return defer . succeed ( l )
return defer . succeed ( None )
2016-04-24 10:42:42 +02:00
def _get_lbry_file_by_file_name ( self , file_name ) :
2016-04-24 23:51:24 +02:00
for l in self . lbry_file_manager . lbry_files :
if l . file_name == file_name :
return defer . succeed ( l )
return defer . succeed ( None )
def _get_lbry_file ( self , search_by , val , return_json = True ) :
def _log_get_lbry_file ( f ) :
2016-04-26 04:35:21 +02:00
if f and val :
2016-04-24 23:51:24 +02:00
log . info ( " Found LBRY file for " + search_by + " : " + val )
2016-04-26 04:35:21 +02:00
elif val :
2016-04-24 23:51:24 +02:00
log . info ( " Did not find LBRY file for " + search_by + " : " + val )
return f
2016-04-24 10:42:42 +02:00
2016-04-24 23:51:24 +02:00
def _get_json_for_return ( f ) :
2016-05-06 04:26:25 +02:00
def _get_file_status ( file_status ) :
message = STREAM_STAGES [ 2 ] [ 1 ] % ( file_status . name , file_status . num_completed , file_status . num_known , file_status . running_status )
return defer . succeed ( message )
2016-05-04 11:20:38 +02:00
def _generate_reply ( size ) :
2016-04-24 10:42:42 +02:00
if f . key :
2016-04-26 04:35:21 +02:00
key = binascii . b2a_hex ( f . key )
2016-04-24 10:42:42 +02:00
else :
2016-04-26 04:35:21 +02:00
key = None
2016-05-04 11:20:38 +02:00
if os . path . isfile ( os . path . join ( self . download_directory , f . file_name ) ) :
written_file = file ( os . path . join ( self . download_directory , f . file_name ) )
written_file . seek ( 0 , os . SEEK_END )
written_bytes = written_file . tell ( )
written_file . close ( )
else :
written_bytes = False
if search_by == " name " :
if val in self . streams . keys ( ) :
status = self . streams [ val ] . code
2016-05-06 03:45:25 +02:00
elif f in self . lbry_file_manager . lbry_files :
# if f.stopped:
# status = STREAM_STAGES[3]
# else:
status = STREAM_STAGES [ 2 ]
2016-05-04 11:20:38 +02:00
else :
status = [ False , False ]
else :
status = [ False , False ]
2016-05-06 04:26:25 +02:00
if status [ 0 ] == DOWNLOAD_RUNNING_CODE :
d = f . status ( )
d . addCallback ( _get_file_status )
2016-05-30 21:49:25 +02:00
d . addCallback ( lambda message : { ' completed ' : f . completed , ' file_name ' : f . file_name ,
' download_directory ' : f . download_directory ,
' download_path ' : os . path . join ( f . download_directory , f . file_name ) ,
2016-06-03 09:45:46 +02:00
' mime_type ' : mimetypes . guess_type ( os . path . join ( f . download_directory , f . file_name ) ) [ 0 ] ,
2016-05-30 21:49:25 +02:00
' key ' : key ,
2016-05-06 04:26:25 +02:00
' points_paid ' : f . points_paid , ' stopped ' : f . stopped ,
' stream_hash ' : f . stream_hash ,
' stream_name ' : f . stream_name ,
' suggested_file_name ' : f . suggested_file_name ,
' upload_allowed ' : f . upload_allowed , ' sd_hash ' : f . sd_hash ,
2016-08-05 00:44:12 +02:00
' lbry_uri ' : f . uri , ' txid ' : f . txid , ' claim_id ' : f . claim_id ,
2016-05-06 04:26:25 +02:00
' total_bytes ' : size ,
' written_bytes ' : written_bytes , ' code ' : status [ 0 ] ,
' message ' : message } )
else :
d = defer . succeed ( { ' completed ' : f . completed , ' file_name ' : f . file_name , ' key ' : key ,
2016-05-30 21:49:25 +02:00
' download_directory ' : f . download_directory ,
' download_path ' : os . path . join ( f . download_directory , f . file_name ) ,
2016-06-03 09:45:46 +02:00
' mime_type ' : mimetypes . guess_type ( os . path . join ( f . download_directory , f . file_name ) ) [ 0 ] ,
2016-05-06 04:26:25 +02:00
' points_paid ' : f . points_paid , ' stopped ' : f . stopped , ' stream_hash ' : f . stream_hash ,
' stream_name ' : f . stream_name , ' suggested_file_name ' : f . suggested_file_name ,
' upload_allowed ' : f . upload_allowed , ' sd_hash ' : f . sd_hash , ' total_bytes ' : size ,
2016-08-05 00:44:12 +02:00
' written_bytes ' : written_bytes , ' lbry_uri ' : f . uri , ' txid ' : f . txid , ' claim_id ' : f . claim_id ,
2016-05-11 01:03:14 +02:00
' code ' : status [ 0 ] , ' message ' : status [ 1 ] } )
2016-05-06 04:26:25 +02:00
return d
2016-05-04 11:20:38 +02:00
2016-05-11 07:48:23 +02:00
def _add_metadata ( message ) :
def _add_to_dict ( metadata ) :
message [ ' metadata ' ] = metadata
return defer . succeed ( message )
if f . txid :
d = self . _resolve_name ( f . uri )
2016-07-27 07:51:09 +02:00
d . addCallbacks ( _add_to_dict , lambda _ : _add_to_dict ( " Pending confirmation " ) )
2016-05-11 07:48:23 +02:00
else :
2016-05-11 10:44:23 +02:00
d = defer . succeed ( message )
2016-05-11 07:48:23 +02:00
return d
2016-05-04 11:20:38 +02:00
if f :
d = f . get_total_bytes ( )
d . addCallback ( _generate_reply )
2016-05-11 07:48:23 +02:00
d . addCallback ( _add_metadata )
2016-05-04 11:20:38 +02:00
return d
2016-04-24 10:42:42 +02:00
else :
return False
if search_by == " name " :
d = self . _get_lbry_file_by_uri ( val )
elif search_by == " sd_hash " :
d = self . _get_lbry_file_by_sd_hash ( val )
elif search_by == " file_name " :
d = self . _get_lbry_file_by_file_name ( val )
2016-08-05 05:10:15 +02:00
# d.addCallback(_log_get_lbry_file)
2016-04-24 23:51:24 +02:00
if return_json :
d . addCallback ( _get_json_for_return )
2016-04-24 10:42:42 +02:00
return d
2016-04-26 04:35:21 +02:00
def _get_lbry_files ( self ) :
d = defer . DeferredList ( [ self . _get_lbry_file ( ' sd_hash ' , l . sd_hash ) for l in self . lbry_file_manager . lbry_files ] )
return d
2016-04-21 05:50:13 +02:00
def _log_to_slack ( self , msg ) :
URL = " https://hooks.slack.com/services/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 "
msg = platform . platform ( ) + " : " + base58 . b58encode ( self . lbryid ) [ : 20 ] + " , " + msg
requests . post ( URL , json . dumps ( { " text " : msg } ) )
return defer . succeed ( None )
2016-05-30 21:49:25 +02:00
def _run_scripts ( self ) :
if len ( [ k for k in self . startup_scripts if ' run_once ' in k . keys ( ) ] ) :
log . info ( " Removing one time startup scripts " )
2016-06-04 20:18:58 +02:00
remaining_scripts = [ s for s in self . startup_scripts if ' run_once ' not in s . keys ( ) ]
startup_scripts = self . startup_scripts
self . startup_scripts = self . session_settings [ ' startup_scripts ' ] = remaining_scripts
2016-05-30 21:49:25 +02:00
f = open ( self . daemon_conf , " w " )
2016-06-04 20:18:58 +02:00
f . write ( json . dumps ( self . session_settings ) )
2016-05-30 21:49:25 +02:00
f . close ( )
2016-06-04 20:18:58 +02:00
for script in startup_scripts :
2016-05-30 21:49:25 +02:00
if script [ ' script_name ' ] == ' migrateto025 ' :
log . info ( " Running migrator to 0.2.5 " )
from lbrynet . lbrynet_daemon . daemon_scripts . migrateto025 import run as run_migrate
run_migrate ( self )
if script [ ' script_name ' ] == ' Autofetcher ' :
log . info ( " Starting autofetcher script " )
from lbrynet . lbrynet_daemon . daemon_scripts . Autofetcher import run as run_autofetcher
run_autofetcher ( self )
return defer . succeed ( None )
2016-07-28 22:12:20 +02:00
def _search ( self , search ) :
2016-08-03 00:34:48 +02:00
return self . lighthouse_client . search ( search )
2016-07-28 22:12:20 +02:00
2016-03-24 03:27:48 +01:00
def _render_response ( self , result , code ) :
2016-03-29 22:42:47 +02:00
return defer . succeed ( { ' result ' : result , ' code ' : code } )
2016-03-24 03:27:48 +01:00
def jsonrpc_is_running ( self ) :
"""
2016-04-13 20:47:34 +02:00
Check if lbrynet daemon is running
Args :
None
Returns : true if daemon completed startup , otherwise false
2016-03-24 03:27:48 +01:00
"""
2016-07-16 08:15:58 +02:00
log . info ( " is_running: " + str ( self . announced_startup ) )
2016-04-09 05:38:57 +02:00
2016-03-29 22:42:47 +02:00
if self . announced_startup :
2016-03-24 03:27:48 +01:00
return self . _render_response ( True , OK_CODE )
2016-03-03 00:32:48 +01:00
else :
2016-03-24 03:27:48 +01:00
return self . _render_response ( False , OK_CODE )
2016-02-19 02:41:23 +01:00
2016-04-09 07:33:27 +02:00
def jsonrpc_daemon_status ( self ) :
"""
2016-04-13 20:47:34 +02:00
Get lbrynet daemon status information
Args :
None
Returns :
2016-04-20 08:56:29 +02:00
' message ' : startup status message
' code ' : status_code
' progress ' : progress , only used in loading_wallet
' is_lagging ' : flag set to indicate lag , if set message will contain relevant message
2016-04-09 07:33:27 +02:00
"""
2016-04-09 08:09:30 +02:00
2016-04-20 10:24:32 +02:00
r = { ' code ' : self . startup_status [ 0 ] , ' message ' : self . startup_status [ 1 ] ,
' progress ' : None , ' is_lagging ' : None , ' problem_code ' : None }
2016-04-18 19:21:53 +02:00
if self . connection_problem :
2016-04-20 10:24:32 +02:00
r [ ' problem_code ' ] = self . connection_problem [ 0 ]
2016-04-18 19:21:53 +02:00
r [ ' message ' ] = self . connection_problem [ 1 ]
r [ ' is_lagging ' ] = True
elif self . startup_status [ 0 ] == LOADING_WALLET_CODE :
2016-05-14 23:36:30 +02:00
if self . wallet_type == ' lbryum ' :
if self . session . wallet . blocks_behind_alert != 0 :
r [ ' message ' ] = r [ ' message ' ] % ( str ( self . session . wallet . blocks_behind_alert ) + " blocks behind " )
r [ ' progress ' ] = self . session . wallet . catchup_progress
else :
r [ ' message ' ] = " Catching up with the blockchain "
r [ ' progress ' ] = 0
2016-05-06 20:31:44 +02:00
else :
r [ ' message ' ] = " Catching up with the blockchain "
r [ ' progress ' ] = 0
2016-07-16 08:15:58 +02:00
log . info ( " daemon status: " + str ( r ) )
2016-04-09 07:33:27 +02:00
return self . _render_response ( r , OK_CODE )
2016-04-09 10:00:31 +02:00
def jsonrpc_is_first_run ( self ) :
2016-04-09 21:43:10 +02:00
"""
2016-04-13 20:47:34 +02:00
Check if this is the first time lbrynet daemon has been run
Args :
None
Returns :
True if first run , otherwise False
2016-04-09 21:43:10 +02:00
"""
2016-07-16 08:15:58 +02:00
log . info ( " Check if is first run " )
2016-04-07 09:12:09 +02:00
try :
d = self . session . wallet . is_first_run ( )
except :
d = defer . fail ( None )
2016-04-09 21:43:10 +02:00
d . addCallbacks ( lambda r : self . _render_response ( r , OK_CODE ) , lambda _ : self . _render_response ( None , OK_CODE ) )
2016-04-07 09:12:09 +02:00
return d
2016-04-09 10:15:58 +02:00
def jsonrpc_get_start_notice ( self ) :
2016-04-09 21:43:10 +02:00
"""
2016-04-13 20:47:34 +02:00
Get special message to be displayed at startup
2016-04-09 21:43:10 +02:00
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
Startup message , such as first run notification
"""
2016-04-09 22:14:28 +02:00
2016-07-16 08:15:58 +02:00
log . info ( " Get startup notice " )
2016-04-09 22:14:28 +02:00
2016-04-09 22:18:06 +02:00
if self . first_run and not self . session . wallet . wallet_balance :
2016-04-09 22:14:28 +02:00
return self . _render_response ( self . startup_message , OK_CODE )
elif self . first_run :
return self . _render_response ( None , OK_CODE )
else :
self . _render_response ( self . startup_message , OK_CODE )
2016-04-12 04:28:46 +02:00
def jsonrpc_version ( self ) :
"""
Get lbry version information
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
2016-04-16 04:31:06 +02:00
" platform " : platform string
" os_release " : os release string
" os_system " : os name
" lbrynet_version: " : lbrynet_version ,
" lbryum_version: " : lbryum_version ,
" ui_version " : commit hash of ui version being used
2016-04-14 04:07:27 +02:00
" remote_lbrynet " : most recent lbrynet version available from github
" remote_lbryum " : most recent lbryum version available from github
2016-04-12 04:28:46 +02:00
"""
2016-05-30 21:49:25 +02:00
platform_info = self . _get_platform ( )
2016-04-12 04:28:46 +02:00
msg = {
2016-05-30 21:49:25 +02:00
' platform ' : platform_info [ ' platform ' ] ,
' os_release ' : platform_info [ ' os_release ' ] ,
' os_system ' : platform_info [ ' os_system ' ] ,
2016-04-16 04:31:06 +02:00
' lbrynet_version ' : lbrynet_version ,
' lbryum_version ' : lbryum_version ,
' ui_version ' : self . ui_version ,
2016-07-30 04:36:30 +02:00
' remote_lbrynet ' : self . git_lbrynet_version ,
' remote_lbryum ' : self . git_lbryum_version ,
' lbrynet_update_available ' : utils . version_is_greater_than ( self . git_lbrynet_version , lbrynet_version ) ,
' lbryum_update_available ' : utils . version_is_greater_than ( self . git_lbryum_version , lbryum_version ) ,
2016-04-12 04:28:46 +02:00
}
2016-04-12 05:13:26 +02:00
2016-07-16 08:15:58 +02:00
log . info ( " Get version info: " + json . dumps ( msg ) )
2016-04-12 04:28:46 +02:00
return self . _render_response ( msg , OK_CODE )
2016-04-09 10:15:58 +02:00
2016-03-24 03:27:48 +01:00
def jsonrpc_get_settings ( self ) :
2016-01-07 16:24:55 +01:00
"""
2016-04-13 20:47:34 +02:00
Get lbrynet daemon settings
Args :
None
Returns :
' run_on_startup ' : bool ,
' data_rate ' : float ,
' max_key_fee ' : float ,
2016-04-22 08:45:05 +02:00
' download_directory ' : string ,
2016-04-13 20:47:34 +02:00
' max_upload ' : float , 0.0 for unlimited
' max_download ' : float , 0.0 for unlimited
' upload_log ' : bool ,
' search_timeout ' : float ,
2016-04-22 08:45:05 +02:00
' download_timeout ' : int
2016-04-13 20:47:34 +02:00
' max_search_results ' : int ,
' wallet_type ' : string ,
' delete_blobs_on_remove ' : bool ,
' peer_port ' : int ,
' dht_node_port ' : int ,
' use_upnp ' : bool ,
' start_lbrycrdd ' : bool ,
2016-01-07 16:24:55 +01:00
"""
2016-01-12 18:42:32 +01:00
2016-07-16 08:15:58 +02:00
log . info ( " Get daemon settings " )
2016-03-24 03:27:48 +01:00
return self . _render_response ( self . session_settings , OK_CODE )
2016-01-07 16:24:55 +01:00
2016-03-24 03:27:48 +01:00
def jsonrpc_set_settings ( self , p ) :
2016-01-24 06:10:22 +01:00
"""
2016-04-13 20:47:34 +02:00
Set lbrynet daemon settings
Args :
' run_on_startup ' : bool ,
' data_rate ' : float ,
' max_key_fee ' : float ,
2016-04-22 08:45:05 +02:00
' download_directory ' : string ,
2016-04-13 20:47:34 +02:00
' max_upload ' : float , 0.0 for unlimited
' max_download ' : float , 0.0 for unlimited
' upload_log ' : bool ,
2016-04-23 00:18:17 +02:00
' download_timeout ' : int
2016-04-13 20:47:34 +02:00
Returns :
settings dict
2016-01-24 06:10:22 +01:00
"""
2016-04-22 08:45:05 +02:00
def _log_settings_change ( ) :
2016-07-16 08:15:58 +02:00
log . info ( " Set daemon settings to " + json . dumps ( self . session_settings ) )
2016-04-07 09:12:09 +02:00
2016-03-24 03:27:48 +01:00
d = self . _update_settings ( p )
2016-05-31 23:38:00 +02:00
d . addErrback ( lambda err : log . info ( err . getTraceback ( ) ) )
2016-04-22 08:45:05 +02:00
d . addCallback ( lambda _ : _log_settings_change ( ) )
2016-04-07 09:12:09 +02:00
d . addCallback ( lambda _ : self . _render_response ( self . session_settings , OK_CODE ) )
2016-01-07 16:24:55 +01:00
2016-04-07 09:12:09 +02:00
return d
2016-01-07 16:24:55 +01:00
2016-04-13 20:47:34 +02:00
def jsonrpc_help ( self , p = None ) :
"""
Function to retrieve docstring for API function
Args :
optional ' function ' : function to retrieve documentation for
optional ' callable_during_startup ' :
Returns :
if given a function , returns given documentation
if given callable_during_startup flag , returns list of functions callable during the startup sequence
if no params are given , returns the list of callable functions
"""
if not p :
2016-04-16 01:37:27 +02:00
return self . _render_response ( self . _listFunctions ( ) , OK_CODE )
2016-04-13 20:47:34 +02:00
elif ' callable_during_start ' in p . keys ( ) :
return self . _render_response ( ALLOWED_DURING_STARTUP , OK_CODE )
elif ' function ' in p . keys ( ) :
func_path = p [ ' function ' ]
function = self . _getFunction ( func_path )
return self . _render_response ( function . __doc__ , OK_CODE )
else :
return self . _render_response ( self . jsonrpc_help . __doc__ , OK_CODE )
2016-03-24 03:27:48 +01:00
def jsonrpc_get_balance ( self ) :
2015-12-06 23:32:17 +01:00
"""
2016-04-13 20:47:34 +02:00
Get balance
2016-03-24 03:27:48 +01:00
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
balance , float
2015-12-06 23:32:17 +01:00
"""
2015-12-14 18:59:44 +01:00
2016-07-16 08:15:58 +02:00
log . info ( " Get balance " )
2016-03-29 22:42:47 +02:00
return self . _render_response ( float ( self . session . wallet . wallet_balance ) , OK_CODE )
2015-12-06 23:32:17 +01:00
2016-03-24 03:27:48 +01:00
def jsonrpc_stop ( self ) :
2015-12-06 23:32:17 +01:00
"""
2016-01-24 06:10:22 +01:00
Stop lbrynet - daemon
2016-03-24 03:27:48 +01:00
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
shutdown message
2015-12-06 23:32:17 +01:00
"""
2016-01-11 21:52:38 +01:00
def _disp_shutdown ( ) :
2016-03-24 03:27:48 +01:00
log . info ( " Shutting down lbrynet daemon " )
2016-01-11 21:52:38 +01:00
d = self . _shutdown ( )
d . addCallback ( lambda _ : _disp_shutdown ( ) )
2016-04-20 08:56:29 +02:00
d . addCallback ( lambda _ : reactor . callLater ( 0.0 , reactor . stop ) )
2016-01-11 21:52:38 +01:00
2016-03-24 03:27:48 +01:00
return self . _render_response ( " Shutting down " , OK_CODE )
2015-12-06 23:32:17 +01:00
2016-03-24 03:27:48 +01:00
def jsonrpc_get_lbry_files ( self ) :
2015-12-15 03:49:01 +01:00
"""
Get LBRY files
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
List of lbry files :
' completed ' : bool
' file_name ' : string
' key ' : hex string
' points_paid ' : float
' stopped ' : bool
' stream_hash ' : base 58 string
' stream_name ' : string
' suggested_file_name ' : string
' upload_allowed ' : bool
2016-04-24 10:42:42 +02:00
' sd_hash ' : string
2015-12-15 03:49:01 +01:00
"""
2016-04-26 04:35:21 +02:00
d = self . _get_lbry_files ( )
d . addCallback ( lambda r : [ d [ 1 ] for d in r ] )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) if len ( r ) else self . _render_response ( False , OK_CODE ) )
return d
2015-12-15 03:49:01 +01:00
2016-04-24 10:42:42 +02:00
def jsonrpc_get_lbry_file ( self , p ) :
"""
Get lbry file
Args :
' name ' : get file by lbry uri ,
' sd_hash ' : get file by the hash in the name claim ,
' file_name ' : get file by its name in the downloads folder ,
Returns :
' completed ' : bool
' file_name ' : string
' key ' : hex string
' points_paid ' : float
' stopped ' : bool
' stream_hash ' : base 58 string
' stream_name ' : string
' suggested_file_name ' : string
' upload_allowed ' : bool
' sd_hash ' : string
"""
if p . keys ( ) [ 0 ] in [ ' name ' , ' sd_hash ' , ' file_name ' ] :
search_type = p . keys ( ) [ 0 ]
d = self . _get_lbry_file ( search_type , p [ search_type ] )
else :
d = defer . fail ( )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
2016-03-24 03:27:48 +01:00
def jsonrpc_resolve_name ( self , p ) :
2015-12-06 23:32:17 +01:00
"""
Resolve stream info from a LBRY uri
2015-12-15 03:49:01 +01:00
2016-04-13 20:47:34 +02:00
Args :
' name ' : name to look up , string , do not include lbry : / / prefix
Returns :
metadata from name claim
2015-12-06 23:32:17 +01:00
"""
2016-04-13 20:47:34 +02:00
2016-07-27 21:02:54 +02:00
force = p . get ( ' force ' , False )
2016-07-27 03:46:04 +02:00
if ' name ' in p :
2016-04-13 20:47:34 +02:00
name = p [ ' name ' ]
else :
return self . _render_response ( None , BAD_REQUEST )
2015-12-06 23:32:17 +01:00
2016-07-27 03:46:04 +02:00
d = self . _resolve_name ( name , force_refresh = force )
2016-04-26 04:35:21 +02:00
d . addCallbacks ( lambda info : self . _render_response ( info , OK_CODE ) , lambda _ : server . failure )
2015-12-06 23:32:17 +01:00
return d
2016-07-28 20:55:17 +02:00
def jsonrpc_get_claim_info ( self , p ) :
2016-07-28 22:12:20 +02:00
"""
Resolve claim info from a LBRY uri
Args :
' name ' : name to look up , string , do not include lbry : / / prefix
Returns :
txid , amount , value , n , height
"""
2016-07-28 20:55:17 +02:00
def _convert_amount_to_float ( r ) :
2016-08-09 03:32:39 +02:00
if not r :
return False
else :
r [ ' amount ' ] = float ( r [ ' amount ' ] ) / 10 * * 8
return r
2016-07-28 20:55:17 +02:00
name = p [ ' name ' ]
2016-08-08 08:37:08 +02:00
txid = p . get ( ' txid ' , None )
d = self . session . wallet . get_claim_info ( name , txid )
2016-07-28 20:55:17 +02:00
d . addCallback ( _convert_amount_to_float )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
2016-07-21 17:44:59 +02:00
def _process_get_parameters ( self , p ) :
""" Extract info from input parameters and fill in default values for `get` call. """
# TODO: this process can be abstracted s.t. each method
# can spec what parameters it expects and how to set default values
timeout = p . get ( ' timeout ' , self . download_timeout )
download_directory = p . get ( ' download_directory ' , self . download_directory )
2016-07-21 18:05:06 +02:00
file_name = p . get ( ' file_name ' )
stream_info = p . get ( ' stream_info ' )
2016-07-21 17:44:59 +02:00
sd_hash = get_sd_hash ( stream_info )
wait_for_write = p . get ( ' wait_for_write ' , True )
name = p . get ( ' name ' )
return Parameters (
timeout = timeout ,
download_directory = download_directory ,
file_name = file_name ,
stream_info = stream_info ,
sd_hash = sd_hash ,
wait_for_write = wait_for_write ,
name = name
)
2016-03-24 03:27:48 +01:00
def jsonrpc_get ( self , p ) :
2016-07-21 17:44:59 +02:00
""" Download stream from a LBRY uri.
2015-12-15 03:49:01 +01:00
2016-04-13 20:47:34 +02:00
Args :
' name ' : name to download , string
2016-05-05 03:25:46 +02:00
' download_directory ' : optional , path to directory where file will be saved , string
' file_name ' : optional , a user specified name for the downloaded file
' stream_info ' : optional , specified stream info overrides name
2016-07-21 19:32:50 +02:00
' timeout ' : optional
2016-07-21 17:44:59 +02:00
' wait_for_write ' : optional , defaults to True
2016-04-13 20:47:34 +02:00
Returns :
' stream_hash ' : hex string
' path ' : path of download
2015-12-06 23:32:17 +01:00
"""
2016-07-21 17:44:59 +02:00
params = self . _process_get_parameters ( p )
if not params . name :
return server . failure
if params . name in self . waiting_on :
return server . failure
d = self . _download_name ( name = params . name ,
timeout = params . timeout ,
download_directory = params . download_directory ,
stream_info = params . stream_info ,
file_name = params . file_name ,
wait_for_write = params . wait_for_write )
2016-07-21 18:05:06 +02:00
d . addCallback ( get_output_callback ( params ) )
2016-07-21 17:44:59 +02:00
d . addCallback ( lambda message : self . _render_response ( message , OK_CODE ) )
2015-12-08 06:37:49 +01:00
return d
2016-04-24 23:51:24 +02:00
def jsonrpc_stop_lbry_file ( self , p ) :
"""
Stop lbry file
Args :
' name ' : stop file by lbry uri ,
' sd_hash ' : stop file by the hash in the name claim ,
' file_name ' : stop file by its name in the downloads folder ,
Returns :
confirmation message
"""
def _stop_file ( f ) :
d = self . lbry_file_manager . toggle_lbry_file_running ( f )
d . addCallback ( lambda _ : " Stopped LBRY file " )
return d
if p . keys ( ) [ 0 ] in [ ' name ' , ' sd_hash ' , ' file_name ' ] :
search_type = p . keys ( ) [ 0 ]
d = self . _get_lbry_file ( search_type , p [ search_type ] , return_json = False )
d . addCallback ( lambda l : _stop_file ( l ) if not l . stopped else " LBRY file wasn ' t running " )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
def jsonrpc_start_lbry_file ( self , p ) :
"""
Stop lbry file
Args :
' name ' : stop file by lbry uri ,
' sd_hash ' : stop file by the hash in the name claim ,
' file_name ' : stop file by its name in the downloads folder ,
Returns :
confirmation message
"""
def _start_file ( f ) :
d = self . lbry_file_manager . toggle_lbry_file_running ( f )
return defer . succeed ( " Started LBRY file " )
if p . keys ( ) [ 0 ] in [ ' name ' , ' sd_hash ' , ' file_name ' ] :
search_type = p . keys ( ) [ 0 ]
d = self . _get_lbry_file ( search_type , p [ search_type ] , return_json = False )
d . addCallback ( lambda l : _start_file ( l ) if l . stopped else " LBRY file was already running " )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
2016-01-17 05:06:24 +01:00
2016-07-28 22:12:20 +02:00
def jsonrpc_get_est_cost ( self , p ) :
"""
Get estimated cost for a lbry uri
Args :
' name ' : lbry uri
Returns :
estimated cost
"""
name = p [ ' name ' ]
d = self . _get_est_cost ( name )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
2016-03-24 03:27:48 +01:00
def jsonrpc_search_nametrie ( self , p ) :
2016-01-24 06:10:22 +01:00
"""
2016-07-30 04:49:50 +02:00
Search the nametrie for claims
2016-01-24 06:10:22 +01:00
2016-04-13 20:47:34 +02:00
Args :
' search ' : search query , string
Returns :
List of search results
2016-01-24 06:10:22 +01:00
"""
2016-08-08 08:37:08 +02:00
# TODO: change this function to "search"
2016-07-30 04:49:50 +02:00
2016-04-13 20:47:34 +02:00
if ' search ' in p . keys ( ) :
search = p [ ' search ' ]
else :
return self . _render_response ( None , BAD_REQUEST )
2016-03-24 03:27:48 +01:00
2016-08-02 23:33:57 +02:00
# TODO: have ui accept the actual outputs
2016-01-17 05:06:24 +01:00
def _clean ( n ) :
t = [ ]
for i in n :
2016-08-02 22:45:06 +02:00
td = { k : i [ ' value ' ] [ k ] for k in i [ ' value ' ] }
td [ ' cost_est ' ] = float ( i [ ' cost ' ] )
td [ ' thumbnail ' ] = i [ ' value ' ] . get ( ' thumbnail ' , " img/Free-speech-flag.svg " )
td [ ' name ' ] = i [ ' name ' ]
t . append ( td )
2016-01-17 05:06:24 +01:00
return t
2016-07-30 04:44:38 +02:00
log . info ( ' Search: %s ' % search )
2016-01-17 05:06:24 +01:00
2016-07-28 22:12:20 +02:00
d = self . _search ( search )
2016-01-17 05:06:24 +01:00
d . addCallback ( _clean )
2016-03-29 22:42:47 +02:00
d . addCallback ( lambda results : self . _render_response ( results , OK_CODE ) )
2016-01-17 05:06:24 +01:00
return d
2016-03-24 03:27:48 +01:00
def jsonrpc_delete_lbry_file ( self , p ) :
"""
Delete a lbry file
2016-01-21 04:00:28 +01:00
2016-04-13 20:47:34 +02:00
Args :
' file_name ' : downloaded file name , string
Returns :
confirmation message
2016-03-24 03:27:48 +01:00
"""
2016-01-21 04:00:28 +01:00
2016-05-30 21:49:25 +02:00
if ' delete_target_file ' in p . keys ( ) :
delete_file = p [ ' delete_target_file ' ]
else :
delete_file = True
2016-04-24 23:51:24 +02:00
def _delete_file ( f ) :
file_name = f . file_name
2016-05-30 21:49:25 +02:00
d = self . _delete_lbry_file ( f , delete_file = delete_file )
2016-04-24 23:51:24 +02:00
d . addCallback ( lambda _ : " Deleted LBRY file " + file_name )
2016-04-07 09:12:09 +02:00
return d
2016-01-21 04:00:28 +01:00
2016-05-30 21:49:25 +02:00
if ' name ' in p . keys ( ) or ' sd_hash ' in p . keys ( ) or ' file_name ' in p . keys ( ) :
search_type = [ k for k in p . keys ( ) if k != ' delete_target_file ' ] [ 0 ]
2016-04-24 23:51:24 +02:00
d = self . _get_lbry_file ( search_type , p [ search_type ] , return_json = False )
2016-05-01 05:43:02 +02:00
d . addCallback ( lambda l : _delete_file ( l ) if l else False )
2016-04-24 23:51:24 +02:00
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
2016-03-24 03:27:48 +01:00
def jsonrpc_publish ( self , p ) :
"""
2016-06-27 23:07:59 +02:00
Make a new name claim and publish associated data to lbrynet
2016-01-21 04:00:28 +01:00
2016-04-13 20:47:34 +02:00
Args :
' name ' : name to be claimed , string
' file_path ' : path to file to be associated with name , string
' bid ' : amount of credits to commit in this claim , float
2016-06-27 23:07:59 +02:00
' metadata ' : metadata dictionary
optional ' fee '
2016-04-13 20:47:34 +02:00
Returns :
2016-06-27 23:07:59 +02:00
Claim txid
2016-03-24 03:27:48 +01:00
"""
2016-07-25 07:40:26 +02:00
2016-08-09 03:32:39 +02:00
def _set_address ( address , currency , m ) :
2016-08-07 02:02:51 +02:00
log . info ( " Generated new address for key fee: " + str ( address ) )
2016-08-09 03:32:39 +02:00
m [ ' fee ' ] [ currency ] [ ' address ' ] = address
return m
2016-08-07 02:02:51 +02:00
2016-06-27 23:07:59 +02:00
name = p [ ' name ' ]
2016-08-07 02:02:51 +02:00
2016-08-09 03:32:39 +02:00
log . info ( " Publish: " )
log . info ( p )
2016-07-29 00:48:29 +02:00
try :
verify_name_characters ( name )
2016-08-07 02:02:51 +02:00
except AssertionError :
2016-07-29 00:48:29 +02:00
log . error ( " Bad name " )
return defer . fail ( InvalidNameError ( " Bad name " ) )
2016-08-07 02:02:51 +02:00
2016-06-27 23:07:59 +02:00
bid = p [ ' bid ' ]
2016-07-25 07:40:26 +02:00
2016-08-07 02:02:51 +02:00
try :
metadata = Metadata ( p [ ' metadata ' ] )
make_lbry_file = False
except AssertionError :
make_lbry_file = True
metadata = p [ ' metadata ' ]
file_path = p [ ' file_path ' ]
2016-07-25 07:40:26 +02:00
2016-07-27 09:00:44 +02:00
if not self . pending_claim_checker . running :
self . pending_claim_checker . start ( 30 )
2016-07-27 07:13:41 +02:00
d = self . _resolve_name ( name , force_refresh = True )
2016-07-27 07:39:23 +02:00
d . addErrback ( lambda _ : None )
2016-07-27 02:52:05 +02:00
2016-06-27 23:07:59 +02:00
if ' fee ' in p :
2016-07-25 07:40:26 +02:00
metadata [ ' fee ' ] = p [ ' fee ' ]
2016-07-26 23:20:51 +02:00
assert len ( metadata [ ' fee ' ] ) == 1 , " Too many fees "
for c in metadata [ ' fee ' ] :
if ' address ' not in metadata [ ' fee ' ] [ c ] :
2016-07-27 02:52:05 +02:00
d . addCallback ( lambda _ : self . session . wallet . get_new_address ( ) )
2016-08-09 03:32:39 +02:00
d . addCallback ( lambda addr : _set_address ( addr , c , metadata ) )
else :
d . addCallback ( lambda _ : metadata )
2016-08-07 02:02:51 +02:00
if make_lbry_file :
pub = Publisher ( self . session , self . lbry_file_manager , self . session . wallet )
2016-08-09 03:32:39 +02:00
d . addCallback ( lambda meta : pub . start ( name , file_path , bid , meta ) )
2016-08-07 02:02:51 +02:00
else :
2016-08-09 03:32:39 +02:00
d . addCallback ( lambda meta : self . session . wallet . claim_name ( name , bid , meta ) )
2016-07-27 09:00:44 +02:00
d . addCallback ( lambda txid : self . _add_to_pending_claims ( name , txid ) )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
2016-01-21 04:00:28 +01:00
return d
2015-12-06 23:32:17 +01:00
2016-03-24 03:27:48 +01:00
def jsonrpc_abandon_name ( self , p ) :
"""
2016-04-13 20:47:34 +02:00
Abandon a name and reclaim credits from the claim
2016-03-24 03:27:48 +01:00
2016-04-13 20:47:34 +02:00
Args :
' txid ' : txid of claim , string
Return :
Confirmation message
2016-03-24 03:27:48 +01:00
"""
2016-04-13 20:47:34 +02:00
if ' txid ' in p . keys ( ) :
txid = p [ ' txid ' ]
else :
return server . failure
2016-03-24 03:27:48 +01:00
2016-04-07 09:12:09 +02:00
def _disp ( x ) :
2016-07-16 08:15:58 +02:00
log . info ( " Abandoned name claim tx " + str ( x ) )
2016-04-07 09:12:09 +02:00
return self . _render_response ( x , OK_CODE )
2016-01-26 02:28:05 +01:00
d = defer . Deferred ( )
2016-04-13 20:47:34 +02:00
d . addCallback ( lambda _ : self . session . wallet . abandon_name ( txid ) )
2016-04-07 09:12:09 +02:00
d . addCallback ( _disp )
2016-01-26 02:28:05 +01:00
d . callback ( None )
return d
2016-08-08 08:32:56 +02:00
def jsonrpc_support_claim ( self , p ) :
"""
Support a name claim
Args :
' name ' : name
' claim_id ' : claim id of claim to support
' amount ' : amount to support by
Return :
txid
"""
name = p [ ' name ' ]
claim_id = p [ ' claim_id ' ]
amount = p [ ' amount ' ]
d = self . session . wallet . support_claim ( name , claim_id , amount )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
2016-03-24 03:27:48 +01:00
def jsonrpc_get_name_claims ( self ) :
"""
2016-04-13 20:47:34 +02:00
Get my name claims
2016-03-24 03:27:48 +01:00
2016-04-13 20:47:34 +02:00
Args :
None
Returns
list of name claims
2016-03-24 03:27:48 +01:00
"""
2016-04-13 20:47:34 +02:00
2016-01-26 02:28:05 +01:00
def _clean ( claims ) :
for c in claims :
for k in c . keys ( ) :
2016-02-11 14:32:48 +01:00
if isinstance ( c [ k ] , Decimal ) :
2016-01-26 02:28:05 +01:00
c [ k ] = float ( c [ k ] )
2016-03-29 22:42:47 +02:00
return defer . succeed ( claims )
2016-01-26 02:28:05 +01:00
d = self . session . wallet . get_name_claims ( )
d . addCallback ( _clean )
2016-03-29 22:42:47 +02:00
d . addCallback ( lambda claims : self . _render_response ( claims , OK_CODE ) )
2016-01-26 02:28:05 +01:00
return d
2016-08-08 07:43:10 +02:00
def jsonrpc_get_claims_for_name ( self , p ) :
"""
Get claims for a name
Args :
' name ' : name
Returns
list of name claims
"""
name = p [ ' name ' ]
d = self . session . wallet . get_claims_for_name ( name )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
2016-04-18 05:23:20 +02:00
def jsonrpc_get_transaction_history ( self ) :
"""
Get transaction history
Args :
None
Returns :
list of transactions
"""
d = self . session . wallet . get_history ( )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
def jsonrpc_get_transaction ( self , p ) :
"""
Get a decoded transaction from a txid
Args :
txid : txid hex string
Returns :
JSON formatted transaction
"""
txid = p [ ' txid ' ]
d = self . session . wallet . get_tx_json ( txid )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
def jsonrpc_get_public_key_from_wallet ( self , p ) :
"""
Get public key from wallet address
Args :
wallet : wallet address , base58
Returns :
public key
"""
wallet = p [ ' wallet ' ]
d = self . session . wallet . get_pub_keys ( wallet )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
2016-03-24 03:27:48 +01:00
def jsonrpc_get_time_behind_blockchain ( self ) :
"""
2016-04-13 20:47:34 +02:00
Get number of blocks behind the blockchain
2016-03-24 03:27:48 +01:00
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
number of blocks behind blockchain , int
2016-03-24 03:27:48 +01:00
"""
2016-04-07 09:12:09 +02:00
def _get_time_behind ( ) :
try :
local_height = self . session . wallet . network . get_local_height ( )
remote_height = self . session . wallet . network . get_server_height ( )
return defer . succeed ( remote_height - local_height )
except :
return defer . fail ( )
d = _get_time_behind ( )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
2016-01-26 02:28:05 +01:00
return d
2016-03-24 03:27:48 +01:00
def jsonrpc_get_new_address ( self ) :
"""
Generate a new wallet address
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
new wallet address , base 58 string
2016-03-24 03:27:48 +01:00
"""
2016-04-13 20:47:34 +02:00
2016-02-25 23:17:07 +01:00
def _disp ( address ) :
2016-07-16 08:15:58 +02:00
log . info ( " Got new wallet address: " + address )
2016-03-29 22:42:47 +02:00
return defer . succeed ( address )
2016-02-25 23:17:07 +01:00
d = self . session . wallet . get_new_address ( )
d . addCallback ( _disp )
2016-03-29 22:42:47 +02:00
d . addCallback ( lambda address : self . _render_response ( address , OK_CODE ) )
2016-02-25 23:17:07 +01:00
return d
2016-05-04 05:13:31 +02:00
def jsonrpc_send_amount_to_address ( self , p ) :
"""
Send credits to an address
Args :
amount : the amount to send
address : the address of the recipient
Returns :
True if payment successfully scheduled
"""
if ' amount ' in p . keys ( ) and ' address ' in p . keys ( ) :
amount = p [ ' amount ' ]
address = p [ ' address ' ]
else :
return server . failure
reserved_points = self . session . wallet . reserve_points ( address , amount )
if reserved_points is None :
return defer . fail ( InsufficientFundsError ( ) )
d = self . session . wallet . send_points_to_address ( reserved_points , amount )
d . addCallback ( lambda _ : self . _render_response ( True , OK_CODE ) )
return d
def jsonrpc_get_best_blockhash ( self ) :
"""
Get hash of most recent block
Args :
None
Returns :
Hash of most recent block
"""
d = self . session . wallet . get_best_blockhash ( )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
def jsonrpc_get_block ( self , p ) :
"""
Get contents of a block
Args :
blockhash : hash of the block to look up
Returns :
requested block
"""
if ' blockhash ' in p . keys ( ) :
blockhash = p [ ' blockhash ' ]
else :
return server . failure
d = self . session . wallet . get_block ( blockhash )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
def jsonrpc_get_claims_for_tx ( self , p ) :
"""
Get claims for tx
Args :
txid : txid of a name claim transaction
Returns :
any claims contained in the requested tx
"""
if ' txid ' in p . keys ( ) :
txid = p [ ' txid ' ]
else :
return server . failure
d = self . session . wallet . get_claims_from_tx ( txid )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
2016-08-02 04:31:49 +02:00
return d
def jsonrpc_download_descriptor ( self , p ) :
"""
Download and return a sd blob
Args :
sd_hash
Returns
sd blob , dict
"""
sd_hash = p [ ' sd_hash ' ]
2016-08-03 23:57:26 +02:00
timeout = p . get ( ' timeout ' , DEFAULT_SD_DOWNLOAD_TIMEOUT )
2016-08-02 08:54:04 +02:00
2016-08-03 23:57:26 +02:00
d = self . _download_sd_blob ( sd_hash , timeout )
2016-08-02 08:54:04 +02:00
d . addCallbacks ( lambda r : self . _render_response ( r , OK_CODE ) , lambda _ : self . _render_response ( False , OK_CODE ) )
2016-05-04 05:13:31 +02:00
return d
def jsonrpc_get_nametrie ( self ) :
"""
Get the nametrie
Args :
None
Returns :
Name claim trie
"""
d = self . session . wallet . get_nametrie ( )
d . addCallback ( lambda r : [ i for i in r if ' txid ' in i . keys ( ) ] )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
2016-05-14 23:57:34 +02:00
def jsonrpc_set_miner ( self , p ) :
"""
Start of stop the miner , function only available when lbrycrd is set as the wallet
Args :
run : True / False
Returns :
miner status , True / False
"""
stat = p [ ' run ' ]
if stat :
d = self . session . wallet . start_miner ( )
else :
d = self . session . wallet . stop_miner ( )
d . addCallback ( lambda _ : self . session . wallet . get_miner_status ( ) )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
def jsonrpc_get_miner_status ( self ) :
"""
Get status of miner
Args :
None
Returns :
True / False
"""
d = self . session . wallet . get_miner_status ( )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
2016-05-31 23:38:00 +02:00
def jsonrpc_log ( self , p ) :
2016-04-13 08:35:31 +02:00
"""
2016-05-30 21:49:25 +02:00
Log message
2016-04-13 20:47:34 +02:00
Args :
2016-05-31 23:38:00 +02:00
' message ' : message to be logged
2016-04-13 20:47:34 +02:00
Returns :
2016-05-30 21:49:25 +02:00
True
2016-04-13 08:35:31 +02:00
"""
2016-05-31 23:38:00 +02:00
message = p [ ' message ' ]
log . info ( " API client log request: %s " % message )
2016-05-30 21:49:25 +02:00
return self . _render_response ( True , OK_CODE )
2016-01-24 06:10:22 +01:00
2016-04-18 01:54:04 +02:00
def jsonrpc_upload_log ( self , p = None ) :
2016-04-18 09:41:16 +02:00
"""
Upload log
Args , optional :
2016-04-18 19:21:53 +02:00
' name_prefix ' : prefix to indicate what is requesting the log upload
' exclude_previous ' : true / false , whether or not to exclude previous sessions from upload , defaults on true
2016-05-30 21:49:25 +02:00
Returns :
2016-04-18 09:41:16 +02:00
True
"""
2016-04-18 01:54:04 +02:00
if p :
if ' name_prefix ' in p . keys ( ) :
2016-05-21 01:59:49 +02:00
log_type = p [ ' name_prefix ' ] + ' _api '
2016-05-31 23:38:00 +02:00
elif ' log_type ' in p . keys ( ) :
log_type = p [ ' log_type ' ] + ' _api '
2016-04-18 01:54:04 +02:00
else :
2016-05-21 01:59:49 +02:00
log_type = None
2016-04-20 20:44:26 +02:00
if ' exclude_previous ' in p . keys ( ) :
2016-04-18 01:54:04 +02:00
exclude_previous = p [ ' exclude_previous ' ]
else :
exclude_previous = True
2016-04-20 20:44:26 +02:00
2016-04-21 05:50:13 +02:00
if ' message ' in p . keys ( ) :
2016-07-16 08:15:58 +02:00
log . info ( " Upload log message: " + str ( p [ ' message ' ] ) )
2016-04-21 05:50:13 +02:00
2016-04-20 20:44:26 +02:00
if ' force ' in p . keys ( ) :
force = p [ ' force ' ]
else :
force = False
2016-04-18 01:54:04 +02:00
else :
2016-05-21 01:59:49 +02:00
log_type = " api "
2016-04-18 01:54:04 +02:00
exclude_previous = True
2016-05-21 01:59:49 +02:00
d = self . _upload_log ( log_type = log_type , exclude_previous = exclude_previous , force = force )
2016-04-21 05:50:13 +02:00
if ' message ' in p . keys ( ) :
d . addCallback ( lambda _ : self . _log_to_slack ( p [ ' message ' ] ) )
2016-04-18 01:54:04 +02:00
d . addCallback ( lambda _ : self . _render_response ( True , OK_CODE ) )
return d
2016-05-30 21:49:25 +02:00
def jsonrpc_configure_ui ( self , p ) :
"""
Configure the UI being hosted
Args , optional :
' branch ' : a branch name on lbryio / lbry - web - ui
' path ' : path to a ui folder
"""
2016-07-06 07:27:25 +02:00
if ' check_requirements ' in p :
check_require = p [ ' check_requirements ' ]
2016-05-30 21:49:25 +02:00
else :
2016-07-06 07:27:25 +02:00
check_require = True
if ' path ' in p :
d = self . lbry_ui_manager . setup ( user_specified = p [ ' path ' ] , check_requirements = check_require )
elif ' branch ' in p :
d = self . lbry_ui_manager . setup ( branch = p [ ' branch ' ] , check_requirements = check_require )
else :
d = self . lbry_ui_manager . setup ( check_requirements = check_require )
2016-05-30 21:49:25 +02:00
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
def jsonrpc_reveal ( self , p ) :
"""
2016-07-06 21:57:00 +02:00
Reveal a file or directory in file browser
2016-05-30 21:49:25 +02:00
Args :
2016-07-06 21:57:00 +02:00
' path ' : path to be selected in file browser
2016-05-30 21:49:25 +02:00
Returns :
2016-07-06 21:57:00 +02:00
True , opens file browser
2016-05-30 21:49:25 +02:00
"""
path = p [ ' path ' ]
if sys . platform == " darwin " :
d = threads . deferToThread ( subprocess . Popen , [ ' open ' , ' -R ' , path ] )
else :
2016-07-06 21:57:00 +02:00
# No easy way to reveal specific files on Linux, so just open the containing directory
2016-07-21 17:44:59 +02:00
d = threads . deferToThread ( subprocess . Popen , [ ' xdg-open ' , os . path . dirname ( path ) ] )
2016-07-06 21:57:00 +02:00
2016-05-30 21:49:25 +02:00
d . addCallback ( lambda _ : self . _render_response ( True , OK_CODE ) )
2016-06-28 07:51:05 +02:00
return d
2016-07-20 17:44:06 +02:00
2016-08-03 09:16:06 +02:00
def jsonrpc_get_peers_for_hash ( self , p ) :
"""
Get peers for blob hash
Args :
' blob_hash ' : blob hash
Returns :
List of contacts
"""
blob_hash = p [ ' blob_hash ' ]
d = self . session . peer_finder . find_peers_for_blob ( blob_hash )
d . addCallback ( lambda r : [ [ c . host , c . port , c . is_available ( ) ] for c in r ] )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
return d
2016-08-04 08:22:10 +02:00
2016-07-20 17:44:06 +02:00
def get_lbrynet_version_from_github ( ) :
""" Return the latest released version from github. """
response = requests . get ( ' https://api.github.com/repos/lbryio/lbry/releases/latest ' )
release = response . json ( )
tag = release [ ' tag_name ' ]
# githubs documentation claims this should never happen, but we'll check just in case
if release [ ' prerelease ' ] :
raise Exception ( ' Release {} is a pre-release ' . format ( tag ) )
return get_version_from_tag ( tag )
def get_version_from_tag ( tag ) :
match = re . match ( ' v([ \ d.]+) ' , tag )
if match :
return match . group ( 1 )
else :
raise Exception ( ' Failed to parse version from tag {} ' . format ( tag ) )
2016-07-21 17:44:59 +02:00
def get_sd_hash ( stream_info ) :
if not stream_info :
return None
try :
return stream_info [ ' sources ' ] [ ' lbry_sd_hash ' ]
except KeyError :
return stream_info . get ( ' stream_hash ' )
def get_output_callback ( params ) :
def callback ( l ) :
return {
' stream_hash ' : params . sd_hash if params . stream_info else l . sd_hash ,
' path ' : os . path . join ( params . download_directory , l . file_name )
}
return callback
class _DownloadNameHelper ( object ) :
def __init__ ( self , daemon , name , timeout = DEFAULT_TIMEOUT , download_directory = None ,
file_name = None , wait_for_write = True ) :
self . daemon = daemon
self . name = name
self . timeout = timeout
if not download_directory or not os . path . isdir ( download_directory ) :
self . download_directory = daemon . download_directory
else :
self . download_directory = download_directory
self . file_name = file_name
self . wait_for_write = wait_for_write
def _setup_stream ( self , stream_info ) :
stream_hash = get_sd_hash ( stream_info )
d = self . daemon . _get_lbry_file_by_sd_hash ( stream_hash )
d . addCallback ( self . _add_results_callback ( stream_info ) )
return d
def _add_results_callback ( self , stream_info ) :
def add_results ( l ) :
if l :
if os . path . isfile ( os . path . join ( self . download_directory , l . file_name ) ) :
return defer . succeed ( ( stream_info , l ) )
return defer . succeed ( ( stream_info , None ) )
return add_results
2016-07-21 18:05:06 +02:00
2016-07-21 17:44:59 +02:00
def wait_or_get_stream ( self , args ) :
stream_info , lbry_file = args
if lbry_file :
return self . _wait_on_lbry_file ( lbry_file )
2016-07-21 18:05:06 +02:00
else :
return self . _get_stream ( stream_info )
2016-07-21 17:44:59 +02:00
def _get_stream ( self , stream_info ) :
d = self . daemon . add_stream (
self . name , self . timeout , self . download_directory , self . file_name , stream_info )
if self . wait_for_write :
d . addCallback ( lambda _ : self . _wait_for_write ( ) )
d . addCallback ( lambda _ : self . daemon . streams [ self . name ] . downloader )
return d
def _wait_for_write ( self ) :
d = defer . succeed ( None )
2016-08-05 12:24:48 +02:00
if not self . has_downloader_wrote ( ) :
2016-07-21 17:44:59 +02:00
d . addCallback ( lambda _ : reactor . callLater ( 1 , self . _wait_for_write ) )
return d
2016-08-05 12:24:48 +02:00
def has_downloader_wrote ( self ) :
downloader = self . daemon . streams [ self . name ] . downloader
if not downloader :
return False
return self . get_written_bytes ( downloader . file_name )
2016-07-21 17:44:59 +02:00
def _wait_on_lbry_file ( self , f ) :
2016-07-21 18:05:06 +02:00
written_bytes = self . get_written_bytes ( f . file_name )
if written_bytes :
return defer . succeed ( self . _disp_file ( f ) )
2016-07-21 17:44:59 +02:00
d = defer . succeed ( None )
d . addCallback ( lambda _ : reactor . callLater ( 1 , self . _wait_on_lbry_file , f ) )
return d
def get_written_bytes ( self , file_name ) :
2016-08-05 12:24:48 +02:00
""" Returns the number of bytes written to `file_name`.
Returns False if there were issues reading ` file_name ` .
"""
2016-07-21 17:44:59 +02:00
try :
file_path = os . path . join ( self . download_directory , file_name )
if os . path . isfile ( file_path ) :
written_file = file ( file_path )
written_file . seek ( 0 , os . SEEK_END )
written_bytes = written_file . tell ( )
written_file . close ( )
else :
written_bytes = False
except Exception :
writen_bytes = False
return written_bytes
def _disp_file ( self , f ) :
file_path = os . path . join ( self . download_directory , f . file_name )
2016-07-21 18:05:06 +02:00
log . info ( " Already downloaded: %s --> %s " , f . sd_hash , file_path )
2016-07-21 17:44:59 +02:00
return f
def _remove_from_wait ( self , r ) :
del self . daemon . waiting_on [ self . name ]
return r
class _ResolveNameHelper ( object ) :
def __init__ ( self , daemon , name , force_refresh ) :
self . daemon = daemon
self . name = name
self . force_refresh = force_refresh
def get_deferred ( self ) :
if self . need_fresh_stream ( ) :
log . info ( " Resolving stream info for lbry:// %s " , self . name )
d = self . wallet . get_stream_info_for_name ( self . name )
d . addCallbacks ( self . _cache_stream_info , lambda _ : defer . fail ( UnknownNameError ) )
else :
2016-08-08 20:58:31 +02:00
log . debug ( " Returning cached stream info for lbry:// %s " , self . name )
2016-07-21 17:44:59 +02:00
d = defer . succeed ( self . name_data [ ' claim_metadata ' ] )
return d
@property
def name_data ( self ) :
return self . daemon . name_cache [ self . name ]
@property
def wallet ( self ) :
return self . daemon . session . wallet
def now ( self ) :
return self . daemon . _get_long_count_timestamp ( )
def _add_txid ( self , txid ) :
self . name_data [ ' txid ' ] = txid
return defer . succeed ( None )
def _cache_stream_info ( self , stream_info ) :
self . daemon . name_cache [ self . name ] = {
' claim_metadata ' : stream_info ,
' timestamp ' : self . now ( )
}
d = self . wallet . get_txid_for_name ( self . name )
d . addCallback ( self . _add_txid )
d . addCallback ( lambda _ : self . daemon . _update_claim_cache ( ) )
d . addCallback ( lambda _ : self . name_data [ ' claim_metadata ' ] )
return d
def need_fresh_stream ( self ) :
return self . force_refresh or not self . is_in_cache ( ) or self . is_cached_name_expired ( )
def is_in_cache ( self ) :
return self . name in self . daemon . name_cache
def is_cached_name_expired ( self ) :
time_in_cache = self . now ( ) - self . name_data [ ' timestamp ' ]
return time_in_cache > = self . daemon . cache_time