2016-03-03 00:32:48 +01:00
import locale
2016-02-29 19:25:47 +01:00
import os
import sys
2016-03-24 03:27:48 +01:00
import simplejson as json
2016-02-29 19:25:47 +01:00
import binascii
import subprocess
import logging
2016-04-07 09:12:09 +02:00
import logging . handlers
2016-03-03 00:32:48 +01:00
import requests
2016-04-07 09:12:09 +02:00
import base64
import base58
import platform
import json
2016-03-24 03:27:48 +01:00
from twisted . web import server , resource , static
from twisted . internet import defer , threads , error , reactor
from txjsonrpc import jsonrpclib
from txjsonrpc . web import jsonrpc
2016-03-29 22:42:47 +02:00
from txjsonrpc . web . jsonrpc import Handler
2016-02-29 19:25:47 +01:00
from datetime import datetime
from decimal import Decimal
2016-03-24 03:27:48 +01:00
from appdirs import user_data_dir
2016-04-07 09:12:09 +02:00
from urllib2 import urlopen
2016-02-29 19:25:47 +01:00
2016-04-12 04:28:46 +02:00
from lbrynet import __version__ as lbrynet_version
from lbryum . version import ELECTRUM_VERSION as lbryum_version
2016-01-26 02:28:05 +01:00
from lbrynet . core . PaymentRateManager import PaymentRateManager
from lbrynet . core . server . BlobAvailabilityHandler import BlobAvailabilityHandlerFactory
from lbrynet . core . server . BlobRequestHandler import BlobRequestHandlerFactory
from lbrynet . core . server . ServerProtocol import ServerProtocolFactory
from lbrynet . lbrynet_console . ControlHandlers import get_time_behind_blockchain
2016-01-12 18:42:32 +01:00
from lbrynet . core . Error import UnknownNameError
2015-12-06 23:32:17 +01:00
from lbrynet . lbryfile . StreamDescriptor import LBRYFileStreamType
from lbrynet . lbryfile . client . LBRYFileDownloader import LBRYFileSaverFactory , LBRYFileOpenerFactory
from lbrynet . lbryfile . client . LBRYFileOptions import add_lbry_file_to_sd_identifier
2015-12-14 21:54:33 +01:00
from lbrynet . lbrynet_daemon . LBRYDownloader import GetStream , FetcherDaemon
2016-01-21 04:00:28 +01:00
from lbrynet . lbrynet_daemon . LBRYPublisher import Publisher
2015-12-06 23:32:17 +01:00
from lbrynet . core . utils import generate_id
from lbrynet . lbrynet_console . LBRYSettings import LBRYSettings
2016-02-15 16:12:49 +01:00
from lbrynet . conf import MIN_BLOB_DATA_PAYMENT_RATE , DEFAULT_MAX_SEARCH_RESULTS , KNOWN_DHT_NODES , DEFAULT_MAX_KEY_FEE
2016-03-25 14:32:24 +01:00
from lbrynet . conf import API_CONNECTION_STRING , API_PORT , API_ADDRESS , DEFAULT_TIMEOUT , UI_ADDRESS
2016-01-24 06:10:22 +01:00
from lbrynet . core . StreamDescriptor import StreamDescriptorIdentifier , download_sd_blob
2015-12-06 23:32:17 +01:00
from lbrynet . core . Session import LBRYSession
from lbrynet . core . PTCWallet import PTCWallet
2016-02-23 04:32:07 +01:00
from lbrynet . core . LBRYcrdWallet import LBRYcrdWallet , LBRYumWallet
2015-12-06 23:32:17 +01:00
from lbrynet . lbryfilemanager . LBRYFileManager import LBRYFileManager
from lbrynet . lbryfile . LBRYFileMetadataManager import DBLBRYFileMetadataManager , TempLBRYFileMetadataManager
2016-02-23 04:32:07 +01:00
2016-04-07 09:12:09 +02:00
if sys . platform != " darwin " :
log_dir = os . path . join ( os . path . expanduser ( " ~ " ) , " .lbrynet " )
else :
log_dir = user_data_dir ( " LBRY " )
if not os . path . isdir ( log_dir ) :
os . mkdir ( log_dir )
LOG_FILENAME = os . path . join ( log_dir , ' lbrynet-daemon.log ' )
2016-04-18 01:54:04 +02:00
if os . path . isfile ( LOG_FILENAME ) :
f = open ( LOG_FILENAME , ' r ' )
PREVIOUS_LOG = len ( f . read ( ) )
f . close ( )
else :
PREVIOUS_LOG = 0
2015-12-14 16:22:52 +01:00
log = logging . getLogger ( __name__ )
2016-04-07 09:12:09 +02:00
handler = logging . handlers . RotatingFileHandler ( LOG_FILENAME , maxBytes = 262144 , backupCount = 5 )
log . addHandler ( handler )
2016-04-16 01:37:27 +02:00
log . setLevel ( logging . INFO )
2016-04-07 09:12:09 +02:00
2016-04-09 08:09:30 +02:00
STARTUP_STAGES = [
( ' initializing ' , ' Initializing... ' ) ,
( ' loading_db ' , ' Loading databases... ' ) ,
2016-04-14 22:50:11 +02:00
( ' loading_wallet ' , ' Catching up with the blockchain... %s ' ) ,
2016-04-09 08:09:30 +02:00
( ' loading_file_manager ' , ' Setting up file manager ' ) ,
( ' loading_server ' , ' Starting lbrynet ' ) ,
( ' started ' , ' Started lbrynet ' )
]
2016-04-13 20:47:34 +02:00
ALLOWED_DURING_STARTUP = [ ' is_running ' , ' is_first_run ' ,
' get_time_behind_blockchain ' , ' stop ' ,
' daemon_status ' , ' get_start_notice ' ,
' version ' , ' check_for_new_version ' ]
2016-04-09 08:09:30 +02:00
2016-03-24 03:27:48 +01:00
BAD_REQUEST = 400
NOT_FOUND = 404
OK_CODE = 200
2016-02-16 19:39:08 +01:00
# TODO add login credentials in a conf file
# TODO alert if your copy of a lbry file is out of date with the name record
2016-01-12 00:01:46 +01:00
2015-12-06 23:32:17 +01:00
2016-03-24 03:27:48 +01:00
class LBRYDaemon ( jsonrpc . JSONRPC ) :
2015-12-06 23:32:17 +01:00
"""
2016-03-24 03:27:48 +01:00
LBRYnet daemon , a jsonrpc interface to lbry functions
2015-12-06 23:32:17 +01:00
"""
2016-04-12 04:28:46 +02:00
2016-03-24 03:27:48 +01:00
isLeaf = True
2016-04-12 08:03:57 +02:00
def __init__ ( self , ui_version_info , wallet_type = " lbryum " ) :
2016-04-12 04:28:46 +02:00
jsonrpc . JSONRPC . __init__ ( self )
reactor . addSystemEventTrigger ( ' before ' , ' shutdown ' , self . _shutdown )
2016-04-12 08:03:57 +02:00
self . startup_status = STARTUP_STAGES [ 0 ]
self . startup_message = None
self . announced_startup = False
self . query_handlers = { }
self . ui_version = ui_version_info . replace ( ' \n ' , ' ' )
2016-04-14 04:07:27 +02:00
self . git_lbrynet_version = None
self . git_lbryum_version = None
2016-04-12 08:03:57 +02:00
self . wallet_type = wallet_type
self . session_settings = None
self . first_run = None
2016-04-12 04:28:46 +02:00
self . log_file = LOG_FILENAME
self . fetcher = None
self . current_db_revision = 1
self . run_server = True
self . session = None
self . known_dht_nodes = KNOWN_DHT_NODES
2016-04-16 04:31:06 +02:00
self . platform_info = {
" processor " : platform . processor ( ) ,
" python_version: " : platform . python_version ( ) ,
" platform " : platform . platform ( ) ,
" os_release " : platform . release ( ) ,
" os_system " : platform . system ( ) ,
" lbrynet_version: " : lbrynet_version ,
" lbryum_version: " : lbryum_version ,
" ui_version " : self . ui_version ,
}
2016-04-18 01:54:04 +02:00
try :
self . platform_info [ ' ip ' ] = json . load ( urlopen ( ' http://jsonip.com ' ) ) [ ' ip ' ]
except :
self . platform_info [ ' ip ' ] = " Could not determine "
2016-04-12 08:03:57 +02:00
2016-04-12 04:28:46 +02:00
if os . name == " nt " :
from lbrynet . winhelpers . knownpaths import get_path , FOLDERID , UserHandle
self . download_directory = get_path ( FOLDERID . Downloads , UserHandle . current )
2016-04-12 08:03:57 +02:00
self . db_dir = os . path . join ( get_path ( FOLDERID . RoamingAppData , UserHandle . current ) , " lbrynet " )
self . lbrycrdd_path = " lbrycrdd.exe "
if wallet_type == " lbrycrd " :
self . wallet_dir = os . path . join ( get_path ( FOLDERID . RoamingAppData , UserHandle . current ) , " lbrycrd " )
else :
self . wallet_dir = os . path . join ( get_path ( FOLDERID . RoamingAppData , UserHandle . current ) , " lbryum " )
2016-04-12 04:28:46 +02:00
elif sys . platform == " darwin " :
self . download_directory = os . path . join ( os . path . expanduser ( " ~ " ) , ' Downloads ' )
2016-04-12 08:03:57 +02:00
self . db_dir = user_data_dir ( " LBRY " )
self . lbrycrdd_path = " ./lbrycrdd "
2016-04-12 04:28:46 +02:00
if wallet_type == " lbrycrd " :
self . wallet_dir = user_data_dir ( " lbrycrd " )
else :
self . wallet_dir = user_data_dir ( " LBRY " )
else :
self . download_directory = os . getcwd ( )
2016-04-12 08:03:57 +02:00
self . db_dir = os . path . join ( os . path . expanduser ( " ~ " ) , " .lbrynet " )
self . lbrycrdd_path = " ./lbrycrdd "
2016-04-12 04:28:46 +02:00
if wallet_type == " lbrycrd " :
self . wallet_dir = os . path . join ( os . path . expanduser ( " ~ " ) , " .lbrycrd " )
else :
self . wallet_dir = os . path . join ( os . path . expanduser ( " ~ " ) , " .lbryum " )
2016-04-12 08:03:57 +02:00
self . created_data_dir = False
if not os . path . exists ( self . db_dir ) :
os . mkdir ( self . db_dir )
self . created_data_dir = True
self . blobfile_dir = os . path . join ( self . db_dir , " blobfiles " )
self . lbrycrd_conf = os . path . join ( self . wallet_dir , " lbrycrd.conf " )
self . autofetcher_conf = os . path . join ( self . wallet_dir , " autofetcher.conf " )
2016-04-12 04:28:46 +02:00
self . daemon_conf = os . path . join ( self . db_dir , ' daemon_settings.json ' )
self . wallet_conf = os . path . join ( self . wallet_dir , " lbrycrd.conf " )
self . wallet_user = None
self . wallet_password = None
2016-04-12 08:03:57 +02:00
2016-04-12 04:28:46 +02:00
self . sd_identifier = StreamDescriptorIdentifier ( )
self . stream_info_manager = TempLBRYFileMetadataManager ( )
2016-04-12 08:03:57 +02:00
self . settings = LBRYSettings ( self . db_dir )
2016-04-12 04:28:46 +02:00
self . blob_request_payment_rate_manager = None
self . lbry_file_metadata_manager = None
self . lbry_file_manager = None
2016-04-12 08:03:57 +02:00
#defaults for settings otherwise loaded from daemon_settings.json
2016-04-12 04:28:46 +02:00
self . default_settings = {
' run_on_startup ' : False ,
' data_rate ' : MIN_BLOB_DATA_PAYMENT_RATE ,
2016-04-12 08:03:57 +02:00
' max_key_fee ' : DEFAULT_MAX_KEY_FEE ,
2016-04-12 04:28:46 +02:00
' default_download_directory ' : self . download_directory ,
' max_upload ' : 0.0 ,
' max_download ' : 0.0 ,
2016-04-12 08:03:57 +02:00
' upload_log ' : True ,
' search_timeout ' : 3.0 ,
' max_search_results ' : DEFAULT_MAX_SEARCH_RESULTS ,
' wallet_type ' : wallet_type ,
' delete_blobs_on_remove ' : True ,
' peer_port ' : 3333 ,
' dht_node_port ' : 4444 ,
' use_upnp ' : True ,
' start_lbrycrdd ' : True ,
2016-04-18 01:54:04 +02:00
' requested_first_run_credits ' : False
2016-04-12 04:28:46 +02:00
}
2016-03-24 03:27:48 +01:00
def render ( self , request ) :
request . content . seek ( 0 , 0 )
# Unmarshal the JSON-RPC data.
content = request . content . read ( )
parsed = jsonrpclib . loads ( content )
functionPath = parsed . get ( " method " )
args = parsed . get ( ' params ' )
2016-04-12 04:28:46 +02:00
#TODO convert args to correct types if possible
2016-03-24 03:27:48 +01:00
id = parsed . get ( ' id ' )
version = parsed . get ( ' jsonrpc ' )
if version :
version = int ( float ( version ) )
elif id and not version :
version = jsonrpclib . VERSION_1
else :
version = jsonrpclib . VERSION_PRE1
# XXX this all needs to be re-worked to support logic for multiple
# versions...
2016-03-29 22:42:47 +02:00
if not self . announced_startup :
2016-04-13 20:47:34 +02:00
if functionPath not in ALLOWED_DURING_STARTUP :
2016-04-02 04:11:11 +02:00
return server . failure
2016-03-29 22:42:47 +02:00
2016-03-24 03:27:48 +01:00
try :
function = self . _getFunction ( functionPath )
except jsonrpclib . Fault , f :
self . _cbRender ( f , request , id , version )
else :
2016-03-28 22:49:51 +02:00
request . setHeader ( " Access-Control-Allow-Origin " , " * " )
2016-03-24 03:27:48 +01:00
request . setHeader ( " content-type " , " text/json " )
2016-04-14 05:10:17 +02:00
if args == [ { } ] :
2016-03-24 03:27:48 +01:00
d = defer . maybeDeferred ( function )
else :
d = defer . maybeDeferred ( function , * args )
d . addErrback ( self . _ebRender , id )
d . addCallback ( self . _cbRender , request , id , version )
return server . NOT_DONE_YET
2015-12-06 23:32:17 +01:00
2016-03-29 22:42:47 +02:00
def _cbRender ( self , result , request , id , version ) :
if isinstance ( result , Handler ) :
result = result . result
if isinstance ( result , dict ) :
result = result [ ' result ' ]
if version == jsonrpclib . VERSION_PRE1 :
if not isinstance ( result , jsonrpclib . Fault ) :
result = ( result , )
# Convert the result (python) to JSON-RPC
try :
s = jsonrpclib . dumps ( result , version = version )
except :
f = jsonrpclib . Fault ( self . FAILURE , " can ' t serialize output " )
s = jsonrpclib . dumps ( f , version = version )
request . setHeader ( " content-length " , str ( len ( s ) ) )
request . write ( s )
request . finish ( )
def _ebRender ( self , failure , id ) :
if isinstance ( failure . value , jsonrpclib . Fault ) :
return failure . value
log . error ( failure )
return jsonrpclib . Fault ( self . FAILURE , " error " )
2016-04-12 04:28:46 +02:00
def setup ( self ) :
2016-04-07 09:12:09 +02:00
def _log_starting_vals ( ) :
def _get_lbry_files_json ( ) :
2016-04-14 04:07:27 +02:00
r = self . _get_lbry_files ( )
2016-04-07 09:12:09 +02:00
return json . dumps ( r )
2016-04-14 04:07:27 +02:00
def _get_lbryum_version ( ) :
r = urlopen ( " https://raw.githubusercontent.com/lbryio/lbryum/master/lib/version.py " ) . read ( ) . split ( ' \n ' )
version = next ( line . split ( " = " ) [ 1 ] . split ( " # " ) [ 0 ] . replace ( " " , " " )
for line in r if " ELECTRUM_VERSION " in line )
version = version . replace ( " ' " , " " )
log . info ( " remote lbryum " + str ( version ) + " > local lbryum " + str ( lbryum_version ) + " = " + str (
version > lbryum_version ) )
return version
def _get_lbrynet_version ( ) :
2016-04-18 01:54:04 +02:00
r = urlopen ( " https://raw.githubusercontent.com/lbryio/lbry/master/lbrynet/__init__.py " ) . read ( ) . split ( ' \n ' )
2016-04-14 04:07:27 +02:00
vs = next ( i for i in r if ' version = ' in i ) . split ( " = " ) [ 1 ] . replace ( " " , " " )
vt = tuple ( int ( x ) for x in vs [ 1 : - 1 ] . split ( ' , ' ) )
vr = " . " . join ( [ str ( x ) for x in vt ] )
log . info ( " remote lbrynet " + str ( vr ) + " > local lbrynet " + str ( lbrynet_version ) + " = " + str (
vr > lbrynet_version ) )
return vr
self . git_lbrynet_version = _get_lbrynet_version ( )
self . git_lbryum_version = _get_lbryum_version ( )
2016-04-07 09:12:09 +02:00
log . info ( " LBRY Files: " + _get_lbry_files_json ( ) )
log . info ( " Starting balance: " + str ( self . session . wallet . wallet_balance ) )
2016-03-29 22:42:47 +02:00
return defer . succeed ( None )
2016-04-07 09:12:09 +02:00
def _announce_startup ( ) :
2016-04-18 01:54:04 +02:00
def _announce ( ) :
self . announced_startup = True
self . startup_status = STARTUP_STAGES [ 5 ]
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Started lbrynet-daemon " )
if self . first_run :
d = self . _upload_log ( name_prefix = " fr " )
else :
d = self . _upload_log ( exclude_previous = True , name_prefix = " start " )
if float ( self . session . wallet . wallet_balance ) == 0.0 :
d . addCallback ( lambda _ : self . _check_first_run ( ) )
d . addCallback ( self . _show_first_run_result )
d . addCallback ( lambda _ : _announce ( ) )
return d
2016-01-11 21:52:38 +01:00
2016-03-24 03:27:48 +01:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Starting lbrynet-daemon " )
2016-02-25 23:17:07 +01:00
2015-12-06 23:32:17 +01:00
d = defer . Deferred ( )
2016-04-07 09:12:09 +02:00
d . addCallback ( lambda _ : self . _initial_setup ( ) )
d . addCallback ( self . _set_daemon_settings )
2015-12-14 16:22:52 +01:00
d . addCallback ( lambda _ : threads . deferToThread ( self . _setup_data_directory ) )
d . addCallback ( lambda _ : self . _check_db_migration ( ) )
2015-12-06 23:32:17 +01:00
d . addCallback ( lambda _ : self . _get_settings ( ) )
2016-01-20 03:07:16 +01:00
d . addCallback ( lambda _ : self . _get_lbrycrdd_path ( ) )
2015-12-06 23:32:17 +01:00
d . addCallback ( lambda _ : self . _get_session ( ) )
d . addCallback ( lambda _ : add_lbry_file_to_sd_identifier ( self . sd_identifier ) )
d . addCallback ( lambda _ : self . _setup_stream_identifier ( ) )
d . addCallback ( lambda _ : self . _setup_lbry_file_manager ( ) )
d . addCallback ( lambda _ : self . _setup_lbry_file_opener ( ) )
2016-01-26 02:28:05 +01:00
d . addCallback ( lambda _ : self . _setup_query_handlers ( ) )
d . addCallback ( lambda _ : self . _setup_server ( ) )
2015-12-14 18:59:44 +01:00
d . addCallback ( lambda _ : self . _setup_fetcher ( ) )
2016-04-07 09:12:09 +02:00
d . addCallback ( lambda _ : _log_starting_vals ( ) )
2016-03-29 22:42:47 +02:00
d . addCallback ( lambda _ : _announce_startup ( ) )
2015-12-06 23:32:17 +01:00
d . callback ( None )
return defer . succeed ( None )
2016-03-24 03:27:48 +01:00
def _initial_setup ( self ) :
2016-04-07 09:12:09 +02:00
def _log_platform ( ) :
2016-04-16 04:31:06 +02:00
log . info ( " Platform: " + json . dumps ( self . platform_info ) )
2016-04-07 09:12:09 +02:00
return defer . succeed ( None )
def _load_daemon_conf ( ) :
if os . path . isfile ( self . daemon_conf ) :
2016-04-18 01:54:04 +02:00
f = open ( self . daemon_conf , " r " )
loaded_settings = json . loads ( f . read ( ) )
f . close ( )
missing_settings = { }
for k in self . default_settings . keys ( ) :
if k not in loaded_settings . keys ( ) :
missing_settings [ k ] = self . default_settings [ k ]
if missing_settings != { } :
for k in missing_settings . keys ( ) :
log . info ( " Adding missing setting: " + k + " with default value: " + str ( missing_settings [ k ] ) )
loaded_settings [ k ] = missing_settings [ k ]
f = open ( self . daemon_conf , " w " )
f . write ( json . dumps ( loaded_settings ) )
f . close ( )
rsettings = loaded_settings
2016-04-07 09:12:09 +02:00
else :
log . info ( " Writing default settings : " + json . dumps ( self . default_settings ) + " --> " + str ( self . daemon_conf ) )
f = open ( self . daemon_conf , " w " )
f . write ( json . dumps ( self . default_settings ) )
f . close ( )
2016-04-18 01:54:04 +02:00
rsettings = self . default_settings
for k in rsettings . keys ( ) :
self . __dict__ [ k ] = rsettings [ k ]
return rsettings
2016-02-16 19:39:08 +01:00
2016-04-07 09:12:09 +02:00
d = _log_platform ( )
d . addCallback ( lambda _ : _load_daemon_conf ( ) )
return d
def _set_daemon_settings ( self , settings ) :
self . session_settings = settings
2016-02-16 19:39:08 +01:00
return defer . succeed ( None )
2016-01-26 02:28:05 +01:00
def _start_server ( self ) :
if self . peer_port is not None :
server_factory = ServerProtocolFactory ( self . session . rate_limiter ,
self . query_handlers ,
self . session . peer_manager )
try :
self . lbry_server_port = reactor . listenTCP ( self . peer_port , server_factory )
except error . CannotListenError as e :
import traceback
log . error ( " Couldn ' t bind to port %d . %s " , self . peer_port , traceback . format_exc ( ) )
raise ValueError ( " %s lbrynet may already be running on your computer. " , str ( e ) )
return defer . succeed ( True )
def _stop_server ( self ) :
if self . lbry_server_port is not None :
self . lbry_server_port , p = None , self . lbry_server_port
return defer . maybeDeferred ( p . stopListening )
else :
return defer . succeed ( True )
def _setup_server ( self ) :
def restore_running_status ( running ) :
if running is True :
return self . _start_server ( )
return defer . succeed ( True )
2016-04-12 04:28:46 +02:00
self . startup_status = STARTUP_STAGES [ 4 ]
2016-04-09 07:33:27 +02:00
2016-01-26 02:28:05 +01:00
dl = self . settings . get_server_running_status ( )
dl . addCallback ( restore_running_status )
return dl
def _setup_query_handlers ( self ) :
handlers = [
2016-02-16 19:39:08 +01:00
# CryptBlobInfoQueryHandlerFactory(self.lbry_file_metadata_manager, self.session.wallet,
2016-01-26 02:28:05 +01:00
# self._server_payment_rate_manager),
BlobAvailabilityHandlerFactory ( self . session . blob_manager ) ,
2016-02-16 19:39:08 +01:00
# BlobRequestHandlerFactory(self.session.blob_manager, self.session.wallet,
2016-01-26 02:28:05 +01:00
# self._server_payment_rate_manager),
self . session . wallet . get_wallet_info_query_handler_factory ( ) ,
]
def get_blob_request_handler_factory ( rate ) :
self . blob_request_payment_rate_manager = PaymentRateManager (
self . session . base_payment_rate_manager , rate
)
handlers . append ( BlobRequestHandlerFactory ( self . session . blob_manager , self . session . wallet ,
self . blob_request_payment_rate_manager ) )
d1 = self . settings . get_server_data_payment_rate ( )
d1 . addCallback ( get_blob_request_handler_factory )
dl = defer . DeferredList ( [ d1 ] )
dl . addCallback ( lambda _ : self . _add_query_handlers ( handlers ) )
return dl
def _add_query_handlers ( self , query_handlers ) :
def _set_query_handlers ( statuses ) :
from future_builtins import zip
for handler , ( success , status ) in zip ( query_handlers , statuses ) :
if success is True :
self . query_handlers [ handler ] = status
ds = [ ]
for handler in query_handlers :
ds . append ( self . settings . get_query_handler_status ( handler . get_primary_query_identifier ( ) ) )
dl = defer . DeferredList ( ds )
dl . addCallback ( _set_query_handlers )
return dl
2016-04-18 01:54:04 +02:00
def _upload_log ( self , name_prefix = None , exclude_previous = False ) :
if name_prefix :
name_prefix = name_prefix + " - " + platform . system ( )
else :
name_prefix = platform . system ( )
2016-04-07 09:12:09 +02:00
if self . session_settings [ ' upload_log ' ] :
LOG_URL = " https://lbry.io/log-upload "
2016-04-18 01:54:04 +02:00
if exclude_previous :
f = open ( self . log_file , " r " )
f . seek ( PREVIOUS_LOG )
log_contents = f . read ( )
f . close ( )
else :
f = open ( self . log_file , " r " )
log_contents = f . read ( )
f . close ( )
2016-04-07 09:12:09 +02:00
t = datetime . now ( )
2016-04-18 01:54:04 +02:00
log_name = name_prefix + " - " + base58 . b58encode ( self . lbryid ) [ : 20 ] + " - " + str ( t . month ) + " - " + str ( t . day ) + " - " + str ( t . year ) + " - " + str ( t . hour ) + " - " + str ( t . minute )
params = { ' name ' : log_name , ' log ' : log_contents }
2016-04-07 09:12:09 +02:00
requests . post ( LOG_URL , params )
return defer . succeed ( None )
2016-01-11 21:52:38 +01:00
def _shutdown ( self ) :
2016-03-24 03:27:48 +01:00
log . info ( " Closing lbrynet session " )
2016-04-18 01:54:04 +02:00
log . info ( " Status at time of shutdown: " + self . startup_status [ 0 ] )
d = self . _upload_log ( name_prefix = " close " , exclude_previous = False if self . first_run else True )
2016-04-07 09:12:09 +02:00
d . addCallback ( lambda _ : self . _stop_server ( ) )
d . addErrback ( lambda err : log . info ( " Bad server shutdown: " + err . getTraceback ( ) ) )
2016-01-11 21:52:38 +01:00
if self . session is not None :
2016-01-26 02:28:05 +01:00
d . addCallback ( lambda _ : self . session . shut_down ( ) )
2016-04-07 09:12:09 +02:00
d . addErrback ( lambda err : log . info ( " Bad session shutdown: " + err . getTraceback ( ) ) )
2016-01-11 21:52:38 +01:00
return d
2016-03-24 03:27:48 +01:00
def _update_settings ( self , settings ) :
2016-04-07 09:12:09 +02:00
for k in settings . keys ( ) :
if k == ' run_on_startup ' :
if type ( settings [ ' run_on_startup ' ] ) is bool :
self . session_settings [ ' run_on_startup ' ] = settings [ ' run_on_startup ' ]
else :
return defer . fail ( )
elif k == ' data_rate ' :
if type ( settings [ ' data_rate ' ] ) is float :
self . session_settings [ ' data_rate ' ] = settings [ ' data_rate ' ]
2016-04-09 05:38:57 +02:00
elif type ( settings [ ' data_rate ' ] ) is int :
self . session_settings [ ' data_rate ' ] = float ( settings [ ' data_rate ' ] )
2016-04-07 09:12:09 +02:00
else :
return defer . fail ( )
elif k == ' max_key_fee ' :
if type ( settings [ ' max_key_fee ' ] ) is float :
self . session_settings [ ' max_key_fee ' ] = settings [ ' max_key_fee ' ]
2016-04-09 05:38:57 +02:00
elif type ( settings [ ' max_key_fee ' ] ) is int :
self . session_settings [ ' max_key_fee ' ] = float ( settings [ ' max_key_fee ' ] )
2016-04-07 09:12:09 +02:00
else :
return defer . fail ( )
elif k == ' default_download_directory ' :
if type ( settings [ ' default_download_directory ' ] ) is unicode :
if os . path . isdir ( settings [ ' default_download_directory ' ] ) :
self . session_settings [ ' default_download_directory ' ] = settings [ ' default_download_directory ' ]
else :
pass
else :
return defer . fail ( )
elif k == ' max_upload ' :
if type ( settings [ ' max_upload ' ] ) is float :
self . session_settings [ ' max_upload ' ] = settings [ ' max_upload ' ]
2016-04-09 05:38:57 +02:00
elif type ( settings [ ' max_upload ' ] ) is int :
self . session_settings [ ' max_upload ' ] = float ( settings [ ' max_upload ' ] )
2016-04-07 09:12:09 +02:00
else :
return defer . fail ( )
elif k == ' max_download ' :
if type ( settings [ ' max_download ' ] ) is float :
self . session_settings [ ' max_download ' ] = settings [ ' max_download ' ]
2016-04-09 05:38:57 +02:00
if type ( settings [ ' max_download ' ] ) is int :
self . session_settings [ ' max_download ' ] = float ( settings [ ' max_download ' ] )
2016-04-07 09:12:09 +02:00
else :
return defer . fail ( )
elif k == ' upload_log ' :
if type ( settings [ ' upload_log ' ] ) is bool :
self . session_settings [ ' upload_log ' ] = settings [ ' upload_log ' ]
else :
return defer . fail ( )
f = open ( self . daemon_conf , " w " )
f . write ( json . dumps ( self . session_settings ) )
f . close ( )
2016-03-24 03:27:48 +01:00
return defer . succeed ( True )
2016-01-07 16:24:55 +01:00
2015-12-14 18:59:44 +01:00
def _setup_fetcher ( self ) :
self . fetcher = FetcherDaemon ( self . session , self . lbry_file_manager , self . lbry_file_metadata_manager ,
self . session . wallet , self . sd_identifier , self . autofetcher_conf )
return defer . succeed ( None )
2015-12-14 16:22:52 +01:00
def _setup_data_directory ( self ) :
2016-04-09 08:09:30 +02:00
self . startup_status = STARTUP_STAGES [ 1 ]
2016-03-24 03:27:48 +01:00
log . info ( " Loading databases... " )
2015-12-14 16:22:52 +01:00
if self . created_data_dir :
db_revision = open ( os . path . join ( self . db_dir , " db_revision " ) , mode = ' w ' )
db_revision . write ( str ( self . current_db_revision ) )
db_revision . close ( )
log . debug ( " Created the db revision file: %s " , str ( os . path . join ( self . db_dir , " db_revision " ) ) )
if not os . path . exists ( self . blobfile_dir ) :
os . mkdir ( self . blobfile_dir )
log . debug ( " Created the blobfile directory: %s " , str ( self . blobfile_dir ) )
def _check_db_migration ( self ) :
old_revision = 0
db_revision_file = os . path . join ( self . db_dir , " db_revision " )
if os . path . exists ( db_revision_file ) :
old_revision = int ( open ( db_revision_file ) . read ( ) . strip ( ) )
if old_revision < self . current_db_revision :
from lbrynet . db_migrator import dbmigrator
2016-03-24 03:27:48 +01:00
log . info ( " Upgrading your databases... " )
2015-12-14 16:22:52 +01:00
d = threads . deferToThread ( dbmigrator . migrate_db , self . db_dir , old_revision , self . current_db_revision )
def print_success ( old_dirs ) :
success_string = " Finished upgrading the databases. It is now safe to delete the "
success_string + = " following directories, if you feel like it. It won ' t make any "
success_string + = " difference. \n Anyway here they are: "
for i , old_dir in enumerate ( old_dirs ) :
success_string + = old_dir
if i + 1 < len ( old_dir ) :
success_string + = " , "
2016-03-24 03:27:48 +01:00
log . info ( success_string )
2015-12-14 16:22:52 +01:00
d . addCallback ( print_success )
return d
return defer . succeed ( True )
2015-12-06 23:32:17 +01:00
def _get_settings ( self ) :
d = self . settings . start ( )
d . addCallback ( lambda _ : self . settings . get_lbryid ( ) )
2016-01-26 02:28:05 +01:00
d . addCallback ( self . _set_lbryid )
2016-01-20 03:07:16 +01:00
d . addCallback ( lambda _ : self . _get_lbrycrdd_path ( ) )
2015-12-06 23:32:17 +01:00
return d
2016-01-26 02:28:05 +01:00
def _set_lbryid ( self , lbryid ) :
2015-12-06 23:32:17 +01:00
if lbryid is None :
return self . _make_lbryid ( )
else :
2016-04-07 09:12:09 +02:00
log . info ( " LBRY ID: " + base58 . b58encode ( lbryid ) )
2015-12-06 23:32:17 +01:00
self . lbryid = lbryid
def _make_lbryid ( self ) :
self . lbryid = generate_id ( )
2016-04-07 09:12:09 +02:00
log . info ( " Generated new LBRY ID: " + base58 . b58encode ( self . lbryid ) )
2015-12-06 23:32:17 +01:00
d = self . settings . save_lbryid ( self . lbryid )
return d
def _setup_lbry_file_manager ( self ) :
2016-04-12 04:28:46 +02:00
self . startup_status = STARTUP_STAGES [ 3 ]
2015-12-06 23:32:17 +01:00
self . lbry_file_metadata_manager = DBLBRYFileMetadataManager ( self . db_dir )
d = self . lbry_file_metadata_manager . setup ( )
def set_lbry_file_manager ( ) :
2016-03-24 03:27:48 +01:00
self . lbry_file_manager = LBRYFileManager ( self . session ,
self . lbry_file_metadata_manager ,
self . sd_identifier ,
delete_data = True )
2015-12-06 23:32:17 +01:00
return self . lbry_file_manager . setup ( )
d . addCallback ( lambda _ : set_lbry_file_manager ( ) )
return d
def _get_session ( self ) :
def get_default_data_rate ( ) :
d = self . settings . get_default_data_payment_rate ( )
2016-02-16 19:39:08 +01:00
d . addCallback ( lambda rate : { " default_data_payment_rate " : rate if rate is not None else
2016-03-24 03:27:48 +01:00
MIN_BLOB_DATA_PAYMENT_RATE } )
2015-12-06 23:32:17 +01:00
return d
def get_wallet ( ) :
if self . wallet_type == " lbrycrd " :
2016-02-25 23:17:07 +01:00
log . info ( " Using lbrycrd wallet " )
2015-12-06 23:32:17 +01:00
lbrycrdd_path = None
if self . start_lbrycrdd is True :
lbrycrdd_path = self . lbrycrdd_path
if not lbrycrdd_path :
lbrycrdd_path = self . default_lbrycrdd_path
2016-01-07 16:24:55 +01:00
d = defer . succeed ( LBRYcrdWallet ( self . db_dir , wallet_dir = self . wallet_dir , wallet_conf = self . lbrycrd_conf ,
2015-12-06 23:32:17 +01:00
lbrycrdd_path = lbrycrdd_path ) )
2016-02-23 04:32:07 +01:00
elif self . wallet_type == " lbryum " :
2016-02-25 23:17:07 +01:00
log . info ( " Using lbryum wallet " )
2016-02-23 04:32:07 +01:00
d = defer . succeed ( LBRYumWallet ( self . db_dir ) )
elif self . wallet_type == " ptc " :
2016-02-25 23:17:07 +01:00
log . info ( " Using PTC wallet " )
2015-12-06 23:32:17 +01:00
d = defer . succeed ( PTCWallet ( self . db_dir ) )
2016-02-23 04:32:07 +01:00
else :
d = defer . fail ( )
2015-12-06 23:32:17 +01:00
d . addCallback ( lambda wallet : { " wallet " : wallet } )
return d
d1 = get_default_data_rate ( )
d2 = get_wallet ( )
def combine_results ( results ) :
r = { }
for success , result in results :
if success is True :
r . update ( result )
return r
def create_session ( results ) :
self . session = LBRYSession ( results [ ' default_data_payment_rate ' ] , db_dir = self . db_dir , lbryid = self . lbryid ,
blob_dir = self . blobfile_dir , dht_node_port = self . dht_node_port ,
known_dht_nodes = self . known_dht_nodes , peer_port = self . peer_port ,
use_upnp = self . use_upnp , wallet = results [ ' wallet ' ] )
2016-04-12 04:28:46 +02:00
self . startup_status = STARTUP_STAGES [ 2 ]
2015-12-06 23:32:17 +01:00
dl = defer . DeferredList ( [ d1 , d2 ] , fireOnOneErrback = True )
dl . addCallback ( combine_results )
dl . addCallback ( create_session )
dl . addCallback ( lambda _ : self . session . setup ( ) )
2016-04-18 01:54:04 +02:00
2015-12-06 23:32:17 +01:00
return dl
2016-03-03 00:32:48 +01:00
def _check_first_run ( self ) :
2016-04-07 09:12:09 +02:00
def _set_first_run_false ( ) :
2016-04-09 04:23:37 +02:00
log . info ( " Not first run " )
2016-04-07 09:12:09 +02:00
self . first_run = False
return 0.0
2016-03-20 02:24:44 +01:00
d = self . session . wallet . is_first_run ( )
2016-04-18 01:54:04 +02:00
d . addCallback ( lambda is_first_run : self . _do_first_run ( ) if is_first_run or not self . requested_first_run_credits
else _set_first_run_false ( ) )
2016-03-03 00:32:48 +01:00
return d
def _do_first_run ( self ) :
def send_request ( url , data ) :
2016-04-09 04:23:37 +02:00
log . info ( " Requesting first run credits " )
2016-03-03 00:32:48 +01:00
r = requests . post ( url , json = data )
if r . status_code == 200 :
2016-04-18 01:54:04 +02:00
self . requested_first_run_credits = True
self . session_settings [ ' requested_first_run_credits ' ] = True
f = open ( self . daemon_conf , " w " )
f . write ( json . dumps ( self . session_settings ) )
f . close ( )
2016-03-03 00:32:48 +01:00
return r . json ( ) [ ' credits_sent ' ]
return 0.0
def log_error ( err ) :
log . warning ( " unable to request free credits. %s " , err . getErrorMessage ( ) )
return 0.0
def request_credits ( address ) :
url = " http://credreq.lbry.io/requestcredits "
data = { " address " : address }
d = threads . deferToThread ( send_request , url , data )
d . addErrback ( log_error )
return d
2016-04-18 01:54:04 +02:00
self . first_run = True
d = self . session . wallet . get_new_address ( )
2016-03-03 00:32:48 +01:00
d . addCallback ( request_credits )
2016-04-18 01:54:04 +02:00
2016-03-03 00:32:48 +01:00
return d
def _show_first_run_result ( self , credits_received ) :
if credits_received != 0.0 :
points_string = locale . format_string ( " %.2f LBC " , ( round ( credits_received , 2 ) , ) , grouping = True )
2016-04-14 23:01:05 +02:00
self . startup_message = " Thank you for testing the alpha version of LBRY! You have been given %s for free because we love you. Please hang on for a few minutes for the next block to be mined. When you refresh this page and see your credits you ' re ready to go!. " % points_string
2016-03-03 00:32:48 +01:00
else :
2016-04-09 10:35:34 +02:00
self . startup_message = None
2016-03-03 00:32:48 +01:00
2016-01-20 03:07:16 +01:00
def _get_lbrycrdd_path ( self ) :
2015-12-06 23:32:17 +01:00
def get_lbrycrdd_path_conf_file ( ) :
lbrycrdd_path_conf_path = os . path . join ( os . path . expanduser ( " ~ " ) , " .lbrycrddpath.conf " )
if not os . path . exists ( lbrycrdd_path_conf_path ) :
return " "
lbrycrdd_path_conf = open ( lbrycrdd_path_conf_path )
lines = lbrycrdd_path_conf . readlines ( )
return lines
d = threads . deferToThread ( get_lbrycrdd_path_conf_file )
def load_lbrycrdd_path ( conf ) :
for line in conf :
if len ( line . strip ( ) ) and line . strip ( ) [ 0 ] != " # " :
self . lbrycrdd_path = line . strip ( )
d . addCallback ( load_lbrycrdd_path )
return d
def _setup_stream_identifier ( self ) :
file_saver_factory = LBRYFileSaverFactory ( self . session . peer_finder , self . session . rate_limiter ,
self . session . blob_manager , self . stream_info_manager ,
2016-04-07 09:12:09 +02:00
self . session . wallet , self . session_settings [ ' default_download_directory ' ] )
2015-12-06 23:32:17 +01:00
self . sd_identifier . add_stream_downloader_factory ( LBRYFileStreamType , file_saver_factory )
file_opener_factory = LBRYFileOpenerFactory ( self . session . peer_finder , self . session . rate_limiter ,
self . session . blob_manager , self . stream_info_manager ,
self . session . wallet )
self . sd_identifier . add_stream_downloader_factory ( LBRYFileStreamType , file_opener_factory )
return defer . succeed ( None )
def _setup_lbry_file_opener ( self ) :
downloader_factory = LBRYFileOpenerFactory ( self . session . peer_finder , self . session . rate_limiter ,
self . session . blob_manager , self . stream_info_manager ,
self . session . wallet )
self . sd_identifier . add_stream_downloader_factory ( LBRYFileStreamType , downloader_factory )
return defer . succeed ( True )
2016-04-09 04:23:37 +02:00
def _download_name ( self , name , timeout = DEFAULT_TIMEOUT , download_directory = None ) :
if not download_directory :
download_directory = self . session_settings [ ' default_download_directory ' ]
elif not os . path . isdir ( download_directory ) :
download_directory = self . session_settings [ ' default_download_directory ' ]
2016-03-24 03:27:48 +01:00
def _disp_file ( f ) :
2016-04-07 09:12:09 +02:00
file_path = os . path . join ( self . session_settings [ ' default_download_directory ' ] , f . file_name )
2016-03-24 03:27:48 +01:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Already downloaded: " + str ( f . stream_hash ) + " --> " + file_path )
return defer . succeed ( f )
2015-12-15 03:49:01 +01:00
2016-01-20 03:07:16 +01:00
def _get_stream ( name ) :
def _disp ( stream ) :
2016-04-02 04:11:11 +02:00
stream_hash = stream [ ' stream_hash ' ]
if isinstance ( stream_hash , dict ) :
stream_hash = stream_hash [ ' sd_hash ' ]
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Start stream: " + stream_hash )
2016-01-20 03:07:16 +01:00
return stream
2016-01-12 18:42:32 +01:00
2016-01-20 03:07:16 +01:00
d = self . session . wallet . get_stream_info_for_name ( name )
2016-01-12 00:01:46 +01:00
stream = GetStream ( self . sd_identifier , self . session , self . session . wallet , self . lbry_file_manager ,
2016-04-07 09:12:09 +02:00
max_key_fee = self . max_key_fee , data_rate = self . data_rate , timeout = timeout ,
2016-04-09 04:23:37 +02:00
download_directory = download_directory )
2016-01-20 03:07:16 +01:00
d . addCallback ( _disp )
2016-01-12 00:01:46 +01:00
d . addCallback ( lambda stream_info : stream . start ( stream_info ) )
2016-01-20 03:07:16 +01:00
d . addCallback ( lambda _ : self . _path_from_name ( name ) )
2016-01-12 00:01:46 +01:00
return d
2016-01-20 03:07:16 +01:00
d = self . _check_history ( name )
d . addCallback ( lambda lbry_file : _get_stream ( name ) if not lbry_file else _disp_file ( lbry_file ) )
2016-03-24 03:27:48 +01:00
d . addCallback ( lambda _ : self . _path_from_name ( name ) )
d . addErrback ( lambda err : defer . fail ( NOT_FOUND ) )
2015-12-15 03:49:01 +01:00
return d
def _resolve_name ( self , name ) :
d = defer . Deferred ( )
d . addCallback ( lambda _ : self . session . wallet . get_stream_info_for_name ( name ) )
2016-01-20 03:07:16 +01:00
d . addErrback ( lambda _ : defer . fail ( UnknownNameError ) )
2015-12-15 03:49:01 +01:00
return d
2016-01-20 03:07:16 +01:00
def _resolve_name_wc ( self , name ) :
d = defer . Deferred ( )
d . addCallback ( lambda _ : self . session . wallet . get_stream_info_for_name ( name ) )
d . addErrback ( lambda _ : defer . fail ( UnknownNameError ) )
d . callback ( None )
return d
2016-01-12 00:01:46 +01:00
2016-01-20 03:07:16 +01:00
def _check_history ( self , name ) :
def _get_lbry_file ( path ) :
f = open ( path , ' r ' )
l = json . loads ( f . read ( ) )
f . close ( )
2016-03-24 03:27:48 +01:00
2016-01-20 03:07:16 +01:00
file_name = l [ ' stream_name ' ] . decode ( ' hex ' )
2016-04-02 04:11:11 +02:00
2016-03-24 03:27:48 +01:00
for lbry_file in self . lbry_file_manager . lbry_files :
if lbry_file . stream_name == file_name :
if sys . platform == " darwin " :
2016-04-07 09:12:09 +02:00
if os . path . isfile ( os . path . join ( self . session_settings [ ' default_download_directory ' ] , lbry_file . stream_name ) ) :
2016-03-24 03:27:48 +01:00
return lbry_file
else :
return False
else :
return lbry_file
2016-02-11 14:32:48 +01:00
else :
2016-03-24 03:27:48 +01:00
return False
2016-01-20 03:07:16 +01:00
def _check ( info ) :
stream_hash = info [ ' stream_hash ' ]
2016-04-02 04:11:11 +02:00
if isinstance ( stream_hash , dict ) :
stream_hash = stream_hash [ ' sd_hash ' ]
2016-01-20 03:07:16 +01:00
path = os . path . join ( self . blobfile_dir , stream_hash )
if os . path . isfile ( path ) :
2016-02-25 23:17:07 +01:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Search for lbry_file, returning: " + stream_hash )
2016-01-20 03:07:16 +01:00
return defer . succeed ( _get_lbry_file ( path ) )
2016-01-12 00:01:46 +01:00
else :
2016-02-25 23:17:07 +01:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Search for lbry_file didn ' t return anything " )
2016-01-20 03:07:16 +01:00
return defer . succeed ( False )
d = self . _resolve_name ( name )
2016-03-24 03:27:48 +01:00
d . addCallback ( _check )
2016-01-20 03:07:16 +01:00
d . callback ( None )
return d
def _delete_lbry_file ( self , lbry_file ) :
d = self . lbry_file_manager . delete_lbry_file ( lbry_file )
def finish_deletion ( lbry_file ) :
d = lbry_file . delete_data ( )
d . addCallback ( lambda _ : _delete_stream_data ( lbry_file ) )
return d
2016-01-12 18:42:32 +01:00
2016-01-20 03:07:16 +01:00
def _delete_stream_data ( lbry_file ) :
s_h = lbry_file . stream_hash
d = self . lbry_file_manager . get_count_for_stream_hash ( s_h )
# TODO: could possibly be a timing issue here
d . addCallback ( lambda c : self . stream_info_manager . delete_stream ( s_h ) if c == 0 else True )
2016-04-07 09:12:09 +02:00
d . addCallback ( lambda _ : os . remove ( os . path . join ( self . session_settings [ ' default_download_directory ' ] , lbry_file . file_name ) ) if
os . path . isfile ( os . path . join ( self . session_settings [ ' default_download_directory ' ] , lbry_file . file_name ) ) else defer . succeed ( None ) )
2016-01-20 03:07:16 +01:00
return d
d . addCallback ( lambda _ : finish_deletion ( lbry_file ) )
return d
def _path_from_name ( self , name ) :
d = self . _check_history ( name )
d . addCallback ( lambda lbry_file : { ' stream_hash ' : lbry_file . stream_hash ,
2016-04-07 09:12:09 +02:00
' path ' : os . path . join ( self . session_settings [ ' default_download_directory ' ] , lbry_file . file_name ) }
2016-03-24 03:27:48 +01:00
if lbry_file else defer . fail ( UnknownNameError ) )
2016-01-20 03:07:16 +01:00
return d
def _path_from_lbry_file ( self , lbry_file ) :
if lbry_file :
r = { ' stream_hash ' : lbry_file . stream_hash ,
2016-04-07 09:12:09 +02:00
' path ' : os . path . join ( self . session_settings [ ' default_download_directory ' ] , lbry_file . file_name ) }
2016-01-20 03:07:16 +01:00
return defer . succeed ( r )
else :
return defer . fail ( UnknownNameError )
2016-01-12 00:01:46 +01:00
2016-01-24 06:10:22 +01:00
def _get_est_cost ( self , name ) :
def _check_est ( d , name ) :
if type ( d . result ) is float :
2016-03-24 03:27:48 +01:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Cost est for lbry:// " + name + " : " + str ( d . result ) + " LBC " )
2016-01-24 06:10:22 +01:00
else :
2016-03-24 03:27:48 +01:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Timeout estimating cost for lbry:// " + name + " , using key fee " )
2016-01-24 06:10:22 +01:00
d . cancel ( )
return defer . succeed ( None )
def _add_key_fee ( data_cost ) :
d = self . session . wallet . get_stream_info_for_name ( name )
2016-02-15 16:12:49 +01:00
d . addCallback ( lambda info : data_cost + info [ ' key_fee ' ] if ' key_fee ' in info . keys ( ) else data_cost )
2016-01-24 06:10:22 +01:00
return d
d = self . session . wallet . get_stream_info_for_name ( name )
2016-04-02 04:11:11 +02:00
d . addCallback ( lambda info : info [ ' stream_hash ' ] if isinstance ( info [ ' stream_hash ' ] , str )
else info [ ' stream_hash ' ] [ ' sd_hash ' ] )
d . addCallback ( lambda sd_hash : download_sd_blob ( self . session , sd_hash ,
2016-01-26 05:59:57 +01:00
self . blob_request_payment_rate_manager ) )
2016-01-24 06:10:22 +01:00
d . addCallback ( self . sd_identifier . get_metadata_for_sd_blob )
d . addCallback ( lambda metadata : metadata . validator . info_to_show ( ) )
2016-02-19 02:41:23 +01:00
d . addCallback ( lambda info : int ( dict ( info ) [ ' stream_size ' ] ) / 1000000 * self . data_rate )
2016-01-24 06:10:22 +01:00
d . addCallback ( _add_key_fee )
d . addErrback ( lambda _ : _add_key_fee ( 0.0 ) )
2016-02-15 16:12:49 +01:00
reactor . callLater ( self . search_timeout , _check_est , d , name )
2016-01-24 06:10:22 +01:00
return d
2016-04-14 04:07:27 +02:00
def _get_lbry_files ( self ) :
r = [ ]
for f in self . lbry_file_manager . lbry_files :
if f . key :
t = { ' completed ' : f . completed , ' file_name ' : f . file_name , ' key ' : binascii . b2a_hex ( f . key ) ,
' points_paid ' : f . points_paid , ' stopped ' : f . stopped , ' stream_hash ' : f . stream_hash ,
' stream_name ' : f . stream_name , ' suggested_file_name ' : f . suggested_file_name ,
' upload_allowed ' : f . upload_allowed }
else :
t = { ' completed ' : f . completed , ' file_name ' : f . file_name , ' key ' : None ,
' points_paid ' : f . points_paid ,
' stopped ' : f . stopped , ' stream_hash ' : f . stream_hash , ' stream_name ' : f . stream_name ,
' suggested_file_name ' : f . suggested_file_name , ' upload_allowed ' : f . upload_allowed }
r . append ( t )
return r
2016-03-24 03:27:48 +01:00
def _render_response ( self , result , code ) :
2016-03-29 22:42:47 +02:00
return defer . succeed ( { ' result ' : result , ' code ' : code } )
2016-03-24 03:27:48 +01:00
def jsonrpc_is_running ( self ) :
"""
2016-04-13 20:47:34 +02:00
Check if lbrynet daemon is running
Args :
None
Returns : true if daemon completed startup , otherwise false
2016-03-24 03:27:48 +01:00
"""
2016-04-09 05:38:57 +02:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] is_running: " + str ( self . announced_startup ) )
2016-03-29 22:42:47 +02:00
if self . announced_startup :
2016-03-24 03:27:48 +01:00
return self . _render_response ( True , OK_CODE )
2016-03-03 00:32:48 +01:00
else :
2016-03-24 03:27:48 +01:00
return self . _render_response ( False , OK_CODE )
2016-02-19 02:41:23 +01:00
2016-04-09 07:33:27 +02:00
def jsonrpc_daemon_status ( self ) :
"""
2016-04-13 20:47:34 +02:00
Get lbrynet daemon status information
Args :
None
Returns :
' status_message ' : startup status message
' status_code ' : status_code
if status_code is ' loading_wallet ' , also contains key ' progress ' : blockchain catchup progress
2016-04-09 07:33:27 +02:00
"""
2016-04-09 08:09:30 +02:00
2016-04-14 06:29:40 +02:00
r = { ' code ' : self . startup_status [ 0 ] , ' message ' : self . startup_status [ 1 ] , ' progress ' : None , ' is_lagging ' : None }
2016-04-14 05:10:17 +02:00
if self . startup_status [ 0 ] == ' loading_wallet ' :
2016-04-14 06:29:40 +02:00
r [ ' is_lagging ' ] = self . session . wallet . is_lagging
2016-04-14 22:50:11 +02:00
if r [ ' is_lagging ' ] == True :
r [ ' message ' ] = " Synchronization with the blockchain is lagging... if this continues try restarting LBRY "
else :
r [ ' message ' ] = r [ ' message ' ] % ( str ( self . session . wallet . blocks_behind_alert ) + " blocks behind " )
r [ ' progress ' ] = self . session . wallet . catchup_progress
2016-04-14 06:29:40 +02:00
2016-04-09 07:33:27 +02:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] daemon status: " + str ( r ) )
return self . _render_response ( r , OK_CODE )
2016-04-09 10:00:31 +02:00
def jsonrpc_is_first_run ( self ) :
2016-04-09 21:43:10 +02:00
"""
2016-04-13 20:47:34 +02:00
Check if this is the first time lbrynet daemon has been run
Args :
None
Returns :
True if first run , otherwise False
2016-04-09 21:43:10 +02:00
"""
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Check if is first run " )
2016-04-07 09:12:09 +02:00
try :
d = self . session . wallet . is_first_run ( )
except :
d = defer . fail ( None )
2016-04-09 21:43:10 +02:00
d . addCallbacks ( lambda r : self . _render_response ( r , OK_CODE ) , lambda _ : self . _render_response ( None , OK_CODE ) )
2016-04-07 09:12:09 +02:00
return d
2016-04-09 10:15:58 +02:00
def jsonrpc_get_start_notice ( self ) :
2016-04-09 21:43:10 +02:00
"""
2016-04-13 20:47:34 +02:00
Get special message to be displayed at startup
2016-04-09 21:43:10 +02:00
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
Startup message , such as first run notification
"""
2016-04-09 22:14:28 +02:00
2016-04-09 21:43:10 +02:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Get startup notice " )
2016-04-09 22:14:28 +02:00
2016-04-09 22:18:06 +02:00
if self . first_run and not self . session . wallet . wallet_balance :
2016-04-09 22:14:28 +02:00
return self . _render_response ( self . startup_message , OK_CODE )
elif self . first_run :
return self . _render_response ( None , OK_CODE )
else :
self . _render_response ( self . startup_message , OK_CODE )
2016-04-12 04:28:46 +02:00
def jsonrpc_version ( self ) :
"""
Get lbry version information
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
2016-04-16 04:31:06 +02:00
" platform " : platform string
" os_release " : os release string
" os_system " : os name
" lbrynet_version: " : lbrynet_version ,
" lbryum_version: " : lbryum_version ,
" ui_version " : commit hash of ui version being used
2016-04-14 04:07:27 +02:00
" remote_lbrynet " : most recent lbrynet version available from github
" remote_lbryum " : most recent lbryum version available from github
2016-04-12 04:28:46 +02:00
"""
msg = {
2016-04-16 04:31:06 +02:00
' platform ' : self . platform_info [ ' platform ' ] ,
' os_release ' : self . platform_info [ ' os_release ' ] ,
' os_system ' : self . platform_info [ ' os_system ' ] ,
' lbrynet_version ' : lbrynet_version ,
' lbryum_version ' : lbryum_version ,
' ui_version ' : self . ui_version ,
' remote_lbrynet ' : self . git_lbrynet_version ,
' remote_lbryum ' : self . git_lbryum_version
2016-04-12 04:28:46 +02:00
}
2016-04-12 05:13:26 +02:00
2016-04-12 04:37:42 +02:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Get version info: " + json . dumps ( msg ) )
2016-04-12 04:28:46 +02:00
return self . _render_response ( msg , OK_CODE )
2016-04-09 10:15:58 +02:00
2016-03-24 03:27:48 +01:00
def jsonrpc_get_settings ( self ) :
2016-01-07 16:24:55 +01:00
"""
2016-04-13 20:47:34 +02:00
Get lbrynet daemon settings
Args :
None
Returns :
' run_on_startup ' : bool ,
' data_rate ' : float ,
' max_key_fee ' : float ,
' default_download_directory ' : string ,
' max_upload ' : float , 0.0 for unlimited
' max_download ' : float , 0.0 for unlimited
' upload_log ' : bool ,
' search_timeout ' : float ,
' max_search_results ' : int ,
' wallet_type ' : string ,
' delete_blobs_on_remove ' : bool ,
' peer_port ' : int ,
' dht_node_port ' : int ,
' use_upnp ' : bool ,
' start_lbrycrdd ' : bool ,
2016-01-07 16:24:55 +01:00
"""
2016-01-12 18:42:32 +01:00
2016-03-24 03:27:48 +01:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Get daemon settings " )
return self . _render_response ( self . session_settings , OK_CODE )
2016-01-07 16:24:55 +01:00
2016-03-24 03:27:48 +01:00
def jsonrpc_set_settings ( self , p ) :
2016-01-24 06:10:22 +01:00
"""
2016-04-13 20:47:34 +02:00
Set lbrynet daemon settings
Args :
' run_on_startup ' : bool ,
' data_rate ' : float ,
' max_key_fee ' : float ,
' default_download_directory ' : string ,
' max_upload ' : float , 0.0 for unlimited
' max_download ' : float , 0.0 for unlimited
' upload_log ' : bool ,
' search_timeout ' : float ,
' max_search_results ' : int ,
' wallet_type ' : string ,
' delete_blobs_on_remove ' : bool ,
' peer_port ' : int ,
' dht_node_port ' : int ,
' use_upnp ' : bool ,
' start_lbrycrdd ' : bool ,
Returns :
settings dict
2016-01-24 06:10:22 +01:00
"""
2016-04-07 09:12:09 +02:00
def _log_settings_change ( params ) :
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Set daemon settings to " + str ( params ) )
2016-03-24 03:27:48 +01:00
d = self . _update_settings ( p )
2016-04-07 09:12:09 +02:00
d . addCallback ( lambda _ : _log_settings_change ( p ) )
d . addCallback ( lambda _ : self . _render_response ( self . session_settings , OK_CODE ) )
2016-01-07 16:24:55 +01:00
2016-04-07 09:12:09 +02:00
return d
2016-01-07 16:24:55 +01:00
2016-04-13 20:47:34 +02:00
def jsonrpc_help ( self , p = None ) :
"""
Function to retrieve docstring for API function
Args :
optional ' function ' : function to retrieve documentation for
optional ' callable_during_startup ' :
Returns :
if given a function , returns given documentation
if given callable_during_startup flag , returns list of functions callable during the startup sequence
if no params are given , returns the list of callable functions
"""
if not p :
2016-04-16 01:37:27 +02:00
return self . _render_response ( self . _listFunctions ( ) , OK_CODE )
2016-04-13 20:47:34 +02:00
elif ' callable_during_start ' in p . keys ( ) :
return self . _render_response ( ALLOWED_DURING_STARTUP , OK_CODE )
elif ' function ' in p . keys ( ) :
func_path = p [ ' function ' ]
function = self . _getFunction ( func_path )
return self . _render_response ( function . __doc__ , OK_CODE )
else :
return self . _render_response ( self . jsonrpc_help . __doc__ , OK_CODE )
2016-03-24 03:27:48 +01:00
def jsonrpc_start_fetcher ( self ) :
2015-12-14 18:59:44 +01:00
"""
2016-04-13 20:47:34 +02:00
Start automatically downloading new name claims as they occur ( off by default )
2016-03-24 03:27:48 +01:00
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
confirmation message
2015-12-14 18:59:44 +01:00
"""
self . fetcher . start ( )
2016-02-25 23:17:07 +01:00
log . info ( ' [ ' + str ( datetime . now ( ) ) + ' ] Start autofetcher ' )
2016-03-24 03:27:48 +01:00
# self._log_to_slack('[' + str(datetime.now()) + '] Start autofetcher')
return self . _render_response ( " Started autofetching claims " , OK_CODE )
2015-12-14 18:59:44 +01:00
2016-03-24 03:27:48 +01:00
def jsonrpc_stop_fetcher ( self ) :
2015-12-14 18:59:44 +01:00
"""
2016-04-13 20:47:34 +02:00
Stop automatically downloading new name claims as they occur
2016-03-24 03:27:48 +01:00
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
confirmation message
2015-12-14 18:59:44 +01:00
"""
self . fetcher . stop ( )
2016-02-25 23:17:07 +01:00
log . info ( ' [ ' + str ( datetime . now ( ) ) + ' ] Stop autofetcher ' )
2016-03-24 03:27:48 +01:00
return self . _render_response ( " Stopped autofetching claims " , OK_CODE )
2015-12-14 18:59:44 +01:00
2016-03-24 03:27:48 +01:00
def jsonrpc_fetcher_status ( self ) :
2015-12-14 18:59:44 +01:00
"""
2016-03-24 03:27:48 +01:00
Get fetcher status
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
True / False
2015-12-14 18:59:44 +01:00
"""
2016-03-24 03:27:48 +01:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Get fetcher status " )
return self . _render_response ( self . fetcher . check_if_running ( ) , OK_CODE )
2015-12-14 18:59:44 +01:00
2016-03-24 03:27:48 +01:00
def jsonrpc_get_balance ( self ) :
2015-12-06 23:32:17 +01:00
"""
2016-04-13 20:47:34 +02:00
Get balance
2016-03-24 03:27:48 +01:00
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
balance , float
2015-12-06 23:32:17 +01:00
"""
2015-12-14 18:59:44 +01:00
2016-03-24 03:27:48 +01:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Get balance " )
2016-03-29 22:42:47 +02:00
return self . _render_response ( float ( self . session . wallet . wallet_balance ) , OK_CODE )
2015-12-06 23:32:17 +01:00
2016-03-24 03:27:48 +01:00
def jsonrpc_stop ( self ) :
2015-12-06 23:32:17 +01:00
"""
2016-01-24 06:10:22 +01:00
Stop lbrynet - daemon
2016-03-24 03:27:48 +01:00
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
shutdown message
2015-12-06 23:32:17 +01:00
"""
2016-01-11 21:52:38 +01:00
def _disp_shutdown ( ) :
2016-03-24 03:27:48 +01:00
log . info ( " Shutting down lbrynet daemon " )
2016-01-11 21:52:38 +01:00
d = self . _shutdown ( )
d . addCallback ( lambda _ : _disp_shutdown ( ) )
2016-01-26 02:28:05 +01:00
d . addCallback ( lambda _ : reactor . callLater ( 1.0 , reactor . stop ) )
2016-01-11 21:52:38 +01:00
2016-03-24 03:27:48 +01:00
return self . _render_response ( " Shutting down " , OK_CODE )
2015-12-06 23:32:17 +01:00
2016-03-24 03:27:48 +01:00
def jsonrpc_get_lbry_files ( self ) :
2015-12-15 03:49:01 +01:00
"""
Get LBRY files
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
List of lbry files :
' completed ' : bool
' file_name ' : string
' key ' : hex string
' points_paid ' : float
' stopped ' : bool
' stream_hash ' : base 58 string
' stream_name ' : string
' suggested_file_name ' : string
' upload_allowed ' : bool
2015-12-15 03:49:01 +01:00
"""
2016-04-14 04:07:27 +02:00
r = self . _get_lbry_files ( )
2016-03-24 03:27:48 +01:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Get LBRY files " )
return self . _render_response ( r , OK_CODE )
2015-12-15 03:49:01 +01:00
2016-03-24 03:27:48 +01:00
def jsonrpc_resolve_name ( self , p ) :
2015-12-06 23:32:17 +01:00
"""
Resolve stream info from a LBRY uri
2015-12-15 03:49:01 +01:00
2016-04-13 20:47:34 +02:00
Args :
' name ' : name to look up , string , do not include lbry : / / prefix
Returns :
metadata from name claim
2015-12-06 23:32:17 +01:00
"""
2016-04-13 20:47:34 +02:00
if ' name ' in p . keys ( ) :
name = p [ ' name ' ]
else :
return self . _render_response ( None , BAD_REQUEST )
2015-12-06 23:32:17 +01:00
def _disp ( info ) :
2016-04-02 04:11:11 +02:00
stream_hash = info [ ' stream_hash ' ]
if isinstance ( stream_hash , dict ) :
stream_hash = stream_hash [ ' sd_hash ' ]
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Resolved info: " + stream_hash )
2016-03-24 03:27:48 +01:00
return self . _render_response ( info , OK_CODE )
2015-12-06 23:32:17 +01:00
2016-04-13 20:47:34 +02:00
d = self . _resolve_name ( name )
2016-04-02 04:11:11 +02:00
d . addCallbacks ( _disp , lambda _ : server . failure )
2015-12-06 23:32:17 +01:00
d . callback ( None )
return d
2016-03-24 03:27:48 +01:00
def jsonrpc_get ( self , p ) :
2015-12-06 23:32:17 +01:00
"""
Download stream from a LBRY uri
2015-12-15 03:49:01 +01:00
2016-04-13 20:47:34 +02:00
Args :
' name ' : name to download , string
optional ' download_directory ' : path to directory where file will be saved , string
Returns :
' stream_hash ' : hex string
' path ' : path of download
2015-12-06 23:32:17 +01:00
"""
2016-03-24 03:27:48 +01:00
if ' timeout ' not in p . keys ( ) :
2016-04-09 04:23:37 +02:00
timeout = DEFAULT_TIMEOUT
else :
timeout = p [ ' timeout ' ]
if ' download_directory ' not in p . keys ( ) :
download_directory = self . session_settings [ ' default_download_directory ' ]
else :
download_directory = p [ ' download_directory ' ]
2015-12-06 23:32:17 +01:00
2016-04-09 04:23:37 +02:00
if ' name ' in p . keys ( ) :
name = p [ ' name ' ]
d = self . _download_name ( name = name , timeout = timeout , download_directory = download_directory )
2016-03-24 03:27:48 +01:00
d . addCallbacks ( lambda message : self . _render_response ( message , OK_CODE ) ,
lambda err : self . _render_response ( ' error ' , NOT_FOUND ) )
2016-01-24 06:26:41 +01:00
else :
2016-03-24 03:27:48 +01:00
d = self . _render_response ( ' error ' , BAD_REQUEST )
2015-12-08 06:37:49 +01:00
return d
2016-03-29 22:42:47 +02:00
# def jsonrpc_stop_lbry_file(self, p):
# params = Bunch(p)
#
# try:
# lbry_file = [f for f in self.lbry_file_manager.lbry_files if f.stream_hash == params.stream_hash][0]
# except IndexError:
# return defer.fail(UnknownNameError)
#
# if not lbry_file.stopped:
# d = self.lbry_file_manager.toggle_lbry_file_running(lbry_file)
# d.addCallback(lambda _: self._render_response("Stream has been stopped", OK_CODE))
# d.addErrback(lambda err: self._render_response(err.getTraceback(), ))
# return d
# else:
# return json.dumps({'result': 'Stream was already stopped'})
#
# def jsonrpc_start_lbry_file(self, p):
# params = Bunch(p)
#
# try:
# lbry_file = [f for f in self.lbry_file_manager.lbry_files if f.stream_hash == params.stream_hash][0]
# except IndexError:
# return defer.fail(UnknownNameError)
#
# if lbry_file.stopped:
# d = self.lbry_file_manager.toggle_lbry_file_running(lbry_file)
# d.callback(None)
# return json.dumps({'result': 'Stream started'})
# else:
# return json.dumps({'result': 'Stream was already running'})
2016-01-17 05:06:24 +01:00
2016-03-24 03:27:48 +01:00
def jsonrpc_search_nametrie ( self , p ) :
2016-01-24 06:10:22 +01:00
"""
2016-04-13 20:47:34 +02:00
Search the nametrie for claims beginning with search ( yes , this is a dumb search , it ' ll be made better)
2016-01-24 06:10:22 +01:00
2016-04-13 20:47:34 +02:00
Args :
' search ' : search query , string
Returns :
List of search results
2016-01-24 06:10:22 +01:00
"""
2016-04-13 20:47:34 +02:00
if ' search ' in p . keys ( ) :
search = p [ ' search ' ]
else :
return self . _render_response ( None , BAD_REQUEST )
2016-03-24 03:27:48 +01:00
2016-01-17 05:06:24 +01:00
def _clean ( n ) :
t = [ ]
for i in n :
if i [ 0 ] :
2016-01-24 06:10:22 +01:00
if i [ 1 ] [ 0 ] [ 0 ] and i [ 1 ] [ 1 ] [ 0 ] and i [ 1 ] [ 2 ] [ 0 ] :
2016-01-17 05:06:24 +01:00
i [ 1 ] [ 0 ] [ 1 ] [ ' value ' ] = str ( i [ 1 ] [ 0 ] [ 1 ] [ ' value ' ] )
2016-01-24 06:10:22 +01:00
t . append ( [ i [ 1 ] [ 0 ] [ 1 ] , i [ 1 ] [ 1 ] [ 1 ] , i [ 1 ] [ 2 ] [ 1 ] ] )
2016-01-17 05:06:24 +01:00
return t
2016-02-17 05:10:26 +01:00
def resolve_claims ( claims ) :
ds = [ ]
for claim in claims :
d1 = defer . succeed ( claim )
d2 = self . _resolve_name_wc ( claim [ ' name ' ] )
d3 = self . _get_est_cost ( claim [ ' name ' ] )
dl = defer . DeferredList ( [ d1 , d2 , d3 ] , consumeErrors = True )
ds . append ( dl )
return defer . DeferredList ( ds )
2016-01-17 05:06:24 +01:00
def _disp ( results ) :
2016-03-24 03:27:48 +01:00
log . info ( ' [ ' + str ( datetime . now ( ) ) + ' ] Found ' + str ( len ( results ) ) + ' search results ' )
consolidated_results = [ ]
2016-02-25 23:17:07 +01:00
for r in results :
2016-03-24 03:27:48 +01:00
t = { }
t . update ( r [ 0 ] )
if ' name ' in r [ 1 ] . keys ( ) :
r [ 1 ] [ ' stream_name ' ] = r [ 1 ] [ ' name ' ]
del r [ 1 ] [ ' name ' ]
t . update ( r [ 1 ] )
t [ ' cost_est ' ] = r [ 2 ]
consolidated_results . append ( t )
# log.info(str(t))
2016-03-29 22:42:47 +02:00
return consolidated_results
2016-03-24 03:27:48 +01:00
2016-04-13 20:47:34 +02:00
log . info ( ' [ ' + str ( datetime . now ( ) ) + ' ] Search nametrie: ' + search )
2016-01-17 05:06:24 +01:00
2016-02-17 05:10:26 +01:00
d = self . session . wallet . get_nametrie ( )
2016-04-13 20:47:34 +02:00
d . addCallback ( lambda trie : [ claim for claim in trie if claim [ ' name ' ] . startswith ( search ) and ' txid ' in claim ] )
2016-02-17 05:10:26 +01:00
d . addCallback ( lambda claims : claims [ : self . max_search_results ] )
d . addCallback ( resolve_claims )
2016-01-17 05:06:24 +01:00
d . addCallback ( _clean )
d . addCallback ( _disp )
2016-03-29 22:42:47 +02:00
d . addCallback ( lambda results : self . _render_response ( results , OK_CODE ) )
2016-01-17 05:06:24 +01:00
return d
2016-03-24 03:27:48 +01:00
def jsonrpc_delete_lbry_file ( self , p ) :
"""
Delete a lbry file
2016-01-21 04:00:28 +01:00
2016-04-13 20:47:34 +02:00
Args :
' file_name ' : downloaded file name , string
Returns :
confirmation message
2016-03-24 03:27:48 +01:00
"""
2016-01-21 04:00:28 +01:00
2016-03-24 03:27:48 +01:00
def _disp ( file_name ) :
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Deleted: " + file_name )
return self . _render_response ( " Deleted: " + file_name , OK_CODE )
2016-01-21 04:00:28 +01:00
2016-04-07 09:12:09 +02:00
if " file_name " in p . keys ( ) :
lbry_files = [ self . _delete_lbry_file ( f ) for f in self . lbry_file_manager . lbry_files
if p [ ' file_name ' ] == f . file_name ]
d = defer . DeferredList ( lbry_files )
d . addCallback ( lambda _ : _disp ( p [ ' file_name ' ] ) )
return d
2016-01-21 04:00:28 +01:00
2016-03-24 03:27:48 +01:00
def jsonrpc_publish ( self , p ) :
"""
Make a new name claim
2016-01-21 04:00:28 +01:00
2016-04-13 20:47:34 +02:00
Args :
' name ' : name to be claimed , string
' file_path ' : path to file to be associated with name , string
' bid ' : amount of credits to commit in this claim , float
optional ' author ' : author , string
optional ' title ' : title , description
optional ' description ' : content description , string
optional ' thumbnail ' : thumbnail image url
optional ' key_fee ' : key fee to be paid to publisher , float ( default 0.0 )
optional ' key_fee_address ' : address for key fee to be sent to , string ( defaults on new address )
optional ' content_license ' : content license string
optional ' sources ' : alternative sources dict , keys ' lbry_sd_hash ' , ' btih ' , ' url '
Returns :
Confirmation message
2016-03-24 03:27:48 +01:00
"""
2016-01-21 04:00:28 +01:00
2016-04-02 04:11:11 +02:00
metadata_fields = [ " name " , " file_path " , " bid " , " author " , " title " ,
" description " , " thumbnail " , " key_fee " , " key_fee_address " ,
" content_license " , " sources " ]
2016-01-21 04:00:28 +01:00
2016-04-02 04:11:11 +02:00
for k in metadata_fields :
if k not in p . keys ( ) :
p [ k ] = None
2016-01-21 04:00:28 +01:00
2016-04-02 04:11:11 +02:00
pub = Publisher ( self . session , self . lbry_file_manager , self . session . wallet )
d = pub . start ( p [ ' name ' ] ,
p [ ' file_path ' ] ,
p [ ' bid ' ] ,
title = p [ ' title ' ] ,
description = p [ ' description ' ] ,
thumbnail = p [ ' thumbnail ' ] ,
key_fee = p [ ' key_fee ' ] ,
key_fee_address = p [ ' key_fee_address ' ] ,
content_license = p [ ' content_license ' ] ,
author = p [ ' author ' ] ,
sources = p [ ' sources ' ] )
2016-03-24 03:27:48 +01:00
d . addCallbacks ( lambda msg : self . _render_response ( msg , OK_CODE ) ,
lambda err : self . _render_response ( err . getTraceback ( ) , BAD_REQUEST ) )
2016-01-21 04:00:28 +01:00
return d
2015-12-06 23:32:17 +01:00
2016-03-24 03:27:48 +01:00
def jsonrpc_abandon_name ( self , p ) :
"""
2016-04-13 20:47:34 +02:00
Abandon a name and reclaim credits from the claim
2016-03-24 03:27:48 +01:00
2016-04-13 20:47:34 +02:00
Args :
' txid ' : txid of claim , string
Return :
Confirmation message
2016-03-24 03:27:48 +01:00
"""
2016-04-13 20:47:34 +02:00
if ' txid ' in p . keys ( ) :
txid = p [ ' txid ' ]
else :
return server . failure
2016-03-24 03:27:48 +01:00
2016-04-07 09:12:09 +02:00
def _disp ( x ) :
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Abandoned name claim tx " + str ( x ) )
return self . _render_response ( x , OK_CODE )
2016-01-26 02:28:05 +01:00
d = defer . Deferred ( )
2016-04-13 20:47:34 +02:00
d . addCallback ( lambda _ : self . session . wallet . abandon_name ( txid ) )
2016-04-07 09:12:09 +02:00
d . addCallback ( _disp )
2016-01-26 02:28:05 +01:00
d . callback ( None )
return d
2016-03-24 03:27:48 +01:00
def jsonrpc_get_name_claims ( self ) :
"""
2016-04-13 20:47:34 +02:00
Get my name claims
2016-03-24 03:27:48 +01:00
2016-04-13 20:47:34 +02:00
Args :
None
Returns
list of name claims
2016-03-24 03:27:48 +01:00
"""
2016-04-13 20:47:34 +02:00
2016-01-26 02:28:05 +01:00
def _clean ( claims ) :
for c in claims :
for k in c . keys ( ) :
2016-02-11 14:32:48 +01:00
if isinstance ( c [ k ] , Decimal ) :
2016-01-26 02:28:05 +01:00
c [ k ] = float ( c [ k ] )
2016-03-29 22:42:47 +02:00
return defer . succeed ( claims )
2016-01-26 02:28:05 +01:00
d = self . session . wallet . get_name_claims ( )
d . addCallback ( _clean )
2016-03-29 22:42:47 +02:00
d . addCallback ( lambda claims : self . _render_response ( claims , OK_CODE ) )
2016-01-26 02:28:05 +01:00
return d
2016-03-24 03:27:48 +01:00
def jsonrpc_get_time_behind_blockchain ( self ) :
"""
2016-04-13 20:47:34 +02:00
Get number of blocks behind the blockchain
2016-03-24 03:27:48 +01:00
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
number of blocks behind blockchain , int
2016-03-24 03:27:48 +01:00
"""
2016-04-07 09:12:09 +02:00
def _get_time_behind ( ) :
try :
local_height = self . session . wallet . network . get_local_height ( )
remote_height = self . session . wallet . network . get_server_height ( )
return defer . succeed ( remote_height - local_height )
except :
return defer . fail ( )
d = _get_time_behind ( )
d . addCallback ( lambda r : self . _render_response ( r , OK_CODE ) )
2016-01-26 02:28:05 +01:00
return d
2016-03-24 03:27:48 +01:00
def jsonrpc_get_new_address ( self ) :
"""
Generate a new wallet address
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
new wallet address , base 58 string
2016-03-24 03:27:48 +01:00
"""
2016-04-13 20:47:34 +02:00
2016-02-25 23:17:07 +01:00
def _disp ( address ) :
2016-03-24 03:27:48 +01:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Got new wallet address: " + address )
2016-03-29 22:42:47 +02:00
return defer . succeed ( address )
2016-02-25 23:17:07 +01:00
d = self . session . wallet . get_new_address ( )
d . addCallback ( _disp )
2016-03-29 22:42:47 +02:00
d . addCallback ( lambda address : self . _render_response ( address , OK_CODE ) )
2016-02-25 23:17:07 +01:00
return d
2016-03-24 03:27:48 +01:00
# def jsonrpc_update_name(self, metadata):
2016-02-11 14:32:48 +01:00
# def _disp(x):
# print x
# return x
#
# metadata = json.loads(metadata)
#
# required = ['name', 'file_path', 'bid']
#
# for r in required:
# if not r in metadata.keys():
# return defer.fail()
#
# d = defer.Deferred()
# d.addCallback(lambda _: self.session.wallet.update_name(metadata))
# d.addCallback(_disp)
# d.callback(None)
#
# return d
2016-03-24 03:27:48 +01:00
def jsonrpc_toggle_fetcher_verbose ( self ) :
2016-04-13 20:47:34 +02:00
"""
Toggle fetcher verbose mode
Args :
None
Returns :
Fetcher verbose status , bool
"""
2016-01-27 17:02:57 +01:00
if self . fetcher . verbose :
self . fetcher . verbose = False
else :
self . fetcher . verbose = True
2016-03-24 03:27:48 +01:00
return self . _render_response ( self . fetcher . verbose , OK_CODE )
2016-01-27 17:02:57 +01:00
2016-03-24 03:27:48 +01:00
def jsonrpc_check_for_new_version ( self ) :
2016-04-13 08:35:31 +02:00
"""
Checks local version against versions in __init__ . py and version . py in the lbrynet and lbryum repos
2016-04-13 20:47:34 +02:00
Args :
None
Returns :
true / false , true meaning that there is a new version available
2016-04-13 08:35:31 +02:00
"""
2016-04-14 04:07:27 +02:00
2016-04-13 08:35:31 +02:00
def _check_version ( ) :
2016-04-14 04:07:27 +02:00
if ( lbrynet_version > = self . git_lbrynet_version ) and ( lbryum_version > = self . git_lbryum_version ) :
2016-04-13 08:35:31 +02:00
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Up to date " )
return self . _render_response ( False , OK_CODE )
else :
log . info ( " [ " + str ( datetime . now ( ) ) + " ] Updates available " )
return self . _render_response ( True , OK_CODE )
2016-04-12 05:13:26 +02:00
2016-04-13 08:35:31 +02:00
return _check_version ( )
2016-01-24 06:10:22 +01:00
2016-04-18 01:54:04 +02:00
def jsonrpc_upload_log ( self , p = None ) :
if p :
if ' name_prefix ' in p . keys ( ) :
prefix = p [ ' name_prefix ' ] + ' _api '
else :
prefix = None
if ' exclude_previous ' in p . keys :
exclude_previous = p [ ' exclude_previous ' ]
else :
exclude_previous = True
else :
prefix = " api "
exclude_previous = True
d = self . _upload_log ( name_prefix = prefix , exclude_previous = exclude_previous )
d . addCallback ( lambda _ : self . _render_response ( True , OK_CODE ) )
return d
2016-02-29 19:25:47 +01:00
2016-03-24 03:27:48 +01:00
class LBRYDaemonCommandHandler ( object ) :
def __init__ ( self , command ) :
self . _api = jsonrpc . Proxy ( API_CONNECTION_STRING )
self . command = command
def run ( self , params = None ) :
if params :
d = self . _api . callRemote ( self . command , params )
else :
d = self . _api . callRemote ( self . command )
return d
2016-02-29 19:25:47 +01:00
2016-03-24 03:27:48 +01:00
class LBRYindex ( resource . Resource ) :
def __init__ ( self , ui_dir ) :
resource . Resource . __init__ ( self )
self . ui_dir = ui_dir
2016-02-29 19:25:47 +01:00
2016-03-24 03:27:48 +01:00
isLeaf = False
2016-02-19 02:41:23 +01:00
2016-03-24 03:27:48 +01:00
def _delayed_render ( self , request , results ) :
request . write ( str ( results ) )
request . finish ( )
2016-02-25 23:17:07 +01:00
2016-03-24 03:27:48 +01:00
def getChild ( self , name , request ) :
if name == ' ' :
return self
return resource . Resource . getChild ( self , name , request )
2016-02-23 04:32:07 +01:00
2016-03-24 03:27:48 +01:00
def render_GET ( self , request ) :
return static . File ( os . path . join ( self . ui_dir , " index.html " ) ) . render_GET ( request )
2015-12-06 23:32:17 +01:00
2016-02-19 02:45:16 +01:00
2016-03-24 03:27:48 +01:00
class LBRYFileRender ( resource . Resource ) :
isLeaf = False
2016-02-29 19:25:47 +01:00
2016-03-24 03:27:48 +01:00
def render_GET ( self , request ) :
if ' name ' in request . args . keys ( ) :
api = jsonrpc . Proxy ( API_CONNECTION_STRING )
2016-04-18 01:54:04 +02:00
if request . args [ ' name ' ] [ 0 ] != ' lbry ' :
d = api . callRemote ( " get " , { ' name ' : request . args [ ' name ' ] [ 0 ] } )
d . addCallback ( lambda results : static . File ( results [ ' path ' ] ) . render_GET ( request ) )
else :
request . redirect ( UI_ADDRESS )
request . finish ( )
2016-03-24 03:27:48 +01:00
return server . NOT_DONE_YET
else :
2016-04-09 05:38:57 +02:00
return server . failure