2016-12-21 19:43:13 +01:00
|
|
|
import base58
|
2016-10-31 22:19:19 +01:00
|
|
|
import json
|
2016-10-28 22:12:51 +02:00
|
|
|
import logging
|
2016-08-22 00:44:16 +02:00
|
|
|
import os
|
2016-09-21 09:49:52 +02:00
|
|
|
import sys
|
2016-10-31 22:19:19 +01:00
|
|
|
import yaml
|
2016-11-16 20:38:43 +01:00
|
|
|
import envparse
|
2017-01-17 04:23:20 +01:00
|
|
|
from appdirs import user_data_dir
|
2016-12-21 19:43:13 +01:00
|
|
|
from lbrynet.core import utils
|
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
ENV_NAMESPACE = 'LBRY_'
|
2016-12-21 19:43:13 +01:00
|
|
|
|
2016-11-10 20:26:21 +01:00
|
|
|
LBRYCRD_WALLET = 'lbrycrd'
|
|
|
|
LBRYUM_WALLET = 'lbryum'
|
|
|
|
PTC_WALLET = 'ptc'
|
2016-10-28 22:12:51 +02:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
PROTOCOL_PREFIX = 'lbry'
|
|
|
|
APP_NAME = 'LBRY'
|
2016-10-28 22:12:51 +02:00
|
|
|
|
2016-09-21 09:49:52 +02:00
|
|
|
LINUX = 1
|
|
|
|
DARWIN = 2
|
|
|
|
WINDOWS = 3
|
2017-01-17 04:23:20 +01:00
|
|
|
KB = 2 ** 10
|
|
|
|
MB = 2 ** 20
|
|
|
|
|
|
|
|
DEFAULT_DHT_NODES = [
|
|
|
|
('lbrynet1.lbry.io', 4444),
|
|
|
|
('lbrynet2.lbry.io', 4444),
|
|
|
|
('lbrynet3.lbry.io', 4444)
|
|
|
|
]
|
2016-09-21 09:49:52 +02:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
settings_decoders = {
|
|
|
|
'.json': json.loads,
|
|
|
|
'.yml': yaml.load
|
|
|
|
}
|
2016-10-23 07:17:24 +02:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
settings_encoders = {
|
|
|
|
'.json': json.dumps,
|
|
|
|
'.yml': yaml.safe_dump
|
|
|
|
}
|
|
|
|
|
|
|
|
if sys.platform.startswith('darwin'):
|
2016-09-21 09:49:52 +02:00
|
|
|
platform = DARWIN
|
2017-01-17 04:23:20 +01:00
|
|
|
default_download_directory = os.path.join(os.path.expanduser('~'), 'Downloads')
|
|
|
|
default_data_dir = user_data_dir('LBRY')
|
|
|
|
default_lbryum_dir = os.path.join(os.path.expanduser('~'), '.lbryum')
|
|
|
|
elif sys.platform.startswith('win'):
|
2016-10-19 06:39:19 +02:00
|
|
|
platform = WINDOWS
|
2016-10-19 06:12:44 +02:00
|
|
|
from lbrynet.winhelpers.knownpaths import get_path, FOLDERID, UserHandle
|
2017-01-17 04:23:20 +01:00
|
|
|
|
2016-10-19 06:12:44 +02:00
|
|
|
default_download_directory = get_path(FOLDERID.Downloads, UserHandle.current)
|
2016-10-27 17:49:28 +02:00
|
|
|
default_data_dir = os.path.join(
|
2017-01-17 04:23:20 +01:00
|
|
|
get_path(FOLDERID.RoamingAppData, UserHandle.current), 'lbrynet')
|
2016-10-27 17:49:28 +02:00
|
|
|
default_lbryum_dir = os.path.join(
|
2017-01-17 04:23:20 +01:00
|
|
|
get_path(FOLDERID.RoamingAppData, UserHandle.current), 'lbryum')
|
2016-10-19 06:39:19 +02:00
|
|
|
else:
|
|
|
|
platform = LINUX
|
2017-01-17 04:23:20 +01:00
|
|
|
default_download_directory = os.path.join(os.path.expanduser('~'), 'Downloads')
|
|
|
|
default_data_dir = os.path.join(os.path.expanduser('~'), '.lbrynet')
|
|
|
|
default_lbryum_dir = os.path.join(os.path.expanduser('~'), '.lbryum')
|
2016-10-19 06:12:44 +02:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
ICON_PATH = 'icons' if platform is WINDOWS else 'app.icns'
|
2016-12-21 20:55:43 +01:00
|
|
|
|
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
def server_port(server_and_port):
|
|
|
|
server, port = server_and_port.split(':')
|
|
|
|
return server, int(port)
|
2016-10-26 09:16:33 +02:00
|
|
|
|
|
|
|
|
2016-11-16 20:38:43 +01:00
|
|
|
class Env(envparse.Env):
|
|
|
|
"""An Env parser that automatically namespaces the variables with LBRY"""
|
2017-01-17 04:23:20 +01:00
|
|
|
|
2016-11-16 20:38:43 +01:00
|
|
|
def __init__(self, **schema):
|
|
|
|
self.original_schema = schema
|
|
|
|
my_schema = {
|
|
|
|
self._convert_key(key): self._convert_value(value)
|
|
|
|
for key, value in schema.items()
|
2017-01-17 18:29:09 +01:00
|
|
|
}
|
2016-11-16 20:38:43 +01:00
|
|
|
envparse.Env.__init__(self, **my_schema)
|
|
|
|
|
|
|
|
def __call__(self, key, *args, **kwargs):
|
|
|
|
my_key = self._convert_key(key)
|
|
|
|
return super(Env, self).__call__(my_key, *args, **kwargs)
|
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
@staticmethod
|
|
|
|
def _convert_key(key):
|
|
|
|
return ENV_NAMESPACE + key.upper()
|
2016-11-16 20:38:43 +01:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
@staticmethod
|
|
|
|
def _convert_value(value):
|
2017-01-17 18:29:09 +01:00
|
|
|
""" Allow value to be specified as a tuple or list.
|
|
|
|
|
|
|
|
If you do this, the tuple/list must be of the
|
|
|
|
form (cast, default) or (cast, default, subcast)
|
2016-11-16 20:38:43 +01:00
|
|
|
"""
|
|
|
|
if isinstance(value, (tuple, list)):
|
|
|
|
new_value = {'cast': value[0], 'default': value[1]}
|
|
|
|
if len(value) == 3:
|
|
|
|
new_value['subcast'] = value[2]
|
|
|
|
return new_value
|
|
|
|
return value
|
|
|
|
|
2017-01-17 18:29:09 +01:00
|
|
|
TYPE_DEFAULT = 'default'
|
|
|
|
TYPE_PERSISTED = 'persisted'
|
|
|
|
TYPE_ENV = 'env'
|
|
|
|
TYPE_CLI = 'cli'
|
|
|
|
TYPE_RUNTIME = 'runtime'
|
2016-11-16 20:38:43 +01:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
FIXED_SETTINGS = {
|
|
|
|
'ANALYTICS_ENDPOINT': 'https://api.segment.io/v1',
|
|
|
|
'ANALYTICS_TOKEN': 'Ax5LZzR1o3q3Z3WjATASDwR5rKyHH0qOIRIbLmMXn2H=',
|
|
|
|
'API_ADDRESS': 'lbryapi',
|
|
|
|
'APP_NAME': APP_NAME,
|
|
|
|
'BLOBFILES_DIR': 'blobfiles',
|
|
|
|
'BLOB_SIZE': 2 * MB,
|
|
|
|
'CRYPTSD_FILE_EXTENSION': '.cryptsd',
|
|
|
|
'CURRENCIES': {
|
|
|
|
'BTC': {'type': 'crypto'},
|
|
|
|
'LBC': {'type': 'crypto'},
|
|
|
|
'USD': {'type': 'fiat'},
|
|
|
|
},
|
|
|
|
'DB_REVISION_FILE_NAME': 'db_revision',
|
|
|
|
'ICON_PATH': ICON_PATH,
|
|
|
|
'LOGGLY_TOKEN': 'LJEzATH4AzRgAwxjAP00LwZ2YGx3MwVgZTMuBQZ3MQuxLmOv',
|
|
|
|
'LOG_FILE_NAME': 'lbrynet.log',
|
|
|
|
'LOG_POST_URL': 'https://lbry.io/log-upload',
|
|
|
|
'MAX_BLOB_REQUEST_SIZE': 64 * KB,
|
|
|
|
'MAX_HANDSHAKE_SIZE': 64 * KB,
|
|
|
|
'MAX_REQUEST_SIZE': 64 * KB,
|
|
|
|
'MAX_RESPONSE_INFO_SIZE': 64 * KB,
|
|
|
|
'MAX_BLOB_INFOS_TO_REQUEST': 20,
|
|
|
|
'PROTOCOL_PREFIX': PROTOCOL_PREFIX,
|
|
|
|
'SLACK_WEBHOOK': ('nUE0pUZ6Yl9bo29epl5moTSwnl5wo20ip2IlqzywMKZiIQSFZR5'
|
|
|
|
'AHx4mY0VmF0WQZ1ESEP9kMHZlp1WzJwWOoKN3ImR1M2yUAaMyqGZ='),
|
|
|
|
'SOURCE_TYPES': ['lbry_sd_hash', 'url', 'btih'],
|
|
|
|
'WALLET_TYPES': [LBRYUM_WALLET, LBRYCRD_WALLET],
|
|
|
|
}
|
2016-11-16 20:38:43 +01:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
ADJUSTABLE_SETTINGS = {
|
2016-11-22 21:40:52 +01:00
|
|
|
# By default, daemon will block all cross origin requests
|
|
|
|
# but if this is set, this value will be used for the
|
|
|
|
# Access-Control-Allow-Origin. For example
|
|
|
|
# set to '*' to allow all requests, or set to 'http://localhost:8080'
|
|
|
|
# if you're running a test UI on that port
|
2017-01-17 04:23:20 +01:00
|
|
|
'allowed_origin': (str, ''),
|
|
|
|
|
|
|
|
# Changing this value is not-advised as it could potentially
|
|
|
|
# expose the lbrynet daemon to the outside world which would
|
|
|
|
# give an attacker access to your wallet and you could lose
|
|
|
|
# all of your credits.
|
|
|
|
'api_host': (str, 'localhost'),
|
|
|
|
|
|
|
|
'api_port': (int, 5279),
|
|
|
|
'bittrex_feed': (str, 'https://bittrex.com/api/v1.1/public/getmarkethistory'),
|
|
|
|
'cache_time': (int, 150),
|
|
|
|
'check_ui_requirements': (bool, True),
|
|
|
|
'data_dir': (str, default_data_dir),
|
|
|
|
'data_rate': (float, .0001), # points/megabyte
|
|
|
|
'default_ui_branch': (str, 'master'),
|
|
|
|
'delete_blobs_on_remove': (bool, True),
|
|
|
|
'dht_node_port': (int, 4444),
|
|
|
|
'download_directory': (str, default_download_directory),
|
|
|
|
'download_timeout': (int, 30),
|
|
|
|
'host_ui': (bool, True),
|
|
|
|
'is_generous_host': (bool, True),
|
|
|
|
'known_dht_nodes': (list, DEFAULT_DHT_NODES, server_port),
|
|
|
|
|
|
|
|
# TODO: this should not be configured; move it elsewhere
|
|
|
|
'last_version': (dict, {'lbrynet': '0.0.1', 'lbryum': '0.0.1'}),
|
|
|
|
|
|
|
|
'lbryum_wallet_dir': (str, default_lbryum_dir),
|
|
|
|
'local_ui_path': (str, ''),
|
|
|
|
'max_connections_per_stream': (int, 5),
|
|
|
|
'max_download': (float, 0.0),
|
|
|
|
|
2016-11-22 18:50:54 +01:00
|
|
|
# TODO: this field is more complicated than it needs to be because
|
|
|
|
# it goes through a Fee validator when loaded by the exchange rate
|
|
|
|
# manager. Look into refactoring the exchange rate conversion to
|
|
|
|
# take in a simpler form.
|
|
|
|
#
|
2016-11-16 20:38:43 +01:00
|
|
|
# TODO: writing json on the cmd line is a pain, come up with a nicer
|
2017-01-17 04:23:20 +01:00
|
|
|
# parser for this data structure. (maybe MAX_KEY_FEE': USD:25
|
|
|
|
'max_key_fee': (json.loads, {'USD': {'amount': 25.0, 'address': ''}}),
|
|
|
|
|
|
|
|
'max_search_results': (int, 25),
|
|
|
|
'max_upload': (float, 0.0),
|
|
|
|
'min_info_rate': (float, .02), # points/1000 infos
|
|
|
|
'min_valuable_hash_rate': (float, .05), # points/1000 infos
|
|
|
|
'min_valuable_info_rate': (float, .05), # points/1000 infos
|
|
|
|
'peer_port': (int, 3333),
|
|
|
|
'pointtrader_server': (str, 'http://127.0.0.1:2424'),
|
|
|
|
'reflector_port': (int, 5566),
|
2017-03-16 22:48:28 +01:00
|
|
|
'reflect_uploads': (bool, True),
|
2017-01-17 04:23:20 +01:00
|
|
|
'reflector_servers': (list, [('reflector.lbry.io', 5566)], server_port),
|
|
|
|
'run_on_startup': (bool, False),
|
|
|
|
'run_reflector_server': (bool, False),
|
|
|
|
'sd_download_timeout': (int, 3),
|
2017-03-29 17:06:04 +02:00
|
|
|
'share_debug_info': (bool, True), # whether to share diagnostic info with LBRY
|
2017-02-16 05:38:33 +01:00
|
|
|
'peer_search_timeout': (int, 3),
|
2017-01-17 04:23:20 +01:00
|
|
|
'search_servers': (list, ['lighthouse1.lbry.io:50005']),
|
|
|
|
'search_timeout': (float, 5.0),
|
|
|
|
'startup_scripts': (list, []),
|
|
|
|
'ui_branch': (str, 'master'),
|
|
|
|
'use_auth_http': (bool, False),
|
|
|
|
'use_upnp': (bool, True),
|
|
|
|
'wallet': (str, LBRYUM_WALLET),
|
|
|
|
}
|
2016-10-14 08:13:37 +02:00
|
|
|
|
2016-10-23 07:17:24 +02:00
|
|
|
|
2017-01-17 18:29:09 +01:00
|
|
|
class Config(object):
|
|
|
|
def __init__(self, fixed_defaults, adjustable_defaults, persisted_settings=None,
|
|
|
|
environment=None, cli_settings=None):
|
|
|
|
|
2017-02-02 16:23:17 +01:00
|
|
|
self._installation_id = None
|
2017-01-17 04:23:20 +01:00
|
|
|
self._session_id = base58.b58encode(utils.generate_id())
|
2016-10-31 22:19:19 +01:00
|
|
|
|
2017-01-17 18:29:09 +01:00
|
|
|
self._fixed_defaults = fixed_defaults
|
|
|
|
self._adjustable_defaults = adjustable_defaults
|
|
|
|
|
|
|
|
self._data = {
|
|
|
|
TYPE_DEFAULT: {}, # defaults
|
|
|
|
TYPE_PERSISTED: {}, # stored settings from daemon_settings.yml (or from a db, etc)
|
|
|
|
TYPE_ENV: {}, # settings from environment variables
|
|
|
|
TYPE_CLI: {}, # command-line arguments
|
|
|
|
TYPE_RUNTIME: {}, # set during runtime (using self.set(), etc)
|
|
|
|
}
|
|
|
|
|
|
|
|
# the order in which a piece of data is searched for. earlier types override later types
|
|
|
|
self._search_order = (
|
|
|
|
TYPE_RUNTIME, TYPE_CLI, TYPE_ENV, TYPE_PERSISTED, TYPE_DEFAULT
|
|
|
|
)
|
|
|
|
|
|
|
|
self._data[TYPE_DEFAULT].update(self._fixed_defaults)
|
|
|
|
self._data[TYPE_DEFAULT].update(
|
|
|
|
{k: v[1] for (k, v) in self._adjustable_defaults.iteritems()})
|
2016-10-31 22:19:19 +01:00
|
|
|
|
2017-01-17 18:29:09 +01:00
|
|
|
if persisted_settings is None:
|
|
|
|
persisted_settings = {}
|
|
|
|
self._validate_settings(persisted_settings)
|
|
|
|
self._data[TYPE_PERSISTED].update(persisted_settings)
|
|
|
|
|
|
|
|
env_settings = self._parse_environment(environment)
|
|
|
|
self._validate_settings(env_settings)
|
|
|
|
self._data[TYPE_ENV].update(env_settings)
|
|
|
|
|
|
|
|
if cli_settings is None:
|
|
|
|
cli_settings = {}
|
|
|
|
self._validate_settings(cli_settings)
|
|
|
|
self._data[TYPE_CLI].update(cli_settings)
|
2016-10-31 22:19:19 +01:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
def __repr__(self):
|
2017-01-17 18:29:09 +01:00
|
|
|
return self.get_current_settings_dict().__repr__()
|
2016-11-05 19:23:48 +01:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
def __iter__(self):
|
2017-01-17 18:29:09 +01:00
|
|
|
for k in self._data[TYPE_DEFAULT].iterkeys():
|
2017-01-17 04:23:20 +01:00
|
|
|
yield k
|
2016-11-05 19:23:48 +01:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
def __getitem__(self, name):
|
|
|
|
return self.get(name)
|
|
|
|
|
|
|
|
def __setitem__(self, name, value):
|
|
|
|
return self.set(name, value)
|
|
|
|
|
|
|
|
def __contains__(self, name):
|
2017-01-17 18:29:09 +01:00
|
|
|
return name in self._data[TYPE_DEFAULT]
|
2017-01-17 04:23:20 +01:00
|
|
|
|
2017-01-17 18:29:09 +01:00
|
|
|
@staticmethod
|
|
|
|
def _parse_environment(environment):
|
2017-01-17 04:23:20 +01:00
|
|
|
env_settings = {}
|
2017-01-17 18:29:09 +01:00
|
|
|
if environment is not None:
|
2017-01-17 04:23:20 +01:00
|
|
|
assert isinstance(environment, Env)
|
|
|
|
for opt in environment.original_schema:
|
2017-01-20 19:38:49 +01:00
|
|
|
if environment(opt) is not None:
|
|
|
|
env_settings[opt] = environment(opt)
|
2017-01-17 04:23:20 +01:00
|
|
|
return env_settings
|
|
|
|
|
2017-01-17 18:29:09 +01:00
|
|
|
def _assert_valid_data_type(self, data_type):
|
|
|
|
assert data_type in self._data, KeyError('{} in is not a valid data type'.format(data_type))
|
|
|
|
|
2017-01-20 19:38:49 +01:00
|
|
|
def get_valid_setting_names(self):
|
|
|
|
return self._data[TYPE_DEFAULT].keys()
|
|
|
|
|
2017-01-17 18:29:09 +01:00
|
|
|
def _is_valid_setting(self, name):
|
2017-01-20 19:38:49 +01:00
|
|
|
return name in self.get_valid_setting_names()
|
2017-01-17 18:29:09 +01:00
|
|
|
|
|
|
|
def _assert_valid_setting(self, name):
|
|
|
|
assert self._is_valid_setting(name), \
|
2017-01-20 19:38:49 +01:00
|
|
|
KeyError('{} is not a valid setting'.format(name))
|
2017-01-17 18:29:09 +01:00
|
|
|
|
|
|
|
def _validate_settings(self, data):
|
2017-01-20 19:38:49 +01:00
|
|
|
invalid_settings = set(data.keys()) - set(self.get_valid_setting_names())
|
|
|
|
if len(invalid_settings) > 0:
|
|
|
|
raise KeyError('invalid settings: {}'.format(', '.join(invalid_settings)))
|
2017-01-17 18:29:09 +01:00
|
|
|
|
|
|
|
def _assert_editable_setting(self, name):
|
|
|
|
self._assert_valid_setting(name)
|
|
|
|
assert name not in self._fixed_defaults, \
|
2017-01-20 19:38:49 +01:00
|
|
|
ValueError('{} is not an editable setting'.format(name))
|
2017-01-17 18:29:09 +01:00
|
|
|
|
|
|
|
def get(self, name, data_type=None):
|
|
|
|
"""Get a config value
|
|
|
|
|
|
|
|
Args:
|
|
|
|
name: the name of the value to get
|
|
|
|
data_type: if given, get the value from a specific data set (see below)
|
|
|
|
|
|
|
|
Returns: the config value for the given name
|
|
|
|
|
|
|
|
If data_type is None, get() will search for the given name in each data set, in
|
|
|
|
order of precedence. It will return the first value it finds. This is the "effective"
|
|
|
|
value of a config name. For example, ENV values take precedence over DEFAULT values,
|
|
|
|
so if a value is present in ENV and in DEFAULT, the ENV value will be returned
|
|
|
|
"""
|
|
|
|
self._assert_valid_setting(name)
|
|
|
|
if data_type is not None:
|
|
|
|
self._assert_valid_data_type(data_type)
|
|
|
|
return self._data[data_type][name]
|
|
|
|
for data_type in self._search_order:
|
|
|
|
if name in self._data[data_type]:
|
|
|
|
return self._data[data_type][name]
|
|
|
|
raise KeyError('{} is not a valid setting'.format(name))
|
|
|
|
|
|
|
|
def set(self, name, value, data_types=(TYPE_RUNTIME,)):
|
|
|
|
"""Set a config value
|
|
|
|
|
|
|
|
Args:
|
|
|
|
name: the name of the value to set
|
|
|
|
value: the value
|
|
|
|
data_types: what type(s) of data this is
|
|
|
|
|
|
|
|
Returns: None
|
|
|
|
|
|
|
|
By default, this sets the RUNTIME value of a config. If you wish to set other
|
|
|
|
data types (e.g. PERSISTED values to save to a file, CLI values from parsed
|
|
|
|
command-line options, etc), you can specify that with the data_types param
|
|
|
|
"""
|
|
|
|
self._assert_editable_setting(name)
|
|
|
|
for data_type in data_types:
|
|
|
|
self._assert_valid_data_type(data_type)
|
|
|
|
self._data[data_type][name] = value
|
|
|
|
|
|
|
|
def update(self, updated_settings, data_types=(TYPE_RUNTIME,)):
|
2017-01-17 04:23:20 +01:00
|
|
|
for k, v in updated_settings.iteritems():
|
|
|
|
try:
|
2017-01-17 18:29:09 +01:00
|
|
|
self.set(k, v, data_types=data_types)
|
2017-01-17 04:23:20 +01:00
|
|
|
except (KeyError, AssertionError):
|
|
|
|
pass
|
2016-10-31 22:19:19 +01:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
def get_current_settings_dict(self):
|
2017-01-17 18:29:09 +01:00
|
|
|
current_settings = {}
|
2017-01-20 19:38:49 +01:00
|
|
|
for key in self.get_valid_setting_names():
|
|
|
|
current_settings[key] = self.get(key)
|
2017-01-17 18:29:09 +01:00
|
|
|
return current_settings
|
2016-10-31 22:19:19 +01:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
def get_adjustable_settings_dict(self):
|
|
|
|
return {
|
2017-01-20 19:38:49 +01:00
|
|
|
key: val for key, val in self.get_current_settings_dict().iteritems()
|
|
|
|
if key in self._adjustable_defaults
|
2017-01-17 18:29:09 +01:00
|
|
|
}
|
2017-01-17 04:23:20 +01:00
|
|
|
|
|
|
|
def save_conf_file_settings(self):
|
|
|
|
path = self.get_conf_filename()
|
|
|
|
ext = os.path.splitext(path)[1]
|
|
|
|
encoder = settings_encoders.get(ext, False)
|
2017-01-17 18:29:09 +01:00
|
|
|
assert encoder is not False, 'Unknown settings format %s' % ext
|
2017-01-17 04:23:20 +01:00
|
|
|
with open(path, 'w') as settings_file:
|
2017-01-17 18:29:09 +01:00
|
|
|
settings_file.write(encoder(self._data[TYPE_PERSISTED]))
|
2017-01-17 04:23:20 +01:00
|
|
|
|
|
|
|
def load_conf_file_settings(self):
|
|
|
|
path = self.get_conf_filename()
|
|
|
|
ext = os.path.splitext(path)[1]
|
|
|
|
decoder = settings_decoders.get(ext, False)
|
2017-01-17 18:29:09 +01:00
|
|
|
assert decoder is not False, 'Unknown settings format %s' % ext
|
2017-01-17 04:23:20 +01:00
|
|
|
try:
|
|
|
|
with open(path, 'r') as settings_file:
|
|
|
|
data = settings_file.read()
|
|
|
|
decoded = self._fix_old_conf_file_settings(decoder(data))
|
|
|
|
log.info('Loaded settings file: %s', path)
|
2017-01-17 18:29:09 +01:00
|
|
|
self._validate_settings(decoded)
|
|
|
|
self._data[TYPE_PERSISTED].update(decoded)
|
2017-01-17 04:23:20 +01:00
|
|
|
except (IOError, OSError) as err:
|
|
|
|
log.info('%s: Failed to update settings from %s', err, path)
|
|
|
|
|
2017-01-20 18:39:54 +01:00
|
|
|
def _fix_old_conf_file_settings(self, settings_dict):
|
2017-01-17 04:23:20 +01:00
|
|
|
if 'API_INTERFACE' in settings_dict:
|
|
|
|
settings_dict['api_host'] = settings_dict['API_INTERFACE']
|
|
|
|
del settings_dict['API_INTERFACE']
|
|
|
|
if 'startup_scripts' in settings_dict:
|
|
|
|
del settings_dict['startup_scripts']
|
2017-03-29 17:06:04 +02:00
|
|
|
if 'upload_log' in settings_dict:
|
|
|
|
settings_dict['share_debug_info'] = settings_dict['upload_log']
|
|
|
|
del settings_dict['upload_log']
|
2017-01-20 18:39:54 +01:00
|
|
|
for key in settings_dict.keys():
|
|
|
|
if not self._is_valid_setting(key):
|
|
|
|
log.warning('Ignoring invalid conf file setting: %s', key)
|
|
|
|
del settings_dict[key]
|
2017-01-17 04:23:20 +01:00
|
|
|
return settings_dict
|
2016-10-31 22:19:19 +01:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
def ensure_data_dir(self):
|
|
|
|
# although there is a risk of a race condition here we don't
|
|
|
|
# expect there to be multiple processes accessing this
|
|
|
|
# directory so the risk can be ignored
|
|
|
|
if not os.path.isdir(self['data_dir']):
|
|
|
|
os.makedirs(self['data_dir'])
|
|
|
|
return self['data_dir']
|
2016-10-31 22:19:19 +01:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
def get_log_filename(self):
|
|
|
|
"""
|
|
|
|
Return the log file for this platform.
|
|
|
|
Also ensure the containing directory exists.
|
|
|
|
"""
|
|
|
|
return os.path.join(self.ensure_data_dir(), self['LOG_FILE_NAME'])
|
2016-10-31 22:19:19 +01:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
def get_api_connection_string(self):
|
|
|
|
return 'http://%s:%i/%s' % (self['api_host'], self['api_port'], self['API_ADDRESS'])
|
2016-12-01 14:53:58 +01:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
def get_ui_address(self):
|
|
|
|
return 'http://%s:%i' % (self['api_host'], self['api_port'])
|
2016-10-23 07:17:24 +02:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
def get_db_revision_filename(self):
|
|
|
|
return os.path.join(self.ensure_data_dir(), self['DB_REVISION_FILE_NAME'])
|
2016-10-23 07:17:24 +02:00
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
def get_conf_filename(self):
|
|
|
|
data_dir = self.ensure_data_dir()
|
|
|
|
yml_path = os.path.join(data_dir, 'daemon_settings.yml')
|
|
|
|
json_path = os.path.join(data_dir, 'daemon_settings.json')
|
|
|
|
if os.path.isfile(yml_path):
|
|
|
|
return yml_path
|
|
|
|
elif os.path.isfile(json_path):
|
|
|
|
return json_path
|
|
|
|
else:
|
|
|
|
return yml_path
|
|
|
|
|
2017-02-02 16:23:17 +01:00
|
|
|
def get_installation_id(self):
|
|
|
|
install_id_filename = os.path.join(self.ensure_data_dir(), "install_id")
|
|
|
|
if not self._installation_id:
|
|
|
|
if os.path.isfile(install_id_filename):
|
|
|
|
with open(install_id_filename, "r") as install_id_file:
|
|
|
|
self._installation_id = install_id_file.read()
|
|
|
|
if not self._installation_id:
|
|
|
|
self._installation_id = base58.b58encode(utils.generate_id())
|
|
|
|
with open(install_id_filename, "w") as install_id_file:
|
|
|
|
install_id_file.write(self._installation_id)
|
|
|
|
return self._installation_id
|
2017-01-17 04:23:20 +01:00
|
|
|
|
|
|
|
def get_session_id(self):
|
|
|
|
return self._session_id
|
|
|
|
|
|
|
|
|
|
|
|
# type: Config
|
2016-12-21 19:43:13 +01:00
|
|
|
settings = None
|
|
|
|
|
|
|
|
|
2017-01-17 18:29:09 +01:00
|
|
|
def get_default_env():
|
2017-01-20 19:38:49 +01:00
|
|
|
env_defaults = {}
|
|
|
|
for k, v in ADJUSTABLE_SETTINGS.iteritems():
|
|
|
|
if len(v) == 3:
|
|
|
|
env_defaults[k] = (v[0], None, v[2])
|
|
|
|
else:
|
|
|
|
env_defaults[k] = (v[0], None)
|
|
|
|
return Env(**env_defaults)
|
2017-01-17 18:29:09 +01:00
|
|
|
|
|
|
|
|
2017-01-17 04:23:20 +01:00
|
|
|
def initialize_settings(load_conf_file=True):
|
2016-12-21 19:43:13 +01:00
|
|
|
global settings
|
2017-01-17 04:23:20 +01:00
|
|
|
if settings is None:
|
2017-01-17 18:29:09 +01:00
|
|
|
settings = Config(FIXED_SETTINGS, ADJUSTABLE_SETTINGS,
|
|
|
|
environment=get_default_env())
|
2017-02-02 16:23:17 +01:00
|
|
|
settings.installation_id = settings.get_installation_id()
|
2017-01-17 04:23:20 +01:00
|
|
|
if load_conf_file:
|
|
|
|
settings.load_conf_file_settings()
|
2017-02-02 16:23:17 +01:00
|
|
|
|