upgrade pylint and fix lint errors

This commit is contained in:
Lex Berezhny 2021-08-20 22:36:35 -04:00
parent 69ea65835d
commit dc1d9e1c84
17 changed files with 39 additions and 35 deletions

View file

@ -4,8 +4,8 @@ install:
pip install -e . pip install -e .
tools: tools:
pip install mypy==0.701 pylint==2.4.4 pip install pylint==2.10.0
pip install coverage astroid pylint pip install coverage
lint: lint:
pylint --rcfile=setup.cfg lbry pylint --rcfile=setup.cfg lbry

View file

@ -203,7 +203,7 @@ class MaxKeyFee(Setting[dict]):
) )
parser.add_argument( parser.add_argument(
self.no_cli_name, self.no_cli_name,
help=f"Disable maximum key fee check.", help="Disable maximum key fee check.",
dest=self.name, dest=self.name,
const=None, const=None,
action="store_const", action="store_const",

View file

@ -556,7 +556,7 @@ class KademliaProtocol(DatagramProtocol):
address[0], address[1], OLD_PROTOCOL_ERRORS[error_datagram.response] address[0], address[1], OLD_PROTOCOL_ERRORS[error_datagram.response]
) )
def datagram_received(self, datagram: bytes, address: typing.Tuple[str, int]) -> None: # pylint: disable=arguments-differ def datagram_received(self, datagram: bytes, address: typing.Tuple[str, int]) -> None: # pylint: disable=arguments-renamed
try: try:
message = decode_datagram(datagram) message = decode_datagram(datagram)
except (ValueError, TypeError, DecodeError): except (ValueError, TypeError, DecodeError):

View file

@ -50,7 +50,7 @@ class KBucket:
self.peers.append(peer) self.peers.append(peer)
return True return True
else: else:
for i in range(len(self.peers)): for i, _ in enumerate(self.peers):
local_peer = self.peers[i] local_peer = self.peers[i]
if local_peer.node_id == peer.node_id: if local_peer.node_id == peer.node_id:
self.peers.remove(local_peer) self.peers.remove(local_peer)

View file

@ -181,7 +181,7 @@ def decode_datagram(datagram: bytes) -> typing.Union[RequestDatagram, ResponseDa
def make_compact_ip(address: str) -> bytearray: def make_compact_ip(address: str) -> bytearray:
compact_ip = reduce(lambda buff, x: buff + bytearray([int(x)]), address.split('.'), bytearray()) compact_ip = reduce(lambda buff, x: buff + bytearray([int(x)]), address.split('.'), bytearray())
if len(compact_ip) != 4: if len(compact_ip) != 4:
raise ValueError(f"invalid IPv4 length") raise ValueError("invalid IPv4 length")
return compact_ip return compact_ip
@ -190,7 +190,7 @@ def make_compact_address(node_id: bytes, address: str, port: int) -> bytearray:
if not 0 < port < 65536: if not 0 < port < 65536:
raise ValueError(f'Invalid port: {port}') raise ValueError(f'Invalid port: {port}')
if len(node_id) != constants.HASH_BITS // 8: if len(node_id) != constants.HASH_BITS // 8:
raise ValueError(f"invalid node node_id length") raise ValueError("invalid node node_id length")
return compact_ip + port.to_bytes(2, 'big') + node_id return compact_ip + port.to_bytes(2, 'big') + node_id
@ -201,5 +201,5 @@ def decode_compact_address(compact_address: bytes) -> typing.Tuple[bytes, str, i
if not 0 < port < 65536: if not 0 < port < 65536:
raise ValueError(f'Invalid port: {port}') raise ValueError(f'Invalid port: {port}')
if len(node_id) != constants.HASH_BITS // 8: if len(node_id) != constants.HASH_BITS // 8:
raise ValueError(f"invalid node node_id length") raise ValueError("invalid node node_id length")
return node_id, address, port return node_id, address, port

View file

@ -63,7 +63,7 @@ class ErrorClass:
@staticmethod @staticmethod
def get_fields(args): def get_fields(args):
if len(args) > 1: if len(args) > 1:
return f''.join(f'\n{INDENT*2}self.{field} = {field}' for field in args[1:]) return ''.join(f'\n{INDENT*2}self.{field} = {field}' for field in args[1:])
return '' return ''
@staticmethod @staticmethod

View file

@ -101,7 +101,7 @@ class ArgumentParser(argparse.ArgumentParser):
self._optionals.title = 'Options' self._optionals.title = 'Options'
if group_name is None: if group_name is None:
self.epilog = ( self.epilog = (
f"Run 'lbrynet COMMAND --help' for more information on a command or group." "Run 'lbrynet COMMAND --help' for more information on a command or group."
) )
else: else:
self.epilog = ( self.epilog = (

View file

@ -1701,9 +1701,9 @@ class Daemon(metaclass=JSONRPCServerType):
'change': {'gap': change_gap, 'maximum_uses_per_address': change_max_uses}, 'change': {'gap': change_gap, 'maximum_uses_per_address': change_max_uses},
'receiving': {'gap': receiving_gap, 'maximum_uses_per_address': receiving_max_uses}, 'receiving': {'gap': receiving_gap, 'maximum_uses_per_address': receiving_max_uses},
} }
for chain_name in address_changes: for chain_name, changes in address_changes.items():
chain = getattr(account, chain_name) chain = getattr(account, chain_name)
for attr, value in address_changes[chain_name].items(): for attr, value in changes.items():
if value is not None: if value is not None:
setattr(chain, attr, value) setattr(chain, attr, value)
change_made = True change_made = True
@ -2248,7 +2248,7 @@ class Daemon(metaclass=JSONRPCServerType):
raise Exception(f"Could not find claim with url '{url}'.") raise Exception(f"Could not find claim with url '{url}'.")
else: else:
# TODO: use error from lbry.error # TODO: use error from lbry.error
raise Exception(f"Missing argument claim_id or url. ") raise Exception("Missing argument claim_id or url.")
if not allow_duplicate_purchase and txo.purchase_receipt: if not allow_duplicate_purchase and txo.purchase_receipt:
# TODO: use error from lbry.error # TODO: use error from lbry.error
raise Exception( raise Exception(
@ -4099,7 +4099,7 @@ class Daemon(metaclass=JSONRPCServerType):
raise Exception(f"Could not find collection with url '{url}'.") raise Exception(f"Could not find collection with url '{url}'.")
else: else:
# TODO: use error from lbry.error # TODO: use error from lbry.error
raise Exception(f"Missing argument claim_id or url. ") raise Exception("Missing argument claim_id or url.")
page_num, page_size = abs(page), min(abs(page_size), 50) page_num, page_size = abs(page), min(abs(page_size), 50)
items = await self.ledger.resolve_collection(txo, page_size * (page_num - 1), page_size) items = await self.ledger.resolve_collection(txo, page_size * (page_num - 1), page_size)
@ -5072,7 +5072,7 @@ class Daemon(metaclass=JSONRPCServerType):
'buckets': {} 'buckets': {}
} }
for i in range(len(self.dht_node.protocol.routing_table.buckets)): for i, _ in enumerate(self.dht_node.protocol.routing_table.buckets):
result['buckets'][i] = [] result['buckets'][i] = []
for peer in self.dht_node.protocol.routing_table.buckets[i].peers: for peer in self.dht_node.protocol.routing_table.buckets[i].peers:
host = { host = {

View file

@ -123,7 +123,7 @@ class JSONResponseEncoder(JSONEncoder):
self.ledger = ledger self.ledger = ledger
self.include_protobuf = include_protobuf self.include_protobuf = include_protobuf
def default(self, obj): # pylint: disable=method-hidden,arguments-differ,too-many-return-statements def default(self, obj): # pylint: disable=method-hidden,arguments-renamed,too-many-return-statements
if isinstance(obj, Account): if isinstance(obj, Account):
return self.encode_account(obj) return self.encode_account(obj)
if isinstance(obj, Wallet): if isinstance(obj, Wallet):

View file

@ -617,7 +617,7 @@ class SQLiteStorage(SQLiteMixin):
).fetchall() ).fetchall()
download_dir = binascii.hexlify(self.conf.download_dir.encode()).decode() download_dir = binascii.hexlify(self.conf.download_dir.encode()).decode()
transaction.executemany( transaction.executemany(
f"update file set download_directory=? where stream_hash=?", "update file set download_directory=? where stream_hash=?",
((download_dir, stream_hash) for stream_hash in stream_hashes) ((download_dir, stream_hash) for stream_hash in stream_hashes)
).fetchall() ).fetchall()
await self.db.run_with_foreign_keys_disabled(_recover) await self.db.run_with_foreign_keys_disabled(_recover)
@ -861,6 +861,6 @@ class SQLiteStorage(SQLiteMixin):
transaction.execute('delete from peer').fetchall() transaction.execute('delete from peer').fetchall()
transaction.executemany( transaction.executemany(
'insert into peer(node_id, address, udp_port, tcp_port) values (?, ?, ?, ?)', 'insert into peer(node_id, address, udp_port, tcp_port) values (?, ?, ?, ?)',
tuple([(binascii.hexlify(p.node_id), p.address, p.udp_port, p.tcp_port) for p in peers]) ((binascii.hexlify(p.node_id), p.address, p.udp_port, p.tcp_port) for p in peers)
).fetchall() ).fetchall()
return await self.db.run(_save_kademlia_peers) return await self.db.run(_save_kademlia_peers)

View file

@ -69,8 +69,8 @@ class VideoFileAnalyzer:
version = str(e) version = str(e)
if code != 0 or not version.startswith("ffmpeg"): if code != 0 or not version.startswith("ffmpeg"):
log.warning("Unable to run ffmpeg, but it was requested. Code: %d; Message: %s", code, version) log.warning("Unable to run ffmpeg, but it was requested. Code: %d; Message: %s", code, version)
raise FileNotFoundError(f"Unable to locate or run ffmpeg or ffprobe. Please install FFmpeg " raise FileNotFoundError("Unable to locate or run ffmpeg or ffprobe. Please install FFmpeg "
f"and ensure that it is callable via PATH or conf.ffmpeg_path") "and ensure that it is callable via PATH or conf.ffmpeg_path")
log.debug("Using %s at %s", version.splitlines()[0].split(" Copyright")[0], self._which_ffmpeg) log.debug("Using %s at %s", version.splitlines()[0].split(" Copyright")[0], self._which_ffmpeg)
return version return version

View file

@ -254,7 +254,7 @@ class ManagedStream(ManagedDownloadSource):
self.finished_writing.clear() self.finished_writing.clear()
self.started_writing.clear() self.started_writing.clear()
try: try:
open(output_path, 'wb').close() open(output_path, 'wb').close() # pylint: disable=consider-using-with
async for blob_info, decrypted in self._aiter_read_stream(connection_id=self.SAVING_ID): async for blob_info, decrypted in self._aiter_read_stream(connection_id=self.SAVING_ID):
log.info("write blob %i/%i", blob_info.blob_num + 1, len(self.descriptor.blobs) - 1) log.info("write blob %i/%i", blob_info.blob_num + 1, len(self.descriptor.blobs) - 1)
await self.loop.run_in_executor(None, self._write_decrypted_blob, output_path, decrypted) await self.loop.run_in_executor(None, self._write_decrypted_blob, output_path, decrypted)

View file

@ -155,7 +155,7 @@ def async_timed_cache(duration: int):
async def _inner(*args, **kwargs) -> typing.Any: async def _inner(*args, **kwargs) -> typing.Any:
loop = asyncio.get_running_loop() loop = asyncio.get_running_loop()
time_now = loop.time() time_now = loop.time()
key = tuple([args, tuple([tuple([k, kwargs[k]]) for k in kwargs])]) key = (args, tuple(kwargs.items()))
if key in cache and (time_now - cache[key][1] < duration): if key in cache and (time_now - cache[key][1] < duration):
return cache[key][0] return cache[key][0]
to_cache = await func(*args, **kwargs) to_cache = await func(*args, **kwargs)
@ -173,7 +173,7 @@ def cache_concurrent(async_fn):
@functools.wraps(async_fn) @functools.wraps(async_fn)
async def wrapper(*args, **kwargs): async def wrapper(*args, **kwargs):
key = tuple([args, tuple([tuple([k, kwargs[k]]) for k in kwargs])]) key = (args, tuple(kwargs.items()))
cache[key] = cache.get(key) or asyncio.create_task(async_fn(*args, **kwargs)) cache[key] = cache.get(key) or asyncio.create_task(async_fn(*args, **kwargs))
try: try:
return await cache[key] return await cache[key]
@ -342,7 +342,7 @@ def lru_cache_concurrent(cache_size: typing.Optional[int] = None,
@functools.wraps(async_fn) @functools.wraps(async_fn)
async def _inner(*args, **kwargs): async def _inner(*args, **kwargs):
key = tuple([args, tuple([tuple([k, kwargs[k]]) for k in kwargs])]) key = (args, tuple(kwargs.items()))
if key in lru_cache: if key in lru_cache:
return lru_cache.get(key) return lru_cache.get(key)

View file

@ -250,7 +250,7 @@ class Account:
generator_name = address_generator.get('name', HierarchicalDeterministic.name) generator_name = address_generator.get('name', HierarchicalDeterministic.name)
self.address_generator = self.address_generators[generator_name] self.address_generator = self.address_generators[generator_name]
self.receiving, self.change = self.address_generator.from_dict(self, address_generator) self.receiving, self.change = self.address_generator.from_dict(self, address_generator)
self.address_managers = {am.chain_number: am for am in {self.receiving, self.change}} self.address_managers = {am.chain_number: am for am in (self.receiving, self.change)}
self.channel_keys = channel_keys self.channel_keys = channel_keys
ledger.add_account(self) ledger.add_account(self)
wallet.add_account(self) wallet.add_account(self)

View file

@ -46,9 +46,11 @@ class _KeyBase:
if len(raw_serkey) != 33: if len(raw_serkey) != 33:
raise ValueError('raw_serkey must have length 33') raise ValueError('raw_serkey must have length 33')
return (ver_bytes + bytes((self.depth,)) return (
ver_bytes + bytes((self.depth,))
+ self.parent_fingerprint() + self.n.to_bytes(4, 'big') + self.parent_fingerprint() + self.n.to_bytes(4, 'big')
+ self.chain_code + raw_serkey) + self.chain_code + raw_serkey
)
def identifier(self): def identifier(self):
raise NotImplementedError raise NotImplementedError

View file

@ -82,10 +82,10 @@ class AIOSQLite:
"read_count", "Number of database reads", namespace="daemon_database" "read_count", "Number of database reads", namespace="daemon_database"
) )
acquire_write_lock_metric = Histogram( acquire_write_lock_metric = Histogram(
f'write_lock_acquired', 'Time to acquire the write lock', namespace="daemon_database", buckets=HISTOGRAM_BUCKETS 'write_lock_acquired', 'Time to acquire the write lock', namespace="daemon_database", buckets=HISTOGRAM_BUCKETS
) )
held_write_lock_metric = Histogram( held_write_lock_metric = Histogram(
f'write_lock_held', 'Length of time the write lock is held for', namespace="daemon_database", 'write_lock_held', 'Length of time the write lock is held for', namespace="daemon_database",
buckets=HISTOGRAM_BUCKETS buckets=HISTOGRAM_BUCKETS
) )
@ -506,7 +506,7 @@ def _get_spendable_utxos(transaction: sqlite3.Connection, accounts: List, decode
amount_to_reserve: int, reserved_amount: int, floor: int, ceiling: int, amount_to_reserve: int, reserved_amount: int, floor: int, ceiling: int,
fee_per_byte: int) -> int: fee_per_byte: int) -> int:
accounts_fmt = ",".join(["?"] * len(accounts)) accounts_fmt = ",".join(["?"] * len(accounts))
txo_query = f""" txo_query = """
SELECT tx.txid, txo.txoid, tx.raw, tx.height, txo.position as nout, tx.is_verified, txo.amount FROM txo SELECT tx.txid, txo.txoid, tx.raw, tx.height, txo.position as nout, tx.is_verified, txo.amount FROM txo
INNER JOIN account_address USING (address) INNER JOIN account_address USING (address)
LEFT JOIN txi USING (txoid) LEFT JOIN txi USING (txoid)

View file

@ -21,6 +21,7 @@ disable=
c-extension-no-member, c-extension-no-member,
fixme, fixme,
broad-except, broad-except,
raise-missing-from,
no-else-return, no-else-return,
cyclic-import, cyclic-import,
missing-docstring, missing-docstring,
@ -37,5 +38,6 @@ disable=
too-many-public-methods, too-many-public-methods,
too-many-return-statements, too-many-return-statements,
too-many-instance-attributes, too-many-instance-attributes,
unspecified-encoding,
protected-access, protected-access,
unused-argument unused-argument