From ba80c0e5942ab885dbc41faa734534404e98147f Mon Sep 17 00:00:00 2001 From: Lex Berezhny Date: Sat, 21 Jul 2018 17:11:44 -0400 Subject: [PATCH] py2(iteritems, itervalues) -> py3(items, values) --- lbrynet/conf.py | 4 ++-- lbrynet/core/Wallet.py | 8 ++++---- lbrynet/core/client/DownloadManager.py | 4 ++-- lbrynet/core/client/StandaloneBlobDownloader.py | 2 +- lbrynet/core/client/StreamProgressManager.py | 6 +++--- lbrynet/daemon/Daemon.py | 4 ++-- lbrynet/daemon/DaemonConsole.py | 2 +- lbrynet/database/migrator/migrate5to6.py | 6 +++--- lbrynet/database/storage.py | 2 +- scripts/seed_node.py | 4 ++-- tests/unit/dht/test_routingtable.py | 2 +- tests/unit/lbrynet_daemon/test_docs.py | 2 +- 12 files changed, 23 insertions(+), 23 deletions(-) diff --git a/lbrynet/conf.py b/lbrynet/conf.py index 2bc3223cd..64a200a3a 100644 --- a/lbrynet/conf.py +++ b/lbrynet/conf.py @@ -340,7 +340,7 @@ class Config(object): return self.get_current_settings_dict().__repr__() def __iter__(self): - for k in self._data[TYPE_DEFAULT].iterkeys(): + for k in self._data[TYPE_DEFAULT].keys(): yield k def __getitem__(self, name): @@ -477,7 +477,7 @@ class Config(object): def get_adjustable_settings_dict(self): return { - key: val for key, val in self.get_current_settings_dict().iteritems() + key: val for key, val in self.get_current_settings_dict().items() if key in self._adjustable_defaults } diff --git a/lbrynet/core/Wallet.py b/lbrynet/core/Wallet.py index 338232a5f..e2e633c45 100644 --- a/lbrynet/core/Wallet.py +++ b/lbrynet/core/Wallet.py @@ -326,7 +326,7 @@ class Wallet(object): tx_heights = yield DeferredDict({txid: self.get_height_for_txid(txid) for txid in pending_outpoints}, consumeErrors=True) outpoint_heights = {} - for txid, outputs in pending_outpoints.iteritems(): + for txid, outputs in pending_outpoints.items(): if txid in tx_heights: for nout in outputs: outpoint_heights["%s:%i" % (txid, nout)] = tx_heights[txid] @@ -442,7 +442,7 @@ class Wallet(object): result = {} batch_results = yield self._get_values_for_uris(page, page_size, *uris) to_save = [] - for uri, resolve_results in batch_results.iteritems(): + for uri, resolve_results in batch_results.items(): try: result[uri] = self._handle_claim_result(resolve_results) to_save.append(result[uri]) @@ -454,7 +454,7 @@ class Wallet(object): @defer.inlineCallbacks def get_claims_by_ids(self, *claim_ids): claims = yield self._get_claims_by_claimids(*claim_ids) - for claim in claims.itervalues(): + for claim in claims.values(): yield self.save_claim(claim) defer.returnValue(claims) @@ -1129,7 +1129,7 @@ class LBRYumWallet(Wallet): return payto_out['txid'] log.debug("Doing send many. payments to send: %s", str(payments_to_send)) - d = self._run_cmd_as_defer_succeed('payto', payments_to_send.iteritems()) + d = self._run_cmd_as_defer_succeed('payto', payments_to_send.items()) d.addCallback(lambda out: handle_payto_out(out)) return d diff --git a/lbrynet/core/client/DownloadManager.py b/lbrynet/core/client/DownloadManager.py index 30839ed0c..9c1673fe9 100644 --- a/lbrynet/core/client/DownloadManager.py +++ b/lbrynet/core/client/DownloadManager.py @@ -79,14 +79,14 @@ class DownloadManager(object): return self.blob_handler.handle_blob(self.blobs[blob_num], self.blob_infos[blob_num]) def calculate_total_bytes(self): - return sum([bi.length for bi in self.blob_infos.itervalues()]) + return sum([bi.length for bi in self.blob_infos.values()]) def calculate_bytes_left_to_output(self): if not self.blobs: return self.calculate_total_bytes() else: to_be_outputted = [ - b for n, b in self.blobs.iteritems() + b for n, b in self.blobs.items() if n >= self.progress_manager.last_blob_outputted ] return sum([b.length for b in to_be_outputted if b.length is not None]) diff --git a/lbrynet/core/client/StandaloneBlobDownloader.py b/lbrynet/core/client/StandaloneBlobDownloader.py index b7166afa4..45bd30f8b 100644 --- a/lbrynet/core/client/StandaloneBlobDownloader.py +++ b/lbrynet/core/client/StandaloneBlobDownloader.py @@ -69,7 +69,7 @@ class SingleProgressManager(object): def needed_blobs(self): blobs = self.download_manager.blobs assert len(blobs) == 1 - return [b for b in blobs.itervalues() if not b.get_is_verified()] + return [b for b in blobs.values() if not b.get_is_verified()] class DummyBlobHandler(object): diff --git a/lbrynet/core/client/StreamProgressManager.py b/lbrynet/core/client/StreamProgressManager.py index 5fc36d451..b8cf32811 100644 --- a/lbrynet/core/client/StreamProgressManager.py +++ b/lbrynet/core/client/StreamProgressManager.py @@ -101,15 +101,15 @@ class FullStreamProgressManager(StreamProgressManager): if not blobs: return 0 else: - for i in range(max(blobs.iterkeys())): + for i in range(max(blobs.keys())): if self._done(i, blobs): return i - return max(blobs.iterkeys()) + 1 + return max(blobs.keys()) + 1 def needed_blobs(self): blobs = self.download_manager.blobs return [ - b for n, b in blobs.iteritems() + b for n, b in blobs.items() if not b.get_is_verified() and not n in self.provided_blob_nums ] diff --git a/lbrynet/daemon/Daemon.py b/lbrynet/daemon/Daemon.py index b3a0c22f5..cfe21382f 100644 --- a/lbrynet/daemon/Daemon.py +++ b/lbrynet/daemon/Daemon.py @@ -915,7 +915,7 @@ class Daemon(AuthJSONRPCServer): 'auto_renew_claim_height_delta': int } - for key, setting_type in setting_types.iteritems(): + for key, setting_type in setting_types.items(): if key in new_settings: if isinstance(new_settings[key], setting_type): conf.settings.update({key: new_settings[key]}, @@ -2916,7 +2916,7 @@ class Daemon(AuthJSONRPCServer): hosts = {} if datastore_len: - for k, v in data_store.iteritems(): + for k, v in data_store.items(): for contact, value, lastPublished, originallyPublished, originalPublisherID in v: if contact in hosts: blobs = hosts[contact] diff --git a/lbrynet/daemon/DaemonConsole.py b/lbrynet/daemon/DaemonConsole.py index 65442e751..6c1c9d4a7 100644 --- a/lbrynet/daemon/DaemonConsole.py +++ b/lbrynet/daemon/DaemonConsole.py @@ -122,7 +122,7 @@ def get_methods(daemon): _fn.__doc__ = fn.__doc__ return {name: _fn} - for method_name, method in daemon.callable_methods.iteritems(): + for method_name, method in daemon.callable_methods.items(): locs.update(wrapped(method_name, method)) return locs diff --git a/lbrynet/database/migrator/migrate5to6.py b/lbrynet/database/migrator/migrate5to6.py index 82518e81c..ca03d3fc8 100644 --- a/lbrynet/database/migrator/migrate5to6.py +++ b/lbrynet/database/migrator/migrate5to6.py @@ -247,7 +247,7 @@ def do_migration(db_dir): claim_queries = {} # : claim query tuple # get the claim queries ready, only keep those with associated files - for outpoint, sd_hash in file_outpoints.iteritems(): + for outpoint, sd_hash in file_outpoints.items(): if outpoint in claim_outpoint_queries: claim_queries[sd_hash] = claim_outpoint_queries[outpoint] @@ -260,7 +260,7 @@ def do_migration(db_dir): claim_arg_tup[7], claim_arg_tup[6], claim_arg_tup[8], smart_decode(claim_arg_tup[8]).certificate_id, claim_arg_tup[5], claim_arg_tup[4] ) - for sd_hash, claim_arg_tup in claim_queries.iteritems() if claim_arg_tup + for sd_hash, claim_arg_tup in claim_queries.items() if claim_arg_tup ] # sd_hash, (txid, nout, claim_id, name, sequence, address, height, amount, serialized) ) @@ -268,7 +268,7 @@ def do_migration(db_dir): damaged_stream_sds = [] # import the files and get sd hashes of streams to attempt recovering - for sd_hash, file_query in file_args.iteritems(): + for sd_hash, file_query in file_args.items(): failed_sd = _import_file(*file_query) if failed_sd: damaged_stream_sds.append(failed_sd) diff --git a/lbrynet/database/storage.py b/lbrynet/database/storage.py index d67d91bef..fc3804671 100644 --- a/lbrynet/database/storage.py +++ b/lbrynet/database/storage.py @@ -839,7 +839,7 @@ class SQLiteStorage(WalletDatabase): def save_claim_tx_heights(self, claim_tx_heights): def _save_claim_heights(transaction): - for outpoint, height in claim_tx_heights.iteritems(): + for outpoint, height in claim_tx_heights.items(): transaction.execute( "update claim set height=? where claim_outpoint=? and height=-1", (height, outpoint) diff --git a/scripts/seed_node.py b/scripts/seed_node.py index c94d55de0..1af6e850b 100644 --- a/scripts/seed_node.py +++ b/scripts/seed_node.py @@ -59,7 +59,7 @@ def format_contact(contact): def format_datastore(node): datastore = deepcopy(node._dataStore._dict) result = {} - for key, values in datastore.iteritems(): + for key, values in datastore.items(): contacts = [] for (contact, value, last_published, originally_published, original_publisher_id) in values: contact_dict = format_contact(contact) @@ -201,7 +201,7 @@ class MultiSeedRPCServer(AuthJSONRPCServer): nodes = [] for node_id in [n.node_id.encode('hex') for n in self._nodes]: routing_info = yield self.jsonrpc_node_routing_table(node_id=node_id) - for index, bucket in routing_info.iteritems(): + for index, bucket in routing_info.items(): if ip_address in map(lambda c: c['address'], bucket['contacts']): nodes.append(node_id) break diff --git a/tests/unit/dht/test_routingtable.py b/tests/unit/dht/test_routingtable.py index 3b33dfa5a..7bbaf5818 100644 --- a/tests/unit/dht/test_routingtable.py +++ b/tests/unit/dht/test_routingtable.py @@ -228,7 +228,7 @@ class TreeRoutingTableTest(unittest.TestCase): # # math.log(bucket.rangeMax, 2)) + ")" # # for c in bucket.getContacts(): # # print " contact " + str(c.id) -# # for key, bucket in self.table._replacementCache.iteritems(): +# # for key, bucket in self.table._replacementCache.items(): # # print "Replacement Cache for Bucket " + str(key) # # for c in bucket: # # print " contact " + str(c.id) diff --git a/tests/unit/lbrynet_daemon/test_docs.py b/tests/unit/lbrynet_daemon/test_docs.py index ba246ed95..8e42e0993 100644 --- a/tests/unit/lbrynet_daemon/test_docs.py +++ b/tests/unit/lbrynet_daemon/test_docs.py @@ -6,7 +6,7 @@ from lbrynet.daemon.Daemon import Daemon class DaemonDocsTests(unittest.TestCase): def test_can_parse_api_method_docs(self): failures = [] - for name, fn in Daemon.callable_methods.iteritems(): + for name, fn in Daemon.callable_methods.items(): try: docopt.docopt(fn.__doc__, ()) except docopt.DocoptLanguageError as err: