lbrycrd/test/functional/feature_maxuploadtarget.py

165 lines
6.5 KiB
Python
Raw Permalink Normal View History

2016-03-19 20:58:06 +01:00
#!/usr/bin/env python3
2018-07-27 00:36:45 +02:00
# Copyright (c) 2015-2018 The Bitcoin Core developers
2016-03-19 20:58:06 +01:00
# Distributed under the MIT software license, see the accompanying
2015-09-18 21:59:55 +02:00
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
"""Test behavior of -maxuploadtarget.
2015-09-18 21:59:55 +02:00
* Verify that getdata requests for old blocks (>1week) are dropped
if uploadtarget has been reached.
2018-03-18 15:26:45 +01:00
* Verify that getdata requests for recent blocks are respected even
2015-09-18 21:59:55 +02:00
if uploadtarget has been reached.
* Verify that the upload counters are reset after 24 hours.
"""
from collections import defaultdict
import time
2018-07-07 00:10:35 +02:00
from test_framework.messages import CInv, msg_getdata
from test_framework.mininode import P2PInterface
from test_framework.test_framework import BitcoinTestFramework
2018-07-07 00:10:35 +02:00
from test_framework.util import assert_equal, mine_large_block
2015-09-18 21:59:55 +02:00
class TestP2PConn(P2PInterface):
2015-09-18 21:59:55 +02:00
def __init__(self):
super().__init__()
self.block_receive_map = defaultdict(int)
2015-09-18 21:59:55 +02:00
def on_inv(self, message):
2015-09-18 21:59:55 +02:00
pass
def on_block(self, message):
2015-09-18 21:59:55 +02:00
message.block.calc_sha256()
self.block_receive_map[message.block.sha256] += 1
2015-09-18 21:59:55 +02:00
class MaxUploadTest(BitcoinTestFramework):
def set_test_params(self):
self.setup_clean_chain = True
2016-05-15 12:20:15 +02:00
self.num_nodes = 1
self.extra_args = [["-maxuploadtarget=800"]]
# Cache for utxos, as the listunspent may take a long time later in the test
self.utxo_cache = []
def skip_test_if_missing_module(self):
self.skip_if_no_wallet()
2015-09-18 21:59:55 +02:00
def run_test(self):
# Before we connect anything, we first set the time on the node
# to be in the past, otherwise things break because the CNode
# time counters can't be reset backward after initialization
old_time = int(time.time() - 2*60*60*24*7)
self.nodes[0].setmocktime(old_time)
# Generate some old blocks
self.nodes[0].generate(130)
# p2p_conns[0] will only request old blocks
# p2p_conns[1] will only request new blocks
# p2p_conns[2] will test resetting the counters
p2p_conns = []
2015-09-18 21:59:55 +02:00
for _ in range(3):
p2p_conns.append(self.nodes[0].add_p2p_connection(TestP2PConn()))
2015-09-18 21:59:55 +02:00
# Now mine a big block
mine_large_block(self.nodes[0], self.utxo_cache)
2015-09-18 21:59:55 +02:00
# Store the hash; we'll request this later
big_old_block = self.nodes[0].getbestblockhash()
old_block_size = self.nodes[0].getblock(big_old_block, True)['size']
big_old_block = int(big_old_block, 16)
# Advance to two days ago
self.nodes[0].setmocktime(int(time.time()) - 2*60*60*24)
# Mine one more block, so that the prior block looks old
mine_large_block(self.nodes[0], self.utxo_cache)
2015-09-18 21:59:55 +02:00
# We'll be requesting this new block too
big_new_block = self.nodes[0].getbestblockhash()
big_new_block = int(big_new_block, 16)
# p2p_conns[0] will test what happens if we just keep requesting the
2015-09-18 21:59:55 +02:00
# the same big old block too many times (expect: disconnect)
getdata_request = msg_getdata()
getdata_request.inv.append(CInv(2, big_old_block))
max_bytes_per_day = 800*1024*1024
daily_buffer = 144 * 4000000
max_bytes_available = max_bytes_per_day - daily_buffer
success_count = max_bytes_available // old_block_size
# 576MB will be reserved for relaying new blocks, so expect this to
# succeed for ~235 tries.
2016-03-19 20:58:06 +01:00
for i in range(success_count):
p2p_conns[0].send_message(getdata_request)
p2p_conns[0].sync_with_ping()
assert_equal(p2p_conns[0].block_receive_map[big_old_block], i+1)
2015-09-18 21:59:55 +02:00
assert_equal(len(self.nodes[0].getpeerinfo()), 3)
# At most a couple more tries should succeed (depending on how long
2015-09-18 21:59:55 +02:00
# the test has been running so far).
2016-03-19 20:58:06 +01:00
for i in range(3):
p2p_conns[0].send_message(getdata_request)
p2p_conns[0].wait_for_disconnect()
2015-09-18 21:59:55 +02:00
assert_equal(len(self.nodes[0].getpeerinfo()), 2)
2017-03-08 00:46:17 +01:00
self.log.info("Peer 0 disconnected after downloading old block too many times")
2015-09-18 21:59:55 +02:00
# Requesting the current block on p2p_conns[1] should succeed indefinitely,
2015-09-18 21:59:55 +02:00
# even when over the max upload target.
# We'll try 800 times
2015-09-18 21:59:55 +02:00
getdata_request.inv = [CInv(2, big_new_block)]
for i in range(800):
p2p_conns[1].send_message(getdata_request)
p2p_conns[1].sync_with_ping()
assert_equal(p2p_conns[1].block_receive_map[big_new_block], i+1)
2015-09-18 21:59:55 +02:00
2017-03-08 00:46:17 +01:00
self.log.info("Peer 1 able to repeatedly download new block")
2015-09-18 21:59:55 +02:00
# But if p2p_conns[1] tries for an old block, it gets disconnected too.
2015-09-18 21:59:55 +02:00
getdata_request.inv = [CInv(2, big_old_block)]
p2p_conns[1].send_message(getdata_request)
p2p_conns[1].wait_for_disconnect()
2015-09-18 21:59:55 +02:00
assert_equal(len(self.nodes[0].getpeerinfo()), 1)
2017-03-08 00:46:17 +01:00
self.log.info("Peer 1 disconnected after trying to download old block")
2015-09-18 21:59:55 +02:00
2017-03-08 00:46:17 +01:00
self.log.info("Advancing system time on node to clear counters...")
2015-09-18 21:59:55 +02:00
# If we advance the time by 24 hours, then the counters should reset,
# and p2p_conns[2] should be able to retrieve the old block.
2015-09-18 21:59:55 +02:00
self.nodes[0].setmocktime(int(time.time()))
p2p_conns[2].sync_with_ping()
p2p_conns[2].send_message(getdata_request)
p2p_conns[2].sync_with_ping()
assert_equal(p2p_conns[2].block_receive_map[big_old_block], 1)
2015-09-18 21:59:55 +02:00
2017-03-08 00:46:17 +01:00
self.log.info("Peer 2 able to download old block")
2015-09-18 21:59:55 +02:00
self.nodes[0].disconnect_p2ps()
2015-09-18 21:59:55 +02:00
#stop and start node 0 with 1MB maxuploadtarget, whitelist 127.0.0.1
2017-03-08 00:46:17 +01:00
self.log.info("Restarting nodes with -whitelist=127.0.0.1")
self.stop_node(0)
self.start_node(0, ["-whitelist=127.0.0.1", "-maxuploadtarget=1"])
# Reconnect to self.nodes[0]
self.nodes[0].add_p2p_connection(TestP2PConn())
#retrieve 20 blocks which should be enough to break the 1MB limit
getdata_request.inv = [CInv(2, big_new_block)]
2016-03-19 20:58:06 +01:00
for i in range(20):
self.nodes[0].p2p.send_message(getdata_request)
self.nodes[0].p2p.sync_with_ping()
assert_equal(self.nodes[0].p2p.block_receive_map[big_new_block], i+1)
getdata_request.inv = [CInv(2, big_old_block)]
self.nodes[0].p2p.send_and_ping(getdata_request)
assert_equal(len(self.nodes[0].getpeerinfo()), 1) #node is still connected because of the whitelist
self.log.info("Peer still connected after trying to download old block (whitelisted)")
2015-09-18 21:59:55 +02:00
if __name__ == '__main__':
MaxUploadTest().main()