mirror of
https://github.com/bitcoin/bitcoin.git
synced 2025-02-23 07:15:29 +01:00
This ensures that the disconnect happens for the expected reason and also makes it easier to navigate between implementation and test code, i.e. both the questions "do we have test coverage for this disconnect?" (from an implementation reader's perspective) and "where is the code handling this disconnect?" (from a test reader's perspective) can be answered simply by grep-ping the corresponding debug message.
198 lines
8.2 KiB
Python
Executable file
198 lines
8.2 KiB
Python
Executable file
#!/usr/bin/env python3
|
|
# Copyright (c) 2015-2022 The Bitcoin Core developers
|
|
# Distributed under the MIT software license, see the accompanying
|
|
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
|
|
"""Test behavior of -maxuploadtarget.
|
|
|
|
* Verify that getdata requests for old blocks (>1week) are dropped
|
|
if uploadtarget has been reached.
|
|
* Verify that getdata requests for recent blocks are respected even
|
|
if uploadtarget has been reached.
|
|
* Verify that the upload counters are reset after 24 hours.
|
|
"""
|
|
from collections import defaultdict
|
|
import time
|
|
|
|
from test_framework.messages import (
|
|
CInv,
|
|
MSG_BLOCK,
|
|
msg_getdata,
|
|
)
|
|
from test_framework.p2p import P2PInterface
|
|
from test_framework.test_framework import BitcoinTestFramework
|
|
from test_framework.util import (
|
|
assert_equal,
|
|
mine_large_block,
|
|
)
|
|
from test_framework.wallet import MiniWallet
|
|
|
|
|
|
UPLOAD_TARGET_MB = 800
|
|
|
|
|
|
class TestP2PConn(P2PInterface):
|
|
def __init__(self):
|
|
super().__init__()
|
|
self.block_receive_map = defaultdict(int)
|
|
|
|
def on_inv(self, message):
|
|
pass
|
|
|
|
def on_block(self, message):
|
|
message.block.calc_sha256()
|
|
self.block_receive_map[message.block.sha256] += 1
|
|
|
|
class MaxUploadTest(BitcoinTestFramework):
|
|
|
|
def set_test_params(self):
|
|
self.setup_clean_chain = True
|
|
self.num_nodes = 1
|
|
self.extra_args = [[
|
|
f"-maxuploadtarget={UPLOAD_TARGET_MB}M",
|
|
"-datacarriersize=100000",
|
|
]]
|
|
self.supports_cli = False
|
|
|
|
def assert_uploadtarget_state(self, *, target_reached, serve_historical_blocks):
|
|
"""Verify the node's current upload target state via the `getnettotals` RPC call."""
|
|
uploadtarget = self.nodes[0].getnettotals()["uploadtarget"]
|
|
assert_equal(uploadtarget["target_reached"], target_reached)
|
|
assert_equal(uploadtarget["serve_historical_blocks"], serve_historical_blocks)
|
|
|
|
def run_test(self):
|
|
# Initially, neither historical blocks serving limit nor total limit are reached
|
|
self.assert_uploadtarget_state(target_reached=False, serve_historical_blocks=True)
|
|
|
|
# Before we connect anything, we first set the time on the node
|
|
# to be in the past, otherwise things break because the CNode
|
|
# time counters can't be reset backward after initialization
|
|
old_time = int(time.time() - 2*60*60*24*7)
|
|
self.nodes[0].setmocktime(old_time)
|
|
|
|
# Generate some old blocks
|
|
self.wallet = MiniWallet(self.nodes[0])
|
|
self.generate(self.wallet, 130)
|
|
|
|
# p2p_conns[0] will only request old blocks
|
|
# p2p_conns[1] will only request new blocks
|
|
# p2p_conns[2] will test resetting the counters
|
|
p2p_conns = []
|
|
|
|
for _ in range(3):
|
|
p2p_conns.append(self.nodes[0].add_p2p_connection(TestP2PConn()))
|
|
|
|
# Now mine a big block
|
|
mine_large_block(self, self.wallet, self.nodes[0])
|
|
|
|
# Store the hash; we'll request this later
|
|
big_old_block = self.nodes[0].getbestblockhash()
|
|
old_block_size = self.nodes[0].getblock(big_old_block, True)['size']
|
|
big_old_block = int(big_old_block, 16)
|
|
|
|
# Advance to two days ago
|
|
self.nodes[0].setmocktime(int(time.time()) - 2*60*60*24)
|
|
|
|
# Mine one more block, so that the prior block looks old
|
|
mine_large_block(self, self.wallet, self.nodes[0])
|
|
|
|
# We'll be requesting this new block too
|
|
big_new_block = self.nodes[0].getbestblockhash()
|
|
big_new_block = int(big_new_block, 16)
|
|
|
|
# p2p_conns[0] will test what happens if we just keep requesting the
|
|
# the same big old block too many times (expect: disconnect)
|
|
|
|
getdata_request = msg_getdata()
|
|
getdata_request.inv.append(CInv(MSG_BLOCK, big_old_block))
|
|
|
|
max_bytes_per_day = UPLOAD_TARGET_MB * 1024 *1024
|
|
daily_buffer = 144 * 4000000
|
|
max_bytes_available = max_bytes_per_day - daily_buffer
|
|
success_count = max_bytes_available // old_block_size
|
|
|
|
# 576MB will be reserved for relaying new blocks, so expect this to
|
|
# succeed for ~235 tries.
|
|
for i in range(success_count):
|
|
p2p_conns[0].send_and_ping(getdata_request)
|
|
assert_equal(p2p_conns[0].block_receive_map[big_old_block], i+1)
|
|
|
|
assert_equal(len(self.nodes[0].getpeerinfo()), 3)
|
|
# At most a couple more tries should succeed (depending on how long
|
|
# the test has been running so far).
|
|
with self.nodes[0].assert_debug_log(expected_msgs=["historical block serving limit reached, disconnect peer"]):
|
|
for _ in range(3):
|
|
p2p_conns[0].send_message(getdata_request)
|
|
p2p_conns[0].wait_for_disconnect()
|
|
assert_equal(len(self.nodes[0].getpeerinfo()), 2)
|
|
self.log.info("Peer 0 disconnected after downloading old block too many times")
|
|
|
|
# Historical blocks serving limit is reached by now, but total limit still isn't
|
|
self.assert_uploadtarget_state(target_reached=False, serve_historical_blocks=False)
|
|
|
|
# Requesting the current block on p2p_conns[1] should succeed indefinitely,
|
|
# even when over the max upload target.
|
|
# We'll try 800 times
|
|
getdata_request.inv = [CInv(MSG_BLOCK, big_new_block)]
|
|
for i in range(800):
|
|
p2p_conns[1].send_and_ping(getdata_request)
|
|
assert_equal(p2p_conns[1].block_receive_map[big_new_block], i+1)
|
|
|
|
# Both historical blocks serving limit and total limit are reached
|
|
self.assert_uploadtarget_state(target_reached=True, serve_historical_blocks=False)
|
|
|
|
self.log.info("Peer 1 able to repeatedly download new block")
|
|
|
|
# But if p2p_conns[1] tries for an old block, it gets disconnected too.
|
|
getdata_request.inv = [CInv(MSG_BLOCK, big_old_block)]
|
|
with self.nodes[0].assert_debug_log(expected_msgs=["historical block serving limit reached, disconnect peer"]):
|
|
p2p_conns[1].send_message(getdata_request)
|
|
p2p_conns[1].wait_for_disconnect()
|
|
assert_equal(len(self.nodes[0].getpeerinfo()), 1)
|
|
|
|
self.log.info("Peer 1 disconnected after trying to download old block")
|
|
|
|
self.log.info("Advancing system time on node to clear counters...")
|
|
|
|
# If we advance the time by 24 hours, then the counters should reset,
|
|
# and p2p_conns[2] should be able to retrieve the old block.
|
|
self.nodes[0].setmocktime(int(time.time()))
|
|
p2p_conns[2].sync_with_ping()
|
|
p2p_conns[2].send_and_ping(getdata_request)
|
|
assert_equal(p2p_conns[2].block_receive_map[big_old_block], 1)
|
|
self.assert_uploadtarget_state(target_reached=False, serve_historical_blocks=True)
|
|
|
|
self.log.info("Peer 2 able to download old block")
|
|
|
|
self.nodes[0].disconnect_p2ps()
|
|
|
|
self.log.info("Restarting node 0 with download permission and 1MB maxuploadtarget")
|
|
self.restart_node(0, ["-whitelist=download@127.0.0.1", "-maxuploadtarget=1"])
|
|
# Total limit isn't reached after restart, but 1 MB is too small to serve historical blocks
|
|
self.assert_uploadtarget_state(target_reached=False, serve_historical_blocks=False)
|
|
|
|
# Reconnect to self.nodes[0]
|
|
peer = self.nodes[0].add_p2p_connection(TestP2PConn())
|
|
|
|
#retrieve 20 blocks which should be enough to break the 1MB limit
|
|
getdata_request.inv = [CInv(MSG_BLOCK, big_new_block)]
|
|
for i in range(20):
|
|
peer.send_and_ping(getdata_request)
|
|
assert_equal(peer.block_receive_map[big_new_block], i+1)
|
|
|
|
# Total limit is exceeded
|
|
self.assert_uploadtarget_state(target_reached=True, serve_historical_blocks=False)
|
|
|
|
getdata_request.inv = [CInv(MSG_BLOCK, big_old_block)]
|
|
peer.send_and_ping(getdata_request)
|
|
|
|
self.log.info("Peer still connected after trying to download old block (download permission)")
|
|
peer_info = self.nodes[0].getpeerinfo()
|
|
assert_equal(len(peer_info), 1) # node is still connected
|
|
assert_equal(peer_info[0]['permissions'], ['download'])
|
|
|
|
self.log.info("Test passing an unparsable value to -maxuploadtarget throws an error")
|
|
self.stop_node(0)
|
|
self.nodes[0].assert_start_raises_init_error(extra_args=["-maxuploadtarget=abc"], expected_msg="Error: Unable to parse -maxuploadtarget: 'abc'")
|
|
|
|
if __name__ == '__main__':
|
|
MaxUploadTest().main()
|