Browse Source
0.1317a073a
Add RPC test for -maxuploadtarget (Suhas Daftuar)872fee3
Introduce -maxuploadtarget (Jonas Schnelli)
Wladimir J. van der Laan
9 years ago
7 changed files with 394 additions and 1 deletions
@ -0,0 +1,248 @@ |
|||||||
|
#!/usr/bin/env python2 |
||||||
|
# |
||||||
|
# Distributed under the MIT/X11 software license, see the accompanying |
||||||
|
# file COPYING or http://www.opensource.org/licenses/mit-license.php. |
||||||
|
# |
||||||
|
|
||||||
|
from test_framework.mininode import * |
||||||
|
from test_framework.test_framework import BitcoinTestFramework |
||||||
|
from test_framework.util import * |
||||||
|
from test_framework.comptool import wait_until |
||||||
|
import time |
||||||
|
|
||||||
|
''' |
||||||
|
Test behavior of -maxuploadtarget. |
||||||
|
|
||||||
|
* Verify that getdata requests for old blocks (>1week) are dropped |
||||||
|
if uploadtarget has been reached. |
||||||
|
* Verify that getdata requests for recent blocks are respecteved even |
||||||
|
if uploadtarget has been reached. |
||||||
|
* Verify that the upload counters are reset after 24 hours. |
||||||
|
''' |
||||||
|
|
||||||
|
# TestNode: bare-bones "peer". Used mostly as a conduit for a test to sending |
||||||
|
# p2p messages to a node, generating the messages in the main testing logic. |
||||||
|
class TestNode(NodeConnCB): |
||||||
|
def __init__(self): |
||||||
|
NodeConnCB.__init__(self) |
||||||
|
self.create_callback_map() |
||||||
|
self.connection = None |
||||||
|
self.ping_counter = 1 |
||||||
|
self.last_pong = msg_pong() |
||||||
|
self.block_receive_map = {} |
||||||
|
|
||||||
|
def add_connection(self, conn): |
||||||
|
self.connection = conn |
||||||
|
self.peer_disconnected = False |
||||||
|
|
||||||
|
def on_inv(self, conn, message): |
||||||
|
pass |
||||||
|
|
||||||
|
# Track the last getdata message we receive (used in the test) |
||||||
|
def on_getdata(self, conn, message): |
||||||
|
self.last_getdata = message |
||||||
|
|
||||||
|
def on_block(self, conn, message): |
||||||
|
message.block.calc_sha256() |
||||||
|
try: |
||||||
|
self.block_receive_map[message.block.sha256] += 1 |
||||||
|
except KeyError as e: |
||||||
|
self.block_receive_map[message.block.sha256] = 1 |
||||||
|
|
||||||
|
# Spin until verack message is received from the node. |
||||||
|
# We use this to signal that our test can begin. This |
||||||
|
# is called from the testing thread, so it needs to acquire |
||||||
|
# the global lock. |
||||||
|
def wait_for_verack(self): |
||||||
|
def veracked(): |
||||||
|
return self.verack_received |
||||||
|
return wait_until(veracked, timeout=10) |
||||||
|
|
||||||
|
def wait_for_disconnect(self): |
||||||
|
def disconnected(): |
||||||
|
return self.peer_disconnected |
||||||
|
return wait_until(disconnected, timeout=10) |
||||||
|
|
||||||
|
# Wrapper for the NodeConn's send_message function |
||||||
|
def send_message(self, message): |
||||||
|
self.connection.send_message(message) |
||||||
|
|
||||||
|
def on_pong(self, conn, message): |
||||||
|
self.last_pong = message |
||||||
|
|
||||||
|
def on_close(self, conn): |
||||||
|
self.peer_disconnected = True |
||||||
|
|
||||||
|
# Sync up with the node after delivery of a block |
||||||
|
def sync_with_ping(self, timeout=30): |
||||||
|
def received_pong(): |
||||||
|
return (self.last_pong.nonce == self.ping_counter) |
||||||
|
self.connection.send_message(msg_ping(nonce=self.ping_counter)) |
||||||
|
success = wait_until(received_pong, timeout) |
||||||
|
self.ping_counter += 1 |
||||||
|
return success |
||||||
|
|
||||||
|
class MaxUploadTest(BitcoinTestFramework): |
||||||
|
def __init__(self): |
||||||
|
self.utxo = [] |
||||||
|
|
||||||
|
# Some pre-processing to create a bunch of OP_RETURN txouts to insert into transactions we create |
||||||
|
# So we have big transactions and full blocks to fill up our block files |
||||||
|
# create one script_pubkey |
||||||
|
script_pubkey = "6a4d0200" #OP_RETURN OP_PUSH2 512 bytes |
||||||
|
for i in xrange (512): |
||||||
|
script_pubkey = script_pubkey + "01" |
||||||
|
# concatenate 128 txouts of above script_pubkey which we'll insert before the txout for change |
||||||
|
self.txouts = "81" |
||||||
|
for k in xrange(128): |
||||||
|
# add txout value |
||||||
|
self.txouts = self.txouts + "0000000000000000" |
||||||
|
# add length of script_pubkey |
||||||
|
self.txouts = self.txouts + "fd0402" |
||||||
|
# add script_pubkey |
||||||
|
self.txouts = self.txouts + script_pubkey |
||||||
|
|
||||||
|
def add_options(self, parser): |
||||||
|
parser.add_option("--testbinary", dest="testbinary", |
||||||
|
default=os.getenv("BITCOIND", "bitcoind"), |
||||||
|
help="bitcoind binary to test") |
||||||
|
|
||||||
|
def setup_chain(self): |
||||||
|
initialize_chain_clean(self.options.tmpdir, 2) |
||||||
|
|
||||||
|
def setup_network(self): |
||||||
|
# Start a node with maxuploadtarget of 200 MB (/24h) |
||||||
|
self.nodes = [] |
||||||
|
self.nodes.append(start_node(0, self.options.tmpdir, ["-debug", "-maxuploadtarget=200", "-blockmaxsize=999000"])) |
||||||
|
|
||||||
|
def mine_full_block(self, node, address): |
||||||
|
# Want to create a full block |
||||||
|
# We'll generate a 66k transaction below, and 14 of them is close to the 1MB block limit |
||||||
|
for j in xrange(14): |
||||||
|
if len(self.utxo) < 14: |
||||||
|
self.utxo = node.listunspent() |
||||||
|
inputs=[] |
||||||
|
outputs = {} |
||||||
|
t = self.utxo.pop() |
||||||
|
inputs.append({ "txid" : t["txid"], "vout" : t["vout"]}) |
||||||
|
remchange = t["amount"] - Decimal("0.001000") |
||||||
|
outputs[address]=remchange |
||||||
|
# Create a basic transaction that will send change back to ourself after account for a fee |
||||||
|
# And then insert the 128 generated transaction outs in the middle rawtx[92] is where the # |
||||||
|
# of txouts is stored and is the only thing we overwrite from the original transaction |
||||||
|
rawtx = node.createrawtransaction(inputs, outputs) |
||||||
|
newtx = rawtx[0:92] |
||||||
|
newtx = newtx + self.txouts |
||||||
|
newtx = newtx + rawtx[94:] |
||||||
|
# Appears to be ever so slightly faster to sign with SIGHASH_NONE |
||||||
|
signresult = node.signrawtransaction(newtx,None,None,"NONE") |
||||||
|
txid = node.sendrawtransaction(signresult["hex"], True) |
||||||
|
# Mine a full sized block which will be these transactions we just created |
||||||
|
node.generate(1) |
||||||
|
|
||||||
|
def run_test(self): |
||||||
|
# Before we connect anything, we first set the time on the node |
||||||
|
# to be in the past, otherwise things break because the CNode |
||||||
|
# time counters can't be reset backward after initialization |
||||||
|
old_time = int(time.time() - 2*60*60*24*7) |
||||||
|
self.nodes[0].setmocktime(old_time) |
||||||
|
|
||||||
|
# Generate some old blocks |
||||||
|
self.nodes[0].generate(130) |
||||||
|
|
||||||
|
# test_nodes[0] will only request old blocks |
||||||
|
# test_nodes[1] will only request new blocks |
||||||
|
# test_nodes[2] will test resetting the counters |
||||||
|
test_nodes = [] |
||||||
|
connections = [] |
||||||
|
|
||||||
|
for i in xrange(3): |
||||||
|
test_nodes.append(TestNode()) |
||||||
|
connections.append(NodeConn('127.0.0.1', p2p_port(0), self.nodes[0], test_nodes[i])) |
||||||
|
test_nodes[i].add_connection(connections[i]) |
||||||
|
|
||||||
|
NetworkThread().start() # Start up network handling in another thread |
||||||
|
[x.wait_for_verack() for x in test_nodes] |
||||||
|
|
||||||
|
# Test logic begins here |
||||||
|
|
||||||
|
# Now mine a big block |
||||||
|
self.mine_full_block(self.nodes[0], self.nodes[0].getnewaddress()) |
||||||
|
|
||||||
|
# Store the hash; we'll request this later |
||||||
|
big_old_block = self.nodes[0].getbestblockhash() |
||||||
|
old_block_size = self.nodes[0].getblock(big_old_block, True)['size'] |
||||||
|
big_old_block = int(big_old_block, 16) |
||||||
|
|
||||||
|
# Advance to two days ago |
||||||
|
self.nodes[0].setmocktime(int(time.time()) - 2*60*60*24) |
||||||
|
|
||||||
|
# Mine one more block, so that the prior block looks old |
||||||
|
self.mine_full_block(self.nodes[0], self.nodes[0].getnewaddress()) |
||||||
|
|
||||||
|
# We'll be requesting this new block too |
||||||
|
big_new_block = self.nodes[0].getbestblockhash() |
||||||
|
new_block_size = self.nodes[0].getblock(big_new_block)['size'] |
||||||
|
big_new_block = int(big_new_block, 16) |
||||||
|
|
||||||
|
# test_nodes[0] will test what happens if we just keep requesting the |
||||||
|
# the same big old block too many times (expect: disconnect) |
||||||
|
|
||||||
|
getdata_request = msg_getdata() |
||||||
|
getdata_request.inv.append(CInv(2, big_old_block)) |
||||||
|
|
||||||
|
max_bytes_per_day = 200*1024*1024 |
||||||
|
max_bytes_available = max_bytes_per_day - 144*1000000 |
||||||
|
success_count = max_bytes_available / old_block_size |
||||||
|
|
||||||
|
# 144MB will be reserved for relaying new blocks, so expect this to |
||||||
|
# succeed for ~70 tries. |
||||||
|
for i in xrange(success_count): |
||||||
|
test_nodes[0].send_message(getdata_request) |
||||||
|
test_nodes[0].sync_with_ping() |
||||||
|
assert_equal(test_nodes[0].block_receive_map[big_old_block], i+1) |
||||||
|
|
||||||
|
assert_equal(len(self.nodes[0].getpeerinfo()), 3) |
||||||
|
# At most a couple more tries should succeed (depending on how long |
||||||
|
# the test has been running so far). |
||||||
|
for i in xrange(3): |
||||||
|
test_nodes[0].send_message(getdata_request) |
||||||
|
test_nodes[0].wait_for_disconnect() |
||||||
|
assert_equal(len(self.nodes[0].getpeerinfo()), 2) |
||||||
|
print "Peer 0 disconnected after downloading old block too many times" |
||||||
|
|
||||||
|
# Requesting the current block on test_nodes[1] should succeed indefinitely, |
||||||
|
# even when over the max upload target. |
||||||
|
# We'll try 200 times |
||||||
|
getdata_request.inv = [CInv(2, big_new_block)] |
||||||
|
for i in xrange(200): |
||||||
|
test_nodes[1].send_message(getdata_request) |
||||||
|
test_nodes[1].sync_with_ping() |
||||||
|
assert_equal(test_nodes[1].block_receive_map[big_new_block], i+1) |
||||||
|
|
||||||
|
print "Peer 1 able to repeatedly download new block" |
||||||
|
|
||||||
|
# But if test_nodes[1] tries for an old block, it gets disconnected too. |
||||||
|
getdata_request.inv = [CInv(2, big_old_block)] |
||||||
|
test_nodes[1].send_message(getdata_request) |
||||||
|
test_nodes[1].wait_for_disconnect() |
||||||
|
assert_equal(len(self.nodes[0].getpeerinfo()), 1) |
||||||
|
|
||||||
|
print "Peer 1 disconnected after trying to download old block" |
||||||
|
|
||||||
|
print "Advancing system time on node to clear counters..." |
||||||
|
|
||||||
|
# If we advance the time by 24 hours, then the counters should reset, |
||||||
|
# and test_nodes[2] should be able to retrieve the old block. |
||||||
|
self.nodes[0].setmocktime(int(time.time())) |
||||||
|
test_nodes[2].sync_with_ping() |
||||||
|
test_nodes[2].send_message(getdata_request) |
||||||
|
test_nodes[2].sync_with_ping() |
||||||
|
assert_equal(test_nodes[2].block_receive_map[big_old_block], 1) |
||||||
|
|
||||||
|
print "Peer 2 able to download old block" |
||||||
|
|
||||||
|
[c.disconnect_node() for c in connections] |
||||||
|
|
||||||
|
if __name__ == '__main__': |
||||||
|
MaxUploadTest().main() |
Loading…
Reference in new issue