Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 3 additions & 3 deletions test/functional/feature_maxuploadtarget.py
Original file line number Diff line number Diff line change
Expand Up @@ -46,8 +46,8 @@ def set_test_params(self):
self.setup_clean_chain = True
self.num_nodes = 1
self.extra_args = [[
"-maxuploadtarget=400",
"-acceptnonstdtxn=1",
"-maxuploadtarget=800M",
"-datacarriersize=100000",
]]
self.supports_cli = False

Expand Down Expand Up @@ -100,7 +100,7 @@ def run_test(self):
getdata_request = msg_getdata()
getdata_request.inv.append(CInv(MSG_BLOCK, big_old_block))

max_bytes_per_day = 400*1024*1024
max_bytes_per_day = 800*1024*1024
daily_buffer = 144 * MAX_BLOCK_SIZE
max_bytes_available = max_bytes_per_day - daily_buffer
success_count = max_bytes_available // old_block_size
Expand Down
2 changes: 1 addition & 1 deletion test/functional/mempool_limit.py
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@ def set_test_params(self):
self.setup_clean_chain = True
self.num_nodes = 1
self.extra_args = [[
"-acceptnonstdtxn=1",
"-datacarriersize=100000",
"-maxmempool=5",
"-spendzeroconfchange=0",
]]
Expand Down
2 changes: 1 addition & 1 deletion test/functional/mining_prioritisetransaction.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@ def set_test_params(self):
self.num_nodes = 2
self.extra_args = [[
"-printpriority=1",
"-acceptnonstdtxn=1",
"-datacarriersize=100000",
]] * self.num_nodes
self.supports_cli = False

Expand Down
19 changes: 5 additions & 14 deletions test/functional/test_framework/util.py
Original file line number Diff line number Diff line change
Expand Up @@ -602,22 +602,13 @@ def chain_transaction(node, parent_txids, vouts, value, fee, num_outputs):


# Create large OP_RETURN txouts that can be appended to a transaction
# to make it large (helper for constructing large transactions).
# to make it large (helper for constructing large transactions). The
# total serialized size of the txouts is about 66k vbytes.
def gen_return_txouts():
# Some pre-processing to create a bunch of OP_RETURN txouts to insert into transactions we create
# So we have big transactions (and therefore can't fit very many into each block)
# create one script_pubkey
script_pubkey = "6a4d0200" # OP_RETURN OP_PUSH2 512 bytes
for _ in range(512):
script_pubkey = script_pubkey + "01"
# concatenate 128 txouts of above script_pubkey which we'll insert before the txout for change
txouts = []
from .messages import CTxOut
txout = CTxOut()
txout.nValue = 0
txout.scriptPubKey = bytes.fromhex(script_pubkey)
for _ in range(128):
txouts.append(txout)
from .script import CScript, OP_RETURN
txouts = [CTxOut(nValue=0, scriptPubKey=CScript([OP_RETURN, b'\x01'*67437]))]
assert_equal(sum([len(txout.serialize()) for txout in txouts]), 67456)
return txouts


Expand Down
Loading