|
|
|
#!/usr/bin/env python3
|
|
|
|
# Copyright (c) 2014-2016 The Bitcoin Core developers
|
|
|
|
# Distributed under the MIT software license, see the accompanying
|
|
|
|
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
|
|
|
|
"""Run regression test suite.
|
|
|
|
|
|
|
|
This module calls down into individual test cases via subprocess. It will
|
|
|
|
forward all unrecognized arguments onto the individual test scripts.
|
|
|
|
|
|
|
|
Functional tests are disabled on Windows by default. Use --force to run them anyway.
|
|
|
|
|
|
|
|
For a description of arguments recognized by test scripts, see
|
|
|
|
`test/functional/test_framework/test_framework.py:BitcoinTestFramework.main`.
|
|
|
|
|
|
|
|
"""
|
|
|
|
|
|
|
|
import argparse
|
|
|
|
import configparser
|
|
|
|
import os
|
|
|
|
import time
|
|
|
|
import shutil
|
|
|
|
import sys
|
|
|
|
import subprocess
|
|
|
|
import tempfile
|
|
|
|
import re
|
|
|
|
import logging
|
|
|
|
|
|
|
|
# Formatting. Default colors to empty strings.
|
|
|
|
BOLD, BLUE, RED, GREY = ("", ""), ("", ""), ("", ""), ("", "")
|
|
|
|
TICK = "✓ "
|
|
|
|
CROSS = "✖ "
|
|
|
|
CIRCLE = "○ "
|
|
|
|
if os.name == 'posix':
|
|
|
|
# primitive formatting on supported
|
|
|
|
# terminal via ANSI escape sequences:
|
|
|
|
BOLD = ('\033[0m', '\033[1m')
|
|
|
|
BLUE = ('\033[0m', '\033[0;34m')
|
|
|
|
RED = ('\033[0m', '\033[0;31m')
|
|
|
|
GREY = ('\033[0m', '\033[1;30m')
|
|
|
|
|
|
|
|
TEST_EXIT_PASSED = 0
|
|
|
|
TEST_EXIT_SKIPPED = 77
|
|
|
|
|
|
|
|
BASE_SCRIPTS= [
|
|
|
|
# Scripts that are run by the travis build process.
|
|
|
|
# Longest test should go first, to favor running tests in parallel
|
|
|
|
'wallet-hd.py',
|
|
|
|
'walletbackup.py',
|
|
|
|
# vv Tests less than 5m vv
|
|
|
|
'p2p-fullblocktest.py',
|
|
|
|
'fundrawtransaction.py',
|
|
|
|
'p2p-compactblocks.py',
|
|
|
|
'segwit.py',
|
|
|
|
# vv Tests less than 2m vv
|
|
|
|
'wallet.py',
|
|
|
|
'wallet-accounts.py',
|
|
|
|
'p2p-segwit.py',
|
|
|
|
'wallet-dump.py',
|
|
|
|
'listtransactions.py',
|
|
|
|
# vv Tests less than 60s vv
|
|
|
|
'sendheaders.py',
|
|
|
|
'zapwallettxes.py',
|
|
|
|
'importmulti.py',
|
|
|
|
'mempool_limit.py',
|
|
|
|
'merkle_blocks.py',
|
|
|
|
'receivedby.py',
|
|
|
|
'abandonconflict.py',
|
|
|
|
'bip68-112-113-p2p.py',
|
|
|
|
'rawtransactions.py',
|
|
|
|
'reindex.py',
|
|
|
|
# vv Tests less than 30s vv
|
|
|
|
'mempool_resurrect_test.py',
|
|
|
|
'txn_doublespend.py --mineblock',
|
|
|
|
'txn_clone.py',
|
|
|
|
'getchaintips.py',
|
|
|
|
'rest.py',
|
|
|
|
'mempool_spendcoinbase.py',
|
|
|
|
'mempool_reorg.py',
|
|
|
|
'httpbasics.py',
|
|
|
|
'multi_rpc.py',
|
|
|
|
'proxy_test.py',
|
|
|
|
'signrawtransactions.py',
|
|
|
|
'nodehandling.py',
|
|
|
|
'decodescript.py',
|
|
|
|
'blockchain.py',
|
|
|
|
'disablewallet.py',
|
|
|
|
'net.py',
|
|
|
|
'keypool.py',
|
|
|
|
'p2p-mempool.py',
|
|
|
|
'prioritise_transaction.py',
|
|
|
|
'invalidblockrequest.py',
|
|
|
|
'invalidtxrequest.py',
|
|
|
|
'p2p-versionbits-warning.py',
|
|
|
|
'preciousblock.py',
|
|
|
|
'importprunedfunds.py',
|
|
|
|
'signmessages.py',
|
|
|
|
'nulldummy.py',
|
|
|
|
'import-rescan.py',
|
|
|
|
'bumpfee.py',
|
|
|
|
'rpcnamedargs.py',
|
|
|
|
'listsinceblock.py',
|
|
|
|
'p2p-leaktests.py',
|
|
|
|
]
|
|
|
|
|
|
|
|
ZMQ_SCRIPTS = [
|
|
|
|
# ZMQ test can only be run if bitcoin was built with zmq-enabled.
|
|
|
|
# call test_runner.py with -nozmq to explicitly exclude these tests.
|
|
|
|
'zmq_test.py']
|
|
|
|
|
|
|
|
EXTENDED_SCRIPTS = [
|
|
|
|
# These tests are not run by the travis build process.
|
|
|
|
# Longest test should go first, to favor running tests in parallel
|
|
|
|
'pruning.py',
|
|
|
|
# vv Tests less than 20m vv
|
|
|
|
'smartfees.py',
|
|
|
|
# vv Tests less than 5m vv
|
|
|
|
'maxuploadtarget.py',
|
|
|
|
'mempool_packages.py',
|
|
|
|
# vv Tests less than 2m vv
|
|
|
|
'bip68-sequence.py',
|
|
|
|
'getblocktemplate_longpoll.py',
|
|
|
|
'p2p-timeouts.py',
|
|
|
|
# vv Tests less than 60s vv
|
|
|
|
'bip9-softforks.py',
|
|
|
|
'p2p-feefilter.py',
|
|
|
|
'rpcbind_test.py',
|
|
|
|
# vv Tests less than 30s vv
|
|
|
|
'assumevalid.py',
|
|
|
|
'bip65-cltv.py',
|
|
|
|
'bip65-cltv-p2p.py',
|
|
|
|
'bipdersig-p2p.py',
|
|
|
|
'bipdersig.py',
|
|
|
|
'getblocktemplate_proposals.py',
|
|
|
|
'txn_doublespend.py',
|
|
|
|
'txn_clone.py --mineblock',
|
|
|
|
'forknotify.py',
|
|
|
|
'invalidateblock.py',
|
|
|
|
'maxblocksinflight.py',
|
|
|
|
'p2p-acceptblock.py',
|
|
|
|
'replace-by-fee.py',
|
|
|
|
]
|
|
|
|
|
|
|
|
ALL_SCRIPTS = BASE_SCRIPTS + ZMQ_SCRIPTS + EXTENDED_SCRIPTS
|
|
|
|
|
|
|
|
NON_SCRIPTS = [
|
|
|
|
# These are python files that live in the functional tests directory, but are not test scripts.
|
|
|
|
"combine_logs.py",
|
|
|
|
"create_cache.py",
|
|
|
|
"test_runner.py",
|
|
|
|
]
|
|
|
|
|
|
|
|
def main():
|
|
|
|
# Parse arguments and pass through unrecognised args
|
|
|
|
parser = argparse.ArgumentParser(add_help=False,
|
|
|
|
usage='%(prog)s [test_runner.py options] [script options] [scripts]',
|
|
|
|
description=__doc__,
|
|
|
|
epilog='''
|
|
|
|
Help text and arguments for individual test script:''',
|
|
|
|
formatter_class=argparse.RawTextHelpFormatter)
|
|
|
|
parser.add_argument('--coverage', action='store_true', help='generate a basic coverage report for the RPC interface')
|
|
|
|
parser.add_argument('--exclude', '-x', help='specify a comma-seperated-list of scripts to exclude. Do not include the .py extension in the name.')
|
|
|
|
parser.add_argument('--extended', action='store_true', help='run the extended test suite in addition to the basic tests')
|
|
|
|
parser.add_argument('--force', '-f', action='store_true', help='run tests even on platforms where they are disabled by default (e.g. windows).')
|
|
|
|
parser.add_argument('--help', '-h', '-?', action='store_true', help='print help text and exit')
|
|
|
|
parser.add_argument('--jobs', '-j', type=int, default=4, help='how many test scripts to run in parallel. Default=4.')
|
|
|
|
parser.add_argument('--quiet', '-q', action='store_true', help='only print results summary and failure logs')
|
|
|
|
parser.add_argument('--nozmq', action='store_true', help='do not run the zmq tests')
|
|
|
|
args, unknown_args = parser.parse_known_args()
|
|
|
|
|
|
|
|
# Create a set to store arguments and create the passon string
|
|
|
|
tests = set(arg for arg in unknown_args if arg[:2] != "--")
|
|
|
|
passon_args = [arg for arg in unknown_args if arg[:2] == "--"]
|
|
|
|
|
|
|
|
# Read config generated by configure.
|
|
|
|
config = configparser.ConfigParser()
|
|
|
|
config.read_file(open(os.path.dirname(__file__) + "/config.ini"))
|
|
|
|
|
|
|
|
# Set up logging
|
|
|
|
logging_level = logging.INFO if args.quiet else logging.DEBUG
|
|
|
|
logging.basicConfig(format='%(message)s', level=logging_level)
|
|
|
|
|
|
|
|
enable_wallet = config["components"].getboolean("ENABLE_WALLET")
|
|
|
|
enable_utils = config["components"].getboolean("ENABLE_UTILS")
|
|
|
|
enable_bitcoind = config["components"].getboolean("ENABLE_BITCOIND")
|
|
|
|
enable_zmq = config["components"].getboolean("ENABLE_ZMQ") and not args.nozmq
|
|
|
|
|
|
|
|
if config["environment"]["EXEEXT"] == ".exe" and not args.force:
|
|
|
|
# https://github.com/bitcoin/bitcoin/commit/d52802551752140cf41f0d9a225a43e84404d3e9
|
|
|
|
# https://github.com/bitcoin/bitcoin/pull/5677#issuecomment-136646964
|
|
|
|
print("Tests currently disabled on Windows by default. Use --force option to enable")
|
|
|
|
sys.exit(0)
|
|
|
|
|
|
|
|
if not (enable_wallet and enable_utils and enable_bitcoind):
|
|
|
|
print("No functional tests to run. Wallet, utils, and bitcoind must all be enabled")
|
|
|
|
print("Rerun `configure` with -enable-wallet, -with-utils and -with-daemon and rerun make")
|
|
|
|
sys.exit(0)
|
|
|
|
|
|
|
|
# python3-zmq may not be installed. Handle this gracefully and with some helpful info
|
|
|
|
if enable_zmq:
|
|
|
|
try:
|
|
|
|
import zmq
|
|
|
|
except ImportError:
|
|
|
|
print("ERROR: \"import zmq\" failed. Use -nozmq to run without the ZMQ tests."
|
|
|
|
"To run zmq tests, see dependency info in /test/README.md.")
|
|
|
|
raise
|
|
|
|
|
|
|
|
# Build list of tests
|
|
|
|
if tests:
|
|
|
|
# Individual tests have been specified. Run specified tests that exist
|
|
|
|
# in the ALL_SCRIPTS list. Accept the name with or without .py extension.
|
|
|
|
test_list = [t for t in ALL_SCRIPTS if
|
|
|
|
(t in tests or re.sub(".py$", "", t) in tests)]
|
|
|
|
else:
|
|
|
|
# No individual tests have been specified. Run base tests, and
|
|
|
|
# optionally ZMQ tests and extended tests.
|
|
|
|
test_list = BASE_SCRIPTS
|
|
|
|
if enable_zmq:
|
|
|
|
test_list += ZMQ_SCRIPTS
|
|
|
|
if args.extended:
|
|
|
|
test_list += EXTENDED_SCRIPTS
|
|
|
|
# TODO: BASE_SCRIPTS and EXTENDED_SCRIPTS are sorted by runtime
|
|
|
|
# (for parallel running efficiency). This combined list will is no
|
|
|
|
# longer sorted.
|
|
|
|
|
|
|
|
# Remove the test cases that the user has explicitly asked to exclude.
|
|
|
|
if args.exclude:
|
|
|
|
for exclude_test in args.exclude.split(','):
|
|
|
|
if exclude_test + ".py" in test_list:
|
|
|
|
test_list.remove(exclude_test + ".py")
|
|
|
|
|
|
|
|
if not test_list:
|
|
|
|
print("No valid test scripts specified. Check that your test is in one "
|
|
|
|
"of the test lists in test_runner.py, or run test_runner.py with no arguments to run all tests")
|
|
|
|
sys.exit(0)
|
|
|
|
|
|
|
|
if args.help:
|
|
|
|
# Print help for test_runner.py, then print help of the first script (with args removed) and exit.
|
|
|
|
parser.print_help()
|
|
|
|
subprocess.check_call([(config["environment"]["SRCDIR"] + '/test/functional/' + test_list[0].split()[0])] + ['-h'])
|
|
|
|
sys.exit(0)
|
|
|
|
|
|
|
|
check_script_list(config["environment"]["SRCDIR"])
|
|
|
|
|
|
|
|
run_tests(test_list, config["environment"]["SRCDIR"], config["environment"]["BUILDDIR"], config["environment"]["EXEEXT"], args.jobs, args.coverage, passon_args)
|
|
|
|
|
|
|
|
def run_tests(test_list, src_dir, build_dir, exeext, jobs=1, enable_coverage=False, args=[]):
|
|
|
|
|
|
|
|
#Set env vars
|
|
|
|
if "BITCOIND" not in os.environ:
|
|
|
|
os.environ["BITCOIND"] = build_dir + '/src/bitcoind' + exeext
|
|
|
|
|
|
|
|
tests_dir = src_dir + '/test/functional/'
|
|
|
|
|
|
|
|
flags = ["--srcdir={}/src".format(build_dir)] + args
|
|
|
|
flags.append("--cachedir=%s/test/cache" % build_dir)
|
|
|
|
|
|
|
|
if enable_coverage:
|
|
|
|
coverage = RPCCoverage()
|
|
|
|
flags.append(coverage.flag)
|
|
|
|
logging.debug("Initializing coverage directory at %s" % coverage.dir)
|
|
|
|
else:
|
|
|
|
coverage = None
|
|
|
|
|
|
|
|
if len(test_list) > 1 and jobs > 1:
|
|
|
|
# Populate cache
|
|
|
|
subprocess.check_output([tests_dir + 'create_cache.py'] + flags)
|
|
|
|
|
|
|
|
#Run Tests
|
|
|
|
job_queue = TestHandler(jobs, tests_dir, test_list, flags)
|
|
|
|
time0 = time.time()
|
|
|
|
test_results = []
|
|
|
|
|
|
|
|
max_len_name = len(max(test_list, key=len))
|
|
|
|
|
|
|
|
for _ in range(len(test_list)):
|
|
|
|
test_result, stdout, stderr = job_queue.get_next()
|
|
|
|
test_results.append(test_result)
|
|
|
|
|
|
|
|
if test_result.status == "Passed":
|
|
|
|
logging.debug("\n%s%s%s passed, Duration: %s s" % (BOLD[1], test_result.name, BOLD[0], test_result.time))
|
|
|
|
elif test_result.status == "Skipped":
|
|
|
|
logging.debug("\n%s%s%s skipped" % (BOLD[1], test_result.name, BOLD[0]))
|
|
|
|
else:
|
|
|
|
print("\n%s%s%s failed, Duration: %s s\n" % (BOLD[1], test_result.name, BOLD[0], test_result.time))
|
|
|
|
print(BOLD[1] + 'stdout:\n' + BOLD[0] + stdout + '\n')
|
|
|
|
print(BOLD[1] + 'stderr:\n' + BOLD[0] + stderr + '\n')
|
|
|
|
|
|
|
|
print_results(test_results, max_len_name, (int(time.time() - time0)))
|
|
|
|
|
|
|
|
if coverage:
|
|
|
|
coverage.report_rpc_coverage()
|
|
|
|
|
|
|
|
logging.debug("Cleaning up coverage data")
|
|
|
|
coverage.cleanup()
|
|
|
|
|
|
|
|
all_passed = all(map(lambda test_result: test_result.status == "Passed", test_results))
|
|
|
|
|
|
|
|
sys.exit(not all_passed)
|
|
|
|
|
|
|
|
def print_results(test_results, max_len_name, runtime):
|
|
|
|
results = "\n" + BOLD[1] + "%s | %s | %s\n\n" % ("TEST".ljust(max_len_name), "STATUS ", "DURATION") + BOLD[0]
|
|
|
|
|
|
|
|
test_results.sort(key=lambda result: result.name.lower())
|
|
|
|
all_passed = True
|
|
|
|
time_sum = 0
|
|
|
|
|
|
|
|
for test_result in test_results:
|
|
|
|
all_passed = all_passed and test_result.status != "Failed"
|
|
|
|
time_sum += test_result.time
|
|
|
|
test_result.padding = max_len_name
|
|
|
|
results += str(test_result)
|
|
|
|
|
|
|
|
status = TICK + "Passed" if all_passed else CROSS + "Failed"
|
|
|
|
results += BOLD[1] + "\n%s | %s | %s s (accumulated) \n" % ("ALL".ljust(max_len_name), status.ljust(9), time_sum) + BOLD[0]
|
|
|
|
results += "Runtime: %s s\n" % (runtime)
|
|
|
|
print(results)
|
|
|
|
|
|
|
|
class TestHandler:
|
|
|
|
"""
|
|
|
|
Trigger the testscrips passed in via the list.
|
|
|
|
"""
|
|
|
|
|
|
|
|
def __init__(self, num_tests_parallel, tests_dir, test_list=None, flags=None):
|
|
|
|
assert(num_tests_parallel >= 1)
|
|
|
|
self.num_jobs = num_tests_parallel
|
|
|
|
self.tests_dir = tests_dir
|
|
|
|
self.test_list = test_list
|
|
|
|
self.flags = flags
|
|
|
|
self.num_running = 0
|
|
|
|
# In case there is a graveyard of zombie bitcoinds, we can apply a
|
|
|
|
# pseudorandom offset to hopefully jump over them.
|
|
|
|
# (625 is PORT_RANGE/MAX_NODES)
|
|
|
|
self.portseed_offset = int(time.time() * 1000) % 625
|
|
|
|
self.jobs = []
|
|
|
|
|
|
|
|
def get_next(self):
|
|
|
|
while self.num_running < self.num_jobs and self.test_list:
|
|
|
|
# Add tests
|
|
|
|
self.num_running += 1
|
|
|
|
t = self.test_list.pop(0)
|
|
|
|
port_seed = ["--portseed={}".format(len(self.test_list) + self.portseed_offset)]
|
|
|
|
log_stdout = tempfile.SpooledTemporaryFile(max_size=2**16)
|
|
|
|
log_stderr = tempfile.SpooledTemporaryFile(max_size=2**16)
|
|
|
|
test_argv = t.split()
|
|
|
|
self.jobs.append((t,
|
|
|
|
time.time(),
|
|
|
|
subprocess.Popen([self.tests_dir + test_argv[0]] + test_argv[1:] + self.flags + port_seed,
|
|
|
|
universal_newlines=True,
|
|
|
|
stdout=log_stdout,
|
|
|
|
stderr=log_stderr),
|
|
|
|
log_stdout,
|
|
|
|
log_stderr))
|
|
|
|
if not self.jobs:
|
|
|
|
raise IndexError('pop from empty list')
|
|
|
|
while True:
|
|
|
|
# Return first proc that finishes
|
|
|
|
time.sleep(.5)
|
|
|
|
for j in self.jobs:
|
|
|
|
(name, time0, proc, log_out, log_err) = j
|
|
|
|
if proc.poll() is not None:
|
|
|
|
log_out.seek(0), log_err.seek(0)
|
|
|
|
[stdout, stderr] = [l.read().decode('utf-8') for l in (log_out, log_err)]
|
|
|
|
log_out.close(), log_err.close()
|
|
|
|
if proc.returncode == TEST_EXIT_PASSED and stderr == "":
|
|
|
|
status = "Passed"
|
|
|
|
elif proc.returncode == TEST_EXIT_SKIPPED:
|
|
|
|
status = "Skipped"
|
|
|
|
else:
|
|
|
|
status = "Failed"
|
|
|
|
self.num_running -= 1
|
|
|
|
self.jobs.remove(j)
|
|
|
|
|
|
|
|
return TestResult(name, status, int(time.time() - time0)), stdout, stderr
|
|
|
|
print('.', end='', flush=True)
|
|
|
|
|
|
|
|
class TestResult():
|
|
|
|
def __init__(self, name, status, time):
|
|
|
|
self.name = name
|
|
|
|
self.status = status
|
|
|
|
self.time = time
|
|
|
|
self.padding = 0
|
|
|
|
|
|
|
|
def __repr__(self):
|
|
|
|
if self.status == "Passed":
|
|
|
|
color = BLUE
|
|
|
|
glyph = TICK
|
|
|
|
elif self.status == "Failed":
|
|
|
|
color = RED
|
|
|
|
glyph = CROSS
|
|
|
|
elif self.status == "Skipped":
|
|
|
|
color = GREY
|
|
|
|
glyph = CIRCLE
|
|
|
|
|
|
|
|
return color[1] + "%s | %s%s | %s s\n" % (self.name.ljust(self.padding), glyph, self.status.ljust(7), self.time) + color[0]
|
|
|
|
|
|
|
|
|
|
|
|
def check_script_list(src_dir):
|
|
|
|
"""Check scripts directory.
|
|
|
|
|
|
|
|
Check that there are no scripts in the functional tests directory which are
|
|
|
|
not being run by pull-tester.py."""
|
|
|
|
script_dir = src_dir + '/test/functional/'
|
|
|
|
python_files = set([t for t in os.listdir(script_dir) if t[-3:] == ".py"])
|
|
|
|
missed_tests = list(python_files - set(map(lambda x: x.split()[0], ALL_SCRIPTS + NON_SCRIPTS)))
|
|
|
|
if len(missed_tests) != 0:
|
|
|
|
print("The following scripts are not being run:" + str(missed_tests))
|
|
|
|
print("Check the test lists in test_runner.py")
|
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
class RPCCoverage(object):
|
|
|
|
"""
|
|
|
|
Coverage reporting utilities for test_runner.
|
|
|
|
|
|
|
|
Coverage calculation works by having each test script subprocess write
|
|
|
|
coverage files into a particular directory. These files contain the RPC
|
|
|
|
commands invoked during testing, as well as a complete listing of RPC
|
|
|
|
commands per `bitcoin-cli help` (`rpc_interface.txt`).
|
|
|
|
|
|
|
|
After all tests complete, the commands run are combined and diff'd against
|
|
|
|
the complete list to calculate uncovered RPC commands.
|
|
|
|
|
|
|
|
See also: test/functional/test_framework/coverage.py
|
|
|
|
|
|
|
|
"""
|
|
|
|
def __init__(self):
|
|
|
|
self.dir = tempfile.mkdtemp(prefix="coverage")
|
|
|
|
self.flag = '--coveragedir=%s' % self.dir
|
|
|
|
|
|
|
|
def report_rpc_coverage(self):
|
|
|
|
"""
|
|
|
|
Print out RPC commands that were unexercised by tests.
|
|
|
|
|
|
|
|
"""
|
|
|
|
uncovered = self._get_uncovered_rpc_commands()
|
|
|
|
|
|
|
|
if uncovered:
|
|
|
|
print("Uncovered RPC commands:")
|
|
|
|
print("".join((" - %s\n" % i) for i in sorted(uncovered)))
|
|
|
|
else:
|
|
|
|
print("All RPC commands covered.")
|
|
|
|
|
|
|
|
def cleanup(self):
|
|
|
|
return shutil.rmtree(self.dir)
|
|
|
|
|
|
|
|
def _get_uncovered_rpc_commands(self):
|
|
|
|
"""
|
|
|
|
Return a set of currently untested RPC commands.
|
|
|
|
|
|
|
|
"""
|
|
|
|
# This is shared from `test/functional/test-framework/coverage.py`
|
|
|
|
reference_filename = 'rpc_interface.txt'
|
|
|
|
coverage_file_prefix = 'coverage.'
|
|
|
|
|
|
|
|
coverage_ref_filename = os.path.join(self.dir, reference_filename)
|
|
|
|
coverage_filenames = set()
|
|
|
|
all_cmds = set()
|
|
|
|
covered_cmds = set()
|
|
|
|
|
|
|
|
if not os.path.isfile(coverage_ref_filename):
|
|
|
|
raise RuntimeError("No coverage reference found")
|
|
|
|
|
|
|
|
with open(coverage_ref_filename, 'r') as f:
|
|
|
|
all_cmds.update([i.strip() for i in f.readlines()])
|
|
|
|
|
|
|
|
for root, dirs, files in os.walk(self.dir):
|
|
|
|
for filename in files:
|
|
|
|
if filename.startswith(coverage_file_prefix):
|
|
|
|
coverage_filenames.add(os.path.join(root, filename))
|
|
|
|
|
|
|
|
for filename in coverage_filenames:
|
|
|
|
with open(filename, 'r') as f:
|
|
|
|
covered_cmds.update([i.strip() for i in f.readlines()])
|
|
|
|
|
|
|
|
return all_cmds - covered_cmds
|
|
|
|
|
|
|
|
|
|
|
|
if __name__ == '__main__':
|
|
|
|
main()
|