summaryrefslogtreecommitdiffstats
path: root/debian/uncrustify-trinity/uncrustify-trinity-0.78.0/tests/test_uncrustify/utilities.py
diff options
context:
space:
mode:
Diffstat (limited to 'debian/uncrustify-trinity/uncrustify-trinity-0.78.0/tests/test_uncrustify/utilities.py')
-rw-r--r--debian/uncrustify-trinity/uncrustify-trinity-0.78.0/tests/test_uncrustify/utilities.py210
1 files changed, 0 insertions, 210 deletions
diff --git a/debian/uncrustify-trinity/uncrustify-trinity-0.78.0/tests/test_uncrustify/utilities.py b/debian/uncrustify-trinity/uncrustify-trinity-0.78.0/tests/test_uncrustify/utilities.py
deleted file mode 100644
index 8570e6d3..00000000
--- a/debian/uncrustify-trinity/uncrustify-trinity-0.78.0/tests/test_uncrustify/utilities.py
+++ /dev/null
@@ -1,210 +0,0 @@
-# Logic for listing and running tests.
-#
-# * @author Ben Gardner October 2009
-# * @author Guy Maurel October 2015
-# * @author Matthew Woehlke June 2018
-#
-
-import argparse
-import os
-import subprocess
-import sys
-
-from .ansicolor import printc
-from .config import config, all_tests, FAIL_ATTRS, PASS_ATTRS, SKIP_ATTRS
-from .failure import (Failure, MismatchFailure, UnexpectedlyPassingFailure,
- UnstableFailure)
-from .test import FormatTest
-
-
-# -----------------------------------------------------------------------------
-def _add_common_arguments(parser):
- parser.add_argument('-c', '--show-commands', action='store_true',
- help='show commands')
-
- parser.add_argument('-v', '--verbose', action='store_true',
- help='show detailed test information')
-
- parser.add_argument('-d', '--diff', action='store_true',
- help='show diff on failure')
-
- parser.add_argument('-x', '--xdiff', action='store_true',
- help='show diff on expected failure')
-
- parser.add_argument('-g', '--debug', action='store_true',
- help='generate debug files (.log, .unc)')
-
- parser.add_argument('-e', '--executable', type=str, required=True,
- metavar='PATH',
- help='uncrustify executable to test')
-
- parser.add_argument('--git', type=str, default=config.git_exe,
- metavar='PATH',
- help='git executable to use to generate diffs')
-
- parser.add_argument('--result-dir', type=str, default=os.getcwd(),
- metavar='DIR',
- help='location to which results will be written')
-
-
-# -----------------------------------------------------------------------------
-def add_test_arguments(parser):
- _add_common_arguments(parser)
-
- parser.add_argument("name", type=str, metavar='NAME')
- parser.add_argument("--lang", type=str, required=True)
- parser.add_argument("--input", type=str, required=True)
- parser.add_argument("--config", type=str, required=True)
- parser.add_argument("--expected", type=str, required=True)
- parser.add_argument("--rerun-config", type=str, metavar='INPUT')
- parser.add_argument("--rerun-expected", type=str, metavar='CONFIG')
- parser.add_argument("--xfail", action='store_true')
-
-
-# -----------------------------------------------------------------------------
-def add_source_tests_arguments(parser):
- _add_common_arguments(parser)
-
- parser.add_argument('-p', '--show-all', action='store_true',
- help='show passed/skipped tests')
-
-
-# -----------------------------------------------------------------------------
-def add_format_tests_arguments(parser):
- _add_common_arguments(parser)
-
- parser.add_argument('-p', '--show-all', action='store_true',
- help='show passed/skipped tests')
-
- parser.add_argument('-r', '--select', metavar='CASE(S)', type=str,
- help='select tests to be executed')
-
- parser.add_argument('tests', metavar='TEST', type=str, nargs='*',
- default=all_tests,
- help='test(s) to run (default all)')
-
- # Arguments for generating the CTest script; users should not use these
- # directly
- parser.add_argument("--write-ctest", type=str, help=argparse.SUPPRESS)
- parser.add_argument("--cmake-config", type=str, help=argparse.SUPPRESS)
- parser.add_argument("--python", type=str, help=argparse.SUPPRESS)
-
-
-# -----------------------------------------------------------------------------
-def parse_args(parser):
- args = parser.parse_args()
-
- if args.git is not None:
- config.git_exe = args.git
-
- config.uncrustify_exe = args.executable
- if not os.path.exists(config.uncrustify_exe):
- msg = 'Specified uncrustify executable {!r} does not exist'.format(
- config.uncrustify_exe)
- printc("FAILED: ", msg, **FAIL_ATTRS)
- sys.exit(-1)
-
- # Do a sanity check on the executable
- try:
- with open(os.devnull, 'w') as bitbucket:
- subprocess.check_call([config.uncrustify_exe, '--help'],
- stdout=bitbucket)
- except Exception as exc:
- msg = ('Specified uncrustify executable {!r} ' +
- 'does not appear to be usable: {!s}').format(
- config.uncrustify_exe, exc)
- printc("FAILED: ", msg, **FAIL_ATTRS)
- sys.exit(-1)
-
- return args
-
-
-# -----------------------------------------------------------------------------
-def run_tests(tests, args, selector=None):
- pass_count = 0
- fail_count = 0
- mismatch_count = 0
- unstable_count = 0
- unexpectedly_passing_count = 0
-
- for test in tests:
- if selector is not None and not selector.test(test.test_name):
- if args.show_all:
- printc("SKIPPED: ", test.test_name, **SKIP_ATTRS)
- continue
-
- try:
- test.run(args)
- if args.show_all:
- outcome = 'XFAILED' if test.test_xfail else 'PASSED'
- printc('{}: '.format(outcome), test.test_name, **PASS_ATTRS)
- pass_count += 1
- except UnstableFailure:
- unstable_count += 1
- except MismatchFailure:
- mismatch_count += 1
- except UnexpectedlyPassingFailure:
- unexpectedly_passing_count += 1
- except Failure:
- fail_count += 1
-
- return {
- 'passing': pass_count,
- 'failing': fail_count,
- 'mismatch': mismatch_count,
- 'unstable': unstable_count,
- 'xpass': unexpectedly_passing_count
- }
-
-
-# -----------------------------------------------------------------------------
-def report(counts):
- total = sum(counts.values())
- print('{passing} / {total} tests passed'.format(total=total, **counts))
- if counts['failing'] > 0:
- printc('{failing} tests failed to execute'.format(**counts), 'f',
- **FAIL_ATTRS)
- if counts['mismatch'] > 0:
- printc(
- '{mismatch} tests did not match the expected output'.format(
- **counts), 'm',
- **FAIL_ATTRS)
- if counts['unstable'] > 0:
- printc('{unstable} tests were unstable'.format(**counts), 'u',
- **FAIL_ATTRS)
- if counts['xpass'] > 0:
- printc('{xpass} tests passed but were expected to fail'.format(**counts),
- 'x', **FAIL_ATTRS)
-
-
-# -----------------------------------------------------------------------------
-def read_format_tests(filename, group):
- tests = []
-
- print("Processing " + filename)
- with open(filename, 'rt') as f:
- for line_number, line in enumerate(f, 1):
- line = line.strip()
- if not len(line):
- continue
- if line.startswith('#'):
- continue
-
- test = FormatTest()
- test.build_from_declaration(line, group, line_number)
- tests.append(test)
-
- return tests
-
-
-# -----------------------------------------------------------------------------
-def fixup_ctest_path(path, config):
- if config is None:
- return path
-
- dirname, basename = os.path.split(path)
- if os.path.basename(dirname).lower() == config.lower():
- dirname, junk = os.path.split(dirname)
- return os.path.join(dirname, '${CTEST_CONFIGURATION_TYPE}', basename)
-
- return path