1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266
|
################################################################################
import test.unittest as unittest
import sys, os, re, StringIO, time, optparse
from inspect import getdoc, getmembers, isclass
from pprint import pformat
import unittest_patch
from unittest_patch import StringIOContents
################################################################################
def prepare_test_env():
main_dir = os.path.split(os.path.abspath(__file__))[0]
test_subdir = os.path.join(main_dir, 'test')
sys.path.insert(0, test_subdir)
fake_test_subdir = os.path.join(test_subdir, 'run_tests__tests')
return main_dir, test_subdir, fake_test_subdir
main_dir, test_subdir, fake_test_subdir = prepare_test_env()
import test_utils
################################################################################
# Set the command line options
#
# options are shared with run_tests.py so make sure not to conflict
# in time more will be added here
opt_parser = optparse.OptionParser()
opt_parser.add_option (
"-i", "--incomplete", action = 'store_true',
help = "fail incomplete tests" )
opt_parser.add_option (
"-s", "--subprocess", action = 'store_true',
help = "run test suites in subprocesses (default: same process)" )
opt_parser.add_option (
"-d", "--dump", action = 'store_true',
help = "dump failures/errors as dict ready to eval" )
opt_parser.add_option (
"-T", "--timings", type = 'int', default = 1, metavar = 'T',
help = "get timings for individual tests.\n"
"Run test T times, giving average time")
opt_parser.add_option (
"-e", "--exclude", default = '',
help = "exclude tests containing any of TAGS" )
opt_parser.add_option (
"-w", "--show_output", action = 'store_true',
help = "show silenced stderr/stdout on errors" )
opt_parser.add_option (
"-a", "--all", action = 'store_true',
help = "dump all results not just errors eg. -da" )
opt_parser.add_option (
"-H", "--human", action = 'store_true',
help = "dump results as dict ready to eval if unsure "
"that pieced together results are correct "
"(subprocess mode)" )
opt_parser.add_option (
"-m", "--multi_thread", metavar = 'THREADS', type = 'int',
help = "run subprocessed tests in x THREADS" )
opt_parser.add_option (
"-t", "--time_out", metavar = 'SECONDS', type = 'int',
help = "kill stalled subprocessed tests after SECONDS" )
opt_parser.add_option (
"-f", "--fake", metavar = "DIR",
help = "run fake tests in run_tests__tests/$DIR" )
opt_parser.add_option (
"-p", "--python", metavar = "PYTHON",
help = "path to python excutable to run subproccesed tests\n"
"default (sys.executable): %s" % sys.executable)
################################################################################
# Human readable output
#
COMPLETE_FAILURE_TEMPLATE = """
======================================================================
ERROR: all_tests_for (%(module)s.AllTestCases)
----------------------------------------------------------------------
Traceback (most recent call last):
File "test\%(module)s.py", line 1, in all_tests_for
subprocess completely failed with return code of %(return_code)s
cmd: %(cmd)s
test_env: %(test_env)s
working_dir: %(working_dir)s
return (top 5 lines):
%(raw_return)s
""" # Leave that last empty line else build page regex won't match
# Text also needs to be vertically compressed
RAN_TESTS_DIV = (70 * "-") + "\nRan"
DOTS = re.compile("^([FE.]*)$", re.MULTILINE)
def combine_results(all_results, t):
"""
Return pieced together results in a form fit for human consumption. Don't
rely on results if piecing together subprocessed results (single process
mode is fine). Was originally meant for that purpose but was found to be
unreliable. See options.dump or options.human for reliable results.
"""
all_dots = ''
failures = []
for module, results in sorted(all_results.items()):
output, return_code, raw_return = map (
results.get, ('output','return_code', 'raw_return')
)
if not output or (return_code and RAN_TESTS_DIV not in output):
# would this effect the original dict? TODO
results['raw_return'] = ''.join(raw_return.splitlines(1)[:5])
failures.append( COMPLETE_FAILURE_TEMPLATE % results )
all_dots += 'E'
continue
dots = DOTS.search(output).group(1)
all_dots += dots
if 'E' in dots or 'F' in dots:
failures.append( output[len(dots)+1:].split(RAN_TESTS_DIV)[0] )
total_fails, total_errors = map(all_dots.count, 'FE')
total_tests = len(all_dots)
combined = [all_dots]
if failures: combined += [''.join(failures).lstrip('\n')[:-1]]
combined += ["%s %s tests in %.3fs\n" % (RAN_TESTS_DIV, total_tests, t)]
if not failures: combined += ['OK\n']
else: combined += [
'FAILED (%s)\n' % ', '.join (
(total_fails and ["failures=%s" % total_fails] or []) +
(total_errors and ["errors=%s" % total_errors] or [])
)]
return total_tests, '\n'.join(combined)
################################################################################
TEST_RESULTS_START = "<--!! TEST RESULTS START HERE !!-->"
TEST_RESULTS_RE = re.compile('%s\n(.*)' % TEST_RESULTS_START, re.DOTALL | re.M)
def get_test_results(raw_return):
test_results = TEST_RESULTS_RE.search(raw_return)
if test_results:
try: return eval(test_results.group(1))
except: raise Exception (
"BUGGY TEST RESULTS EVAL:\n %s" % test_results.group(1)
)
################################################################################
# ERRORS
# TODO
def make_complete_failure_error(result):
return (
"ERROR: all_tests_for (%s.AllTestCases)" % result['module'],
"Complete Failure (ret code: %s)" % result['return_code'],
result['test_file'],
'1',
)
# For combined results, plural
def test_failures(results):
errors = {}
total = sum(v.get('num_tests', 0) for v in results.values())
for module, result in results.items():
num_errors = (
len(result.get('failures', [])) + len(result.get('errors', []))
)
if num_errors is 0 and result.get('return_code'):
result.update(RESULTS_TEMPLATE)
result['errors'].append(make_complete_failure_error(result))
num_errors += 1
total += 1
if num_errors: errors.update({module:result})
return total, errors
def combined_errs(results):
for result in results.values():
combined_errs = result['errors'] + result['failures']
for err in combined_errs:
yield err
################################################################################
# For complete failures (+ namespace saving)
def from_namespace(ns, template):
if isinstance(template, dict):
return dict((i, ns.get(i, template[i])) for i in template)
return dict((i, ns[i]) for i in template)
RESULTS_TEMPLATE = {
'output' : '',
'num_tests' : 0,
'failures' : [],
'errors' : [],
'tests' : {},
}
################################################################################
def run_test(module, options):
suite = unittest.TestSuite()
test_utils.fail_incomplete_tests = options.incomplete
m = __import__(module)
if m.unittest is not unittest:
raise ImportError(
"%s is not using correct unittest\n\n" % module +
"should be: %s\n is using: %s" % (unittest.__file__,
m.unittest.__file__)
)
print 'loading', module
test = unittest.defaultTestLoader.loadTestsFromName(module)
suite.addTest(test)
output = StringIO.StringIO()
runner = unittest.TextTestRunner(stream = output)
results = runner.run(suite)
output = StringIOContents(output)
num_tests = results.testsRun
failures = results.failures
errors = results.errors
tests = results.tests
results = {module:from_namespace(locals(), RESULTS_TEMPLATE)}
if options.subprocess:
print TEST_RESULTS_START
print pformat(results)
else:
return results
################################################################################
if __name__ == '__main__':
options, args = opt_parser.parse_args()
unittest_patch.patch(options)
if not args: sys.exit('Called from run_tests.py, use that')
run_test(args[0], options)
################################################################################
|