forked from organicmaps/organicmaps
run_desktop_tests.py removed -o option
Signed-off-by: Alexander Borsuk <me@alex.bio>
This commit is contained in:
parent
ffd255578e
commit
baf49c8311
1 changed files with 24 additions and 64 deletions
|
@ -1,18 +1,18 @@
|
|||
#!/usr/bin/env python
|
||||
#!/usr/bin/env python2
|
||||
|
||||
"""
|
||||
This script is mainly for running autotests on the build server, however, it
|
||||
This script is mainly for running autotests on the build server, however, it
|
||||
can also be used by engineers to run the tests locally on their machines.
|
||||
|
||||
It takes as optional parameters the path to the folder containing the test
|
||||
executables (which must have names ending in _tests), and a list of tests that
|
||||
It takes as optional parameters the path to the folder containing the test
|
||||
executables (which must have names ending in _tests), and a list of tests that
|
||||
need to be skipped, this list must be comma separated and contain no spaces. E.g.:
|
||||
|
||||
./run_desktop_tests.py -f ./omim-build-release -e drape_tests,some_other_tests
|
||||
|
||||
The script outputs the console output of the tests. It also checks the error
|
||||
code of each test suite, and after all the tests are executed, it prints the
|
||||
list of the failed tests, passed tests, skipped tests and tests that could not
|
||||
The script outputs the console output of the tests. It also checks the error
|
||||
code of each test suite, and after all the tests are executed, it prints the
|
||||
list of the failed tests, passed tests, skipped tests and tests that could not
|
||||
be found, i.e. the tests that were specified in the skip list, but do not exist.
|
||||
"""
|
||||
|
||||
|
@ -46,7 +46,7 @@ class TestRunner:
|
|||
def print_pretty(self, result, tests):
|
||||
if not tests:
|
||||
return
|
||||
|
||||
|
||||
logging.info("\n{result}".format(result=result.upper()))
|
||||
|
||||
for test in tests:
|
||||
|
@ -56,7 +56,6 @@ class TestRunner:
|
|||
def set_global_vars(self):
|
||||
|
||||
parser = OptionParser()
|
||||
parser.add_option("-o", "--output", dest="output", default="testlog.log", help="resulting log file. Default testlog.log")
|
||||
parser.add_option("-f", "--folder", dest="folder", default="omim-build-release/out/release", help="specify the folder where the tests reside (absolute path or relative to the location of this script)")
|
||||
parser.add_option("-d", "--data_path", dest="data_path", help="Path to data files (passed to the test executables as --data_path=<value>)")
|
||||
parser.add_option("-u", "--user_resource_path", dest="resource_path", help="Path to resources, styles and classificators (passed to the test executables as --user_resource_path=<value>)")
|
||||
|
@ -65,7 +64,6 @@ class TestRunner:
|
|||
parser.add_option("-b", "--boost_tests", dest="boost_tests", action="store_true", default=False, help="Treat all the tests as boost tests (their output is different and it must be processed differently).")
|
||||
parser.add_option("-k", "--keep_alive", dest="keep_alive", action="store_true", default=False, help="Keep the server alive after the end of the test. Because the server sometimes fails to start, this reduces the probability of false test failures on CI servers.")
|
||||
|
||||
|
||||
(options, args) = parser.parse_args()
|
||||
|
||||
self.skiplist = set()
|
||||
|
@ -74,7 +72,7 @@ class TestRunner:
|
|||
for tests in options.skiplist:
|
||||
for test in tests.split(","):
|
||||
self.skiplist.add(test)
|
||||
|
||||
|
||||
for tests in options.runlist:
|
||||
self.runlist.extend(tests.split(","))
|
||||
|
||||
|
@ -82,14 +80,13 @@ class TestRunner:
|
|||
|
||||
if self.runlist:
|
||||
logging.warn("-i or -b option found, the -e option will be ignored")
|
||||
|
||||
|
||||
self.workspace_path = options.folder
|
||||
self.logfile = options.output
|
||||
self.data_path = (" --data_path={0}".format(options.data_path) if options.data_path else "")
|
||||
self.user_resource_path = (" --user_resource_path={0}".format(options.resource_path) if options.resource_path else "")
|
||||
|
||||
self.keep_alive = options.keep_alive
|
||||
|
||||
|
||||
|
||||
def start_server(self):
|
||||
server = testserver.TestServer()
|
||||
|
@ -106,11 +103,11 @@ class TestRunner:
|
|||
logging.info("Failed to stop the server...")
|
||||
|
||||
def categorize_tests(self):
|
||||
|
||||
|
||||
tests_to_run = list()
|
||||
local_skiplist = list()
|
||||
not_found = list()
|
||||
|
||||
|
||||
test_files_in_dir = filter(lambda x: x.endswith("_tests"), listdir(self.workspace_path))
|
||||
|
||||
on_disk = lambda x: x in test_files_in_dir
|
||||
|
@ -123,7 +120,7 @@ class TestRunner:
|
|||
else:
|
||||
tests_to_run = filter(on_disk, self.runlist)
|
||||
shuffle(tests_to_run)
|
||||
|
||||
|
||||
not_found = filter(not_on_disk, self.runlist)
|
||||
|
||||
# now let's move the tests that need a server either to the beginning or the end of the tests_to_run list
|
||||
|
@ -135,7 +132,7 @@ class TestRunner:
|
|||
tests_with_server.remove(test)
|
||||
|
||||
return {TO_RUN:tests_to_run, SKIP:local_skiplist, NOT_FOUND:not_found, WITH_SERVER:tests_with_server}
|
||||
|
||||
|
||||
|
||||
def test_file_with_keys(self, test_file):
|
||||
boost_keys = " --report_format=xml --report_level=detailed --log_level=test_suite --log_format=xml " if self.boost_tests else ""
|
||||
|
@ -147,11 +144,9 @@ class TestRunner:
|
|||
passed = list()
|
||||
|
||||
for test_file in tests_to_run:
|
||||
|
||||
self.log_exec_file(test_file)
|
||||
|
||||
test_file_with_keys = self.test_file_with_keys(test_file)
|
||||
|
||||
|
||||
logging.info(test_file_with_keys)
|
||||
|
||||
# Fix float parsing
|
||||
|
@ -160,8 +155,8 @@ class TestRunner:
|
|||
current_env = environ
|
||||
current_env['LC_NUMERIC'] = 'C'
|
||||
#
|
||||
process = subprocess.Popen("{tests_path}/{test_file} 2>> {logfile}".
|
||||
format(tests_path=self.workspace_path, test_file=test_file_with_keys, logfile=self.logfile),
|
||||
process = subprocess.Popen("{tests_path}/{test_file}".
|
||||
format(tests_path=self.workspace_path, test_file=test_file_with_keys),
|
||||
env=current_env,
|
||||
shell=True,
|
||||
stdout=subprocess.PIPE)
|
||||
|
@ -173,49 +168,12 @@ class TestRunner:
|
|||
failed.append(test_file)
|
||||
else:
|
||||
passed.append(test_file)
|
||||
|
||||
self.log_exec_file(test_file, result=process.returncode)
|
||||
|
||||
return {FAILED: failed, PASSED: passed}
|
||||
|
||||
|
||||
def log_exec_file(self, filename, result=None):
|
||||
if self.boost_tests:
|
||||
return
|
||||
|
||||
logstring = "BEGIN" if result is None else "END" #can be 0 or None. If we omit the explicit check for None, we get wrong result
|
||||
resstring = (" | result: {returncode}".format(returncode=result) if result is not None else "")
|
||||
with open(self.logfile, "a") as logf:
|
||||
logf.write("\n{logstring}: {filename}{resstring}\n".format(logstring=logstring, filename=filename, resstring=resstring))
|
||||
|
||||
|
||||
def rm_log_file(self):
|
||||
try:
|
||||
remove(self.logfile)
|
||||
except OSError:
|
||||
pass
|
||||
|
||||
|
||||
def __init__(self):
|
||||
self.set_global_vars()
|
||||
self.rm_log_file()
|
||||
|
||||
|
||||
def merge_dicts_of_lists(self, one, two):
|
||||
if not one:
|
||||
return two
|
||||
if not two:
|
||||
return one
|
||||
|
||||
ret = one.copy()
|
||||
|
||||
for key, value in two.iteritems():
|
||||
if key in one:
|
||||
ret[key] = ret[key].append(two[key])
|
||||
else:
|
||||
ret[key] = two[key]
|
||||
|
||||
return ret
|
||||
|
||||
|
||||
def execute(self):
|
||||
|
@ -224,12 +182,14 @@ class TestRunner:
|
|||
to_run_and_with_server_keys = [TO_RUN, WITH_SERVER]
|
||||
random.shuffle(to_run_and_with_server_keys)
|
||||
|
||||
results = dict()
|
||||
results = {FAILED: [], PASSED: []}
|
||||
|
||||
for key in to_run_and_with_server_keys:
|
||||
if key == WITH_SERVER and categorized_tests[WITH_SERVER]:
|
||||
self.start_server()
|
||||
results = self.merge_dicts_of_lists(results, self.run_tests(categorized_tests[key]))
|
||||
result = self.run_tests(categorized_tests[key])
|
||||
results[FAILED].extend(result[FAILED])
|
||||
results[PASSED].extend(result[PASSED])
|
||||
if key == WITH_SERVER and categorized_tests[WITH_SERVER]:
|
||||
self.stop_server()
|
||||
|
||||
|
@ -237,7 +197,7 @@ class TestRunner:
|
|||
self.print_pretty("skipped", categorized_tests[SKIP])
|
||||
self.print_pretty("passed", results[PASSED])
|
||||
self.print_pretty("not found", categorized_tests[NOT_FOUND])
|
||||
|
||||
return len(results[FAILED]) > 0 and 1 or 0
|
||||
|
||||
def tests_on_disk(path):
|
||||
return filter(lambda x: x.endswith("_tests"), listdir(path))
|
||||
|
@ -245,4 +205,4 @@ def tests_on_disk(path):
|
|||
|
||||
if __name__ == "__main__":
|
||||
runner = TestRunner()
|
||||
runner.execute()
|
||||
quit(runner.execute())
|
||||
|
|
Loading…
Add table
Reference in a new issue