analyze_outcomes: reset name of TestLog to Results
Signed-off-by: Valerio Setti <valerio.setti@nordicsemi.no>
diff --git a/tests/scripts/analyze_outcomes.py b/tests/scripts/analyze_outcomes.py
index 8ddbf6c..95f0cc6 100755
--- a/tests/scripts/analyze_outcomes.py
+++ b/tests/scripts/analyze_outcomes.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python3
+ #!/usr/bin/env python3
"""Analyze the test outcomes from a full CI run.
@@ -15,7 +15,7 @@
import check_test_cases
-class TestLog:
+class Results:
"""Process analysis results."""
def __init__(self):
@@ -56,27 +56,27 @@
"""
return len(self.successes) + len(self.failures)
-def execute_reference_driver_tests(log: TestLog, ref_component, driver_component, \
- outcome_file) -> TestLog:
+def execute_reference_driver_tests(results: Results, ref_component, driver_component, \
+ outcome_file) -> Results:
"""Run the tests specified in ref_component and driver_component. Results
are stored in the output_file and they will be used for the following
coverage analysis"""
# If the outcome file already exists, we assume that the user wants to
# perform the comparison analysis again without repeating the tests.
if os.path.exists(outcome_file):
- log.info("Outcome file (" + outcome_file + ") already exists. " + \
+ results.info("Outcome file (" + outcome_file + ") already exists. " + \
"Tests will be skipped.")
- return log
+ return results
shell_command = "tests/scripts/all.sh --outcome-file " + outcome_file + \
" " + ref_component + " " + driver_component
- log.info("Running: " + shell_command)
+ results.info("Running: " + shell_command)
ret_val = subprocess.run(shell_command.split(), check=False).returncode
if ret_val != 0:
- log.error("failed to run reference/driver components")
+ results.error("failed to run reference/driver components")
- return log
+ return results
def analyze_coverage(results, outcomes, allow_list, full_coverage):
"""Check that all available test cases are executed at least once."""
@@ -95,7 +95,7 @@
else:
results.warning('Allow listed test case was executed: {}', key)
-def analyze_driver_vs_reference(log: TestLog, outcomes,
+def analyze_driver_vs_reference(results: Results, outcomes,
component_ref, component_driver,
ignored_suites, ignored_test=None):
"""Check that all tests executed in the reference component are also
@@ -130,15 +130,15 @@
if component_ref in entry:
reference_test_passed = True
if(reference_test_passed and not driver_test_passed):
- log.error(key)
+ results.error(key)
- return log
+ return results
-def analyze_outcomes(log: TestLog, outcomes, args) -> TestLog:
+def analyze_outcomes(results: Results, outcomes, args) -> Results:
"""Run all analyses on the given outcome collection."""
- analyze_coverage(log, outcomes, args['allow_list'],
+ analyze_coverage(results, outcomes, args['allow_list'],
args['full_coverage'])
- return log
+ return results
def read_outcome_file(outcome_file):
"""Parse an outcome file and return an outcome collection.
@@ -161,30 +161,30 @@
outcomes[key].failures.append(setup)
return outcomes
-def do_analyze_coverage(log: TestLog, outcome_file, args) -> TestLog:
+def do_analyze_coverage(results: Results, outcome_file, args) -> Results:
"""Perform coverage analysis."""
- log.info("\n*** Analyze coverage ***\n")
+ results.info("\n*** Analyze coverage ***\n")
outcomes = read_outcome_file(outcome_file)
- log = analyze_outcomes(log, outcomes, args)
- return log
+ results = analyze_outcomes(results, outcomes, args)
+ return results
-def do_analyze_driver_vs_reference(log: TestLog, outcome_file, args) -> TestLog:
+def do_analyze_driver_vs_reference(results: Results, outcome_file, args) -> Results:
"""Perform driver vs reference analyze."""
- log.info("\n*** Analyze driver {} vs reference {} ***\n".format(
+ results.info("\n*** Analyze driver {} vs reference {} ***\n".format(
args['component_driver'], args['component_ref']))
- log = execute_reference_driver_tests(log, args['component_ref'], \
+ results = execute_reference_driver_tests(results, args['component_ref'], \
args['component_driver'], outcome_file)
ignored_suites = ['test_suite_' + x for x in args['ignored_suites']]
outcomes = read_outcome_file(outcome_file)
- log = analyze_driver_vs_reference(log, outcomes,
- args['component_ref'], args['component_driver'],
- ignored_suites, args['ignored_tests'])
+ results = analyze_driver_vs_reference(results, outcomes,
+ args['component_ref'], args['component_driver'],
+ ignored_suites, args['ignored_tests'])
- return log
+ return results
# List of tasks with a function that can handle this task and additional arguments if required
KNOWN_TASKS = {
@@ -649,7 +649,7 @@
}
def main():
- main_log = TestLog()
+ main_results = Results()
try:
parser = argparse.ArgumentParser(description=__doc__)
@@ -688,14 +688,13 @@
for task in tasks_list:
test_function = KNOWN_TASKS[task]['test_function']
test_args = KNOWN_TASKS[task]['args']
- main_log = test_function(main_log, options.outcomes, test_args)
-
- main_log.info("Overall results:\n" + \
- "{} warnings\n".format(main_log.warning_count) + \
- "{} errors\n".format(main_log.error_count))
+ main_results = test_function(main_results, options.outcomes, test_args)
- main_log.print_output()
- sys.exit(0 if (main_log.error_count == 0) else 2)
+ main_results.info("Overall results:\n" + \
+ "{} warnings\n".format(main_results.warning_count) + \
+ "{} errors\n".format(main_results.error_count))
+
+ sys.exit(0 if (main_results.error_count == 0) else 2)
except Exception: # pylint: disable=broad-except
# Print the backtrace and exit explicitly with our chosen status.