blob: ed005fec16d909f89f72eb95f16592bf4e439e0a [file] [log] [blame]
# Copyright 2023 Google LLC
# Copyright lowRISC contributors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
r"""
Class describing simulation configuration object
"""
import collections
from datetime import datetime, timezone
import fnmatch
import json
import logging as log
import os
import re
import shutil
import subprocess
import sys
from collections import OrderedDict
from pathlib import Path
from typing import Optional
from Deploy import CompileSim, CovAnalyze, CovMerge, CovReport, CovUnr, RunTest
from FlowCfg import FlowCfg
from Modes import BuildModes, Modes, Regressions, RunModes, Tests
from SimResults import SimResults
from tabulate import tabulate
from dvsim_matcha.Testplan import Testplan
from utils import TS_FORMAT, VERBOSE, rm_path
# This affects the bucketizer failure report.
_MAX_UNIQUE_TESTS = 5
_MAX_TEST_RESEEDS = 2
class SimCfg(FlowCfg):
"""Simulation configuration object
A simulation configuration class holds key information required for building
a DV regression framework.
"""
flow = 'sim'
# TODO: Find a way to set these in sim cfg instead
ignored_wildcards = [
"build_mode", "index", "test", "seed", "uvm_test", "uvm_test_seq",
"cov_db_dirs", "sw_images", "sw_build_device", "sw_build_cmd",
"sw_build_opts"
]
def __init__(self, flow_cfg_file, hjson_data, args, mk_config):
# Options set from command line
self.tool = args.tool
self.build_opts = []
self.build_opts.extend(args.build_opts)
self.en_build_modes = args.build_modes.copy()
self.run_opts = []
self.run_opts.extend(args.run_opts)
self.en_run_modes = []
self.en_run_modes.extend(args.run_modes)
self.build_unique = args.build_unique
self.build_seed = args.build_seed
self.build_only = args.build_only
self.run_only = args.run_only
self.reseed_ovrd = args.reseed
self.reseed_multiplier = args.reseed_multiplier
# Waves must be of type string, since it may be used as substitution
# variable in the HJson cfg files.
self.waves = args.waves or 'none'
self.max_waves = args.max_waves
self.cov = args.cov
self.cov_merge_previous = args.cov_merge_previous
self.profile = args.profile or '(cfg uses profile without --profile)'
self.xprop_off = args.xprop_off
self.no_rerun = args.no_rerun
self.verbosity = None # set in _expand
self.verbose = args.verbose
self.dry_run = args.dry_run
self.map_full_testplan = args.map_full_testplan
# Set default sim modes for unpacking
if args.gui:
self.en_build_modes.append("gui")
if args.waves is not None:
self.en_build_modes.append("waves")
else:
self.en_build_modes.append("waves_off")
if self.cov is True:
self.en_build_modes.append("cov")
if args.profile is not None:
self.en_build_modes.append("profile")
if self.xprop_off is not True:
self.en_build_modes.append("xprop")
if self.build_seed:
self.en_build_modes.append("build_seed")
# Options built from cfg_file files
self.project = ""
self.flow = ""
self.flow_makefile = ""
self.pre_build_cmds = []
self.post_build_cmds = []
self.build_dir = ""
self.pre_run_cmds = []
self.post_run_cmds = []
self.run_dir = ""
self.sw_images = []
self.sw_build_opts = []
self.pass_patterns = []
self.fail_patterns = []
self.name = ""
self.variant = ""
self.dut = ""
self.tb = ""
self.testplan = ""
self.fusesoc_core = ""
self.ral_spec = ""
self.build_modes = []
self.run_modes = []
self.regressions = []
self.supported_wave_formats = None
# Options from tools - for building and running tests
self.build_cmd = ""
self.flist_gen_cmd = ""
self.flist_gen_opts = []
self.flist_file = ""
self.run_cmd = ""
# Generated data structures
self.variant_name = ""
self.links = {}
self.build_list = []
self.run_list = []
self.cov_merge_deploy = None
self.cov_report_deploy = None
self.results_summary = OrderedDict()
super().__init__(flow_cfg_file, hjson_data, args, mk_config)
# Handle titan_root
if args.titan_root:
self.css_file = os.path.join(args.titan_root, "util/dvsim",
"style.css")
def _expand(self):
# Choose a wave format now. Note that this has to happen after parsing
# the configuration format because our choice might depend on the
# chosen tool.
self.waves = self._resolve_waves()
# If build_unique is set, then add current timestamp to uniquify it
if self.build_unique:
self.build_dir += "_" + self.timestamp
# If the user specified a verbosity on the command line then
# self.args.verbosity will be n, l, m, h or d. Set self.verbosity now.
# We will actually have loaded some other verbosity level from the
# config file, but that won't have any effect until expansion so we can
# safely switch it out now.
if self.args.verbosity is not None:
self.verbosity = self.args.verbosity
super()._expand()
if self.variant:
self.variant_name = self.name + "/" + self.variant
else:
self.variant_name = self.name
# Set the title for simulation results.
self.results_title = self.variant_name.upper() + " Simulation Results"
# Stuff below only pertains to individual cfg (not primary cfg)
# or individual selected cfgs (if select_cfgs is configured via command line)
# TODO: find a better way to support select_cfgs
if not self.is_primary_cfg and (not self.select_cfgs or
self.name in self.select_cfgs):
# If self.tool is None at this point, there was no --tool argument on
# the command line, and there is no default tool set in the config
# file. That's ok if this is a primary config (where the
# sub-configurations can choose tools themselves), but not otherwise.
if self.tool is None:
log.error(
'Config file does not specify a default tool, '
'and there was no --tool argument on the command line.')
sys.exit(1)
# Print scratch_path at the start:
log.info("[scratch_path]: [%s] [%s]", self.name, self.scratch_path)
# Set directories with links for ease of debug / triage.
self.links = {
"D": self.scratch_path + "/" + "dispatched",
"P": self.scratch_path + "/" + "passed",
"F": self.scratch_path + "/" + "failed",
"K": self.scratch_path + "/" + "killed"
}
# Use the default build mode for tests that do not specify it
if not hasattr(self, "build_mode"):
self.build_mode = 'default'
# Set the primary build mode. The coverage associated to this build
# is the main coverage. Some tools need this information. This is
# of significance only when there are multiple builds. If there is
# only one build, and its not the primary_build_mode, then we
# update the primary_build_mode to match what is built.
if not hasattr(self, "primary_build_mode"):
self.primary_build_mode = self.build_mode
# Create objects from raw dicts - build_modes, sim_modes, run_modes,
# tests and regressions, only if not a primary cfg obj
self._create_objects()
def _resolve_waves(self):
'''Choose and return a wave format, if waves are enabled.
This is called after reading the config file. This method is used to
update the value of class member 'waves', which must be of type string,
since it is used as a substitution variable in the parsed HJson dict.
If waves are not enabled, or if this is a primary cfg, then return
'none'. 'tool', which must be set at this point, supports a limited
list of wave formats (supplied with 'supported_wave_formats' key).
'''
if self.waves == 'none' or self.is_primary_cfg:
return 'none'
assert self.tool is not None
# If the user has specified their preferred wave format, use it. As
# a sanity check, error out if the chosen tool doesn't support the
# format, but only if we know about the tool. If not, we'll just assume
# they know what they're doing.
if self.supported_wave_formats and \
self.waves not in self.supported_wave_formats:
log.error('Chosen tool ({}) does not support wave format '
'{!r}.'.format(self.tool, self.waves))
sys.exit(1)
return self.waves
# Purge the output directories. This operates on self.
def _purge(self):
assert self.scratch_path
log.info("Purging scratch path %s", self.scratch_path)
rm_path(self.scratch_path)
def _create_objects(self):
# Create build and run modes objects
self.build_modes = Modes.create_modes(BuildModes, self.build_modes)
self.run_modes = Modes.create_modes(RunModes, self.run_modes)
# Walk through build modes enabled on the CLI and append the opts
for en_build_mode in self.en_build_modes:
build_mode_obj = Modes.find_mode(en_build_mode, self.build_modes)
if build_mode_obj is not None:
self.pre_build_cmds.extend(build_mode_obj.pre_build_cmds)
self.post_build_cmds.extend(build_mode_obj.post_build_cmds)
self.build_opts.extend(build_mode_obj.build_opts)
self.pre_run_cmds.extend(build_mode_obj.pre_run_cmds)
self.post_run_cmds.extend(build_mode_obj.post_run_cmds)
self.run_opts.extend(build_mode_obj.run_opts)
self.sw_images.extend(build_mode_obj.sw_images)
self.sw_build_opts.extend(build_mode_obj.sw_build_opts)
else:
log.error(
"Mode \"%s\" enabled on the the command line is not defined",
en_build_mode)
sys.exit(1)
# Walk through run modes enabled on the CLI and append the opts
for en_run_mode in self.en_run_modes:
run_mode_obj = Modes.find_mode(en_run_mode, self.run_modes)
if run_mode_obj is not None:
self.pre_run_cmds.extend(run_mode_obj.pre_run_cmds)
self.post_run_cmds.extend(run_mode_obj.post_run_cmds)
self.run_opts.extend(run_mode_obj.run_opts)
self.sw_images.extend(run_mode_obj.sw_images)
self.sw_build_opts.extend(run_mode_obj.sw_build_opts)
else:
log.error(
"Mode \"%s\" enabled on the the command line is not defined",
en_run_mode)
sys.exit(1)
# Create tests from given list of items
self.tests = Tests.create_tests(self.tests, self)
# Regressions
# Parse testplan if provided.
if self.testplan != "":
self.testplan = Testplan(self.testplan,
repo_top=Path(self.proj_root),
titan_root=Path(self.titan_root))
# Extract tests in each stage and add them as regression target.
self.regressions.extend(self.testplan.get_stage_regressions())
else:
# Create a dummy testplan with no entries.
self.testplan = Testplan(None, name=self.name)
# Create regressions
self.regressions = Regressions.create_regressions(
self.regressions, self, self.tests)
def _print_list(self):
for list_item in self.list_items:
log.info("---- List of %s in %s ----", list_item, self.variant_name)
if hasattr(self, list_item):
items = getattr(self, list_item)
for item in items:
log.info(item)
else:
log.error("Item %s does not exist!", list_item)
def _create_build_and_run_list(self):
'''Generates a list of deployable objects from the provided items.
Tests to be run are provided with --items switch. These can be glob-
style patterns. This method finds regressions and tests that match
these patterns.
'''
def _match_items(items: list, patterns: list):
hits = []
matched = set()
for pattern in patterns:
item_hits = fnmatch.filter(items, pattern)
if item_hits:
hits += item_hits
matched.add(pattern)
return hits, matched
# Process regressions first.
regr_map = {regr.name: regr for regr in self.regressions}
regr_hits, items_matched = _match_items(regr_map.keys(), self.items)
regrs = [regr_map[regr] for regr in regr_hits]
for regr in regrs:
overlap = bool([t for t in regr.tests if t in self.run_list])
if overlap:
log.warning(
f"Regression {regr.name} added to be run has tests that "
"overlap with other regressions also being run. This can "
"result in conflicting build / run time opts to be set, "
"resulting in unexpected results. Skipping.")
continue
self.run_list += regr.tests
# Merge regression's build and run opts with its tests and their
# build_modes.
regr.merge_regression_opts()
# Process individual tests, skipping the ones already added from
# regressions.
test_map = {
test.name: test
for test in self.tests if test not in self.run_list
}
test_hits, items_matched_ = _match_items(test_map.keys(), self.items)
self.run_list += [test_map[test] for test in test_hits]
items_matched |= items_matched_
# Check if all items have been processed.
for item in set(self.items) - items_matched:
log.warning(f"Item {item} did not match any regressions or "
f"tests in {self.flow_cfg_file}.")
# Merge the global build and run opts
Tests.merge_global_opts(self.run_list, self.pre_build_cmds,
self.post_build_cmds, self.build_opts,
self.pre_run_cmds, self.post_run_cmds,
self.run_opts, self.sw_images,
self.sw_build_opts)
# Process reseed override and create the build_list
build_list_names = []
for test in self.run_list:
# Override reseed if available.
if self.reseed_ovrd is not None:
test.reseed = self.reseed_ovrd
# Apply reseed multiplier if set on the command line. This is
# always positive but might not be an integer. Round to nearest,
# but make sure there's always at least one iteration.
scaled = round(test.reseed * self.reseed_multiplier)
test.reseed = max(1, scaled)
# Create the unique set of builds needed.
if test.build_mode.name not in build_list_names:
self.build_list.append(test.build_mode)
build_list_names.append(test.build_mode.name)
def _create_dirs(self):
'''Create initial set of directories
'''
for link in self.links.keys():
rm_path(self.links[link])
os.makedirs(self.links[link])
def _expand_run_list(self, build_map):
'''Generate a list of tests to be run
For each test in tests, we add it test.reseed times. The ordering is
interleaved so that we run through all of the tests as soon as
possible. If there are multiple tests and they have different reseed
values, they are "fully interleaved" at the start (so if there are
tests A, B with reseed values of 5 and 2, respectively, then the list
will be ABABAAA).
build_map is a dictionary mapping a build mode to a CompileSim object.
'''
tagged = []
for test in self.run_list:
build_job = build_map[test.build_mode]
for idx in range(test.reseed):
tagged.append((idx, RunTest(idx, test, build_job, self)))
# Stably sort the tagged list by the 1st coordinate.
tagged.sort(key=lambda x: x[0])
# Return the sorted list of RunTest objects, discarding the indices by
# which we sorted it.
return [run for _, run in tagged]
def _create_deploy_objects(self):
'''Create deploy objects from the build and run lists.
'''
# Create the build and run list first
self._create_build_and_run_list()
self.builds = []
build_map = {}
for build_mode_obj in self.build_list:
new_build = CompileSim(build_mode_obj, self)
# It is possible for tests to supply different build modes, but
# those builds may differ only under specific circumstances,
# such as coverage being enabled. If coverage is not enabled,
# then they may be completely identical. In that case, we can
# save compute resources by removing the extra duplicated
# builds. We discard the new_build if it is equivalent to an
# existing one.
is_unique = True
for build in self.builds:
if build.is_equivalent_job(new_build):
# Discard `new_build` since it is equivalent to build. If
# `new_build` is the same as `primary_build_mode`, update
# `primary_build_mode` to match `build`.
if new_build.name == self.primary_build_mode:
self.primary_build_mode = build.name
new_build = build
is_unique = False
break
if is_unique:
self.builds.append(new_build)
build_map[build_mode_obj] = new_build
# If there is only one build, set primary_build_mode to it.
if len(self.builds) == 1:
self.primary_build_mode = self.builds[0].name
# Check self.primary_build_mode is set correctly.
build_mode_names = set(b.name for b in self.builds)
if self.primary_build_mode not in build_mode_names:
log.error(f"\"primary_build_mode: {self.primary_build_mode}\" "
f"in {self.name} cfg is invalid. Please pick from "
f"{build_mode_names}.")
sys.exit(1)
# Update all tests to use the updated (uniquified) build modes.
for test in self.run_list:
if test.build_mode.name != build_map[test.build_mode].name:
test.build_mode = Modes.find_mode(
build_map[test.build_mode].name, self.build_modes)
self.runs = ([]
if self.build_only else self._expand_run_list(build_map))
# In GUI mode, only allow one test to run.
if self.gui and len(self.runs) > 1:
self.runs = self.runs[:1]
log.warning("In GUI mode, only one test is allowed to run. "
"Picking {}".format(self.runs[0].full_name))
# Add builds to the list of things to run, only if --run-only switch
# is not passed.
self.deploy = []
if not self.run_only:
self.deploy += self.builds
if not self.build_only:
self.deploy += self.runs
# Create cov_merge and cov_report objects, so long as we've got at
# least one run to do.
if self.cov and self.runs:
self.cov_merge_deploy = CovMerge(self.runs, self)
self.cov_report_deploy = CovReport(self.cov_merge_deploy, self)
self.deploy += [self.cov_merge_deploy, self.cov_report_deploy]
# Create initial set of directories before kicking off the regression.
self._create_dirs()
def _cov_analyze(self):
'''Use the last regression coverage data to open up the GUI tool to
analyze the coverage.
'''
# Create initial set of directories, such as dispatched, passed etc.
self._create_dirs()
cov_analyze_deploy = CovAnalyze(self)
self.deploy = [cov_analyze_deploy]
def cov_analyze(self):
'''Public facing API for analyzing coverage.
'''
for item in self.cfgs:
item._cov_analyze()
def _cov_unr(self):
'''Use the last regression coverage data to generate unreachable
coverage exclusions.
'''
# TODO, Only support VCS
if self.tool not in ['vcs', 'xcelium']:
log.error("Only VCS and Xcelium are supported for the UNR flow.")
sys.exit(1)
# Create initial set of directories, such as dispatched, passed etc.
self._create_dirs()
cov_unr_deploy = CovUnr(self)
self.deploy = [cov_unr_deploy]
def cov_unr(self):
'''Public facing API for analyzing coverage.
'''
for item in self.cfgs:
item._cov_unr()
def _gen_json_results(self, run_results):
"""Returns the run results as json-formatted dictionary.
"""
def _empty_str_as_none(s: str) -> Optional[str]:
"""Map an empty string to None and retain the value of a non-empty
string.
This is intended to clearly distinguish an empty string, which may
or may not be an valid value, from an invalid value.
"""
return s if s != "" else None
def _pct_str_to_float(s: str) -> Optional[float]:
"""Map a percentage value stored in a string with ` %` suffix to a
float or to None if the conversion to Float fails.
"""
try:
return float(s[:-2])
except ValueError:
return None
def _test_result_to_dict(tr) -> dict:
"""Map a test result entry to a dict."""
job_time_s = (tr.job_runtime.with_unit('s').get()[0]
if tr.job_runtime is not None
else None)
sim_time_us = (tr.simulated_time.with_unit('us').get()[0]
if tr.simulated_time is not None
else None)
pass_rate = tr.passing * 100.0 / tr.total if tr.total > 0 else 0
return {
'name': tr.name,
'max_runtime_s': job_time_s,
'simulated_time_us': sim_time_us,
'passing_runs': tr.passing,
'total_runs': tr.total,
'pass_rate': pass_rate,
}
results = dict()
# Describe name of hardware block targeted by this run and optionally
# the variant of the hardware block.
results['block_name'] = self.name.lower()
results['block_variant'] = _empty_str_as_none(self.variant.lower())
# The timestamp for this run has been taken with `utcnow()` and is
# stored in a custom format. Store it in standard ISO format with
# explicit timezone annotation.
timestamp = datetime.strptime(self.timestamp, TS_FORMAT)
timestamp = timestamp.replace(tzinfo=timezone.utc)
results['report_timestamp'] = timestamp.isoformat()
# Extract Git properties.
m = re.search(r'https://github.com/.+?/tree/([0-9a-fA-F]+)',
self.revision)
results['git_revision'] = m.group(1) if m else None
results['git_branch_name'] = _empty_str_as_none(self.branch)
# Describe type of report and tool used.
results['report_type'] = 'simulation'
results['tool'] = self.tool.lower()
# Create dictionary to store results.
results['results'] = {
'testpoints': [],
'unmapped_tests': [],
'testplan_stage_summary': [],
'coverage': dict(),
'failure_buckets': [],
}
# If the testplan does not yet have test results mapped to testpoints,
# map them now.
sim_results = SimResults(self.deploy, run_results)
if not self.testplan.test_results_mapped:
self.testplan.map_test_results(test_results=sim_results.table)
# Extract results of testpoints and tests into the `testpoints` field.
for tp in self.testplan.testpoints:
# Ignore testpoints that contain unmapped tests, because those will
# be handled separately.
if tp.name in ["Unmapped tests", "N.A."]:
continue
# Extract test results for this testpoint.
tests = []
for tr in tp.test_results:
# Ignore test results with zero total runs unless we are told
# to "map the full testplan".
if tr.total == 0 and not self.map_full_testplan:
continue
# Map test result metrics and append it to the collecting list.
tests.append(_test_result_to_dict(tr))
# Ignore testpoints for which no tests have been run unless we are
# told to "map the full testplan".
if len(tests) == 0 and not self.map_full_testplan:
continue
# Append testpoint to results.
results['results']['testpoints'].append({
'name': tp.name,
'stage': tp.stage,
'tests': tests,
})
# Extract unmapped tests.
unmapped_trs = [tr for tr in sim_results.table if not tr.mapped]
for tr in unmapped_trs:
results['results']['unmapped_tests'].append(
_test_result_to_dict(tr))
# Extract summary of testplan stages.
if self.map_full_testplan:
for k, d in self.testplan.progress.items():
results['results']['testplan_stage_summary'].append({
'name': k,
'total_tests': d['total'],
'written_tests': d['written'],
'passing_tests': d['passing'],
'pass_rate': _pct_str_to_float(d['progress']),
})
# Extract coverage results if coverage has been collected in this run.
if self.cov_report_deploy is not None:
cov = self.cov_report_deploy.cov_results_dict
for k, v in cov.items():
results['results']['coverage'][k.lower()] = _pct_str_to_float(v)
# Extract failure buckets.
if sim_results.buckets:
by_tests = sorted(sim_results.buckets.items(),
key=lambda i: len(i[1]),
reverse=True)
for bucket, tests in by_tests:
unique_tests = collections.defaultdict(list)
for (test, line, context) in tests:
if not isinstance(test, RunTest):
continue
unique_tests[test.name].append((test, line, context))
fts = []
for test_name, test_runs in unique_tests.items():
frs = []
for test, line, context in test_runs:
frs.append({
'seed': test.seed,
'failure_message': {
'log_file_path': test.get_log_path(),
'log_file_line_num': line,
'text': ''.join(context),
},
})
fts.append({
'name': test_name,
'failing_runs': frs,
})
results['results']['failure_buckets'].append({
'identifier': bucket,
'failing_tests': fts,
})
# Store the `results` dictionary in this object.
self.results_dict = results
# Return the `results` dictionary as json string.
return json.dumps(self.results_dict)
def _gen_results(self, run_results):
'''
The function is called after the regression has completed. It collates the
status of all run targets and generates a dict. It parses the testplan and
maps the generated result to the testplan entries to generate a final table
(list). It also prints the full list of failures for debug / triage. If cov
is enabled, then the summary coverage report is also generated. The final
result is in markdown format.
'''
def indent_by(level):
return " " * (4 * level)
def create_failure_message(test, line, context):
message = [f"{indent_by(2)}* {test.qual_name}\\"]
if line:
message.append(
f"{indent_by(2)} Line {line}, in log " +
test.get_log_path())
else:
message.append(f"{indent_by(2)} Log {test.get_log_path()}")
if context:
message.append("")
lines = [f"{indent_by(4)}{c.rstrip()}" for c in context]
message.extend(lines)
message.append("")
return message
def create_bucket_report(buckets):
"""Creates a report based on the given buckets.
The buckets are sorted by descending number of failures. Within
buckets this also group tests by unqualified name, and just a few
failures are shown per unqualified name.
Args:
buckets: A dictionary by bucket containing triples
(test, line, context).
Returns:
A list of text lines for the report.
"""
by_tests = sorted(buckets.items(),
key=lambda i: len(i[1]),
reverse=True)
fail_msgs = ["\n## Failure Buckets", ""]
for bucket, tests in by_tests:
fail_msgs.append(f"* `{bucket}` has {len(tests)} failures:")
unique_tests = collections.defaultdict(list)
for (test, line, context) in tests:
unique_tests[test.name].append((test, line, context))
for name, test_reseeds in list(unique_tests.items())[
:_MAX_UNIQUE_TESTS]:
fail_msgs.append(f"{indent_by(1)}* Test {name} has "
f"{len(test_reseeds)} failures.")
for test, line, context in test_reseeds[:_MAX_TEST_RESEEDS]:
fail_msgs.extend(
create_failure_message(test, line, context))
if len(test_reseeds) > _MAX_TEST_RESEEDS:
fail_msgs.append(
f"{indent_by(2)}* ... and "
f"{len(test_reseeds) - _MAX_TEST_RESEEDS} "
"more failures.")
if len(unique_tests) > _MAX_UNIQUE_TESTS:
fail_msgs.append(
f"{indent_by(1)}* ... and "
f"{len(unique_tests) - _MAX_UNIQUE_TESTS} more tests.")
fail_msgs.append("")
return fail_msgs
deployed_items = self.deploy
results = SimResults(deployed_items, run_results)
# Generate results table for runs.
results_str = "## " + self.results_title + "\n"
results_str += "### " + self.timestamp_long + "\n"
if self.revision:
results_str += "### " + self.revision + "\n"
results_str += "### Branch: " + self.branch + "\n"
# Add path to testplan, only if it has entries (i.e., its not dummy).
if self.testplan.testpoints:
if hasattr(self, "testplan_doc_path"):
testplan = "https://{}/{}".format(self.doc_server,
self.testplan_doc_path)
else:
testplan = "https://{}/{}".format(self.doc_server,
self.rel_path)
testplan = testplan.replace("/dv", "/doc/dv/#testplan")
results_str += f"### [Testplan]({testplan})\n"
results_str += f"### Simulator: {self.tool.upper()}\n"
# Print the build seed used for clarity.
if self.build_seed and not self.run_only:
results_str += ("### Build randomization enabled with "
f"--build-seed {self.build_seed}\n")
if not results.table:
results_str += "No results to display.\n"
else:
# Map regr results to the testplan entries.
if not self.testplan.test_results_mapped:
self.testplan.map_test_results(test_results=results.table)
results_str += self.testplan.get_test_results_table(
map_full_testplan=self.map_full_testplan)
if self.map_full_testplan:
results_str += self.testplan.get_progress_table()
self.results_summary = self.testplan.get_test_results_summary()
# Append coverage results if coverage was enabled.
if self.cov_report_deploy is not None:
report_status = run_results[self.cov_report_deploy]
if report_status == "P":
results_str += "\n## Coverage Results\n"
# Link the dashboard page using "cov_report_page" value.
if hasattr(self, "cov_report_page"):
results_str += "\n### [Coverage Dashboard]"
if self.args.publish:
cov_report_page_path = "cov_report"
else:
cov_report_page_path = self.cov_report_dir
cov_report_page_path += "/" + self.cov_report_page
results_str += "({})\n\n".format(cov_report_page_path)
results_str += self.cov_report_deploy.cov_results
self.results_summary[
"Coverage"] = self.cov_report_deploy.cov_total
else:
self.results_summary["Coverage"] = "--"
if results.buckets:
self.errors_seen = True
results_str += "\n".join(create_bucket_report(results.buckets))
self.results_md = results_str
return results_str
def gen_results_summary(self):
'''Generate the summary results table.
This method is specific to the primary cfg. It summarizes the results
from each individual cfg in a markdown table.
Prints the generated summary markdown text to stdout and returns it.
'''
lines = [f"## {self.results_title} (Summary)"]
lines += [f"### {self.timestamp_long}"]
if self.revision:
lines += [f"### {self.revision}"]
lines += [f"### Branch: {self.branch}"]
table = []
header = []
for cfg in self.cfgs:
row = cfg.results_summary
if row:
# convert name entry to relative link
row = cfg.results_summary
row["Name"] = cfg._get_results_page_link(
self.results_dir,
row["Name"])
# If header is set, ensure its the same for all cfgs.
if header:
assert header == cfg.results_summary.keys()
else:
header = cfg.results_summary.keys()
table.append(row.values())
if table:
assert header
colalign = ("center", ) * len(header)
table_txt = tabulate(table,
headers=header,
tablefmt="pipe",
colalign=colalign)
lines += ["", table_txt, ""]
else:
lines += ["\nNo results to display.\n"]
self.results_summary_md = "\n".join(lines)
print(self.results_summary_md)
return self.results_summary_md
def _publish_results(self):
'''Publish coverage results to the opentitan web server.'''
super()._publish_results()
if self.cov_report_deploy is not None:
results_server_dir_url = self.results_server_dir.replace(
self.results_server_prefix, "https://")
log.info("Publishing coverage results to %s",
results_server_dir_url)
cmd = (self.results_server_cmd + " -m cp -R " +
self.cov_report_dir + " " + self.results_server_dir)
try:
cmd_output = subprocess.run(args=cmd,
shell=True,
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT)
log.log(VERBOSE, cmd_output.stdout.decode("utf-8"))
except Exception as e:
log.error("%s: Failed to publish results:\n\"%s\"", e,
str(cmd))