|  | #!/usr/bin/env python3 | 
|  | # Copyright lowRISC contributors. | 
|  | # Licensed under the Apache License, Version 2.0, see LICENSE for details. | 
|  | # SPDX-License-Identifier: Apache-2.0 | 
|  | r"""Top Module Generator | 
|  | """ | 
|  | import argparse | 
|  | import logging as log | 
|  | import os | 
|  | import random | 
|  | import shutil | 
|  | import sys | 
|  | import tempfile | 
|  | from collections import OrderedDict | 
|  | from copy import deepcopy | 
|  | from io import StringIO | 
|  | from pathlib import Path | 
|  | from typing import Dict, List, Optional, Tuple | 
|  | from itertools import chain | 
|  |  | 
|  | import hjson | 
|  | import tlgen | 
|  | from ipgen import (IpBlockRenderer, IpConfig, IpDescriptionOnlyRenderer, | 
|  | IpTemplate, TemplateRenderError) | 
|  | from mako import exceptions | 
|  | from mako.template import Template | 
|  | from reggen import access, gen_rtl, gen_sec_cm_testplan, window | 
|  | from reggen.inter_signal import InterSignal | 
|  | from reggen.ip_block import IpBlock | 
|  | from reggen.countermeasure import CounterMeasure | 
|  | from reggen.lib import check_list | 
|  | from topgen import entropy_buffer_generator as ebg | 
|  | from topgen import get_hjsonobj_xbars | 
|  | from topgen import intermodule as im | 
|  | from topgen import lib as lib | 
|  | from topgen import merge_top, search_ips, strong_random, validate_top | 
|  | from topgen.c_test import TopGenCTest | 
|  | from topgen.rust import TopGenRust | 
|  | from topgen.clocks import Clocks | 
|  | from topgen.gen_dv import gen_dv | 
|  | from topgen.gen_top_docs import gen_top_docs | 
|  | from topgen.merge import connect_clocks, create_alert_lpgs, extract_clocks | 
|  | from topgen.resets import Resets | 
|  | from topgen.top import Top | 
|  |  | 
|  | # Common header for generated files | 
|  | warnhdr = """// | 
|  | // ------------------- W A R N I N G: A U T O - G E N E R A T E D   C O D E !! -------------------// | 
|  | // PLEASE DO NOT HAND-EDIT THIS FILE. IT HAS BEEN AUTO-GENERATED WITH THE FOLLOWING COMMAND: | 
|  | """ | 
|  | genhdr = """// Copyright lowRISC contributors. | 
|  | // Licensed under the Apache License, Version 2.0, see LICENSE for details. | 
|  | // SPDX-License-Identifier: Apache-2.0 | 
|  | """ + warnhdr | 
|  |  | 
|  | GENCMD = ("// util/topgen.py -t hw/top_{topname}/data/top_{topname}.hjson\n" | 
|  | "// -o hw/top_{topname}") | 
|  |  | 
|  | SRCTREE_TOP = Path(__file__).parent.parent.resolve() | 
|  |  | 
|  | TOPGEN_TEMPLATE_PATH = Path(__file__).parent / "topgen/templates" | 
|  |  | 
|  | # Size and path to the entropy buffer. | 
|  | # This buffer is generated using Mersenne Twister PRNG seeded with rnd_cnst_seed | 
|  | # and deleted in the end." | 
|  | # This buffer will not be created If a different one is provided by args.entropy_buffer. | 
|  | # Module strong_random fetches entropy from the buffer to generate random bit-vectors | 
|  | # and permutations. | 
|  | BUFFER_SIZE = 20000 | 
|  |  | 
|  | PATH_TO_BUFFER = "util/topgen/entropy_buffer.txt" | 
|  |  | 
|  |  | 
|  | def ipgen_render(template_name: str, topname: str, params: Dict, | 
|  | out_path: Path): | 
|  | """ Render an IP template for a specific toplevel using ipgen. | 
|  |  | 
|  | The generated IP block is placed in the "ip_autogen" directory of the | 
|  | toplevel. | 
|  |  | 
|  | Aborts the program execution in case of an error. | 
|  | """ | 
|  | module_name = params.get("module_instance_name", template_name) | 
|  | instance_name = f"top_{topname}_{module_name}" | 
|  | ip_template = IpTemplate.from_template_path( | 
|  | SRCTREE_TOP / "hw/ip_templates" / template_name) | 
|  |  | 
|  | try: | 
|  | ip_config = IpConfig(ip_template.params, instance_name, params) | 
|  | except ValueError as e: | 
|  | log.error(f"Unable to render IP template {template_name!r}: {str(e)}") | 
|  | sys.exit(1) | 
|  |  | 
|  | try: | 
|  | renderer = IpBlockRenderer(ip_template, ip_config) | 
|  | renderer.render(out_path / "ip_autogen" / module_name, | 
|  | overwrite_output_dir=True) | 
|  | except TemplateRenderError as e: | 
|  | log.error(e.verbose_str()) | 
|  | sys.exit(1) | 
|  |  | 
|  |  | 
|  | def generate_top(top, name_to_block, tpl_filename, **kwargs): | 
|  | top_tpl = Template(filename=tpl_filename) | 
|  |  | 
|  | try: | 
|  | return top_tpl.render(top=top, name_to_block=name_to_block, **kwargs) | 
|  | except:  # noqa: E722 | 
|  | log.error(exceptions.text_error_template().render()) | 
|  | return "" | 
|  |  | 
|  |  | 
|  | def generate_xbars(top, out_path): | 
|  | topname = top["name"] | 
|  | gencmd = ("// util/topgen.py -t hw/top_{topname}/data/top_{topname}.hjson " | 
|  | "-o hw/top_{topname}/\n\n".format(topname=topname)) | 
|  |  | 
|  | for obj in top["xbar"]: | 
|  | xbar_path = out_path / "ip/xbar_{}/data/autogen".format(obj["name"]) | 
|  | xbar_path.mkdir(parents=True, exist_ok=True) | 
|  | xbar = tlgen.validate(obj) | 
|  | xbar.ip_path = "hw/top_" + top["name"] + "/ip/{dut}" | 
|  |  | 
|  | # Generate output of crossbar with complete fields | 
|  | xbar_hjson_path = xbar_path / "xbar_{}.gen.hjson".format(xbar.name) | 
|  | xbar_hjson_path.write_text(genhdr + gencmd + | 
|  | hjson.dumps(obj, for_json=True) + '\n') | 
|  |  | 
|  | if not tlgen.elaborate(xbar): | 
|  | log.error("Elaboration failed." + repr(xbar)) | 
|  |  | 
|  | try: | 
|  | results = tlgen.generate(xbar, "top_" + top["name"]) | 
|  | except:  # noqa: E722 | 
|  | log.error(exceptions.text_error_template().render()) | 
|  |  | 
|  | ip_path = out_path / "ip/xbar_{}".format(obj["name"]) | 
|  |  | 
|  | for filename, filecontent in results: | 
|  | filepath = ip_path / filename | 
|  | filepath.parent.mkdir(parents=True, exist_ok=True) | 
|  | with filepath.open(mode="w", encoding="UTF-8") as fout: | 
|  | fout.write(filecontent) | 
|  |  | 
|  | dv_path = out_path / "ip/xbar_{}/dv/autogen".format(obj["name"]) | 
|  | dv_path.mkdir(parents=True, exist_ok=True) | 
|  |  | 
|  | # generate testbench for xbar | 
|  | tlgen.generate_tb(xbar, dv_path, "top_" + top["name"]) | 
|  |  | 
|  | # Read back the comportable IP and amend to Xbar | 
|  | xbar_ipfile = ip_path / ("data/autogen/xbar_%s.hjson" % obj["name"]) | 
|  | with xbar_ipfile.open() as fxbar: | 
|  | xbar_ipobj = hjson.load(fxbar, | 
|  | use_decimal=True, | 
|  | object_pairs_hook=OrderedDict) | 
|  |  | 
|  | r_inter_signal_list = check_list( | 
|  | xbar_ipobj.get("inter_signal_list", []), | 
|  | "inter_signal_list field") | 
|  | obj["inter_signal_list"] = [ | 
|  | InterSignal.from_raw( | 
|  | "entry {} of the inter_signal_list field".format(idx + 1), | 
|  | entry) for idx, entry in enumerate(r_inter_signal_list) | 
|  | ] | 
|  |  | 
|  |  | 
|  | def generate_alert_handler(top, out_path): | 
|  | topname = top["name"] | 
|  |  | 
|  | # default values | 
|  | esc_cnt_dw = 32 | 
|  | accu_cnt_dw = 16 | 
|  | async_on = [] | 
|  | # leave this constant | 
|  | n_classes = 4 | 
|  | # low power groups | 
|  | n_lpg = 1 | 
|  | lpg_map = [] | 
|  |  | 
|  | # Count number of alerts and LPGs | 
|  | n_alerts = sum([x["width"] if "width" in x else 1 for x in top["alert"]]) | 
|  | n_lpg = len(top["alert_lpgs"]) | 
|  | n_lpg_width = n_lpg.bit_length() | 
|  | # format used to print out indices in binary format | 
|  | async_on_format = "1'b{:01b}" | 
|  | lpg_idx_format = str(n_lpg_width) + "'d{:d}" | 
|  |  | 
|  | # Double check that all these values are greated than 0 | 
|  | if esc_cnt_dw < 1: | 
|  | raise ValueError("esc_cnt_dw must be larger than 0") | 
|  | if accu_cnt_dw < 1: | 
|  | raise ValueError("accu_cnt_dw must be larger than 0") | 
|  | if n_lpg < 1: | 
|  | raise ValueError("n_lpg must be larger than 0") | 
|  |  | 
|  | if n_alerts < 1: | 
|  | # set number of alerts to 1 such that the config is still valid | 
|  | # that input will be tied off | 
|  | n_alerts = 1 | 
|  | log.warning("no alerts are defined in the system") | 
|  | else: | 
|  | async_on = [] | 
|  | lpg_map = [] | 
|  | for alert in top["alert"]: | 
|  | for k in range(alert["width"]): | 
|  | async_on.append(async_on_format.format(int(alert["async"]))) | 
|  | lpg_map.append(lpg_idx_format.format(int(alert["lpg_idx"]))) | 
|  |  | 
|  | params = { | 
|  | "n_alerts": n_alerts, | 
|  | "esc_cnt_dw": esc_cnt_dw, | 
|  | "accu_cnt_dw": accu_cnt_dw, | 
|  | "async_on": async_on, | 
|  | "n_classes": n_classes, | 
|  | "n_lpg": n_lpg, | 
|  | "lpg_map": lpg_map, | 
|  | } | 
|  |  | 
|  | ipgen_render("alert_handler", topname, params, out_path) | 
|  |  | 
|  |  | 
|  | def generate_plic(top, out_path): | 
|  | topname = top["name"] | 
|  | params = {} | 
|  |  | 
|  | # Count number of interrupts | 
|  | # Interrupt source 0 is tied to 0 to conform RISC-V PLIC spec. | 
|  | # So, total number of interrupts are the number of entries in the list + 1 | 
|  | params["src"] = sum( | 
|  | [x["width"] if "width" in x else 1 for x in top["interrupt"]]) + 1 | 
|  |  | 
|  | # Target and priority: Currently fixed | 
|  | params["target"] = int(top["num_cores"], 0) if "num_cores" in top else 1 | 
|  | params["prio"] = 3 | 
|  |  | 
|  | ipgen_render("rv_plic", topname, params, out_path) | 
|  |  | 
|  |  | 
|  | # TODO: For generated IPs that are generated legacy style (i.e., without IPgen) | 
|  | # we have to search both the source and destination RTL directories, since not | 
|  | # all files are copied over. This is a workaround which can be removed once | 
|  | # all generated IPs have transitioned to IPgen. | 
|  | def generate_regfile_from_path(hjson_path: Path, | 
|  | generated_rtl_path: Path, | 
|  | original_rtl_path: Path = None): | 
|  | """Generate RTL register file from path and check countermeasure labels""" | 
|  | obj = IpBlock.from_path(str(hjson_path), []) | 
|  |  | 
|  | # If this block has countermeasures, we grep for RTL annotations in | 
|  | # all .sv implementation files and check whether they match up | 
|  | # with what is defined inside the Hjson. | 
|  | sv_files = generated_rtl_path.glob("*.sv") | 
|  | if original_rtl_path is not None: | 
|  | sv_files = chain(sv_files, original_rtl_path.glob("*.sv")) | 
|  | rtl_names = CounterMeasure.search_rtl_files(sv_files) | 
|  | obj.check_cm_annotations(rtl_names, str(hjson_path)) | 
|  | gen_rtl.gen_rtl(obj, str(generated_rtl_path)) | 
|  | if gen_sec_cm_testplan.gen_sec_cm_testplan(obj, hjson_path.parent): | 
|  | sys.exit(1) | 
|  |  | 
|  |  | 
|  | def generate_pinmux(top, out_path): | 
|  |  | 
|  | topname = top["name"] | 
|  | pinmux = top["pinmux"] | 
|  |  | 
|  | # Generation without pinmux and pinout configuration is not supported. | 
|  | assert "pinmux" in top | 
|  | assert "pinout" in top | 
|  |  | 
|  | # Get number of wakeup detectors | 
|  | if "num_wkup_detect" in pinmux: | 
|  | num_wkup_detect = pinmux["num_wkup_detect"] | 
|  | else: | 
|  | num_wkup_detect = 1 | 
|  |  | 
|  | if num_wkup_detect <= 0: | 
|  | # TODO: add support for no wakeup counter case | 
|  | log.error("Topgen does currently not support generation of a top " + | 
|  | "without DIOs.") | 
|  | return | 
|  |  | 
|  | if "wkup_cnt_width" in pinmux: | 
|  | wkup_cnt_width = pinmux["wkup_cnt_width"] | 
|  | else: | 
|  | wkup_cnt_width = 8 | 
|  |  | 
|  | if wkup_cnt_width <= 1: | 
|  | log.error("Wakeup counter width must be greater equal 2.") | 
|  | return | 
|  |  | 
|  | # MIO Pads | 
|  | n_mio_pads = pinmux["io_counts"]["muxed"]["pads"] | 
|  |  | 
|  | # Total inputs/outputs | 
|  | # Reuse the counts from the merge phase | 
|  | n_mio_periph_in = (pinmux["io_counts"]["muxed"]["inouts"] + | 
|  | pinmux["io_counts"]["muxed"]["inputs"]) | 
|  | n_mio_periph_out = (pinmux["io_counts"]["muxed"]["inouts"] + | 
|  | pinmux["io_counts"]["muxed"]["outputs"]) | 
|  | n_dio_periph_in = (pinmux["io_counts"]["dedicated"]["inouts"] + | 
|  | pinmux["io_counts"]["dedicated"]["inputs"]) | 
|  | n_dio_periph_out = (pinmux["io_counts"]["dedicated"]["inouts"] + | 
|  | pinmux["io_counts"]["dedicated"]["outputs"]) | 
|  | n_dio_pads = (pinmux["io_counts"]["dedicated"]["inouts"] + | 
|  | pinmux["io_counts"]["dedicated"]["inputs"] + | 
|  | pinmux["io_counts"]["dedicated"]["outputs"]) | 
|  |  | 
|  | # TODO: derive this value | 
|  | attr_dw = 13 | 
|  |  | 
|  | # Generation with zero MIO/DIO pads is currently not supported. | 
|  | assert (n_mio_pads > 0) | 
|  | assert (n_dio_pads > 0) | 
|  |  | 
|  | log.info("Generating pinmux with following info from hjson:") | 
|  | log.info("attr_dw:         %d" % attr_dw) | 
|  | log.info("num_wkup_detect: %d" % num_wkup_detect) | 
|  | log.info("wkup_cnt_width:  %d" % wkup_cnt_width) | 
|  | log.info("n_mio_periph_in:  %d" % n_mio_periph_in) | 
|  | log.info("n_mio_periph_out: %d" % n_mio_periph_out) | 
|  | log.info("n_dio_periph_in:  %d" % n_dio_periph_in) | 
|  | log.info("n_dio_periph_out: %d" % n_dio_periph_out) | 
|  | log.info("n_dio_pads:       %d" % n_dio_pads) | 
|  |  | 
|  | # Target path | 
|  | #   rtl: pinmux_reg_pkg.sv & pinmux_reg_top.sv | 
|  | #   data: pinmux.hjson | 
|  | rtl_path = out_path / "ip/pinmux/rtl/autogen" | 
|  | rtl_path.mkdir(parents=True, exist_ok=True) | 
|  | data_path = out_path / "ip/pinmux/data/autogen" | 
|  | data_path.mkdir(parents=True, exist_ok=True) | 
|  |  | 
|  | # Template path | 
|  | tpl_path = Path( | 
|  | __file__).resolve().parent / "../hw/ip/pinmux/data/pinmux.hjson.tpl" | 
|  | original_rtl_path = Path( | 
|  | __file__).resolve().parent / "../hw/ip/pinmux/rtl" | 
|  |  | 
|  | # Generate register package and RTLs | 
|  | gencmd = ("// util/topgen.py -t hw/top_{topname}/data/top_{topname}.hjson " | 
|  | "-o hw/top_{topname}/\n\n".format(topname=topname)) | 
|  |  | 
|  | hjson_gen_path = data_path / "pinmux.hjson" | 
|  |  | 
|  | out = StringIO() | 
|  | with tpl_path.open(mode="r", encoding="UTF-8") as fin: | 
|  | hjson_tpl = Template(fin.read()) | 
|  | try: | 
|  | out = hjson_tpl.render( | 
|  | n_mio_periph_in=n_mio_periph_in, | 
|  | n_mio_periph_out=n_mio_periph_out, | 
|  | n_mio_pads=n_mio_pads, | 
|  | # each DIO has in, out and oe wires | 
|  | # some of these have to be tied off in the | 
|  | # top, depending on the type. | 
|  | n_dio_periph_in=n_dio_pads, | 
|  | n_dio_periph_out=n_dio_pads, | 
|  | n_dio_pads=n_dio_pads, | 
|  | attr_dw=attr_dw, | 
|  | n_wkup_detect=num_wkup_detect, | 
|  | wkup_cnt_width=wkup_cnt_width) | 
|  | except:  # noqa: E722 | 
|  | log.error(exceptions.text_error_template().render()) | 
|  | log.info("PINMUX HJSON: %s" % out) | 
|  |  | 
|  | if out == "": | 
|  | log.error("Cannot generate pinmux HJSON") | 
|  | return | 
|  |  | 
|  | with hjson_gen_path.open(mode="w", encoding="UTF-8") as fout: | 
|  | fout.write(genhdr + gencmd + out) | 
|  |  | 
|  | # Generate reg file | 
|  | generate_regfile_from_path(hjson_gen_path, rtl_path, original_rtl_path) | 
|  |  | 
|  |  | 
|  | def generate_clkmgr(top, cfg_path, out_path): | 
|  |  | 
|  | # Target paths | 
|  | rtl_path = out_path / "ip/clkmgr/rtl/autogen" | 
|  | rtl_path.mkdir(parents=True, exist_ok=True) | 
|  | data_path = out_path / "ip/clkmgr/data/autogen" | 
|  | data_path.mkdir(parents=True, exist_ok=True) | 
|  |  | 
|  | # Template paths | 
|  | hjson_tpl = cfg_path / "../ip/clkmgr/data/clkmgr.hjson.tpl" | 
|  | rtl_tpl = cfg_path / "../ip/clkmgr/data/clkmgr.sv.tpl" | 
|  | pkg_tpl = cfg_path / "../ip/clkmgr/data/clkmgr_pkg.sv.tpl" | 
|  | original_rtl_path = cfg_path / "../ip/clkmgr/rtl" | 
|  |  | 
|  | hjson_out = data_path / "clkmgr.hjson" | 
|  | rtl_out = rtl_path / "clkmgr.sv" | 
|  | pkg_out = rtl_path / "clkmgr_pkg.sv" | 
|  |  | 
|  | tpls = [hjson_tpl, rtl_tpl, pkg_tpl] | 
|  | outputs = [hjson_out, rtl_out, pkg_out] | 
|  | names = ["clkmgr.hjson", "clkmgr.sv", "clkmgr_pkg.sv"] | 
|  |  | 
|  | clocks = top["clocks"] | 
|  | assert isinstance(clocks, Clocks) | 
|  |  | 
|  | typed_clocks = clocks.typed_clocks() | 
|  | hint_names = typed_clocks.hint_names() | 
|  |  | 
|  | for idx, tpl in enumerate(tpls): | 
|  | out = "" | 
|  | with tpl.open(mode="r", encoding="UTF-8") as fin: | 
|  | tpl = Template(fin.read()) | 
|  | try: | 
|  | out = tpl.render(cfg=top, | 
|  | clocks=clocks, | 
|  | typed_clocks=typed_clocks, | 
|  | hint_names=hint_names) | 
|  | except:  # noqa: E722 | 
|  | log.error(exceptions.text_error_template().render()) | 
|  |  | 
|  | if out == "": | 
|  | log.error("Cannot generate {}".format(names[idx])) | 
|  | return | 
|  |  | 
|  | with outputs[idx].open(mode="w", encoding="UTF-8") as fout: | 
|  | fout.write(genhdr + out) | 
|  |  | 
|  | # Generate reg files | 
|  | generate_regfile_from_path(hjson_out, rtl_path, original_rtl_path) | 
|  |  | 
|  |  | 
|  | # generate pwrmgr | 
|  | def generate_pwrmgr(top, out_path): | 
|  | log.info("Generating pwrmgr") | 
|  |  | 
|  | # Count number of wakeups | 
|  | n_wkups = len(top["wakeups"]) | 
|  | log.info("Found {} wakeup signals".format(n_wkups)) | 
|  |  | 
|  | # Count number of reset requests | 
|  | n_rstreqs = len(top["reset_requests"]["peripheral"]) | 
|  | log.info("Found {} reset request signals".format(n_rstreqs)) | 
|  |  | 
|  | if n_wkups < 1: | 
|  | n_wkups = 1 | 
|  | log.warning( | 
|  | "The design has no wakeup sources. Low power not supported.") | 
|  |  | 
|  | if n_rstreqs < 1: | 
|  | n_rstreqs = 1 | 
|  | log.warning("The design has no reset request sources. " | 
|  | "Reset requests are not supported.") | 
|  |  | 
|  | # Define target path | 
|  | rtl_path = out_path / "ip/pwrmgr/rtl/autogen" | 
|  | rtl_path.mkdir(parents=True, exist_ok=True) | 
|  | doc_path = out_path / "ip/pwrmgr/data/autogen" | 
|  | doc_path.mkdir(parents=True, exist_ok=True) | 
|  |  | 
|  | # So, read template files from ip directory. | 
|  | tpl_path = Path(__file__).resolve().parent / "../hw/ip/pwrmgr/data" | 
|  | hjson_tpl_path = tpl_path / "pwrmgr.hjson.tpl" | 
|  | original_rtl_path = Path(__file__).resolve().parent / "../hw/ip/pwrmgr/rtl" | 
|  |  | 
|  | # Render and write out hjson | 
|  | out = StringIO() | 
|  | with hjson_tpl_path.open(mode="r", encoding="UTF-8") as fin: | 
|  | hjson_tpl = Template(fin.read()) | 
|  | try: | 
|  | out = hjson_tpl.render(NumWkups=n_wkups, | 
|  | Wkups=top["wakeups"], | 
|  | rst_reqs=top["reset_requests"], | 
|  | NumRstReqs=n_rstreqs) | 
|  |  | 
|  | except:  # noqa: E722 | 
|  | log.error(exceptions.text_error_template().render()) | 
|  | log.info("pwrmgr hjson: %s" % out) | 
|  |  | 
|  | if out == "": | 
|  | log.error("Cannot generate pwrmgr config file") | 
|  | return | 
|  |  | 
|  | hjson_path = doc_path / "pwrmgr.hjson" | 
|  | with hjson_path.open(mode="w", encoding="UTF-8") as fout: | 
|  | fout.write(genhdr + out) | 
|  |  | 
|  | # Generate reg files | 
|  | generate_regfile_from_path(hjson_path, rtl_path, original_rtl_path) | 
|  |  | 
|  |  | 
|  | def get_rst_ni(top): | 
|  | rstmgrs = [m for m in top['module'] if m['type'] == 'rstmgr'] | 
|  | return rstmgrs[0]["reset_connections"] | 
|  |  | 
|  |  | 
|  | # generate rstmgr | 
|  | def generate_rstmgr(topcfg, out_path): | 
|  | log.info("Generating rstmgr") | 
|  |  | 
|  | # Define target path | 
|  | rtl_path = out_path / "ip/rstmgr/rtl/autogen" | 
|  | rtl_path.mkdir(parents=True, exist_ok=True) | 
|  | doc_path = out_path / "ip/rstmgr/data/autogen" | 
|  | doc_path.mkdir(parents=True, exist_ok=True) | 
|  | tpl_path = Path(__file__).resolve().parent / "../hw/ip/rstmgr/data" | 
|  | original_rtl_path = Path(__file__).resolve().parent / "../hw/ip/rstmgr/rtl" | 
|  |  | 
|  | # Read template files from ip directory. | 
|  | tpls = [] | 
|  | outputs = [] | 
|  | names = ["rstmgr.hjson", "rstmgr.sv", "rstmgr_pkg.sv"] | 
|  |  | 
|  | for x in names: | 
|  | tpls.append(tpl_path / Path(x + ".tpl")) | 
|  | if "hjson" in x: | 
|  | outputs.append(doc_path / Path(x)) | 
|  | else: | 
|  | outputs.append(rtl_path / Path(x)) | 
|  |  | 
|  | # Parameters needed for generation | 
|  | reset_obj = topcfg["resets"] | 
|  |  | 
|  | # The original resets dict is transformed to the reset class | 
|  | assert isinstance(reset_obj, Resets) | 
|  |  | 
|  | # unique clocks | 
|  | clks = reset_obj.get_clocks() | 
|  |  | 
|  | # resets sent to reset struct | 
|  | output_rsts = reset_obj.get_top_resets() | 
|  |  | 
|  | # sw controlled resets | 
|  | sw_rsts = reset_obj.get_sw_resets() | 
|  |  | 
|  | # rst_ni | 
|  | rst_ni = get_rst_ni(topcfg) | 
|  |  | 
|  | # leaf resets | 
|  | leaf_rsts = reset_obj.get_generated_resets() | 
|  |  | 
|  | # Number of reset requests | 
|  | n_rstreqs = len(topcfg["reset_requests"]["peripheral"]) | 
|  |  | 
|  | # Generate templated files | 
|  | for idx, t in enumerate(tpls): | 
|  | out = StringIO() | 
|  | with t.open(mode="r", encoding="UTF-8") as fin: | 
|  | tpl = Template(fin.read()) | 
|  | try: | 
|  | out = tpl.render(clks=clks, | 
|  | reqs=topcfg["reset_requests"], | 
|  | power_domains=topcfg["power"]["domains"], | 
|  | num_rstreqs=n_rstreqs, | 
|  | sw_rsts=sw_rsts, | 
|  | output_rsts=output_rsts, | 
|  | leaf_rsts=leaf_rsts, | 
|  | rst_ni = rst_ni['rst_ni']['name'], | 
|  | export_rsts=topcfg["exported_rsts"], | 
|  | reset_obj=topcfg["resets"]) | 
|  |  | 
|  | except:  # noqa: E722 | 
|  | log.error(exceptions.text_error_template().render()) | 
|  |  | 
|  | if out == "": | 
|  | log.error("Cannot generate {}".format(names[idx])) | 
|  | return | 
|  |  | 
|  | with outputs[idx].open(mode="w", encoding="UTF-8") as fout: | 
|  | fout.write(genhdr + out) | 
|  |  | 
|  | # Generate reg files | 
|  | hjson_path = outputs[0] | 
|  | generate_regfile_from_path(hjson_path, rtl_path, original_rtl_path) | 
|  |  | 
|  |  | 
|  | # generate flash | 
|  | def generate_flash(topcfg, out_path): | 
|  | log.info("Generating flash") | 
|  |  | 
|  | # Define target path | 
|  | rtl_path = out_path / "ip/flash_ctrl/rtl/autogen" | 
|  | rtl_path.mkdir(parents=True, exist_ok=True) | 
|  | doc_path = out_path / "ip/flash_ctrl/data/autogen" | 
|  | doc_path.mkdir(parents=True, exist_ok=True) | 
|  | tpl_path = Path(__file__).resolve().parent / "../hw/ip/flash_ctrl/data" | 
|  | original_rtl_path = Path( | 
|  | __file__).resolve().parent / "../hw/ip/flash_ctrl/rtl" | 
|  |  | 
|  | # Read template files from ip directory. | 
|  | tpls = [] | 
|  | outputs = [] | 
|  | names = [ | 
|  | "flash_ctrl.hjson", "flash_ctrl.sv", "flash_ctrl_pkg.sv", | 
|  | "flash_ctrl_region_cfg.sv" | 
|  | ] | 
|  |  | 
|  | for x in names: | 
|  | tpls.append(tpl_path / Path(x + ".tpl")) | 
|  | if "hjson" in x: | 
|  | outputs.append(doc_path / Path(x)) | 
|  | else: | 
|  | outputs.append(rtl_path / Path(x)) | 
|  |  | 
|  | # Parameters needed for generation | 
|  | flash_mems = [ | 
|  | module for module in topcfg["module"] if module["type"] == "flash_ctrl" | 
|  | ] | 
|  | if len(flash_mems) > 1: | 
|  | log.error("This design does not currently support multiple flashes") | 
|  | return | 
|  |  | 
|  | cfg = flash_mems[0]["memory"]["mem"]["config"] | 
|  |  | 
|  | # Generate templated files | 
|  | for idx, t in enumerate(tpls): | 
|  | out = StringIO() | 
|  | with t.open(mode="r", encoding="UTF-8") as fin: | 
|  | tpl = Template(fin.read()) | 
|  | try: | 
|  | out = tpl.render(cfg=cfg) | 
|  |  | 
|  | except:  # noqa: E722 | 
|  | log.error(exceptions.text_error_template().render()) | 
|  |  | 
|  | if out == "": | 
|  | log.error("Cannot generate {}".format(names[idx])) | 
|  | return | 
|  |  | 
|  | with outputs[idx].open(mode="w", encoding="UTF-8") as fout: | 
|  | fout.write(genhdr + out) | 
|  |  | 
|  | # Generate reg files | 
|  | hjson_path = outputs[0] | 
|  | generate_regfile_from_path(hjson_path, rtl_path, original_rtl_path) | 
|  |  | 
|  |  | 
|  | def generate_top_only(top_only_dict, out_path, topname, alt_hjson_path): | 
|  | log.info("Generating top only modules") | 
|  |  | 
|  | for ip, reggen_only in top_only_dict.items(): | 
|  |  | 
|  | if reggen_only and alt_hjson_path is not None: | 
|  | hjson_dir = Path(alt_hjson_path) | 
|  | else: | 
|  | hjson_dir = Path(__file__).resolve( | 
|  | ).parent / f"../hw/top_{topname}/ip/{ip}/data/" | 
|  |  | 
|  | hjson_path = hjson_dir / f"{ip}.hjson" | 
|  |  | 
|  | genrtl_dir = out_path / "ip/{}/rtl".format(ip) | 
|  | genrtl_dir.mkdir(parents=True, exist_ok=True) | 
|  | log.info("Generating top modules {}, hjson: {}, output: {}".format( | 
|  | ip, hjson_path, genrtl_dir)) | 
|  |  | 
|  | # Generate reg files | 
|  | generate_regfile_from_path(hjson_path, genrtl_dir) | 
|  |  | 
|  |  | 
|  | def generate_top_ral(top: Dict[str, object], name_to_block: Dict[str, IpBlock], | 
|  | dv_base_names: List[str], out_path: str): | 
|  | # construct top ral block | 
|  | regwidth = int(top["datawidth"]) | 
|  | assert regwidth % 8 == 0 | 
|  | addrsep = regwidth // 8 | 
|  |  | 
|  | # Generate a map from instance name to the block that it instantiates, | 
|  | # together with a map of interface addresses. | 
|  | inst_to_block = {}  # type: Dict[str, str] | 
|  | if_addrs = {}  # type: Dict[Tuple[str, Optional[str]], int] | 
|  | attrs = {}  # type: Dict[str, str] | 
|  |  | 
|  | for module in top["module"]: | 
|  | inst_name = module["name"] | 
|  | block_name = module["type"] | 
|  | block = name_to_block[block_name] | 
|  | if "attr" in module: | 
|  | if module["attr"] not in ["templated", "ipgen", "reggen_top", | 
|  | "reggen_only"]: | 
|  | raise ValueError("Unsupported value for attr field of {}: {!r}" | 
|  | .format(inst_name, module["attr"])) | 
|  | attrs[inst_name] = module["attr"] | 
|  |  | 
|  | inst_to_block[inst_name] = block_name | 
|  | for if_name in block.reg_blocks.keys(): | 
|  | if_addr = int(module["base_addrs"][if_name], 0) | 
|  | if_addrs[(inst_name, if_name)] = if_addr | 
|  |  | 
|  | # Collect up the memories to add | 
|  | mems = [] | 
|  | for item in list(top.get("memory", [])): | 
|  | mems.append(create_mem(item, addrsep, regwidth)) | 
|  |  | 
|  | # Top-level may override the mem setting. Store the new type to name_to_block | 
|  | # If no other instance uses the orignal type, delete it | 
|  | original_types = set() | 
|  | for module in top["module"]: | 
|  | if "memory" in module.keys() and len(module["memory"]) > 0: | 
|  | newtype = "{}_{}".format(module["type"], module["name"]) | 
|  | assert newtype not in name_to_block | 
|  |  | 
|  | block = deepcopy(name_to_block[module["type"]]) | 
|  | name_to_block[newtype] = block | 
|  | inst_to_block[module["name"]] = newtype | 
|  |  | 
|  | original_types.add(module["type"]) | 
|  |  | 
|  | for mem_name, item in module["memory"].items(): | 
|  | assert block.reg_blocks[mem_name] | 
|  | assert len(block.reg_blocks[mem_name].windows) <= 1 | 
|  | item["name"] = mem_name | 
|  |  | 
|  | win = create_mem(item, addrsep, regwidth) | 
|  | if len(block.reg_blocks[mem_name].windows) > 0: | 
|  | blk_win = block.reg_blocks[mem_name].windows[0] | 
|  |  | 
|  | # Top can only add new info for mem, shouldn't overwrite | 
|  | # existing configuration | 
|  | assert win.items == blk_win.items | 
|  | assert win.byte_write == blk_win.byte_write | 
|  | assert win.data_intg_passthru == blk_win.data_intg_passthru | 
|  |  | 
|  | block.reg_blocks[mem_name].windows[0] = win | 
|  | else: | 
|  | block.reg_blocks[mem_name].windows.append(win) | 
|  |  | 
|  | for t in original_types: | 
|  | if t not in inst_to_block.values(): | 
|  | del name_to_block[t] | 
|  |  | 
|  | chip = Top(regwidth, name_to_block, inst_to_block, if_addrs, mems, attrs) | 
|  |  | 
|  | # generate the top ral model with template | 
|  | return gen_dv(chip, dv_base_names, str(out_path)) | 
|  |  | 
|  |  | 
|  | def create_mem(item, addrsep, regwidth): | 
|  | byte_write = ("byte_write" in item and | 
|  | item["byte_write"].lower() == "true") | 
|  | data_intg_passthru = ("data_intg_passthru" in item and | 
|  | item["data_intg_passthru"].lower() == "true") | 
|  | size_in_bytes = int(item["size"], 0) | 
|  | num_regs = size_in_bytes // addrsep | 
|  | swaccess = access.SWAccess("top-level memory", item.get("swaccess", "rw")) | 
|  |  | 
|  | return window.Window(name=item["name"], | 
|  | desc="(generated from top-level)", | 
|  | unusual=False, | 
|  | byte_write=byte_write, | 
|  | data_intg_passthru=data_intg_passthru, | 
|  | validbits=regwidth, | 
|  | items=num_regs, | 
|  | size_in_bytes=size_in_bytes, | 
|  | offset=int(item.get("base_addr", "0"), 0), | 
|  | swaccess=swaccess) | 
|  |  | 
|  |  | 
|  | def _process_top(topcfg, args, cfg_path, out_path, pass_idx): | 
|  | # Create generated list | 
|  | # These modules are generated through topgen | 
|  | templated_list = lib.get_templated_modules(topcfg) | 
|  | log.info("Templated list is {}".format(templated_list)) | 
|  |  | 
|  | ipgen_list = lib.get_ipgen_modules(topcfg) | 
|  | log.info("Ip gen list is {}".format(ipgen_list)) | 
|  |  | 
|  | generated_list = templated_list + ipgen_list | 
|  |  | 
|  | # These modules are NOT generated but belong to a specific top | 
|  | # and therefore not part of "hw/ip" | 
|  | top_only_dict = { | 
|  | module["type"]: lib.is_reggen_only(module) | 
|  | for module in topcfg["module"] if lib.is_top_reggen(module) | 
|  | } | 
|  | log.info("Filtered dict is {}".format(top_only_dict)) | 
|  |  | 
|  | topname = topcfg["name"] | 
|  |  | 
|  | # Sweep the IP directory and gather the config files | 
|  | ip_dir = Path(__file__).parents[1] / "hw/ip" | 
|  | ips = search_ips(ip_dir) | 
|  |  | 
|  | # exclude filtered IPs (to use top_${topname} one) and | 
|  | exclude_list = generated_list + list(top_only_dict.keys()) | 
|  | ips = [x for x in ips if not x.parents[1].name in exclude_list] | 
|  |  | 
|  | # Hack alert | 
|  | # Generate clkmgr.hjson here so that it can be included below | 
|  | # Unlike other generated hjsons, clkmgr thankfully does not require | 
|  | # ip.hjson information.  All the information is embedded within | 
|  | # the top hjson file | 
|  | topcfg["clocks"] = Clocks(topcfg["clocks"]) | 
|  | extract_clocks(topcfg) | 
|  | generate_clkmgr(topcfg, cfg_path, out_path) | 
|  |  | 
|  | # It may require two passes to check if the module is needed. | 
|  | # TODO: first run of topgen will fail due to the absent of rv_plic. | 
|  | # It needs to run up to amend_interrupt in merge_top function | 
|  | # then creates rv_plic.hjson then run xbar generation. | 
|  | hjson_dir = Path(args.topcfg).parent | 
|  |  | 
|  | for ip in generated_list: | 
|  | # For modules that are generated prior to gathering, we need to take it from | 
|  | # the output path.  For modules not generated before, it may exist in a | 
|  | # pre-defined area already. | 
|  | log.info("Appending {}".format(ip)) | 
|  | if ip in ipgen_list: | 
|  | ip_relpath = "ip_autogen" | 
|  | desc_file_relpath = "data" | 
|  | else: | 
|  | ip_relpath = "ip" | 
|  | desc_file_relpath = "data/autogen" | 
|  |  | 
|  | if ip == "clkmgr" or (pass_idx > 0): | 
|  | ip_hjson = (Path(out_path) / ip_relpath / ip / desc_file_relpath / | 
|  | f"{ip}.hjson") | 
|  | else: | 
|  | ip_hjson = (hjson_dir.parent / ip_relpath / ip / | 
|  | desc_file_relpath / f"{ip}.hjson") | 
|  | ips.append(ip_hjson) | 
|  |  | 
|  | for ip, reggen_only in top_only_dict.items(): | 
|  | log.info("Appending {}".format(ip)) | 
|  |  | 
|  | if reggen_only and args.hjson_path: | 
|  | ip_hjson = Path(args.hjson_path) / f"{ip}.hjson" | 
|  | else: | 
|  | ip_hjson = hjson_dir.parent / f"ip/{ip}/data/{ip}.hjson" | 
|  |  | 
|  | ips.append(ip_hjson) | 
|  |  | 
|  | # load Hjson and pass validate from reggen | 
|  | try: | 
|  | ip_objs = [] | 
|  | for ip_desc_file in ips: | 
|  | ip_name = ip_desc_file.stem | 
|  | # Skip if it is not in the module list | 
|  | if ip_name not in [ip["type"] for ip in topcfg["module"]]: | 
|  | log.info("Skip module %s as it isn't in the top module list" % | 
|  | ip_name) | 
|  | continue | 
|  |  | 
|  | # The auto-generated hjson might not yet exist. It will be created | 
|  | # later, see generate_{ip_name}() calls below. For the initial | 
|  | # validation, use the Hjson file with default values. | 
|  | # TODO: All of this is a rather ugly hack that we need to get rid | 
|  | # of as soon as we don't arbitrarily template IP description Hjson | 
|  | # files any more. | 
|  | if ip_name in generated_list and not ip_desc_file.is_file(): | 
|  | if ip_name in ipgen_list: | 
|  | log.info( | 
|  | "To-be-auto-generated Hjson %s does not yet exist. " | 
|  | "Falling back to the default configuration of template " | 
|  | "%s for initial validation." % (ip_desc_file, ip_name)) | 
|  |  | 
|  | tpl_path = SRCTREE_TOP / "hw/ip_templates" / ip_name | 
|  | ip_template = IpTemplate.from_template_path(tpl_path) | 
|  | ip_config = IpConfig(ip_template.params, | 
|  | f"top_{topname}_{ip_name}") | 
|  |  | 
|  | try: | 
|  | ip_desc = IpDescriptionOnlyRenderer( | 
|  | ip_template, ip_config).render() | 
|  | except TemplateRenderError as e: | 
|  | log.error(e.verbose_str()) | 
|  | sys.exit(1) | 
|  | s = "default description of IP template {}".format(ip_name) | 
|  | ip_objs.append(IpBlock.from_text(ip_desc, [], s)) | 
|  | else: | 
|  | # TODO: Remove this block as soon as all IP templates use | 
|  | # ipgen. | 
|  | template_hjson_file = ip_dir / "{}/data/{}.hjson".format( | 
|  | ip_name, ip_name) | 
|  | log.info( | 
|  | "To-be-auto-generated Hjson %s does not yet exist. " | 
|  | "Falling back to Hjson description file %s shipped " | 
|  | "with the IP template for initial validation." % | 
|  | (ip_desc_file, template_hjson_file)) | 
|  |  | 
|  | ip_objs.append( | 
|  | IpBlock.from_path(str(template_hjson_file), [])) | 
|  | else: | 
|  | ip_objs.append(IpBlock.from_path(str(ip_desc_file), [])) | 
|  |  | 
|  | except ValueError: | 
|  | raise SystemExit(sys.exc_info()[1]) | 
|  |  | 
|  | name_to_block = {}  # type: Dict[str, IpBlock] | 
|  | for block in ip_objs: | 
|  | lblock = block.name.lower() | 
|  | assert lblock not in name_to_block | 
|  | name_to_block[lblock] = block | 
|  |  | 
|  | # Read in alias files one-by-one, peek inside to figure out which IP block | 
|  | # they belong to and apply the alias file to that IP block. | 
|  | if args.alias_files: | 
|  | for alias in args.alias_files: | 
|  | with open(alias, 'r', encoding='utf-8') as handle: | 
|  | raw = hjson.loads(handle.read(), use_decimal=True) | 
|  | if 'alias_target' not in raw: | 
|  | raise ValueError('Missing alias_target key ' | 
|  | 'in alias file {}.'.format(alias)) | 
|  | alias_target = raw['alias_target'].lower() | 
|  | if alias_target not in name_to_block: | 
|  | raise ValueError('Alias target {} is not defined.' | 
|  | .format(alias_target)) | 
|  | where = 'alias file at {}'.format(alias) | 
|  | name_to_block[alias_target].alias_from_raw(False, raw, where) | 
|  |  | 
|  | connect_clocks(topcfg, name_to_block) | 
|  |  | 
|  | # Read the crossbars under the top directory | 
|  | xbar_objs = get_hjsonobj_xbars(hjson_dir) | 
|  |  | 
|  | log.info("Detected crossbars: %s" % | 
|  | (", ".join([x["name"] for x in xbar_objs]))) | 
|  |  | 
|  | topcfg, error = validate_top(topcfg, ip_objs, xbar_objs) | 
|  | if error != 0: | 
|  | raise SystemExit("Error occured while validating top.hjson") | 
|  |  | 
|  | completecfg = merge_top(topcfg, name_to_block, xbar_objs) | 
|  |  | 
|  | # Generate flash controller and flash memory | 
|  | generate_flash(topcfg, out_path) | 
|  |  | 
|  | # Generate PLIC | 
|  | if not args.no_plic and \ | 
|  | not args.alert_handler_only and \ | 
|  | not args.xbar_only: | 
|  | generate_plic(completecfg, out_path) | 
|  | if args.plic_only: | 
|  | sys.exit() | 
|  |  | 
|  | # Create Alert Handler LPGs before | 
|  | # generating the Alert Handler | 
|  | create_alert_lpgs(topcfg, name_to_block) | 
|  |  | 
|  | # Generate Alert Handler | 
|  | if not args.xbar_only: | 
|  | generate_alert_handler(completecfg, out_path) | 
|  | if args.alert_handler_only: | 
|  | sys.exit() | 
|  |  | 
|  | # Generate Pinmux | 
|  | generate_pinmux(completecfg, out_path) | 
|  |  | 
|  | # Generate Pwrmgr | 
|  | generate_pwrmgr(completecfg, out_path) | 
|  |  | 
|  | # Generate rstmgr | 
|  | generate_rstmgr(completecfg, out_path) | 
|  |  | 
|  | # Generate top only modules | 
|  | # These modules are not templated, but are not in hw/ip | 
|  | generate_top_only(top_only_dict, out_path, topname, args.hjson_path) | 
|  |  | 
|  | return completecfg, name_to_block | 
|  |  | 
|  |  | 
|  | def main(): | 
|  | parser = argparse.ArgumentParser(prog="topgen") | 
|  | parser.add_argument("--topcfg", | 
|  | "-t", | 
|  | required=True, | 
|  | help="`top_{name}.hjson` file.") | 
|  | parser.add_argument( | 
|  | "--outdir", | 
|  | "-o", | 
|  | help="""Target TOP directory. | 
|  | Module is created under rtl/. (default: dir(topcfg)/..) | 
|  | """)  # yapf: disable | 
|  | parser.add_argument( | 
|  | "--hjson-path", | 
|  | help=""" | 
|  | If defined, topgen uses supplied path to search for ip hjson. | 
|  | This applies only to ip's with the `reggen_only` attribute. | 
|  | If an hjson is located both in the conventional path and the alternate | 
|  | path, the alternate path has priority. | 
|  | """) | 
|  | parser.add_argument("--verbose", "-v", action="store_true", help="Verbose") | 
|  |  | 
|  | # Generator options: 'no' series. cannot combined with 'only' series | 
|  | parser.add_argument( | 
|  | "--no-top", | 
|  | action="store_true", | 
|  | help="If defined, topgen doesn't generate top_{name} RTLs.") | 
|  | parser.add_argument( | 
|  | "--no-xbar", | 
|  | action="store_true", | 
|  | help="If defined, topgen doesn't generate crossbar RTLs.") | 
|  | parser.add_argument( | 
|  | "--no-plic", | 
|  | action="store_true", | 
|  | help="If defined, topgen doesn't generate the interrup controller RTLs." | 
|  | ) | 
|  |  | 
|  | # Generator options: 'only' series. cannot combined with 'no' series | 
|  | parser.add_argument( | 
|  | "--top-only", | 
|  | action="store_true", | 
|  | help="If defined, the tool generates top RTL only")  # yapf:disable | 
|  | parser.add_argument( | 
|  | "--xbar-only", | 
|  | action="store_true", | 
|  | help="If defined, the tool generates crossbar RTLs only") | 
|  | parser.add_argument( | 
|  | "--plic-only", | 
|  | action="store_true", | 
|  | help="If defined, the tool generates RV_PLIC RTL and Hjson only") | 
|  | parser.add_argument( | 
|  | "--alert-handler-only", | 
|  | action="store_true", | 
|  | help="If defined, the tool generates alert handler hjson only") | 
|  | # Generator options: generate dv ral model | 
|  | parser.add_argument( | 
|  | "--top_ral", | 
|  | "-r", | 
|  | default=False, | 
|  | action="store_true", | 
|  | help="If set, the tool generates top level RAL model for DV") | 
|  | parser.add_argument( | 
|  | "--alias-files", | 
|  | nargs="+", | 
|  | type=Path, | 
|  | default=None, | 
|  | help=""" | 
|  | If defined, topgen uses supplied alias hjson file(s) to override the | 
|  | generic register definitions when building the RAL model. This | 
|  | argument is only relevant in conjunction with the `--top_ral` switch. | 
|  | """) | 
|  | parser.add_argument( | 
|  | "--dv-base-names", | 
|  | nargs="+", | 
|  | help="Names or prefix for the DV register classes from which " | 
|  | "the register models are derived.") | 
|  | # Generator options for compile time random netlist constants | 
|  | parser.add_argument( | 
|  | "--rnd_cnst_seed", | 
|  | type=int, | 
|  | metavar="<seed>", | 
|  | help="Custom seed for RNG to compute netlist constants.") | 
|  | parser.add_argument( | 
|  | "--entropy_buffer", | 
|  | help="A file with entropy.") | 
|  | # Miscellaneous: only return the list of blocks and exit. | 
|  | parser.add_argument("--get_blocks", | 
|  | default=False, | 
|  | action="store_true", | 
|  | help="Only return the list of blocks and exit.") | 
|  |  | 
|  | args = parser.parse_args() | 
|  |  | 
|  | # check combinations | 
|  | if args.top_ral: | 
|  | args.no_top = True | 
|  |  | 
|  | if (args.no_top or args.no_xbar or | 
|  | args.no_plic) and (args.top_only or args.xbar_only or | 
|  | args.plic_only or args.alert_handler_only): | 
|  | log.error( | 
|  | "'no' series options cannot be used with 'only' series options") | 
|  | raise SystemExit(sys.exc_info()[1]) | 
|  |  | 
|  | # Don't print warnings when querying the list of blocks. | 
|  | log_level = (log.ERROR | 
|  | if args.get_blocks else log.DEBUG if args.verbose else None) | 
|  |  | 
|  | log.basicConfig(format="%(levelname)s: %(message)s", level=log_level) | 
|  |  | 
|  | if not args.outdir: | 
|  | outdir = Path(args.topcfg).parent / ".." | 
|  | log.info("TOP directory not given. Use %s", (outdir)) | 
|  | elif not Path(args.outdir).is_dir(): | 
|  | log.error("'--outdir' should point to writable directory") | 
|  | raise SystemExit(sys.exc_info()[1]) | 
|  | else: | 
|  | outdir = Path(args.outdir) | 
|  |  | 
|  | if args.hjson_path is not None: | 
|  | log.info(f"Alternate hjson path is {args.hjson_path}") | 
|  |  | 
|  | out_path = Path(outdir) | 
|  | cfg_path = Path(args.topcfg).parents[1] | 
|  |  | 
|  | try: | 
|  | with open(args.topcfg, "r") as ftop: | 
|  | topcfg = hjson.load(ftop, | 
|  | use_decimal=True, | 
|  | object_pairs_hook=OrderedDict) | 
|  | except ValueError: | 
|  | raise SystemExit(sys.exc_info()[1]) | 
|  |  | 
|  | # Initialize RNG for compile-time netlist constants. | 
|  | if args.entropy_buffer: | 
|  | if args.rnd_cnst_seed: | 
|  | log.error("'entropy_buffer' option cannot be used with 'rnd_cnst_seed option'") | 
|  | # error out | 
|  | raise SystemExit(sys.exc_info()[1]) | 
|  | else: | 
|  | # generate entropy from a buffer | 
|  | strong_random.load(SRCTREE_TOP / args.entropy_buffer) | 
|  | else: | 
|  | # If specified, override the seed for random netlist constant computation. | 
|  | if args.rnd_cnst_seed: | 
|  | log.warning("Commandline override of rnd_cnst_seed with {}.".format( | 
|  | args.rnd_cnst_seed)) | 
|  | topcfg["rnd_cnst_seed"] = args.rnd_cnst_seed | 
|  | # Otherwise, we either take it from the top_{topname}.hjson if present, or | 
|  | # randomly generate a new seed if not. | 
|  | else: | 
|  | random.seed() | 
|  | new_seed = random.getrandbits(64) | 
|  | if topcfg.setdefault("rnd_cnst_seed", new_seed) == new_seed: | 
|  | log.warning( | 
|  | "No rnd_cnst_seed specified, setting to {}.".format(new_seed)) | 
|  | ebg.gen_buffer(BUFFER_SIZE, SRCTREE_TOP / PATH_TO_BUFFER, False, topcfg["rnd_cnst_seed"]) | 
|  | strong_random.load(SRCTREE_TOP / PATH_TO_BUFFER) | 
|  |  | 
|  | # TODO, long term, the levels of dependency should be automatically determined instead | 
|  | # of hardcoded.  The following are a few examples: | 
|  | # Example 1: pinmux depends on amending all modules before calculating the correct number of | 
|  | #            pins. | 
|  | #            This would be 1 level of dependency and require 2 passes. | 
|  | # Example 2: pinmux depends on amending all modules, and pwrmgr depends on pinmux generation to | 
|  | #            know correct number of wakeups.  This would be 2 levels of dependency and require 3 | 
|  | #            passes. | 
|  | # | 
|  | # How does mulit-pass work? | 
|  | # In example 1, the first pass gathers all modules and merges them.  However, the merge process | 
|  | # uses a stale pinmux.  The correct pinmux is then generated using the merged configuration. The | 
|  | # second pass now merges all the correct modules (including the generated pinmux) and creates | 
|  | # the final merged config. | 
|  | # | 
|  | # In example 2, the first pass gathers all modules and merges them.  However, the merge process | 
|  | # uses a stale pinmux and pwrmgr. The correct pinmux is then generated using the merged | 
|  | # configuration.  However, since pwrmgr is dependent on this new pinmux, it is still generated | 
|  | # incorrectly.  The second pass merge now has an updated pinmux but stale pwrmgr.  The correct | 
|  | # pwrmgr can now be generated.  The final pass then merges all the correct modules and creates | 
|  | # the final configuration. | 
|  | # | 
|  | # This fix is related to #2083 | 
|  | process_dependencies = 1 | 
|  |  | 
|  | # topgen generates IP blocks and associated Hjson configuration in multiple | 
|  | # steps. After each step, the IP Hjson configuration is read back and then | 
|  | # combined into the toplevel configuration. To generate the chip-level RAL, | 
|  | # we need to run the full generation step, but ultimately only care about | 
|  | # the toplevel configuration (a single Hjson file). Since we don't have a | 
|  | # better way at the moment dump all output into a temporary directory, and | 
|  | # delete it after the fact, retaining only the toplevel configuration. | 
|  | if args.top_ral: | 
|  | out_path_gen = Path(tempfile.mkdtemp()) | 
|  | else: | 
|  | out_path_gen = out_path | 
|  |  | 
|  | for pass_idx in range(process_dependencies + 1): | 
|  | log.debug("Generation pass {}".format(pass_idx)) | 
|  | if pass_idx < process_dependencies: | 
|  | cfg_copy = deepcopy(topcfg) | 
|  | _process_top(cfg_copy, args, cfg_path, out_path_gen, pass_idx) | 
|  | else: | 
|  | completecfg, name_to_block = _process_top(topcfg, args, cfg_path, | 
|  | out_path_gen, pass_idx) | 
|  |  | 
|  | topname = topcfg["name"] | 
|  |  | 
|  | if not args.entropy_buffer: | 
|  | # Delete entropy buffer since it is no longer needed. | 
|  | # This buffer can always be re-generated from the seed using entropy_buffer_generator | 
|  | os.remove(SRCTREE_TOP / PATH_TO_BUFFER) | 
|  |  | 
|  | # Create the chip-level RAL only | 
|  | if args.top_ral: | 
|  | # See above: we only need `completeconfig` and `name_to_block`, not all | 
|  | # the other files (e.g. RTL files) generated through topgen. | 
|  | shutil.rmtree(out_path_gen, ignore_errors=True) | 
|  |  | 
|  | exit_code = generate_top_ral(completecfg, name_to_block, | 
|  | args.dv_base_names, out_path) | 
|  | sys.exit(exit_code) | 
|  |  | 
|  | if args.get_blocks: | 
|  | print("\n".join(name_to_block.keys())) | 
|  | sys.exit(0) | 
|  |  | 
|  | # Generate xbars | 
|  | if not args.no_xbar or args.xbar_only: | 
|  | generate_xbars(completecfg, out_path) | 
|  |  | 
|  | # All IPs are generated. Connect phase now | 
|  | # Find {memory, module} <-> {xbar} connections first. | 
|  | im.autoconnect(completecfg, name_to_block) | 
|  |  | 
|  | # Generic Inter-module connection | 
|  | im.elab_intermodule(completecfg) | 
|  |  | 
|  | # Generate top.gen.hjson right before rendering | 
|  | genhjson_dir = out_path / "data/autogen" | 
|  | genhjson_dir.mkdir(parents=True, exist_ok=True) | 
|  | genhjson_path = genhjson_dir / ("top_%s.gen.hjson" % completecfg["name"]) | 
|  |  | 
|  | # Header for HJSON | 
|  | if args.entropy_buffer: | 
|  | gencmd = """// | 
|  | // util/topgen.py -t hw/top_{topname}/data/top_{topname}.hjson \\ | 
|  | //                -o hw/top_{topname}/ \\ | 
|  | //                --hjson-only \\ | 
|  | //                --entropy-buffer {path} | 
|  | """.format(topname=topname, path = args.entropy_buffer) | 
|  | else: | 
|  | gencmd = """// | 
|  | // util/topgen.py -t hw/top_{topname}/data/top_{topname}.hjson \\ | 
|  | //                -o hw/top_{topname}/ \\ | 
|  | //                --hjson-only \\ | 
|  | //                --rnd_cnst_seed {seed} | 
|  | """.format(topname=topname, seed=completecfg["rnd_cnst_seed"]) | 
|  |  | 
|  | genhjson_path.write_text(genhdr + gencmd + | 
|  | hjson.dumps(completecfg, for_json=True) + '\n') | 
|  |  | 
|  | if not args.no_top or args.top_only: | 
|  |  | 
|  | def render_template(template_path: str, rendered_path: Path, | 
|  | **other_info): | 
|  | template_contents = generate_top(completecfg, name_to_block, | 
|  | str(template_path), **other_info) | 
|  |  | 
|  | rendered_path.parent.mkdir(exist_ok=True, parents=True) | 
|  | with rendered_path.open(mode="w", encoding="UTF-8") as fout: | 
|  | fout.write(template_contents) | 
|  |  | 
|  | # Header for SV files | 
|  | if args.entropy_buffer: | 
|  | gencmd = warnhdr + """// | 
|  | // util/topgen.py -t hw/top_{topname}/data/top_{topname}.hjson \\ | 
|  | //                -o hw/top_{topname}/ \\ | 
|  | //                --entropy-buffer {path} | 
|  | """.format(topname=topname, path = args.entropy_buffer) | 
|  | else: | 
|  | gencmd = warnhdr + """// | 
|  | // util/topgen.py -t hw/top_{topname}/data/top_{topname}.hjson \\ | 
|  | //                -o hw/top_{topname}/ \\ | 
|  | //                --rnd_cnst_seed {seed} | 
|  | """.format(topname=topname, seed=completecfg["rnd_cnst_seed"]) | 
|  |  | 
|  | # SystemVerilog Top: | 
|  | # "toplevel.sv.tpl" -> "rtl/autogen/top_{topname}.sv" | 
|  | render_template(TOPGEN_TEMPLATE_PATH / "toplevel.sv.tpl", | 
|  | out_path / f"rtl/autogen/top_{topname}.sv", | 
|  | gencmd=gencmd) | 
|  |  | 
|  | # Multiple chip-levels (ASIC, FPGA, Verilator, etc) | 
|  | for target in topcfg["targets"]: | 
|  | target_name = target["name"] | 
|  | render_template(TOPGEN_TEMPLATE_PATH / "chiplevel.sv.tpl", | 
|  | out_path / | 
|  | f"rtl/autogen/chip_{topname}_{target_name}.sv", | 
|  | gencmd=gencmd, | 
|  | target=target) | 
|  |  | 
|  | # The C / SV file needs some complex information, so we initialize this | 
|  | # object to store it. | 
|  | c_helper = TopGenCTest(completecfg, name_to_block) | 
|  |  | 
|  | # The Rust file needs some complex information, so we initialize this | 
|  | # object to store it. | 
|  | rs_helper = TopGenRust(completecfg, name_to_block) | 
|  |  | 
|  | # "toplevel_pkg.sv.tpl" -> "rtl/autogen/top_{topname}_pkg.sv" | 
|  | render_template(TOPGEN_TEMPLATE_PATH / "toplevel_pkg.sv.tpl", | 
|  | out_path / f"rtl/autogen/top_{topname}_pkg.sv", | 
|  | helper=c_helper, | 
|  | gencmd=gencmd) | 
|  |  | 
|  | # compile-time random netlist constants | 
|  | render_template(TOPGEN_TEMPLATE_PATH / "toplevel_rnd_cnst_pkg.sv.tpl", | 
|  | out_path / | 
|  | f"rtl/autogen/top_{topname}_rnd_cnst_pkg.sv", | 
|  | gencmd=gencmd) | 
|  |  | 
|  | # C Header + C File + Clang-format file | 
|  |  | 
|  | # Since SW does not use FuseSoC and instead expects those files always | 
|  | # to be in hw/top_{topname}/sw/autogen, we currently create these files | 
|  | # twice: | 
|  | # - Once under out_path/sw/autogen | 
|  | # - Once under hw/top_{topname}/sw/autogen | 
|  | root_paths = [out_path.resolve(), SRCTREE_TOP] | 
|  | out_paths = [ | 
|  | out_path.resolve(), | 
|  | (SRCTREE_TOP / "hw/top_{}/".format(topname)).resolve() | 
|  | ] | 
|  | for idx, path in enumerate(out_paths): | 
|  | # "clang-format" -> "sw/autogen/.clang-format" | 
|  | cformat_tplpath = TOPGEN_TEMPLATE_PATH / "clang-format" | 
|  | cformat_dir = path / "sw/autogen" | 
|  | cformat_dir.mkdir(parents=True, exist_ok=True) | 
|  | cformat_path = cformat_dir / ".clang-format" | 
|  | cformat_path.write_text(cformat_tplpath.read_text()) | 
|  |  | 
|  | # Save the header macro prefix into `c_helper` | 
|  | rel_header_dir = cformat_dir.relative_to(root_paths[idx]) | 
|  | c_helper.header_macro_prefix = ( | 
|  | "OPENTITAN_" + str(rel_header_dir).replace("/", "_").upper()) | 
|  |  | 
|  | # "top_{topname}.h.tpl" -> "sw/autogen/top_{topname}.h" | 
|  | cheader_path = cformat_dir / f"top_{topname}.h" | 
|  | render_template(TOPGEN_TEMPLATE_PATH / "toplevel.h.tpl", | 
|  | cheader_path, | 
|  | helper=c_helper) | 
|  |  | 
|  | # Save the relative header path into `c_helper` | 
|  | rel_header_path = cheader_path.relative_to(root_paths[idx]) | 
|  | c_helper.header_path = str(rel_header_path) | 
|  |  | 
|  | # "toplevel.c.tpl" -> "sw/autogen/top_{topname}.c" | 
|  | render_template(TOPGEN_TEMPLATE_PATH / "toplevel.c.tpl", | 
|  | cformat_dir / f"top_{topname}.c", | 
|  | helper=c_helper) | 
|  |  | 
|  | # "toplevel_memory.ld.tpl" -> "sw/autogen/top_{topname}_memory.ld" | 
|  | render_template(TOPGEN_TEMPLATE_PATH / "toplevel_memory.ld.tpl", | 
|  | cformat_dir / f"top_{topname}_memory.ld") | 
|  |  | 
|  | # "toplevel_memory.h.tpl" -> "sw/autogen/top_{topname}_memory.h" | 
|  | memory_cheader_path = cformat_dir / f"top_{topname}_memory.h" | 
|  | render_template(TOPGEN_TEMPLATE_PATH / "toplevel_memory.h.tpl", | 
|  | memory_cheader_path, | 
|  | helper=c_helper) | 
|  |  | 
|  | # Rust File + Clang-format file | 
|  |  | 
|  | # Since SW does not use FuseSoC and instead expects those files always | 
|  | # to be in hw/top_{topname}/sw/autogen, we currently create these files | 
|  | # twice: | 
|  | # - Once under out_path/sw/autogen | 
|  | # - Once under hw/top_{topname}/sw/autogen | 
|  | root_paths = [out_path.resolve(), SRCTREE_TOP] | 
|  | out_paths = [ | 
|  | out_path.resolve(), | 
|  | (SRCTREE_TOP / "hw/top_{}/".format(topname)).resolve() | 
|  | ] | 
|  | for idx, path in enumerate(out_paths): | 
|  | # "clang-format" -> "sw/autogen/.clang-format" | 
|  | cformat_tplpath = TOPGEN_TEMPLATE_PATH / "clang-format" | 
|  | cformat_dir = path / "sw/autogen" | 
|  | cformat_dir.mkdir(parents=True, exist_ok=True) | 
|  | cformat_path = cformat_dir / ".clang-format" | 
|  | cformat_path.write_text(cformat_tplpath.read_text()) | 
|  |  | 
|  | # Save the header macro prefix into `rs_helper` | 
|  | rel_header_dir = cformat_dir.relative_to(root_paths[idx]) | 
|  | rs_helper.header_macro_prefix = ( | 
|  | "OPENTITAN_" + str(rel_header_dir).replace("/", "_").upper()) | 
|  |  | 
|  | # "top_{topname}.rs.tpl" -> "sw/autogen/top_{topname}.rs" | 
|  | cheader_path = cformat_dir / f"top_{topname}.rs" | 
|  | render_template(TOPGEN_TEMPLATE_PATH / "toplevel.rs.tpl", | 
|  | cheader_path, | 
|  | helper=rs_helper) | 
|  |  | 
|  | # TODO Rust don't have header this path | 
|  | # should be used to generate rust modules. | 
|  | # Save the relative header path into `rs_helper` | 
|  | rel_header_path = cheader_path.relative_to(root_paths[idx]) | 
|  | rs_helper.header_path = str(rel_header_path) | 
|  |  | 
|  | # "toplevel_memory.ld.tpl" -> "sw/autogen/top_{topname}_memory.ld" | 
|  | render_template(TOPGEN_TEMPLATE_PATH / "toplevel_memory.ld.tpl", | 
|  | cformat_dir / f"top_{topname}_memory.ld") | 
|  |  | 
|  | # TODO megre to one memmory file | 
|  | # "toplevel_memory.rs.tpl" -> "sw/autogen/top_{topname}_memory.rs.h" | 
|  | memory_cheader_path = cformat_dir / f"top_{topname}_memory.rs" | 
|  | render_template(TOPGEN_TEMPLATE_PATH / "toplevel_memory.rs.tpl", | 
|  | memory_cheader_path, | 
|  | helper=rs_helper) | 
|  |  | 
|  | # generate chip level xbar and alert_handler TB | 
|  | tb_files = [ | 
|  | "xbar_env_pkg__params.sv", "tb__xbar_connect.sv", | 
|  | "tb__alert_handler_connect.sv", "xbar_tgl_excl.cfg", | 
|  | "rstmgr_tgl_excl.cfg" | 
|  | ] | 
|  | for fname in tb_files: | 
|  | tpl_fname = "%s.tpl" % (fname) | 
|  | xbar_chip_data_path = TOPGEN_TEMPLATE_PATH / tpl_fname | 
|  | template_contents = generate_top(completecfg, name_to_block, | 
|  | str(xbar_chip_data_path), gencmd=gencmd) | 
|  |  | 
|  | rendered_dir = out_path / "dv/autogen" | 
|  | rendered_dir.mkdir(parents=True, exist_ok=True) | 
|  | rendered_path = rendered_dir / fname | 
|  |  | 
|  | with rendered_path.open(mode="w", encoding="UTF-8") as fout: | 
|  | fout.write(template_contents) | 
|  |  | 
|  | # generate parameters for chip-level environment package | 
|  | tpl_fname = "chip_env_pkg__params.sv.tpl" | 
|  | alert_handler_chip_data_path = TOPGEN_TEMPLATE_PATH / tpl_fname | 
|  | template_contents = generate_top(completecfg, name_to_block, | 
|  | str(alert_handler_chip_data_path)) | 
|  |  | 
|  | rendered_dir = out_path / "dv/env/autogen" | 
|  | rendered_dir.mkdir(parents=True, exist_ok=True) | 
|  | rendered_path = rendered_dir / "chip_env_pkg__params.sv" | 
|  |  | 
|  | with rendered_path.open(mode="w", encoding="UTF-8") as fout: | 
|  | fout.write(template_contents) | 
|  |  | 
|  | # generate documentation for toplevel | 
|  | gen_top_docs(completecfg, c_helper, out_path) | 
|  |  | 
|  | # Auto-generate tests in "sw/device/tests/autogen" area. | 
|  | gencmd = warnhdr + GENCMD.format(topname=topname) | 
|  | for fname in ["plic_all_irqs_test.c", "alert_test.c", "BUILD"]: | 
|  | outfile = SRCTREE_TOP / "sw/device/tests/autogen" / fname | 
|  | render_template(TOPGEN_TEMPLATE_PATH / f"{fname}.tpl", | 
|  | outfile, | 
|  | helper=c_helper, | 
|  | gencmd=gencmd) | 
|  |  | 
|  |  | 
|  | if __name__ == "__main__": | 
|  | main() |