| #!/usr/bin/env python3 |
| # Copyright lowRISC contributors. |
| # Licensed under the Apache License, Version 2.0, see LICENSE for details. |
| # SPDX-License-Identifier: Apache-2.0 |
| r"""Top Module Generator |
| """ |
| import argparse |
| import logging as log |
| import os |
| import random |
| import shutil |
| import sys |
| import tempfile |
| from collections import OrderedDict |
| from copy import deepcopy |
| from io import StringIO |
| from pathlib import Path |
| from typing import Dict, List, Optional, Tuple |
| from itertools import chain |
| |
| import hjson |
| import tlgen |
| from ipgen import (IpBlockRenderer, IpConfig, IpDescriptionOnlyRenderer, |
| IpTemplate, TemplateRenderError) |
| from mako import exceptions |
| from mako.template import Template |
| from reggen import access, gen_rtl, gen_sec_cm_testplan, window |
| from reggen.inter_signal import InterSignal |
| from reggen.ip_block import IpBlock |
| from reggen.countermeasure import CounterMeasure |
| from reggen.lib import check_list |
| from topgen import entropy_buffer_generator as ebg |
| from topgen import get_hjsonobj_xbars |
| from topgen import intermodule as im |
| from topgen import lib as lib |
| from topgen import merge_top, search_ips, strong_random, validate_top |
| from topgen.c_test import TopGenCTest |
| from topgen.rust import TopGenRust |
| from topgen.clocks import Clocks |
| from topgen.gen_dv import gen_dv |
| from topgen.gen_top_docs import gen_top_docs |
| from topgen.merge import connect_clocks, create_alert_lpgs, extract_clocks |
| from topgen.resets import Resets |
| from topgen.top import Top |
| |
| # Common header for generated files |
| warnhdr = """// |
| // ------------------- W A R N I N G: A U T O - G E N E R A T E D C O D E !! -------------------// |
| // PLEASE DO NOT HAND-EDIT THIS FILE. IT HAS BEEN AUTO-GENERATED WITH THE FOLLOWING COMMAND: |
| """ |
| genhdr = """// Copyright lowRISC contributors. |
| // Licensed under the Apache License, Version 2.0, see LICENSE for details. |
| // SPDX-License-Identifier: Apache-2.0 |
| """ + warnhdr |
| |
| GENCMD = ("// util/topgen.py -t hw/top_{topname}/data/top_{topname}.hjson\n" |
| "// -o hw/top_{topname}") |
| |
| SRCTREE_TOP = Path(__file__).parent.parent.resolve() |
| |
| TOPGEN_TEMPLATE_PATH = Path(__file__).parent / "topgen/templates" |
| |
| # Size and path to the entropy buffer. |
| # This buffer is generated using Mersenne Twister PRNG seeded with rnd_cnst_seed |
| # and deleted in the end." |
| # This buffer will not be created If a different one is provided by args.entropy_buffer. |
| # Module strong_random fetches entropy from the buffer to generate random bit-vectors |
| # and permutations. |
| BUFFER_SIZE = 20000 |
| |
| PATH_TO_BUFFER = "util/topgen/entropy_buffer.txt" |
| |
| |
| def ipgen_render(template_name: str, topname: str, params: Dict, |
| out_path: Path): |
| """ Render an IP template for a specific toplevel using ipgen. |
| |
| The generated IP block is placed in the "ip_autogen" directory of the |
| toplevel. |
| |
| Aborts the program execution in case of an error. |
| """ |
| module_name = params.get("module_instance_name", template_name) |
| instance_name = f"top_{topname}_{module_name}" |
| ip_template = IpTemplate.from_template_path( |
| SRCTREE_TOP / "hw/ip_templates" / template_name) |
| |
| try: |
| ip_config = IpConfig(ip_template.params, instance_name, params) |
| except ValueError as e: |
| log.error(f"Unable to render IP template {template_name!r}: {str(e)}") |
| sys.exit(1) |
| |
| try: |
| renderer = IpBlockRenderer(ip_template, ip_config) |
| renderer.render(out_path / "ip_autogen" / module_name, |
| overwrite_output_dir=True) |
| except TemplateRenderError as e: |
| log.error(e.verbose_str()) |
| sys.exit(1) |
| |
| |
| def generate_top(top, name_to_block, tpl_filename, **kwargs): |
| top_tpl = Template(filename=tpl_filename) |
| |
| try: |
| return top_tpl.render(top=top, name_to_block=name_to_block, **kwargs) |
| except: # noqa: E722 |
| log.error(exceptions.text_error_template().render()) |
| return "" |
| |
| |
| def generate_xbars(top, out_path): |
| topname = top["name"] |
| gencmd = ("// util/topgen.py -t hw/top_{topname}/data/top_{topname}.hjson " |
| "-o hw/top_{topname}/\n\n".format(topname=topname)) |
| |
| for obj in top["xbar"]: |
| xbar_path = out_path / "ip/xbar_{}/data/autogen".format(obj["name"]) |
| xbar_path.mkdir(parents=True, exist_ok=True) |
| xbar = tlgen.validate(obj) |
| xbar.ip_path = "hw/top_" + top["name"] + "/ip/{dut}" |
| |
| # Generate output of crossbar with complete fields |
| xbar_hjson_path = xbar_path / "xbar_{}.gen.hjson".format(xbar.name) |
| xbar_hjson_path.write_text(genhdr + gencmd + |
| hjson.dumps(obj, for_json=True) + '\n') |
| |
| if not tlgen.elaborate(xbar): |
| log.error("Elaboration failed." + repr(xbar)) |
| |
| try: |
| results = tlgen.generate(xbar, "top_" + top["name"]) |
| except: # noqa: E722 |
| log.error(exceptions.text_error_template().render()) |
| |
| ip_path = out_path / "ip/xbar_{}".format(obj["name"]) |
| |
| for filename, filecontent in results: |
| filepath = ip_path / filename |
| filepath.parent.mkdir(parents=True, exist_ok=True) |
| with filepath.open(mode="w", encoding="UTF-8") as fout: |
| fout.write(filecontent) |
| |
| dv_path = out_path / "ip/xbar_{}/dv/autogen".format(obj["name"]) |
| dv_path.mkdir(parents=True, exist_ok=True) |
| |
| # generate testbench for xbar |
| tlgen.generate_tb(xbar, dv_path, "top_" + top["name"]) |
| |
| # Read back the comportable IP and amend to Xbar |
| xbar_ipfile = ip_path / ("data/autogen/xbar_%s.hjson" % obj["name"]) |
| with xbar_ipfile.open() as fxbar: |
| xbar_ipobj = hjson.load(fxbar, |
| use_decimal=True, |
| object_pairs_hook=OrderedDict) |
| |
| r_inter_signal_list = check_list( |
| xbar_ipobj.get("inter_signal_list", []), |
| "inter_signal_list field") |
| obj["inter_signal_list"] = [ |
| InterSignal.from_raw( |
| "entry {} of the inter_signal_list field".format(idx + 1), |
| entry) for idx, entry in enumerate(r_inter_signal_list) |
| ] |
| |
| |
| def generate_alert_handler(top, out_path): |
| topname = top["name"] |
| |
| # default values |
| esc_cnt_dw = 32 |
| accu_cnt_dw = 16 |
| async_on = [] |
| # leave this constant |
| n_classes = 4 |
| # low power groups |
| n_lpg = 1 |
| lpg_map = [] |
| |
| # Count number of alerts and LPGs |
| n_alerts = sum([x["width"] if "width" in x else 1 for x in top["alert"]]) |
| n_lpg = len(top["alert_lpgs"]) |
| n_lpg_width = n_lpg.bit_length() |
| # format used to print out indices in binary format |
| async_on_format = "1'b{:01b}" |
| lpg_idx_format = str(n_lpg_width) + "'d{:d}" |
| |
| # Double check that all these values are greated than 0 |
| if esc_cnt_dw < 1: |
| raise ValueError("esc_cnt_dw must be larger than 0") |
| if accu_cnt_dw < 1: |
| raise ValueError("accu_cnt_dw must be larger than 0") |
| if n_lpg < 1: |
| raise ValueError("n_lpg must be larger than 0") |
| |
| if n_alerts < 1: |
| # set number of alerts to 1 such that the config is still valid |
| # that input will be tied off |
| n_alerts = 1 |
| log.warning("no alerts are defined in the system") |
| else: |
| async_on = [] |
| lpg_map = [] |
| for alert in top["alert"]: |
| for k in range(alert["width"]): |
| async_on.append(async_on_format.format(int(alert["async"]))) |
| lpg_map.append(lpg_idx_format.format(int(alert["lpg_idx"]))) |
| |
| params = { |
| "n_alerts": n_alerts, |
| "esc_cnt_dw": esc_cnt_dw, |
| "accu_cnt_dw": accu_cnt_dw, |
| "async_on": async_on, |
| "n_classes": n_classes, |
| "n_lpg": n_lpg, |
| "lpg_map": lpg_map, |
| } |
| |
| ipgen_render("alert_handler", topname, params, out_path) |
| |
| |
| def generate_plic(top, out_path): |
| topname = top["name"] |
| params = {} |
| |
| # Count number of interrupts |
| # Interrupt source 0 is tied to 0 to conform RISC-V PLIC spec. |
| # So, total number of interrupts are the number of entries in the list + 1 |
| params["src"] = sum( |
| [x["width"] if "width" in x else 1 for x in top["interrupt"]]) + 1 |
| |
| # Target and priority: Currently fixed |
| params["target"] = int(top["num_cores"], 0) if "num_cores" in top else 1 |
| params["prio"] = 3 |
| |
| ipgen_render("rv_plic", topname, params, out_path) |
| |
| |
| # TODO: For generated IPs that are generated legacy style (i.e., without IPgen) |
| # we have to search both the source and destination RTL directories, since not |
| # all files are copied over. This is a workaround which can be removed once |
| # all generated IPs have transitioned to IPgen. |
| def generate_regfile_from_path(hjson_path: Path, |
| generated_rtl_path: Path, |
| original_rtl_path: Path = None): |
| """Generate RTL register file from path and check countermeasure labels""" |
| obj = IpBlock.from_path(str(hjson_path), []) |
| |
| # If this block has countermeasures, we grep for RTL annotations in |
| # all .sv implementation files and check whether they match up |
| # with what is defined inside the Hjson. |
| sv_files = generated_rtl_path.glob("*.sv") |
| if original_rtl_path is not None: |
| sv_files = chain(sv_files, original_rtl_path.glob("*.sv")) |
| rtl_names = CounterMeasure.search_rtl_files(sv_files) |
| obj.check_cm_annotations(rtl_names, str(hjson_path)) |
| gen_rtl.gen_rtl(obj, str(generated_rtl_path)) |
| if gen_sec_cm_testplan.gen_sec_cm_testplan(obj, hjson_path.parent): |
| sys.exit(1) |
| |
| |
| def generate_pinmux(top, out_path): |
| |
| topname = top["name"] |
| pinmux = top["pinmux"] |
| |
| # Generation without pinmux and pinout configuration is not supported. |
| assert "pinmux" in top |
| assert "pinout" in top |
| |
| # Get number of wakeup detectors |
| if "num_wkup_detect" in pinmux: |
| num_wkup_detect = pinmux["num_wkup_detect"] |
| else: |
| num_wkup_detect = 1 |
| |
| if num_wkup_detect <= 0: |
| # TODO: add support for no wakeup counter case |
| log.error("Topgen does currently not support generation of a top " + |
| "without DIOs.") |
| return |
| |
| if "wkup_cnt_width" in pinmux: |
| wkup_cnt_width = pinmux["wkup_cnt_width"] |
| else: |
| wkup_cnt_width = 8 |
| |
| if wkup_cnt_width <= 1: |
| log.error("Wakeup counter width must be greater equal 2.") |
| return |
| |
| # MIO Pads |
| n_mio_pads = pinmux["io_counts"]["muxed"]["pads"] |
| |
| # Total inputs/outputs |
| # Reuse the counts from the merge phase |
| n_mio_periph_in = (pinmux["io_counts"]["muxed"]["inouts"] + |
| pinmux["io_counts"]["muxed"]["inputs"]) |
| n_mio_periph_out = (pinmux["io_counts"]["muxed"]["inouts"] + |
| pinmux["io_counts"]["muxed"]["outputs"]) |
| n_dio_periph_in = (pinmux["io_counts"]["dedicated"]["inouts"] + |
| pinmux["io_counts"]["dedicated"]["inputs"]) |
| n_dio_periph_out = (pinmux["io_counts"]["dedicated"]["inouts"] + |
| pinmux["io_counts"]["dedicated"]["outputs"]) |
| n_dio_pads = (pinmux["io_counts"]["dedicated"]["inouts"] + |
| pinmux["io_counts"]["dedicated"]["inputs"] + |
| pinmux["io_counts"]["dedicated"]["outputs"]) |
| |
| # TODO: derive this value |
| attr_dw = 13 |
| |
| # Generation with zero MIO/DIO pads is currently not supported. |
| assert (n_mio_pads > 0) |
| assert (n_dio_pads > 0) |
| |
| log.info("Generating pinmux with following info from hjson:") |
| log.info("attr_dw: %d" % attr_dw) |
| log.info("num_wkup_detect: %d" % num_wkup_detect) |
| log.info("wkup_cnt_width: %d" % wkup_cnt_width) |
| log.info("n_mio_periph_in: %d" % n_mio_periph_in) |
| log.info("n_mio_periph_out: %d" % n_mio_periph_out) |
| log.info("n_dio_periph_in: %d" % n_dio_periph_in) |
| log.info("n_dio_periph_out: %d" % n_dio_periph_out) |
| log.info("n_dio_pads: %d" % n_dio_pads) |
| |
| # Target path |
| # rtl: pinmux_reg_pkg.sv & pinmux_reg_top.sv |
| # data: pinmux.hjson |
| rtl_path = out_path / "ip/pinmux/rtl/autogen" |
| rtl_path.mkdir(parents=True, exist_ok=True) |
| data_path = out_path / "ip/pinmux/data/autogen" |
| data_path.mkdir(parents=True, exist_ok=True) |
| |
| # Template path |
| tpl_path = Path( |
| __file__).resolve().parent / "../hw/ip/pinmux/data/pinmux.hjson.tpl" |
| original_rtl_path = Path( |
| __file__).resolve().parent / "../hw/ip/pinmux/rtl" |
| |
| # Generate register package and RTLs |
| gencmd = ("// util/topgen.py -t hw/top_{topname}/data/top_{topname}.hjson " |
| "-o hw/top_{topname}/\n\n".format(topname=topname)) |
| |
| hjson_gen_path = data_path / "pinmux.hjson" |
| |
| out = StringIO() |
| with tpl_path.open(mode="r", encoding="UTF-8") as fin: |
| hjson_tpl = Template(fin.read()) |
| try: |
| out = hjson_tpl.render( |
| n_mio_periph_in=n_mio_periph_in, |
| n_mio_periph_out=n_mio_periph_out, |
| n_mio_pads=n_mio_pads, |
| # each DIO has in, out and oe wires |
| # some of these have to be tied off in the |
| # top, depending on the type. |
| n_dio_periph_in=n_dio_pads, |
| n_dio_periph_out=n_dio_pads, |
| n_dio_pads=n_dio_pads, |
| attr_dw=attr_dw, |
| n_wkup_detect=num_wkup_detect, |
| wkup_cnt_width=wkup_cnt_width) |
| except: # noqa: E722 |
| log.error(exceptions.text_error_template().render()) |
| log.info("PINMUX HJSON: %s" % out) |
| |
| if out == "": |
| log.error("Cannot generate pinmux HJSON") |
| return |
| |
| with hjson_gen_path.open(mode="w", encoding="UTF-8") as fout: |
| fout.write(genhdr + gencmd + out) |
| |
| # Generate reg file |
| generate_regfile_from_path(hjson_gen_path, rtl_path, original_rtl_path) |
| |
| |
| def generate_clkmgr(top, cfg_path, out_path): |
| |
| # Target paths |
| rtl_path = out_path / "ip/clkmgr/rtl/autogen" |
| rtl_path.mkdir(parents=True, exist_ok=True) |
| data_path = out_path / "ip/clkmgr/data/autogen" |
| data_path.mkdir(parents=True, exist_ok=True) |
| |
| # Template paths |
| hjson_tpl = cfg_path / "../ip/clkmgr/data/clkmgr.hjson.tpl" |
| rtl_tpl = cfg_path / "../ip/clkmgr/data/clkmgr.sv.tpl" |
| pkg_tpl = cfg_path / "../ip/clkmgr/data/clkmgr_pkg.sv.tpl" |
| original_rtl_path = cfg_path / "../ip/clkmgr/rtl" |
| |
| hjson_out = data_path / "clkmgr.hjson" |
| rtl_out = rtl_path / "clkmgr.sv" |
| pkg_out = rtl_path / "clkmgr_pkg.sv" |
| |
| tpls = [hjson_tpl, rtl_tpl, pkg_tpl] |
| outputs = [hjson_out, rtl_out, pkg_out] |
| names = ["clkmgr.hjson", "clkmgr.sv", "clkmgr_pkg.sv"] |
| |
| clocks = top["clocks"] |
| assert isinstance(clocks, Clocks) |
| |
| typed_clocks = clocks.typed_clocks() |
| hint_names = typed_clocks.hint_names() |
| |
| for idx, tpl in enumerate(tpls): |
| out = "" |
| with tpl.open(mode="r", encoding="UTF-8") as fin: |
| tpl = Template(fin.read()) |
| try: |
| out = tpl.render(cfg=top, |
| clocks=clocks, |
| typed_clocks=typed_clocks, |
| hint_names=hint_names) |
| except: # noqa: E722 |
| log.error(exceptions.text_error_template().render()) |
| |
| if out == "": |
| log.error("Cannot generate {}".format(names[idx])) |
| return |
| |
| with outputs[idx].open(mode="w", encoding="UTF-8") as fout: |
| fout.write(genhdr + out) |
| |
| # Generate reg files |
| generate_regfile_from_path(hjson_out, rtl_path, original_rtl_path) |
| |
| |
| # generate pwrmgr |
| def generate_pwrmgr(top, out_path): |
| log.info("Generating pwrmgr") |
| |
| # Count number of wakeups |
| n_wkups = len(top["wakeups"]) |
| log.info("Found {} wakeup signals".format(n_wkups)) |
| |
| # Count number of reset requests |
| n_rstreqs = len(top["reset_requests"]["peripheral"]) |
| log.info("Found {} reset request signals".format(n_rstreqs)) |
| |
| if n_wkups < 1: |
| n_wkups = 1 |
| log.warning( |
| "The design has no wakeup sources. Low power not supported.") |
| |
| if n_rstreqs < 1: |
| n_rstreqs = 1 |
| log.warning("The design has no reset request sources. " |
| "Reset requests are not supported.") |
| |
| # Define target path |
| rtl_path = out_path / "ip/pwrmgr/rtl/autogen" |
| rtl_path.mkdir(parents=True, exist_ok=True) |
| doc_path = out_path / "ip/pwrmgr/data/autogen" |
| doc_path.mkdir(parents=True, exist_ok=True) |
| |
| # So, read template files from ip directory. |
| tpl_path = Path(__file__).resolve().parent / "../hw/ip/pwrmgr/data" |
| hjson_tpl_path = tpl_path / "pwrmgr.hjson.tpl" |
| original_rtl_path = Path(__file__).resolve().parent / "../hw/ip/pwrmgr/rtl" |
| |
| # Render and write out hjson |
| out = StringIO() |
| with hjson_tpl_path.open(mode="r", encoding="UTF-8") as fin: |
| hjson_tpl = Template(fin.read()) |
| try: |
| out = hjson_tpl.render(NumWkups=n_wkups, |
| Wkups=top["wakeups"], |
| rst_reqs=top["reset_requests"], |
| NumRstReqs=n_rstreqs) |
| |
| except: # noqa: E722 |
| log.error(exceptions.text_error_template().render()) |
| log.info("pwrmgr hjson: %s" % out) |
| |
| if out == "": |
| log.error("Cannot generate pwrmgr config file") |
| return |
| |
| hjson_path = doc_path / "pwrmgr.hjson" |
| with hjson_path.open(mode="w", encoding="UTF-8") as fout: |
| fout.write(genhdr + out) |
| |
| # Generate reg files |
| generate_regfile_from_path(hjson_path, rtl_path, original_rtl_path) |
| |
| |
| def get_rst_ni(top): |
| rstmgrs = [m for m in top['module'] if m['type'] == 'rstmgr'] |
| return rstmgrs[0]["reset_connections"] |
| |
| |
| # generate rstmgr |
| def generate_rstmgr(topcfg, out_path): |
| log.info("Generating rstmgr") |
| |
| # Define target path |
| rtl_path = out_path / "ip/rstmgr/rtl/autogen" |
| rtl_path.mkdir(parents=True, exist_ok=True) |
| doc_path = out_path / "ip/rstmgr/data/autogen" |
| doc_path.mkdir(parents=True, exist_ok=True) |
| tpl_path = Path(__file__).resolve().parent / "../hw/ip/rstmgr/data" |
| original_rtl_path = Path(__file__).resolve().parent / "../hw/ip/rstmgr/rtl" |
| |
| # Read template files from ip directory. |
| tpls = [] |
| outputs = [] |
| names = ["rstmgr.hjson", "rstmgr.sv", "rstmgr_pkg.sv"] |
| |
| for x in names: |
| tpls.append(tpl_path / Path(x + ".tpl")) |
| if "hjson" in x: |
| outputs.append(doc_path / Path(x)) |
| else: |
| outputs.append(rtl_path / Path(x)) |
| |
| # Parameters needed for generation |
| reset_obj = topcfg["resets"] |
| |
| # The original resets dict is transformed to the reset class |
| assert isinstance(reset_obj, Resets) |
| |
| # unique clocks |
| clks = reset_obj.get_clocks() |
| |
| # resets sent to reset struct |
| output_rsts = reset_obj.get_top_resets() |
| |
| # sw controlled resets |
| sw_rsts = reset_obj.get_sw_resets() |
| |
| # rst_ni |
| rst_ni = get_rst_ni(topcfg) |
| |
| # leaf resets |
| leaf_rsts = reset_obj.get_generated_resets() |
| |
| # Number of reset requests |
| n_rstreqs = len(topcfg["reset_requests"]["peripheral"]) |
| |
| # Generate templated files |
| for idx, t in enumerate(tpls): |
| out = StringIO() |
| with t.open(mode="r", encoding="UTF-8") as fin: |
| tpl = Template(fin.read()) |
| try: |
| out = tpl.render(clks=clks, |
| reqs=topcfg["reset_requests"], |
| power_domains=topcfg["power"]["domains"], |
| num_rstreqs=n_rstreqs, |
| sw_rsts=sw_rsts, |
| output_rsts=output_rsts, |
| leaf_rsts=leaf_rsts, |
| rst_ni = rst_ni['rst_ni']['name'], |
| export_rsts=topcfg["exported_rsts"], |
| reset_obj=topcfg["resets"]) |
| |
| except: # noqa: E722 |
| log.error(exceptions.text_error_template().render()) |
| |
| if out == "": |
| log.error("Cannot generate {}".format(names[idx])) |
| return |
| |
| with outputs[idx].open(mode="w", encoding="UTF-8") as fout: |
| fout.write(genhdr + out) |
| |
| # Generate reg files |
| hjson_path = outputs[0] |
| generate_regfile_from_path(hjson_path, rtl_path, original_rtl_path) |
| |
| |
| # generate flash |
| def generate_flash(topcfg, out_path): |
| log.info("Generating flash") |
| |
| # Define target path |
| rtl_path = out_path / "ip/flash_ctrl/rtl/autogen" |
| rtl_path.mkdir(parents=True, exist_ok=True) |
| doc_path = out_path / "ip/flash_ctrl/data/autogen" |
| doc_path.mkdir(parents=True, exist_ok=True) |
| tpl_path = Path(__file__).resolve().parent / "../hw/ip/flash_ctrl/data" |
| original_rtl_path = Path( |
| __file__).resolve().parent / "../hw/ip/flash_ctrl/rtl" |
| |
| # Read template files from ip directory. |
| tpls = [] |
| outputs = [] |
| names = [ |
| "flash_ctrl.hjson", "flash_ctrl.sv", "flash_ctrl_pkg.sv", |
| "flash_ctrl_region_cfg.sv" |
| ] |
| |
| for x in names: |
| tpls.append(tpl_path / Path(x + ".tpl")) |
| if "hjson" in x: |
| outputs.append(doc_path / Path(x)) |
| else: |
| outputs.append(rtl_path / Path(x)) |
| |
| # Parameters needed for generation |
| flash_mems = [ |
| module for module in topcfg["module"] if module["type"] == "flash_ctrl" |
| ] |
| if len(flash_mems) > 1: |
| log.error("This design does not currently support multiple flashes") |
| return |
| |
| cfg = flash_mems[0]["memory"]["mem"]["config"] |
| |
| # Generate templated files |
| for idx, t in enumerate(tpls): |
| out = StringIO() |
| with t.open(mode="r", encoding="UTF-8") as fin: |
| tpl = Template(fin.read()) |
| try: |
| out = tpl.render(cfg=cfg) |
| |
| except: # noqa: E722 |
| log.error(exceptions.text_error_template().render()) |
| |
| if out == "": |
| log.error("Cannot generate {}".format(names[idx])) |
| return |
| |
| with outputs[idx].open(mode="w", encoding="UTF-8") as fout: |
| fout.write(genhdr + out) |
| |
| # Generate reg files |
| hjson_path = outputs[0] |
| generate_regfile_from_path(hjson_path, rtl_path, original_rtl_path) |
| |
| |
| def generate_top_only(top_only_dict, out_path, topname, alt_hjson_path): |
| log.info("Generating top only modules") |
| |
| for ip, reggen_only in top_only_dict.items(): |
| |
| if reggen_only and alt_hjson_path is not None: |
| hjson_dir = Path(alt_hjson_path) |
| else: |
| hjson_dir = Path(__file__).resolve( |
| ).parent / f"../hw/top_{topname}/ip/{ip}/data/" |
| |
| hjson_path = hjson_dir / f"{ip}.hjson" |
| |
| genrtl_dir = out_path / "ip/{}/rtl".format(ip) |
| genrtl_dir.mkdir(parents=True, exist_ok=True) |
| log.info("Generating top modules {}, hjson: {}, output: {}".format( |
| ip, hjson_path, genrtl_dir)) |
| |
| # Generate reg files |
| generate_regfile_from_path(hjson_path, genrtl_dir) |
| |
| |
| def generate_top_ral(top: Dict[str, object], name_to_block: Dict[str, IpBlock], |
| dv_base_names: List[str], out_path: str): |
| # construct top ral block |
| regwidth = int(top["datawidth"]) |
| assert regwidth % 8 == 0 |
| addrsep = regwidth // 8 |
| |
| # Generate a map from instance name to the block that it instantiates, |
| # together with a map of interface addresses. |
| inst_to_block = {} # type: Dict[str, str] |
| if_addrs = {} # type: Dict[Tuple[str, Optional[str]], int] |
| attrs = {} # type: Dict[str, str] |
| |
| for module in top["module"]: |
| inst_name = module["name"] |
| block_name = module["type"] |
| block = name_to_block[block_name] |
| if "attr" in module: |
| if module["attr"] not in ["templated", "ipgen", "reggen_top", |
| "reggen_only"]: |
| raise ValueError("Unsupported value for attr field of {}: {!r}" |
| .format(inst_name, module["attr"])) |
| attrs[inst_name] = module["attr"] |
| |
| inst_to_block[inst_name] = block_name |
| for if_name in block.reg_blocks.keys(): |
| if_addr = int(module["base_addrs"][if_name], 0) |
| if_addrs[(inst_name, if_name)] = if_addr |
| |
| # Collect up the memories to add |
| mems = [] |
| for item in list(top.get("memory", [])): |
| mems.append(create_mem(item, addrsep, regwidth)) |
| |
| # Top-level may override the mem setting. Store the new type to name_to_block |
| # If no other instance uses the orignal type, delete it |
| original_types = set() |
| for module in top["module"]: |
| if "memory" in module.keys() and len(module["memory"]) > 0: |
| newtype = "{}_{}".format(module["type"], module["name"]) |
| assert newtype not in name_to_block |
| |
| block = deepcopy(name_to_block[module["type"]]) |
| name_to_block[newtype] = block |
| inst_to_block[module["name"]] = newtype |
| |
| original_types.add(module["type"]) |
| |
| for mem_name, item in module["memory"].items(): |
| assert block.reg_blocks[mem_name] |
| assert len(block.reg_blocks[mem_name].windows) <= 1 |
| item["name"] = mem_name |
| |
| win = create_mem(item, addrsep, regwidth) |
| if len(block.reg_blocks[mem_name].windows) > 0: |
| blk_win = block.reg_blocks[mem_name].windows[0] |
| |
| # Top can only add new info for mem, shouldn't overwrite |
| # existing configuration |
| assert win.items == blk_win.items |
| assert win.byte_write == blk_win.byte_write |
| assert win.data_intg_passthru == blk_win.data_intg_passthru |
| |
| block.reg_blocks[mem_name].windows[0] = win |
| else: |
| block.reg_blocks[mem_name].windows.append(win) |
| |
| for t in original_types: |
| if t not in inst_to_block.values(): |
| del name_to_block[t] |
| |
| chip = Top(regwidth, name_to_block, inst_to_block, if_addrs, mems, attrs) |
| |
| # generate the top ral model with template |
| return gen_dv(chip, dv_base_names, str(out_path)) |
| |
| |
| def create_mem(item, addrsep, regwidth): |
| byte_write = ("byte_write" in item and |
| item["byte_write"].lower() == "true") |
| data_intg_passthru = ("data_intg_passthru" in item and |
| item["data_intg_passthru"].lower() == "true") |
| size_in_bytes = int(item["size"], 0) |
| num_regs = size_in_bytes // addrsep |
| swaccess = access.SWAccess("top-level memory", item.get("swaccess", "rw")) |
| |
| return window.Window(name=item["name"], |
| desc="(generated from top-level)", |
| unusual=False, |
| byte_write=byte_write, |
| data_intg_passthru=data_intg_passthru, |
| validbits=regwidth, |
| items=num_regs, |
| size_in_bytes=size_in_bytes, |
| offset=int(item.get("base_addr", "0"), 0), |
| swaccess=swaccess) |
| |
| |
| def _process_top(topcfg, args, cfg_path, out_path, pass_idx): |
| # Create generated list |
| # These modules are generated through topgen |
| templated_list = lib.get_templated_modules(topcfg) |
| log.info("Templated list is {}".format(templated_list)) |
| |
| ipgen_list = lib.get_ipgen_modules(topcfg) |
| log.info("Ip gen list is {}".format(ipgen_list)) |
| |
| generated_list = templated_list + ipgen_list |
| |
| # These modules are NOT generated but belong to a specific top |
| # and therefore not part of "hw/ip" |
| top_only_dict = { |
| module["type"]: lib.is_reggen_only(module) |
| for module in topcfg["module"] if lib.is_top_reggen(module) |
| } |
| log.info("Filtered dict is {}".format(top_only_dict)) |
| |
| topname = topcfg["name"] |
| |
| # Sweep the IP directory and gather the config files |
| ip_dir = Path(__file__).parents[1] / "hw/ip" |
| ips = search_ips(ip_dir) |
| |
| # exclude filtered IPs (to use top_${topname} one) and |
| exclude_list = generated_list + list(top_only_dict.keys()) |
| ips = [x for x in ips if not x.parents[1].name in exclude_list] |
| |
| # Hack alert |
| # Generate clkmgr.hjson here so that it can be included below |
| # Unlike other generated hjsons, clkmgr thankfully does not require |
| # ip.hjson information. All the information is embedded within |
| # the top hjson file |
| topcfg["clocks"] = Clocks(topcfg["clocks"]) |
| extract_clocks(topcfg) |
| generate_clkmgr(topcfg, cfg_path, out_path) |
| |
| # It may require two passes to check if the module is needed. |
| # TODO: first run of topgen will fail due to the absent of rv_plic. |
| # It needs to run up to amend_interrupt in merge_top function |
| # then creates rv_plic.hjson then run xbar generation. |
| hjson_dir = Path(args.topcfg).parent |
| |
| for ip in generated_list: |
| # For modules that are generated prior to gathering, we need to take it from |
| # the output path. For modules not generated before, it may exist in a |
| # pre-defined area already. |
| log.info("Appending {}".format(ip)) |
| if ip in ipgen_list: |
| ip_relpath = "ip_autogen" |
| desc_file_relpath = "data" |
| else: |
| ip_relpath = "ip" |
| desc_file_relpath = "data/autogen" |
| |
| if ip == "clkmgr" or (pass_idx > 0): |
| ip_hjson = (Path(out_path) / ip_relpath / ip / desc_file_relpath / |
| f"{ip}.hjson") |
| else: |
| ip_hjson = (hjson_dir.parent / ip_relpath / ip / |
| desc_file_relpath / f"{ip}.hjson") |
| ips.append(ip_hjson) |
| |
| for ip, reggen_only in top_only_dict.items(): |
| log.info("Appending {}".format(ip)) |
| |
| if reggen_only and args.hjson_path: |
| ip_hjson = Path(args.hjson_path) / f"{ip}.hjson" |
| else: |
| ip_hjson = hjson_dir.parent / f"ip/{ip}/data/{ip}.hjson" |
| |
| ips.append(ip_hjson) |
| |
| # load Hjson and pass validate from reggen |
| try: |
| ip_objs = [] |
| for ip_desc_file in ips: |
| ip_name = ip_desc_file.stem |
| # Skip if it is not in the module list |
| if ip_name not in [ip["type"] for ip in topcfg["module"]]: |
| log.info("Skip module %s as it isn't in the top module list" % |
| ip_name) |
| continue |
| |
| # The auto-generated hjson might not yet exist. It will be created |
| # later, see generate_{ip_name}() calls below. For the initial |
| # validation, use the Hjson file with default values. |
| # TODO: All of this is a rather ugly hack that we need to get rid |
| # of as soon as we don't arbitrarily template IP description Hjson |
| # files any more. |
| if ip_name in generated_list and not ip_desc_file.is_file(): |
| if ip_name in ipgen_list: |
| log.info( |
| "To-be-auto-generated Hjson %s does not yet exist. " |
| "Falling back to the default configuration of template " |
| "%s for initial validation." % (ip_desc_file, ip_name)) |
| |
| tpl_path = SRCTREE_TOP / "hw/ip_templates" / ip_name |
| ip_template = IpTemplate.from_template_path(tpl_path) |
| ip_config = IpConfig(ip_template.params, |
| f"top_{topname}_{ip_name}") |
| |
| try: |
| ip_desc = IpDescriptionOnlyRenderer( |
| ip_template, ip_config).render() |
| except TemplateRenderError as e: |
| log.error(e.verbose_str()) |
| sys.exit(1) |
| s = "default description of IP template {}".format(ip_name) |
| ip_objs.append(IpBlock.from_text(ip_desc, [], s)) |
| else: |
| # TODO: Remove this block as soon as all IP templates use |
| # ipgen. |
| template_hjson_file = ip_dir / "{}/data/{}.hjson".format( |
| ip_name, ip_name) |
| log.info( |
| "To-be-auto-generated Hjson %s does not yet exist. " |
| "Falling back to Hjson description file %s shipped " |
| "with the IP template for initial validation." % |
| (ip_desc_file, template_hjson_file)) |
| |
| ip_objs.append( |
| IpBlock.from_path(str(template_hjson_file), [])) |
| else: |
| ip_objs.append(IpBlock.from_path(str(ip_desc_file), [])) |
| |
| except ValueError: |
| raise SystemExit(sys.exc_info()[1]) |
| |
| name_to_block = {} # type: Dict[str, IpBlock] |
| for block in ip_objs: |
| lblock = block.name.lower() |
| assert lblock not in name_to_block |
| name_to_block[lblock] = block |
| |
| # Read in alias files one-by-one, peek inside to figure out which IP block |
| # they belong to and apply the alias file to that IP block. |
| if args.alias_files: |
| for alias in args.alias_files: |
| with open(alias, 'r', encoding='utf-8') as handle: |
| raw = hjson.loads(handle.read(), use_decimal=True) |
| if 'alias_target' not in raw: |
| raise ValueError('Missing alias_target key ' |
| 'in alias file {}.'.format(alias)) |
| alias_target = raw['alias_target'].lower() |
| if alias_target not in name_to_block: |
| raise ValueError('Alias target {} is not defined.' |
| .format(alias_target)) |
| where = 'alias file at {}'.format(alias) |
| name_to_block[alias_target].alias_from_raw(False, raw, where) |
| |
| connect_clocks(topcfg, name_to_block) |
| |
| # Read the crossbars under the top directory |
| xbar_objs = get_hjsonobj_xbars(hjson_dir) |
| |
| log.info("Detected crossbars: %s" % |
| (", ".join([x["name"] for x in xbar_objs]))) |
| |
| topcfg, error = validate_top(topcfg, ip_objs, xbar_objs) |
| if error != 0: |
| raise SystemExit("Error occured while validating top.hjson") |
| |
| completecfg = merge_top(topcfg, name_to_block, xbar_objs) |
| |
| # Generate flash controller and flash memory |
| generate_flash(topcfg, out_path) |
| |
| # Generate PLIC |
| if not args.no_plic and \ |
| not args.alert_handler_only and \ |
| not args.xbar_only: |
| generate_plic(completecfg, out_path) |
| if args.plic_only: |
| sys.exit() |
| |
| # Create Alert Handler LPGs before |
| # generating the Alert Handler |
| create_alert_lpgs(topcfg, name_to_block) |
| |
| # Generate Alert Handler |
| if not args.xbar_only: |
| generate_alert_handler(completecfg, out_path) |
| if args.alert_handler_only: |
| sys.exit() |
| |
| # Generate Pinmux |
| generate_pinmux(completecfg, out_path) |
| |
| # Generate Pwrmgr |
| generate_pwrmgr(completecfg, out_path) |
| |
| # Generate rstmgr |
| generate_rstmgr(completecfg, out_path) |
| |
| # Generate top only modules |
| # These modules are not templated, but are not in hw/ip |
| generate_top_only(top_only_dict, out_path, topname, args.hjson_path) |
| |
| return completecfg, name_to_block |
| |
| |
| def main(): |
| parser = argparse.ArgumentParser(prog="topgen") |
| parser.add_argument("--topcfg", |
| "-t", |
| required=True, |
| help="`top_{name}.hjson` file.") |
| parser.add_argument( |
| "--outdir", |
| "-o", |
| help="""Target TOP directory. |
| Module is created under rtl/. (default: dir(topcfg)/..) |
| """) # yapf: disable |
| parser.add_argument( |
| "--hjson-path", |
| help=""" |
| If defined, topgen uses supplied path to search for ip hjson. |
| This applies only to ip's with the `reggen_only` attribute. |
| If an hjson is located both in the conventional path and the alternate |
| path, the alternate path has priority. |
| """) |
| parser.add_argument("--verbose", "-v", action="store_true", help="Verbose") |
| |
| # Generator options: 'no' series. cannot combined with 'only' series |
| parser.add_argument( |
| "--no-top", |
| action="store_true", |
| help="If defined, topgen doesn't generate top_{name} RTLs.") |
| parser.add_argument( |
| "--no-xbar", |
| action="store_true", |
| help="If defined, topgen doesn't generate crossbar RTLs.") |
| parser.add_argument( |
| "--no-plic", |
| action="store_true", |
| help="If defined, topgen doesn't generate the interrup controller RTLs." |
| ) |
| |
| # Generator options: 'only' series. cannot combined with 'no' series |
| parser.add_argument( |
| "--top-only", |
| action="store_true", |
| help="If defined, the tool generates top RTL only") # yapf:disable |
| parser.add_argument( |
| "--xbar-only", |
| action="store_true", |
| help="If defined, the tool generates crossbar RTLs only") |
| parser.add_argument( |
| "--plic-only", |
| action="store_true", |
| help="If defined, the tool generates RV_PLIC RTL and Hjson only") |
| parser.add_argument( |
| "--alert-handler-only", |
| action="store_true", |
| help="If defined, the tool generates alert handler hjson only") |
| # Generator options: generate dv ral model |
| parser.add_argument( |
| "--top_ral", |
| "-r", |
| default=False, |
| action="store_true", |
| help="If set, the tool generates top level RAL model for DV") |
| parser.add_argument( |
| "--alias-files", |
| nargs="+", |
| type=Path, |
| default=None, |
| help=""" |
| If defined, topgen uses supplied alias hjson file(s) to override the |
| generic register definitions when building the RAL model. This |
| argument is only relevant in conjunction with the `--top_ral` switch. |
| """) |
| parser.add_argument( |
| "--dv-base-names", |
| nargs="+", |
| help="Names or prefix for the DV register classes from which " |
| "the register models are derived.") |
| # Generator options for compile time random netlist constants |
| parser.add_argument( |
| "--rnd_cnst_seed", |
| type=int, |
| metavar="<seed>", |
| help="Custom seed for RNG to compute netlist constants.") |
| parser.add_argument( |
| "--entropy_buffer", |
| help="A file with entropy.") |
| # Miscellaneous: only return the list of blocks and exit. |
| parser.add_argument("--get_blocks", |
| default=False, |
| action="store_true", |
| help="Only return the list of blocks and exit.") |
| |
| args = parser.parse_args() |
| |
| # check combinations |
| if args.top_ral: |
| args.no_top = True |
| |
| if (args.no_top or args.no_xbar or |
| args.no_plic) and (args.top_only or args.xbar_only or |
| args.plic_only or args.alert_handler_only): |
| log.error( |
| "'no' series options cannot be used with 'only' series options") |
| raise SystemExit(sys.exc_info()[1]) |
| |
| # Don't print warnings when querying the list of blocks. |
| log_level = (log.ERROR |
| if args.get_blocks else log.DEBUG if args.verbose else None) |
| |
| log.basicConfig(format="%(levelname)s: %(message)s", level=log_level) |
| |
| if not args.outdir: |
| outdir = Path(args.topcfg).parent / ".." |
| log.info("TOP directory not given. Use %s", (outdir)) |
| elif not Path(args.outdir).is_dir(): |
| log.error("'--outdir' should point to writable directory") |
| raise SystemExit(sys.exc_info()[1]) |
| else: |
| outdir = Path(args.outdir) |
| |
| if args.hjson_path is not None: |
| log.info(f"Alternate hjson path is {args.hjson_path}") |
| |
| out_path = Path(outdir) |
| cfg_path = Path(args.topcfg).parents[1] |
| |
| try: |
| with open(args.topcfg, "r") as ftop: |
| topcfg = hjson.load(ftop, |
| use_decimal=True, |
| object_pairs_hook=OrderedDict) |
| except ValueError: |
| raise SystemExit(sys.exc_info()[1]) |
| |
| # Initialize RNG for compile-time netlist constants. |
| if args.entropy_buffer: |
| if args.rnd_cnst_seed: |
| log.error("'entropy_buffer' option cannot be used with 'rnd_cnst_seed option'") |
| # error out |
| raise SystemExit(sys.exc_info()[1]) |
| else: |
| # generate entropy from a buffer |
| strong_random.load(SRCTREE_TOP / args.entropy_buffer) |
| else: |
| # If specified, override the seed for random netlist constant computation. |
| if args.rnd_cnst_seed: |
| log.warning("Commandline override of rnd_cnst_seed with {}.".format( |
| args.rnd_cnst_seed)) |
| topcfg["rnd_cnst_seed"] = args.rnd_cnst_seed |
| # Otherwise, we either take it from the top_{topname}.hjson if present, or |
| # randomly generate a new seed if not. |
| else: |
| random.seed() |
| new_seed = random.getrandbits(64) |
| if topcfg.setdefault("rnd_cnst_seed", new_seed) == new_seed: |
| log.warning( |
| "No rnd_cnst_seed specified, setting to {}.".format(new_seed)) |
| ebg.gen_buffer(BUFFER_SIZE, SRCTREE_TOP / PATH_TO_BUFFER, False, topcfg["rnd_cnst_seed"]) |
| strong_random.load(SRCTREE_TOP / PATH_TO_BUFFER) |
| |
| # TODO, long term, the levels of dependency should be automatically determined instead |
| # of hardcoded. The following are a few examples: |
| # Example 1: pinmux depends on amending all modules before calculating the correct number of |
| # pins. |
| # This would be 1 level of dependency and require 2 passes. |
| # Example 2: pinmux depends on amending all modules, and pwrmgr depends on pinmux generation to |
| # know correct number of wakeups. This would be 2 levels of dependency and require 3 |
| # passes. |
| # |
| # How does mulit-pass work? |
| # In example 1, the first pass gathers all modules and merges them. However, the merge process |
| # uses a stale pinmux. The correct pinmux is then generated using the merged configuration. The |
| # second pass now merges all the correct modules (including the generated pinmux) and creates |
| # the final merged config. |
| # |
| # In example 2, the first pass gathers all modules and merges them. However, the merge process |
| # uses a stale pinmux and pwrmgr. The correct pinmux is then generated using the merged |
| # configuration. However, since pwrmgr is dependent on this new pinmux, it is still generated |
| # incorrectly. The second pass merge now has an updated pinmux but stale pwrmgr. The correct |
| # pwrmgr can now be generated. The final pass then merges all the correct modules and creates |
| # the final configuration. |
| # |
| # This fix is related to #2083 |
| process_dependencies = 1 |
| |
| # topgen generates IP blocks and associated Hjson configuration in multiple |
| # steps. After each step, the IP Hjson configuration is read back and then |
| # combined into the toplevel configuration. To generate the chip-level RAL, |
| # we need to run the full generation step, but ultimately only care about |
| # the toplevel configuration (a single Hjson file). Since we don't have a |
| # better way at the moment dump all output into a temporary directory, and |
| # delete it after the fact, retaining only the toplevel configuration. |
| if args.top_ral: |
| out_path_gen = Path(tempfile.mkdtemp()) |
| else: |
| out_path_gen = out_path |
| |
| for pass_idx in range(process_dependencies + 1): |
| log.debug("Generation pass {}".format(pass_idx)) |
| if pass_idx < process_dependencies: |
| cfg_copy = deepcopy(topcfg) |
| _process_top(cfg_copy, args, cfg_path, out_path_gen, pass_idx) |
| else: |
| completecfg, name_to_block = _process_top(topcfg, args, cfg_path, |
| out_path_gen, pass_idx) |
| |
| topname = topcfg["name"] |
| |
| if not args.entropy_buffer: |
| # Delete entropy buffer since it is no longer needed. |
| # This buffer can always be re-generated from the seed using entropy_buffer_generator |
| os.remove(SRCTREE_TOP / PATH_TO_BUFFER) |
| |
| # Create the chip-level RAL only |
| if args.top_ral: |
| # See above: we only need `completeconfig` and `name_to_block`, not all |
| # the other files (e.g. RTL files) generated through topgen. |
| shutil.rmtree(out_path_gen, ignore_errors=True) |
| |
| exit_code = generate_top_ral(completecfg, name_to_block, |
| args.dv_base_names, out_path) |
| sys.exit(exit_code) |
| |
| if args.get_blocks: |
| print("\n".join(name_to_block.keys())) |
| sys.exit(0) |
| |
| # Generate xbars |
| if not args.no_xbar or args.xbar_only: |
| generate_xbars(completecfg, out_path) |
| |
| # All IPs are generated. Connect phase now |
| # Find {memory, module} <-> {xbar} connections first. |
| im.autoconnect(completecfg, name_to_block) |
| |
| # Generic Inter-module connection |
| im.elab_intermodule(completecfg) |
| |
| # Generate top.gen.hjson right before rendering |
| genhjson_dir = out_path / "data/autogen" |
| genhjson_dir.mkdir(parents=True, exist_ok=True) |
| genhjson_path = genhjson_dir / ("top_%s.gen.hjson" % completecfg["name"]) |
| |
| # Header for HJSON |
| if args.entropy_buffer: |
| gencmd = """// |
| // util/topgen.py -t hw/top_{topname}/data/top_{topname}.hjson \\ |
| // -o hw/top_{topname}/ \\ |
| // --hjson-only \\ |
| // --entropy-buffer {path} |
| """.format(topname=topname, path = args.entropy_buffer) |
| else: |
| gencmd = """// |
| // util/topgen.py -t hw/top_{topname}/data/top_{topname}.hjson \\ |
| // -o hw/top_{topname}/ \\ |
| // --hjson-only \\ |
| // --rnd_cnst_seed {seed} |
| """.format(topname=topname, seed=completecfg["rnd_cnst_seed"]) |
| |
| genhjson_path.write_text(genhdr + gencmd + |
| hjson.dumps(completecfg, for_json=True) + '\n') |
| |
| if not args.no_top or args.top_only: |
| |
| def render_template(template_path: str, rendered_path: Path, |
| **other_info): |
| template_contents = generate_top(completecfg, name_to_block, |
| str(template_path), **other_info) |
| |
| rendered_path.parent.mkdir(exist_ok=True, parents=True) |
| with rendered_path.open(mode="w", encoding="UTF-8") as fout: |
| fout.write(template_contents) |
| |
| # Header for SV files |
| if args.entropy_buffer: |
| gencmd = warnhdr + """// |
| // util/topgen.py -t hw/top_{topname}/data/top_{topname}.hjson \\ |
| // -o hw/top_{topname}/ \\ |
| // --entropy-buffer {path} |
| """.format(topname=topname, path = args.entropy_buffer) |
| else: |
| gencmd = warnhdr + """// |
| // util/topgen.py -t hw/top_{topname}/data/top_{topname}.hjson \\ |
| // -o hw/top_{topname}/ \\ |
| // --rnd_cnst_seed {seed} |
| """.format(topname=topname, seed=completecfg["rnd_cnst_seed"]) |
| |
| # SystemVerilog Top: |
| # "toplevel.sv.tpl" -> "rtl/autogen/top_{topname}.sv" |
| render_template(TOPGEN_TEMPLATE_PATH / "toplevel.sv.tpl", |
| out_path / f"rtl/autogen/top_{topname}.sv", |
| gencmd=gencmd) |
| |
| # Multiple chip-levels (ASIC, FPGA, Verilator, etc) |
| for target in topcfg["targets"]: |
| target_name = target["name"] |
| render_template(TOPGEN_TEMPLATE_PATH / "chiplevel.sv.tpl", |
| out_path / |
| f"rtl/autogen/chip_{topname}_{target_name}.sv", |
| gencmd=gencmd, |
| target=target) |
| |
| # The C / SV file needs some complex information, so we initialize this |
| # object to store it. |
| c_helper = TopGenCTest(completecfg, name_to_block) |
| |
| # The Rust file needs some complex information, so we initialize this |
| # object to store it. |
| rs_helper = TopGenRust(completecfg, name_to_block) |
| |
| # "toplevel_pkg.sv.tpl" -> "rtl/autogen/top_{topname}_pkg.sv" |
| render_template(TOPGEN_TEMPLATE_PATH / "toplevel_pkg.sv.tpl", |
| out_path / f"rtl/autogen/top_{topname}_pkg.sv", |
| helper=c_helper, |
| gencmd=gencmd) |
| |
| # compile-time random netlist constants |
| render_template(TOPGEN_TEMPLATE_PATH / "toplevel_rnd_cnst_pkg.sv.tpl", |
| out_path / |
| f"rtl/autogen/top_{topname}_rnd_cnst_pkg.sv", |
| gencmd=gencmd) |
| |
| # C Header + C File + Clang-format file |
| |
| # Since SW does not use FuseSoC and instead expects those files always |
| # to be in hw/top_{topname}/sw/autogen, we currently create these files |
| # twice: |
| # - Once under out_path/sw/autogen |
| # - Once under hw/top_{topname}/sw/autogen |
| root_paths = [out_path.resolve(), SRCTREE_TOP] |
| out_paths = [ |
| out_path.resolve(), |
| (SRCTREE_TOP / "hw/top_{}/".format(topname)).resolve() |
| ] |
| for idx, path in enumerate(out_paths): |
| # "clang-format" -> "sw/autogen/.clang-format" |
| cformat_tplpath = TOPGEN_TEMPLATE_PATH / "clang-format" |
| cformat_dir = path / "sw/autogen" |
| cformat_dir.mkdir(parents=True, exist_ok=True) |
| cformat_path = cformat_dir / ".clang-format" |
| cformat_path.write_text(cformat_tplpath.read_text()) |
| |
| # Save the header macro prefix into `c_helper` |
| rel_header_dir = cformat_dir.relative_to(root_paths[idx]) |
| c_helper.header_macro_prefix = ( |
| "OPENTITAN_" + str(rel_header_dir).replace("/", "_").upper()) |
| |
| # "top_{topname}.h.tpl" -> "sw/autogen/top_{topname}.h" |
| cheader_path = cformat_dir / f"top_{topname}.h" |
| render_template(TOPGEN_TEMPLATE_PATH / "toplevel.h.tpl", |
| cheader_path, |
| helper=c_helper) |
| |
| # Save the relative header path into `c_helper` |
| rel_header_path = cheader_path.relative_to(root_paths[idx]) |
| c_helper.header_path = str(rel_header_path) |
| |
| # "toplevel.c.tpl" -> "sw/autogen/top_{topname}.c" |
| render_template(TOPGEN_TEMPLATE_PATH / "toplevel.c.tpl", |
| cformat_dir / f"top_{topname}.c", |
| helper=c_helper) |
| |
| # "toplevel_memory.ld.tpl" -> "sw/autogen/top_{topname}_memory.ld" |
| render_template(TOPGEN_TEMPLATE_PATH / "toplevel_memory.ld.tpl", |
| cformat_dir / f"top_{topname}_memory.ld") |
| |
| # "toplevel_memory.h.tpl" -> "sw/autogen/top_{topname}_memory.h" |
| memory_cheader_path = cformat_dir / f"top_{topname}_memory.h" |
| render_template(TOPGEN_TEMPLATE_PATH / "toplevel_memory.h.tpl", |
| memory_cheader_path, |
| helper=c_helper) |
| |
| # Rust File + Clang-format file |
| |
| # Since SW does not use FuseSoC and instead expects those files always |
| # to be in hw/top_{topname}/sw/autogen, we currently create these files |
| # twice: |
| # - Once under out_path/sw/autogen |
| # - Once under hw/top_{topname}/sw/autogen |
| root_paths = [out_path.resolve(), SRCTREE_TOP] |
| out_paths = [ |
| out_path.resolve(), |
| (SRCTREE_TOP / "hw/top_{}/".format(topname)).resolve() |
| ] |
| for idx, path in enumerate(out_paths): |
| # "clang-format" -> "sw/autogen/.clang-format" |
| cformat_tplpath = TOPGEN_TEMPLATE_PATH / "clang-format" |
| cformat_dir = path / "sw/autogen" |
| cformat_dir.mkdir(parents=True, exist_ok=True) |
| cformat_path = cformat_dir / ".clang-format" |
| cformat_path.write_text(cformat_tplpath.read_text()) |
| |
| # Save the header macro prefix into `rs_helper` |
| rel_header_dir = cformat_dir.relative_to(root_paths[idx]) |
| rs_helper.header_macro_prefix = ( |
| "OPENTITAN_" + str(rel_header_dir).replace("/", "_").upper()) |
| |
| # "top_{topname}.rs.tpl" -> "sw/autogen/top_{topname}.rs" |
| cheader_path = cformat_dir / f"top_{topname}.rs" |
| render_template(TOPGEN_TEMPLATE_PATH / "toplevel.rs.tpl", |
| cheader_path, |
| helper=rs_helper) |
| |
| # TODO Rust don't have header this path |
| # should be used to generate rust modules. |
| # Save the relative header path into `rs_helper` |
| rel_header_path = cheader_path.relative_to(root_paths[idx]) |
| rs_helper.header_path = str(rel_header_path) |
| |
| # "toplevel_memory.ld.tpl" -> "sw/autogen/top_{topname}_memory.ld" |
| render_template(TOPGEN_TEMPLATE_PATH / "toplevel_memory.ld.tpl", |
| cformat_dir / f"top_{topname}_memory.ld") |
| |
| # TODO megre to one memmory file |
| # "toplevel_memory.rs.tpl" -> "sw/autogen/top_{topname}_memory.rs.h" |
| memory_cheader_path = cformat_dir / f"top_{topname}_memory.rs" |
| render_template(TOPGEN_TEMPLATE_PATH / "toplevel_memory.rs.tpl", |
| memory_cheader_path, |
| helper=rs_helper) |
| |
| # generate chip level xbar and alert_handler TB |
| tb_files = [ |
| "xbar_env_pkg__params.sv", "tb__xbar_connect.sv", |
| "tb__alert_handler_connect.sv", "xbar_tgl_excl.cfg", |
| "rstmgr_tgl_excl.cfg" |
| ] |
| for fname in tb_files: |
| tpl_fname = "%s.tpl" % (fname) |
| xbar_chip_data_path = TOPGEN_TEMPLATE_PATH / tpl_fname |
| template_contents = generate_top(completecfg, name_to_block, |
| str(xbar_chip_data_path), gencmd=gencmd) |
| |
| rendered_dir = out_path / "dv/autogen" |
| rendered_dir.mkdir(parents=True, exist_ok=True) |
| rendered_path = rendered_dir / fname |
| |
| with rendered_path.open(mode="w", encoding="UTF-8") as fout: |
| fout.write(template_contents) |
| |
| # generate parameters for chip-level environment package |
| tpl_fname = "chip_env_pkg__params.sv.tpl" |
| alert_handler_chip_data_path = TOPGEN_TEMPLATE_PATH / tpl_fname |
| template_contents = generate_top(completecfg, name_to_block, |
| str(alert_handler_chip_data_path)) |
| |
| rendered_dir = out_path / "dv/env/autogen" |
| rendered_dir.mkdir(parents=True, exist_ok=True) |
| rendered_path = rendered_dir / "chip_env_pkg__params.sv" |
| |
| with rendered_path.open(mode="w", encoding="UTF-8") as fout: |
| fout.write(template_contents) |
| |
| # generate documentation for toplevel |
| gen_top_docs(completecfg, c_helper, out_path) |
| |
| # Auto-generate tests in "sw/device/tests/autogen" area. |
| gencmd = warnhdr + GENCMD.format(topname=topname) |
| for fname in ["plic_all_irqs_test.c", "alert_test.c", "BUILD"]: |
| outfile = SRCTREE_TOP / "sw/device/tests/autogen" / fname |
| render_template(TOPGEN_TEMPLATE_PATH / f"{fname}.tpl", |
| outfile, |
| helper=c_helper, |
| gencmd=gencmd) |
| |
| |
| if __name__ == "__main__": |
| main() |