|  | # Copyright lowRISC contributors. | 
|  | # Licensed under the Apache License, Version 2.0, see LICENSE for details. | 
|  | # SPDX-License-Identifier: Apache-2.0 | 
|  | import logging as log | 
|  | from enum import Enum | 
|  | from collections import OrderedDict | 
|  |  | 
|  | from reggen.validate import check_keys | 
|  |  | 
|  | # For the reference | 
|  | # val_types = { | 
|  | #     'd': ["int", "integer (binary 0b, octal 0o, decimal, hex 0x)"], | 
|  | #     'x': ["xint", "x for undefined otherwise int"], | 
|  | #     'b': [ | 
|  | #         "bitrange", "bit number as decimal integer, \ | 
|  | #                     or bit-range as decimal integers msb:lsb" | 
|  | #     ], | 
|  | #     'l': ["list", "comma separated list enclosed in `[]`"], | 
|  | #     'ln': ["name list", 'comma separated list enclosed in `[]` of '\ | 
|  | #            'one or more groups that have just name and dscr keys.'\ | 
|  | #            ' e.g. `{ name: "name", desc: "description"}`'], | 
|  | #     'lnw': ["name list+", 'name list that optionally contains a width'], | 
|  | #     'lp': ["parameter list", 'parameter list having default value optionally'], | 
|  | #     'g': ["group", "comma separated group of key:value enclosed in `{}`"], | 
|  | #     's': ["string", "string, typically short"], | 
|  | #     't': ["text", "string, may be multi-line enclosed in `'''` "\ | 
|  | #           "may use `**bold**`, `*italic*` or `!!Reg` markup"], | 
|  | #     'T': ["tuple", "tuple enclosed in ()"], | 
|  | #     'pi': ["python int", "Native Python type int (generated)"], | 
|  | #     'pb': ["python Bool", "Native Python type Bool (generated)"], | 
|  | #     'pl': ["python list", "Native Python type list (generated)"], | 
|  | #     'pe': ["python enum", "Native Python type enum (generated)"] | 
|  | # } | 
|  |  | 
|  | # Required/optional field in top hjson | 
|  | top_required = { | 
|  | 'name': ['s', 'Top name'], | 
|  | 'type': ['s', 'type of hjson. Shall be "top" always'], | 
|  | 'clocks': ['g', 'group of clock properties'], | 
|  | 'resets': ['l', 'list of resets'], | 
|  | 'module': ['l', 'list of modules to instantiate'], | 
|  | 'memory': [ | 
|  | 'l', 'list of memories. At least one memory is needed to run \ | 
|  | the software' | 
|  | ], | 
|  | 'debug_mem_base_addr': | 
|  | ['d', 'Base address of RV_DM. Planned to move to \ | 
|  | module'], | 
|  | 'xbar': ['l', 'List of the xbar used in the top'], | 
|  | } | 
|  |  | 
|  | top_optional = { | 
|  | 'interrupt_module': ['l', 'list of the modules that connects to rv_plic'], | 
|  | 'interrupt': ['lnw', 'interrupts (generated)'], | 
|  | 'alert_module': | 
|  | ['l', 'list of the modules that connects to alert_handler'], | 
|  | 'alert': ['lnw', 'alerts (generated)'], | 
|  | 'alert_async': ['l', 'async alerts (generated)'], | 
|  | 'pinmux': ['g', 'pinmux definition if doesn\'t exist, tool uses defaults'], | 
|  | 'padctrl': | 
|  | ['g', 'PADS instantiation, if doesn\'t exist, tool creates direct output'], | 
|  | 'inter_module': ['g', 'define the signal connections between the modules'], | 
|  | 'num_cores': ['pn', "number of computing units"], | 
|  | 'datawidth': ['pn', "default data width"], | 
|  | } | 
|  |  | 
|  | top_added = {} | 
|  |  | 
|  | pinmux_required = {} | 
|  | pinmux_optional = { | 
|  | 'num_mio': [ | 
|  | 'd', 'Number of Multiplexed IOs' | 
|  | ' If padctrl is used, this value will be replaced with #pads' | 
|  | ' - #DIO' | 
|  | ], | 
|  | 'dio_modules': ['l', 'List of Dedicated IOs.'], | 
|  | 'mio_modules': ['l', 'List of Multiplexed IPs/IOs'], | 
|  | 'nc_modules': ['l', 'List of NotConnected IOs'], | 
|  | } | 
|  | pinmux_added = { | 
|  | 'inputs': ['l', 'Full list of SoC inputs, `module_name.sig_name`'], | 
|  | 'outputs': ['l', 'Full list of SoC outputs, `module_name.sig_name`'], | 
|  | } | 
|  |  | 
|  | padctrl_required = {} | 
|  | padctrl_optional = { | 
|  | 'pads': ['l', 'List of pads'], | 
|  | 'attr_default': ['l', 'List of the attribute'] | 
|  | } | 
|  | padctrl_added = {} | 
|  |  | 
|  | clock_srcs_required = { | 
|  | 'name': ['s', 'name of clock group'], | 
|  | 'aon': ['s', 'yes, no. aon attribute of a clock'], | 
|  | 'freq': ['s', 'frequency of clock in Hz'], | 
|  | } | 
|  |  | 
|  | clock_srcs_optional = { | 
|  | 'derived': ['s', 'whether clock is derived'], | 
|  | 'params': ['s', 'extra clock parameters'] | 
|  | } | 
|  |  | 
|  | derived_clock_srcs_required = { | 
|  | 'name': ['s', 'name of clock group'], | 
|  | 'aon': ['s', 'yes, no. aon attribute of a clock'], | 
|  | 'freq': ['s', 'frequency of clock in Hz'], | 
|  | 'src': ['s', 'source clock'], | 
|  | 'div': ['d', 'ratio between source clock and derived clock'], | 
|  | } | 
|  |  | 
|  | clock_groups_required = { | 
|  | 'name': ['s', 'name of clock group'], | 
|  | 'src': ['s', 'yes, no. This clock group is directly from source'], | 
|  | 'sw_cg': ['s', 'yes, no, hint. Software clock gate attributes'], | 
|  | } | 
|  | clock_groups_optional = { | 
|  | 'unique': ['s', 'whether clocks in the group are unique'], | 
|  | 'clocks': ['g', 'groups of clock name to source'], | 
|  | } | 
|  | clock_groups_added = {} | 
|  |  | 
|  |  | 
|  | class TargetType(Enum): | 
|  | MODULE = "module" | 
|  | XBAR = "xbar" | 
|  |  | 
|  |  | 
|  | class Target: | 
|  | """Target class informs the checkers if we are validating a module or xbar | 
|  | """ | 
|  |  | 
|  | # The type of this target | 
|  | target_type = "" | 
|  |  | 
|  | # The key to search against | 
|  | key = "" | 
|  |  | 
|  | def __init__(self, target_type): | 
|  | self.target_type = target_type | 
|  | if target_type == TargetType.MODULE: | 
|  | self.key = "type" | 
|  | else: | 
|  | self.key = "name" | 
|  |  | 
|  |  | 
|  | # Check to see if each module/xbar defined in top.hjson exists as ip/xbar.hjson | 
|  | # Also check to make sure there are not multiple definitions of ip/xbar.hjson for each | 
|  | # top level definition | 
|  | # If it does, return a dictionary of instance names to index in ip/xbarobjs | 
|  | def check_target(top, objs, tgtobj): | 
|  | error = 0 | 
|  | idxs = OrderedDict() | 
|  |  | 
|  | for i in range(len(objs)): | 
|  | log.info("%d Order is %s" % (i, objs[i]['name'].lower())) | 
|  |  | 
|  | tgt_type = tgtobj.target_type.value | 
|  | inst_key = tgtobj.key | 
|  |  | 
|  | for cfg in top[tgt_type]: | 
|  | cfg_name = cfg['name'].lower() | 
|  | log.info("Checking target %s %s" % (tgt_type, cfg_name)) | 
|  | tgt_def = [o for o in objs if cfg[inst_key] == o['name'].lower()] | 
|  | error += check_def(tgt_def, cfg_name) | 
|  | if error: | 
|  | log.error("Target existence check failed") | 
|  | break | 
|  | else: | 
|  | idxs[cfg_name] = objs.index(tgt_def[0]) | 
|  |  | 
|  | log.info("Current state %s" % idxs) | 
|  | return error, idxs | 
|  |  | 
|  |  | 
|  | def check_padctrl(top, prefix): | 
|  | error = check_keys(top["padctrl"], padctrl_required, padctrl_optional, | 
|  | padctrl_added, prefix + " PadControl") | 
|  | return error | 
|  |  | 
|  |  | 
|  | def check_pinmux(top, prefix): | 
|  | return 0 | 
|  |  | 
|  |  | 
|  | # check for inconsistent clock group definitions | 
|  | def check_clock_groups(top): | 
|  |  | 
|  | # default empty assignment | 
|  | if "groups" not in top['clocks']: | 
|  | top['clocks']['groups'] = [] | 
|  |  | 
|  | error = 0 | 
|  | for group in top['clocks']['groups']: | 
|  | error = check_keys(group, clock_groups_required, clock_groups_optional, | 
|  | clock_groups_added, "Clock Groups") | 
|  |  | 
|  | # Check sw_cg values are valid | 
|  | if group['sw_cg'] not in ['yes', 'no', 'hint']: | 
|  | log.error("Incorrect attribute for sw_cg: {}".format( | 
|  | group['sw_cg'])) | 
|  | error += 1 | 
|  |  | 
|  | # Check combination of src and sw are valid | 
|  | if group['src'] == 'yes' and group['sw_cg'] != 'no': | 
|  | log.error("Invalid combination of src and sw_cg: {} and {}".format( | 
|  | group['src'], group['sw_cg'])) | 
|  | error += 1 | 
|  |  | 
|  | # Check combination of sw_cg and unique are valid | 
|  | unique = group['unique'] if 'unique' in group else 'no' | 
|  | if group['sw_cg'] == 'no' and unique != 'no': | 
|  | log.error( | 
|  | "Incorrect attribute combination.  When sw_cg is no, unique must be no" | 
|  | ) | 
|  | error += 1 | 
|  |  | 
|  | if error: | 
|  | break | 
|  |  | 
|  | return error | 
|  |  | 
|  |  | 
|  | def check_clocks_resets(top, ipobjs, ip_idxs, xbarobjs, xbar_idxs): | 
|  |  | 
|  | error = 0 | 
|  |  | 
|  | # check clock fields are all there | 
|  | ext_srcs = [] | 
|  | for src in top['clocks']['srcs']: | 
|  | check_keys(src, clock_srcs_required, clock_srcs_optional, {}, "Clock source") | 
|  | ext_srcs.append(src['name']) | 
|  |  | 
|  | # check derived clock sources | 
|  | log.info("Collected clocks are {}".format(ext_srcs)) | 
|  | for src in top['clocks']['derived_srcs']: | 
|  | check_keys(src, derived_clock_srcs_required, {}, {}, "Derived clocks") | 
|  | try: | 
|  | ext_srcs.index(src['src']) | 
|  | except Exception: | 
|  | error += 1 | 
|  | log.error("{} is not a valid src for {}".format(src['src'], src['name'])) | 
|  |  | 
|  | # all defined clock/reset nets | 
|  | reset_nets = [reset['name'] for reset in top['resets']['nodes']] | 
|  | clock_srcs = [clock['name'] for clock in top['clocks']['srcs'] + | 
|  | top['clocks']['derived_srcs']] | 
|  |  | 
|  | # Check clock/reset port connection for all IPs | 
|  | for ipcfg in top['module']: | 
|  | ipcfg_name = ipcfg['name'].lower() | 
|  | log.info("Checking clock/resets for %s" % ipcfg_name) | 
|  | error += validate_reset(ipcfg, ipobjs[ip_idxs[ipcfg_name]], reset_nets) | 
|  | error += validate_clock(ipcfg, ipobjs[ip_idxs[ipcfg_name]], clock_srcs) | 
|  |  | 
|  | if error: | 
|  | log.error("module clock/reset checking failed") | 
|  | break | 
|  |  | 
|  | # Check clock/reset port connection for all xbars | 
|  | for xbarcfg in top['xbar']: | 
|  | xbarcfg_name = xbarcfg['name'].lower() | 
|  | log.info("Checking clock/resets for xbar %s" % xbarcfg_name) | 
|  | error += validate_reset(xbarcfg, xbarobjs[xbar_idxs[xbarcfg_name]], | 
|  | reset_nets, "xbar") | 
|  | error += validate_clock(xbarcfg, xbarobjs[xbar_idxs[xbarcfg_name]], | 
|  | clock_srcs, "xbar") | 
|  |  | 
|  | if error: | 
|  | log.error("xbar clock/reset checking failed") | 
|  | break | 
|  |  | 
|  | return error | 
|  |  | 
|  |  | 
|  | def check_def(inst_def, name): | 
|  | error = 0 | 
|  | if not inst_def: | 
|  | log.error("Could not find %s.hjson" % name) | 
|  | error += 1 | 
|  |  | 
|  | if len(inst_def) > 1: | 
|  | log.error("Duplicate %s.hjson" % name) | 
|  | error += 1 | 
|  |  | 
|  | return error | 
|  |  | 
|  |  | 
|  | # Checks the following | 
|  | # For each defined reset connection in top*.hjson, there exists a defined port at the destination | 
|  | # and defined reset net | 
|  | # There are the same number of defined connections as there are ports | 
|  | def validate_reset(top, inst, reset_nets, prefix=""): | 
|  | # Gather inst port list | 
|  | error = 0 | 
|  | inst_port_list = [] | 
|  | if 'reset_primary' not in inst.keys(): | 
|  | log.info("%s %s does not have a reset_primary defined, default used" % | 
|  | (prefix, inst['name'])) | 
|  | inst_port_list.append("rst_ni") | 
|  | else: | 
|  | inst_port_list.append(inst['reset_primary']) | 
|  |  | 
|  | if 'other_reset_list' in inst.keys(): | 
|  | inst_port_list.extend(inst['other_reset_list']) | 
|  | log.info("%s %s resets are %s" % | 
|  | (prefix, inst['name'].lower(), inst_port_list)) | 
|  |  | 
|  | if len(top['reset_connections'].keys()) != len(inst_port_list): | 
|  | error += 1 | 
|  | log.error("%s %s mismatched number of reset ports and nets" % | 
|  | (prefix, inst['name'])) | 
|  |  | 
|  | missing_port = [ | 
|  | port for port in top['reset_connections'].keys() | 
|  | if port not in inst_port_list | 
|  | ] | 
|  |  | 
|  | if missing_port: | 
|  | error += 1 | 
|  | log.error("%s %s Following reset ports do not exist:" % | 
|  | (prefix, inst['name'])) | 
|  | [log.error("%s" % port) for port in missing_port] | 
|  |  | 
|  | missing_net = [ | 
|  | net for port, net in top['reset_connections'].items() | 
|  | if net not in reset_nets | 
|  | ] | 
|  |  | 
|  | if missing_net: | 
|  | error += 1 | 
|  | log.error("%s %s Following reset nets do not exist:" % | 
|  | (prefix, inst['name'])) | 
|  | [log.error("%s" % net) for net in missing_net] | 
|  |  | 
|  | return error | 
|  |  | 
|  |  | 
|  | # Checks the following | 
|  | # For each defined clock_src in top*.hjson, there exists a defined port at the destination | 
|  | # and defined clock source | 
|  | # There are the same number of defined connections as there are ports | 
|  | def validate_clock(top, inst, clock_srcs, prefix=""): | 
|  | # Gather inst port list | 
|  | error = 0 | 
|  | inst_port_list = [] | 
|  | if 'clock_primary' not in inst.keys(): | 
|  | log.info("%s %s does not have a clock_primary defined, default used" % | 
|  | (prefix, inst['name'])) | 
|  | inst_port_list.append("clk_i") | 
|  | else: | 
|  | inst_port_list.append(inst['clock_primary']) | 
|  |  | 
|  | if 'other_clock_list' in inst.keys(): | 
|  | inst_port_list.extend(inst['other_clock_list']) | 
|  | log.info("%s %s clocks are %s" % | 
|  | (prefix, inst['name'].lower(), inst_port_list)) | 
|  |  | 
|  | if len(top['clock_srcs'].keys()) != len(inst_port_list): | 
|  | error += 1 | 
|  | log.error("%s %s mismatched number of clock ports and nets" % | 
|  | (prefix, inst['name'])) | 
|  |  | 
|  | missing_port = [ | 
|  | port for port in top['clock_srcs'].keys() if port not in inst_port_list | 
|  | ] | 
|  |  | 
|  | if missing_port: | 
|  | error += 1 | 
|  | log.error("%s %s Following clock ports do not exist:" % | 
|  | (prefix, inst['name'])) | 
|  | [log.error("%s" % port) for port in missing_port] | 
|  |  | 
|  | missing_net = [ | 
|  | net for port, net in top['clock_srcs'].items() if net not in clock_srcs | 
|  | ] | 
|  |  | 
|  | if missing_net: | 
|  | error += 1 | 
|  | log.error("%s %s Following clock nets do not exist:" % | 
|  | (prefix, inst['name'])) | 
|  | [log.error("%s" % net) for net in missing_net] | 
|  |  | 
|  | return error | 
|  |  | 
|  |  | 
|  | def validate_top(top, ipobjs, xbarobjs): | 
|  | # return as it is for now | 
|  | error = check_keys(top, top_required, top_optional, top_added, "top") | 
|  |  | 
|  | if error != 0: | 
|  | log.error("Top HJSON has top level errors. Aborting") | 
|  | return top, error | 
|  |  | 
|  | component = top['name'] | 
|  |  | 
|  | # MODULE check | 
|  | err, ip_idxs = check_target(top, ipobjs, Target(TargetType.MODULE)) | 
|  | error += err | 
|  |  | 
|  | # XBAR check | 
|  | err, xbar_idxs = check_target(top, xbarobjs, Target(TargetType.XBAR)) | 
|  | error += err | 
|  |  | 
|  | # MEMORY check | 
|  |  | 
|  | # Clock / Reset check | 
|  | error += check_clocks_resets(top, ipobjs, ip_idxs, xbarobjs, xbar_idxs) | 
|  |  | 
|  | # Clock group check | 
|  | error += check_clock_groups(top) | 
|  |  | 
|  | # RV_PLIC check | 
|  |  | 
|  | # PINMUX & PADS check | 
|  | if "padctrl" not in top: | 
|  | log.warning("padsctrl field doesn't exist in top. Skipping pads \ | 
|  | generation. Top input/output are directly connected from \ | 
|  | peripherals.") | 
|  | # Pads configuration check | 
|  | else: | 
|  | error += check_padctrl(top, component) | 
|  |  | 
|  | if "pinmux" not in top: | 
|  | log.warning("Top {} has no 'pinmux' field. Please consider specifying \ | 
|  | pinmux and pads configuration") | 
|  | top["pinmux"] = OrderedDict() | 
|  | # checking pinmux after pads as dio connects to PAD | 
|  |  | 
|  | error += check_pinmux(top, component) | 
|  |  | 
|  | return top, error |