| # Copyright lowRISC contributors. |
| # Licensed under the Apache License, Version 2.0, see LICENSE for details. |
| # SPDX-License-Identifier: Apache-2.0 |
| r"""Contains life cycle state encoding class which is |
| used to generate new life cycle encodings. |
| """ |
| import logging as log |
| import random |
| from collections import OrderedDict |
| |
| from lib.common import (check_int, ecc_encode, get_hd, hd_histogram, |
| is_valid_codeword, scatter_bits, random_or_hexvalue) |
| |
| |
| def _is_incremental_codeword(word1, word2): |
| '''Test whether word2 is incremental wrt word1.''' |
| if len(word1) != len(word2): |
| log.error('Words are not of equal size') |
| exit(1) |
| |
| _word1 = int(word1, 2) |
| _word2 = int(word2, 2) |
| |
| # This basically checks that the second word does not |
| # clear any bits that are set to 1 in the first word. |
| return ((_word1 & _word2) == _word1) |
| |
| |
| def _get_incremental_codewords(config, base_ecc, existing_words): |
| '''Get all possible incremental codewords fulfilling the constraints.''' |
| |
| base_data = base_ecc[config['secded']['ecc_width']:] |
| |
| # We only need to spin through data bits that have not been set yet. |
| # Hence, we first count how many bits are zero (and hence still |
| # modifyable). Then, we enumerate all possible combinations and scatter |
| # the bits of the enumerated values into the correct bit positions using |
| # the scatter_bits() function. |
| incr_cands = [] |
| free_bits = base_data.count('0') |
| for k in range(1, 2**free_bits): |
| # Get incremental dataword by scattering the enumeration bits |
| # into the zero bit positions in base_data. |
| incr_cand = scatter_bits(base_data, |
| format(k, '0' + str(free_bits) + 'b')) |
| incr_cand_ecc = ecc_encode(config, incr_cand) |
| |
| # Dataword is correct by construction, but we need to check whether |
| # the ECC bits are incremental. |
| if _is_incremental_codeword(base_ecc, incr_cand_ecc): |
| # Check whether the candidate fulfills the maximum |
| # Hamming weight constraint. |
| if incr_cand_ecc.count('1') <= config['max_hw']: |
| # Check Hamming distance wrt all existing words. |
| for w in existing_words + [base_ecc]: |
| if get_hd(incr_cand_ecc, w) < config['min_hd']: |
| break |
| else: |
| incr_cands.append(incr_cand_ecc) |
| |
| return incr_cands |
| |
| |
| def _get_new_state_word_pair(config, existing_words): |
| '''Randomly generate a new incrementally writable word pair''' |
| while 1: |
| # Draw a random number and check whether it is unique and whether |
| # the Hamming weight is in range. |
| width = config['secded']['data_width'] |
| ecc_width = config['secded']['ecc_width'] |
| base = random.getrandbits(width) |
| base = format(base, '0' + str(width) + 'b') |
| base_cand_ecc = ecc_encode(config, base) |
| # disallow all-zero and all-one states |
| pop_cnt = base_cand_ecc.count('1') |
| if pop_cnt >= config['min_hw'] and pop_cnt <= config['max_hw']: |
| |
| # Check Hamming distance wrt all existing words |
| for w in existing_words: |
| if get_hd(base_cand_ecc, w) < config['min_hd']: |
| break |
| else: |
| # Get encoded incremental candidates. |
| incr_cands_ecc = _get_incremental_codewords( |
| config, base_cand_ecc, existing_words) |
| # there are valid candidates, draw one at random. |
| # otherwise we just start over. |
| if incr_cands_ecc: |
| incr_cand_ecc = random.choice(incr_cands_ecc) |
| log.info('word {}: {}|{} -> {}|{}'.format( |
| int(len(existing_words) / 2), |
| base_cand_ecc[ecc_width:], base_cand_ecc[0:ecc_width], |
| incr_cand_ecc[ecc_width:], incr_cand_ecc[0:ecc_width])) |
| existing_words.append(base_cand_ecc) |
| existing_words.append(incr_cand_ecc) |
| return (base_cand_ecc, incr_cand_ecc) |
| |
| |
| def _validate_words(config, words): |
| '''Validate generated words (base and incremental).''' |
| for k, w in enumerate(words): |
| # Check whether word is valid wrt to ECC polynomial. |
| if not is_valid_codeword(config, w): |
| log.error('Codeword {} at index {} is not valid'.format(w, k)) |
| exit(1) |
| # Check that word fulfills the Hamming weight constraints. |
| pop_cnt = w.count('1') |
| if pop_cnt < config['min_hw'] or pop_cnt > config['max_hw']: |
| log.error( |
| 'Codeword {} at index {} has wrong Hamming weight'.format( |
| w, k)) |
| exit(1) |
| # Check Hamming distance wrt to all other existing words. |
| # If the constraint is larger than 0 this implies uniqueness. |
| if k < len(words) - 1: |
| for k2, w2 in enumerate(words[k + 1:]): |
| if get_hd(w, w2) < config['min_hd']: |
| log.error( |
| 'Hamming distance between codeword {} at index {} ' |
| 'and codeword {} at index {} is too low.'.format( |
| w, k, w2, k + 1 + k2)) |
| exit(1) |
| |
| |
| class LcStEnc(): |
| '''Life cycle state encoding generator class |
| |
| The constructor expects the parsed configuration |
| hjson to be passed in. |
| ''' |
| |
| # This holds the config dict. |
| config = {} |
| # Holds generated life cycle words. |
| gen = { |
| 'ab_words': [], |
| 'cd_words': [], |
| 'ef_words': [], |
| 'stats': [], |
| } |
| |
| def __init__(self, config): |
| '''The constructor validates the configuration dict.''' |
| |
| log.info('') |
| log.info('Generate life cycle state') |
| log.info('') |
| |
| if 'seed' not in config: |
| log.error('Missing seed in configuration') |
| exit(1) |
| |
| if 'secded' not in config: |
| log.error('Missing secded configuration') |
| exit(1) |
| |
| if 'tokens' not in config: |
| log.error('Missing token configuration') |
| exit(1) |
| |
| config['secded'].setdefault('data_width', 0) |
| config['secded'].setdefault('ecc_width', 0) |
| config['secded'].setdefault('ecc_matrix', [[]]) |
| config.setdefault('num_ab_words', 0) |
| config.setdefault('num_cd_words', 0) |
| config.setdefault('num_ef_words', 0) |
| config.setdefault('min_hw', 0) |
| config.setdefault('max_hw', 0) |
| config.setdefault('min_hd', 0) |
| config.setdefault('token_size', 128) |
| |
| config['seed'] = check_int(config['seed']) |
| |
| log.info('Seed: {0:x}'.format(config['seed'])) |
| log.info('') |
| |
| config['secded']['data_width'] = check_int( |
| config['secded']['data_width']) |
| config['secded']['ecc_width'] = check_int( |
| config['secded']['ecc_width']) |
| config['num_ab_words'] = check_int(config['num_ab_words']) |
| config['num_cd_words'] = check_int(config['num_cd_words']) |
| config['num_ef_words'] = check_int(config['num_ef_words']) |
| config['min_hw'] = check_int(config['min_hw']) |
| config['max_hw'] = check_int(config['max_hw']) |
| config['min_hd'] = check_int(config['min_hd']) |
| config['token_size'] = check_int(config['token_size']) |
| |
| total_width = config['secded']['data_width'] + config['secded'][ |
| 'ecc_width'] |
| |
| if config['min_hw'] >= total_width or \ |
| config['max_hw'] > total_width or \ |
| config['min_hw'] >= config['max_hw']: |
| log.error('Hamming weight constraints are inconsistent.') |
| exit(1) |
| |
| if config['max_hw'] - config['min_hw'] + 1 < config['min_hd']: |
| log.error('Hamming distance constraint is inconsistent.') |
| exit(1) |
| |
| if config['secded']['ecc_width'] != len( |
| config['secded']['ecc_matrix']): |
| log.error('ECC matrix does not have correct number of rows') |
| exit(1) |
| |
| log.info('SECDED Matrix:') |
| for i, l in enumerate(config['secded']['ecc_matrix']): |
| log.info('ECC Bit {} Fanin: {}'.format(i, l)) |
| for j, e in enumerate(l): |
| e = check_int(e) |
| config['secded']['ecc_matrix'][i][j] = e |
| |
| log.info('') |
| |
| hashed_tokens = [] |
| for token in config['tokens']: |
| random_or_hexvalue(token, 'value', config['token_size']) |
| hashed_token = OrderedDict() |
| hashed_token['name'] = token['name'] + 'Hashed' |
| # TODO: plug in PRESENT-based hashing algo or KMAC |
| hashed_token['value'] = token['value'] |
| hashed_tokens.append(hashed_token) |
| |
| config['tokens'] += hashed_tokens |
| |
| print(config['tokens']) |
| self.config = config |
| |
| # Re-initialize with seed to make results reproducible. |
| random.seed(int(self.config['seed'])) |
| |
| # Generate new encoding words |
| word_types = ['ab_words', 'cd_words', 'ef_words'] |
| existing_words = [] |
| for w in word_types: |
| while len(self.gen[w]) < self.config['num_' + w]: |
| new_word = _get_new_state_word_pair(self.config, |
| existing_words) |
| self.gen[w].append(new_word) |
| |
| # Validate words (this must not fail at this point). |
| _validate_words(self.config, existing_words) |
| |
| # Print out HD histogram |
| self.gen['stats'] = hd_histogram(existing_words) |
| |
| log.info('') |
| log.info('Hamming distance histogram:') |
| log.info('') |
| for bar in self.gen['stats']["bars"]: |
| log.info(bar) |
| log.info('') |
| log.info('Minimum HD: {}'.format(self.gen['stats']['min_hd'])) |
| log.info('Maximum HD: {}'.format(self.gen['stats']['max_hd'])) |
| log.info('Minimum HW: {}'.format(self.gen['stats']['min_hw'])) |
| log.info('Maximum HW: {}'.format(self.gen['stats']['max_hw'])) |
| |
| log.info('') |
| log.info('Successfully generated life cycle state.') |
| log.info('') |