| #!/usr/bin/env python3 | 
 | # Copyright lowRISC contributors. | 
 | # Licensed under the Apache License, Version 2.0, see LICENSE for details. | 
 | # SPDX-License-Identifier: Apache-2.0 | 
 |  | 
 | '''A tool to copy source code from upstream into this repository. | 
 |  | 
 | For an introduction to using this tool, see doc/ug/vendor_hw.md in this | 
 | repository (on the internet at https://docs.opentitan.org/doc/ug/vendor_hw/). | 
 |  | 
 | For full documentation, see doc/rm/vendor_in_tool.md (on the internet at | 
 | https://docs.opentitan.org/doc/rm/vendor_in_tool). | 
 |  | 
 | ''' | 
 |  | 
 | import argparse | 
 | import fnmatch | 
 | import logging as log | 
 | import os | 
 | import re | 
 | import shutil | 
 | import subprocess | 
 | import sys | 
 | import tempfile | 
 | import textwrap | 
 | from pathlib import Path | 
 |  | 
 | import hjson | 
 |  | 
 | EXCLUDE_ALWAYS = ['.git'] | 
 |  | 
 | LOCK_FILE_HEADER = """// Copyright lowRISC contributors. | 
 | // Licensed under the Apache License, Version 2.0, see LICENSE for details. | 
 | // SPDX-License-Identifier: Apache-2.0 | 
 |  | 
 | // This file is generated by the util/vendor script. Please do not modify it | 
 | // manually. | 
 |  | 
 | """ | 
 |  | 
 | # Keys in the description (configuration) file which can be overridden through | 
 | # the command line. | 
 | OVERRIDABLE_DESC_KEYS = [ | 
 |     'patch_repo.url', | 
 |     'patch_repo.rev_base', | 
 |     'patch_repo.rev_patched', | 
 |     'upstream.url', | 
 |     'upstream.rev', | 
 | ] | 
 |  | 
 | verbose = False | 
 |  | 
 |  | 
 | def git_is_clean_workdir(git_workdir): | 
 |     """Check if the git working directory is clean (no unstaged or staged changes)""" | 
 |     cmd = ['git', 'status', '--untracked-files=no', '--porcelain'] | 
 |     modified_files = subprocess.run(cmd, | 
 |                                     cwd=str(git_workdir), | 
 |                                     check=True, | 
 |                                     stdout=subprocess.PIPE, | 
 |                                     stderr=subprocess.PIPE).stdout.strip() | 
 |     return not modified_files | 
 |  | 
 |  | 
 | def github_qualify_references(log, repo_userorg, repo_name): | 
 |     """ Replace "unqualified" GitHub references with "fully qualified" one | 
 |  | 
 |     GitHub automatically links issues and pull requests if they have a specific | 
 |     format. Links can be qualified with the user/org name and the repository | 
 |     name, or unqualified, if they only contain the issue or pull request number. | 
 |  | 
 |     This function converts all unqualified references to qualified ones. | 
 |  | 
 |     See https://help.github.com/en/articles/autolinked-references-and-urls#issues-and-pull-requests | 
 |     for a documentation of all supported formats. | 
 |     """ | 
 |  | 
 |     r = re.compile(r"(^|[^\w])(?:#|[gG][hH]-)(\d+)\b") | 
 |     repl_str = r'\1%s/%s#\2' % (repo_userorg, repo_name) | 
 |     return [r.sub(repl_str, line) for line in log] | 
 |  | 
 |  | 
 | def test_github_qualify_references(): | 
 |     repo_userorg = 'lowRISC' | 
 |     repo_name = 'ibex' | 
 |  | 
 |     # Unqualified references, should be replaced | 
 |     items_unqualified = [ | 
 |         '#28', | 
 |         'GH-27', | 
 |         'klaus #27', | 
 |         'Fixes #27', | 
 |         'Fixes #27 and #28', | 
 |         '(#27)', | 
 |         'something (#27) done', | 
 |         '#27 and (GH-38)', | 
 |     ] | 
 |     exp_items_unqualified = [ | 
 |         'lowRISC/ibex#28', | 
 |         'lowRISC/ibex#27', | 
 |         'klaus lowRISC/ibex#27', | 
 |         'Fixes lowRISC/ibex#27', | 
 |         'Fixes lowRISC/ibex#27 and lowRISC/ibex#28', | 
 |         '(lowRISC/ibex#27)', | 
 |         'something (lowRISC/ibex#27) done', | 
 |         'lowRISC/ibex#27 and (lowRISC/ibex#38)', | 
 |     ] | 
 |     assert github_qualify_references(items_unqualified, repo_userorg, | 
 |                                      repo_name) == exp_items_unqualified | 
 |  | 
 |     # Qualified references, should stay as they are | 
 |     items_qualified = [ | 
 |         'Fixes lowrisc/ibex#27', | 
 |         'lowrisc/ibex#2', | 
 |     ] | 
 |     assert github_qualify_references(items_qualified, repo_userorg, | 
 |                                      repo_name) == items_qualified | 
 |  | 
 |     # Invalid references, should stay as they are | 
 |     items_invalid = [ | 
 |         'something#27', | 
 |         'lowrisc/ibex#', | 
 |     ] | 
 |     assert github_qualify_references(items_invalid, repo_userorg, | 
 |                                      repo_name) == items_invalid | 
 |  | 
 |  | 
 | def test_github_parse_url(): | 
 |     assert github_parse_url('https://example.com/something/asdf.git') is None | 
 |     assert github_parse_url('https://github.com/lowRISC/ibex.git') == ( | 
 |         'lowRISC', 'ibex') | 
 |     assert github_parse_url('https://github.com/lowRISC/ibex') == ('lowRISC', | 
 |                                                                    'ibex') | 
 |     assert github_parse_url('git@github.com:lowRISC/ibex.git') == ('lowRISC', | 
 |                                                                    'ibex') | 
 |  | 
 |  | 
 | def github_parse_url(github_repo_url): | 
 |     """Parse a GitHub repository URL into its parts. | 
 |  | 
 |     Return a tuple (userorg, name), or None if the parsing failed. | 
 |     """ | 
 |  | 
 |     regex = r"(?:@github\.com\:|\/github\.com\/)([a-zA-Z\d-]+)\/([a-zA-Z\d-]+)(?:\.git)?$" | 
 |     m = re.search(regex, github_repo_url) | 
 |     if m is None: | 
 |         return None | 
 |     return (m.group(1), m.group(2)) | 
 |  | 
 |  | 
 | def produce_shortlog(clone_dir, mapping, old_rev, new_rev): | 
 |     """ Produce a list of changes between two revisions, one revision per line | 
 |  | 
 |     Merges are excluded""" | 
 |  | 
 |     # If mapping is None, we want to list all changes below clone_dir. | 
 |     # Otherwise, we want to list changes in each 'source' in the mapping. Since | 
 |     # these strings are paths relative to clone_dir, we can just pass them all | 
 |     # to git and let it figure out what to do. | 
 |     subdirs = (['.'] if mapping is None | 
 |                else [m.from_path for m in mapping.items]) | 
 |  | 
 |     cmd = (['git', '-C', str(clone_dir), 'log', | 
 |            '--pretty=format:%s (%aN)', '--no-merges', | 
 |             old_rev + '..' + new_rev] + | 
 |            subdirs) | 
 |     try: | 
 |         proc = subprocess.run(cmd, | 
 |                               cwd=str(clone_dir), | 
 |                               check=True, | 
 |                               stdout=subprocess.PIPE, | 
 |                               stderr=subprocess.PIPE, | 
 |                               universal_newlines=True) | 
 |         return proc.stdout.splitlines() | 
 |     except subprocess.CalledProcessError as e: | 
 |         log.error("Unable to capture shortlog: %s", e.stderr) | 
 |         return "" | 
 |  | 
 |  | 
 | def format_list_to_str(list, width=70): | 
 |     """ Create Markdown-style formatted string from a list of strings """ | 
 |     wrapper = textwrap.TextWrapper(initial_indent="* ", | 
 |                                    subsequent_indent="  ", | 
 |                                    width=width) | 
 |     return '\n'.join([wrapper.fill(s) for s in list]) | 
 |  | 
 |  | 
 | class JsonError(Exception): | 
 |     '''An error class for when data in the source HJSON is bad''' | 
 |     def __init__(self, path, msg): | 
 |         self.path = path | 
 |         self.msg = msg | 
 |  | 
 |     def __str__(self): | 
 |         return 'In hjson at {}, {}'.format(self.path, self.msg) | 
 |  | 
 |  | 
 | def get_field(path, where, data, name, expected_type=dict, optional=False, constructor=None): | 
 |     value = data.get(name) | 
 |     if value is None: | 
 |         if not optional: | 
 |             raise JsonError(path, '{}, missing {!r} field.'.format(where, name)) | 
 |         return None | 
 |  | 
 |     if not isinstance(value, expected_type): | 
 |         raise JsonError(path, | 
 |                         '{}, the {!r} field is {!r}, but should be of type {!r}.' | 
 |                         .format(where, name, value, expected_type.__name__)) | 
 |  | 
 |     return value if constructor is None else constructor(value) | 
 |  | 
 |  | 
 | class Upstream: | 
 |     '''A class representing the 'upstream' field in a config or lock file''' | 
 |     def __init__(self, path, data): | 
 |         # Fields: 'url', 'rev', 'only_subdir' (optional). All should be strings. | 
 |         where = 'in upstream dict' | 
 |         self.url = get_field(path, where, data, 'url', str) | 
 |         self.rev = get_field(path, where, data, 'rev', str) | 
 |         self.only_subdir = get_field(path, where, data, | 
 |                                      'only_subdir', str, optional=True) | 
 |  | 
 |     def as_dict(self): | 
 |         data = {'url': self.url, 'rev': self.rev} | 
 |         if self.only_subdir is not None: | 
 |             data['only_subdir'] = self.only_subdir | 
 |         return data | 
 |  | 
 |  | 
 | class PatchRepo: | 
 |     '''A class representing the 'patch_repo' field in a config file''' | 
 |     def __init__(self, path, data): | 
 |         # Fields: 'url', 'rev_base', 'rev_patched'. All should be strings. | 
 |         where = 'in patch_repo dict' | 
 |         self.url = get_field(path, where, data, 'url', str) | 
 |         self.rev_base = get_field(path, where, data, 'rev_base', str) | 
 |         self.rev_patched = get_field(path, where, data, 'rev_patched', str) | 
 |  | 
 |  | 
 | class Mapping1: | 
 |     '''A class to represent a single item in the 'mapping' field in a config file''' | 
 |     def __init__(self, from_path, to_path, patch_dir): | 
 |         self.from_path = from_path | 
 |         self.to_path = to_path | 
 |         self.patch_dir = patch_dir | 
 |  | 
 |     @staticmethod | 
 |     def check_path(idx, name, val): | 
 |         '''Check that a path is valid for use in a mapping. | 
 |  | 
 |         This spots things like '../../../foo' or '/etc/passwd'. | 
 |         ''' | 
 |         val = os.path.normpath(val) | 
 |         if val.startswith('/') or val.startswith('..'): | 
 |             raise JsonError(val, | 
 |                             'Mapping entry {} has a bad path for {!r} ' | 
 |                             '(must be a relative path that doesn\'t ' | 
 |                             'escape the directory)' | 
 |                             .format(idx + 1, name)) | 
 |         return Path(val) | 
 |  | 
 |     @staticmethod | 
 |     def make(path, idx, data): | 
 |         if isinstance(data, str): | 
 |             # A string S is interpreted as {from: S, to: S}. | 
 |             from_path = Mapping1.check_path(idx, 'from', data) | 
 |             to_path = from_path | 
 |             patch_dir = None | 
 |         else: | 
 |             # Otherwise, we expect a dictionary with 'from', 'to' and | 
 |             # (optionally) 'patch_dir'. | 
 |             assert isinstance(data, dict) | 
 |  | 
 |             def get_path(name, optional=False): | 
 |                 val = get_field(path, 'in mapping entry {}'.format(idx + 1), | 
 |                                 data, name, | 
 |                                 expected_type=str, optional=optional) | 
 |                 if val is None: | 
 |                     return None | 
 |  | 
 |                 return Mapping1.check_path(idx, name, val) | 
 |  | 
 |             from_path = get_path('from') | 
 |             to_path = get_path('to') | 
 |             patch_dir = get_path('patch_dir', optional=True) | 
 |  | 
 |         return Mapping1(from_path, to_path, patch_dir) | 
 |  | 
 |     @staticmethod | 
 |     def make_default(have_patch_dir): | 
 |         '''Make a default mapping1, which copies everything straight through''' | 
 |         return Mapping1(Path('.'), Path('.'), | 
 |                         Path('.') if have_patch_dir else None) | 
 |  | 
 |     @staticmethod | 
 |     def apply_patch(basedir, patchfile): | 
 |         cmd = ['git', 'apply', '--directory', str(basedir), '-p1', | 
 |                str(patchfile)] | 
 |         if verbose: | 
 |             cmd += ['--verbose'] | 
 |         subprocess.run(cmd, check=True) | 
 |  | 
 |     def import_from_upstream(self, upstream_path, | 
 |                              target_path, exclude_files, patch_dir): | 
 |         '''Copy from the upstream checkout to target_path''' | 
 |         from_path = upstream_path / self.from_path | 
 |         to_path = target_path / self.to_path | 
 |  | 
 |         # Make sure the target directory actually exists | 
 |         to_path.parent.mkdir(exist_ok=True, parents=True) | 
 |  | 
 |         # Copy src to dst recursively. For directories, we can use | 
 |         # shutil.copytree. This doesn't support files, though, so we have to | 
 |         # check for them first. | 
 |         if from_path.is_file(): | 
 |             shutil.copy(str(from_path), str(to_path)) | 
 |         else: | 
 |             ignore = ignore_patterns(str(upstream_path), *exclude_files) | 
 |             shutil.copytree(str(from_path), str(to_path), ignore=ignore) | 
 |  | 
 |         # Apply any patches to the copied files. If self.patch_dir is None, | 
 |         # there are none to apply. Otherwise, resolve it relative to patch_dir. | 
 |         if self.patch_dir is not None: | 
 |             patches = (patch_dir / self.patch_dir).glob('*.patch') | 
 |             for patch in sorted(patches): | 
 |                 log.info("Applying patch {} at {}".format(patch, to_path)) | 
 |                 Mapping1.apply_patch(to_path, patch) | 
 |  | 
 |  | 
 | class Mapping: | 
 |     '''A class representing the 'mapping' field in a config file | 
 |  | 
 |     This should be a list of dicts. | 
 |     ''' | 
 |     def __init__(self, items): | 
 |         self.items = items | 
 |  | 
 |     @staticmethod | 
 |     def make(path, data): | 
 |         items = [] | 
 |         assert isinstance(data, list) | 
 |         for idx, elt in enumerate(data): | 
 |             if not (isinstance(elt, dict) or isinstance(elt, str)): | 
 |                 raise JsonError(path, | 
 |                                 f'Mapping element {elt!r} is not a dict.') | 
 |             items.append(Mapping1.make(path, idx, elt)) | 
 |  | 
 |         return Mapping(items) | 
 |  | 
 |     def has_patch_dir(self): | 
 |         '''Check whether at least one item defines a patch dir''' | 
 |         for item in self.items: | 
 |             if item.patch_dir is not None: | 
 |                 return True | 
 |         return False | 
 |  | 
 |  | 
 | class LockDesc: | 
 |     '''A class representing the contents of a lock file''' | 
 |     def __init__(self, handle): | 
 |         data = hjson.loads(handle.read(), use_decimal=True) | 
 |         self.upstream = get_field(handle.name, 'at top-level', data, 'upstream', | 
 |                                   constructor=lambda data: Upstream(handle.name, data)) | 
 |  | 
 |  | 
 | class Desc: | 
 |     '''A class representing the configuration file''' | 
 |  | 
 |     def __init__(self, handle, desc_overrides): | 
 |  | 
 |         # Ensure description file matches our naming rules (otherwise we don't | 
 |         # know the name for the lockfile). This regex checks that we have the | 
 |         # right suffix and a nonempty name. | 
 |         if not re.match(r'.+\.vendor\.hjson', handle.name): | 
 |             raise ValueError("Description file names must have a .vendor.hjson suffix.") | 
 |  | 
 |         data = hjson.loads(handle.read(), use_decimal=True) | 
 |         where = 'at top-level' | 
 |  | 
 |         self.apply_overrides(data, desc_overrides) | 
 |  | 
 |         path = Path(handle.name) | 
 |  | 
 |         def take_path(p): | 
 |             return path.parent / p | 
 |  | 
 |         self.path = path | 
 |         self.name = get_field(path, where, data, 'name', expected_type=str) | 
 |         self.target_dir = get_field(path, where, data, 'target_dir', | 
 |                                     expected_type=str, constructor=take_path) | 
 |         self.upstream = get_field(path, where, data, 'upstream', | 
 |                                   constructor=lambda data: Upstream(path, data)) | 
 |         self.patch_dir = get_field(path, where, data, 'patch_dir', | 
 |                                    optional=True, expected_type=str, constructor=take_path) | 
 |         self.patch_repo = get_field(path, where, data, 'patch_repo', | 
 |                                     optional=True, | 
 |                                     constructor=lambda data: PatchRepo(path, data)) | 
 |         self.exclude_from_upstream = (get_field(path, where, data, 'exclude_from_upstream', | 
 |                                                 optional=True, expected_type=list) or | 
 |                                       []) | 
 |         self.mapping = get_field(path, where, data, 'mapping', optional=True, | 
 |                                  expected_type=list, | 
 |                                  constructor=lambda data: Mapping.make(path, data)) | 
 |  | 
 |         # Add default exclusions | 
 |         self.exclude_from_upstream += EXCLUDE_ALWAYS | 
 |  | 
 |         # It doesn't make sense to define a patch_repo, but not a patch_dir | 
 |         # (where should we put the patches that we get?) | 
 |         if self.patch_repo is not None and self.patch_dir is None: | 
 |             raise JsonError(path, 'Has patch_repo but not patch_dir.') | 
 |  | 
 |         # We don't currently support a patch_repo and a mapping (just because | 
 |         # we haven't written the code to generate the patches across subdirs | 
 |         # yet). Tracked in issue #2317. | 
 |         if self.patch_repo is not None and self.mapping is not None: | 
 |             raise JsonError(path, | 
 |                             "vendor.py doesn't currently support patch_repo " | 
 |                             "and mapping at the same time (see issue #2317).") | 
 |  | 
 |         # If a patch_dir is defined and there is no mapping, we will look in | 
 |         # that directory for patches and apply them in (the only) directory | 
 |         # that we copy stuff into. | 
 |         # | 
 |         # If there is a mapping check that there is a patch_dir if and only if | 
 |         # least one mapping entry uses it. | 
 |         if self.mapping is not None: | 
 |             if self.patch_dir is not None: | 
 |                 if not self.mapping.has_patch_dir(): | 
 |                     raise JsonError(path, 'Has patch_dir, but no mapping item uses it.') | 
 |             else: | 
 |                 if self.mapping.has_patch_dir(): | 
 |                     raise JsonError(path, | 
 |                                     'Has a mapping item with a patch directory, ' | 
 |                                     'but there is no global patch_dir key.') | 
 |  | 
 |         # Check that exclude_from_upstream really is a list of strings. Most of | 
 |         # this type-checking is in the constructors for field types, but we | 
 |         # don't have a "ExcludeList" class, so have to do it explicitly here. | 
 |         for efu in self.exclude_from_upstream: | 
 |             if not isinstance(efu, str): | 
 |                 raise JsonError(path, | 
 |                                 'exclude_from_upstream has entry {}, which is not a string.' | 
 |                                 .format(efu)) | 
 |  | 
 |     def apply_overrides(self, desc_data, desc_overrides): | 
 |         """ Apply overrides from command line to configuration file data | 
 |  | 
 |         Updates are applied to the desc_data reference.""" | 
 |  | 
 |         for key, value in desc_overrides: | 
 |             log.info("Overriding description key {!r} with value {!r}".format( | 
 |                 key, value)) | 
 |             ref = desc_data | 
 |             split_keys = key.split('.') | 
 |             for key_part in split_keys[:-1]: | 
 |                 if key_part not in ref: | 
 |                     ref[key_part] = {} | 
 |                 ref = ref[key_part] | 
 |             ref[split_keys[-1]] = value | 
 |  | 
 |     def lock_file_path(self): | 
 |         desc_file_stem = self.path.name.rsplit('.', 2)[0] | 
 |         return self.path.with_name(desc_file_stem + '.lock.hjson') | 
 |  | 
 |     def import_from_upstream(self, upstream_path): | 
 |         log.info('Copying upstream sources to {}'.format(self.target_dir)) | 
 |  | 
 |         # Remove existing directories before importing them again | 
 |         shutil.rmtree(str(self.target_dir), ignore_errors=True) | 
 |  | 
 |         items = (self.mapping.items if self.mapping is not None | 
 |                  else [Mapping1.make_default(self.patch_dir is not None)]) | 
 |         for map1 in items: | 
 |             map1.import_from_upstream(upstream_path, | 
 |                                       self.target_dir, | 
 |                                       self.exclude_from_upstream, | 
 |                                       self.patch_dir) | 
 |  | 
 |  | 
 | def refresh_patches(desc): | 
 |     if desc.patch_repo is None: | 
 |         log.fatal('Unable to refresh patches, patch_repo not set in config.') | 
 |         sys.exit(1) | 
 |  | 
 |     log.info('Refreshing patches in {}'.format(desc.patch_dir)) | 
 |  | 
 |     # remove existing patches | 
 |     for patch in desc.patch_dir.glob('*.patch'): | 
 |         os.unlink(str(patch)) | 
 |  | 
 |     # get current patches | 
 |     _export_patches(desc.patch_repo.url, desc.patch_dir, | 
 |                     desc.patch_repo.rev_base, | 
 |                     desc.patch_repo.rev_patched) | 
 |  | 
 |  | 
 | def _export_patches(patchrepo_clone_url, target_patch_dir, upstream_rev, | 
 |                     patched_rev): | 
 |     with tempfile.TemporaryDirectory() as clone_dir: | 
 |         clone_git_repo(patchrepo_clone_url, clone_dir, patched_rev) | 
 |         rev_range = 'origin/' + upstream_rev + '..' + 'origin/' + patched_rev | 
 |         cmd = [ | 
 |             'git', | 
 |             'format-patch', | 
 |             '--no-signature', | 
 |             '--no-stat', | 
 |             '-o', | 
 |             str(target_patch_dir.resolve()), | 
 |             rev_range | 
 |         ] | 
 |         if not verbose: | 
 |             cmd += ['-q'] | 
 |         subprocess.run(cmd, cwd=str(clone_dir), check=True) | 
 |  | 
 |  | 
 | def ignore_patterns(base_dir, *patterns): | 
 |     """Similar to shutil.ignore_patterns, but with support for directory excludes.""" | 
 |     def _rel_to_base(path, name): | 
 |         return os.path.relpath(os.path.join(path, name), base_dir) | 
 |  | 
 |     def _ignore_patterns(path, names): | 
 |         ignored_names = [] | 
 |         for pattern in patterns: | 
 |             pattern_matches = [ | 
 |                 n for n in names | 
 |                 if fnmatch.fnmatch(_rel_to_base(path, n), pattern) | 
 |             ] | 
 |             ignored_names.extend(pattern_matches) | 
 |         return set(ignored_names) | 
 |  | 
 |     return _ignore_patterns | 
 |  | 
 |  | 
 | def clone_git_repo(repo_url, clone_dir, rev='master'): | 
 |     log.info('Cloning upstream repository %s @ %s', repo_url, rev) | 
 |  | 
 |     # Clone the whole repository | 
 |     cmd = ['git', 'clone', '--no-single-branch'] | 
 |     if not verbose: | 
 |         cmd += ['-q'] | 
 |     cmd += [repo_url, str(clone_dir)] | 
 |     subprocess.run(cmd, check=True) | 
 |  | 
 |     # Check out exactly the revision requested | 
 |     cmd = ['git', '-C', str(clone_dir), 'checkout', '--force', rev] | 
 |     if not verbose: | 
 |         cmd += ['-q'] | 
 |     subprocess.run(cmd, check=True) | 
 |  | 
 |     # Get revision information | 
 |     cmd = ['git', '-C', str(clone_dir), 'rev-parse', 'HEAD'] | 
 |     rev = subprocess.run(cmd, | 
 |                          stdout=subprocess.PIPE, | 
 |                          stderr=subprocess.PIPE, | 
 |                          check=True, | 
 |                          universal_newlines=True).stdout.strip() | 
 |     log.info('Cloned at revision %s', rev) | 
 |     return rev | 
 |  | 
 |  | 
 | def git_get_short_rev(clone_dir, rev): | 
 |     """ Get the shortened SHA-1 hash for a revision """ | 
 |     cmd = ['git', '-C', str(clone_dir), 'rev-parse', '--short', rev] | 
 |     short_rev = subprocess.run(cmd, | 
 |                                stdout=subprocess.PIPE, | 
 |                                stderr=subprocess.PIPE, | 
 |                                check=True, | 
 |                                universal_newlines=True).stdout.strip() | 
 |     return short_rev | 
 |  | 
 |  | 
 | def git_add_commit(paths, commit_msg): | 
 |     """ Stage and commit all changes in paths""" | 
 |  | 
 |     assert paths | 
 |     base_dir = paths[0].parent | 
 |  | 
 |     # Stage all changes | 
 |     # | 
 |     # Rather than figuring out GIT_DIR properly, we cheat and use "git -C" to | 
 |     # pretend that we're running in base_dir. Of course, the elements of paths | 
 |     # are relative to our actual working directory. Rather than do anything | 
 |     # clever, we just resolve them to absolute paths as we go. | 
 |     abs_paths = [p.resolve() for p in paths] | 
 |     subprocess.run(['git', '-C', base_dir, 'add'] + abs_paths, check=True) | 
 |  | 
 |     cmd_commit = ['git', '-C', base_dir, 'commit', '-s', '-F', '-'] | 
 |     try: | 
 |         subprocess.run(cmd_commit, | 
 |                        check=True, | 
 |                        universal_newlines=True, | 
 |                        input=commit_msg) | 
 |     except subprocess.CalledProcessError: | 
 |         log.warning("Unable to create commit. Are there no changes?") | 
 |  | 
 |  | 
 | def define_arg_type(arg): | 
 |     """Sanity-check and return a config file override argument""" | 
 |     try: | 
 |         (key, value) = [v.strip() for v in arg.split('=', 2)] | 
 |     except Exception: | 
 |         raise argparse.ArgumentTypeError( | 
 |             'unable to parse {!r}: configuration overrides must be in the form key=value' | 
 |             .format(arg)) | 
 |  | 
 |     if key not in OVERRIDABLE_DESC_KEYS: | 
 |         raise argparse.ArgumentTypeError( | 
 |             'invalid configuration override: key {!r} cannot be overwritten' | 
 |             .format(key)) | 
 |     return (key, value) | 
 |  | 
 |  | 
 | def main(argv): | 
 |     parser = argparse.ArgumentParser(prog="vendor", description=__doc__) | 
 |     parser.add_argument( | 
 |         '--update', | 
 |         '-U', | 
 |         dest='update', | 
 |         action='store_true', | 
 |         help='Update locked version of repository with upstream changes') | 
 |     parser.add_argument('--refresh-patches', | 
 |                         action='store_true', | 
 |                         help='Refresh the patches from the patch repository') | 
 |     parser.add_argument('--commit', | 
 |                         '-c', | 
 |                         action='store_true', | 
 |                         help='Commit the changes') | 
 |     parser.add_argument('--desc-override', | 
 |                         '-D', | 
 |                         dest="desc_overrides", | 
 |                         action="append", | 
 |                         type=define_arg_type, | 
 |                         default=[], | 
 |                         help='Override a setting in the description file. ' | 
 |                              'Format: -Dsome.key=value. ' | 
 |                              'Can be used multiple times.') | 
 |     parser.add_argument('desc_file', | 
 |                         metavar='file', | 
 |                         type=argparse.FileType('r', encoding='UTF-8'), | 
 |                         help='vendoring description file (*.vendor.hjson)') | 
 |     parser.add_argument('--verbose', '-v', action='store_true', help='Verbose') | 
 |     args = parser.parse_args() | 
 |  | 
 |     global verbose | 
 |     verbose = args.verbose | 
 |     if (verbose): | 
 |         log.basicConfig(format="%(levelname)s: %(message)s", level=log.DEBUG) | 
 |     else: | 
 |         log.basicConfig(format="%(levelname)s: %(message)s") | 
 |  | 
 |     # Load input files (desc file; lock file) and check syntax etc. | 
 |     try: | 
 |         # Load description file | 
 |         desc = Desc(args.desc_file, args.desc_overrides) | 
 |         lock_file_path = desc.lock_file_path() | 
 |  | 
 |         # Try to load lock file (which might not exist) | 
 |         try: | 
 |             with open(str(lock_file_path), 'r') as lock_file: | 
 |                 lock = LockDesc(lock_file) | 
 |         except FileNotFoundError: | 
 |             lock = None | 
 |     except (JsonError, ValueError) as err: | 
 |         log.fatal(str(err)) | 
 |         raise SystemExit(1) | 
 |  | 
 |     # Check for a clean working directory when commit is requested | 
 |     if args.commit: | 
 |         if not git_is_clean_workdir(desc.path.parent): | 
 |             log.fatal("A clean git working directory is required for " | 
 |                       "--commit/-c. git stash your changes and try again.") | 
 |             raise SystemExit(1) | 
 |  | 
 |     if lock is None and not args.update: | 
 |         log.warning("No lock file at {}, so will update upstream repo." | 
 |                     .format(str(desc.lock_file_path()))) | 
 |         args.update = True | 
 |  | 
 |     # If we have a lock file and we're not in update mode, override desc's | 
 |     # upstream field with the one from the lock file. Keep track of whether the | 
 |     # URL differs (in which case, we can't get a shortlog) | 
 |     changed_url = False | 
 |     if lock is not None: | 
 |         changed_url = desc.upstream.url != lock.upstream.url | 
 |         if not args.update: | 
 |             desc.upstream = lock.upstream | 
 |  | 
 |     if args.refresh_patches: | 
 |         refresh_patches(desc) | 
 |  | 
 |     with tempfile.TemporaryDirectory() as clone_dir: | 
 |         # clone upstream repository | 
 |         upstream_new_rev = clone_git_repo(desc.upstream.url, clone_dir, rev=desc.upstream.rev) | 
 |  | 
 |         if not args.update: | 
 |             if upstream_new_rev != lock.upstream.rev: | 
 |                 log.fatal( | 
 |                     "Revision mismatch. Unable to re-clone locked version of repository." | 
 |                 ) | 
 |                 log.fatal("Attempted revision: %s", desc.upstream.rev) | 
 |                 log.fatal("Re-cloned revision: %s", upstream_new_rev) | 
 |                 raise SystemExit(1) | 
 |  | 
 |         clone_subdir = Path(clone_dir) | 
 |         if desc.upstream.only_subdir is not None: | 
 |             clone_subdir = clone_subdir / desc.upstream.only_subdir | 
 |             if not clone_subdir.is_dir(): | 
 |                 log.fatal("subdir '{}' does not exist in repo" | 
 |                           .format(desc.upstream.only_subdir)) | 
 |                 raise SystemExit(1) | 
 |  | 
 |         # copy selected files from upstream repo and apply patches as necessary | 
 |         desc.import_from_upstream(clone_subdir) | 
 |  | 
 |         # get shortlog | 
 |         get_shortlog = args.update | 
 |         if args.update: | 
 |             if lock is None: | 
 |                 get_shortlog = False | 
 |                 log.warning("No lock file %s: unable to summarize changes.", str(lock_file_path)) | 
 |             elif changed_url: | 
 |                 get_shortlog = False | 
 |                 log.warning("The repository URL changed since the last run. " | 
 |                             "Unable to get log of changes.") | 
 |  | 
 |         shortlog = None | 
 |         if get_shortlog: | 
 |             shortlog = produce_shortlog(clone_subdir, desc.mapping, | 
 |                                         lock.upstream.rev, upstream_new_rev) | 
 |  | 
 |             # Ensure fully-qualified issue/PR references for GitHub repos | 
 |             gh_repo_info = github_parse_url(desc.upstream.url) | 
 |             if gh_repo_info: | 
 |                 shortlog = github_qualify_references(shortlog, gh_repo_info[0], | 
 |                                                      gh_repo_info[1]) | 
 |  | 
 |             log.info("Changes since the last import:\n" + | 
 |                      format_list_to_str(shortlog)) | 
 |  | 
 |         # write lock file | 
 |         if args.update: | 
 |             lock_data = {} | 
 |             lock_data['upstream'] = desc.upstream.as_dict() | 
 |             lock_data['upstream']['rev'] = upstream_new_rev | 
 |             with open(str(lock_file_path), 'w', encoding='UTF-8') as f: | 
 |                 f.write(LOCK_FILE_HEADER) | 
 |                 hjson.dump(lock_data, f) | 
 |                 f.write("\n") | 
 |                 log.info("Wrote lock file %s", str(lock_file_path)) | 
 |  | 
 |         # Commit changes | 
 |         if args.commit: | 
 |             sha_short = git_get_short_rev(clone_subdir, upstream_new_rev) | 
 |  | 
 |             repo_info = github_parse_url(desc.upstream.url) | 
 |             if repo_info is not None: | 
 |                 sha_short = "%s/%s@%s" % (repo_info[0], repo_info[1], | 
 |                                           sha_short) | 
 |  | 
 |             commit_msg_subject = 'Update %s to %s' % (desc.name, sha_short) | 
 |             intro = ('Update code from {}upstream repository {} to revision {}' | 
 |                      .format(('' if desc.upstream.only_subdir is None else | 
 |                               'subdir {} in '.format(desc.upstream.only_subdir)), | 
 |                              desc.upstream.url, | 
 |                              upstream_new_rev)) | 
 |             commit_msg_body = textwrap.fill(intro, width=70) | 
 |  | 
 |             if shortlog: | 
 |                 commit_msg_body += "\n\n" | 
 |                 commit_msg_body += format_list_to_str(shortlog, width=70) | 
 |  | 
 |             commit_msg = commit_msg_subject + "\n\n" + commit_msg_body | 
 |  | 
 |             commit_paths = [] | 
 |             commit_paths.append(desc.target_dir) | 
 |             if args.refresh_patches: | 
 |                 commit_paths.append(desc.patch_dir) | 
 |             commit_paths.append(lock_file_path) | 
 |  | 
 |             git_add_commit(commit_paths, commit_msg) | 
 |  | 
 |     log.info('Import finished') | 
 |  | 
 |  | 
 | if __name__ == '__main__': | 
 |     try: | 
 |         main(sys.argv) | 
 |     except subprocess.CalledProcessError as e: | 
 |         log.fatal("Called program '%s' returned with %d.\n" | 
 |                   "STDOUT:\n%s\n" | 
 |                   "STDERR:\n%s\n" % | 
 |                   (" ".join(e.cmd), e.returncode, e.stdout, e.stderr)) | 
 |         raise | 
 |     except KeyboardInterrupt: | 
 |         log.info("Aborting operation on user request.") | 
 |         sys.exit(1) |