|  | #!/usr/bin/env python3 | 
|  | # | 
|  | # SPDX-License-Identifier: GPL-2.0-only | 
|  | # | 
|  |  | 
|  | import os | 
|  | import sys | 
|  | import shutil | 
|  | import errno | 
|  | import time | 
|  |  | 
|  | def mkdir(d): | 
|  | try: | 
|  | os.makedirs(d) | 
|  | except OSError as e: | 
|  | if e.errno != errno.EEXIST: | 
|  | raise e | 
|  |  | 
|  | # extract the hash from past the last colon to last underscore | 
|  | def extract_sha(filename): | 
|  | return filename.split(':')[7].split('_')[0] | 
|  |  | 
|  | # get all files in a directory, extract hash and make | 
|  | # a map from hash to list of file with that hash | 
|  | def map_sha_to_files(dir_, prefix, sha_map): | 
|  | sstate_prefix_path = dir_ + '/' + prefix + '/' | 
|  | if not os.path.exists(sstate_prefix_path): | 
|  | return | 
|  | sstate_files = os.listdir(sstate_prefix_path) | 
|  | for f in sstate_files: | 
|  | try: | 
|  | sha = extract_sha(f) | 
|  | if sha not in sha_map: | 
|  | sha_map[sha] = [] | 
|  | sha_map[sha].append(sstate_prefix_path + f) | 
|  | except IndexError: | 
|  | continue | 
|  |  | 
|  | # given a prefix build a map of hash to list of files | 
|  | def build_sha_cache(prefix): | 
|  | sha_map = {} | 
|  |  | 
|  | sstate_dir = sys.argv[2] | 
|  | map_sha_to_files(sstate_dir, prefix, sha_map) | 
|  |  | 
|  | native_sstate_dir = sys.argv[2] + '/' + sys.argv[4] | 
|  | map_sha_to_files(native_sstate_dir, prefix, sha_map) | 
|  |  | 
|  | return sha_map | 
|  |  | 
|  | if len(sys.argv) < 5: | 
|  | print("Incorrect number of arguments specified") | 
|  | print("syntax: gen-lockedsig-cache <locked-sigs.inc> <input-cachedir> <output-cachedir> <nativelsbstring> [filterfile]") | 
|  | sys.exit(1) | 
|  |  | 
|  | filterlist = [] | 
|  | if len(sys.argv) > 5: | 
|  | print('Reading filter file %s' % sys.argv[5]) | 
|  | with open(sys.argv[5]) as f: | 
|  | for l in f.readlines(): | 
|  | if ":" in l: | 
|  | filterlist.append(l.rstrip()) | 
|  |  | 
|  | print('Reading %s' % sys.argv[1]) | 
|  | sigs = [] | 
|  | with open(sys.argv[1]) as f: | 
|  | for l in f.readlines(): | 
|  | if ":" in l: | 
|  | task, sig = l.split()[0].rsplit(':', 1) | 
|  | if filterlist and not task in filterlist: | 
|  | print('Filtering out %s' % task) | 
|  | else: | 
|  | sigs.append(sig) | 
|  |  | 
|  | print('Gathering file list') | 
|  | start_time = time.perf_counter() | 
|  | files = set() | 
|  | sstate_content_cache = {} | 
|  | for s in sigs: | 
|  | prefix = s[:2] | 
|  | prefix2 = s[2:4] | 
|  | if prefix not in sstate_content_cache: | 
|  | sstate_content_cache[prefix] = {} | 
|  | if prefix2 not in sstate_content_cache[prefix]: | 
|  | sstate_content_cache[prefix][prefix2] = build_sha_cache(prefix + "/" + prefix2) | 
|  |  | 
|  | if s in sstate_content_cache[prefix][prefix2]: | 
|  | for f in sstate_content_cache[prefix][prefix2][s]: | 
|  | files.add(f) | 
|  |  | 
|  | elapsed = time.perf_counter() - start_time | 
|  | print("Gathering file list took %.1fs" % elapsed) | 
|  |  | 
|  | print('Processing files') | 
|  | for f in files: | 
|  | sys.stdout.write('Processing %s... ' % f) | 
|  | _, ext = os.path.splitext(f) | 
|  | if not ext in ['.tgz', '.siginfo', '.sig']: | 
|  | # Most likely a temp file, skip it | 
|  | print('skipping') | 
|  | continue | 
|  | dst = os.path.join(sys.argv[3], os.path.relpath(f, sys.argv[2])) | 
|  | destdir = os.path.dirname(dst) | 
|  | mkdir(destdir) | 
|  |  | 
|  | src = os.path.realpath(f) | 
|  | if os.path.exists(dst): | 
|  | os.remove(dst) | 
|  | if (os.stat(src).st_dev == os.stat(destdir).st_dev): | 
|  | print('linking') | 
|  | try: | 
|  | os.link(src, dst) | 
|  | except OSError as e: | 
|  | print('hard linking failed, copying') | 
|  | shutil.copyfile(src, dst) | 
|  | else: | 
|  | print('copying') | 
|  | shutil.copyfile(src, dst) | 
|  |  | 
|  | print('Done!') |