blob: 2a0ecf57e16ed44f0331bddae318ec0e011d0536 [file] [log] [blame]
Brad Bishopc342db32019-05-15 21:57:59 -04001#
Patrick Williams92b42cb2022-09-03 06:53:57 -05002# Copyright BitBake Contributors
3#
Brad Bishopc342db32019-05-15 21:57:59 -04004# SPDX-License-Identifier: GPL-2.0-only
5#
6
Patrick Williamsc124f4f2015-09-15 14:41:29 -05007import hashlib
8import logging
9import os
10import re
11import tempfile
Patrick Williamsc0f7c042017-02-23 20:41:17 -060012import pickle
Patrick Williamsc124f4f2015-09-15 14:41:29 -050013import bb.data
Brad Bishop6e60e8b2018-02-01 10:27:11 -050014import difflib
15import simplediff
Andrew Geisslereff27472021-10-29 15:35:00 -050016import json
Andrew Geissler517393d2023-01-13 08:55:19 -060017import types
Andrew Geisslereff27472021-10-29 15:35:00 -050018import bb.compress.zstd
Patrick Williamsd8c66bc2016-06-20 12:57:21 -050019from bb.checksum import FileChecksumCache
Brad Bishop08902b02019-08-20 09:16:51 -040020from bb import runqueue
Brad Bishopa34c0302019-09-23 22:34:48 -040021import hashserv
Andrew Geissler475cb722020-07-10 16:00:51 -050022import hashserv.client
Patrick Williamsc124f4f2015-09-15 14:41:29 -050023
24logger = logging.getLogger('BitBake.SigGen')
Andrew Geissler82c905d2020-04-13 13:39:40 -050025hashequiv_logger = logging.getLogger('BitBake.SigGen.HashEquiv')
Patrick Williamsc124f4f2015-09-15 14:41:29 -050026
Patrick Williams169d7bc2024-01-05 11:33:25 -060027#find_siginfo and find_siginfo_version are set by the metadata siggen
28# The minimum version of the find_siginfo function we need
29find_siginfo_minversion = 2
30
31def check_siggen_version(siggen):
32 if not hasattr(siggen, "find_siginfo_version"):
33 bb.fatal("Siggen from metadata (OE-Core?) is too old, please update it (no version found)")
34 if siggen.find_siginfo_version < siggen.find_siginfo_minversion:
35 bb.fatal("Siggen from metadata (OE-Core?) is too old, please update it (%s vs %s)" % (siggen.find_siginfo_version, siggen.find_siginfo_minversion))
36
Andrew Geisslereff27472021-10-29 15:35:00 -050037class SetEncoder(json.JSONEncoder):
38 def default(self, obj):
Andrew Geissler517393d2023-01-13 08:55:19 -060039 if isinstance(obj, set) or isinstance(obj, frozenset):
Andrew Geisslereff27472021-10-29 15:35:00 -050040 return dict(_set_object=list(sorted(obj)))
41 return json.JSONEncoder.default(self, obj)
42
43def SetDecoder(dct):
44 if '_set_object' in dct:
Andrew Geissler517393d2023-01-13 08:55:19 -060045 return frozenset(dct['_set_object'])
Andrew Geisslereff27472021-10-29 15:35:00 -050046 return dct
47
Patrick Williamsc124f4f2015-09-15 14:41:29 -050048def init(d):
Patrick Williamsc0f7c042017-02-23 20:41:17 -060049 siggens = [obj for obj in globals().values()
Patrick Williamsc124f4f2015-09-15 14:41:29 -050050 if type(obj) is type and issubclass(obj, SignatureGenerator)]
51
Brad Bishop6e60e8b2018-02-01 10:27:11 -050052 desired = d.getVar("BB_SIGNATURE_HANDLER") or "noop"
Patrick Williamsc124f4f2015-09-15 14:41:29 -050053 for sg in siggens:
54 if desired == sg.name:
55 return sg(d)
Patrick Williamsc124f4f2015-09-15 14:41:29 -050056 else:
57 logger.error("Invalid signature generator '%s', using default 'noop'\n"
58 "Available generators: %s", desired,
59 ', '.join(obj.name for obj in siggens))
60 return SignatureGenerator(d)
61
62class SignatureGenerator(object):
63 """
64 """
65 name = "noop"
66
67 def __init__(self, data):
Brad Bishop37a0e4d2017-12-04 01:01:44 -050068 self.basehash = {}
Patrick Williamsc124f4f2015-09-15 14:41:29 -050069 self.taskhash = {}
Andrew Geissler82c905d2020-04-13 13:39:40 -050070 self.unihash = {}
Patrick Williamsc124f4f2015-09-15 14:41:29 -050071 self.runtaskdeps = {}
72 self.file_checksum_values = {}
Patrick Williamsd8c66bc2016-06-20 12:57:21 -050073 self.taints = {}
Brad Bishop08902b02019-08-20 09:16:51 -040074 self.unitaskhashes = {}
Andrew Geissler82c905d2020-04-13 13:39:40 -050075 self.tidtopn = {}
76 self.setscenetasks = set()
Patrick Williamsc124f4f2015-09-15 14:41:29 -050077
78 def finalise(self, fn, d, varient):
79 return
80
Andrew Geissler82c905d2020-04-13 13:39:40 -050081 def postparsing_clean_cache(self):
82 return
83
Andrew Geissler517393d2023-01-13 08:55:19 -060084 def setup_datacache(self, datacaches):
85 self.datacaches = datacaches
86
87 def setup_datacache_from_datastore(self, mcfn, d):
88 # In task context we have no cache so setup internal data structures
89 # from the fully parsed data store provided
90
91 mc = d.getVar("__BBMULTICONFIG", False) or ""
92 tasks = d.getVar('__BBTASKS', False)
93
94 self.datacaches = {}
95 self.datacaches[mc] = types.SimpleNamespace()
96 setattr(self.datacaches[mc], "stamp", {})
97 self.datacaches[mc].stamp[mcfn] = d.getVar('STAMP')
98 setattr(self.datacaches[mc], "stamp_extrainfo", {})
99 self.datacaches[mc].stamp_extrainfo[mcfn] = {}
100 for t in tasks:
101 flag = d.getVarFlag(t, "stamp-extra-info")
102 if flag:
103 self.datacaches[mc].stamp_extrainfo[mcfn][t] = flag
104
Patrick Williams73bd93f2024-02-20 08:07:48 -0600105 def get_cached_unihash(self, tid):
106 return None
107
Brad Bishop08902b02019-08-20 09:16:51 -0400108 def get_unihash(self, tid):
Patrick Williams73bd93f2024-02-20 08:07:48 -0600109 unihash = self.get_cached_unihash(tid)
110 if unihash:
111 return unihash
Brad Bishop08902b02019-08-20 09:16:51 -0400112 return self.taskhash[tid]
Brad Bishop19323692019-04-05 15:28:33 -0400113
Patrick Williams73bd93f2024-02-20 08:07:48 -0600114 def get_unihashes(self, tids):
115 return {tid: self.get_unihash(tid) for tid in tids}
116
Andrew Geissler5a43b432020-06-13 10:46:56 -0500117 def prep_taskhash(self, tid, deps, dataCaches):
Andrew Geissler82c905d2020-04-13 13:39:40 -0500118 return
119
Andrew Geissler5a43b432020-06-13 10:46:56 -0500120 def get_taskhash(self, tid, deps, dataCaches):
Brad Bishop08902b02019-08-20 09:16:51 -0400121 self.taskhash[tid] = hashlib.sha256(tid.encode("utf-8")).hexdigest()
122 return self.taskhash[tid]
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500123
Patrick Williamsd8c66bc2016-06-20 12:57:21 -0500124 def writeout_file_checksum_cache(self):
125 """Write/update the file checksum cache onto disk"""
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500126 return
127
Andrew Geissler517393d2023-01-13 08:55:19 -0600128 def stampfile_base(self, mcfn):
129 mc = bb.runqueue.mc_from_tid(mcfn)
130 return self.datacaches[mc].stamp[mcfn]
131
132 def stampfile_mcfn(self, taskname, mcfn, extrainfo=True):
133 mc = bb.runqueue.mc_from_tid(mcfn)
134 stamp = self.datacaches[mc].stamp[mcfn]
135 if not stamp:
136 return
137
138 stamp_extrainfo = ""
139 if extrainfo:
140 taskflagname = taskname
141 if taskname.endswith("_setscene"):
142 taskflagname = taskname.replace("_setscene", "")
143 stamp_extrainfo = self.datacaches[mc].stamp_extrainfo[mcfn].get(taskflagname) or ""
144
145 return self.stampfile(stamp, mcfn, taskname, stamp_extrainfo)
146
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500147 def stampfile(self, stampbase, file_name, taskname, extrainfo):
148 return ("%s.%s.%s" % (stampbase, taskname, extrainfo)).rstrip('.')
149
Andrew Geissler517393d2023-01-13 08:55:19 -0600150 def stampcleanmask_mcfn(self, taskname, mcfn):
151 mc = bb.runqueue.mc_from_tid(mcfn)
152 stamp = self.datacaches[mc].stamp[mcfn]
153 if not stamp:
154 return []
155
156 taskflagname = taskname
157 if taskname.endswith("_setscene"):
158 taskflagname = taskname.replace("_setscene", "")
159 stamp_extrainfo = self.datacaches[mc].stamp_extrainfo[mcfn].get(taskflagname) or ""
160
161 return self.stampcleanmask(stamp, mcfn, taskname, stamp_extrainfo)
162
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500163 def stampcleanmask(self, stampbase, file_name, taskname, extrainfo):
164 return ("%s.%s.%s" % (stampbase, taskname, extrainfo)).rstrip('.')
165
Andrew Geissler517393d2023-01-13 08:55:19 -0600166 def dump_sigtask(self, mcfn, task, stampbase, runtime):
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500167 return
168
Andrew Geissler517393d2023-01-13 08:55:19 -0600169 def invalidate_task(self, task, mcfn):
170 mc = bb.runqueue.mc_from_tid(mcfn)
171 stamp = self.datacaches[mc].stamp[mcfn]
172 bb.utils.remove(stamp)
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500173
174 def dump_sigs(self, dataCache, options):
175 return
176
177 def get_taskdata(self):
Andrew Geissler82c905d2020-04-13 13:39:40 -0500178 return (self.runtaskdeps, self.taskhash, self.unihash, self.file_checksum_values, self.taints, self.basehash, self.unitaskhashes, self.tidtopn, self.setscenetasks)
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500179
180 def set_taskdata(self, data):
Andrew Geissler82c905d2020-04-13 13:39:40 -0500181 self.runtaskdeps, self.taskhash, self.unihash, self.file_checksum_values, self.taints, self.basehash, self.unitaskhashes, self.tidtopn, self.setscenetasks = data
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500182
Brad Bishopd7bf8c12018-02-25 22:55:05 -0500183 def reset(self, data):
184 self.__init__(data)
185
Brad Bishop08902b02019-08-20 09:16:51 -0400186 def get_taskhashes(self):
Andrew Geissler82c905d2020-04-13 13:39:40 -0500187 return self.taskhash, self.unihash, self.unitaskhashes, self.tidtopn
Brad Bishop08902b02019-08-20 09:16:51 -0400188
189 def set_taskhashes(self, hashes):
Andrew Geissler82c905d2020-04-13 13:39:40 -0500190 self.taskhash, self.unihash, self.unitaskhashes, self.tidtopn = hashes
Brad Bishop08902b02019-08-20 09:16:51 -0400191
192 def save_unitaskhashes(self):
193 return
194
Andrew Geissler78b72792022-06-14 06:47:25 -0500195 def copy_unitaskhashes(self, targetdir):
196 return
197
Brad Bishopa34c0302019-09-23 22:34:48 -0400198 def set_setscene_tasks(self, setscene_tasks):
199 return
Brad Bishopd7bf8c12018-02-25 22:55:05 -0500200
Andrew Geissler9aee5002022-03-30 16:27:02 +0000201 def exit(self):
202 return
203
Patrick Williams2a254922023-08-11 09:48:11 -0500204def build_pnid(mc, pn, taskname):
205 if mc:
206 return "mc:" + mc + ":" + pn + ":" + taskname
207 return pn + ":" + taskname
208
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500209class SignatureGeneratorBasic(SignatureGenerator):
210 """
211 """
212 name = "basic"
213
214 def __init__(self, data):
215 self.basehash = {}
216 self.taskhash = {}
Andrew Geissler82c905d2020-04-13 13:39:40 -0500217 self.unihash = {}
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500218 self.runtaskdeps = {}
219 self.file_checksum_values = {}
Patrick Williamsf1e5d692016-03-30 15:21:19 -0500220 self.taints = {}
Andrew Geissler82c905d2020-04-13 13:39:40 -0500221 self.setscenetasks = set()
Andrew Geissler7e0e3c02022-02-25 20:34:39 +0000222 self.basehash_ignore_vars = set((data.getVar("BB_BASEHASH_IGNORE_VARS") or "").split())
223 self.taskhash_ignore_tasks = None
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500224 self.init_rundepcheck(data)
Brad Bishop6e60e8b2018-02-01 10:27:11 -0500225 checksum_cache_file = data.getVar("BB_HASH_CHECKSUM_CACHE_FILE")
Patrick Williamsd8c66bc2016-06-20 12:57:21 -0500226 if checksum_cache_file:
227 self.checksum_cache = FileChecksumCache()
228 self.checksum_cache.init_cache(data, checksum_cache_file)
229 else:
230 self.checksum_cache = None
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500231
Andrew Geissler82c905d2020-04-13 13:39:40 -0500232 self.unihash_cache = bb.cache.SimpleCache("3")
Brad Bishop08902b02019-08-20 09:16:51 -0400233 self.unitaskhashes = self.unihash_cache.init_cache(data, "bb_unihashes.dat", {})
Andrew Geissler82c905d2020-04-13 13:39:40 -0500234 self.localdirsexclude = (data.getVar("BB_SIGNATURE_LOCAL_DIRS_EXCLUDE") or "CVS .bzr .git .hg .osc .p4 .repo .svn").split()
235 self.tidtopn = {}
Brad Bishop08902b02019-08-20 09:16:51 -0400236
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500237 def init_rundepcheck(self, data):
Andrew Geissler7e0e3c02022-02-25 20:34:39 +0000238 self.taskhash_ignore_tasks = data.getVar("BB_TASKHASH_IGNORE_TASKS") or None
239 if self.taskhash_ignore_tasks:
240 self.twl = re.compile(self.taskhash_ignore_tasks)
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500241 else:
242 self.twl = None
243
Andrew Geissler517393d2023-01-13 08:55:19 -0600244 def _build_data(self, mcfn, d):
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500245
Brad Bishop6e60e8b2018-02-01 10:27:11 -0500246 ignore_mismatch = ((d.getVar("BB_HASH_IGNORE_MISMATCH") or '') == '1')
Andrew Geissler7e0e3c02022-02-25 20:34:39 +0000247 tasklist, gendeps, lookupcache = bb.data.generate_dependencies(d, self.basehash_ignore_vars)
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500248
Andrew Geissler517393d2023-01-13 08:55:19 -0600249 taskdeps, basehash = bb.data.generate_dependency_hash(tasklist, gendeps, lookupcache, self.basehash_ignore_vars, mcfn)
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500250
251 for task in tasklist:
Andrew Geissler517393d2023-01-13 08:55:19 -0600252 tid = mcfn + ":" + task
Brad Bishop08902b02019-08-20 09:16:51 -0400253 if not ignore_mismatch and tid in self.basehash and self.basehash[tid] != basehash[tid]:
254 bb.error("When reparsing %s, the basehash value changed from %s to %s. The metadata is not deterministic and this needs to be fixed." % (tid, self.basehash[tid], basehash[tid]))
Brad Bishopc342db32019-05-15 21:57:59 -0400255 bb.error("The following commands may help:")
256 cmd = "$ bitbake %s -c%s" % (d.getVar('PN'), task)
257 # Make sure sigdata is dumped before run printdiff
258 bb.error("%s -Snone" % cmd)
259 bb.error("Then:")
260 bb.error("%s -Sprintdiff\n" % cmd)
Brad Bishop08902b02019-08-20 09:16:51 -0400261 self.basehash[tid] = basehash[tid]
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500262
Andrew Geissler517393d2023-01-13 08:55:19 -0600263 return taskdeps, gendeps, lookupcache
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500264
Brad Bishopa34c0302019-09-23 22:34:48 -0400265 def set_setscene_tasks(self, setscene_tasks):
Andrew Geissler82c905d2020-04-13 13:39:40 -0500266 self.setscenetasks = set(setscene_tasks)
Brad Bishopa34c0302019-09-23 22:34:48 -0400267
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500268 def finalise(self, fn, d, variant):
269
Patrick Williamsc0f7c042017-02-23 20:41:17 -0600270 mc = d.getVar("__BBMULTICONFIG", False) or ""
Andrew Geissler517393d2023-01-13 08:55:19 -0600271 mcfn = fn
Patrick Williamsc0f7c042017-02-23 20:41:17 -0600272 if variant or mc:
Andrew Geissler517393d2023-01-13 08:55:19 -0600273 mcfn = bb.cache.realfn2virtual(fn, variant, mc)
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500274
275 try:
Andrew Geissler517393d2023-01-13 08:55:19 -0600276 taskdeps, gendeps, lookupcache = self._build_data(mcfn, d)
Brad Bishop6e60e8b2018-02-01 10:27:11 -0500277 except bb.parse.SkipRecipe:
278 raise
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500279 except:
Andrew Geissler517393d2023-01-13 08:55:19 -0600280 bb.warn("Error during finalise of %s" % mcfn)
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500281 raise
282
Andrew Geissler517393d2023-01-13 08:55:19 -0600283 basehashes = {}
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500284 for task in taskdeps:
Andrew Geissler517393d2023-01-13 08:55:19 -0600285 basehashes[task] = self.basehash[mcfn + ":" + task]
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500286
Andrew Geissler517393d2023-01-13 08:55:19 -0600287 d.setVar("__siggen_basehashes", basehashes)
288 d.setVar("__siggen_gendeps", gendeps)
289 d.setVar("__siggen_varvals", lookupcache)
290 d.setVar("__siggen_taskdeps", taskdeps)
291
Andrew Geissler5082cc72023-09-11 08:41:39 -0400292 #Slow but can be useful for debugging mismatched basehashes
293 #self.setup_datacache_from_datastore(mcfn, d)
294 #for task in taskdeps:
295 # self.dump_sigtask(mcfn, task, d.getVar("STAMP"), False)
296
Andrew Geissler517393d2023-01-13 08:55:19 -0600297 def setup_datacache_from_datastore(self, mcfn, d):
298 super().setup_datacache_from_datastore(mcfn, d)
299
300 mc = bb.runqueue.mc_from_tid(mcfn)
301 for attr in ["siggen_varvals", "siggen_taskdeps", "siggen_gendeps"]:
302 if not hasattr(self.datacaches[mc], attr):
303 setattr(self.datacaches[mc], attr, {})
304 self.datacaches[mc].siggen_varvals[mcfn] = d.getVar("__siggen_varvals")
305 self.datacaches[mc].siggen_taskdeps[mcfn] = d.getVar("__siggen_taskdeps")
306 self.datacaches[mc].siggen_gendeps[mcfn] = d.getVar("__siggen_gendeps")
Andrew Geissler82c905d2020-04-13 13:39:40 -0500307
Andrew Geissler5a43b432020-06-13 10:46:56 -0500308 def rundep_check(self, fn, recipename, task, dep, depname, dataCaches):
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500309 # Return True if we should keep the dependency, False to drop it
Andrew Geissler7e0e3c02022-02-25 20:34:39 +0000310 # We only manipulate the dependencies for packages not in the ignore
311 # list
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500312 if self.twl and not self.twl.search(recipename):
313 # then process the actual dependencies
314 if self.twl.search(depname):
315 return False
316 return True
317
318 def read_taint(self, fn, task, stampbase):
319 taint = None
320 try:
321 with open(stampbase + '.' + task + '.taint', 'r') as taintf:
322 taint = taintf.read()
323 except IOError:
324 pass
325 return taint
326
Andrew Geissler5a43b432020-06-13 10:46:56 -0500327 def prep_taskhash(self, tid, deps, dataCaches):
Brad Bishop1a4b7ee2018-12-16 17:11:34 -0800328
Andrew Geissler517393d2023-01-13 08:55:19 -0600329 (mc, _, task, mcfn) = bb.runqueue.split_tid_mcfn(tid)
Brad Bishop1a4b7ee2018-12-16 17:11:34 -0800330
Andrew Geissler5a43b432020-06-13 10:46:56 -0500331 self.basehash[tid] = dataCaches[mc].basetaskhash[tid]
Brad Bishop08902b02019-08-20 09:16:51 -0400332 self.runtaskdeps[tid] = []
333 self.file_checksum_values[tid] = []
Andrew Geissler517393d2023-01-13 08:55:19 -0600334 recipename = dataCaches[mc].pkg_fn[mcfn]
Andrew Geissler82c905d2020-04-13 13:39:40 -0500335
336 self.tidtopn[tid] = recipename
Patrick Williams2a254922023-08-11 09:48:11 -0500337 # save hashfn for deps into siginfo?
338 for dep in deps:
339 (depmc, _, deptask, depmcfn) = bb.runqueue.split_tid_mcfn(dep)
340 dep_pn = dataCaches[depmc].pkg_fn[depmcfn]
Andrew Geissler82c905d2020-04-13 13:39:40 -0500341
Patrick Williams2a254922023-08-11 09:48:11 -0500342 if not self.rundep_check(mcfn, recipename, task, dep, dep_pn, dataCaches):
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500343 continue
Patrick Williams2a254922023-08-11 09:48:11 -0500344
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500345 if dep not in self.taskhash:
Brad Bishop1a4b7ee2018-12-16 17:11:34 -0800346 bb.fatal("%s is not in taskhash, caller isn't calling in dependency order?" % dep)
Patrick Williams2a254922023-08-11 09:48:11 -0500347
348 dep_pnid = build_pnid(depmc, dep_pn, deptask)
349 self.runtaskdeps[tid].append((dep_pnid, dep))
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500350
Andrew Geissler517393d2023-01-13 08:55:19 -0600351 if task in dataCaches[mc].file_checksums[mcfn]:
Patrick Williamsd8c66bc2016-06-20 12:57:21 -0500352 if self.checksum_cache:
Andrew Geissler517393d2023-01-13 08:55:19 -0600353 checksums = self.checksum_cache.get_checksums(dataCaches[mc].file_checksums[mcfn][task], recipename, self.localdirsexclude)
Patrick Williamsd8c66bc2016-06-20 12:57:21 -0500354 else:
Andrew Geissler517393d2023-01-13 08:55:19 -0600355 checksums = bb.fetch2.get_file_checksums(dataCaches[mc].file_checksums[mcfn][task], recipename, self.localdirsexclude)
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500356 for (f,cs) in checksums:
Brad Bishop08902b02019-08-20 09:16:51 -0400357 self.file_checksum_values[tid].append((f,cs))
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500358
Andrew Geissler517393d2023-01-13 08:55:19 -0600359 taskdep = dataCaches[mc].task_deps[mcfn]
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500360 if 'nostamp' in taskdep and task in taskdep['nostamp']:
361 # Nostamp tasks need an implicit taint so that they force any dependent tasks to run
Andrew Geissler82c905d2020-04-13 13:39:40 -0500362 if tid in self.taints and self.taints[tid].startswith("nostamp:"):
363 # Don't reset taint value upon every call
364 pass
365 else:
366 import uuid
367 taint = str(uuid.uuid4())
368 self.taints[tid] = "nostamp:" + taint
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500369
Andrew Geissler517393d2023-01-13 08:55:19 -0600370 taint = self.read_taint(mcfn, task, dataCaches[mc].stamp[mcfn])
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500371 if taint:
Brad Bishop08902b02019-08-20 09:16:51 -0400372 self.taints[tid] = taint
373 logger.warning("%s is tainted from a forced run" % tid)
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500374
Andrew Geissler82c905d2020-04-13 13:39:40 -0500375 return
376
Andrew Geissler5a43b432020-06-13 10:46:56 -0500377 def get_taskhash(self, tid, deps, dataCaches):
Andrew Geissler82c905d2020-04-13 13:39:40 -0500378
379 data = self.basehash[tid]
Patrick Williams2a254922023-08-11 09:48:11 -0500380 for dep in sorted(self.runtaskdeps[tid]):
381 data += self.get_unihash(dep[1])
Andrew Geissler82c905d2020-04-13 13:39:40 -0500382
Andrew Geissler5082cc72023-09-11 08:41:39 -0400383 for (f, cs) in sorted(self.file_checksum_values[tid], key=clean_checksum_file_path):
Andrew Geissler82c905d2020-04-13 13:39:40 -0500384 if cs:
Andrew Geissler595f6302022-01-24 19:11:47 +0000385 if "/./" in f:
Patrick Williams864cc432023-02-09 14:54:44 -0600386 data += "./" + f.split("/./")[1]
387 data += cs
Andrew Geissler82c905d2020-04-13 13:39:40 -0500388
389 if tid in self.taints:
390 if self.taints[tid].startswith("nostamp:"):
Patrick Williams864cc432023-02-09 14:54:44 -0600391 data += self.taints[tid][8:]
Andrew Geissler82c905d2020-04-13 13:39:40 -0500392 else:
Patrick Williams864cc432023-02-09 14:54:44 -0600393 data += self.taints[tid]
Andrew Geissler82c905d2020-04-13 13:39:40 -0500394
Brad Bishop19323692019-04-05 15:28:33 -0400395 h = hashlib.sha256(data.encode("utf-8")).hexdigest()
Brad Bishop08902b02019-08-20 09:16:51 -0400396 self.taskhash[tid] = h
Patrick Williams213cb262021-08-07 19:21:33 -0500397 #d.setVar("BB_TASKHASH:task-%s" % task, taskhash[task])
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500398 return h
399
Patrick Williamsd8c66bc2016-06-20 12:57:21 -0500400 def writeout_file_checksum_cache(self):
401 """Write/update the file checksum cache onto disk"""
402 if self.checksum_cache:
403 self.checksum_cache.save_extras()
404 self.checksum_cache.save_merge()
405 else:
406 bb.fetch2.fetcher_parse_save()
407 bb.fetch2.fetcher_parse_done()
408
Brad Bishop08902b02019-08-20 09:16:51 -0400409 def save_unitaskhashes(self):
410 self.unihash_cache.save(self.unitaskhashes)
411
Andrew Geissler78b72792022-06-14 06:47:25 -0500412 def copy_unitaskhashes(self, targetdir):
413 self.unihash_cache.copyfile(targetdir)
414
Andrew Geissler517393d2023-01-13 08:55:19 -0600415 def dump_sigtask(self, mcfn, task, stampbase, runtime):
416 tid = mcfn + ":" + task
417 mc = bb.runqueue.mc_from_tid(mcfn)
Patrick Williamsd8c66bc2016-06-20 12:57:21 -0500418 referencestamp = stampbase
419 if isinstance(runtime, str) and runtime.startswith("customfile"):
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500420 sigfile = stampbase
Patrick Williamsd8c66bc2016-06-20 12:57:21 -0500421 referencestamp = runtime[11:]
Brad Bishop08902b02019-08-20 09:16:51 -0400422 elif runtime and tid in self.taskhash:
Brad Bishop00e122a2019-10-05 11:10:57 -0400423 sigfile = stampbase + "." + task + ".sigdata" + "." + self.get_unihash(tid)
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500424 else:
Brad Bishop08902b02019-08-20 09:16:51 -0400425 sigfile = stampbase + "." + task + ".sigbasedata" + "." + self.basehash[tid]
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500426
Andrew Geisslerc3d88e42020-10-02 09:45:00 -0500427 with bb.utils.umask(0o002):
428 bb.utils.mkdirhier(os.path.dirname(sigfile))
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500429
430 data = {}
Patrick Williamsd8c66bc2016-06-20 12:57:21 -0500431 data['task'] = task
Andrew Geissler7e0e3c02022-02-25 20:34:39 +0000432 data['basehash_ignore_vars'] = self.basehash_ignore_vars
433 data['taskhash_ignore_tasks'] = self.taskhash_ignore_tasks
Andrew Geissler517393d2023-01-13 08:55:19 -0600434 data['taskdeps'] = self.datacaches[mc].siggen_taskdeps[mcfn][task]
Brad Bishop08902b02019-08-20 09:16:51 -0400435 data['basehash'] = self.basehash[tid]
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500436 data['gendeps'] = {}
437 data['varvals'] = {}
Andrew Geissler517393d2023-01-13 08:55:19 -0600438 data['varvals'][task] = self.datacaches[mc].siggen_varvals[mcfn][task]
439 for dep in self.datacaches[mc].siggen_taskdeps[mcfn][task]:
Andrew Geissler7e0e3c02022-02-25 20:34:39 +0000440 if dep in self.basehash_ignore_vars:
Andrew Geissler5082cc72023-09-11 08:41:39 -0400441 continue
Andrew Geissler517393d2023-01-13 08:55:19 -0600442 data['gendeps'][dep] = self.datacaches[mc].siggen_gendeps[mcfn][dep]
443 data['varvals'][dep] = self.datacaches[mc].siggen_varvals[mcfn][dep]
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500444
Brad Bishop08902b02019-08-20 09:16:51 -0400445 if runtime and tid in self.taskhash:
Patrick Williams2a254922023-08-11 09:48:11 -0500446 data['runtaskdeps'] = [dep[0] for dep in sorted(self.runtaskdeps[tid])]
Andrew Geissler595f6302022-01-24 19:11:47 +0000447 data['file_checksum_values'] = []
Andrew Geissler5082cc72023-09-11 08:41:39 -0400448 for f,cs in sorted(self.file_checksum_values[tid], key=clean_checksum_file_path):
Andrew Geissler595f6302022-01-24 19:11:47 +0000449 if "/./" in f:
450 data['file_checksum_values'].append(("./" + f.split("/./")[1], cs))
451 else:
452 data['file_checksum_values'].append((os.path.basename(f), cs))
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500453 data['runtaskhashes'] = {}
Patrick Williams2a254922023-08-11 09:48:11 -0500454 for dep in self.runtaskdeps[tid]:
455 data['runtaskhashes'][dep[0]] = self.get_unihash(dep[1])
Brad Bishop08902b02019-08-20 09:16:51 -0400456 data['taskhash'] = self.taskhash[tid]
Brad Bishop00e122a2019-10-05 11:10:57 -0400457 data['unihash'] = self.get_unihash(tid)
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500458
Andrew Geissler517393d2023-01-13 08:55:19 -0600459 taint = self.read_taint(mcfn, task, referencestamp)
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500460 if taint:
461 data['taint'] = taint
462
Brad Bishop08902b02019-08-20 09:16:51 -0400463 if runtime and tid in self.taints:
464 if 'nostamp:' in self.taints[tid]:
465 data['taint'] = self.taints[tid]
Patrick Williamsf1e5d692016-03-30 15:21:19 -0500466
Brad Bishop37a0e4d2017-12-04 01:01:44 -0500467 computed_basehash = calc_basehash(data)
Brad Bishop08902b02019-08-20 09:16:51 -0400468 if computed_basehash != self.basehash[tid]:
469 bb.error("Basehash mismatch %s versus %s for %s" % (computed_basehash, self.basehash[tid], tid))
470 if runtime and tid in self.taskhash:
Brad Bishop37a0e4d2017-12-04 01:01:44 -0500471 computed_taskhash = calc_taskhash(data)
Brad Bishop08902b02019-08-20 09:16:51 -0400472 if computed_taskhash != self.taskhash[tid]:
473 bb.error("Taskhash mismatch %s versus %s for %s" % (computed_taskhash, self.taskhash[tid], tid))
474 sigfile = sigfile.replace(self.taskhash[tid], computed_taskhash)
Brad Bishop37a0e4d2017-12-04 01:01:44 -0500475
Patrick Williams92b42cb2022-09-03 06:53:57 -0500476 fd, tmpfile = bb.utils.mkstemp(dir=os.path.dirname(sigfile), prefix="sigtask.")
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500477 try:
Andrew Geisslereff27472021-10-29 15:35:00 -0500478 with bb.compress.zstd.open(fd, "wt", encoding="utf-8", num_threads=1) as f:
479 json.dump(data, f, sort_keys=True, separators=(",", ":"), cls=SetEncoder)
480 f.flush()
Patrick Williamsc0f7c042017-02-23 20:41:17 -0600481 os.chmod(tmpfile, 0o664)
Andrew Geisslerc926e172021-05-07 16:11:35 -0500482 bb.utils.rename(tmpfile, sigfile)
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500483 except (OSError, IOError) as err:
484 try:
485 os.unlink(tmpfile)
486 except OSError:
487 pass
488 raise err
489
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500490class SignatureGeneratorBasicHash(SignatureGeneratorBasic):
491 name = "basichash"
492
Brad Bishop08902b02019-08-20 09:16:51 -0400493 def get_stampfile_hash(self, tid):
494 if tid in self.taskhash:
495 return self.taskhash[tid]
Brad Bishop19323692019-04-05 15:28:33 -0400496
497 # If task is not in basehash, then error
Brad Bishop08902b02019-08-20 09:16:51 -0400498 return self.basehash[tid]
Brad Bishop19323692019-04-05 15:28:33 -0400499
Andrew Geissler517393d2023-01-13 08:55:19 -0600500 def stampfile(self, stampbase, mcfn, taskname, extrainfo, clean=False):
501 if taskname.endswith("_setscene"):
502 tid = mcfn + ":" + taskname[:-9]
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500503 else:
Andrew Geissler517393d2023-01-13 08:55:19 -0600504 tid = mcfn + ":" + taskname
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500505 if clean:
506 h = "*"
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500507 else:
Brad Bishop08902b02019-08-20 09:16:51 -0400508 h = self.get_stampfile_hash(tid)
Brad Bishop19323692019-04-05 15:28:33 -0400509
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500510 return ("%s.%s.%s.%s" % (stampbase, taskname, h, extrainfo)).rstrip('.')
511
Andrew Geissler517393d2023-01-13 08:55:19 -0600512 def stampcleanmask(self, stampbase, mcfn, taskname, extrainfo):
513 return self.stampfile(stampbase, mcfn, taskname, extrainfo, clean=True)
Brad Bishop1a4b7ee2018-12-16 17:11:34 -0800514
Andrew Geissler517393d2023-01-13 08:55:19 -0600515 def invalidate_task(self, task, mcfn):
516 bb.note("Tainting hash to force rebuild of task %s, %s" % (mcfn, task))
517
518 mc = bb.runqueue.mc_from_tid(mcfn)
519 stamp = self.datacaches[mc].stamp[mcfn]
520
521 taintfn = stamp + '.' + task + '.taint'
522
523 import uuid
524 bb.utils.mkdirhier(os.path.dirname(taintfn))
525 # The specific content of the taint file is not really important,
526 # we just need it to be random, so a random UUID is used
527 with open(taintfn, 'w') as taintf:
528 taintf.write(str(uuid.uuid4()))
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500529
Brad Bishop08902b02019-08-20 09:16:51 -0400530class SignatureGeneratorUniHashMixIn(object):
Andrew Geissler82c905d2020-04-13 13:39:40 -0500531 def __init__(self, data):
532 self.extramethod = {}
Patrick Williams73bd93f2024-02-20 08:07:48 -0600533 # NOTE: The cache only tracks hashes that exist. Hashes that don't
534 # exist are always queries from the server since it is possible for
535 # hashes to appear over time, but much less likely for them to
536 # disappear
537 self.unihash_exists_cache = set()
Patrick Williams03514f12024-04-05 07:04:11 -0500538 self.username = None
539 self.password = None
Andrew Geissler82c905d2020-04-13 13:39:40 -0500540 super().__init__(data)
541
Brad Bishop08902b02019-08-20 09:16:51 -0400542 def get_taskdata(self):
Patrick Williams03514f12024-04-05 07:04:11 -0500543 return (self.server, self.method, self.extramethod, self.max_parallel, self.username, self.password) + super().get_taskdata()
Brad Bishop08902b02019-08-20 09:16:51 -0400544
545 def set_taskdata(self, data):
Patrick Williams03514f12024-04-05 07:04:11 -0500546 self.server, self.method, self.extramethod, self.max_parallel, self.username, self.password = data[:6]
547 super().set_taskdata(data[6:])
548
549 def get_hashserv_creds(self):
550 if self.username and self.password:
551 return {
552 "username": self.username,
553 "password": self.password,
554 }
555
556 return {}
Brad Bishop08902b02019-08-20 09:16:51 -0400557
Brad Bishopa34c0302019-09-23 22:34:48 -0400558 def client(self):
559 if getattr(self, '_client', None) is None:
Patrick Williams03514f12024-04-05 07:04:11 -0500560 self._client = hashserv.create_client(self.server, **self.get_hashserv_creds())
Brad Bishopa34c0302019-09-23 22:34:48 -0400561 return self._client
562
Patrick Williams73bd93f2024-02-20 08:07:48 -0600563 def client_pool(self):
564 if getattr(self, '_client_pool', None) is None:
Patrick Williams03514f12024-04-05 07:04:11 -0500565 self._client_pool = hashserv.client.ClientPool(self.server, self.max_parallel, **self.get_hashserv_creds())
Patrick Williams73bd93f2024-02-20 08:07:48 -0600566 return self._client_pool
567
Andrew Geissler9aee5002022-03-30 16:27:02 +0000568 def reset(self, data):
Patrick Williams73bd93f2024-02-20 08:07:48 -0600569 self.__close_clients()
Andrew Geissler9aee5002022-03-30 16:27:02 +0000570 return super().reset(data)
571
572 def exit(self):
Patrick Williams73bd93f2024-02-20 08:07:48 -0600573 self.__close_clients()
574 return super().exit()
575
576 def __close_clients(self):
Andrew Geissler9aee5002022-03-30 16:27:02 +0000577 if getattr(self, '_client', None) is not None:
578 self._client.close()
579 self._client = None
Patrick Williams73bd93f2024-02-20 08:07:48 -0600580 if getattr(self, '_client_pool', None) is not None:
581 self._client_pool.close()
582 self._client_pool = None
Andrew Geissler9aee5002022-03-30 16:27:02 +0000583
Brad Bishop08902b02019-08-20 09:16:51 -0400584 def get_stampfile_hash(self, tid):
585 if tid in self.taskhash:
586 # If a unique hash is reported, use it as the stampfile hash. This
587 # ensures that if a task won't be re-run if the taskhash changes,
588 # but it would result in the same output hash
Andrew Geissler82c905d2020-04-13 13:39:40 -0500589 unihash = self._get_unihash(tid)
Brad Bishop08902b02019-08-20 09:16:51 -0400590 if unihash is not None:
591 return unihash
592
593 return super().get_stampfile_hash(tid)
594
595 def set_unihash(self, tid, unihash):
Andrew Geissler82c905d2020-04-13 13:39:40 -0500596 (mc, fn, taskname, taskfn) = bb.runqueue.split_tid_mcfn(tid)
597 key = mc + ":" + self.tidtopn[tid] + ":" + taskname
598 self.unitaskhashes[key] = (self.taskhash[tid], unihash)
599 self.unihash[tid] = unihash
600
601 def _get_unihash(self, tid, checkkey=None):
602 if tid not in self.tidtopn:
603 return None
604 (mc, fn, taskname, taskfn) = bb.runqueue.split_tid_mcfn(tid)
605 key = mc + ":" + self.tidtopn[tid] + ":" + taskname
606 if key not in self.unitaskhashes:
607 return None
608 if not checkkey:
609 checkkey = self.taskhash[tid]
610 (key, unihash) = self.unitaskhashes[key]
611 if key != checkkey:
612 return None
613 return unihash
Brad Bishop08902b02019-08-20 09:16:51 -0400614
Patrick Williams73bd93f2024-02-20 08:07:48 -0600615 def get_cached_unihash(self, tid):
Brad Bishop08902b02019-08-20 09:16:51 -0400616 taskhash = self.taskhash[tid]
617
Brad Bishopa34c0302019-09-23 22:34:48 -0400618 # If its not a setscene task we can return
619 if self.setscenetasks and tid not in self.setscenetasks:
Andrew Geissler82c905d2020-04-13 13:39:40 -0500620 self.unihash[tid] = None
Brad Bishopa34c0302019-09-23 22:34:48 -0400621 return taskhash
622
Brad Bishop08902b02019-08-20 09:16:51 -0400623 # TODO: This cache can grow unbounded. It probably only needs to keep
624 # for each task
Andrew Geissler82c905d2020-04-13 13:39:40 -0500625 unihash = self._get_unihash(tid)
Brad Bishop08902b02019-08-20 09:16:51 -0400626 if unihash is not None:
Andrew Geissler82c905d2020-04-13 13:39:40 -0500627 self.unihash[tid] = unihash
Brad Bishop08902b02019-08-20 09:16:51 -0400628 return unihash
629
Patrick Williams73bd93f2024-02-20 08:07:48 -0600630 return None
Brad Bishop08902b02019-08-20 09:16:51 -0400631
Patrick Williams73bd93f2024-02-20 08:07:48 -0600632 def _get_method(self, tid):
633 method = self.method
634 if tid in self.extramethod:
635 method = method + self.extramethod[tid]
636
637 return method
638
639 def unihashes_exist(self, query):
640 if len(query) == 0:
641 return {}
642
643 uncached_query = {}
644 result = {}
645 for key, unihash in query.items():
646 if unihash in self.unihash_exists_cache:
647 result[key] = True
648 else:
649 uncached_query[key] = unihash
650
651 if self.max_parallel <= 1 or len(uncached_query) <= 1:
652 # No parallelism required. Make the query serially with the single client
653 uncached_result = {
654 key: self.client().unihash_exists(value) for key, value in uncached_query.items()
655 }
656 else:
657 uncached_result = self.client_pool().unihashes_exist(uncached_query)
658
659 for key, exists in uncached_result.items():
660 if exists:
661 self.unihash_exists_cache.add(query[key])
662 result[key] = exists
663
664 return result
665
666 def get_unihash(self, tid):
667 return self.get_unihashes([tid])[tid]
668
669 def get_unihashes(self, tids):
670 """
671 For a iterable of tids, returns a dictionary that maps each tid to a
672 unihash
673 """
674 result = {}
675 queries = {}
676 query_result = {}
677
678 for tid in tids:
679 unihash = self.get_cached_unihash(tid)
680 if unihash:
681 result[tid] = unihash
682 else:
683 queries[tid] = (self._get_method(tid), self.taskhash[tid])
684
685 if len(queries) == 0:
686 return result
687
688 if self.max_parallel <= 1 or len(queries) <= 1:
689 # No parallelism required. Make the query serially with the single client
690 for tid, args in queries.items():
691 query_result[tid] = self.client().get_unihash(*args)
692 else:
693 query_result = self.client_pool().get_unihashes(queries)
694
695 for tid, unihash in query_result.items():
696 # In the absence of being able to discover a unique hash from the
697 # server, make it be equivalent to the taskhash. The unique "hash" only
698 # really needs to be a unique string (not even necessarily a hash), but
699 # making it match the taskhash has a few advantages:
700 #
701 # 1) All of the sstate code that assumes hashes can be the same
702 # 2) It provides maximal compatibility with builders that don't use
703 # an equivalency server
704 # 3) The value is easy for multiple independent builders to derive the
705 # same unique hash from the same input. This means that if the
706 # independent builders find the same taskhash, but it isn't reported
707 # to the server, there is a better chance that they will agree on
708 # the unique hash.
709 taskhash = self.taskhash[tid]
710 if unihash:
Brad Bishop08902b02019-08-20 09:16:51 -0400711 # A unique hash equal to the taskhash is not very interesting,
712 # so it is reported it at debug level 2. If they differ, that
713 # is much more interesting, so it is reported at debug level 1
Andrew Geissler6aa7eec2023-03-03 12:41:14 -0600714 hashequiv_logger.bbdebug((1, 2)[unihash == taskhash], 'Found unihash %s in place of %s for %s from %s' % (unihash, taskhash, tid, self.server))
Brad Bishop08902b02019-08-20 09:16:51 -0400715 else:
Andrew Geisslerd1e89492021-02-12 15:35:20 -0600716 hashequiv_logger.debug2('No reported unihash for %s:%s from %s' % (tid, taskhash, self.server))
Patrick Williams73bd93f2024-02-20 08:07:48 -0600717 unihash = taskhash
Brad Bishop08902b02019-08-20 09:16:51 -0400718
Patrick Williams73bd93f2024-02-20 08:07:48 -0600719
720 self.set_unihash(tid, unihash)
721 self.unihash[tid] = unihash
722 result[tid] = unihash
723
724 return result
Brad Bishop08902b02019-08-20 09:16:51 -0400725
726 def report_unihash(self, path, task, d):
Brad Bishop08902b02019-08-20 09:16:51 -0400727 import importlib
728
729 taskhash = d.getVar('BB_TASKHASH')
730 unihash = d.getVar('BB_UNIHASH')
731 report_taskdata = d.getVar('SSTATE_HASHEQUIV_REPORT_TASKDATA') == '1'
732 tempdir = d.getVar('T')
Andrew Geissler517393d2023-01-13 08:55:19 -0600733 mcfn = d.getVar('BB_FILENAME')
734 tid = mcfn + ':do_' + task
Andrew Geissler82c905d2020-04-13 13:39:40 -0500735 key = tid + ':' + taskhash
Brad Bishop00e122a2019-10-05 11:10:57 -0400736
737 if self.setscenetasks and tid not in self.setscenetasks:
738 return
Brad Bishop08902b02019-08-20 09:16:51 -0400739
Andrew Geissler7e0e3c02022-02-25 20:34:39 +0000740 # This can happen if locked sigs are in action. Detect and just exit
Andrew Geissler82c905d2020-04-13 13:39:40 -0500741 if taskhash != self.taskhash[tid]:
742 return
743
Brad Bishop08902b02019-08-20 09:16:51 -0400744 # Sanity checks
Andrew Geissler82c905d2020-04-13 13:39:40 -0500745 cache_unihash = self._get_unihash(tid, checkkey=taskhash)
Brad Bishop08902b02019-08-20 09:16:51 -0400746 if cache_unihash is None:
747 bb.fatal('%s not in unihash cache. Please report this error' % key)
748
749 if cache_unihash != unihash:
750 bb.fatal("Cache unihash %s doesn't match BB_UNIHASH %s" % (cache_unihash, unihash))
751
752 sigfile = None
753 sigfile_name = "depsig.do_%s.%d" % (task, os.getpid())
754 sigfile_link = "depsig.do_%s" % task
755
756 try:
757 sigfile = open(os.path.join(tempdir, sigfile_name), 'w+b')
758
759 locs = {'path': path, 'sigfile': sigfile, 'task': task, 'd': d}
760
761 if "." in self.method:
762 (module, method) = self.method.rsplit('.', 1)
763 locs['method'] = getattr(importlib.import_module(module), method)
764 outhash = bb.utils.better_eval('method(path, sigfile, task, d)', locs)
765 else:
766 outhash = bb.utils.better_eval(self.method + '(path, sigfile, task, d)', locs)
767
768 try:
Brad Bishopa34c0302019-09-23 22:34:48 -0400769 extra_data = {}
770
771 owner = d.getVar('SSTATE_HASHEQUIV_OWNER')
772 if owner:
773 extra_data['owner'] = owner
Brad Bishop08902b02019-08-20 09:16:51 -0400774
775 if report_taskdata:
776 sigfile.seek(0)
777
Brad Bishopa34c0302019-09-23 22:34:48 -0400778 extra_data['PN'] = d.getVar('PN')
779 extra_data['PV'] = d.getVar('PV')
780 extra_data['PR'] = d.getVar('PR')
781 extra_data['task'] = task
782 extra_data['outhash_siginfo'] = sigfile.read().decode('utf-8')
Brad Bishop08902b02019-08-20 09:16:51 -0400783
Andrew Geissler82c905d2020-04-13 13:39:40 -0500784 method = self.method
785 if tid in self.extramethod:
786 method = method + self.extramethod[tid]
787
788 data = self.client().report_unihash(taskhash, method, outhash, unihash, extra_data)
Brad Bishopa34c0302019-09-23 22:34:48 -0400789 new_unihash = data['unihash']
Brad Bishop08902b02019-08-20 09:16:51 -0400790
791 if new_unihash != unihash:
Andrew Geisslerd1e89492021-02-12 15:35:20 -0600792 hashequiv_logger.debug('Task %s unihash changed %s -> %s by server %s' % (taskhash, unihash, new_unihash, self.server))
Andrew Geissler517393d2023-01-13 08:55:19 -0600793 bb.event.fire(bb.runqueue.taskUniHashUpdate(mcfn + ':do_' + task, new_unihash), d)
Andrew Geissler82c905d2020-04-13 13:39:40 -0500794 self.set_unihash(tid, new_unihash)
795 d.setVar('BB_UNIHASH', new_unihash)
Brad Bishop08902b02019-08-20 09:16:51 -0400796 else:
Andrew Geisslerd1e89492021-02-12 15:35:20 -0600797 hashequiv_logger.debug('Reported task %s as unihash %s to %s' % (taskhash, unihash, self.server))
Andrew Geisslerc926e172021-05-07 16:11:35 -0500798 except ConnectionError as e:
Brad Bishopa34c0302019-09-23 22:34:48 -0400799 bb.warn('Error contacting Hash Equivalence Server %s: %s' % (self.server, str(e)))
Brad Bishop08902b02019-08-20 09:16:51 -0400800 finally:
801 if sigfile:
802 sigfile.close()
803
804 sigfile_link_path = os.path.join(tempdir, sigfile_link)
805 bb.utils.remove(sigfile_link_path)
806
807 try:
808 os.symlink(sigfile_name, sigfile_link_path)
809 except OSError:
810 pass
811
Andrew Geissler82c905d2020-04-13 13:39:40 -0500812 def report_unihash_equiv(self, tid, taskhash, wanted_unihash, current_unihash, datacaches):
813 try:
814 extra_data = {}
815 method = self.method
816 if tid in self.extramethod:
817 method = method + self.extramethod[tid]
818
819 data = self.client().report_unihash_equiv(taskhash, method, wanted_unihash, extra_data)
820 hashequiv_logger.verbose('Reported task %s as unihash %s to %s (%s)' % (tid, wanted_unihash, self.server, str(data)))
821
822 if data is None:
823 bb.warn("Server unable to handle unihash report")
824 return False
825
826 finalunihash = data['unihash']
827
828 if finalunihash == current_unihash:
829 hashequiv_logger.verbose('Task %s unihash %s unchanged by server' % (tid, finalunihash))
830 elif finalunihash == wanted_unihash:
831 hashequiv_logger.verbose('Task %s unihash changed %s -> %s as wanted' % (tid, current_unihash, finalunihash))
832 self.set_unihash(tid, finalunihash)
833 return True
834 else:
835 # TODO: What to do here?
836 hashequiv_logger.verbose('Task %s unihash reported as unwanted hash %s' % (tid, finalunihash))
837
Andrew Geisslerc926e172021-05-07 16:11:35 -0500838 except ConnectionError as e:
Andrew Geissler82c905d2020-04-13 13:39:40 -0500839 bb.warn('Error contacting Hash Equivalence Server %s: %s' % (self.server, str(e)))
840
841 return False
Brad Bishop08902b02019-08-20 09:16:51 -0400842
843#
844# Dummy class used for bitbake-selftest
845#
846class SignatureGeneratorTestEquivHash(SignatureGeneratorUniHashMixIn, SignatureGeneratorBasicHash):
847 name = "TestEquivHash"
848 def init_rundepcheck(self, data):
849 super().init_rundepcheck(data)
Brad Bishopa34c0302019-09-23 22:34:48 -0400850 self.server = data.getVar('BB_HASHSERVE')
Brad Bishop08902b02019-08-20 09:16:51 -0400851 self.method = "sstate_output_hash"
Patrick Williams73bd93f2024-02-20 08:07:48 -0600852 self.max_parallel = 1
Brad Bishop08902b02019-08-20 09:16:51 -0400853
Andrew Geissler5082cc72023-09-11 08:41:39 -0400854def clean_checksum_file_path(file_checksum_tuple):
855 f, cs = file_checksum_tuple
856 if "/./" in f:
857 return "./" + f.split("/./")[1]
858 return f
859
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500860def dump_this_task(outfile, d):
861 import bb.parse
Andrew Geissler517393d2023-01-13 08:55:19 -0600862 mcfn = d.getVar("BB_FILENAME")
Brad Bishop6e60e8b2018-02-01 10:27:11 -0500863 task = "do_" + d.getVar("BB_CURRENTTASK")
Andrew Geissler517393d2023-01-13 08:55:19 -0600864 referencestamp = bb.parse.siggen.stampfile_base(mcfn)
865 bb.parse.siggen.dump_sigtask(mcfn, task, outfile, "customfile:" + referencestamp)
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500866
Brad Bishop6e60e8b2018-02-01 10:27:11 -0500867def init_colors(enable_color):
868 """Initialise colour dict for passing to compare_sigfiles()"""
869 # First set up the colours
Brad Bishop1a4b7ee2018-12-16 17:11:34 -0800870 colors = {'color_title': '\033[1m',
871 'color_default': '\033[0m',
872 'color_add': '\033[0;32m',
873 'color_remove': '\033[0;31m',
Brad Bishop6e60e8b2018-02-01 10:27:11 -0500874 }
875 # Leave all keys present but clear the values
876 if not enable_color:
877 for k in colors.keys():
878 colors[k] = ''
879 return colors
880
881def worddiff_str(oldstr, newstr, colors=None):
882 if not colors:
883 colors = init_colors(False)
884 diff = simplediff.diff(oldstr.split(' '), newstr.split(' '))
885 ret = []
886 for change, value in diff:
887 value = ' '.join(value)
888 if change == '=':
889 ret.append(value)
890 elif change == '+':
891 item = '{color_add}{{+{value}+}}{color_default}'.format(value=value, **colors)
892 ret.append(item)
893 elif change == '-':
894 item = '{color_remove}[-{value}-]{color_default}'.format(value=value, **colors)
895 ret.append(item)
896 whitespace_note = ''
897 if oldstr != newstr and ' '.join(oldstr.split()) == ' '.join(newstr.split()):
898 whitespace_note = ' (whitespace changed)'
899 return '"%s"%s' % (' '.join(ret), whitespace_note)
900
901def list_inline_diff(oldlist, newlist, colors=None):
902 if not colors:
903 colors = init_colors(False)
904 diff = simplediff.diff(oldlist, newlist)
905 ret = []
906 for change, value in diff:
907 value = ' '.join(value)
908 if change == '=':
909 ret.append("'%s'" % value)
910 elif change == '+':
911 item = '{color_add}+{value}{color_default}'.format(value=value, **colors)
912 ret.append(item)
913 elif change == '-':
914 item = '{color_remove}-{value}{color_default}'.format(value=value, **colors)
915 ret.append(item)
916 return '[%s]' % (', '.join(ret))
917
Andrew Geissler7e0e3c02022-02-25 20:34:39 +0000918# Handled renamed fields
919def handle_renames(data):
920 if 'basewhitelist' in data:
921 data['basehash_ignore_vars'] = data['basewhitelist']
922 del data['basewhitelist']
923 if 'taskwhitelist' in data:
924 data['taskhash_ignore_tasks'] = data['taskwhitelist']
925 del data['taskwhitelist']
926
927
Brad Bishop6e60e8b2018-02-01 10:27:11 -0500928def compare_sigfiles(a, b, recursecb=None, color=False, collapsed=False):
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500929 output = []
930
Brad Bishop6e60e8b2018-02-01 10:27:11 -0500931 colors = init_colors(color)
932 def color_format(formatstr, **values):
933 """
934 Return colour formatted string.
935 NOTE: call with the format string, not an already formatted string
936 containing values (otherwise you could have trouble with { and }
937 characters)
938 """
939 if not formatstr.endswith('{color_default}'):
940 formatstr += '{color_default}'
941 # In newer python 3 versions you can pass both of these directly,
942 # but we only require 3.4 at the moment
943 formatparams = {}
944 formatparams.update(colors)
945 formatparams.update(values)
946 return formatstr.format(**formatparams)
947
Patrick Williams73bd93f2024-02-20 08:07:48 -0600948 try:
949 with bb.compress.zstd.open(a, "rt", encoding="utf-8", num_threads=1) as f:
950 a_data = json.load(f, object_hook=SetDecoder)
951 except (TypeError, OSError) as err:
952 bb.error("Failed to open sigdata file '%s': %s" % (a, str(err)))
953 raise err
954 try:
955 with bb.compress.zstd.open(b, "rt", encoding="utf-8", num_threads=1) as f:
956 b_data = json.load(f, object_hook=SetDecoder)
957 except (TypeError, OSError) as err:
958 bb.error("Failed to open sigdata file '%s': %s" % (b, str(err)))
959 raise err
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500960
Andrew Geissler7e0e3c02022-02-25 20:34:39 +0000961 for data in [a_data, b_data]:
962 handle_renames(data)
963
964 def dict_diff(a, b, ignored_vars=set()):
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500965 sa = set(a.keys())
966 sb = set(b.keys())
967 common = sa & sb
968 changed = set()
969 for i in common:
Andrew Geissler7e0e3c02022-02-25 20:34:39 +0000970 if a[i] != b[i] and i not in ignored_vars:
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500971 changed.add(i)
972 added = sb - sa
973 removed = sa - sb
974 return changed, added, removed
975
976 def file_checksums_diff(a, b):
977 from collections import Counter
Andrew Geisslereff27472021-10-29 15:35:00 -0500978
979 # Convert lists back to tuples
980 a = [(f[0], f[1]) for f in a]
981 b = [(f[0], f[1]) for f in b]
982
Patrick Williamsc124f4f2015-09-15 14:41:29 -0500983 # Compare lists, ensuring we can handle duplicate filenames if they exist
984 removedcount = Counter(a)
985 removedcount.subtract(b)
986 addedcount = Counter(b)
987 addedcount.subtract(a)
988 added = []
989 for x in b:
990 if addedcount[x] > 0:
991 addedcount[x] -= 1
992 added.append(x)
993 removed = []
994 changed = []
995 for x in a:
996 if removedcount[x] > 0:
997 removedcount[x] -= 1
998 for y in added:
999 if y[0] == x[0]:
1000 changed.append((x[0], x[1], y[1]))
1001 added.remove(y)
1002 break
1003 else:
1004 removed.append(x)
1005 added = [x[0] for x in added]
1006 removed = [x[0] for x in removed]
1007 return changed, added, removed
1008
Andrew Geissler7e0e3c02022-02-25 20:34:39 +00001009 if 'basehash_ignore_vars' in a_data and a_data['basehash_ignore_vars'] != b_data['basehash_ignore_vars']:
1010 output.append(color_format("{color_title}basehash_ignore_vars changed{color_default} from '%s' to '%s'") % (a_data['basehash_ignore_vars'], b_data['basehash_ignore_vars']))
1011 if a_data['basehash_ignore_vars'] and b_data['basehash_ignore_vars']:
1012 output.append("changed items: %s" % a_data['basehash_ignore_vars'].symmetric_difference(b_data['basehash_ignore_vars']))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001013
Andrew Geissler7e0e3c02022-02-25 20:34:39 +00001014 if 'taskhash_ignore_tasks' in a_data and a_data['taskhash_ignore_tasks'] != b_data['taskhash_ignore_tasks']:
1015 output.append(color_format("{color_title}taskhash_ignore_tasks changed{color_default} from '%s' to '%s'") % (a_data['taskhash_ignore_tasks'], b_data['taskhash_ignore_tasks']))
1016 if a_data['taskhash_ignore_tasks'] and b_data['taskhash_ignore_tasks']:
1017 output.append("changed items: %s" % a_data['taskhash_ignore_tasks'].symmetric_difference(b_data['taskhash_ignore_tasks']))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001018
1019 if a_data['taskdeps'] != b_data['taskdeps']:
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001020 output.append(color_format("{color_title}Task dependencies changed{color_default} from:\n%s\nto:\n%s") % (sorted(a_data['taskdeps']), sorted(b_data['taskdeps'])))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001021
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001022 if a_data['basehash'] != b_data['basehash'] and not collapsed:
1023 output.append(color_format("{color_title}basehash changed{color_default} from %s to %s") % (a_data['basehash'], b_data['basehash']))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001024
Andrew Geissler7e0e3c02022-02-25 20:34:39 +00001025 changed, added, removed = dict_diff(a_data['gendeps'], b_data['gendeps'], a_data['basehash_ignore_vars'] & b_data['basehash_ignore_vars'])
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001026 if changed:
Patrick Williams93c203f2021-10-06 16:15:23 -05001027 for dep in sorted(changed):
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001028 output.append(color_format("{color_title}List of dependencies for variable %s changed from '{color_default}%s{color_title}' to '{color_default}%s{color_title}'") % (dep, a_data['gendeps'][dep], b_data['gendeps'][dep]))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001029 if a_data['gendeps'][dep] and b_data['gendeps'][dep]:
1030 output.append("changed items: %s" % a_data['gendeps'][dep].symmetric_difference(b_data['gendeps'][dep]))
1031 if added:
Patrick Williams93c203f2021-10-06 16:15:23 -05001032 for dep in sorted(added):
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001033 output.append(color_format("{color_title}Dependency on variable %s was added") % (dep))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001034 if removed:
Patrick Williams93c203f2021-10-06 16:15:23 -05001035 for dep in sorted(removed):
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001036 output.append(color_format("{color_title}Dependency on Variable %s was removed") % (dep))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001037
1038
1039 changed, added, removed = dict_diff(a_data['varvals'], b_data['varvals'])
1040 if changed:
Patrick Williams93c203f2021-10-06 16:15:23 -05001041 for dep in sorted(changed):
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001042 oldval = a_data['varvals'][dep]
1043 newval = b_data['varvals'][dep]
1044 if newval and oldval and ('\n' in oldval or '\n' in newval):
1045 diff = difflib.unified_diff(oldval.splitlines(), newval.splitlines(), lineterm='')
1046 # Cut off the first two lines, since we aren't interested in
1047 # the old/new filename (they are blank anyway in this case)
1048 difflines = list(diff)[2:]
1049 if color:
1050 # Add colour to diff output
1051 for i, line in enumerate(difflines):
1052 if line.startswith('+'):
1053 line = color_format('{color_add}{line}', line=line)
1054 difflines[i] = line
1055 elif line.startswith('-'):
1056 line = color_format('{color_remove}{line}', line=line)
1057 difflines[i] = line
1058 output.append(color_format("{color_title}Variable {var} value changed:{color_default}\n{diff}", var=dep, diff='\n'.join(difflines)))
1059 elif newval and oldval and (' ' in oldval or ' ' in newval):
1060 output.append(color_format("{color_title}Variable {var} value changed:{color_default}\n{diff}", var=dep, diff=worddiff_str(oldval, newval, colors)))
1061 else:
1062 output.append(color_format("{color_title}Variable {var} value changed from '{color_default}{oldval}{color_title}' to '{color_default}{newval}{color_title}'{color_default}", var=dep, oldval=oldval, newval=newval))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001063
Patrick Williamsc0f7c042017-02-23 20:41:17 -06001064 if not 'file_checksum_values' in a_data:
Andrew Geisslereff27472021-10-29 15:35:00 -05001065 a_data['file_checksum_values'] = []
Patrick Williamsc0f7c042017-02-23 20:41:17 -06001066 if not 'file_checksum_values' in b_data:
Andrew Geisslereff27472021-10-29 15:35:00 -05001067 b_data['file_checksum_values'] = []
Patrick Williamsc0f7c042017-02-23 20:41:17 -06001068
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001069 changed, added, removed = file_checksums_diff(a_data['file_checksum_values'], b_data['file_checksum_values'])
1070 if changed:
1071 for f, old, new in changed:
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001072 output.append(color_format("{color_title}Checksum for file %s changed{color_default} from %s to %s") % (f, old, new))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001073 if added:
1074 for f in added:
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001075 output.append(color_format("{color_title}Dependency on checksum of file %s was added") % (f))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001076 if removed:
1077 for f in removed:
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001078 output.append(color_format("{color_title}Dependency on checksum of file %s was removed") % (f))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001079
Patrick Williamsc0f7c042017-02-23 20:41:17 -06001080 if not 'runtaskdeps' in a_data:
1081 a_data['runtaskdeps'] = {}
1082 if not 'runtaskdeps' in b_data:
1083 b_data['runtaskdeps'] = {}
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001084
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001085 if not collapsed:
1086 if len(a_data['runtaskdeps']) != len(b_data['runtaskdeps']):
1087 changed = ["Number of task dependencies changed"]
1088 else:
1089 changed = []
1090 for idx, task in enumerate(a_data['runtaskdeps']):
1091 a = a_data['runtaskdeps'][idx]
1092 b = b_data['runtaskdeps'][idx]
1093 if a_data['runtaskhashes'][a] != b_data['runtaskhashes'][b] and not collapsed:
Patrick Williams2a254922023-08-11 09:48:11 -05001094 changed.append("%s with hash %s\n changed to\n%s with hash %s" % (a, a_data['runtaskhashes'][a], b, b_data['runtaskhashes'][b]))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001095
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001096 if changed:
Patrick Williams2a254922023-08-11 09:48:11 -05001097 clean_a = a_data['runtaskdeps']
1098 clean_b = b_data['runtaskdeps']
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001099 if clean_a != clean_b:
1100 output.append(color_format("{color_title}runtaskdeps changed:{color_default}\n%s") % list_inline_diff(clean_a, clean_b, colors))
1101 else:
1102 output.append(color_format("{color_title}runtaskdeps changed:"))
1103 output.append("\n".join(changed))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001104
1105
1106 if 'runtaskhashes' in a_data and 'runtaskhashes' in b_data:
Patrick Williams2a254922023-08-11 09:48:11 -05001107 a = a_data['runtaskhashes']
1108 b = b_data['runtaskhashes']
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001109 changed, added, removed = dict_diff(a, b)
1110 if added:
Patrick Williams93c203f2021-10-06 16:15:23 -05001111 for dep in sorted(added):
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001112 bdep_found = False
1113 if removed:
1114 for bdep in removed:
1115 if b[dep] == a[bdep]:
1116 #output.append("Dependency on task %s was replaced by %s with same hash" % (dep, bdep))
1117 bdep_found = True
1118 if not bdep_found:
Patrick Williams03907ee2022-05-01 06:28:52 -05001119 output.append(color_format("{color_title}Dependency on task %s was added{color_default} with hash %s") % (dep, b[dep]))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001120 if removed:
Patrick Williams93c203f2021-10-06 16:15:23 -05001121 for dep in sorted(removed):
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001122 adep_found = False
1123 if added:
1124 for adep in added:
1125 if b[adep] == a[dep]:
1126 #output.append("Dependency on task %s was replaced by %s with same hash" % (adep, dep))
1127 adep_found = True
1128 if not adep_found:
Patrick Williams03907ee2022-05-01 06:28:52 -05001129 output.append(color_format("{color_title}Dependency on task %s was removed{color_default} with hash %s") % (dep, a[dep]))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001130 if changed:
Patrick Williams93c203f2021-10-06 16:15:23 -05001131 for dep in sorted(changed):
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001132 if not collapsed:
Patrick Williams03907ee2022-05-01 06:28:52 -05001133 output.append(color_format("{color_title}Hash for task dependency %s changed{color_default} from %s to %s") % (dep, a[dep], b[dep]))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001134 if callable(recursecb):
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001135 recout = recursecb(dep, a[dep], b[dep])
1136 if recout:
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001137 if collapsed:
1138 output.extend(recout)
1139 else:
Brad Bishop1a4b7ee2018-12-16 17:11:34 -08001140 # If a dependent hash changed, might as well print the line above and then defer to the changes in
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001141 # that hash since in all likelyhood, they're the same changes this task also saw.
1142 output = [output[-1]] + recout
Andrew Geisslerd5838332022-05-27 11:33:10 -05001143 break
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001144
1145 a_taint = a_data.get('taint', None)
1146 b_taint = b_data.get('taint', None)
1147 if a_taint != b_taint:
Brad Bishop96ff1982019-08-19 13:50:42 -04001148 if a_taint and a_taint.startswith('nostamp:'):
Brad Bishopc342db32019-05-15 21:57:59 -04001149 a_taint = a_taint.replace('nostamp:', 'nostamp(uuid4):')
Brad Bishop96ff1982019-08-19 13:50:42 -04001150 if b_taint and b_taint.startswith('nostamp:'):
Brad Bishopc342db32019-05-15 21:57:59 -04001151 b_taint = b_taint.replace('nostamp:', 'nostamp(uuid4):')
Brad Bishop6e60e8b2018-02-01 10:27:11 -05001152 output.append(color_format("{color_title}Taint (by forced/invalidated task) changed{color_default} from %s to %s") % (a_taint, b_taint))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001153
1154 return output
1155
1156
Patrick Williamsd8c66bc2016-06-20 12:57:21 -05001157def calc_basehash(sigdata):
1158 task = sigdata['task']
1159 basedata = sigdata['varvals'][task]
1160
1161 if basedata is None:
1162 basedata = ''
1163
1164 alldeps = sigdata['taskdeps']
Andrew Geissler517393d2023-01-13 08:55:19 -06001165 for dep in sorted(alldeps):
Patrick Williamsd8c66bc2016-06-20 12:57:21 -05001166 basedata = basedata + dep
1167 val = sigdata['varvals'][dep]
1168 if val is not None:
1169 basedata = basedata + str(val)
1170
Brad Bishop19323692019-04-05 15:28:33 -04001171 return hashlib.sha256(basedata.encode("utf-8")).hexdigest()
Patrick Williamsd8c66bc2016-06-20 12:57:21 -05001172
1173def calc_taskhash(sigdata):
1174 data = sigdata['basehash']
1175
1176 for dep in sigdata['runtaskdeps']:
1177 data = data + sigdata['runtaskhashes'][dep]
1178
1179 for c in sigdata['file_checksum_values']:
Brad Bishop37a0e4d2017-12-04 01:01:44 -05001180 if c[1]:
Andrew Geissler595f6302022-01-24 19:11:47 +00001181 if "./" in c[0]:
1182 data = data + c[0]
Brad Bishop37a0e4d2017-12-04 01:01:44 -05001183 data = data + c[1]
Patrick Williamsd8c66bc2016-06-20 12:57:21 -05001184
1185 if 'taint' in sigdata:
1186 if 'nostamp:' in sigdata['taint']:
1187 data = data + sigdata['taint'][8:]
1188 else:
1189 data = data + sigdata['taint']
1190
Brad Bishop19323692019-04-05 15:28:33 -04001191 return hashlib.sha256(data.encode("utf-8")).hexdigest()
Patrick Williamsd8c66bc2016-06-20 12:57:21 -05001192
1193
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001194def dump_sigfile(a):
1195 output = []
1196
Patrick Williams73bd93f2024-02-20 08:07:48 -06001197 try:
1198 with bb.compress.zstd.open(a, "rt", encoding="utf-8", num_threads=1) as f:
1199 a_data = json.load(f, object_hook=SetDecoder)
1200 except (TypeError, OSError) as err:
1201 bb.error("Failed to open sigdata file '%s': %s" % (a, str(err)))
1202 raise err
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001203
Andrew Geissler7e0e3c02022-02-25 20:34:39 +00001204 handle_renames(a_data)
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001205
Andrew Geissler7e0e3c02022-02-25 20:34:39 +00001206 output.append("basehash_ignore_vars: %s" % (sorted(a_data['basehash_ignore_vars'])))
1207
1208 output.append("taskhash_ignore_tasks: %s" % (sorted(a_data['taskhash_ignore_tasks'] or [])))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001209
1210 output.append("Task dependencies: %s" % (sorted(a_data['taskdeps'])))
1211
1212 output.append("basehash: %s" % (a_data['basehash']))
1213
Andrew Geissler595f6302022-01-24 19:11:47 +00001214 for dep in sorted(a_data['gendeps']):
1215 output.append("List of dependencies for variable %s is %s" % (dep, sorted(a_data['gendeps'][dep])))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001216
Andrew Geissler595f6302022-01-24 19:11:47 +00001217 for dep in sorted(a_data['varvals']):
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001218 output.append("Variable %s value is %s" % (dep, a_data['varvals'][dep]))
1219
1220 if 'runtaskdeps' in a_data:
Andrew Geissler595f6302022-01-24 19:11:47 +00001221 output.append("Tasks this task depends on: %s" % (sorted(a_data['runtaskdeps'])))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001222
1223 if 'file_checksum_values' in a_data:
Andrew Geissler595f6302022-01-24 19:11:47 +00001224 output.append("This task depends on the checksums of files: %s" % (sorted(a_data['file_checksum_values'])))
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001225
1226 if 'runtaskhashes' in a_data:
Andrew Geissler595f6302022-01-24 19:11:47 +00001227 for dep in sorted(a_data['runtaskhashes']):
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001228 output.append("Hash for dependent task %s is %s" % (dep, a_data['runtaskhashes'][dep]))
1229
1230 if 'taint' in a_data:
Brad Bishopc342db32019-05-15 21:57:59 -04001231 if a_data['taint'].startswith('nostamp:'):
1232 msg = a_data['taint'].replace('nostamp:', 'nostamp(uuid4):')
1233 else:
1234 msg = a_data['taint']
1235 output.append("Tainted (by forced/invalidated task): %s" % msg)
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001236
Patrick Williamsd8c66bc2016-06-20 12:57:21 -05001237 if 'task' in a_data:
1238 computed_basehash = calc_basehash(a_data)
1239 output.append("Computed base hash is %s and from file %s" % (computed_basehash, a_data['basehash']))
1240 else:
1241 output.append("Unable to compute base hash")
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001242
Patrick Williamsd8c66bc2016-06-20 12:57:21 -05001243 computed_taskhash = calc_taskhash(a_data)
1244 output.append("Computed task hash is %s" % computed_taskhash)
Patrick Williamsc124f4f2015-09-15 14:41:29 -05001245
1246 return output