blob: 5b98f1e9252f3fa24830243db84d9c0c2e1209c4 [file] [log] [blame]
#!/usr/bin/env python3
"""
This script determines the given package's openbmc dependencies from its
configure.ac file where it downloads, configures, builds, and installs each of
these dependencies. Then the given package is configured, built, and installed
prior to executing its unit tests.
"""
import argparse
import multiprocessing
import os
import platform
import re
import shutil
import subprocess
import sys
from subprocess import CalledProcessError, check_call
from tempfile import TemporaryDirectory
from urllib.parse import urljoin
from git import Repo
# interpreter is not used directly but this resolves dependency ordering
# that would be broken if we didn't include it.
from mesonbuild import interpreter # noqa: F401
from mesonbuild import coredata, optinterpreter
from mesonbuild.mesonlib import OptionKey
from mesonbuild.mesonlib import version_compare as meson_version_compare
class DepTree:
"""
Represents package dependency tree, where each node is a DepTree with a
name and DepTree children.
"""
def __init__(self, name):
"""
Create new DepTree.
Parameter descriptions:
name Name of new tree node.
"""
self.name = name
self.children = list()
def AddChild(self, name):
"""
Add new child node to current node.
Parameter descriptions:
name Name of new child
"""
new_child = DepTree(name)
self.children.append(new_child)
return new_child
def AddChildNode(self, node):
"""
Add existing child node to current node.
Parameter descriptions:
node Tree node to add
"""
self.children.append(node)
def RemoveChild(self, name):
"""
Remove child node.
Parameter descriptions:
name Name of child to remove
"""
for child in self.children:
if child.name == name:
self.children.remove(child)
return
def GetNode(self, name):
"""
Return node with matching name. Return None if not found.
Parameter descriptions:
name Name of node to return
"""
if self.name == name:
return self
for child in self.children:
node = child.GetNode(name)
if node:
return node
return None
def GetParentNode(self, name, parent_node=None):
"""
Return parent of node with matching name. Return none if not found.
Parameter descriptions:
name Name of node to get parent of
parent_node Parent of current node
"""
if self.name == name:
return parent_node
for child in self.children:
found_node = child.GetParentNode(name, self)
if found_node:
return found_node
return None
def GetPath(self, name, path=None):
"""
Return list of node names from head to matching name.
Return None if not found.
Parameter descriptions:
name Name of node
path List of node names from head to current node
"""
if not path:
path = []
if self.name == name:
path.append(self.name)
return path
for child in self.children:
match = child.GetPath(name, path + [self.name])
if match:
return match
return None
def GetPathRegex(self, name, regex_str, path=None):
"""
Return list of node paths that end in name, or match regex_str.
Return empty list if not found.
Parameter descriptions:
name Name of node to search for
regex_str Regex string to match node names
path Path of node names from head to current node
"""
new_paths = []
if not path:
path = []
match = re.match(regex_str, self.name)
if (self.name == name) or (match):
new_paths.append(path + [self.name])
for child in self.children:
return_paths = None
full_path = path + [self.name]
return_paths = child.GetPathRegex(name, regex_str, full_path)
for i in return_paths:
new_paths.append(i)
return new_paths
def MoveNode(self, from_name, to_name):
"""
Mode existing from_name node to become child of to_name node.
Parameter descriptions:
from_name Name of node to make a child of to_name
to_name Name of node to make parent of from_name
"""
parent_from_node = self.GetParentNode(from_name)
from_node = self.GetNode(from_name)
parent_from_node.RemoveChild(from_name)
to_node = self.GetNode(to_name)
to_node.AddChildNode(from_node)
def ReorderDeps(self, name, regex_str):
"""
Reorder dependency tree. If tree contains nodes with names that
match 'name' and 'regex_str', move 'regex_str' nodes that are
to the right of 'name' node, so that they become children of the
'name' node.
Parameter descriptions:
name Name of node to look for
regex_str Regex string to match names to
"""
name_path = self.GetPath(name)
if not name_path:
return
paths = self.GetPathRegex(name, regex_str)
is_name_in_paths = False
name_index = 0
for i in range(len(paths)):
path = paths[i]
if path[-1] == name:
is_name_in_paths = True
name_index = i
break
if not is_name_in_paths:
return
for i in range(name_index + 1, len(paths)):
path = paths[i]
if name in path:
continue
from_name = path[-1]
self.MoveNode(from_name, name)
def GetInstallList(self):
"""
Return post-order list of node names.
Parameter descriptions:
"""
install_list = []
for child in self.children:
child_install_list = child.GetInstallList()
install_list.extend(child_install_list)
install_list.append(self.name)
return install_list
def PrintTree(self, level=0):
"""
Print pre-order node names with indentation denoting node depth level.
Parameter descriptions:
level Current depth level
"""
INDENT_PER_LEVEL = 4
print(" " * (level * INDENT_PER_LEVEL) + self.name)
for child in self.children:
child.PrintTree(level + 1)
def check_call_cmd(*cmd):
"""
Verbose prints the directory location the given command is called from and
the command, then executes the command using check_call.
Parameter descriptions:
dir Directory location command is to be called from
cmd List of parameters constructing the complete command
"""
printline(os.getcwd(), ">", " ".join(cmd))
check_call(cmd)
def clone_pkg(pkg, branch):
"""
Clone the given openbmc package's git repository from gerrit into
the WORKSPACE location
Parameter descriptions:
pkg Name of the package to clone
branch Branch to clone from pkg
"""
pkg_dir = os.path.join(WORKSPACE, pkg)
if os.path.exists(os.path.join(pkg_dir, ".git")):
return pkg_dir
pkg_repo = urljoin("https://gerrit.openbmc.org/openbmc/", pkg)
os.mkdir(pkg_dir)
printline(pkg_dir, "> git clone", pkg_repo, branch, "./")
try:
# first try the branch
clone = Repo.clone_from(pkg_repo, pkg_dir, branch=branch)
repo_inst = clone.working_dir
except Exception:
printline("Input branch not found, default to master")
clone = Repo.clone_from(pkg_repo, pkg_dir, branch="master")
repo_inst = clone.working_dir
return repo_inst
def make_target_exists(target):
"""
Runs a check against the makefile in the current directory to determine
if the target exists so that it can be built.
Parameter descriptions:
target The make target we are checking
"""
try:
cmd = ["make", "-n", target]
with open(os.devnull, "w") as devnull:
check_call(cmd, stdout=devnull, stderr=devnull)
return True
except CalledProcessError:
return False
make_parallel = [
"make",
# Run enough jobs to saturate all the cpus
"-j",
str(multiprocessing.cpu_count()),
# Don't start more jobs if the load avg is too high
"-l",
str(multiprocessing.cpu_count()),
# Synchronize the output so logs aren't intermixed in stdout / stderr
"-O",
]
def build_and_install(name, build_for_testing=False):
"""
Builds and installs the package in the environment. Optionally
builds the examples and test cases for package.
Parameter description:
name The name of the package we are building
build_for_testing Enable options related to testing on the package?
"""
os.chdir(os.path.join(WORKSPACE, name))
# Refresh dynamic linker run time bindings for dependencies
check_call_cmd("sudo", "-n", "--", "ldconfig")
pkg = Package()
if build_for_testing:
pkg.test()
else:
pkg.install()
def build_dep_tree(name, pkgdir, dep_added, head, branch, dep_tree=None):
"""
For each package (name), starting with the package to be unit tested,
extract its dependencies. For each package dependency defined, recursively
apply the same strategy
Parameter descriptions:
name Name of the package
pkgdir Directory where package source is located
dep_added Current dict of dependencies and added status
head Head node of the dependency tree
branch Branch to clone from pkg
dep_tree Current dependency tree node
"""
if not dep_tree:
dep_tree = head
with open("/tmp/depcache", "r") as depcache:
cache = depcache.readline()
# Read out pkg dependencies
pkg = Package(name, pkgdir)
build = pkg.build_system()
if not build:
raise Exception(f"Unable to find build system for {name}.")
for dep in set(build.dependencies()):
if dep in cache:
continue
# Dependency package not already known
if dep_added.get(dep) is None:
print(f"Adding {dep} dependency to {name}.")
# Dependency package not added
new_child = dep_tree.AddChild(dep)
dep_added[dep] = False
dep_pkgdir = clone_pkg(dep, branch)
# Determine this dependency package's
# dependencies and add them before
# returning to add this package
dep_added = build_dep_tree(
dep, dep_pkgdir, dep_added, head, branch, new_child
)
else:
# Dependency package known and added
if dep_added[dep]:
continue
else:
# Cyclic dependency failure
raise Exception("Cyclic dependencies found in " + name)
if not dep_added[name]:
dep_added[name] = True
return dep_added
def run_cppcheck():
if not os.path.exists(os.path.join("build", "compile_commands.json")):
return None
with TemporaryDirectory() as cpp_dir:
# http://cppcheck.sourceforge.net/manual.pdf
try:
check_call_cmd(
"cppcheck",
"-j",
str(multiprocessing.cpu_count()),
"--enable=style,performance,portability,missingInclude",
"--suppress=useStlAlgorithm",
"--suppress=unusedStructMember",
"--suppress=postfixOperator",
"--suppress=unreadVariable",
"--suppress=knownConditionTrueFalse",
"--library=googletest",
"--project=build/compile_commands.json",
f"--cppcheck-build-dir={cpp_dir}",
)
except subprocess.CalledProcessError:
print("cppcheck found errors")
def is_valgrind_safe():
"""
Returns whether it is safe to run valgrind on our platform
"""
src = "unit-test-vg.c"
exe = "./unit-test-vg"
with open(src, "w") as h:
h.write("#include <errno.h>\n")
h.write("#include <stdio.h>\n")
h.write("#include <stdlib.h>\n")
h.write("#include <string.h>\n")
h.write("int main() {\n")
h.write("char *heap_str = malloc(16);\n")
h.write('strcpy(heap_str, "RandString");\n')
h.write('int res = strcmp("RandString", heap_str);\n')
h.write("free(heap_str);\n")
h.write("char errstr[64];\n")
h.write("strerror_r(EINVAL, errstr, sizeof(errstr));\n")
h.write('printf("%s\\n", errstr);\n')
h.write("return res;\n")
h.write("}\n")
try:
with open(os.devnull, "w") as devnull:
check_call(
["gcc", "-O2", "-o", exe, src], stdout=devnull, stderr=devnull
)
check_call(
["valgrind", "--error-exitcode=99", exe],
stdout=devnull,
stderr=devnull,
)
return True
except Exception:
sys.stderr.write("###### Platform is not valgrind safe ######\n")
return False
finally:
os.remove(src)
os.remove(exe)
def is_sanitize_safe():
"""
Returns whether it is safe to run sanitizers on our platform
"""
src = "unit-test-sanitize.c"
exe = "./unit-test-sanitize"
with open(src, "w") as h:
h.write("int main() { return 0; }\n")
try:
with open(os.devnull, "w") as devnull:
check_call(
[
"gcc",
"-O2",
"-fsanitize=address",
"-fsanitize=undefined",
"-o",
exe,
src,
],
stdout=devnull,
stderr=devnull,
)
check_call([exe], stdout=devnull, stderr=devnull)
# TODO - Sanitizer not working on ppc64le
# https://github.com/openbmc/openbmc-build-scripts/issues/31
if platform.processor() == "ppc64le":
sys.stderr.write("###### ppc64le is not sanitize safe ######\n")
return False
else:
return True
except Exception:
sys.stderr.write("###### Platform is not sanitize safe ######\n")
return False
finally:
os.remove(src)
os.remove(exe)
def maybe_make_valgrind():
"""
Potentially runs the unit tests through valgrind for the package
via `make check-valgrind`. If the package does not have valgrind testing
then it just skips over this.
"""
# Valgrind testing is currently broken by an aggressive strcmp optimization
# that is inlined into optimized code for POWER by gcc 7+. Until we find
# a workaround, just don't run valgrind tests on POWER.
# https://github.com/openbmc/openbmc/issues/3315
if not is_valgrind_safe():
sys.stderr.write("###### Skipping valgrind ######\n")
return
if not make_target_exists("check-valgrind"):
return
try:
cmd = make_parallel + ["check-valgrind"]
check_call_cmd(*cmd)
except CalledProcessError:
for root, _, files in os.walk(os.getcwd()):
for f in files:
if re.search("test-suite-[a-z]+.log", f) is None:
continue
check_call_cmd("cat", os.path.join(root, f))
raise Exception("Valgrind tests failed")
def maybe_make_coverage():
"""
Potentially runs the unit tests through code coverage for the package
via `make check-code-coverage`. If the package does not have code coverage
testing then it just skips over this.
"""
if not make_target_exists("check-code-coverage"):
return
# Actually run code coverage
try:
cmd = make_parallel + ["check-code-coverage"]
check_call_cmd(*cmd)
except CalledProcessError:
raise Exception("Code coverage failed")
class BuildSystem(object):
"""
Build systems generally provide the means to configure, build, install and
test software. The BuildSystem class defines a set of interfaces on top of
which Autotools, Meson, CMake and possibly other build system drivers can
be implemented, separating out the phases to control whether a package
should merely be installed or also tested and analyzed.
"""
def __init__(self, package, path):
"""Initialise the driver with properties independent of the build
system
Keyword arguments:
package: The name of the package. Derived from the path if None
path: The path to the package. Set to the working directory if None
"""
self.path = "." if not path else path
realpath = os.path.realpath(self.path)
self.package = package if package else os.path.basename(realpath)
self.build_for_testing = False
def probe(self):
"""Test if the build system driver can be applied to the package
Return True if the driver can drive the package's build system,
otherwise False.
Generally probe() is implemented by testing for the presence of the
build system's configuration file(s).
"""
raise NotImplementedError
def dependencies(self):
"""Provide the package's dependencies
Returns a list of dependencies. If no dependencies are required then an
empty list must be returned.
Generally dependencies() is implemented by analysing and extracting the
data from the build system configuration.
"""
raise NotImplementedError
def configure(self, build_for_testing):
"""Configure the source ready for building
Should raise an exception if configuration failed.
Keyword arguments:
build_for_testing: Mark the package as being built for testing rather
than for installation as a dependency for the
package under test. Setting to True generally
implies that the package will be configured to build
with debug information, at a low level of
optimisation and possibly with sanitizers enabled.
Generally configure() is implemented by invoking the build system
tooling to generate Makefiles or equivalent.
"""
raise NotImplementedError
def build(self):
"""Build the software ready for installation and/or testing
Should raise an exception if the build fails
Generally build() is implemented by invoking `make` or `ninja`.
"""
raise NotImplementedError
def install(self):
"""Install the software ready for use
Should raise an exception if installation fails
Like build(), install() is generally implemented by invoking `make` or
`ninja`.
"""
raise NotImplementedError
def test(self):
"""Build and run the test suite associated with the package
Should raise an exception if the build or testing fails.
Like install(), test() is generally implemented by invoking `make` or
`ninja`.
"""
raise NotImplementedError
def analyze(self):
"""Run any supported analysis tools over the codebase
Should raise an exception if analysis fails.
Some analysis tools such as scan-build need injection into the build
system. analyze() provides the necessary hook to implement such
behaviour. Analyzers independent of the build system can also be
specified here but at the cost of possible duplication of code between
the build system driver implementations.
"""
raise NotImplementedError
class Autotools(BuildSystem):
def __init__(self, package=None, path=None):
super(Autotools, self).__init__(package, path)
def probe(self):
return os.path.isfile(os.path.join(self.path, "configure.ac"))
def dependencies(self):
configure_ac = os.path.join(self.path, "configure.ac")
contents = ""
# Prepend some special function overrides so we can parse out
# dependencies
for macro in DEPENDENCIES.keys():
contents += (
"m4_define(["
+ macro
+ "], ["
+ macro
+ "_START$"
+ str(DEPENDENCIES_OFFSET[macro] + 1)
+ macro
+ "_END])\n"
)
with open(configure_ac, "rt") as f:
contents += f.read()
autoconf_cmdline = ["autoconf", "-Wno-undefined", "-"]
autoconf_process = subprocess.Popen(
autoconf_cmdline,
stdin=subprocess.PIPE,
stdout=subprocess.PIPE,
stderr=subprocess.PIPE,
)
document = contents.encode("utf-8")
(stdout, stderr) = autoconf_process.communicate(input=document)
if not stdout:
print(stderr)
raise Exception("Failed to run autoconf for parsing dependencies")
# Parse out all of the dependency text
matches = []
for macro in DEPENDENCIES.keys():
pattern = "(" + macro + ")_START(.*?)" + macro + "_END"
for match in re.compile(pattern).finditer(stdout.decode("utf-8")):
matches.append((match.group(1), match.group(2)))
# Look up dependencies from the text
found_deps = []
for macro, deptext in matches:
for potential_dep in deptext.split(" "):
for known_dep in DEPENDENCIES[macro].keys():
if potential_dep.startswith(known_dep):
found_deps.append(DEPENDENCIES[macro][known_dep])
return found_deps
def _configure_feature(self, flag, enabled):
"""
Returns an configure flag as a string
Parameters:
flag The name of the flag
enabled Whether the flag is enabled or disabled
"""
return "--" + ("enable" if enabled else "disable") + "-" + flag
def configure(self, build_for_testing):
self.build_for_testing = build_for_testing
conf_flags = [
self._configure_feature("silent-rules", False),
self._configure_feature("examples", build_for_testing),
self._configure_feature("tests", build_for_testing),
self._configure_feature("itests", INTEGRATION_TEST),
]
conf_flags.extend(
[
self._configure_feature("code-coverage", build_for_testing),
self._configure_feature("valgrind", build_for_testing),
]
)
# Add any necessary configure flags for package
if CONFIGURE_FLAGS.get(self.package) is not None:
conf_flags.extend(CONFIGURE_FLAGS.get(self.package))
for bootstrap in ["bootstrap.sh", "bootstrap", "autogen.sh"]:
if os.path.exists(bootstrap):
check_call_cmd("./" + bootstrap)
break
check_call_cmd("./configure", *conf_flags)
def build(self):
check_call_cmd(*make_parallel)
def install(self):
check_call_cmd("sudo", "-n", "--", *(make_parallel + ["install"]))
def test(self):
try:
cmd = make_parallel + ["check"]
for i in range(0, args.repeat):
check_call_cmd(*cmd)
maybe_make_valgrind()
maybe_make_coverage()
except CalledProcessError:
for root, _, files in os.walk(os.getcwd()):
if "test-suite.log" not in files:
continue
check_call_cmd("cat", os.path.join(root, "test-suite.log"))
raise Exception("Unit tests failed")
def analyze(self):
run_cppcheck()
class CMake(BuildSystem):
def __init__(self, package=None, path=None):
super(CMake, self).__init__(package, path)
def probe(self):
return os.path.isfile(os.path.join(self.path, "CMakeLists.txt"))
def dependencies(self):
return []
def configure(self, build_for_testing):
self.build_for_testing = build_for_testing
if INTEGRATION_TEST:
check_call_cmd(
"cmake",
"-DCMAKE_EXPORT_COMPILE_COMMANDS=ON",
"-DITESTS=ON",
".",
)
else:
check_call_cmd("cmake", "-DCMAKE_EXPORT_COMPILE_COMMANDS=ON", ".")
def build(self):
check_call_cmd(
"cmake",
"--build",
".",
"--",
"-j",
str(multiprocessing.cpu_count()),
)
def install(self):
pass
def test(self):
if make_target_exists("test"):
check_call_cmd("ctest", ".")
def analyze(self):
if os.path.isfile(".clang-tidy"):
with TemporaryDirectory(prefix="build", dir=".") as build_dir:
# clang-tidy needs to run on a clang-specific build
check_call_cmd(
"cmake",
"-DCMAKE_C_COMPILER=clang",
"-DCMAKE_CXX_COMPILER=clang++",
"-DCMAKE_EXPORT_COMPILE_COMMANDS=ON",
"-H.",
"-B" + build_dir,
)
check_call_cmd(
"run-clang-tidy", "-header-filter=.*", "-p", build_dir
)
maybe_make_valgrind()
maybe_make_coverage()
run_cppcheck()
class Meson(BuildSystem):
def __init__(self, package=None, path=None):
super(Meson, self).__init__(package, path)
def probe(self):
return os.path.isfile(os.path.join(self.path, "meson.build"))
def dependencies(self):
meson_build = os.path.join(self.path, "meson.build")
if not os.path.exists(meson_build):
return []
found_deps = []
for root, dirs, files in os.walk(self.path):
if "meson.build" not in files:
continue
with open(os.path.join(root, "meson.build"), "rt") as f:
build_contents = f.read()
pattern = r"dependency\('([^']*)'.*?\),?\n"
for match in re.finditer(pattern, build_contents):
group = match.group(1)
maybe_dep = DEPENDENCIES["PKG_CHECK_MODULES"].get(group)
if maybe_dep is not None:
found_deps.append(maybe_dep)
return found_deps
def _parse_options(self, options_file):
"""
Returns a set of options defined in the provides meson_options.txt file
Parameters:
options_file The file containing options
"""
oi = optinterpreter.OptionInterpreter("")
oi.process(options_file)
return oi.options
def _configure_boolean(self, val):
"""
Returns the meson flag which signifies the value
True is true which requires the boolean.
False is false which disables the boolean.
Parameters:
val The value being converted
"""
if val is True:
return "true"
elif val is False:
return "false"
else:
raise Exception("Bad meson boolean value")
def _configure_feature(self, val):
"""
Returns the meson flag which signifies the value
True is enabled which requires the feature.
False is disabled which disables the feature.
None is auto which autodetects the feature.
Parameters:
val The value being converted
"""
if val is True:
return "enabled"
elif val is False:
return "disabled"
elif val is None:
return "auto"
else:
raise Exception("Bad meson feature value")
def _configure_option(self, opts, key, val):
"""
Returns the meson flag which signifies the value
based on the type of the opt
Parameters:
opt The meson option which we are setting
val The value being converted
"""
if isinstance(opts[key], coredata.UserBooleanOption):
str_val = self._configure_boolean(val)
elif isinstance(opts[key], coredata.UserFeatureOption):
str_val = self._configure_feature(val)
else:
raise Exception("Unknown meson option type")
return "-D{}={}".format(key, str_val)
def configure(self, build_for_testing):
self.build_for_testing = build_for_testing
meson_options = {}
if os.path.exists("meson_options.txt"):
meson_options = self._parse_options("meson_options.txt")
meson_flags = [
"-Db_colorout=never",
"-Dwerror=true",
"-Dwarning_level=3",
]
if build_for_testing:
meson_flags.append("--buildtype=debug")
else:
meson_flags.append("--buildtype=debugoptimized")
if OptionKey("tests") in meson_options:
meson_flags.append(
self._configure_option(
meson_options, OptionKey("tests"), build_for_testing
)
)
if OptionKey("examples") in meson_options:
meson_flags.append(
self._configure_option(
meson_options, OptionKey("examples"), build_for_testing
)
)
if OptionKey("itests") in meson_options:
meson_flags.append(
self._configure_option(
meson_options, OptionKey("itests"), INTEGRATION_TEST
)
)
if MESON_FLAGS.get(self.package) is not None:
meson_flags.extend(MESON_FLAGS.get(self.package))
try:
check_call_cmd(
"meson", "setup", "--reconfigure", "build", *meson_flags
)
except Exception:
shutil.rmtree("build", ignore_errors=True)
check_call_cmd("meson", "setup", "build", *meson_flags)
def build(self):
check_call_cmd("ninja", "-C", "build")
def install(self):
check_call_cmd("sudo", "-n", "--", "ninja", "-C", "build", "install")
def test(self):
# It is useful to check various settings of the meson.build file
# for compatibility, such as meson_version checks. We shouldn't
# do this in the configure path though because it affects subprojects
# and dependencies as well, but we only want this applied to the
# project-under-test (otherwise an upstream dependency could fail
# this check without our control).
self._extra_meson_checks()
try:
test_args = ("--repeat", str(args.repeat), "-C", "build")
check_call_cmd("meson", "test", "--print-errorlogs", *test_args)
except CalledProcessError:
raise Exception("Unit tests failed")
def _setup_exists(self, setup):
"""
Returns whether the meson build supports the named test setup.
Parameter descriptions:
setup The setup target to check
"""
try:
with open(os.devnull, "w"):
output = subprocess.check_output(
[
"meson",
"test",
"-C",
"build",
"--setup",
setup,
"-t",
"0",
],
stderr=subprocess.STDOUT,
)
except CalledProcessError as e:
output = e.output
output = output.decode("utf-8")
return not re.search("Test setup .* not found from project", output)
def _maybe_valgrind(self):
"""
Potentially runs the unit tests through valgrind for the package
via `meson test`. The package can specify custom valgrind
configurations by utilizing add_test_setup() in a meson.build
"""
if not is_valgrind_safe():
sys.stderr.write("###### Skipping valgrind ######\n")
return
try:
if self._setup_exists("valgrind"):
check_call_cmd(
"meson",
"test",
"-t",
"10",
"-C",
"build",
"--print-errorlogs",
"--setup",
"valgrind",
)
else:
check_call_cmd(
"meson",
"test",
"-t",
"10",
"-C",
"build",
"--print-errorlogs",
"--wrapper",
"valgrind",
)
except CalledProcessError:
raise Exception("Valgrind tests failed")
def analyze(self):
self._maybe_valgrind()
# Run clang-tidy only if the project has a configuration
if os.path.isfile(".clang-tidy"):
os.environ["CXX"] = "clang++"
with TemporaryDirectory(prefix="build", dir=".") as build_dir:
check_call_cmd("meson", "setup", build_dir)
if not os.path.isfile(".openbmc-no-clang"):
check_call_cmd("meson", "compile", "-C", build_dir)
try:
check_call_cmd(
"run-clang-tidy", "-fix", "-format", "-p", build_dir
)
except subprocess.CalledProcessError:
check_call_cmd(
"git", "-C", CODE_SCAN_DIR, "--no-pager", "diff"
)
raise
# Run the basic clang static analyzer otherwise
else:
check_call_cmd("ninja", "-C", "build", "scan-build")
# Run tests through sanitizers
# b_lundef is needed if clang++ is CXX since it resolves the
# asan symbols at runtime only. We don't want to set it earlier
# in the build process to ensure we don't have undefined
# runtime code.
if is_sanitize_safe():
check_call_cmd(
"meson",
"configure",
"build",
"-Db_sanitize=address,undefined",
"-Db_lundef=false",
)
check_call_cmd(
"meson",
"test",
"-C",
"build",
"--print-errorlogs",
"--logbase",
"testlog-ubasan",
)
# TODO: Fix memory sanitizer
# check_call_cmd('meson', 'configure', 'build',
# '-Db_sanitize=memory')
# check_call_cmd('meson', 'test', '-C', 'build'
# '--logbase', 'testlog-msan')
check_call_cmd("meson", "configure", "build", "-Db_sanitize=none")
else:
sys.stderr.write("###### Skipping sanitizers ######\n")
# Run coverage checks
check_call_cmd("meson", "configure", "build", "-Db_coverage=true")
self.test()
# Only build coverage HTML if coverage files were produced
for root, dirs, files in os.walk("build"):
if any([f.endswith(".gcda") for f in files]):
check_call_cmd("ninja", "-C", "build", "coverage-html")
break
check_call_cmd("meson", "configure", "build", "-Db_coverage=false")
run_cppcheck()
def _extra_meson_checks(self):
with open(os.path.join(self.path, "meson.build"), "rt") as f:
build_contents = f.read()
# Find project's specified meson_version.
meson_version = None
pattern = r"meson_version:[^']*'([^']*)'"
for match in re.finditer(pattern, build_contents):
group = match.group(1)
meson_version = group
# C++20 requires at least Meson 0.57 but Meson itself doesn't
# identify this. Add to our unit-test checks so that we don't
# get a meson.build missing this.
pattern = r"'cpp_std=c\+\+20'"
for match in re.finditer(pattern, build_contents):
if not meson_version or not meson_version_compare(
meson_version, ">=0.57"
):
raise Exception(
"C++20 support requires specifying in meson.build: "
+ "meson_version: '>=0.57'"
)
class Package(object):
def __init__(self, name=None, path=None):
self.supported = [Meson, Autotools, CMake]
self.name = name
self.path = path
self.test_only = False
def build_systems(self):
instances = (system(self.name, self.path) for system in self.supported)
return (instance for instance in instances if instance.probe())
def build_system(self, preferred=None):
systems = list(self.build_systems())
if not systems:
return None
if preferred:
return {type(system): system for system in systems}[preferred]
return next(iter(systems))
def install(self, system=None):
if not system:
system = self.build_system()
system.configure(False)
system.build()
system.install()
def _test_one(self, system):
system.configure(True)
system.build()
system.install()
system.test()
if not TEST_ONLY:
system.analyze()
def test(self):
for system in self.build_systems():
self._test_one(system)
def find_file(filename, basedir):
"""
Finds all occurrences of a file (or list of files) in the base
directory and passes them back with their relative paths.
Parameter descriptions:
filename The name of the file (or list of files) to
find
basedir The base directory search in
"""
if not isinstance(filename, list):
filename = [filename]
filepaths = []
for root, dirs, files in os.walk(basedir):
if os.path.split(root)[-1] == "subprojects":
for f in files:
subproject = ".".join(f.split(".")[0:-1])
if f.endswith(".wrap") and subproject in dirs:
# don't find files in meson subprojects with wraps
dirs.remove(subproject)
for f in filename:
if f in files:
filepaths.append(os.path.join(root, f))
return filepaths
if __name__ == "__main__":
# CONFIGURE_FLAGS = [GIT REPO]:[CONFIGURE FLAGS]
CONFIGURE_FLAGS = {
"phosphor-logging": [
"--enable-metadata-processing",
"--enable-openpower-pel-extension",
"YAML_DIR=/usr/local/share/phosphor-dbus-yaml/yaml",
]
}
# MESON_FLAGS = [GIT REPO]:[MESON FLAGS]
MESON_FLAGS = {
"phosphor-dbus-interfaces": [
"-Ddata_com_ibm=true",
"-Ddata_org_open_power=true",
],
"phosphor-logging": ["-Dopenpower-pel-extension=enabled"],
}
# DEPENDENCIES = [MACRO]:[library/header]:[GIT REPO]
DEPENDENCIES = {
"AC_CHECK_LIB": {"mapper": "phosphor-objmgr"},
"AC_CHECK_HEADER": {
"host-ipmid": "phosphor-host-ipmid",
"blobs-ipmid": "phosphor-ipmi-blobs",
"sdbusplus": "sdbusplus",
"sdeventplus": "sdeventplus",
"stdplus": "stdplus",
"gpioplus": "gpioplus",
"phosphor-logging/log.hpp": "phosphor-logging",
},
"AC_PATH_PROG": {"sdbus++": "sdbusplus"},
"PKG_CHECK_MODULES": {
"phosphor-dbus-interfaces": "phosphor-dbus-interfaces",
"libipmid": "phosphor-host-ipmid",
"libipmid-host": "phosphor-host-ipmid",
"sdbusplus": "sdbusplus",
"sdeventplus": "sdeventplus",
"stdplus": "stdplus",
"gpioplus": "gpioplus",
"phosphor-logging": "phosphor-logging",
"phosphor-snmp": "phosphor-snmp",
"ipmiblob": "ipmi-blob-tool",
"hei": "openpower-libhei",
"phosphor-ipmi-blobs": "phosphor-ipmi-blobs",
"libcr51sign": "google-misc",
},
}
# Offset into array of macro parameters MACRO(0, 1, ...N)
DEPENDENCIES_OFFSET = {
"AC_CHECK_LIB": 0,
"AC_CHECK_HEADER": 0,
"AC_PATH_PROG": 1,
"PKG_CHECK_MODULES": 1,
}
# DEPENDENCIES_REGEX = [GIT REPO]:[REGEX STRING]
DEPENDENCIES_REGEX = {"phosphor-logging": r"\S+-dbus-interfaces$"}
# Set command line arguments
parser = argparse.ArgumentParser()
parser.add_argument(
"-w",
"--workspace",
dest="WORKSPACE",
required=True,
help="Workspace directory location(i.e. /home)",
)
parser.add_argument(
"-p",
"--package",
dest="PACKAGE",
required=True,
help="OpenBMC package to be unit tested",
)
parser.add_argument(
"-t",
"--test-only",
dest="TEST_ONLY",
action="store_true",
required=False,
default=False,
help="Only run test cases, no other validation",
)
arg_inttests = parser.add_mutually_exclusive_group()
arg_inttests.add_argument(
"--integration-tests",
dest="INTEGRATION_TEST",
action="store_true",
required=False,
default=True,
help="Enable integration tests [default].",
)
arg_inttests.add_argument(
"--no-integration-tests",
dest="INTEGRATION_TEST",
action="store_false",
required=False,
help="Disable integration tests.",
)
parser.add_argument(
"-v",
"--verbose",
action="store_true",
help="Print additional package status messages",
)
parser.add_argument(
"-r", "--repeat", help="Repeat tests N times", type=int, default=1
)
parser.add_argument(
"-b",
"--branch",
dest="BRANCH",
required=False,
help="Branch to target for dependent repositories",
default="master",
)
parser.add_argument(
"-n",
"--noformat",
dest="FORMAT",
action="store_false",
required=False,
help="Whether or not to run format code",
)
args = parser.parse_args(sys.argv[1:])
WORKSPACE = args.WORKSPACE
UNIT_TEST_PKG = args.PACKAGE
TEST_ONLY = args.TEST_ONLY
INTEGRATION_TEST = args.INTEGRATION_TEST
BRANCH = args.BRANCH
FORMAT_CODE = args.FORMAT
if args.verbose:
def printline(*line):
for arg in line:
print(arg, end=" ")
print()
else:
def printline(*line):
pass
CODE_SCAN_DIR = os.path.join(WORKSPACE, UNIT_TEST_PKG)
# Run format-code.sh, which will in turn call any repo-level formatters.
if FORMAT_CODE:
check_call_cmd(
os.path.join(
WORKSPACE, "openbmc-build-scripts", "scripts", "format-code.sh"
),
CODE_SCAN_DIR,
)
# Check to see if any files changed
check_call_cmd(
"git", "-C", CODE_SCAN_DIR, "--no-pager", "diff", "--exit-code"
)
# Check if this repo has a supported make infrastructure
pkg = Package(UNIT_TEST_PKG, CODE_SCAN_DIR)
if not pkg.build_system():
print("No valid build system, exit")
sys.exit(0)
prev_umask = os.umask(000)
# Determine dependencies and add them
dep_added = dict()
dep_added[UNIT_TEST_PKG] = False
# Create dependency tree
dep_tree = DepTree(UNIT_TEST_PKG)
build_dep_tree(UNIT_TEST_PKG, CODE_SCAN_DIR, dep_added, dep_tree, BRANCH)
# Reorder Dependency Tree
for pkg_name, regex_str in DEPENDENCIES_REGEX.items():
dep_tree.ReorderDeps(pkg_name, regex_str)
if args.verbose:
dep_tree.PrintTree()
install_list = dep_tree.GetInstallList()
# We don't want to treat our package as a dependency
install_list.remove(UNIT_TEST_PKG)
# Install reordered dependencies
for dep in install_list:
build_and_install(dep, False)
# Run package unit tests
build_and_install(UNIT_TEST_PKG, True)
os.umask(prev_umask)
# Run any custom CI scripts the repo has, of which there can be
# multiple of and anywhere in the repository.
ci_scripts = find_file(["run-ci.sh", "run-ci"], CODE_SCAN_DIR)
if ci_scripts:
os.chdir(CODE_SCAN_DIR)
for ci_script in ci_scripts:
check_call_cmd(ci_script)