| # IceCream distributed compiling support |
| # |
| # Stages directories with symlinks from gcc/g++ to icecc, for both |
| # native and cross compilers. Depending on each configure or compile, |
| # the directories are added at the head of the PATH list and ICECC_CXX |
| # and ICEC_CC are set. |
| # |
| # For the cross compiler, creates a tar.gz of our toolchain and sets |
| # ICECC_VERSION accordingly. |
| # |
| # The class now handles all 3 different compile 'stages' (i.e native ,cross-kernel and target) creating the |
| # necessary environment tar.gz file to be used by the remote machines. |
| # It also supports meta-toolchain generation |
| # |
| # If ICECC_PATH is not set in local.conf then the class will try to locate it using 'bb.utils.which' |
| # but nothing is sure ;) |
| # |
| # If ICECC_ENV_EXEC is set in local.conf, then it should point to the icecc-create-env script provided by the user |
| # or the default one provided by icecc-create-env.bb will be used |
| # (NOTE that this is a modified version of the script need it and *not the one that comes with icecc* |
| # |
| # User can specify if specific recipes or recipes belonging to class should not use icecc to distribute |
| # compile jobs to remote machines, but handled locally, by defining ICECC_CLASS_DISABLE and ICECC_RECIPE_DISABLE |
| # with the appropriate values in local.conf. In addition the user can force to enable icecc for recipes |
| # which set an empty PARALLEL_MAKE variable by defining ICECC_RECIPE_ENABLE. |
| # |
| ######################################################################################### |
| #Error checking is kept to minimum so double check any parameters you pass to the class |
| ########################################################################################### |
| |
| BB_BASEHASH_IGNORE_VARS += "ICECC_PARALLEL_MAKE ICECC_DISABLED ICECC_RECIPE_DISABLE \ |
| ICECC_CLASS_DISABLE ICECC_RECIPE_ENABLE ICECC_PATH ICECC_ENV_EXEC \ |
| ICECC_CARET_WORKAROUND ICECC_CFLAGS ICECC_ENV_VERSION \ |
| ICECC_DEBUG ICECC_LOGFILE ICECC_REPEAT_RATE ICECC_PREFERRED_HOST \ |
| ICECC_CLANG_REMOTE_CPP ICECC_IGNORE_UNVERIFIED ICECC_TEST_SOCKET \ |
| ICECC_ENV_DEBUG ICECC_REMOTE_CPP \ |
| " |
| |
| ICECC_ENV_EXEC ?= "${STAGING_BINDIR_NATIVE}/icecc-create-env" |
| |
| HOSTTOOLS_NONFATAL += "icecc patchelf" |
| |
| # This version can be incremented when changes are made to the environment that |
| # invalidate the version on the compile nodes. Changing it will cause a new |
| # environment to be created. |
| # |
| # A useful thing to do for testing Icecream changes locally is to add a |
| # subversion in local.conf: |
| # ICECC_ENV_VERSION:append = "-my-ver-1" |
| ICECC_ENV_VERSION = "2" |
| |
| # Default to disabling the caret workaround, If set to "1" in local.conf, icecc |
| # will locally recompile any files that have warnings, which can adversely |
| # affect performance. |
| # |
| # See: https://github.com/icecc/icecream/issues/190 |
| export ICECC_CARET_WORKAROUND ??= "0" |
| |
| export ICECC_REMOTE_CPP ??= "0" |
| |
| ICECC_CFLAGS = "" |
| CFLAGS += "${ICECC_CFLAGS}" |
| CXXFLAGS += "${ICECC_CFLAGS}" |
| |
| # Debug flags when generating environments |
| ICECC_ENV_DEBUG ??= "" |
| |
| # Disable recipe list contains a list of recipes that can not distribute |
| # compile tasks for one reason or the other. When adding new entry, please |
| # document why (how it failed) so that we can re-evaluate it later e.g. when |
| # there is new version |
| # |
| # libgcc-initial - fails with CPP sanity check error if host sysroot contains |
| # cross gcc built for another target tune/variant |
| # pixman - prng_state: TLS reference mismatches non-TLS reference, possibly due to |
| # pragma omp threadprivate(prng_state) |
| # systemtap - _HelperSDT.c undefs macros and uses the identifiers in macros emitting |
| # inline assembly |
| # target-sdk-provides-dummy - ${HOST_PREFIX} is empty which triggers the "NULL |
| # prefix" error. |
| ICECC_RECIPE_DISABLE += "\ |
| libgcc-initial \ |
| pixman \ |
| systemtap \ |
| target-sdk-provides-dummy \ |
| " |
| |
| # Classes that should not use icecc. When adding new entry, please |
| # document why (how it failed) so that we can re-evaluate it later |
| # |
| # image - Image aren't compiling, but the testing framework for images captures |
| # PARALLEL_MAKE as part of the test environment. Many tests won't use |
| # icecream, but leaving the high level of parallelism can cause them to |
| # consume an unnecessary amount of resources. |
| ICECC_CLASS_DISABLE += "\ |
| image \ |
| " |
| |
| def get_icecc_dep(d): |
| # INHIBIT_DEFAULT_DEPS doesn't apply to the patch command. Whether or not |
| # we need that built is the responsibility of the patch function / class, not |
| # the application. |
| if not d.getVar('INHIBIT_DEFAULT_DEPS'): |
| return "icecc-create-env-native" |
| return "" |
| |
| DEPENDS:prepend = "${@get_icecc_dep(d)} " |
| |
| get_cross_kernel_cc[vardepsexclude] += "KERNEL_CC" |
| def get_cross_kernel_cc(bb,d): |
| if not icecc_is_kernel(bb, d): |
| return None |
| |
| # evaluate the expression by the shell if necessary |
| kernel_cc = d.getVar('KERNEL_CC') |
| if '`' in kernel_cc or '$(' in kernel_cc: |
| import subprocess |
| kernel_cc = subprocess.check_output("echo %s" % kernel_cc, shell=True).decode("utf-8")[:-1] |
| |
| kernel_cc = kernel_cc.replace('ccache', '').strip() |
| kernel_cc = kernel_cc.split(' ')[0] |
| kernel_cc = kernel_cc.strip() |
| return kernel_cc |
| |
| def get_icecc(d): |
| return d.getVar('ICECC_PATH') or bb.utils.which(os.getenv("PATH"), "icecc") |
| |
| def use_icecc(bb,d): |
| if d.getVar('ICECC_DISABLED') == "1": |
| # don't even try it, when explicitly disabled |
| return "no" |
| |
| # allarch recipes don't use compiler |
| if icecc_is_allarch(bb, d): |
| return "no" |
| |
| if icecc_is_cross_canadian(bb, d): |
| return "no" |
| |
| pn = d.getVar('PN') |
| bpn = d.getVar('BPN') |
| |
| # Enable/disable checks are made against BPN, because there is a good |
| # chance that if icecc should be skipped for a recipe, it should be skipped |
| # for all the variants of that recipe. PN is still checked in case a user |
| # specified a more specific recipe. |
| check_pn = set([pn, bpn]) |
| |
| class_disable = (d.getVar('ICECC_CLASS_DISABLE') or "").split() |
| |
| for bbclass in class_disable: |
| if bb.data.inherits_class(bbclass, d): |
| bb.debug(1, "%s: bbclass %s found in disable, disable icecc" % (pn, bbclass)) |
| return "no" |
| |
| disabled_recipes = (d.getVar('ICECC_RECIPE_DISABLE') or "").split() |
| enabled_recipes = (d.getVar('ICECC_RECIPE_ENABLE') or "").split() |
| |
| if check_pn & set(disabled_recipes): |
| bb.debug(1, "%s: found in disable list, disable icecc" % pn) |
| return "no" |
| |
| if check_pn & set(enabled_recipes): |
| bb.debug(1, "%s: found in enabled recipes list, enable icecc" % pn) |
| return "yes" |
| |
| if d.getVar('PARALLEL_MAKE') == "": |
| bb.debug(1, "%s: has empty PARALLEL_MAKE, disable icecc" % pn) |
| return "no" |
| |
| return "yes" |
| |
| def icecc_is_allarch(bb, d): |
| return d.getVar("PACKAGE_ARCH") == "all" |
| |
| def icecc_is_kernel(bb, d): |
| return \ |
| bb.data.inherits_class("kernel", d); |
| |
| def icecc_is_native(bb, d): |
| return \ |
| bb.data.inherits_class("cross", d) or \ |
| bb.data.inherits_class("native", d); |
| |
| def icecc_is_cross_canadian(bb, d): |
| return bb.data.inherits_class("cross-canadian", d) |
| |
| def icecc_dir(bb, d): |
| return d.expand('${TMPDIR}/work-shared/ice') |
| |
| # Don't pollute allarch signatures with TARGET_FPU |
| icecc_version[vardepsexclude] += "TARGET_FPU" |
| def icecc_version(bb, d): |
| if use_icecc(bb, d) == "no": |
| return "" |
| |
| parallel = d.getVar('ICECC_PARALLEL_MAKE') or "" |
| if not d.getVar('PARALLEL_MAKE') == "" and parallel: |
| d.setVar("PARALLEL_MAKE", parallel) |
| |
| # Disable showing the caret in the GCC compiler output if the workaround is |
| # disabled |
| if d.getVar('ICECC_CARET_WORKAROUND') == '0': |
| d.setVar('ICECC_CFLAGS', '-fno-diagnostics-show-caret') |
| |
| if icecc_is_native(bb, d): |
| archive_name = "local-host-env" |
| elif d.expand('${HOST_PREFIX}') == "": |
| bb.fatal(d.expand("${PN}"), " NULL prefix") |
| else: |
| prefix = d.expand('${HOST_PREFIX}' ) |
| distro = d.expand('${DISTRO}') |
| target_sys = d.expand('${TARGET_SYS}') |
| float = d.getVar('TARGET_FPU') or "hard" |
| archive_name = prefix + distro + "-" + target_sys + "-" + float |
| if icecc_is_kernel(bb, d): |
| archive_name += "-kernel" |
| |
| import socket |
| ice_dir = icecc_dir(bb, d) |
| tar_file = os.path.join(ice_dir, "{archive}-{version}-@VERSION@-{hostname}.tar.gz".format( |
| archive=archive_name, |
| version=d.getVar('ICECC_ENV_VERSION'), |
| hostname=socket.gethostname() |
| )) |
| |
| return tar_file |
| |
| def icecc_path(bb,d): |
| if use_icecc(bb, d) == "no": |
| # don't create unnecessary directories when icecc is disabled |
| return |
| |
| staging = os.path.join(d.expand('${STAGING_BINDIR}'), "ice") |
| if icecc_is_kernel(bb, d): |
| staging += "-kernel" |
| |
| return staging |
| |
| def icecc_get_external_tool(bb, d, tool): |
| external_toolchain_bindir = d.expand('${EXTERNAL_TOOLCHAIN}${bindir_cross}') |
| target_prefix = d.expand('${TARGET_PREFIX}') |
| return os.path.join(external_toolchain_bindir, '%s%s' % (target_prefix, tool)) |
| |
| def icecc_get_tool_link(tool, d): |
| import subprocess |
| try: |
| return subprocess.check_output("readlink -f %s" % tool, shell=True).decode("utf-8")[:-1] |
| except subprocess.CalledProcessError as e: |
| bb.note("icecc: one of the tools probably disappeared during recipe parsing, cmd readlink -f %s returned %d:\n%s" % (tool, e.returncode, e.output.decode("utf-8"))) |
| return tool |
| |
| def icecc_get_path_tool(tool, d): |
| # This is a little ugly, but we want to make sure we add an actual |
| # compiler to the toolchain, not ccache. Some distros (e.g. Fedora) |
| # have ccache enabled by default using symlinks PATH, meaning ccache |
| # would be found first when looking for the compiler. |
| paths = os.getenv("PATH").split(':') |
| while True: |
| p, hist = bb.utils.which(':'.join(paths), tool, history=True) |
| if not p or os.path.basename(icecc_get_tool_link(p, d)) != 'ccache': |
| return p |
| paths = paths[len(hist):] |
| |
| return "" |
| |
| # Don't pollute native signatures with target TUNE_PKGARCH through STAGING_BINDIR_TOOLCHAIN |
| icecc_get_tool[vardepsexclude] += "STAGING_BINDIR_TOOLCHAIN" |
| def icecc_get_tool(bb, d, tool): |
| if icecc_is_native(bb, d): |
| return icecc_get_path_tool(tool, d) |
| elif icecc_is_kernel(bb, d): |
| return icecc_get_path_tool(get_cross_kernel_cc(bb, d), d) |
| else: |
| ice_dir = d.expand('${STAGING_BINDIR_TOOLCHAIN}') |
| target_sys = d.expand('${TARGET_SYS}') |
| for p in ice_dir.split(':'): |
| tool_bin = os.path.join(p, "%s-%s" % (target_sys, tool)) |
| if os.path.isfile(tool_bin): |
| return tool_bin |
| external_tool_bin = icecc_get_external_tool(bb, d, tool) |
| if os.path.isfile(external_tool_bin): |
| return external_tool_bin |
| return "" |
| |
| def icecc_get_and_check_tool(bb, d, tool): |
| # Check that g++ or gcc is not a symbolic link to icecc binary in |
| # PATH or icecc-create-env script will silently create an invalid |
| # compiler environment package. |
| t = icecc_get_tool(bb, d, tool) |
| if t: |
| link_path = icecc_get_tool_link(t, d) |
| if link_path == get_icecc(d): |
| bb.error("%s is a symlink to %s in PATH and this prevents icecc from working" % (t, link_path)) |
| return "" |
| else: |
| return t |
| else: |
| return t |
| |
| wait_for_file() { |
| local TIME_ELAPSED=0 |
| local FILE_TO_TEST=$1 |
| local TIMEOUT=$2 |
| until [ -f "$FILE_TO_TEST" ] |
| do |
| TIME_ELAPSED=$(expr $TIME_ELAPSED + 1) |
| if [ $TIME_ELAPSED -gt $TIMEOUT ] |
| then |
| return 1 |
| fi |
| sleep 1 |
| done |
| } |
| |
| def set_icecc_env(): |
| # dummy python version of set_icecc_env |
| return |
| |
| set_icecc_env[vardepsexclude] += "KERNEL_CC" |
| set_icecc_env() { |
| if [ "${@use_icecc(bb, d)}" = "no" ] |
| then |
| return |
| fi |
| ICECC_VERSION="${@icecc_version(bb, d)}" |
| if [ "x${ICECC_VERSION}" = "x" ] |
| then |
| bbwarn "Cannot use icecc: could not get ICECC_VERSION" |
| return |
| fi |
| |
| ICE_PATH="${@icecc_path(bb, d)}" |
| if [ "x${ICE_PATH}" = "x" ] |
| then |
| bbwarn "Cannot use icecc: could not get ICE_PATH" |
| return |
| fi |
| |
| ICECC_BIN="${@get_icecc(d)}" |
| if [ -z "${ICECC_BIN}" ]; then |
| bbwarn "Cannot use icecc: icecc binary not found" |
| return |
| fi |
| if [ -z "$(which patchelf patchelf-uninative)" ]; then |
| bbwarn "Cannot use icecc: patchelf not found" |
| return |
| fi |
| |
| ICECC_CC="${@icecc_get_and_check_tool(bb, d, "gcc")}" |
| ICECC_CXX="${@icecc_get_and_check_tool(bb, d, "g++")}" |
| # cannot use icecc_get_and_check_tool here because it assumes as without target_sys prefix |
| ICECC_WHICH_AS="${@bb.utils.which(os.getenv('PATH'), 'as')}" |
| if [ ! -x "${ICECC_CC}" -o ! -x "${ICECC_CXX}" ] |
| then |
| bbnote "Cannot use icecc: could not get ICECC_CC or ICECC_CXX" |
| return |
| fi |
| |
| ICE_VERSION="$($ICECC_CC -dumpversion)" |
| ICECC_VERSION=$(echo ${ICECC_VERSION} | sed -e "s/@VERSION@/$ICE_VERSION/g") |
| if [ ! -x "${ICECC_ENV_EXEC}" ] |
| then |
| bbwarn "Cannot use icecc: invalid ICECC_ENV_EXEC" |
| return |
| fi |
| |
| # Create symlinks to icecc and wrapper-scripts in the recipe-sysroot directory |
| mkdir -p $ICE_PATH/symlinks |
| if [ -n "${KERNEL_CC}" ]; then |
| compilers="${@get_cross_kernel_cc(bb,d)}" |
| else |
| compilers="${HOST_PREFIX}gcc ${HOST_PREFIX}g++" |
| fi |
| for compiler in $compilers; do |
| ln -sf $ICECC_BIN $ICE_PATH/symlinks/$compiler |
| rm -f $ICE_PATH/$compiler |
| cat <<-__EOF__ > $ICE_PATH/$compiler |
| #!/bin/sh -e |
| export ICECC_VERSION=$ICECC_VERSION |
| export ICECC_CC=$ICECC_CC |
| export ICECC_CXX=$ICECC_CXX |
| $ICE_PATH/symlinks/$compiler "\$@" |
| __EOF__ |
| chmod 775 $ICE_PATH/$compiler |
| done |
| |
| ICECC_AS="$(${ICECC_CC} -print-prog-name=as)" |
| # for target recipes should return something like: |
| # /OE/tmp-eglibc/sysroots/x86_64-linux/usr/libexec/arm920tt-oe-linux-gnueabi/gcc/arm-oe-linux-gnueabi/4.8.2/as |
| # and just "as" for native, if it returns "as" in current directory (for whatever reason) use "as" from PATH |
| if [ "$(dirname "${ICECC_AS}")" = "." ] |
| then |
| ICECC_AS="${ICECC_WHICH_AS}" |
| fi |
| |
| if [ ! -f "${ICECC_VERSION}.done" ] |
| then |
| mkdir -p "$(dirname "${ICECC_VERSION}")" |
| |
| # the ICECC_VERSION generation step must be locked by a mutex |
| # in order to prevent race conditions |
| if flock -n "${ICECC_VERSION}.lock" \ |
| ${ICECC_ENV_EXEC} ${ICECC_ENV_DEBUG} "${ICECC_CC}" "${ICECC_CXX}" "${ICECC_AS}" "${ICECC_VERSION}" |
| then |
| touch "${ICECC_VERSION}.done" |
| elif ! wait_for_file "${ICECC_VERSION}.done" 30 |
| then |
| # locking failed so wait for ${ICECC_VERSION}.done to appear |
| bbwarn "Timeout waiting for ${ICECC_VERSION}.done" |
| return |
| fi |
| fi |
| |
| # Don't let ccache find the icecream compiler links that have been created, otherwise |
| # it can end up invoking icecream recursively. |
| export CCACHE_PATH="$PATH" |
| export CCACHE_DISABLE="1" |
| |
| export PATH="$ICE_PATH:$PATH" |
| |
| bbnote "Using icecc path: $ICE_PATH" |
| bbnote "Using icecc tarball: $ICECC_VERSION" |
| } |
| |
| do_configure[network] = "1" |
| do_configure:prepend() { |
| set_icecc_env |
| } |
| |
| do_compile[network] = "1" |
| do_compile:prepend() { |
| set_icecc_env |
| } |
| |
| do_compile_kernelmodules[network] = "1" |
| do_compile_kernelmodules:prepend() { |
| set_icecc_env |
| } |
| |
| do_install[network] = "1" |
| do_install:prepend() { |
| set_icecc_env |
| } |
| |
| # IceCream is not (currently) supported in the extensible SDK |
| ICECC_SDK_HOST_TASK = "nativesdk-icecc-toolchain" |
| ICECC_SDK_HOST_TASK:task-populate-sdk-ext = "" |
| |
| # Don't include IceCream in uninative tarball |
| ICECC_SDK_HOST_TASK:pn-uninative-tarball = "" |
| |
| # Add the toolchain scripts to the SDK |
| TOOLCHAIN_HOST_TASK:append = " ${ICECC_SDK_HOST_TASK}" |