#!/usr/bin/env bash # # SPDX-License-Identifier: GPL-2.0 # # Copyright (c) 2013-2023 Igor Pecovnik, igor@armbian.com # # This file is a part of the Armbian Build Framework # https://github.com/armbian/build/ function fetch_and_build_host_tools() { call_extension_method "fetch_sources_tools" <<- 'FETCH_SOURCES_TOOLS' *fetch host-side sources needed for tools and build* Run early to fetch_from_repo or otherwise obtain sources for needed tools. FETCH_SOURCES_TOOLS call_extension_method "build_host_tools" <<- 'BUILD_HOST_TOOLS' *build needed tools for the build, host-side* After sources are fetched, build host-side tools needed for the build. BUILD_HOST_TOOLS } # wait_for_package_manager # # * installation will break if we try to install when package manager is running # wait_for_package_manager() { # exit if package manager is running in the back while true; do if [[ "$( fuser /var/lib/dpkg/lock 2> /dev/null echo $? )" != 1 && "$( fuser /var/lib/dpkg/lock-frontend 2> /dev/null echo $? )" != 1 ]]; then display_alert "Package manager is running in the background." "Please wait! Retrying in 30 sec" "wrn" sleep 30 else break fi done } # Install the whitespace-delimited packages listed in the first parameter, in the host (not chroot). # It handles correctly the case where all wanted packages are already installed, and in that case does nothing. # If packages are to be installed, it does an apt-get update first. function install_host_side_packages() { declare wanted_packages_string PKG_TO_INSTALL declare -a currently_installed_packages declare -a missing_packages declare -a currently_provided_packages wanted_packages_string=${*} missing_packages=() # We need to jump through hoops to get the installed packages, due to the fact the "Provided" packages are a bit hidden. # Case in point: "gcc-aarch64-linux-gnu" is provided by "gcc" on native iron # If we don't do this, we keep on trying to apt install something that's already installed. # shellcheck disable=SC2207 # I wanna split, thanks. currently_installed_packages=($(dpkg-query --show --showformat='${Package} ')) # shellcheck disable=SC2207 # I wanna split, thanks. currently_provided_packages=($(dpkg-query --show --showformat='${Provides}\n' | grep -v "^$" | sed -e 's/([^()]*)//g' | sed -e 's|,||g' | tr -s "\n" " ")) for PKG_TO_INSTALL in ${wanted_packages_string}; do # shellcheck disable=SC2076 # I wanna match literally, thanks. if [[ ! " ${currently_installed_packages[*]} " =~ " ${PKG_TO_INSTALL} " ]]; then if [[ ! " ${currently_provided_packages[*]} " =~ " ${PKG_TO_INSTALL} " ]]; then missing_packages+=("${PKG_TO_INSTALL}") fi fi done unset currently_installed_packages unset currently_provided_packages if [[ ${#missing_packages[@]} -gt 0 ]]; then display_alert "Updating apt host-side for installing host-side packages" "${#missing_packages[@]} packages" "info" host_apt_get update display_alert "Installing host-side packages" "${missing_packages[*]}" "info" host_apt_get_install "${missing_packages[@]}" || exit_with_error "Failed to install host packages; make sure you have a sane sources.list." else display_alert "All host-side dependencies/packages already installed." "Skipping host-hide install" "debug" fi return 0 } function is_root_or_sudo_prefix() { declare -n __my_sudo_prefix=${1} # nameref... if [[ "${EUID}" == "0" ]]; then # do not use sudo if we're effectively already root display_alert "EUID=0, so" "we're already root!" "debug" __my_sudo_prefix="" elif [[ -n "$(command -v sudo)" ]]; then # sudo binary found in path, use it. display_alert "EUID is not 0" "sudo binary found, using it" "debug" __my_sudo_prefix="sudo" else # No root and no sudo binary. Bail out exit_with_error "EUID is not 0 and no sudo binary found - Please install sudo or run as root" fi return 0 } # usage: local_apt_deb_cache_prepare "using cache for xxx"; echo "${LOCAL_APT_CACHE_INFO[xyz]}" function local_apt_deb_cache_prepare() { declare when_used="${1}" # validate $RELEASE and $ARCH are set before we start if [[ -z "${RELEASE}" || -z "${ARCH}" ]]; then exit_with_error "RELEASE(=${RELEASE}) and ARCH(=${ARCH}) must be set before calling local_apt_deb_cache_prepare" fi # validate SDCARD is set and valid before we start if [[ -z "${SDCARD}" || ! -d "${SDCARD}" ]]; then exit_with_error "SDCARD(=${SDCARD}) must be set and valid before calling local_apt_deb_cache_prepare" fi if [[ "${USE_LOCAL_APT_DEB_CACHE}" != "yes" ]]; then # Not using the local cache, do nothing. Just return "no" in the first nameref. return 0 fi declare __my_use_yes_or_no="yes" declare __my_apt_cache_host_debs_dir="${SRC}/cache/aptcache/${RELEASE}-${ARCH}" declare __my_apt_cache_host_debs_dir_strap="${SRC}/cache/aptcache/${RELEASE}-${ARCH}/archives" declare __my_apt_cache_host_lists_dir="${SRC}/cache/aptcache/lists/${RELEASE}-${ARCH}" declare -a __my_all_dirs=("${__my_apt_cache_host_debs_dir}" "${__my_apt_cache_host_debs_dir_strap}" "${__my_apt_cache_host_lists_dir}") mkdir -p "${__my_all_dirs[@]}" if [[ "${SHOW_DEBUG}" == "yes" ]]; then # get the size, in bytes, of the cache directory, including subdirs declare -i cache_size # heh, mark var as integer cache_size=$(du -sb "${__my_apt_cache_host_debs_dir}" | cut -f1) display_alert "Size of apt/deb cache ${when_used}" "${cache_size} bytes" "debug" declare -g -i __previous_apt_cache_size if [[ -z "${__previous_apt_cache_size}" ]]; then # first time, set the size to 0 __previous_apt_cache_size=0 else # not first time, check if the size has changed if [[ "${cache_size}" -ne "${__previous_apt_cache_size}" ]]; then display_alert "Local apt cache size changed ${when_used}" "from ${__previous_apt_cache_size} to ${cache_size} bytes" "debug" else display_alert "Local apt cache size unchanged ${when_used}" "at ${cache_size} bytes" "debug" fi fi __previous_apt_cache_size=${cache_size} fi # set a global dictionary for easy usage. declare -g -A LOCAL_APT_CACHE_INFO=( [USE]="${__my_use_yes_or_no}" [HOST_DEBS_DIR]="${__my_apt_cache_host_debs_dir}" # Dir with debs [HOST_DEBOOTSTRAP_CACHE_DIR]="${__my_apt_cache_host_debs_dir_strap}" # Small difference for debootstrap, if compared to apt: we need to pass it the "/archives" subpath to share cache with apt. [HOST_LISTS_DIR]="${__my_apt_cache_host_lists_dir}" [SDCARD_DEBS_DIR]="${SDCARD}/var/cache/apt" [SDCARD_LISTS_DIR]="${SDCARD}/var/lib/apt/lists" [LAST_USED]="${when_used}" ) if [[ "${skip_target_check:-"no"}" != "yes" ]]; then # Lets take the chance here and _warn_ if the _target_ cache is not empty. Skip if dir doesn't exist or is a mountpoint. declare sdcard_var_cache_apt_dir="${SDCARD}/var/cache/apt" if [[ -d "${sdcard_var_cache_apt_dir}" ]]; then if ! mountpoint -q "${sdcard_var_cache_apt_dir}"; then declare -i sdcard_var_cache_apt_files_count sdcard_var_cache_apt_files_count=$(find "${sdcard_var_cache_apt_dir}" -type f | wc -l) if [[ "${sdcard_var_cache_apt_files_count}" -gt 1 ]]; then # 1 cos of lockfile that might or not be there display_alert "WARNING: SDCARD /var/cache/apt dir is not empty" "${when_used} :: ${sdcard_var_cache_apt_dir} (${sdcard_var_cache_apt_files_count} files)" "wrn" run_host_command_logged ls -lahtR "${sdcard_var_cache_apt_dir}" # list the contents so we can try and identify what is polluting it fi fi fi # Same, but for /var/lib/apt/lists declare sdcard_var_lib_apt_lists_dir="${SDCARD}/var/lib/apt/lists" if [[ -d "${sdcard_var_lib_apt_lists_dir}" ]]; then if ! mountpoint -q "${sdcard_var_lib_apt_lists_dir}"; then declare -i sdcard_var_lib_apt_lists_files_count sdcard_var_lib_apt_lists_files_count=$(find "${sdcard_var_lib_apt_lists_dir}" -type f | wc -l) if [[ "${sdcard_var_lib_apt_lists_files_count}" -gt 1 ]]; then # 1 cos of lockfile that might or not be there display_alert "WARNING: SDCARD /var/lib/apt/lists dir is not empty" "${when_used} :: ${sdcard_var_lib_apt_lists_dir} (${sdcard_var_lib_apt_lists_files_count} files)" "wrn" run_host_command_logged ls -lahtR "${sdcard_var_lib_apt_lists_dir}" # list the contents so we can try and identify what is polluting it fi fi fi fi return 0 } # usage: if armbian_is_host_running_systemd; then ... fi function armbian_is_host_running_systemd() { # Detect if systemctl is available in the path if [[ -n "$(command -v systemctl)" ]]; then display_alert "systemctl binary found" "host has systemd installed" "debug" # Detect if systemd is actively running if systemctl is-system-running --quiet; then display_alert "systemctl reports" "systemd is running" "debug" return 0 else display_alert "systemctl binary found" "but systemd is not running" "debug" return 1 fi else display_alert "systemctl binary not found" "host does not have systemd installed" "debug" fi # Not running with systemd, return 1. display_alert "Systemd not detected" "host is not running systemd" "debug" return 1 } # usage: if armbian_is_running_in_container; then ... fi function armbian_is_running_in_container() { # First, check an environment variable. This is passed by the docker launchers, and also set in the Dockerfile, so should be authoritative. if [[ "${ARMBIAN_RUNNING_IN_CONTAINER}" == "yes" ]]; then display_alert "ARMBIAN_RUNNING_IN_CONTAINER is set to 'yes' in the environment" "so we're running in a container/Docker" "debug" return 0 fi # Second, check the hardcoded path `/.dockerenv` -- not all Docker images have this, but if they do, we're pretty sure it is under Docker. if [[ -f "/.dockerenv" ]]; then display_alert "File /.dockerenv exists" "so we're running in a container/Docker" "debug" return 0 fi # Third: if host is actively running systemd (not just installed), it's very _unlikely_ that we're running under Docker. bail. if armbian_is_host_running_systemd; then display_alert "Host is running systemd" "so we're not running in a container/Docker" "debug" return 1 fi # Fourth, if `systemd-detect-virt` is available in the path, and executing it returns "docker", we're pretty sure it is under Docker. if [[ -n "$(command -v systemd-detect-virt)" ]]; then local systemd_detect_virt_output systemd_detect_virt_output="$(systemd-detect-virt)" if [[ "${systemd_detect_virt_output}" == "docker" ]]; then display_alert "systemd-detect-virt says we're running in a container/Docker" "so we're running in a container/Docker" "debug" return 0 else display_alert "systemd-detect-virt says we're running on '${systemd_detect_virt_output}'" "so we're not running in a container/Docker" "debug" fi fi # End of the line. I've nothing else to check here. We're not running in a container/Docker. display_alert "No evidence found that we're running in a container/Docker" "so we're not running in a container/Docker" "debug" return 1 } # This does `mkdir -p` on the parameters, and also sets it to be owned by the correct UID. # Call: armbian_mkdir_p_and_chown_to_user "dir1" "dir2" "dir3/dir4" function mkdir_recursive_and_set_uid_owner() { # loop over args... for dir in "$@"; do mkdir -p "${dir}" reset_uid_owner "${dir}" done } # Call: reset_uid_owner "one/file" "some/directory" "another/file" - is recursive if dir given function reset_uid_owner() { if [[ "x${SET_OWNER_TO_UID}x" == "xx" ]]; then return 0 # Nothing to do. fi # Loop over args.. local arg for arg in "$@"; do display_alert "reset_uid_owner: '${arg}' will be owner id '${SET_OWNER_TO_UID}'" "reset_uid_owner" "debug" if [[ -d "${arg}" ]]; then chown "${SET_OWNER_TO_UID}" "${arg}" find "${arg}" -uid 0 -print0 | xargs --no-run-if-empty -0 chown "${SET_OWNER_TO_UID}" elif [[ -f "${arg}" ]]; then chown "${SET_OWNER_TO_UID}" "${arg}" else display_alert "reset_uid_owner: '${arg}' is not a file or directory" "skipping" "debug" return 1 fi done } # Non recursive version of the above function reset_uid_owner_non_recursive() { if [[ "x${SET_OWNER_TO_UID}x" == "xx" ]]; then return 0 # Nothing to do. fi # Loop over args.. local arg for arg in "$@"; do display_alert "reset_uid_owner_non_recursive: '${arg}' will be owner id '${SET_OWNER_TO_UID}'" "reset_uid_owner_non_recursive" "debug" if [[ -d "${arg}" ]]; then chown "${SET_OWNER_TO_UID}" "${arg}" elif [[ -f "${arg}" ]]; then chown "${SET_OWNER_TO_UID}" "${arg}" else display_alert "reset_uid_owner_non_recursive: '${arg}' is not a file or directory" "skipping" "debug" return 1 fi done } # call: check_dir_for_mount_options "/path/to/dir" "main build dir description" function check_dir_for_mount_options() { declare -r dir="${1}" declare -r description="${2}" declare src_mount_source="" src_mount_opts="" src_mount_opts="$(findmnt -T "${dir}" --output OPTIONS --raw --notruncate --noheadings)" # make sure $src_mount_opts does not contain noexec if [[ "${src_mount_opts}" == *"noexec"* || "${src_mount_opts}" == *"nodev"* ]]; then src_mount_source="$(findmnt -T "${dir}" --output SOURCE --raw --notruncate --noheadings)" display_alert "Directory ${dir} (${description}) is mounted" "from '${src_mount_source}' with options '${src_mount_opts}'" "warn" exit_with_error "Directory ${dir} (${description}) is mounted with the 'noexec' and/or 'nodev' options; this will cause rootfs build failures. Please correct this before trying again." fi display_alert "Checked directory OK for mount options" "${dir} ('${description}')" "info" return 0 } function trap_handler_reset_output_owner() { display_alert "Resetting output directory owner" "${SRC}/output" "debug" reset_uid_owner "${SRC}/output" # For .tmp: do NOT do it recursively. If another build is running in another process, this is destructive if recursive. display_alert "Resetting tmp directory owner" "${SRC}/.tmp" "debug" reset_uid_owner_non_recursive "${SRC}/.tmp" } # Recursive function to find all descendant processes of a given PID. Writes to stdout. function list_descendants_of_pid() { local children children=$(ps -o "pid=" --ppid "$1" | xargs echo -n) for pid in $children; do list_descendants_of_pid "$pid" done echo -n "${children} " } function get_descendants_of_pid_array() { descendants_of_pid_array_result=() # outer scope variable, reset # if not on Linux, just return; "ps" shenanigans we use are not supported if [[ "${OSTYPE}" != "linux-gnu" ]]; then display_alert "get_descendants_of_pid_array: not on Linux, so not supported" "get_descendants_of_pid_array" "debug" return 0 fi local descendants descendants="$(list_descendants_of_pid "$1")" display_alert "Descendants of PID $1: ${descendants}" "string - get_descendants_of_pid_array" "debug" # shellcheck disable=SC2206 # lets expand! descendants_of_pid_array_result=(${descendants}) # outer scope variable display_alert "Descendants of PID $1: ${descendants_of_pid_array_result[*]}" "array = get_descendants_of_pid_array" "debug" } # This essentially runs "sync". It's a bit more complicated than that, though. # If "sync" takes more than 5 seconds, we want to let the user know. # So run sync under "timeout" in a loop, until it completes successfully. # call: wait_for_disk_sync "after writing a huge file to disk" function wait_for_disk_sync() { declare -i timeout_seconds=30 declare -i sync_worked=0 declare -i sync_timeout_count=0 declare -i total_wait=0 while [[ "${sync_worked}" -eq 0 ]]; do declare -i sync_timeout=0 display_alert "wait_for_disk_sync: $*" "sync start, timeout_seconds:${timeout_seconds}" "debug" # run a bash sending all to /dev/null, then send to /dev/null again, so we don't get "Killed" messages bash -c "timeout --signal=9 ${timeout_seconds} sync &> /dev/null" &> /dev/null && sync_worked=1 || sync_timeout=1 display_alert "wait_for_disk_sync: $*" "sync done worked: ${sync_worked} sync_timeout: ${sync_timeout}" "debug" if [[ "${sync_timeout}" -eq 1 ]]; then total_wait=$((total_wait + timeout_seconds)) sync_timeout_count=$((sync_timeout_count + 1)) display_alert "Waiting sync to storage $*" "fsync taking more than ${total_wait}s - be patient" "warn" else if [[ "${sync_timeout_count}" -gt 0 ]]; then display_alert "Sync to storage OK $*" "fsync took more than ${total_wait}s" "info" fi fi done }