build/lib/functions/host/host-utils.sh

391 lines
16 KiB
Bash
Raw Normal View History

#!/usr/bin/env bash
#
# SPDX-License-Identifier: GPL-2.0
#
# Copyright (c) 2013-2023 Igor Pecovnik, igor@armbian.com
#
# This file is a part of the Armbian Build Framework
# https://github.com/armbian/build/
function fetch_and_build_host_tools() {
call_extension_method "fetch_sources_tools" <<- 'FETCH_SOURCES_TOOLS'
*fetch host-side sources needed for tools and build*
Run early to fetch_from_repo or otherwise obtain sources for needed tools.
FETCH_SOURCES_TOOLS
call_extension_method "build_host_tools" <<- 'BUILD_HOST_TOOLS'
*build needed tools for the build, host-side*
After sources are fetched, build host-side tools needed for the build.
BUILD_HOST_TOOLS
}
# wait_for_package_manager
#
# * installation will break if we try to install when package manager is running
#
wait_for_package_manager() {
# exit if package manager is running in the back
while true; do
if [[ "$(
fuser /var/lib/dpkg/lock 2> /dev/null
echo $?
)" != 1 && "$(
fuser /var/lib/dpkg/lock-frontend 2> /dev/null
echo $?
)" != 1 ]]; then
display_alert "Package manager is running in the background." "Please wait! Retrying in 30 sec" "wrn"
sleep 30
else
break
fi
done
}
# Install the whitespace-delimited packages listed in the first parameter, in the host (not chroot).
# It handles correctly the case where all wanted packages are already installed, and in that case does nothing.
# If packages are to be installed, it does an apt-get update first.
function install_host_side_packages() {
declare wanted_packages_string PKG_TO_INSTALL
declare -a currently_installed_packages
declare -a missing_packages
declare -a currently_provided_packages
wanted_packages_string=${*}
missing_packages=()
# We need to jump through hoops to get the installed packages, due to the fact the "Provided" packages are a bit hidden.
# Case in point: "gcc-aarch64-linux-gnu" is provided by "gcc" on native iron
# If we don't do this, we keep on trying to apt install something that's already installed.
# shellcheck disable=SC2207 # I wanna split, thanks.
currently_installed_packages=($(dpkg-query --show --showformat='${Package} '))
# shellcheck disable=SC2207 # I wanna split, thanks.
currently_provided_packages=($(dpkg-query --show --showformat='${Provides}\n' | grep -v "^$" | sed -e 's/([^()]*)//g' | sed -e 's|,||g' | tr -s "\n" " "))
for PKG_TO_INSTALL in ${wanted_packages_string}; do
# shellcheck disable=SC2076 # I wanna match literally, thanks.
if [[ ! " ${currently_installed_packages[*]} " =~ " ${PKG_TO_INSTALL} " ]]; then
if [[ ! " ${currently_provided_packages[*]} " =~ " ${PKG_TO_INSTALL} " ]]; then
missing_packages+=("${PKG_TO_INSTALL}")
fi
fi
done
unset currently_installed_packages
unset currently_provided_packages
if [[ ${#missing_packages[@]} -gt 0 ]]; then
display_alert "Updating apt host-side for installing host-side packages" "${#missing_packages[@]} packages" "info"
host_apt_get update
display_alert "Installing host-side packages" "${missing_packages[*]}" "info"
host_apt_get_install "${missing_packages[@]}" || exit_with_error "Failed to install host packages; make sure you have a sane sources.list."
else
display_alert "All host-side dependencies/packages already installed." "Skipping host-hide install" "debug"
fi
return 0
}
function is_root_or_sudo_prefix() {
declare -n __my_sudo_prefix=${1} # nameref...
if [[ "${EUID}" == "0" ]]; then
# do not use sudo if we're effectively already root
display_alert "EUID=0, so" "we're already root!" "debug"
__my_sudo_prefix=""
elif [[ -n "$(command -v sudo)" ]]; then
# sudo binary found in path, use it.
display_alert "EUID is not 0" "sudo binary found, using it" "debug"
__my_sudo_prefix="sudo"
else
# No root and no sudo binary. Bail out
exit_with_error "EUID is not 0 and no sudo binary found - Please install sudo or run as root"
fi
return 0
}
# usage: local_apt_deb_cache_prepare "using cache for xxx"; echo "${LOCAL_APT_CACHE_INFO[xyz]}"
function local_apt_deb_cache_prepare() {
declare when_used="${1}"
# validate $RELEASE and $ARCH are set before we start
if [[ -z "${RELEASE}" || -z "${ARCH}" ]]; then
exit_with_error "RELEASE(=${RELEASE}) and ARCH(=${ARCH}) must be set before calling local_apt_deb_cache_prepare"
fi
# validate SDCARD is set and valid before we start
if [[ -z "${SDCARD}" || ! -d "${SDCARD}" ]]; then
exit_with_error "SDCARD(=${SDCARD}) must be set and valid before calling local_apt_deb_cache_prepare"
fi
if [[ "${USE_LOCAL_APT_DEB_CACHE}" != "yes" ]]; then
# Not using the local cache, do nothing. Just return "no" in the first nameref.
return 0
fi
declare __my_use_yes_or_no="yes"
declare __my_apt_cache_host_debs_dir="${SRC}/cache/aptcache/${RELEASE}-${ARCH}"
declare __my_apt_cache_host_debs_dir_strap="${SRC}/cache/aptcache/${RELEASE}-${ARCH}/archives"
declare __my_apt_cache_host_lists_dir="${SRC}/cache/aptcache/lists/${RELEASE}-${ARCH}"
declare -a __my_all_dirs=("${__my_apt_cache_host_debs_dir}" "${__my_apt_cache_host_debs_dir_strap}" "${__my_apt_cache_host_lists_dir}")
mkdir -p "${__my_all_dirs[@]}"
if [[ "${SHOW_DEBUG}" == "yes" ]]; then
# get the size, in bytes, of the cache directory, including subdirs
declare -i cache_size # heh, mark var as integer
cache_size=$(du -sb "${__my_apt_cache_host_debs_dir}" | cut -f1)
display_alert "Size of apt/deb cache ${when_used}" "${cache_size} bytes" "debug"
declare -g -i __previous_apt_cache_size
if [[ -z "${__previous_apt_cache_size}" ]]; then
# first time, set the size to 0
__previous_apt_cache_size=0
else
# not first time, check if the size has changed
if [[ "${cache_size}" -ne "${__previous_apt_cache_size}" ]]; then
display_alert "Local apt cache size changed ${when_used}" "from ${__previous_apt_cache_size} to ${cache_size} bytes" "debug"
else
display_alert "Local apt cache size unchanged ${when_used}" "at ${cache_size} bytes" "debug"
fi
fi
__previous_apt_cache_size=${cache_size}
fi
# set a global dictionary for easy usage.
declare -g -A LOCAL_APT_CACHE_INFO=(
[USE]="${__my_use_yes_or_no}"
[HOST_DEBS_DIR]="${__my_apt_cache_host_debs_dir}" # Dir with debs
[HOST_DEBOOTSTRAP_CACHE_DIR]="${__my_apt_cache_host_debs_dir_strap}" # Small difference for debootstrap, if compared to apt: we need to pass it the "/archives" subpath to share cache with apt.
[HOST_LISTS_DIR]="${__my_apt_cache_host_lists_dir}"
[SDCARD_DEBS_DIR]="${SDCARD}/var/cache/apt"
[SDCARD_LISTS_DIR]="${SDCARD}/var/lib/apt/lists"
[LAST_USED]="${when_used}"
)
if [[ "${skip_target_check:-"no"}" != "yes" ]]; then
# Lets take the chance here and _warn_ if the _target_ cache is not empty. Skip if dir doesn't exist or is a mountpoint.
declare sdcard_var_cache_apt_dir="${SDCARD}/var/cache/apt"
if [[ -d "${sdcard_var_cache_apt_dir}" ]]; then
if ! mountpoint -q "${sdcard_var_cache_apt_dir}"; then
declare -i sdcard_var_cache_apt_files_count
sdcard_var_cache_apt_files_count=$(find "${sdcard_var_cache_apt_dir}" -type f | wc -l)
if [[ "${sdcard_var_cache_apt_files_count}" -gt 1 ]]; then # 1 cos of lockfile that might or not be there
display_alert "WARNING: SDCARD /var/cache/apt dir is not empty" "${when_used} :: ${sdcard_var_cache_apt_dir} (${sdcard_var_cache_apt_files_count} files)" "wrn"
run_host_command_logged ls -lahtR "${sdcard_var_cache_apt_dir}" # list the contents so we can try and identify what is polluting it
fi
fi
fi
# Same, but for /var/lib/apt/lists
declare sdcard_var_lib_apt_lists_dir="${SDCARD}/var/lib/apt/lists"
if [[ -d "${sdcard_var_lib_apt_lists_dir}" ]]; then
if ! mountpoint -q "${sdcard_var_lib_apt_lists_dir}"; then
declare -i sdcard_var_lib_apt_lists_files_count
sdcard_var_lib_apt_lists_files_count=$(find "${sdcard_var_lib_apt_lists_dir}" -type f | wc -l)
if [[ "${sdcard_var_lib_apt_lists_files_count}" -gt 1 ]]; then # 1 cos of lockfile that might or not be there
display_alert "WARNING: SDCARD /var/lib/apt/lists dir is not empty" "${when_used} :: ${sdcard_var_lib_apt_lists_dir} (${sdcard_var_lib_apt_lists_files_count} files)" "wrn"
run_host_command_logged ls -lahtR "${sdcard_var_lib_apt_lists_dir}" # list the contents so we can try and identify what is polluting it
fi
fi
fi
fi
return 0
}
# usage: if armbian_is_host_running_systemd; then ... fi
function armbian_is_host_running_systemd() {
# Detect if systemctl is available in the path
if [[ -n "$(command -v systemctl)" ]]; then
display_alert "systemctl binary found" "host has systemd installed" "debug"
# Detect if systemd is actively running
if systemctl is-system-running --quiet; then
display_alert "systemctl reports" "systemd is running" "debug"
return 0
else
display_alert "systemctl binary found" "but systemd is not running" "debug"
return 1
fi
else
display_alert "systemctl binary not found" "host does not have systemd installed" "debug"
fi
# Not running with systemd, return 1.
display_alert "Systemd not detected" "host is not running systemd" "debug"
return 1
}
# usage: if armbian_is_running_in_container; then ... fi
function armbian_is_running_in_container() {
# First, check an environment variable. This is passed by the docker launchers, and also set in the Dockerfile, so should be authoritative.
if [[ "${ARMBIAN_RUNNING_IN_CONTAINER}" == "yes" ]]; then
display_alert "ARMBIAN_RUNNING_IN_CONTAINER is set to 'yes' in the environment" "so we're running in a container/Docker" "debug"
return 0
fi
# Second, check the hardcoded path `/.dockerenv` -- not all Docker images have this, but if they do, we're pretty sure it is under Docker.
if [[ -f "/.dockerenv" ]]; then
display_alert "File /.dockerenv exists" "so we're running in a container/Docker" "debug"
return 0
fi
# Third: if host is actively running systemd (not just installed), it's very _unlikely_ that we're running under Docker. bail.
if armbian_is_host_running_systemd; then
display_alert "Host is running systemd" "so we're not running in a container/Docker" "debug"
return 1
fi
# Fourth, if `systemd-detect-virt` is available in the path, and executing it returns "docker", we're pretty sure it is under Docker.
if [[ -n "$(command -v systemd-detect-virt)" ]]; then
local systemd_detect_virt_output
systemd_detect_virt_output="$(systemd-detect-virt)"
if [[ "${systemd_detect_virt_output}" == "docker" ]]; then
display_alert "systemd-detect-virt says we're running in a container/Docker" "so we're running in a container/Docker" "debug"
return 0
else
display_alert "systemd-detect-virt says we're running on '${systemd_detect_virt_output}'" "so we're not running in a container/Docker" "debug"
fi
fi
# End of the line. I've nothing else to check here. We're not running in a container/Docker.
display_alert "No evidence found that we're running in a container/Docker" "so we're not running in a container/Docker" "debug"
return 1
}
# This does `mkdir -p` on the parameters, and also sets it to be owned by the correct UID.
# Call: armbian_mkdir_p_and_chown_to_user "dir1" "dir2" "dir3/dir4"
function mkdir_recursive_and_set_uid_owner() {
# loop over args...
for dir in "$@"; do
mkdir -p "${dir}"
reset_uid_owner "${dir}"
done
}
# Call: reset_uid_owner "one/file" "some/directory" "another/file" - is recursive if dir given
function reset_uid_owner() {
if [[ "x${SET_OWNER_TO_UID}x" == "xx" ]]; then
return 0 # Nothing to do.
fi
# Loop over args..
local arg
for arg in "$@"; do
display_alert "reset_uid_owner: '${arg}' will be owner id '${SET_OWNER_TO_UID}'" "reset_uid_owner" "debug"
if [[ -d "${arg}" ]]; then
chown "${SET_OWNER_TO_UID}" "${arg}"
find "${arg}" -uid 0 -print0 | xargs --no-run-if-empty -0 chown "${SET_OWNER_TO_UID}"
elif [[ -f "${arg}" ]]; then
chown "${SET_OWNER_TO_UID}" "${arg}"
else
display_alert "reset_uid_owner: '${arg}' is not a file or directory" "skipping" "debug"
return 1
fi
done
}
# Non recursive version of the above
function reset_uid_owner_non_recursive() {
if [[ "x${SET_OWNER_TO_UID}x" == "xx" ]]; then
return 0 # Nothing to do.
fi
# Loop over args..
local arg
for arg in "$@"; do
display_alert "reset_uid_owner_non_recursive: '${arg}' will be owner id '${SET_OWNER_TO_UID}'" "reset_uid_owner_non_recursive" "debug"
if [[ -d "${arg}" ]]; then
chown "${SET_OWNER_TO_UID}" "${arg}"
elif [[ -f "${arg}" ]]; then
chown "${SET_OWNER_TO_UID}" "${arg}"
else
display_alert "reset_uid_owner_non_recursive: '${arg}' is not a file or directory" "skipping" "debug"
return 1
fi
done
}
# call: check_dir_for_mount_options "/path/to/dir" "main build dir description"
function check_dir_for_mount_options() {
declare -r dir="${1}"
declare -r description="${2}"
declare src_mount_source="" src_mount_opts=""
src_mount_opts="$(findmnt -T "${dir}" --output OPTIONS --raw --notruncate --noheadings)"
# make sure $src_mount_opts does not contain noexec
if [[ "${src_mount_opts}" == *"noexec"* || "${src_mount_opts}" == *"nodev"* ]]; then
src_mount_source="$(findmnt -T "${dir}" --output SOURCE --raw --notruncate --noheadings)"
display_alert "Directory ${dir} (${description}) is mounted" "from '${src_mount_source}' with options '${src_mount_opts}'" "warn"
exit_with_error "Directory ${dir} (${description}) is mounted with the 'noexec' and/or 'nodev' options; this will cause rootfs build failures. Please correct this before trying again."
fi
display_alert "Checked directory OK for mount options" "${dir} ('${description}')" "info"
return 0
}
function trap_handler_reset_output_owner() {
display_alert "Resetting output directory owner" "${SRC}/output" "debug"
reset_uid_owner "${SRC}/output"
# For .tmp: do NOT do it recursively. If another build is running in another process, this is destructive if recursive.
display_alert "Resetting tmp directory owner" "${SRC}/.tmp" "debug"
reset_uid_owner_non_recursive "${SRC}/.tmp"
}
# Recursive function to find all descendant processes of a given PID. Writes to stdout.
function list_descendants_of_pid() {
local children
children=$(ps -o "pid=" --ppid "$1" | xargs echo -n)
for pid in $children; do
list_descendants_of_pid "$pid"
done
echo -n "${children} "
}
function get_descendants_of_pid_array() {
descendants_of_pid_array_result=() # outer scope variable, reset
# if not on Linux, just return; "ps" shenanigans we use are not supported
if [[ "${OSTYPE}" != "linux-gnu" ]]; then
display_alert "get_descendants_of_pid_array: not on Linux, so not supported" "get_descendants_of_pid_array" "debug"
return 0
fi
local descendants
descendants="$(list_descendants_of_pid "$1")"
display_alert "Descendants of PID $1: ${descendants}" "string - get_descendants_of_pid_array" "debug"
# shellcheck disable=SC2206 # lets expand!
descendants_of_pid_array_result=(${descendants}) # outer scope variable
display_alert "Descendants of PID $1: ${descendants_of_pid_array_result[*]}" "array = get_descendants_of_pid_array" "debug"
}
# This essentially runs "sync". It's a bit more complicated than that, though.
# If "sync" takes more than 5 seconds, we want to let the user know.
# So run sync under "timeout" in a loop, until it completes successfully.
# call: wait_for_disk_sync "after writing a huge file to disk"
function wait_for_disk_sync() {
declare -i timeout_seconds=30
declare -i sync_worked=0
declare -i sync_timeout_count=0
declare -i total_wait=0
while [[ "${sync_worked}" -eq 0 ]]; do
declare -i sync_timeout=0
display_alert "wait_for_disk_sync: $*" "sync start, timeout_seconds:${timeout_seconds}" "debug"
# run a bash sending all to /dev/null, then send to /dev/null again, so we don't get "Killed" messages
bash -c "timeout --signal=9 ${timeout_seconds} sync &> /dev/null" &> /dev/null && sync_worked=1 || sync_timeout=1
display_alert "wait_for_disk_sync: $*" "sync done worked: ${sync_worked} sync_timeout: ${sync_timeout}" "debug"
if [[ "${sync_timeout}" -eq 1 ]]; then
total_wait=$((total_wait + timeout_seconds))
sync_timeout_count=$((sync_timeout_count + 1))
display_alert "Waiting sync to storage $*" "fsync taking more than ${total_wait}s - be patient" "warn"
else
if [[ "${sync_timeout_count}" -gt 0 ]]; then
display_alert "Sync to storage OK $*" "fsync took more than ${total_wait}s" "info"
fi
fi
done
}