#!/bin/bash # # functions-common - Common functions used by DevStack components # # The canonical copy of this file is maintained in the DevStack repo. # All modifications should be made there and then sync'ed to other repos # as required. # # This file is sorted alphabetically within the function groups. # # - Config Functions # - Control Functions # - Distro Functions # - Git Functions # - OpenStack Functions # - Package Functions # - Process Functions # - Service Functions # - System Functions # # The following variables are assumed to be defined by certain functions: # # - ``ENABLED_SERVICES`` # - ``ERROR_ON_CLONE`` # - ``FILES`` # - ``OFFLINE`` # - ``RECLONE`` # - ``REQUIREMENTS_DIR`` # - ``STACK_USER`` # - ``http_proxy``, ``https_proxy``, ``no_proxy`` # # Save trace setting _XTRACE_FUNCTIONS_COMMON=$(set +o | grep xtrace) set +o xtrace # ensure we don't re-source this in the same environment [[ -z "$_DEVSTACK_FUNCTIONS_COMMON" ]] || return 0 declare -r -g _DEVSTACK_FUNCTIONS_COMMON=1 # Global Config Variables declare -A -g GITREPO declare -A -g GITBRANCH declare -A -g GITDIR KILL_PATH="$(which kill)" # Save these variables to .stackenv STACK_ENV_VARS="BASE_SQL_CONN DATA_DIR DEST ENABLED_SERVICES HOST_IP \ KEYSTONE_SERVICE_URI \ LOGFILE OS_CACERT SERVICE_HOST STACK_USER TLS_IP \ HOST_IPV6 SERVICE_IP_VERSION TUNNEL_ENDPOINT_IP TUNNEL_IP_VERSION" # Saves significant environment variables to .stackenv for later use # Refers to a lot of globals, only TOP_DIR and STACK_ENV_VARS are required to # function, the rest are simply saved and do not cause problems if they are undefined. # save_stackenv [tag] function save_stackenv { local tag=${1:-""} # Save some values we generated for later use time_stamp=$(date "+$TIMESTAMP_FORMAT") echo "# $time_stamp $tag" >$TOP_DIR/.stackenv for i in $STACK_ENV_VARS; do echo $i=${!i} >>$TOP_DIR/.stackenv done } # Update/create user clouds.yaml file. # clouds.yaml will have # - A `devstack` entry for the `demo` user for the `demo` project. # - A `devstack-admin` entry for the `admin` user for the `admin` project. # write_clouds_yaml function write_clouds_yaml { # The location is a variable to allow for easier refactoring later to make it # overridable. There is currently no usecase where doing so makes sense, so # it's not currently configurable. CLOUDS_YAML=/etc/openstack/clouds.yaml sudo mkdir -p $(dirname $CLOUDS_YAML) sudo chown -R $STACK_USER /etc/openstack CA_CERT_ARG='' if [ -f "$SSL_BUNDLE_FILE" ]; then CA_CERT_ARG="--os-cacert $SSL_BUNDLE_FILE" fi # devstack: user with the member role on demo project $PYTHON $TOP_DIR/tools/update_clouds_yaml.py \ --file $CLOUDS_YAML \ --os-cloud devstack \ --os-region-name $REGION_NAME \ $CA_CERT_ARG \ --os-auth-url $KEYSTONE_SERVICE_URI \ --os-username demo \ --os-password $ADMIN_PASSWORD \ --os-project-name demo # devstack-admin: user with the admin role on the admin project $PYTHON $TOP_DIR/tools/update_clouds_yaml.py \ --file $CLOUDS_YAML \ --os-cloud devstack-admin \ --os-region-name $REGION_NAME \ $CA_CERT_ARG \ --os-auth-url $KEYSTONE_SERVICE_URI \ --os-username admin \ --os-password $ADMIN_PASSWORD \ --os-project-name admin # devstack-admin-demo: user with the admin role on the demo project $PYTHON $TOP_DIR/tools/update_clouds_yaml.py \ --file $CLOUDS_YAML \ --os-cloud devstack-admin-demo \ --os-region-name $REGION_NAME \ $CA_CERT_ARG \ --os-auth-url $KEYSTONE_SERVICE_URI \ --os-username admin \ --os-password $ADMIN_PASSWORD \ --os-project-name demo # devstack-alt: user with the member role on alt_demo project $PYTHON $TOP_DIR/tools/update_clouds_yaml.py \ --file $CLOUDS_YAML \ --os-cloud devstack-alt \ --os-region-name $REGION_NAME \ $CA_CERT_ARG \ --os-auth-url $KEYSTONE_SERVICE_URI \ --os-username alt_demo \ --os-password $ADMIN_PASSWORD \ --os-project-name alt_demo # devstack-alt-member: user with the member role on alt_demo project $PYTHON $TOP_DIR/tools/update_clouds_yaml.py \ --file $CLOUDS_YAML \ --os-cloud devstack-alt-member \ --os-region-name $REGION_NAME \ $CA_CERT_ARG \ --os-auth-url $KEYSTONE_SERVICE_URI \ --os-username alt_demo_member \ --os-password $ADMIN_PASSWORD \ --os-project-name alt_demo # devstack-alt-reader: user with the reader role on alt_demo project $PYTHON $TOP_DIR/tools/update_clouds_yaml.py \ --file $CLOUDS_YAML \ --os-cloud devstack-alt-reader \ --os-region-name $REGION_NAME \ $CA_CERT_ARG \ --os-auth-url $KEYSTONE_SERVICE_URI \ --os-username alt_demo_reader \ --os-password $ADMIN_PASSWORD \ --os-project-name alt_demo # devstack-reader: user with the reader role on demo project $PYTHON $TOP_DIR/tools/update_clouds_yaml.py \ --file $CLOUDS_YAML \ --os-cloud devstack-reader \ --os-region-name $REGION_NAME \ $CA_CERT_ARG \ --os-auth-url $KEYSTONE_SERVICE_URI \ --os-username demo_reader \ --os-password $ADMIN_PASSWORD \ --os-project-name demo # devstack-system-admin: user with the admin role on the system $PYTHON $TOP_DIR/tools/update_clouds_yaml.py \ --file $CLOUDS_YAML \ --os-cloud devstack-system-admin \ --os-region-name $REGION_NAME \ $CA_CERT_ARG \ --os-auth-url $KEYSTONE_SERVICE_URI \ --os-username admin \ --os-password $ADMIN_PASSWORD \ --os-system-scope all # devstack-system-member: user with the member role on the system $PYTHON $TOP_DIR/tools/update_clouds_yaml.py \ --file $CLOUDS_YAML \ --os-cloud devstack-system-member \ --os-region-name $REGION_NAME \ $CA_CERT_ARG \ --os-auth-url $KEYSTONE_SERVICE_URI \ --os-username system_member \ --os-password $ADMIN_PASSWORD \ --os-system-scope all # devstack-system-reader: user with the reader role on the system $PYTHON $TOP_DIR/tools/update_clouds_yaml.py \ --file $CLOUDS_YAML \ --os-cloud devstack-system-reader \ --os-region-name $REGION_NAME \ $CA_CERT_ARG \ --os-auth-url $KEYSTONE_SERVICE_URI \ --os-username system_reader \ --os-password $ADMIN_PASSWORD \ --os-system-scope all cat >> $CLOUDS_YAML < # # Normalize config-value provided in variable VAR to either "True" or # "False". If VAR is unset (i.e. $VAR evaluates as empty), the value # of the second argument will be used as the default value. # # Accepts as False: 0 no No NO false False FALSE # Accepts as True: 1 yes Yes YES true True TRUE # # usage: # VAL=$(trueorfalse False VAL) function trueorfalse { local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace local default=$1 if [ -z $2 ]; then die $LINENO "variable to normalize required" fi local testval=${!2:-} case "$testval" in "1" | [yY]es | "YES" | [tT]rue | "TRUE" ) echo "True" ;; "0" | [nN]o | "NO" | [fF]alse | "FALSE" ) echo "False" ;; * ) echo "$default" ;; esac $xtrace } # bool_to_int # # Convert True|False to int 1 or 0 # This function can be used to convert the output of trueorfalse # to an int follow c conventions where false is 0 and 1 it true. function bool_to_int { local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace if [ -z $1 ]; then die $LINENO "Bool value required" fi if [[ $1 == "True" ]] ; then echo '1' else echo '0' fi $xtrace } function isset { [[ -v "$1" ]] } # Control Functions # ================= # Prints backtrace info # filename:lineno:function # backtrace level function backtrace { local level=$1 local deep deep=$((${#BASH_SOURCE[@]} - 1)) echo "[Call Trace]" while [ $level -le $deep ]; do echo "${BASH_SOURCE[$deep]}:${BASH_LINENO[$deep-1]}:${FUNCNAME[$deep-1]}" deep=$((deep - 1)) done } # Prints line number and "message" then exits # die $LINENO "message" function die { local exitcode=$? set +o xtrace local line=$1; shift if [ $exitcode == 0 ]; then exitcode=1 fi backtrace 2 err $line "$*" # Give buffers a second to flush sleep 1 exit $exitcode } # Checks an environment variable is not set or has length 0 OR if the # exit code is non-zero and prints "message" and exits # NOTE: env-var is the variable name without a '$' # die_if_not_set $LINENO env-var "message" function die_if_not_set { local exitcode=$? local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace local line=$1; shift local evar=$1; shift if ! is_set $evar || [ $exitcode != 0 ]; then die $line "$*" fi $xtrace } function deprecated { local text=$1 DEPRECATED_TEXT+="\n$text" echo "WARNING: $text" >&2 } # Prints line number and "message" in error format # err $LINENO "message" function err { local exitcode=$? local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace local msg="[ERROR] ${BASH_SOURCE[2]}:$1 $2" echo "$msg" 1>&2; if [[ -n ${LOGDIR} ]]; then echo "$msg" >> "${LOGDIR}/error.log" fi $xtrace return $exitcode } # Checks an environment variable is not set or has length 0 OR if the # exit code is non-zero and prints "message" # NOTE: env-var is the variable name without a '$' # err_if_not_set $LINENO env-var "message" function err_if_not_set { local exitcode=$? local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace local line=$1; shift local evar=$1; shift if ! is_set $evar || [ $exitcode != 0 ]; then err $line "$*" fi $xtrace return $exitcode } # Exit after outputting a message about the distribution not being supported. # exit_distro_not_supported [optional-string-telling-what-is-missing] function exit_distro_not_supported { if [[ -z "$DISTRO" ]]; then GetDistro fi if [ $# -gt 0 ]; then die $LINENO "Support for $DISTRO is incomplete: no support for $@" else die $LINENO "Support for $DISTRO is incomplete." fi } # Test if the named environment variable is set and not zero length # is_set env-var function is_set { local var=\$"$1" eval "[ -n \"$var\" ]" # For ex.: sh -c "[ -n \"$var\" ]" would be better, but several exercises depends on this } # Prints line number and "message" in warning format # warn $LINENO "message" function warn { local exitcode=$? local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace local msg="[WARNING] ${BASH_SOURCE[2]}:$1 $2" echo "$msg" $xtrace return $exitcode } # Distro Functions # ================ # Determine OS Vendor, Release and Update # # NOTE : For portability, you almost certainly do not want to use # these variables directly! The "is_*" functions defined below this # bundle up compatible platforms under larger umbrellas that we have # determinted are compatible enough (e.g. is_ubuntu covers Ubuntu & # Debian, is_fedora covers RPM-based distros). Higher-level functions # such as "install_package" further abstract things in better ways. # # ``os_VENDOR`` - vendor name: ``Ubuntu``, ``Fedora``, etc # ``os_RELEASE`` - major release: ``22.04`` (Ubuntu), ``23`` (Fedora) # ``os_PACKAGE`` - package type: ``deb`` or ``rpm`` # ``os_CODENAME`` - vendor's codename for release: ``jammy`` declare -g os_VENDOR os_RELEASE os_PACKAGE os_CODENAME # Make a *best effort* attempt to install lsb_release packages for the # user if not available. Note can't use generic install_package* # because they depend on this! function _ensure_lsb_release { if [[ -x $(command -v lsb_release 2>/dev/null) ]]; then return fi if [[ -x $(command -v apt-get 2>/dev/null) ]]; then sudo apt-get install -y lsb-release elif [[ -x $(command -v zypper 2>/dev/null) ]]; then sudo zypper -n install lsb-release elif [[ -x $(command -v dnf 2>/dev/null) ]]; then sudo dnf install -y redhat-lsb-core || sudo dnf install -y openeuler-lsb else die $LINENO "Unable to find or auto-install lsb_release" fi } # GetOSVersion # Set the following variables: # - os_RELEASE # - os_CODENAME # - os_VENDOR # - os_PACKAGE function GetOSVersion { # CentOS Stream 9 and RHEL 9 do not provide lsb_release source /etc/os-release if [[ "${ID}${VERSION}" == "centos9" ]] || [[ "${ID}${VERSION}" =~ "rhel9" ]]; then os_RELEASE=${VERSION_ID} os_CODENAME="n/a" os_VENDOR=$(echo $NAME | tr -d '[:space:]') elif [[ "${ID}${VERSION}" =~ "rocky9" ]]; then os_VENDOR="Rocky" os_RELEASE=${VERSION_ID} else _ensure_lsb_release os_RELEASE=$(lsb_release -r -s) os_CODENAME=$(lsb_release -c -s) os_VENDOR=$(lsb_release -i -s) fi if [[ $os_VENDOR =~ (Debian|Ubuntu) ]]; then os_PACKAGE="deb" else os_PACKAGE="rpm" fi typeset -xr os_VENDOR typeset -xr os_RELEASE typeset -xr os_PACKAGE typeset -xr os_CODENAME } # Translate the OS version values into common nomenclature # Sets global ``DISTRO`` from the ``os_*`` values declare -g DISTRO function GetDistro { GetOSVersion if [[ "$os_VENDOR" =~ (Ubuntu) || "$os_VENDOR" =~ (Debian) ]]; then # 'Everyone' refers to Ubuntu / Debian releases by # the code name adjective DISTRO=$os_CODENAME elif [[ "$os_VENDOR" =~ (Fedora) ]]; then # For Fedora, just use 'f' and the release DISTRO="f$os_RELEASE" elif [[ "$os_VENDOR" =~ (Red.*Hat) || \ "$os_VENDOR" =~ (CentOS) || \ "$os_VENDOR" =~ (AlmaLinux) || \ "$os_VENDOR" =~ (Scientific) || \ "$os_VENDOR" =~ (OracleServer) || \ "$os_VENDOR" =~ (Rocky) || \ "$os_VENDOR" =~ (Virtuozzo) ]]; then # Drop the . release as we assume it's compatible # XXX re-evaluate when we get RHEL10 DISTRO="rhel${os_RELEASE::1}" elif [[ "$os_VENDOR" =~ (openEuler) ]]; then DISTRO="openEuler-$os_RELEASE" else # We can't make a good choice here. Setting a sensible DISTRO # is part of the problem, but not the major issue -- we really # only use DISTRO in the code as a fine-filter. # # The bigger problem is categorising the system into one of # our two big categories as Ubuntu/Debian-ish or # Fedora/CentOS-ish. # # The setting of os_PACKAGE above is only set to "deb" based # on a hard-coded list of vendor names ... thus we will # default to thinking unknown distros are RPM based # (ie. is_ubuntu does not match). But the platform will then # also not match in is_fedora, because that also has a list of # names. # # So, if you are reading this, getting your distro supported # is really about making sure it matches correctly in these # functions. Then you can choose a sensible way to construct # DISTRO based on your distros release approach. die $LINENO "Unable to determine DISTRO, can not continue." fi typeset -xr DISTRO } # Utility function for checking machine architecture # is_arch arch-type function is_arch { [[ "$(uname -m)" == "$1" ]] } # Determine if current distribution is an Oracle distribution # is_oraclelinux function is_oraclelinux { if [[ -z "$os_VENDOR" ]]; then GetOSVersion fi [ "$os_VENDOR" = "OracleServer" ] } # Determine if current distribution is a Fedora-based distribution # (Fedora, RHEL, CentOS, Rocky, etc). # is_fedora function is_fedora { if [[ -z "$os_VENDOR" ]]; then GetOSVersion fi [ "$os_VENDOR" = "Fedora" ] || [ "$os_VENDOR" = "Red Hat" ] || \ [ "$os_VENDOR" = "openEuler" ] || \ [ "$os_VENDOR" = "RedHatEnterpriseServer" ] || \ [ "$os_VENDOR" = "RedHatEnterprise" ] || \ [ "$os_VENDOR" = "RedHatEnterpriseLinux" ] || \ [ "$os_VENDOR" = "Rocky" ] || \ [ "$os_VENDOR" = "CentOS" ] || [ "$os_VENDOR" = "CentOSStream" ] || \ [ "$os_VENDOR" = "AlmaLinux" ] || \ [ "$os_VENDOR" = "OracleServer" ] || [ "$os_VENDOR" = "Virtuozzo" ] } # Determine if current distribution is an Ubuntu-based distribution # It will also detect non-Ubuntu but Debian-based distros # is_ubuntu function is_ubuntu { if [[ -z "$os_PACKAGE" ]]; then GetOSVersion fi [ "$os_PACKAGE" = "deb" ] } # Determine if current distribution is an openEuler distribution # is_openeuler function is_openeuler { if [[ -z "$os_PACKAGE" ]]; then GetOSVersion fi [ "$os_VENDOR" = "openEuler" ] } # Git Functions # ============= # Returns openstack release name for a given branch name # ``get_release_name_from_branch branch-name`` function get_release_name_from_branch { local branch=$1 if [[ $branch =~ "stable/" || $branch =~ "proposed/" ]]; then echo ${branch#*/} else echo "master" fi } # git clone only if directory doesn't exist already. Since ``DEST`` might not # be owned by the installation user, we create the directory and change the # ownership to the proper user. # Set global ``RECLONE=yes`` to simulate a clone when dest-dir exists # Set global ``ERROR_ON_CLONE=True`` to abort execution with an error if the git repo # does not exist (default is False, meaning the repo will be cloned). # Uses globals ``ERROR_ON_CLONE``, ``OFFLINE``, ``RECLONE`` # git_clone remote dest-dir branch function git_clone { local git_remote=$1 local git_dest=$2 local git_ref=$3 local orig_dir orig_dir=$(pwd) local git_clone_flags="" RECLONE=$(trueorfalse False RECLONE) if [[ "${GIT_DEPTH}" -gt 0 ]]; then git_clone_flags="$git_clone_flags --depth $GIT_DEPTH" fi if [[ "$OFFLINE" = "True" ]]; then echo "Running in offline mode, clones already exist" # print out the results so we know what change was used in the logs cd $git_dest git show --oneline | head -1 cd $orig_dir return fi if echo $git_ref | egrep -q "^refs"; then # If our branch name is a gerrit style refs/changes/... if [[ ! -d $git_dest ]]; then if [[ "$ERROR_ON_CLONE" = "True" ]]; then echo "The $git_dest project was not found; if this is a gate job, add" echo "the project to 'required-projects' in the job definition." die $LINENO "ERROR_ON_CLONE is set to True so cloning not allowed in this configuration" fi git_timed clone $git_clone_flags $git_remote $git_dest fi cd $git_dest git_timed fetch $git_remote $git_ref && git checkout FETCH_HEAD else # do a full clone only if the directory doesn't exist if [[ ! -d $git_dest ]]; then if [[ "$ERROR_ON_CLONE" = "True" ]]; then echo "The $git_dest project was not found; if this is a gate job, add" echo "the project to the \$PROJECTS variable in the job definition." die $LINENO "ERROR_ON_CLONE is set to True so cloning not allowed in this configuration" fi git_timed clone --no-checkout $git_clone_flags $git_remote $git_dest cd $git_dest git_timed fetch $git_clone_flags origin $git_ref git_timed checkout FETCH_HEAD elif [[ "$RECLONE" = "True" ]]; then # if it does exist then simulate what clone does if asked to RECLONE cd $git_dest # set the url to pull from and fetch git remote set-url origin $git_remote git_timed fetch origin # remove the existing ignored files (like pyc) as they cause breakage # (due to the py files having older timestamps than our pyc, so python # thinks the pyc files are correct using them) sudo find $git_dest -name '*.pyc' -delete # handle git_ref accordingly to type (tag, branch) if [[ -n "`git show-ref refs/tags/$git_ref`" ]]; then git_update_tag $git_ref elif [[ -n "`git show-ref refs/heads/$git_ref`" ]]; then git_update_branch $git_ref elif [[ -n "`git show-ref refs/remotes/origin/$git_ref`" ]]; then git_update_remote_branch $git_ref else die $LINENO "$git_ref is neither branch nor tag" fi fi fi # NOTE(ianw) 2022-04-13 : commit [1] has broken many assumptions # about how we clone and work with repos. Mark them safe globally # as a work-around. # # NOTE(danms): On bionic (and likely others) git-config may write # ~stackuser/.gitconfig if not run with sudo -H. Using --system # writes these changes to /etc/gitconfig which is more # discoverable anyway. # # [1] https://github.com/git/git/commit/8959555cee7ec045958f9b6dd62e541affb7e7d9 sudo git config --system --add safe.directory ${git_dest} # print out the results so we know what change was used in the logs cd $git_dest git show --oneline | head -1 cd $orig_dir } # A variation on git clone that lets us specify a project by it's # actual name, like oslo.config. This is exceptionally useful in the # library installation case function git_clone_by_name { local name=$1 local repo=${GITREPO[$name]} local dir=${GITDIR[$name]} local branch=${GITBRANCH[$name]} git_clone $repo $dir $branch } # git can sometimes get itself infinitely stuck with transient network # errors or other issues with the remote end. This wraps git in a # timeout/retry loop and is intended to watch over non-local git # processes that might hang. GIT_TIMEOUT, if set, is passed directly # to timeout(1); otherwise the default value of 0 maintains the status # quo of waiting forever. # usage: git_timed function git_timed { local count=0 local timeout=0 if [[ -n "${GIT_TIMEOUT}" ]]; then timeout=${GIT_TIMEOUT} fi time_start "git_timed" until timeout -s SIGINT ${timeout} git "$@"; do # 124 is timeout(1)'s special return code when it reached the # timeout; otherwise assume fatal failure if [[ $? -ne 124 ]]; then die $LINENO "git call failed: [git $@]" fi count=$(($count + 1)) warn $LINENO "timeout ${count} for git call: [git $@]" if [ $count -eq 3 ]; then die $LINENO "Maximum of 3 git retries reached" fi sleep 5 done time_stop "git_timed" } # git update using reference as a branch. # git_update_branch ref function git_update_branch { local git_branch=$1 git checkout -f origin/$git_branch # a local branch might not exist git branch -D $git_branch || true git checkout -b $git_branch } # git update using reference as a branch. # git_update_remote_branch ref function git_update_remote_branch { local git_branch=$1 git checkout -b $git_branch -t origin/$git_branch } # git update using reference as a tag. Be careful editing source at that repo # as working copy will be in a detached mode # git_update_tag ref function git_update_tag { local git_tag=$1 git tag -d $git_tag # fetching given tag only git_timed fetch origin tag $git_tag git checkout -f $git_tag } # OpenStack Functions # =================== # Get the default value for HOST_IP # get_default_host_ip fixed_range floating_range host_ip_iface host_ip function get_default_host_ip { local fixed_range=$1 local floating_range=$2 local host_ip_iface=$3 local host_ip=$4 local af=$5 # Search for an IP unless an explicit is set by ``HOST_IP`` environment variable if [ -z "$host_ip" -o "$host_ip" == "dhcp" ]; then host_ip="" # Find the interface used for the default route host_ip_iface=${host_ip_iface:-$(ip -f $af route | awk '/default/ {print $5}' | head -1)} local host_ips host_ips=$(LC_ALL=C ip -f $af addr show ${host_ip_iface} | sed /temporary/d |awk /$af'/ {split($2,parts,"/"); print parts[1]}') local ip for ip in $host_ips; do # Attempt to filter out IP addresses that are part of the fixed and # floating range. Note that this method only works if the ``netaddr`` # python library is installed. If it is not installed, an error # will be printed and the first IP from the interface will be used. # If that is not correct set ``HOST_IP`` in ``localrc`` to the correct # address. if [[ "$af" == "inet6" ]]; then host_ip=$ip break; fi if ! (address_in_net $ip $fixed_range || address_in_net $ip $floating_range); then host_ip=$ip break; fi done fi echo $host_ip } # Generates hex string from ``size`` byte of pseudo random data # generate_hex_string size function generate_hex_string { local size=$1 hexdump -n "$size" -v -e '/1 "%02x"' /dev/urandom } # Grab a numbered field from python prettytable output # Fields are numbered starting with 1 # Reverse syntax is supported: -1 is the last field, -2 is second to last, etc. # get_field field-number function get_field { local data field while read data; do if [ "$1" -lt 0 ]; then field="(\$(NF$1))" else field="\$$(($1 + 1))" fi echo "$data" | awk -F'[ \t]*\\|[ \t]*' "{print $field}" done } # install default policy # copy over a default policy.json and policy.d for projects function install_default_policy { local project=$1 local project_uc project_uc=$(echo $1|tr a-z A-Z) local conf_dir="${project_uc}_CONF_DIR" # eval conf dir to get the variable conf_dir="${!conf_dir}" local project_dir="${project_uc}_DIR" # eval project dir to get the variable project_dir="${!project_dir}" local sample_conf_dir="${project_dir}/etc/${project}" local sample_policy_dir="${project_dir}/etc/${project}/policy.d" # first copy any policy.json cp -p $sample_conf_dir/policy.json $conf_dir # then optionally copy over policy.d if [[ -d $sample_policy_dir ]]; then cp -r $sample_policy_dir $conf_dir/policy.d fi } # Add a policy to a policy.json file # Do nothing if the policy already exists # ``policy_add policy_file policy_name policy_permissions`` function policy_add { local policy_file=$1 local policy_name=$2 local policy_perm=$3 if grep -q ${policy_name} ${policy_file}; then echo "Policy ${policy_name} already exists in ${policy_file}" return fi # Add a terminating comma to policy lines without one # Remove the closing '}' and all lines following to the end-of-file local tmpfile tmpfile=$(mktemp) uniq ${policy_file} | sed -e ' s/]$/],/ /^[}]/,$d ' > ${tmpfile} # Append policy and closing brace echo " \"${policy_name}\": ${policy_perm}" >>${tmpfile} echo "}" >>${tmpfile} mv ${tmpfile} ${policy_file} } # Gets or creates a domain # Usage: get_or_create_domain function get_or_create_domain { local domain_id domain_id=$( openstack --os-cloud devstack-system-admin domain create $1 \ --description "$2" --or-show \ -f value -c id ) echo $domain_id } # Gets or creates group # Usage: get_or_create_group [] function get_or_create_group { local desc="${3:-}" local group_id # Gets group id group_id=$( # Creates new group with --or-show openstack --os-cloud devstack-system-admin group create $1 \ --domain $2 --description "$desc" --or-show \ -f value -c id ) echo $group_id } # Gets or creates user # Usage: get_or_create_user [] function get_or_create_user { local user_id if [[ ! -z "$4" ]]; then local email="--email=$4" else local email="" fi # Gets user id user_id=$( # Creates new user with --or-show openstack --os-cloud devstack-system-admin user create \ $1 \ --password "$2" \ --domain=$3 \ $email \ --or-show \ -f value -c id ) echo $user_id } # Gets or creates project # Usage: get_or_create_project function get_or_create_project { local project_id project_id=$( # Creates new project with --or-show openstack --os-cloud devstack-system-admin project create $1 \ --domain=$2 \ --or-show -f value -c id ) echo $project_id } # Gets or creates role # Usage: get_or_create_role function get_or_create_role { local role_id role_id=$( # Creates role with --or-show openstack --os-cloud devstack-system-admin role create $1 \ --or-show -f value -c id ) echo $role_id } # Returns the domain parts of a function call if present # Usage: _get_domain_args [ ] function _get_domain_args { local domain domain="" if [[ -n "$1" ]]; then domain="$domain --user-domain $1" fi if [[ -n "$2" ]]; then domain="$domain --project-domain $2" fi echo $domain } # Gets or adds user role to project # Usage: get_or_add_user_project_role [ ] function get_or_add_user_project_role { local user_role_id local domain_args domain_args=$(_get_domain_args $4 $5) # Note this is idempotent so we are safe across multiple # duplicate calls. openstack --os-cloud devstack-system-admin role add $1 \ --user $2 \ --project $3 \ $domain_args user_role_id=$(openstack --os-cloud devstack-system-admin role assignment list \ --role $1 \ --user $2 \ --project $3 \ $domain_args \ -c Role -f value) echo $user_role_id } # Gets or adds user role to domain # Usage: get_or_add_user_domain_role function get_or_add_user_domain_role { local user_role_id # Note this is idempotent so we are safe across multiple # duplicate calls. openstack --os-cloud devstack-system-admin role add $1 \ --user $2 \ --domain $3 user_role_id=$(openstack --os-cloud devstack-system-admin role assignment list \ --role $1 \ --user $2 \ --domain $3 \ -c Role -f value) echo $user_role_id } # Gets or adds user role to system # Usage: get_or_add_user_system_role [] function get_or_add_user_system_role { local user_role_id local domain_args domain_args=$(_get_domain_args $4) # Gets user role id user_role_id=$(openstack --os-cloud devstack-system-admin role assignment list \ --role $1 \ --user $2 \ --system $3 \ $domain_args \ -f value -c Role) if [[ -z "$user_role_id" ]]; then # Adds role to user and get it openstack --os-cloud devstack-system-admin role add $1 \ --user $2 \ --system $3 \ $domain_args user_role_id=$(openstack --os-cloud devstack-system-admin role assignment list \ --role $1 \ --user $2 \ --system $3 \ $domain_args \ -f value -c Role) fi echo $user_role_id } # Gets or adds group role to project # Usage: get_or_add_group_project_role function get_or_add_group_project_role { local group_role_id # Note this is idempotent so we are safe across multiple # duplicate calls. openstack role add $1 \ --group $2 \ --project $3 group_role_id=$(openstack --os-cloud devstack-system-admin role assignment list \ --role $1 \ --group $2 \ --project $3 \ -f value -c Role) echo $group_role_id } # Gets or creates service # Usage: get_or_create_service function get_or_create_service { local service_id # Gets service id service_id=$( # Gets service id openstack --os-cloud devstack-system-admin service show $2 -f value -c id 2>/dev/null || # Creates new service if not exists openstack --os-cloud devstack-system-admin service create \ $2 \ --name $1 \ --description="$3" \ -f value -c id ) echo $service_id } # Create an endpoint with a specific interface # Usage: _get_or_create_endpoint_with_interface function _get_or_create_endpoint_with_interface { local endpoint_id endpoint_id=$(openstack --os-cloud devstack-system-admin endpoint list \ --service $1 \ --interface $2 \ --region $4 \ -c ID -f value) if [[ -z "$endpoint_id" ]]; then # Creates new endpoint endpoint_id=$(openstack --os-cloud devstack-system-admin endpoint create \ $1 $2 $3 --region $4 -f value -c id) fi echo $endpoint_id } # Gets or creates endpoint # Usage: get_or_create_endpoint [adminurl] [internalurl] function get_or_create_endpoint { # NOTE(jamielennnox): when converting to v3 endpoint creation we go from # creating one endpoint with multiple urls to multiple endpoints each with # a different interface. To maintain the existing function interface we # create 3 endpoints and return the id of the public one. In reality # returning the public id will not make a lot of difference as there are no # scenarios currently that use the returned id. Ideally this behaviour # should be pushed out to the service setups and let them create the # endpoints they need. local public_id public_id=$(_get_or_create_endpoint_with_interface $1 public $3 $2) # only create admin/internal urls if provided content for them if [[ -n "$4" ]]; then _get_or_create_endpoint_with_interface $1 admin $4 $2 fi if [[ -n "$5" ]]; then _get_or_create_endpoint_with_interface $1 internal $5 $2 fi # return the public id to indicate success, and this is the endpoint most likely wanted echo $public_id } # Get a URL from the identity service # Usage: get_endpoint_url function get_endpoint_url { echo $(openstack --os-cloud devstack-system-admin endpoint list \ --service $1 --interface $2 \ -c URL -f value) } # check if we are using ironic with hardware # TODO(jroll) this is a kludge left behind when ripping ironic code # out of tree, as it is used by nova and neutron. # figure out a way to refactor nova/neutron code to eliminate this function is_ironic_hardware { is_service_enabled ironic && [[ "$IRONIC_IS_HARDWARE" == "True" ]] && return 0 return 1 } function is_ironic_enforce_scope { is_service_enabled ironic && [[ "$IRONIC_ENFORCE_SCOPE" == "True" || "$ENFORCE_SCOPE" == "True" ]] && return 0 return 1 } function is_ironic_sharded { # todo(JayF): Support >1 shard with multiple n-cpu instances for each is_service_enabled ironic && [[ "$IRONIC_SHARDS" == "1" ]] && return 0 return 1 } # Package Functions # ================= # _get_package_dir function _get_package_dir { local base_dir=$1 local pkg_dir if [[ -z "$base_dir" ]]; then base_dir=$FILES fi if is_ubuntu; then pkg_dir=$base_dir/debs elif is_fedora; then pkg_dir=$base_dir/rpms else exit_distro_not_supported "list of packages" fi echo "$pkg_dir" } # Wrapper for ``apt-get update`` to try multiple times on the update # to address bad package mirrors (which happen all the time). function apt_get_update { # only do this once per run if [[ "$REPOS_UPDATED" == "True" && "$RETRY_UPDATE" != "True" ]]; then return fi # bail if we are offline [[ "$OFFLINE" = "True" ]] && return local sudo="sudo" [[ "$(id -u)" = "0" ]] && sudo="env" # time all the apt operations time_start "apt-get-update" local proxies="http_proxy=${http_proxy:-} https_proxy=${https_proxy:-} no_proxy=${no_proxy:-} " local update_cmd="$sudo $proxies apt-get update" if ! timeout 300 sh -c "while ! $update_cmd; do sleep 30; done"; then die $LINENO "Failed to update apt repos, we're dead now" fi REPOS_UPDATED=True # stop the clock time_stop "apt-get-update" } # Wrapper for ``apt-get`` to set cache and proxy environment variables # Uses globals ``OFFLINE``, ``*_proxy`` # apt_get operation package [package ...] function apt_get { local xtrace result xtrace=$(set +o | grep xtrace) set +o xtrace [[ "$OFFLINE" = "True" || -z "$@" ]] && return local sudo="sudo" [[ "$(id -u)" = "0" ]] && sudo="env" # time all the apt operations time_start "apt-get" $xtrace $sudo DEBIAN_FRONTEND=noninteractive \ http_proxy=${http_proxy:-} https_proxy=${https_proxy:-} \ no_proxy=${no_proxy:-} \ apt-get --option "Dpkg::Options::=--force-confold" --assume-yes "$@" < /dev/null result=$? # stop the clock time_stop "apt-get" return $result } function _parse_package_files { local files_to_parse=$@ if [[ -z "$DISTRO" ]]; then GetDistro fi for fname in ${files_to_parse}; do local OIFS line package distros distro [[ -e $fname ]] || continue OIFS=$IFS IFS=$'\n' for line in $(<${fname}); do if [[ $line =~ "NOPRIME" ]]; then continue fi # Assume we want this package; free-form # comments allowed after a # package=${line%%#*} inst_pkg=1 # Look for # dist:xxx in comment if [[ $line =~ (.*)#.*dist:([^ ]*) ]]; then # We are using BASH regexp matching feature. package=${BASH_REMATCH[1]} distros=${BASH_REMATCH[2]} # In bash ${VAR,,} will lowercase VAR # Look for a match in the distro list if [[ ! ${distros,,} =~ ${DISTRO,,} ]]; then # If no match then skip this package inst_pkg=0 fi fi # Look for # not:xxx in comment if [[ $line =~ (.*)#.*not:([^ ]*) ]]; then # We are using BASH regexp matching feature. package=${BASH_REMATCH[1]} distros=${BASH_REMATCH[2]} # In bash ${VAR,,} will lowercase VAR # Look for a match in the distro list if [[ ${distros,,} =~ ${DISTRO,,} ]]; then # If match then skip this package inst_pkg=0 fi fi if [[ $inst_pkg = 1 ]]; then echo $package fi done IFS=$OIFS done } # get_packages() collects a list of package names of any type from the # prerequisite files in ``files/{debs|rpms}``. The list is intended # to be passed to a package installer such as apt or yum. # # Only packages required for the services in 1st argument will be # included. Two bits of metadata are recognized in the prerequisite files: # # - ``# NOPRIME`` defers installation to be performed later in `stack.sh` # - ``# dist:DISTRO`` or ``dist:DISTRO1,DISTRO2`` limits the selection # of the package to the distros listed. The distro names are case insensitive. # - ``# not:DISTRO`` or ``not:DISTRO1,DISTRO2`` limits the selection # of the package to the distros not listed. The distro names are case insensitive. function get_packages { local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace local services=$@ local package_dir package_dir=$(_get_package_dir) local file_to_parse="" local service="" if [ $# -ne 1 ]; then die $LINENO "get_packages takes a single, comma-separated argument" fi if [[ -z "$package_dir" ]]; then echo "No package directory supplied" return 1 fi for service in ${services//,/ }; do # Allow individual services to specify dependencies if [[ -e ${package_dir}/${service} ]]; then file_to_parse="${file_to_parse} ${package_dir}/${service}" fi # NOTE(sdague) n-api needs glance for now because that's where # glance client is if [[ $service == n-api ]]; then if [[ ! $file_to_parse =~ $package_dir/nova ]]; then file_to_parse="${file_to_parse} ${package_dir}/nova" fi if [[ ! $file_to_parse =~ $package_dir/glance ]]; then file_to_parse="${file_to_parse} ${package_dir}/glance" fi if [[ ! $file_to_parse =~ $package_dir/os-brick ]]; then file_to_parse="${file_to_parse} ${package_dir}/os-brick" fi elif [[ $service == c-* ]]; then if [[ ! $file_to_parse =~ $package_dir/cinder ]]; then file_to_parse="${file_to_parse} ${package_dir}/cinder" fi if [[ ! $file_to_parse =~ $package_dir/os-brick ]]; then file_to_parse="${file_to_parse} ${package_dir}/os-brick" fi elif [[ $service == s-* ]]; then if [[ ! $file_to_parse =~ $package_dir/swift ]]; then file_to_parse="${file_to_parse} ${package_dir}/swift" fi elif [[ $service == n-* ]]; then if [[ ! $file_to_parse =~ $package_dir/nova ]]; then file_to_parse="${file_to_parse} ${package_dir}/nova" fi if [[ ! $file_to_parse =~ $package_dir/os-brick ]]; then file_to_parse="${file_to_parse} ${package_dir}/os-brick" fi elif [[ $service == g-* ]]; then if [[ ! $file_to_parse =~ $package_dir/glance ]]; then file_to_parse="${file_to_parse} ${package_dir}/glance" fi elif [[ $service == key* ]]; then if [[ ! $file_to_parse =~ $package_dir/keystone ]]; then file_to_parse="${file_to_parse} ${package_dir}/keystone" fi elif [[ $service == q-* || $service == neutron-* ]]; then if [[ ! $file_to_parse =~ $package_dir/neutron-common ]]; then file_to_parse="${file_to_parse} ${package_dir}/neutron-common" fi elif [[ $service == ir-* ]]; then if [[ ! $file_to_parse =~ $package_dir/ironic ]]; then file_to_parse="${file_to_parse} ${package_dir}/ironic" fi fi done echo "$(_parse_package_files $file_to_parse)" $xtrace } # get_plugin_packages() collects a list of package names of any type from a # plugin's prerequisite files in ``$PLUGIN/devstack/files/{debs|rpms}``. The # list is intended to be passed to a package installer such as apt or yum. # # Only packages required for enabled and collected plugins will included. # # The same metadata used in the main DevStack prerequisite files may be used # in these prerequisite files, see get_packages() for more info. function get_plugin_packages { local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace local files_to_parse="" local package_dir="" for plugin in ${DEVSTACK_PLUGINS//,/ }; do package_dir="$(_get_package_dir ${GITDIR[$plugin]}/devstack/files)" files_to_parse+=" $package_dir/$plugin" done echo "$(_parse_package_files $files_to_parse)" $xtrace } # Search plugins for a bindep.txt file # # Uses globals ``BINDEP_CMD``, ``GITDIR``, ``DEVSTACK_PLUGINS`` # # Note this is only valid after BINDEP_CMD is setup in stack.sh, and # is thus not really intended to be called externally. function _get_plugin_bindep_packages { local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace local bindep_file local packages for plugin in ${DEVSTACK_PLUGINS//,/ }; do bindep_file=${GITDIR[$plugin]}/devstack/files/bindep.txt if [[ -f ${bindep_file} ]]; then packages+=$($BINDEP_CMD -b --file ${bindep_file} || true) fi done echo "${packages}" $xtrace } # Distro-agnostic package installer # Uses globals ``NO_UPDATE_REPOS``, ``REPOS_UPDATED``, ``RETRY_UPDATE`` # install_package package [package ...] function update_package_repo { NO_UPDATE_REPOS=${NO_UPDATE_REPOS:-False} REPOS_UPDATED=${REPOS_UPDATED:-False} RETRY_UPDATE=${RETRY_UPDATE:-False} if [[ "$NO_UPDATE_REPOS" = "True" ]]; then return 0 fi if is_ubuntu; then apt_get_update fi } function real_install_package { if is_ubuntu; then apt_get install "$@" elif is_fedora; then yum_install "$@" else exit_distro_not_supported "installing packages" fi } # Distro-agnostic package installer # install_package package [package ...] function install_package { update_package_repo if ! real_install_package "$@"; then RETRY_UPDATE=True update_package_repo && real_install_package "$@" fi } # Distro-agnostic function to tell if a package is installed # is_package_installed package [package ...] function is_package_installed { if [[ -z "$@" ]]; then return 1 fi if [[ -z "$os_PACKAGE" ]]; then GetOSVersion fi if [[ "$os_PACKAGE" = "deb" ]]; then dpkg -s "$@" > /dev/null 2> /dev/null elif [[ "$os_PACKAGE" = "rpm" ]]; then rpm --quiet -q "$@" else exit_distro_not_supported "finding if a package is installed" fi } # Distro-agnostic package uninstaller # uninstall_package package [package ...] function uninstall_package { if is_ubuntu; then apt_get purge "$@" elif is_fedora; then sudo dnf remove -y "$@" ||: else exit_distro_not_supported "uninstalling packages" fi } # Wrapper for ``dnf`` to set proxy environment variables # Uses globals ``OFFLINE``, ``*_proxy`` # The name is kept for backwards compatability with external # callers, despite none of our supported platforms using yum # any more. # yum_install package [package ...] function yum_install { local result parse_yum_result [[ "$OFFLINE" = "True" ]] && return time_start "yum_install" sudo_with_proxies dnf install -y "$@" time_stop "yum_install" } # zypper wrapper to set arguments correctly # Uses globals ``OFFLINE``, ``*_proxy`` # zypper_install package [package ...] function zypper_install { [[ "$OFFLINE" = "True" ]] && return local sudo="sudo" [[ "$(id -u)" = "0" ]] && sudo="env" $sudo http_proxy="${http_proxy:-}" https_proxy="${https_proxy:-}" \ no_proxy="${no_proxy:-}" \ zypper --non-interactive install --auto-agree-with-licenses --no-recommends "$@" } # Run bindep and install packages it outputs # # Usage: # install_bindep [profile,profile] # # Note unlike the bindep command itself, profile(s) specified should # be a single, comma-separated string, no spaces. function install_bindep { local file=$1 local profiles=${2:-""} local pkgs if [[ ! -f $file ]]; then warn $LINENO "Can not find bindep file: $file" return fi # converting here makes it much easier to work with passing # arguments profiles=${profiles/,/ /} # Note bindep returns 1 when packages need to be installed, so we # have to ignore it's return for "-e" pkgs=$($DEST/bindep-venv/bin/bindep -b --file $file $profiles || true) if [[ -n "${pkgs}" ]]; then install_package ${pkgs} fi } function write_user_unit_file { local service=$1 local command="$2" local group=$3 local user=$4 local env_vars="$5" local extra="" if [[ -n "$group" ]]; then extra="Group=$group" fi local unitfile="$SYSTEMD_DIR/$service" mkdir -p $SYSTEMD_DIR iniset -sudo $unitfile "Unit" "Description" "Devstack $service" iniset -sudo $unitfile "Service" "Environment" "\"PATH=$PATH\"" iniset -sudo $unitfile "Service" "User" "$user" iniset -sudo $unitfile "Service" "ExecStart" "$command" iniset -sudo $unitfile "Service" "KillMode" "process" iniset -sudo $unitfile "Service" "TimeoutStopSec" "300" iniset -sudo $unitfile "Service" "ExecReload" "$KILL_PATH -HUP \$MAINPID" if [[ -n "$env_vars" ]] ; then iniset -sudo $unitfile "Service" "Environment" "$env_vars" fi if [[ -n "$group" ]]; then iniset -sudo $unitfile "Service" "Group" "$group" fi iniset -sudo $unitfile "Install" "WantedBy" "multi-user.target" # changes to existing units sometimes need a refresh $SYSTEMCTL daemon-reload } function write_uwsgi_user_unit_file { local service=$1 local command="$2" local group=$3 local user=$4 local env_vars="$5" local unitfile="$SYSTEMD_DIR/$service" mkdir -p $SYSTEMD_DIR iniset -sudo $unitfile "Unit" "Description" "Devstack $service" iniset -sudo $unitfile "Service" "Environment" "\"PATH=$PATH\"" iniset -sudo $unitfile "Service" "SyslogIdentifier" "$service" iniset -sudo $unitfile "Service" "User" "$user" iniset -sudo $unitfile "Service" "ExecStart" "$command" iniset -sudo $unitfile "Service" "ExecReload" "$KILL_PATH -HUP \$MAINPID" iniset -sudo $unitfile "Service" "Type" "notify" iniset -sudo $unitfile "Service" "KillMode" "process" iniset -sudo $unitfile "Service" "Restart" "always" iniset -sudo $unitfile "Service" "NotifyAccess" "all" iniset -sudo $unitfile "Service" "RestartForceExitStatus" "100" if [[ -n "$env_vars" ]] ; then iniset -sudo $unitfile "Service" "Environment" "$env_vars" fi if [[ -n "$group" ]]; then iniset -sudo $unitfile "Service" "Group" "$group" fi iniset -sudo $unitfile "Install" "WantedBy" "multi-user.target" # changes to existing units sometimes need a refresh $SYSTEMCTL daemon-reload } function _common_systemd_pitfalls { local cmd=$1 # do some sanity checks on $cmd to see things we don't expect to work if [[ "$cmd" =~ "sudo" ]]; then read -r -d '' msg << EOF || true # read returns 1 for EOF, but it is ok here You are trying to use run_process with sudo, this is not going to work under systemd. If you need to run a service as a user other than \$STACK_USER call it with: run_process \$name \$cmd \$group \$user EOF die $LINENO "$msg" fi if [[ ! "$cmd" =~ ^/ ]]; then read -r -d '' msg << EOF || true # read returns 1 for EOF, but it is ok here The cmd="$cmd" does not start with an absolute path. It will fail to start under systemd. Please update your run_process stanza to have an absolute path. EOF die $LINENO "$msg" fi } # Helper function to build a basic unit file and run it under systemd. function _run_under_systemd { local service=$1 local command="$2" local cmd=$command # sanity check the command _common_systemd_pitfalls "$cmd" local systemd_service="devstack@$service.service" local group=$3 local user=${4:-$STACK_USER} if [[ -z "$user" ]]; then user=$STACK_USER fi local env_vars="$5" if [[ "$command" =~ "uwsgi" ]] ; then if [[ "$GLOBAL_VENV" == "True" ]] ; then cmd="$cmd --venv $DEVSTACK_VENV" fi write_uwsgi_user_unit_file $systemd_service "$cmd" "$group" "$user" "$env_vars" else write_user_unit_file $systemd_service "$cmd" "$group" "$user" "$env_vars" fi $SYSTEMCTL enable $systemd_service $SYSTEMCTL start $systemd_service } # Find out if a process exists by partial name. # is_running name function is_running { local name=$1 ps auxw | grep -v grep | grep ${name} > /dev/null local exitcode=$? # some times I really hate bash reverse binary logic return $exitcode } # Run a single service under screen or directly # If the command includes shell metachatacters (;<>*) it must be run using a shell # If an optional group is provided sg will be used to run the # command as that group. # run_process service "command-line" [group] [user] [env_vars] # env_vars must be a space separated list of variable assigments, ie: "A=1 B=2" function run_process { local service=$1 local command="$2" local group=$3 local user=$4 local env_vars="$5" local name=$service time_start "run_process" if is_service_enabled $service; then _run_under_systemd "$name" "$command" "$group" "$user" "$env_vars" fi time_stop "run_process" } # Stop a service process # If a PID is available use it, kill the whole process group via TERM # If screen is being used kill the screen window; this will catch processes # that did not leave a PID behind # Uses globals ``SERVICE_DIR`` # stop_process service function stop_process { local service=$1 SERVICE_DIR=${SERVICE_DIR:-${DEST}/status} if is_service_enabled $service; then # Only do this for units which appear enabled, this also # catches units that don't really exist for cases like # keystone without a failure. if $SYSTEMCTL is-enabled devstack@$service.service; then $SYSTEMCTL stop devstack@$service.service $SYSTEMCTL disable devstack@$service.service fi fi } # use systemctl to check service status function service_check { local service for service in ${ENABLED_SERVICES//,/ }; do # because some things got renamed like key => keystone if $SYSTEMCTL is-enabled devstack@$service.service; then # no-pager is needed because otherwise status dumps to a # pager when in interactive mode, which will stop a manual # devstack run. $SYSTEMCTL status devstack@$service.service --no-pager fi done } # Plugin Functions # ================= DEVSTACK_PLUGINS=${DEVSTACK_PLUGINS:-""} # enable_plugin [branch] # # ``name`` is an arbitrary name - (aka: glusterfs, nova-docker, zaqar) # ``url`` is a git url # ``branch`` is a gitref. If it's not set, defaults to master function enable_plugin { local name=$1 local url=$2 local branch=${3:-master} if is_plugin_enabled $name; then die $LINENO "Plugin attempted to be enabled twice: ${name} ${url} ${branch}" fi DEVSTACK_PLUGINS+=",$name" GITREPO[$name]=$url GITDIR[$name]=$DEST/$name GITBRANCH[$name]=$branch } # is_plugin_enabled # # Check if the plugin was enabled, e.g. using enable_plugin # # ``name`` The name with which the plugin was enabled function is_plugin_enabled { local name=$1 if [[ ",${DEVSTACK_PLUGINS}," =~ ",${name}," ]]; then return 0 fi return 1 } # fetch_plugins # # clones all plugins function fetch_plugins { local plugins="${DEVSTACK_PLUGINS}" local plugin # short circuit if nothing to do if [[ -z $plugins ]]; then return fi echo "Fetching DevStack plugins" for plugin in ${plugins//,/ }; do git_clone_by_name $plugin done } # load_plugin_settings # # Load settings from plugins in the order that they were registered function load_plugin_settings { local plugins="${DEVSTACK_PLUGINS}" local plugin # short circuit if nothing to do if [[ -z $plugins ]]; then return fi echo "Loading plugin settings" for plugin in ${plugins//,/ }; do local dir=${GITDIR[$plugin]} # source any known settings if [[ -f $dir/devstack/settings ]]; then source $dir/devstack/settings fi done } # plugin_override_defaults # # Run an extremely early setting phase for plugins that allows default # overriding of services. function plugin_override_defaults { local plugins="${DEVSTACK_PLUGINS}" local plugin # short circuit if nothing to do if [[ -z $plugins ]]; then return fi echo "Overriding Configuration Defaults" for plugin in ${plugins//,/ }; do local dir=${GITDIR[$plugin]} # source any overrides if [[ -f $dir/devstack/override-defaults ]]; then # be really verbose that an override is happening, as it # may not be obvious if things fail later. echo "$plugin has overridden the following defaults" cat $dir/devstack/override-defaults source $dir/devstack/override-defaults fi done } # run_plugins # # Run the devstack/plugin.sh in all the plugin directories. These are # run in registration order. function run_plugins { local mode=$1 local phase=$2 local plugins="${DEVSTACK_PLUGINS}" local plugin for plugin in ${plugins//,/ }; do local dir=${GITDIR[$plugin]} if [[ -f $dir/devstack/plugin.sh ]]; then source $dir/devstack/plugin.sh $mode $phase fi done } function run_phase { local mode=$1 local phase=$2 if [[ -d $TOP_DIR/extras.d ]]; then local extra_plugin_file_name for extra_plugin_file_name in $TOP_DIR/extras.d/*.sh; do # NOTE(sdague): only process extras.d for the 3 explicitly # white listed elements in tree. We want these to move out # over time as well, but they are in tree, so we need to # manage that. local exceptions="80-tempest.sh" local extra extra=$(basename $extra_plugin_file_name) if [[ ! ( $exceptions =~ "$extra" ) ]]; then warn "use of extras.d is no longer supported" warn "processing of project $extra is skipped" else [[ -r $extra_plugin_file_name ]] && source $extra_plugin_file_name $mode $phase fi done fi # the source phase corresponds to settings loading in plugins if [[ "$mode" == "source" ]]; then load_plugin_settings verify_disabled_services elif [[ "$mode" == "override_defaults" ]]; then plugin_override_defaults else run_plugins $mode $phase fi } # define_plugin # # This function is a no-op. It allows a plugin to define its name So # that other plugins may reference it by name. It should generally be # the last component of the canonical git repo name. E.g., # openstack/devstack-foo should use "devstack-foo" as the name here. # # This function is currently a noop, but the value may still be used # by external tools (as in plugin_requires) and may be used by # devstack in the future. # # ``name`` is an arbitrary name - (aka: glusterfs, nova-docker, zaqar) function define_plugin { : } # plugin_requires # # This function is a no-op. It is currently used by external tools # (such as the devstack module for Ansible) to automatically generate # local.conf files. It is not currently used by devstack itself to # resolve dependencies. # # ``name`` is an arbitrary name - (aka: glusterfs, nova-docker, zaqar) # ``other`` is the name of another plugin function plugin_requires { : } # Service Functions # ================= # remove extra commas from the input string (i.e. ``ENABLED_SERVICES``) # _cleanup_service_list service-list function _cleanup_service_list { local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace echo "$1" | sed -e ' s/,,/,/g; s/^,//; s/,$// ' $xtrace } # disable_all_services() removes all current services # from ``ENABLED_SERVICES`` to reset the configuration # before a minimal installation # Uses global ``ENABLED_SERVICES`` # disable_all_services function disable_all_services { ENABLED_SERVICES="" } # Remove all services starting with '-'. For example, to install all default # services except rabbit (rabbit) set in ``localrc``: # ENABLED_SERVICES+=",-rabbit" # Uses global ``ENABLED_SERVICES`` # disable_negated_services function disable_negated_services { local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace local to_remove="" local remaining="" local service # build up list of services that should be removed; i.e. they # begin with "-" for service in ${ENABLED_SERVICES//,/ }; do if [[ ${service} == -* ]]; then to_remove+=",${service#-}" else remaining+=",${service}" fi done # go through the service list. if this service appears in the "to # be removed" list, drop it ENABLED_SERVICES=$(remove_disabled_services "$remaining" "$to_remove") $xtrace } # disable_service() prepares the services passed as argument to be # removed from the ``ENABLED_SERVICES`` list, if they are present. # # For example: # disable_service rabbit # # Uses global ``DISABLED_SERVICES`` # disable_service service [service ...] function disable_service { local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace local disabled_svcs="${DISABLED_SERVICES}" local enabled_svcs=",${ENABLED_SERVICES}," local service for service in $@; do disabled_svcs+=",$service" if is_service_enabled $service; then enabled_svcs=${enabled_svcs//,$service,/,} fi done DISABLED_SERVICES=$(_cleanup_service_list "$disabled_svcs") ENABLED_SERVICES=$(_cleanup_service_list "$enabled_svcs") $xtrace } # enable_service() adds the services passed as argument to the # ``ENABLED_SERVICES`` list, if they are not already present. # # For example: # enable_service q-svc # # This function does not know about the special cases # for nova, glance, and neutron built into is_service_enabled(). # Uses global ``ENABLED_SERVICES`` # enable_service service [service ...] function enable_service { local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace local tmpsvcs="${ENABLED_SERVICES}" local service for service in $@; do if [[ ,${DISABLED_SERVICES}, =~ ,${service}, ]]; then warn $LINENO "Attempt to enable_service ${service} when it has been disabled" continue fi if ! is_service_enabled $service; then tmpsvcs+=",$service" fi done ENABLED_SERVICES=$(_cleanup_service_list "$tmpsvcs") disable_negated_services $xtrace } # is_service_enabled() checks if the service(s) specified as arguments are # enabled by the user in ``ENABLED_SERVICES``. # # Multiple services specified as arguments are ``OR``'ed together; the test # is a short-circuit boolean, i.e it returns on the first match. # # There are special cases for some 'catch-all' services:: # **nova** returns true if any service enabled start with **n-** # **cinder** returns true if any service enabled start with **c-** # **glance** returns true if any service enabled start with **g-** # **neutron** returns true if any service enabled start with **q-** # **swift** returns true if any service enabled start with **s-** # **trove** returns true if any service enabled start with **tr-** # For backward compatibility if we have **swift** in ENABLED_SERVICES all the # **s-** services will be enabled. This will be deprecated in the future. # # Uses global ``ENABLED_SERVICES`` # is_service_enabled service [service ...] function is_service_enabled { local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace local enabled=1 local services=$@ local service for service in ${services}; do [[ ,${ENABLED_SERVICES}, =~ ,${service}, ]] && enabled=0 # Look for top-level 'enabled' function for this service if type is_${service}_enabled >/dev/null 2>&1; then # A function exists for this service, use it is_${service}_enabled && enabled=0 fi # TODO(dtroyer): Remove these legacy special-cases after the is_XXX_enabled() # are implemented [[ ${service} == n-cpu-* && ,${ENABLED_SERVICES} =~ ,"n-cpu" ]] && enabled=0 [[ ${service} == "nova" && ,${ENABLED_SERVICES} =~ ,"n-" ]] && enabled=0 [[ ${service} == "glance" && ,${ENABLED_SERVICES} =~ ,"g-" ]] && enabled=0 [[ ${service} == "neutron" && ,${ENABLED_SERVICES} =~ ,"q-" ]] && enabled=0 [[ ${service} == "trove" && ,${ENABLED_SERVICES} =~ ,"tr-" ]] && enabled=0 [[ ${service} == "swift" && ,${ENABLED_SERVICES} =~ ,"s-" ]] && enabled=0 [[ ${service} == s-* && ,${ENABLED_SERVICES} =~ ,"swift" ]] && enabled=0 done $xtrace return $enabled } # remove specified list from the input string # remove_disabled_services service-list remove-list function remove_disabled_services { local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace local service_list=$1 local remove_list=$2 local service local enabled="" for service in ${service_list//,/ }; do local remove local add=1 for remove in ${remove_list//,/ }; do if [[ ${remove} == ${service} ]]; then add=0 break fi done if [[ $add == 1 ]]; then enabled="${enabled},$service" fi done $xtrace _cleanup_service_list "$enabled" } # Toggle enable/disable_service for services that must run exclusive of each other # $1 The name of a variable containing a space-separated list of services # $2 The name of a variable in which to store the enabled service's name # $3 The name of the service to enable function use_exclusive_service { local options=${!1} local selection=$3 local out=$2 [ -z $selection ] || [[ ! "$options" =~ "$selection" ]] && return 1 local opt for opt in $options;do [[ "$opt" = "$selection" ]] && enable_service $opt || disable_service $opt done eval "$out=$selection" return 0 } # Make sure that nothing has manipulated ENABLED_SERVICES in a way # that conflicts with prior calls to disable_service. # Uses global ``ENABLED_SERVICES`` function verify_disabled_services { local service for service in ${ENABLED_SERVICES//,/ }; do if [[ ,${DISABLED_SERVICES}, =~ ,${service}, ]]; then die $LINENO "ENABLED_SERVICES directly modified to overcome 'disable_service ${service}'" fi done } # System Functions # ================ # Only run the command if the target file (the last arg) is not on an # NFS filesystem. function _safe_permission_operation { local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace local args=( $@ ) local last local sudo_cmd local dir_to_check let last="${#args[*]} - 1" local dir_to_check=${args[$last]} if [ ! -d "$dir_to_check" ]; then dir_to_check=`dirname "$dir_to_check"` fi if is_nfs_directory "$dir_to_check" ; then $xtrace return 0 fi sudo_cmd="sudo" $xtrace $sudo_cmd $@ } # Exit 0 if address is in network or 1 if address is not in network # ip-range is in CIDR notation: 1.2.3.4/20 # address_in_net ip-address ip-range function address_in_net { local ip=$1 local range=$2 local masklen=${range#*/} local network network=$(maskip ${range%/*} $(cidr2netmask $masklen)) local subnet subnet=$(maskip $ip $(cidr2netmask $masklen)) [[ $network == $subnet ]] } # Add a user to a group. # add_user_to_group user group function add_user_to_group { local user=$1 local group=$2 sudo usermod -a -G "$group" "$user" } # Convert CIDR notation to a IPv4 netmask # cidr2netmask cidr-bits function cidr2netmask { local maskpat="255 255 255 255" local maskdgt="254 252 248 240 224 192 128" set -- ${maskpat:0:$(( ($1 / 8) * 4 ))}${maskdgt:$(( (7 - ($1 % 8)) * 4 )):3} echo ${1-0}.${2-0}.${3-0}.${4-0} } # Check if this is a valid ipv4 address string function is_ipv4_address { local address=$1 local regex='([0-9]{1,3}\.){3}[0-9]{1,3}' # TODO(clarkb) make this more robust if [[ "$address" =~ $regex ]] ; then return 0 else return 1 fi } # Remove "[]" around urlquoted IPv6 addresses function ipv6_unquote { echo $1 | tr -d [] } # Gracefully cp only if source file/dir exists # cp_it source destination function cp_it { if [ -e $1 ] || [ -d $1 ]; then cp -pRL $1 $2 fi } # HTTP and HTTPS proxy servers are supported via the usual environment variables [1] # ``http_proxy``, ``https_proxy`` and ``no_proxy``. They can be set in # ``localrc`` or on the command line if necessary:: # # [1] http://www.w3.org/Daemon/User/Proxies/ProxyClients.html # # http_proxy=http://proxy.example.com:3128/ no_proxy=repo.example.net ./stack.sh function export_proxy_variables { if isset http_proxy ; then export http_proxy=$http_proxy fi if isset https_proxy ; then export https_proxy=$https_proxy fi if isset no_proxy ; then export no_proxy=$no_proxy fi } # Returns true if the directory is on a filesystem mounted via NFS. function is_nfs_directory { local mount_type mount_type=`stat -f -L -c %T $1` test "$mount_type" == "nfs" } # Return the network portion of the given IP address using netmask # netmask is in the traditional dotted-quad format # maskip ip-address netmask function maskip { local ip=$1 local mask=$2 local l="${ip%.*}"; local r="${ip#*.}"; local n="${mask%.*}"; local m="${mask#*.}" local subnet subnet=$((${ip%%.*}&${mask%%.*})).$((${r%%.*}&${m%%.*})).$((${l##*.}&${n##*.})).$((${ip##*.}&${mask##*.})) echo $subnet } function is_provider_network { if [ "$Q_USE_PROVIDER_NETWORKING" == "True" ]; then return 0 fi return 1 } # Return just the . for the given python interpreter function _get_python_version { local interp=$1 local version # disable erroring out here, otherwise if python 3 doesn't exist we fail hard. if [[ -x $(which $interp 2> /dev/null) ]]; then version=$($interp -c 'import sys; print("%s.%s" % sys.version_info[0:2])') fi echo ${version} } # Return the current python as "python." function python_version { local python_version python_version=$(_get_python_version python2) echo "python${python_version}" } function python3_version { local python3_version python3_version=$(_get_python_version python3) echo "python${python3_version}" } # Service wrapper to restart services # restart_service service-name function restart_service { if [ -x /bin/systemctl ]; then sudo /bin/systemctl restart $1 else sudo service $1 restart fi } # Only change permissions of a file or directory if it is not on an # NFS filesystem. function safe_chmod { _safe_permission_operation chmod $@ } # Only change ownership of a file or directory if it is not on an NFS # filesystem. function safe_chown { _safe_permission_operation chown $@ } # Service wrapper to start services # start_service service-name function start_service { if [ -x /bin/systemctl ]; then sudo /bin/systemctl start $1 else sudo service $1 start fi } # Service wrapper to stop services # stop_service service-name function stop_service { if [ -x /bin/systemctl ]; then sudo /bin/systemctl stop $1 else sudo service $1 stop fi } # Service wrapper to reload services # If the service was not in running state it will start it # reload_service service-name function reload_service { if [ -x /bin/systemctl ]; then sudo /bin/systemctl reload-or-restart $1 else sudo service $1 reload fi } # Test with a finite retry loop. # function test_with_retry { local testcmd=$1 local failmsg=$2 local until=${3:-10} local sleep=${4:-0.5} time_start "test_with_retry" if ! timeout $until sh -c "while ! $testcmd; do sleep $sleep; done"; then die $LINENO "$failmsg" fi time_stop "test_with_retry" } # Like sudo but forwarding http_proxy https_proxy no_proxy environment vars. # If it is run as superuser then sudo is replaced by env. # function sudo_with_proxies { local sudo [[ "$(id -u)" = "0" ]] && sudo="env" || sudo="sudo" $sudo http_proxy="${http_proxy:-}" https_proxy="${https_proxy:-}"\ no_proxy="${no_proxy:-}" "$@" } # Timing infrastructure - figure out where large blocks of time are # used in DevStack # # The timing infrastructure for DevStack is about collecting buckets # of time that are spend in some subtask. For instance, that might be # 'apt', 'pip', 'osc', even database migrations. We do this by a pair # of functions: time_start / time_stop. # # These take a single parameter: $name - which specifies the name of # the bucket to be accounted against. time_totals function spits out # the results. # # Resolution is only in whole seconds, so should be used for long # running activities. declare -A -g _TIME_TOTAL declare -A -g _TIME_START declare -r -g _TIME_BEGIN=$(date +%s) # time_start $name # # starts the clock for a timer by name. Errors if that clock is # already started. function time_start { local name=$1 local start_time=${_TIME_START[$name]} if [[ -n "$start_time" ]]; then die $LINENO "Trying to start the clock on $name, but it's already been started" fi _TIME_START[$name]=$(date +%s%3N) } # time_stop $name # # stops the clock for a timer by name, and accumulate that time in the # global counter for that name. Errors if that clock had not # previously been started. function time_stop { local name local end_time local elapsed_time local total local start_time name=$1 start_time=${_TIME_START[$name]} if [[ -z "$start_time" ]]; then die $LINENO "Trying to stop the clock on $name, but it was never started" fi end_time=$(date +%s%3N) elapsed_time=$(($end_time - $start_time)) total=${_TIME_TOTAL[$name]:-0} # reset the clock so we can start it in the future _TIME_START[$name]="" _TIME_TOTAL[$name]=$(($total + $elapsed_time)) } function oscwrap { local xtrace xtrace=$(set +o | grep xtrace) set +o xtrace local out local rc local start local end # Cannot use timer_start and timer_stop as we run in subshells # and those rely on modifying vars in the same process (which cannot # happen from a subshell. start=$(date +%s%3N) out=$(command openstack "$@") rc=$? end=$(date +%s%3N) echo $((end - start)) >> $OSCWRAP_TIMER_FILE echo "$out" $xtrace return $rc } function install_oscwrap { # File to accumulate our timing data OSCWRAP_TIMER_FILE=$(mktemp) # Bash by default doesn't expand aliases, allow it for the aliases # we want to whitelist. shopt -s expand_aliases # Remove all aliases that might be expanded to preserve old unexpanded # behavior unalias -a # Add only the alias we want for openstack alias openstack=oscwrap } function cleanup_oscwrap { local total=0 total=$(cat $OSCWRAP_TIMER_FILE | $PYTHON -c "import sys; print(sum(int(l) for l in sys.stdin))") _TIME_TOTAL["osc"]=$total rm $OSCWRAP_TIMER_FILE } # time_totals # Print out total time summary function time_totals { local elapsed_time local end_time local len=20 local xtrace local unaccounted_time end_time=$(date +%s) elapsed_time=$(($end_time - $_TIME_BEGIN)) unaccounted_time=$elapsed_time # pad 1st column this far for t in ${!_TIME_TOTAL[*]}; do if [[ ${#t} -gt $len ]]; then len=${#t} fi done cleanup_oscwrap xtrace=$(set +o | grep xtrace) set +o xtrace echo echo "=========================" echo "DevStack Component Timing" echo " (times are in seconds) " echo "=========================" for t in ${!_TIME_TOTAL[*]}; do local v=${_TIME_TOTAL[$t]} # because we're recording in milliseconds v=$(($v / 1000)) printf "%-${len}s %3d\n" "$t" "$v" unaccounted_time=$(($unaccounted_time - $v)) done echo "-------------------------" printf "%-${len}s %3d\n" "Unaccounted time" "$unaccounted_time" echo "=========================" printf "%-${len}s %3d\n" "Total runtime" "$elapsed_time" $xtrace } function clean_pyc_files { # Clean up all *.pyc files if [[ -n "$DEST" ]] && [[ -d "$DEST" ]]; then sudo find $DEST -name "*.pyc" -delete fi } function is_fips_enabled { fips=`cat /proc/sys/crypto/fips_enabled` [ "$fips" == "1" ] } # Restore xtrace $_XTRACE_FUNCTIONS_COMMON # Local variables: # mode: shell-script # End: