From ada886dd43ccc07f48d3a82d8d3d840fe5096c03 Mon Sep 17 00:00:00 2001 From: Ian Wienand Date: Wed, 7 Oct 2015 14:06:26 +1100 Subject: [PATCH] Don't mix declaration and set of locals Ia0957b47187c3dcadd46154b17022c4213781112 proposes to have bashate find instances of setting a local value. The issue is that "local" always returns 0, thus hiding any failure in the commands running to set the variable. This is an automated replacement of such instances Depends-On: I676c805e8f0401f75cc5367eee83b3d880cdef81 Change-Id: I9c8912a8fd596535589b207d7fc553b9d951d3fe --- exercises/neutron-adv-test.sh | 36 ++++++++++----- functions | 9 ++-- functions-common | 36 ++++++++++----- inc/ini-config | 3 +- inc/meta-config | 3 +- inc/python | 18 +++++--- inc/rootwrap | 6 ++- lib/apache | 12 +++-- lib/ceph | 9 ++-- lib/cinder | 6 ++- lib/glance | 6 ++- lib/heat | 3 +- lib/horizon | 9 ++-- lib/ironic | 63 ++++++++++++++++++--------- lib/keystone | 36 ++++++++++----- lib/ldap | 9 ++-- lib/lvm | 6 ++- lib/neutron-legacy | 33 +++++++++----- lib/neutron_plugins/embrane | 3 +- lib/neutron_plugins/ovs_base | 6 ++- lib/nova | 15 ++++--- lib/swift | 42 ++++++++++++------ lib/tempest | 3 +- lib/tls | 3 +- tests/unittest.sh | 24 ++++++---- tools/create_userrc.sh | 3 +- tools/peakmem_tracker.sh | 6 ++- tools/xen/scripts/install-os-vpx.sh | 3 +- tools/xen/scripts/uninstall-os-vpx.sh | 12 +++-- tools/xen/test_functions.sh | 6 ++- 30 files changed, 286 insertions(+), 143 deletions(-) diff --git a/exercises/neutron-adv-test.sh b/exercises/neutron-adv-test.sh index a0de4ccd37..a8fbd86473 100755 --- a/exercises/neutron-adv-test.sh +++ b/exercises/neutron-adv-test.sh @@ -122,41 +122,47 @@ function foreach_tenant_net { } function get_image_id { - local IMAGE_ID=$(openstack image list | egrep " $DEFAULT_IMAGE_NAME " | get_field 1) + local IMAGE_ID + IMAGE_ID=$(openstack image list | egrep " $DEFAULT_IMAGE_NAME " | get_field 1) die_if_not_set $LINENO IMAGE_ID "Failure retrieving IMAGE_ID" echo "$IMAGE_ID" } function get_tenant_id { local TENANT_NAME=$1 - local TENANT_ID=`openstack project list | grep " $TENANT_NAME " | head -n 1 | get_field 1` + local TENANT_ID + TENANT_ID=`openstack project list | grep " $TENANT_NAME " | head -n 1 | get_field 1` die_if_not_set $LINENO TENANT_ID "Failure retrieving TENANT_ID for $TENANT_NAME" echo "$TENANT_ID" } function get_user_id { local USER_NAME=$1 - local USER_ID=`openstack user list | grep $USER_NAME | awk '{print $2}'` + local USER_ID + USER_ID=`openstack user list | grep $USER_NAME | awk '{print $2}'` die_if_not_set $LINENO USER_ID "Failure retrieving USER_ID for $USER_NAME" echo "$USER_ID" } function get_role_id { local ROLE_NAME=$1 - local ROLE_ID=`openstack role list | grep $ROLE_NAME | awk '{print $2}'` + local ROLE_ID + ROLE_ID=`openstack role list | grep $ROLE_NAME | awk '{print $2}'` die_if_not_set $LINENO ROLE_ID "Failure retrieving ROLE_ID for $ROLE_NAME" echo "$ROLE_ID" } function get_network_id { local NETWORK_NAME="$1" - local NETWORK_ID=`neutron net-list -F id -- --name=$NETWORK_NAME | awk "NR==4" | awk '{print $2}'` + local NETWORK_ID + NETWORK_ID=`neutron net-list -F id -- --name=$NETWORK_NAME | awk "NR==4" | awk '{print $2}'` echo $NETWORK_ID } function get_flavor_id { local INSTANCE_TYPE=$1 - local FLAVOR_ID=`nova flavor-list | grep $INSTANCE_TYPE | awk '{print $2}'` + local FLAVOR_ID + FLAVOR_ID=`nova flavor-list | grep $INSTANCE_TYPE | awk '{print $2}'` die_if_not_set $LINENO FLAVOR_ID "Failure retrieving FLAVOR_ID for $INSTANCE_TYPE" echo "$FLAVOR_ID" } @@ -185,13 +191,15 @@ function add_tenant { function remove_tenant { local TENANT=$1 - local TENANT_ID=$(get_tenant_id $TENANT) + local TENANT_ID + TENANT_ID=$(get_tenant_id $TENANT) openstack project delete $TENANT_ID } function remove_user { local USER=$1 - local USER_ID=$(get_user_id $USER) + local USER_ID + USER_ID=$(get_user_id $USER) openstack user delete $USER_ID } @@ -221,9 +229,11 @@ function create_network { local NET_NAME="${TENANT}-net$NUM" local ROUTER_NAME="${TENANT}-router${NUM}" source $TOP_DIR/openrc admin admin - local TENANT_ID=$(get_tenant_id $TENANT) + local TENANT_ID + TENANT_ID=$(get_tenant_id $TENANT) source $TOP_DIR/openrc $TENANT $TENANT - local NET_ID=$(neutron net-create --tenant-id $TENANT_ID $NET_NAME $EXTRA| grep ' id ' | awk '{print $4}' ) + local NET_ID + NET_ID=$(neutron net-create --tenant-id $TENANT_ID $NET_NAME $EXTRA| grep ' id ' | awk '{print $4}' ) die_if_not_set $LINENO NET_ID "Failure creating NET_ID for $TENANT_ID $NET_NAME $EXTRA" neutron subnet-create --ip-version 4 --tenant-id $TENANT_ID --gateway $GATEWAY $NET_ID $CIDR neutron_debug_admin probe-create --device-owner compute $NET_ID @@ -251,7 +261,8 @@ function create_vm { done #TODO (nati) Add multi-nic test #TODO (nati) Add public-net test - local VM_UUID=`nova boot --flavor $(get_flavor_id m1.tiny) \ + local VM_UUID + VM_UUID=`nova boot --flavor $(get_flavor_id m1.tiny) \ --image $(get_image_id) \ $NIC \ $TENANT-server$NUM | grep ' id ' | cut -d"|" -f3 | sed 's/ //g'` @@ -309,7 +320,8 @@ function delete_network { local NUM=$2 local NET_NAME="${TENANT}-net$NUM" source $TOP_DIR/openrc admin admin - local TENANT_ID=$(get_tenant_id $TENANT) + local TENANT_ID + TENANT_ID=$(get_tenant_id $TENANT) #TODO(nati) comment out until l3-agent merged #for res in port subnet net router;do for net_id in `neutron net-list -c id -c name | grep $NET_NAME | awk '{print $2}'`;do diff --git a/functions b/functions index ff95c89ad9..8cf7a25f49 100644 --- a/functions +++ b/functions @@ -264,7 +264,8 @@ function upload_image { ;; *.img) image_name=$(basename "$image" ".img") - local format=$(qemu-img info ${image} | awk '/^file format/ { print $3; exit }') + local format + format=$(qemu-img info ${image} | awk '/^file format/ { print $3; exit }') if [[ ",qcow2,raw,vdi,vmdk,vpc," =~ ",$format," ]]; then disk_format=$format else @@ -405,7 +406,8 @@ function get_instance_ip { local vm_id=$1 local network_name=$2 local nova_result="$(nova show $vm_id)" - local ip=$(echo "$nova_result" | grep "$network_name" | get_field 2) + local ip + ip=$(echo "$nova_result" | grep "$network_name" | get_field 2) if [[ $ip = "" ]];then echo "$nova_result" die $LINENO "[Fail] Coudn't get ipaddress of VM" @@ -455,7 +457,8 @@ function check_path_perm_sanity { # homedir permissions on RHEL and common practice of making DEST in # the stack user's homedir. - local real_path=$(readlink -f $1) + local real_path + real_path=$(readlink -f $1) local rebuilt_path="" for i in $(echo ${real_path} | tr "/" " "); do rebuilt_path=$rebuilt_path"/"$i diff --git a/functions-common b/functions-common index 53b64d668a..c831b261e1 100644 --- a/functions-common +++ b/functions-common @@ -140,7 +140,8 @@ function isset { # backtrace level function backtrace { local level=$1 - local deep=$((${#BASH_SOURCE[@]} - 1)) + local deep + deep=$((${#BASH_SOURCE[@]} - 1)) echo "[Call Trace]" while [ $level -le $deep ]; do echo "${BASH_SOURCE[$deep]}:${BASH_LINENO[$deep-1]}:${FUNCNAME[$deep-1]}" @@ -477,7 +478,8 @@ function git_clone { local git_remote=$1 local git_dest=$2 local git_ref=$3 - local orig_dir=$(pwd) + local orig_dir + orig_dir=$(pwd) local git_clone_flags="" RECLONE=$(trueorfalse False RECLONE) @@ -641,7 +643,8 @@ function get_default_host_ip { host_ip="" # Find the interface used for the default route host_ip_iface=${host_ip_iface:-$(ip -f $af route | awk '/default/ {print $5}' | head -1)} - local host_ips=$(LC_ALL=C ip -f $af addr show ${host_ip_iface} | sed /temporary/d |awk /$af'/ {split($2,parts,"/"); print parts[1]}') + local host_ips + host_ips=$(LC_ALL=C ip -f $af addr show ${host_ip_iface} | sed /temporary/d |awk /$af'/ {split($2,parts,"/"); print parts[1]}') local ip for ip in $host_ips; do # Attempt to filter out IP addresses that are part of the fixed and @@ -690,7 +693,8 @@ function get_field { # copy over a default policy.json and policy.d for projects function install_default_policy { local project=$1 - local project_uc=$(echo $1|tr a-z A-Z) + local project_uc + project_uc=$(echo $1|tr a-z A-Z) local conf_dir="${project_uc}_CONF_DIR" # eval conf dir to get the variable conf_dir="${!conf_dir}" @@ -723,7 +727,8 @@ function policy_add { # Add a terminating comma to policy lines without one # Remove the closing '}' and all lines following to the end-of-file - local tmpfile=$(mktemp) + local tmpfile + tmpfile=$(mktemp) uniq ${policy_file} | sed -e ' s/]$/],/ /^[}]/,$d @@ -945,7 +950,8 @@ function get_or_create_endpoint { # scenarios currently that use the returned id. Ideally this behaviour # should be pushed out to the service setups and let them create the # endpoints they need. - local public_id=$(_get_or_create_endpoint_with_interface $1 public $3 $2) + local public_id + public_id=$(_get_or_create_endpoint_with_interface $1 public $3 $2) _get_or_create_endpoint_with_interface $1 admin $4 $2 _get_or_create_endpoint_with_interface $1 internal $5 $2 @@ -1065,7 +1071,8 @@ function get_packages { xtrace=$(set +o | grep xtrace) set +o xtrace local services=$@ - local package_dir=$(_get_package_dir) + local package_dir + package_dir=$(_get_package_dir) local file_to_parse="" local service="" @@ -1980,8 +1987,10 @@ function address_in_net { local ip=$1 local range=$2 local masklen=${range#*/} - local network=$(maskip ${range%/*} $(cidr2netmask $masklen)) - local subnet=$(maskip $ip $(cidr2netmask $masklen)) + local network + network=$(maskip ${range%/*} $(cidr2netmask $masklen)) + local subnet + subnet=$(maskip $ip $(cidr2netmask $masklen)) [[ $network == $subnet ]] } @@ -2033,7 +2042,8 @@ function export_proxy_variables { # Returns true if the directory is on a filesystem mounted via NFS. function is_nfs_directory { - local mount_type=`stat -f -L -c %T $1` + local mount_type + mount_type=`stat -f -L -c %T $1` test "$mount_type" == "nfs" } @@ -2044,13 +2054,15 @@ function maskip { local ip=$1 local mask=$2 local l="${ip%.*}"; local r="${ip#*.}"; local n="${mask%.*}"; local m="${mask#*.}" - local subnet=$((${ip%%.*}&${mask%%.*})).$((${r%%.*}&${m%%.*})).$((${l##*.}&${n##*.})).$((${ip##*.}&${mask##*.})) + local subnet + subnet=$((${ip%%.*}&${mask%%.*})).$((${r%%.*}&${m%%.*})).$((${l##*.}&${n##*.})).$((${ip##*.}&${mask##*.})) echo $subnet } # Return the current python as "python." function python_version { - local python_version=$(python -c 'import sys; print("%s.%s" % sys.version_info[0:2])') + local python_version + python_version=$(python -c 'import sys; print("%s.%s" % sys.version_info[0:2])') echo "python${python_version}" } diff --git a/inc/ini-config b/inc/ini-config index ba2d827ae9..42a66c63b6 100644 --- a/inc/ini-config +++ b/inc/ini-config @@ -196,7 +196,8 @@ function iniset { $option = $value " "$file" else - local sep=$(echo -ne "\x01") + local sep + sep=$(echo -ne "\x01") # Replace it $sudo sed -i -e '/^\['${section}'\]/,/^\[.*\]/ s'${sep}'^\('${option}'[ \t]*=[ \t]*\).*$'${sep}'\1'"${value}"${sep} "$file" fi diff --git a/inc/meta-config b/inc/meta-config index e5f902d1dd..d74db59bb3 100644 --- a/inc/meta-config +++ b/inc/meta-config @@ -89,7 +89,8 @@ function merge_config_file { # note, configfile might be a variable (note the iniset, etc # created in the mega-awk below is "eval"ed too, so we just leave # it alone. - local real_configfile=$(eval echo $configfile) + local real_configfile + real_configfile=$(eval echo $configfile) if [ ! -f $real_configfile ]; then touch $real_configfile fi diff --git a/inc/python b/inc/python index fe7bba6992..c7ba51a81d 100644 --- a/inc/python +++ b/inc/python @@ -61,7 +61,8 @@ function get_python_exec_prefix { # pip_install_gr packagename function pip_install_gr { local name=$1 - local clean_name=$(get_from_global_requirements $name) + local clean_name + clean_name=$(get_from_global_requirements $name) pip_install $clean_name } @@ -100,7 +101,8 @@ function pip_install { local cmd_pip=$PIP_VIRTUAL_ENV/bin/pip local sudo_pip="env" else - local cmd_pip=$(get_pip_command) + local cmd_pip + cmd_pip=$(get_pip_command) local sudo_pip="sudo -H" fi fi @@ -109,7 +111,8 @@ function pip_install { # Always apply constraints cmd_pip="$cmd_pip -c $REQUIREMENTS_DIR/upper-constraints.txt" - local pip_version=$(python -c "import pip; \ + local pip_version + pip_version=$(python -c "import pip; \ print(pip.__version__.strip('.')[0])") if (( pip_version<6 )); then die $LINENO "Currently installed pip version ${pip_version} does not" \ @@ -143,7 +146,8 @@ function pip_install { # get_from_global_requirements function get_from_global_requirements { local package=$1 - local required_pkg=$(grep -i -h ^${package} $REQUIREMENTS_DIR/global-requirements.txt | cut -d\# -f1) + local required_pkg + required_pkg=$(grep -i -h ^${package} $REQUIREMENTS_DIR/global-requirements.txt | cut -d\# -f1) if [[ $required_pkg == "" ]]; then die $LINENO "Can't find package $package in requirements" fi @@ -222,7 +226,8 @@ function setup_develop { # practical ways. function is_in_projects_txt { local project_dir=$1 - local project_name=$(basename $project_dir) + local project_name + project_name=$(basename $project_dir) grep -q "/$project_name\$" $REQUIREMENTS_DIR/projects.txt } @@ -241,7 +246,8 @@ function setup_package_with_constraints_edit { if [ -n "$REQUIREMENTS_DIR" ]; then # Constrain this package to this project directory from here on out. - local name=$(awk '/^name.*=/ {print $3}' $project_dir/setup.cfg) + local name + name=$(awk '/^name.*=/ {print $3}' $project_dir/setup.cfg) $REQUIREMENTS_DIR/.venv/bin/edit-constraints \ $REQUIREMENTS_DIR/upper-constraints.txt -- $name \ "$flags file://$project_dir#egg=$name" diff --git a/inc/rootwrap b/inc/rootwrap index f91e557e68..63ab59adc7 100644 --- a/inc/rootwrap +++ b/inc/rootwrap @@ -41,7 +41,8 @@ function add_sudo_secure_path { # configure_rootwrap project function configure_rootwrap { local project=$1 - local project_uc=$(echo $1|tr a-z A-Z) + local project_uc + project_uc=$(echo $1|tr a-z A-Z) local bin_dir="${project_uc}_BIN_DIR" bin_dir="${!bin_dir}" local project_dir="${project_uc}_DIR" @@ -60,7 +61,8 @@ function configure_rootwrap { sudo sed -e "s:^filters_path=.*$:filters_path=/etc/${project}/rootwrap.d:" -i /etc/${project}/rootwrap.conf # Set up the rootwrap sudoers - local tempfile=$(mktemp) + local tempfile + tempfile=$(mktemp) # Specify rootwrap.conf as first parameter to rootwrap rootwrap_sudo_cmd="${rootwrap_bin} /etc/${project}/rootwrap.conf *" echo "$STACK_USER ALL=(root) NOPASSWD: $rootwrap_sudo_cmd" >$tempfile diff --git a/lib/apache b/lib/apache index a8e9bc5ad2..17526c74d0 100644 --- a/lib/apache +++ b/lib/apache @@ -72,11 +72,14 @@ function install_apache_wsgi { # various differences between Apache 2.2 and 2.4 that warrant special handling. function get_apache_version { if is_ubuntu; then - local version_str=$(sudo /usr/sbin/apache2ctl -v | awk '/Server version/ {print $3}' | cut -f2 -d/) + local version_str + version_str=$(sudo /usr/sbin/apache2ctl -v | awk '/Server version/ {print $3}' | cut -f2 -d/) elif is_fedora; then - local version_str=$(rpm -qa --queryformat '%{VERSION}' httpd) + local version_str + version_str=$(rpm -qa --queryformat '%{VERSION}' httpd) elif is_suse; then - local version_str=$(rpm -qa --queryformat '%{VERSION}' apache2) + local version_str + version_str=$(rpm -qa --queryformat '%{VERSION}' apache2) else exit_distro_not_supported "cannot determine apache version" fi @@ -115,7 +118,8 @@ function get_apache_version { function apache_site_config_for { local site=$@ if is_ubuntu; then - local apache_version=$(get_apache_version) + local apache_version + apache_version=$(get_apache_version) if [[ "$apache_version" == "2.2" ]]; then # Ubuntu 12.04 - Apache 2.2 echo $APACHE_CONF_DIR/${site} diff --git a/lib/ceph b/lib/ceph index 8e34aa49a4..29d2aca54f 100644 --- a/lib/ceph +++ b/lib/ceph @@ -83,7 +83,8 @@ ATTACH_ENCRYPTED_VOLUME_AVAILABLE=False # ------------ function get_ceph_version { - local ceph_version_str=$(sudo ceph daemon mon.$(hostname) version | cut -d '"' -f 4 | cut -f 1,2 -d '.') + local ceph_version_str + ceph_version_str=$(sudo ceph daemon mon.$(hostname) version | cut -d '"' -f 4 | cut -f 1,2 -d '.') echo $ceph_version_str } @@ -106,7 +107,8 @@ EOF # undefine_virsh_secret() - Undefine Cinder key secret from libvirt function undefine_virsh_secret { if is_service_enabled cinder || is_service_enabled nova; then - local virsh_uuid=$(sudo virsh secret-list | awk '/^ ?[0-9a-z]/ { print $1 }') + local virsh_uuid + virsh_uuid=$(sudo virsh secret-list | awk '/^ ?[0-9a-z]/ { print $1 }') sudo virsh secret-undefine ${virsh_uuid} >/dev/null 2>&1 fi } @@ -219,7 +221,8 @@ EOF done # pools data and metadata were removed in the Giant release so depending on the version we apply different commands - local ceph_version=$(get_ceph_version) + local ceph_version + ceph_version=$(get_ceph_version) # change pool replica size according to the CEPH_REPLICAS set by the user if [[ ${ceph_version%%.*} -eq 0 ]] && [[ ${ceph_version##*.} -lt 87 ]]; then sudo ceph -c ${CEPH_CONF_FILE} osd pool set rbd size ${CEPH_REPLICAS} diff --git a/lib/cinder b/lib/cinder index 10144117ec..73941c6973 100644 --- a/lib/cinder +++ b/lib/cinder @@ -150,7 +150,8 @@ function cleanup_cinder { # ensure the volume group is cleared up because fails might # leave dead volumes in the group if [ "$CINDER_ISCSI_HELPER" = "tgtadm" ]; then - local targets=$(sudo tgtadm --op show --mode target) + local targets + targets=$(sudo tgtadm --op show --mode target) if [ $? -ne 0 ]; then # If tgt driver isn't running this won't work obviously # So check the response and restart if need be @@ -198,7 +199,8 @@ function cleanup_cinder { # _cinder_config_apache_wsgi() - Set WSGI config files function _cinder_config_apache_wsgi { - local cinder_apache_conf=$(apache_site_config_for osapi-volume) + local cinder_apache_conf + cinder_apache_conf=$(apache_site_config_for osapi-volume) local cinder_ssl="" local cinder_certfile="" local cinder_keyfile="" diff --git a/lib/glance b/lib/glance index 7be3a8495c..2eb93a46e6 100644 --- a/lib/glance +++ b/lib/glance @@ -106,7 +106,8 @@ function configure_glance { iniset $GLANCE_REGISTRY_CONF DEFAULT debug $ENABLE_DEBUG_LOG_LEVEL iniset $GLANCE_REGISTRY_CONF DEFAULT bind_host $GLANCE_SERVICE_LISTEN_ADDRESS inicomment $GLANCE_REGISTRY_CONF DEFAULT log_file - local dburl=`database_connection_url glance` + local dburl + dburl=`database_connection_url glance` iniset $GLANCE_REGISTRY_CONF database connection $dburl iniset $GLANCE_REGISTRY_CONF DEFAULT use_syslog $SYSLOG iniset $GLANCE_REGISTRY_CONF DEFAULT workers "$API_WORKERS" @@ -265,7 +266,8 @@ function create_glance_accounts { # required for swift access if is_service_enabled s-proxy; then - local glance_swift_user=$(get_or_create_user "glance-swift" \ + local glance_swift_user + glance_swift_user=$(get_or_create_user "glance-swift" \ "$SERVICE_PASSWORD" "default" "glance-swift@example.com") get_or_add_user_project_role "ResellerAdmin" $glance_swift_user $SERVICE_TENANT_NAME fi diff --git a/lib/heat b/lib/heat index 3e6975ae26..df85c72cfe 100644 --- a/lib/heat +++ b/lib/heat @@ -321,7 +321,8 @@ function build_heat_pip_mirror { echo "" >> $HEAT_PIP_REPO/index.html - local heat_pip_repo_apache_conf=$(apache_site_config_for heat_pip_repo) + local heat_pip_repo_apache_conf + heat_pip_repo_apache_conf=$(apache_site_config_for heat_pip_repo) sudo cp $FILES/apache-heat-pip-repo.template $heat_pip_repo_apache_conf sudo sed -e " diff --git a/lib/horizon b/lib/horizon index b2539d1b7d..6ecd755795 100644 --- a/lib/horizon +++ b/lib/horizon @@ -49,7 +49,8 @@ function _horizon_config_set { sed -e "/^$option/d" -i $local_settings echo -e "\n$option=$value" >> $file elif grep -q "^$section" $file; then - local line=$(sed -ne "/^$section/,/^}/ { /^ *'$option':/ p; }" $file) + local line + line=$(sed -ne "/^$section/,/^}/ { /^ *'$option':/ p; }" $file) if [ -n "$line" ]; then sed -i -e "/^$section/,/^}/ s/^\( *'$option'\) *:.*$/\1: $value,/" $file else @@ -68,7 +69,8 @@ function _horizon_config_set { # cleanup_horizon() - Remove residual data files, anything left over from previous # runs that a clean run would need to clean up function cleanup_horizon { - local horizon_conf=$(apache_site_config_for horizon) + local horizon_conf + horizon_conf=$(apache_site_config_for horizon) sudo rm -f $horizon_conf } @@ -112,7 +114,8 @@ function init_horizon { # Create an empty directory that apache uses as docroot sudo mkdir -p $HORIZON_DIR/.blackhole - local horizon_conf=$(apache_site_config_for horizon) + local horizon_conf + horizon_conf=$(apache_site_config_for horizon) # Configure apache to run horizon sudo sh -c "sed -e \" diff --git a/lib/ironic b/lib/ironic index 40475e0a83..8eb0d80e07 100644 --- a/lib/ironic +++ b/lib/ironic @@ -225,7 +225,8 @@ function _cleanup_ironic_apache_wsgi { # _config_ironic_apache_wsgi() - Set WSGI config files of Ironic function _config_ironic_apache_wsgi { - local ironic_apache_conf=$(apache_site_config_for ironic) + local ironic_apache_conf + ironic_apache_conf=$(apache_site_config_for ironic) sudo cp $FILES/apache-ironic.template $ironic_apache_conf sudo sed -e " s|%PUBLICPORT%|$IRONIC_HTTP_PORT|g; @@ -325,11 +326,13 @@ function configure_ironic_api { function configure_ironic_conductor { cp $IRONIC_DIR/etc/ironic/rootwrap.conf $IRONIC_ROOTWRAP_CONF cp -r $IRONIC_DIR/etc/ironic/rootwrap.d $IRONIC_CONF_DIR - local ironic_rootwrap=$(get_rootwrap_location ironic) + local ironic_rootwrap + ironic_rootwrap=$(get_rootwrap_location ironic) local rootwrap_isudoer_cmd="$ironic_rootwrap $IRONIC_CONF_DIR/rootwrap.conf *" # Set up the rootwrap sudoers for ironic - local tempfile=`mktemp` + local tempfile + tempfile=`mktemp` echo "$STACK_USER ALL=(root) NOPASSWD: $rootwrap_isudoer_cmd" >$tempfile chmod 0440 $tempfile sudo chown root:root $tempfile @@ -370,7 +373,8 @@ function configure_ironic_conductor { fi iniset $IRONIC_CONF_FILE glance swift_endpoint_url http://${HOST_IP}:${SWIFT_DEFAULT_BIND_PORT:-8080} iniset $IRONIC_CONF_FILE glance swift_api_version v1 - local tenant_id=$(get_or_create_project $SERVICE_TENANT_NAME default) + local tenant_id + tenant_id=$(get_or_create_project $SERVICE_TENANT_NAME default) iniset $IRONIC_CONF_FILE glance swift_account AUTH_${tenant_id} iniset $IRONIC_CONF_FILE glance swift_container glance iniset $IRONIC_CONF_FILE glance swift_temp_url_duration 3600 @@ -379,7 +383,8 @@ function configure_ironic_conductor { fi if [[ "$IRONIC_IPXE_ENABLED" == "True" ]] ; then - local pxebin=`basename $IRONIC_PXE_BOOT_IMAGE` + local pxebin + pxebin=`basename $IRONIC_PXE_BOOT_IMAGE` iniset $IRONIC_CONF_FILE pxe ipxe_enabled True iniset $IRONIC_CONF_FILE pxe pxe_config_template '\$pybasedir/drivers/modules/ipxe_config.template' iniset $IRONIC_CONF_FILE pxe pxe_bootfile_name $pxebin @@ -445,7 +450,8 @@ function init_ironic { # _ironic_bm_vm_names() - Generates list of names for baremetal VMs. function _ironic_bm_vm_names { local idx - local num_vms=$(($IRONIC_VM_COUNT - 1)) + local num_vms + num_vms=$(($IRONIC_VM_COUNT - 1)) for idx in $(seq 0 $num_vms); do echo "baremetal${IRONIC_VM_NETWORK_BRIDGE}_${idx}" done @@ -498,22 +504,27 @@ function stop_ironic { } function create_ovs_taps { - local ironic_net_id=$(neutron net-list | grep private | get_field 1) + local ironic_net_id + ironic_net_id=$(neutron net-list | grep private | get_field 1) # Work around: No netns exists on host until a Neutron port is created. We # need to create one in Neutron to know what netns to tap into prior to the # first node booting. - local port_id=$(neutron port-create private | grep " id " | get_field 2) + local port_id + port_id=$(neutron port-create private | grep " id " | get_field 2) # intentional sleep to make sure the tag has been set to port sleep 10 if [[ "$Q_USE_NAMESPACE" = "True" ]]; then - local tapdev=$(sudo ip netns exec qdhcp-${ironic_net_id} ip link list | grep " tap" | cut -d':' -f2 | cut -d'@' -f1 | cut -b2-) + local tapdev + tapdev=$(sudo ip netns exec qdhcp-${ironic_net_id} ip link list | grep " tap" | cut -d':' -f2 | cut -d'@' -f1 | cut -b2-) else - local tapdev=$(sudo ip link list | grep " tap" | cut -d':' -f2 | cut -d'@' -f1 | cut -b2-) + local tapdev + tapdev=$(sudo ip link list | grep " tap" | cut -d':' -f2 | cut -d'@' -f1 | cut -b2-) fi - local tag_id=$(sudo ovs-vsctl show |grep ${tapdev} -A1 -m1 | grep tag | cut -d':' -f2 | cut -b2-) + local tag_id + tag_id=$(sudo ovs-vsctl show |grep ${tapdev} -A1 -m1 | grep tag | cut -d':' -f2 | cut -b2-) # make sure veth pair is not existing, otherwise delete its links sudo ip link show ovs-tap1 && sudo ip link delete ovs-tap1 @@ -570,7 +581,8 @@ function wait_for_nova_resources { } function enroll_nodes { - local chassis_id=$(ironic chassis-create -d "ironic test chassis" | grep " uuid " | get_field 2) + local chassis_id + chassis_id=$(ironic chassis-create -d "ironic test chassis" | grep " uuid " | get_field 2) if ! is_ironic_hardware; then local ironic_node_cpu=$IRONIC_VM_SPECS_CPU @@ -602,10 +614,14 @@ function enroll_nodes { if ! is_ironic_hardware; then local mac_address=$hardware_info elif [[ -z "${IRONIC_DEPLOY_DRIVER##*_ipmitool}" ]]; then - local ipmi_address=$(echo $hardware_info |awk '{print $1}') - local mac_address=$(echo $hardware_info |awk '{print $2}') - local ironic_ipmi_username=$(echo $hardware_info |awk '{print $3}') - local ironic_ipmi_passwd=$(echo $hardware_info |awk '{print $4}') + local ipmi_address + ipmi_address=$(echo $hardware_info |awk '{print $1}') + local mac_address + mac_address=$(echo $hardware_info |awk '{print $2}') + local ironic_ipmi_username + ironic_ipmi_username=$(echo $hardware_info |awk '{print $3}') + local ironic_ipmi_passwd + ironic_ipmi_passwd=$(echo $hardware_info |awk '{print $4}') # Currently we require all hardware platform have same CPU/RAM/DISK info # in future, this can be enhanced to support different type, and then # we create the bare metal flavor with minimum value @@ -617,9 +633,11 @@ function enroll_nodes { # First node created will be used for testing in ironic w/o glance # scenario, so we need to know its UUID. - local standalone_node_uuid=$([ $total_nodes -eq 0 ] && echo "--uuid $IRONIC_NODE_UUID") + local standalone_node_uuid + standalone_node_uuid=$([ $total_nodes -eq 0 ] && echo "--uuid $IRONIC_NODE_UUID") - local node_id=$(ironic node-create $standalone_node_uuid\ + local node_id + node_id=$(ironic node-create $standalone_node_uuid\ --chassis_uuid $chassis_id \ --driver $IRONIC_DEPLOY_DRIVER \ --name node-$total_nodes \ @@ -640,7 +658,8 @@ function enroll_nodes { # NOTE(adam_g): Attempting to use an autogenerated UUID for flavor id here uncovered # bug (LP: #1333852) in Trove. This can be changed to use an auto flavor id when the # bug is fixed in Juno. - local adjusted_disk=$(($ironic_node_disk - $ironic_ephemeral_disk)) + local adjusted_disk + adjusted_disk=$(($ironic_node_disk - $ironic_ephemeral_disk)) nova flavor-create --ephemeral $ironic_ephemeral_disk baremetal 551 $ironic_node_ram $adjusted_disk $ironic_node_cpu nova flavor-key baremetal set "cpu_arch"="x86_64" @@ -771,7 +790,8 @@ function upload_baremetal_ironic_deploy { fi fi - local token=$(openstack token issue -c id -f value) + local token + token=$(openstack token issue -c id -f value) die_if_not_set $LINENO token "Keystone fail to get token" # load them into glance @@ -809,7 +829,8 @@ function prepare_baremetal_basic_ops { function cleanup_baremetal_basic_ops { rm -f $IRONIC_VM_MACS_CSV_FILE if [ -f $IRONIC_KEY_FILE ]; then - local key=$(cat $IRONIC_KEY_FILE.pub) + local key + key=$(cat $IRONIC_KEY_FILE.pub) # remove public key from authorized_keys grep -v "$key" $IRONIC_AUTHORIZED_KEYS_FILE > temp && mv temp $IRONIC_AUTHORIZED_KEYS_FILE chmod 0600 $IRONIC_AUTHORIZED_KEYS_FILE diff --git a/lib/keystone b/lib/keystone index ec28b46341..cdcc13a326 100644 --- a/lib/keystone +++ b/lib/keystone @@ -132,7 +132,8 @@ function _cleanup_keystone_apache_wsgi { # _config_keystone_apache_wsgi() - Set WSGI config files of Keystone function _config_keystone_apache_wsgi { - local keystone_apache_conf=$(apache_site_config_for keystone) + local keystone_apache_conf + keystone_apache_conf=$(apache_site_config_for keystone) local keystone_ssl="" local keystone_certfile="" local keystone_keyfile="" @@ -347,9 +348,12 @@ function configure_keystone_extensions { function create_keystone_accounts { # admin - local admin_tenant=$(get_or_create_project "admin" default) - local admin_user=$(get_or_create_user "admin" "$ADMIN_PASSWORD" default) - local admin_role=$(get_or_create_role "admin") + local admin_tenant + admin_tenant=$(get_or_create_project "admin" default) + local admin_user + admin_user=$(get_or_create_user "admin" "$ADMIN_PASSWORD" default) + local admin_role + admin_role=$(get_or_create_role "admin") get_or_add_user_project_role $admin_role $admin_user $admin_tenant # Create service project/role @@ -365,18 +369,23 @@ function create_keystone_accounts { get_or_create_role ResellerAdmin # The Member role is used by Horizon and Swift so we need to keep it: - local member_role=$(get_or_create_role "Member") + local member_role + member_role=$(get_or_create_role "Member") # another_role demonstrates that an arbitrary role may be created and used # TODO(sleepsonthefloor): show how this can be used for rbac in the future! - local another_role=$(get_or_create_role "anotherrole") + local another_role + another_role=$(get_or_create_role "anotherrole") # invisible tenant - admin can't see this one - local invis_tenant=$(get_or_create_project "invisible_to_admin" default) + local invis_tenant + invis_tenant=$(get_or_create_project "invisible_to_admin" default) # demo - local demo_tenant=$(get_or_create_project "demo" default) - local demo_user=$(get_or_create_user "demo" \ + local demo_tenant + demo_tenant=$(get_or_create_project "demo" default) + local demo_user + demo_user=$(get_or_create_user "demo" \ "$ADMIN_PASSWORD" "default" "demo@example.com") get_or_add_user_project_role $member_role $demo_user $demo_tenant @@ -384,9 +393,11 @@ function create_keystone_accounts { get_or_add_user_project_role $another_role $demo_user $demo_tenant get_or_add_user_project_role $member_role $demo_user $invis_tenant - local admin_group=$(get_or_create_group "admins" \ + local admin_group + admin_group=$(get_or_create_group "admins" \ "default" "openstack admin group") - local non_admin_group=$(get_or_create_group "nonadmins" \ + local non_admin_group + non_admin_group=$(get_or_create_group "nonadmins" \ "default" "non-admin group") get_or_add_group_project_role $member_role $non_admin_group $demo_tenant @@ -415,7 +426,8 @@ function create_keystone_accounts { function create_service_user { local role=${2:-service} - local user=$(get_or_create_user "$1" "$SERVICE_PASSWORD" default) + local user + user=$(get_or_create_user "$1" "$SERVICE_PASSWORD" default) get_or_add_user_project_role "$role" "$user" "$SERVICE_TENANT_NAME" } diff --git a/lib/ldap b/lib/ldap index d2dbc3b728..0414fea639 100644 --- a/lib/ldap +++ b/lib/ldap @@ -82,7 +82,8 @@ function cleanup_ldap { function init_ldap { local keystone_ldif - local tmp_ldap_dir=$(mktemp -d -t ldap.$$.XXXXXXXXXX) + local tmp_ldap_dir + tmp_ldap_dir=$(mktemp -d -t ldap.$$.XXXXXXXXXX) # Remove data but not schemas clear_ldap_state @@ -113,7 +114,8 @@ function install_ldap { echo "Installing LDAP inside function" echo "os_VENDOR is $os_VENDOR" - local tmp_ldap_dir=$(mktemp -d -t ldap.$$.XXXXXXXXXX) + local tmp_ldap_dir + tmp_ldap_dir=$(mktemp -d -t ldap.$$.XXXXXXXXXX) printf "installing OpenLDAP" if is_ubuntu; then @@ -129,7 +131,8 @@ function install_ldap { fi echo "LDAP_PASSWORD is $LDAP_PASSWORD" - local slappass=$(slappasswd -s $LDAP_PASSWORD) + local slappass + slappass=$(slappasswd -s $LDAP_PASSWORD) printf "LDAP secret is $slappass\n" # Create manager.ldif and add to olcdb diff --git a/lib/lvm b/lib/lvm index 8afd543f34..468a99aecc 100644 --- a/lib/lvm +++ b/lib/lvm @@ -56,7 +56,8 @@ function _clean_lvm_backing_file { # If the backing physical device is a loop device, it was probably setup by DevStack if [[ -n "$backing_file" ]] && [[ -e "$backing_file" ]]; then - local vg_dev=$(sudo losetup -j $backing_file | awk -F':' '/'$BACKING_FILE_SUFFIX'/ { print $1}') + local vg_dev + vg_dev=$(sudo losetup -j $backing_file | awk -F':' '/'$BACKING_FILE_SUFFIX'/ { print $1}') sudo losetup -d $vg_dev rm -f $backing_file fi @@ -89,7 +90,8 @@ function _create_lvm_volume_group { if ! sudo vgs $vg; then # Only create if the file doesn't already exists [[ -f $backing_file ]] || truncate -s $size $backing_file - local vg_dev=`sudo losetup -f --show $backing_file` + local vg_dev + vg_dev=`sudo losetup -f --show $backing_file` # Only create volume group if it doesn't already exist if ! sudo vgs $vg; then diff --git a/lib/neutron-legacy b/lib/neutron-legacy index e67bd4ae32..e9f88fbd1b 100644 --- a/lib/neutron-legacy +++ b/lib/neutron-legacy @@ -806,7 +806,8 @@ function _move_neutron_addresses_route { local IP_ADD="" local IP_DEL="" - local DEFAULT_ROUTE_GW=$(ip r | awk "/default.+$from_intf/ { print \$3; exit }") + local DEFAULT_ROUTE_GW + DEFAULT_ROUTE_GW=$(ip r | awk "/default.+$from_intf/ { print \$3; exit }") local ADD_OVS_PORT="" if [[ $af == "inet" ]]; then @@ -1244,7 +1245,8 @@ function _neutron_create_private_subnet_v4 { subnet_params+="--gateway $NETWORK_GATEWAY " subnet_params+="--name $PRIVATE_SUBNET_NAME " subnet_params+="$NET_ID $FIXED_RANGE" - local subnet_id=$(neutron subnet-create $subnet_params | grep ' id ' | get_field 2) + local subnet_id + subnet_id=$(neutron subnet-create $subnet_params | grep ' id ' | get_field 2) die_if_not_set $LINENO subnet_id "Failure creating private IPv4 subnet for $TENANT_ID" echo $subnet_id } @@ -1259,7 +1261,8 @@ function _neutron_create_private_subnet_v6 { subnet_params+="--gateway $IPV6_PRIVATE_NETWORK_GATEWAY " subnet_params+="--name $IPV6_PRIVATE_SUBNET_NAME " subnet_params+="$NET_ID $FIXED_RANGE_V6 $ipv6_modes" - local ipv6_subnet_id=$(neutron subnet-create $subnet_params | grep ' id ' | get_field 2) + local ipv6_subnet_id + ipv6_subnet_id=$(neutron subnet-create $subnet_params | grep ' id ' | get_field 2) die_if_not_set $LINENO ipv6_subnet_id "Failure creating private IPv6 subnet for $TENANT_ID" echo $ipv6_subnet_id } @@ -1272,7 +1275,8 @@ function _neutron_create_public_subnet_v4 { subnet_params+="--name $PUBLIC_SUBNET_NAME " subnet_params+="$EXT_NET_ID $FLOATING_RANGE " subnet_params+="-- --enable_dhcp=False" - local id_and_ext_gw_ip=$(neutron subnet-create $subnet_params | grep -e 'gateway_ip' -e ' id ') + local id_and_ext_gw_ip + id_and_ext_gw_ip=$(neutron subnet-create $subnet_params | grep -e 'gateway_ip' -e ' id ') die_if_not_set $LINENO id_and_ext_gw_ip "Failure creating public IPv4 subnet" echo $id_and_ext_gw_ip } @@ -1284,7 +1288,8 @@ function _neutron_create_public_subnet_v6 { subnet_params+="--name $IPV6_PUBLIC_SUBNET_NAME " subnet_params+="$EXT_NET_ID $IPV6_PUBLIC_RANGE " subnet_params+="-- --enable_dhcp=False" - local ipv6_id_and_ext_gw_ip=$(neutron subnet-create $subnet_params | grep -e 'gateway_ip' -e ' id ') + local ipv6_id_and_ext_gw_ip + ipv6_id_and_ext_gw_ip=$(neutron subnet-create $subnet_params | grep -e 'gateway_ip' -e ' id ') die_if_not_set $LINENO ipv6_id_and_ext_gw_ip "Failure creating an IPv6 public subnet" echo $ipv6_id_and_ext_gw_ip } @@ -1293,8 +1298,10 @@ function _neutron_create_public_subnet_v6 { function _neutron_configure_router_v4 { neutron router-interface-add $ROUTER_ID $SUBNET_ID # Create a public subnet on the external network - local id_and_ext_gw_ip=$(_neutron_create_public_subnet_v4 $EXT_NET_ID) - local ext_gw_ip=$(echo $id_and_ext_gw_ip | get_field 2) + local id_and_ext_gw_ip + id_and_ext_gw_ip=$(_neutron_create_public_subnet_v4 $EXT_NET_ID) + local ext_gw_ip + ext_gw_ip=$(echo $id_and_ext_gw_ip | get_field 2) PUB_SUBNET_ID=$(echo $id_and_ext_gw_ip | get_field 5) # Configure the external network as the default router gateway neutron router-gateway-set $ROUTER_ID $EXT_NET_ID @@ -1331,9 +1338,12 @@ function _neutron_configure_router_v4 { function _neutron_configure_router_v6 { neutron router-interface-add $ROUTER_ID $IPV6_SUBNET_ID # Create a public subnet on the external network - local ipv6_id_and_ext_gw_ip=$(_neutron_create_public_subnet_v6 $EXT_NET_ID) - local ipv6_ext_gw_ip=$(echo $ipv6_id_and_ext_gw_ip | get_field 2) - local ipv6_pub_subnet_id=$(echo $ipv6_id_and_ext_gw_ip | get_field 5) + local ipv6_id_and_ext_gw_ip + ipv6_id_and_ext_gw_ip=$(_neutron_create_public_subnet_v6 $EXT_NET_ID) + local ipv6_ext_gw_ip + ipv6_ext_gw_ip=$(echo $ipv6_id_and_ext_gw_ip | get_field 2) + local ipv6_pub_subnet_id + ipv6_pub_subnet_id=$(echo $ipv6_id_and_ext_gw_ip | get_field 5) # If the external network has not already been set as the default router # gateway when configuring an IPv4 public subnet, do so now @@ -1351,7 +1361,8 @@ function _neutron_configure_router_v6 { die_if_not_set $LINENO IPV6_ROUTER_GW_IP "Failure retrieving IPV6_ROUTER_GW_IP" if is_neutron_ovs_base_plugin && [[ "$Q_USE_NAMESPACE" = "True" ]]; then - local ext_gw_interface=$(_neutron_get_ext_gw_interface) + local ext_gw_interface + ext_gw_interface=$(_neutron_get_ext_gw_interface) local ipv6_cidr_len=${IPV6_PUBLIC_RANGE#*/} # Configure interface for public bridge diff --git a/lib/neutron_plugins/embrane b/lib/neutron_plugins/embrane index 6b4819ef70..2028496ca1 100644 --- a/lib/neutron_plugins/embrane +++ b/lib/neutron_plugins/embrane @@ -10,7 +10,8 @@ set +o xtrace source $TOP_DIR/lib/neutron_plugins/openvswitch function save_function { - local ORIG_FUNC=$(declare -f $1) + local ORIG_FUNC + ORIG_FUNC=$(declare -f $1) local NEW_FUNC="$2${ORIG_FUNC#$1}" eval "$NEW_FUNC" } diff --git a/lib/neutron_plugins/ovs_base b/lib/neutron_plugins/ovs_base index b012683a6f..d3fd198b08 100644 --- a/lib/neutron_plugins/ovs_base +++ b/lib/neutron_plugins/ovs_base @@ -49,8 +49,10 @@ function neutron_ovs_base_cleanup { function _neutron_ovs_base_install_ubuntu_dkms { # install Dynamic Kernel Module Support packages if needed - local kernel_version=$(uname -r) - local kernel_major_minor=`echo $kernel_version | cut -d. -f1-2` + local kernel_version + kernel_version=$(uname -r) + local kernel_major_minor + kernel_major_minor=`echo $kernel_version | cut -d. -f1-2` # From kernel 3.13 on, openvswitch-datapath-dkms is not needed if [ `vercmp_numbers "$kernel_major_minor" "3.13"` -lt "0" ]; then install_package "dkms openvswitch-datapath-dkms linux-headers-$kernel_version" diff --git a/lib/nova b/lib/nova index 9830276232..6e6075cae6 100644 --- a/lib/nova +++ b/lib/nova @@ -202,14 +202,16 @@ function cleanup_nova { clean_iptables # Destroy old instances - local instances=`sudo virsh list --all | grep $INSTANCE_NAME_PREFIX | sed "s/.*\($INSTANCE_NAME_PREFIX[0-9a-fA-F]*\).*/\1/g"` + local instances + instances=`sudo virsh list --all | grep $INSTANCE_NAME_PREFIX | sed "s/.*\($INSTANCE_NAME_PREFIX[0-9a-fA-F]*\).*/\1/g"` if [ ! "$instances" = "" ]; then echo $instances | xargs -n1 sudo virsh destroy || true echo $instances | xargs -n1 sudo virsh undefine --managed-save || true fi # Logout and delete iscsi sessions - local tgts=$(sudo iscsiadm --mode node | grep $VOLUME_NAME_PREFIX | cut -d ' ' -f2) + local tgts + tgts=$(sudo iscsiadm --mode node | grep $VOLUME_NAME_PREFIX | cut -d ' ' -f2) local target for target in $tgts; do sudo iscsiadm --mode node -T $target --logout || true @@ -245,8 +247,10 @@ function _cleanup_nova_apache_wsgi { function _config_nova_apache_wsgi { sudo mkdir -p $NOVA_WSGI_DIR - local nova_apache_conf=$(apache_site_config_for nova-api) - local nova_ec2_apache_conf=$(apache_site_config_for nova-ec2-api) + local nova_apache_conf + nova_apache_conf=$(apache_site_config_for nova-api) + local nova_ec2_apache_conf + nova_ec2_apache_conf=$(apache_site_config_for nova-ec2-api) local nova_ssl="" local nova_certfile="" local nova_keyfile="" @@ -784,7 +788,8 @@ function start_nova_api { export PATH=$NOVA_BIN_DIR:$PATH # If the site is not enabled then we are in a grenade scenario - local enabled_site_file=$(apache_site_config_for nova-api) + local enabled_site_file + enabled_site_file=$(apache_site_config_for nova-api) if [ -f ${enabled_site_file} ] && [ "$NOVA_USE_MOD_WSGI" == "True" ]; then enable_apache_site nova-api enable_apache_site nova-ec2-api diff --git a/lib/swift b/lib/swift index 645bfd7cd9..62f47dce50 100644 --- a/lib/swift +++ b/lib/swift @@ -205,9 +205,12 @@ function _config_swift_apache_wsgi { # copy apache vhost file and set name and port local node_number for node_number in ${SWIFT_REPLICAS_SEQ}; do - local object_port=$(( OBJECT_PORT_BASE + 10 * (node_number - 1) )) - local container_port=$(( CONTAINER_PORT_BASE + 10 * (node_number - 1) )) - local account_port=$(( ACCOUNT_PORT_BASE + 10 * (node_number - 1) )) + local object_port + object_port=$(( OBJECT_PORT_BASE + 10 * (node_number - 1) )) + local container_port + container_port=$(( CONTAINER_PORT_BASE + 10 * (node_number - 1) )) + local account_port + account_port=$(( ACCOUNT_PORT_BASE + 10 * (node_number - 1) )) sudo cp ${SWIFT_DIR}/examples/apache2/object-server.template $(apache_site_config_for object-server-${node_number}) sudo sed -e " @@ -504,7 +507,8 @@ EOF if is_service_enabled keystone; then iniuncomment ${testfile} func_test auth_version - local auth_vers=$(iniget ${testfile} func_test auth_version) + local auth_vers + auth_vers=$(iniget ${testfile} func_test auth_version) iniset ${testfile} func_test auth_host ${KEYSTONE_SERVICE_HOST} iniset ${testfile} func_test auth_port ${KEYSTONE_AUTH_PORT} if [[ $auth_vers == "3" ]]; then @@ -514,7 +518,8 @@ EOF fi fi - local user_group=$(id -g ${STACK_USER}) + local user_group + user_group=$(id -g ${STACK_USER}) sudo install -d -o ${STACK_USER} -g ${user_group} ${SWIFT_DATA_DIR} local swift_log_dir=${SWIFT_DATA_DIR}/logs @@ -540,7 +545,8 @@ function create_swift_disk { # First do a bit of setup by creating the directories and # changing the permissions so we can run it as our user. - local user_group=$(id -g ${STACK_USER}) + local user_group + user_group=$(id -g ${STACK_USER}) sudo install -d -o ${STACK_USER} -g ${user_group} ${SWIFT_DATA_DIR}/{drives,cache,run,logs} # Create a loopback disk and format it to XFS. @@ -607,7 +613,8 @@ function create_swift_accounts { KEYSTONE_CATALOG_BACKEND=${KEYSTONE_CATALOG_BACKEND:-sql} - local another_role=$(get_or_create_role "anotherrole") + local another_role + another_role=$(get_or_create_role "anotherrole") # NOTE(jroll): Swift doesn't need the admin role here, however Ironic uses # temp urls, which break when uploaded by a non-admin role @@ -623,33 +630,40 @@ function create_swift_accounts { "$SWIFT_SERVICE_PROTOCOL://$SERVICE_HOST:8080/v1/AUTH_\$(tenant_id)s" fi - local swift_tenant_test1=$(get_or_create_project swifttenanttest1 default) + local swift_tenant_test1 + swift_tenant_test1=$(get_or_create_project swifttenanttest1 default) die_if_not_set $LINENO swift_tenant_test1 "Failure creating swift_tenant_test1" SWIFT_USER_TEST1=$(get_or_create_user swiftusertest1 $swiftusertest1_password \ "default" "test@example.com") die_if_not_set $LINENO SWIFT_USER_TEST1 "Failure creating SWIFT_USER_TEST1" get_or_add_user_project_role admin $SWIFT_USER_TEST1 $swift_tenant_test1 - local swift_user_test3=$(get_or_create_user swiftusertest3 $swiftusertest3_password \ + local swift_user_test3 + swift_user_test3=$(get_or_create_user swiftusertest3 $swiftusertest3_password \ "default" "test3@example.com") die_if_not_set $LINENO swift_user_test3 "Failure creating swift_user_test3" get_or_add_user_project_role $another_role $swift_user_test3 $swift_tenant_test1 - local swift_tenant_test2=$(get_or_create_project swifttenanttest2 default) + local swift_tenant_test2 + swift_tenant_test2=$(get_or_create_project swifttenanttest2 default) die_if_not_set $LINENO swift_tenant_test2 "Failure creating swift_tenant_test2" - local swift_user_test2=$(get_or_create_user swiftusertest2 $swiftusertest2_password \ + local swift_user_test2 + swift_user_test2=$(get_or_create_user swiftusertest2 $swiftusertest2_password \ "default" "test2@example.com") die_if_not_set $LINENO swift_user_test2 "Failure creating swift_user_test2" get_or_add_user_project_role admin $swift_user_test2 $swift_tenant_test2 - local swift_domain=$(get_or_create_domain swift_test 'Used for swift functional testing') + local swift_domain + swift_domain=$(get_or_create_domain swift_test 'Used for swift functional testing') die_if_not_set $LINENO swift_domain "Failure creating swift_test domain" - local swift_tenant_test4=$(get_or_create_project swifttenanttest4 $swift_domain) + local swift_tenant_test4 + swift_tenant_test4=$(get_or_create_project swifttenanttest4 $swift_domain) die_if_not_set $LINENO swift_tenant_test4 "Failure creating swift_tenant_test4" - local swift_user_test4=$(get_or_create_user swiftusertest4 $swiftusertest4_password \ + local swift_user_test4 + swift_user_test4=$(get_or_create_user swiftusertest4 $swiftusertest4_password \ $swift_domain "test4@example.com") die_if_not_set $LINENO swift_user_test4 "Failure creating swift_user_test4" get_or_add_user_project_role admin $swift_user_test4 $swift_tenant_test4 diff --git a/lib/tempest b/lib/tempest index f4d0a6dab0..e7ea429847 100644 --- a/lib/tempest +++ b/lib/tempest @@ -372,7 +372,8 @@ function configure_tempest { # Compute Features # Run ``verify_tempest_config -ur`` to retrieve enabled extensions on API endpoints # NOTE(mtreinish): This must be done after auth settings are added to the tempest config - local tmp_cfg_file=$(mktemp) + local tmp_cfg_file + tmp_cfg_file=$(mktemp) cd $TEMPEST_DIR tox -revenv -- verify-tempest-config -uro $tmp_cfg_file diff --git a/lib/tls b/lib/tls index 8ff2027819..f4740b88be 100644 --- a/lib/tls +++ b/lib/tls @@ -346,7 +346,8 @@ function make_root_CA { # we need to change it. function fix_system_ca_bundle_path { if is_service_enabled tls-proxy || [ "$USE_SSL" == "True" ]; then - local capath=$(python -c $'try:\n from requests import certs\n print certs.where()\nexcept ImportError: pass') + local capath + capath=$(python -c $'try:\n from requests import certs\n print certs.where()\nexcept ImportError: pass') if [[ ! $capath == "" && ! $capath =~ ^/etc/.* && ! -L $capath ]]; then if is_fedora; then diff --git a/tests/unittest.sh b/tests/unittest.sh index 603652a216..df7a8b4534 100644 --- a/tests/unittest.sh +++ b/tests/unittest.sh @@ -20,8 +20,10 @@ FAILED_FUNCS="" # pass a test, printing out MSG # usage: passed message function passed { - local lineno=$(caller 0 | awk '{print $1}') - local function=$(caller 0 | awk '{print $2}') + local lineno + lineno=$(caller 0 | awk '{print $1}') + local function + function=$(caller 0 | awk '{print $2}') local msg="$1" if [ -z "$msg" ]; then msg="OK" @@ -33,8 +35,10 @@ function passed { # fail a test, printing out MSG # usage: failed message function failed { - local lineno=$(caller 0 | awk '{print $1}') - local function=$(caller 0 | awk '{print $2}') + local lineno + lineno=$(caller 0 | awk '{print $1}') + local function + function=$(caller 0 | awk '{print $2}') local msg="$1" FAILED_FUNCS+="$function:L$lineno\n" echo "ERROR: $function:L$lineno!" @@ -45,8 +49,10 @@ function failed { # assert string comparision of val1 equal val2, printing out msg # usage: assert_equal val1 val2 msg function assert_equal { - local lineno=`caller 0 | awk '{print $1}'` - local function=`caller 0 | awk '{print $2}'` + local lineno + lineno=`caller 0 | awk '{print $1}'` + local function + function=`caller 0 | awk '{print $2}'` local msg=$3 if [ -z "$msg" ]; then @@ -66,8 +72,10 @@ function assert_equal { # assert variable is empty/blank, printing out msg # usage: assert_empty VAR msg function assert_empty { - local lineno=`caller 0 | awk '{print $1}'` - local function=`caller 0 | awk '{print $2}'` + local lineno + lineno=`caller 0 | awk '{print $1}'` + local function + function=`caller 0 | awk '{print $2}'` local msg=$2 if [ -z "$msg" ]; then diff --git a/tools/create_userrc.sh b/tools/create_userrc.sh index de44abbbe5..25f713ca93 100755 --- a/tools/create_userrc.sh +++ b/tools/create_userrc.sh @@ -190,7 +190,8 @@ function add_entry { local user_passwd=$5 # The admin user can see all user's secret AWS keys, it does not looks good - local line=`openstack ec2 credentials list --user $user_id | grep " $project_id "` + local line + line=$(openstack ec2 credentials list --user $user_id | grep " $project_id " || true) if [ -z "$line" ]; then openstack ec2 credentials create --user $user_id --project $project_id 1>&2 line=`openstack ec2 credentials list --user $user_id | grep " $project_id "` diff --git a/tools/peakmem_tracker.sh b/tools/peakmem_tracker.sh index 0d5728a538..ecbd79a0bc 100755 --- a/tools/peakmem_tracker.sh +++ b/tools/peakmem_tracker.sh @@ -41,10 +41,12 @@ function get_mem_available { # snapshot of current usage; i.e. checking the latest entry in the # file will give the peak-memory usage function tracker { - local low_point=$(get_mem_available) + local low_point + low_point=$(get_mem_available) while [ 1 ]; do - local mem_available=$(get_mem_available) + local mem_available + mem_available=$(get_mem_available) if [[ $mem_available -lt $low_point ]]; then low_point=$mem_available diff --git a/tools/xen/scripts/install-os-vpx.sh b/tools/xen/scripts/install-os-vpx.sh index 1ebbeaf564..66f7ef4763 100755 --- a/tools/xen/scripts/install-os-vpx.sh +++ b/tools/xen/scripts/install-os-vpx.sh @@ -100,7 +100,8 @@ create_vif() { local v="$1" echo "Installing VM interface on [$BRIDGE]" - local out_network_uuid=$(find_network "$BRIDGE") + local out_network_uuid + out_network_uuid=$(find_network "$BRIDGE") xe vif-create vm-uuid="$v" network-uuid="$out_network_uuid" device="0" } diff --git a/tools/xen/scripts/uninstall-os-vpx.sh b/tools/xen/scripts/uninstall-os-vpx.sh index 1ed249433a..96dad7e852 100755 --- a/tools/xen/scripts/uninstall-os-vpx.sh +++ b/tools/xen/scripts/uninstall-os-vpx.sh @@ -35,9 +35,12 @@ xe_min() destroy_vdi() { local vbd_uuid="$1" - local type=$(xe_min vbd-list uuid=$vbd_uuid params=type) - local dev=$(xe_min vbd-list uuid=$vbd_uuid params=userdevice) - local vdi_uuid=$(xe_min vbd-list uuid=$vbd_uuid params=vdi-uuid) + local type + type=$(xe_min vbd-list uuid=$vbd_uuid params=type) + local dev + dev=$(xe_min vbd-list uuid=$vbd_uuid params=userdevice) + local vdi_uuid + vdi_uuid=$(xe_min vbd-list uuid=$vbd_uuid params=vdi-uuid) if [ "$type" == 'Disk' ] && [ "$dev" != 'xvda' ] && [ "$dev" != '0' ]; then xe vdi-destroy uuid=$vdi_uuid @@ -47,7 +50,8 @@ destroy_vdi() uninstall() { local vm_uuid="$1" - local power_state=$(xe_min vm-list uuid=$vm_uuid params=power-state) + local power_state + power_state=$(xe_min vm-list uuid=$vm_uuid params=power-state) if [ "$power_state" != "halted" ]; then xe vm-shutdown vm=$vm_uuid force=true diff --git a/tools/xen/test_functions.sh b/tools/xen/test_functions.sh index 924e773c00..324e6a1a1e 100755 --- a/tools/xen/test_functions.sh +++ b/tools/xen/test_functions.sh @@ -165,7 +165,8 @@ EOF function test_get_local_sr { setup_xe_response "uuid123" - local RESULT=$(. mocks && get_local_sr) + local RESULT + RESULT=$(. mocks && get_local_sr) [ "$RESULT" == "uuid123" ] @@ -173,7 +174,8 @@ function test_get_local_sr { } function test_get_local_sr_path { - local RESULT=$(mock_out get_local_sr "uuid1" && get_local_sr_path) + local RESULT + RESULT=$(mock_out get_local_sr "uuid1" && get_local_sr_path) [ "/var/run/sr-mount/uuid1" == "$RESULT" ] }