X-Git-Url: https://git.arvados.org/arvados.git/blobdiff_plain/f43a1f883619d2ec25439145b9b35fec1265b0cc..afbb3f058e8fcf056d6942b59ffeb4d2494b4fd7:/tools/salt-install/provision.sh diff --git a/tools/salt-install/provision.sh b/tools/salt-install/provision.sh index 1ee13752b0..5cd376844d 100755 --- a/tools/salt-install/provision.sh +++ b/tools/salt-install/provision.sh @@ -10,68 +10,84 @@ # # vagrant up -########################################################## -# The 5 letters name you want to give your cluster -CLUSTER="arva2" -DOMAIN="arv.local" - -# The example config you want to use. Currently, only "single_host" is -# available -CONFIG_DIR="single_host" - -# Which release of Arvados repo you want to use -RELEASE="production" -# Which version of Arvados you want to install. Defaults to 'latest' -# in the desired repo -VERSION="latest" - -# Host SSL port where you want to point your browser to access Arvados -# Defaults to 443 for regular runs, and to 8443 when called in Vagrant. -# You can point it to another port if desired -# In Vagrant, make sure it matches what you set in the Vagrantfile -# HOST_SSL_PORT=443 - -# This is a arvados-formula setting. -# If branch is set, the script will switch to it before running salt -# Usually not needed, only used for testing -# BRANCH="master" - -########################################################## -# Usually there's no need to modify things below this line - set -o pipefail +# capture the directory that the script is running from +SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" >/dev/null 2>&1 && pwd )" + usage() { echo >&2 - echo >&2 "Usage: $0 [-h] [-h]" + echo >&2 "Usage: ${0} [-h] [-h]" echo >&2 - echo >&2 "$0 options:" - echo >&2 " -v, --vagrant Run in vagrant and use the /vagrant shared dir" - echo >&2 " -p , --ssl-port SSL port to use for the web applications" - echo >&2 " -h, --help Display this help and exit" + echo >&2 "${0} options:" + echo >&2 " -d, --debug Run salt installation in debug mode" + echo >&2 " -p , --ssl-port SSL port to use for the web applications" + echo >&2 " -c , --config Path to the local.params config file" + echo >&2 " -t, --test Test installation running a CWL workflow" + echo >&2 " -r, --roles List of Arvados roles to apply to the host, comma separated" + echo >&2 " Possible values are:" + echo >&2 " api" + echo >&2 " controller" + echo >&2 " dispatcher" + echo >&2 " keepproxy" + echo >&2 " keepstore" + echo >&2 " keepweb" + echo >&2 " shell" + echo >&2 " webshell" + echo >&2 " websocket" + echo >&2 " workbench" + echo >&2 " workbench2" + echo >&2 " Defaults to applying them all" + echo >&2 " -h, --help Display this help and exit" + echo >&2 " -v, --vagrant Run in vagrant and use the /vagrant shared dir" echo >&2 } arguments() { # NOTE: This requires GNU getopt (part of the util-linux package on Debian-based distros). - TEMP=`getopt -o hvp: \ - --long help,vagrant,ssl-port: \ - -n "$0" -- "$@"` + TEMP=$(getopt -o c:dhp:r:tv \ + --long config:,debug,help,ssl-port:,roles:,test,vagrant \ + -n "${0}" -- "${@}") - if [ $? != 0 ] ; then echo "GNU getopt missing? Use -h for help"; exit 1 ; fi + if [ ${?} != 0 ] ; then echo "GNU getopt missing? Use -h for help"; exit 1 ; fi # Note the quotes around `$TEMP': they are essential! eval set -- "$TEMP" - while [ $# -ge 1 ]; do - case $1 in - -v | --vagrant) - VAGRANT="yes" + while [ ${#} -ge 1 ]; do + case ${1} in + -c | --config) + CONFIG_FILE=${2} + shift 2 + ;; + -d | --debug) + LOG_LEVEL="debug" shift ;; -p | --ssl-port) - HOST_SSL_PORT=${2} + CONTROLLER_EXT_SSL_PORT=${2} shift 2 ;; + -r | --roles) + for i in ${2//,/ } + do + # Verify the role exists + if [[ ! "database,api,controller,keepstore,websocket,keepweb,workbench2,webshell,keepproxy,shell,workbench,dispatcher" == *"$i"* ]]; then + echo "The role '${i}' is not a valid role" + usage + exit 1 + fi + ROLES="${ROLES} ${i}" + done + shift 2 + ;; + -t | --test) + TEST="yes" + shift + ;; + -v | --vagrant) + VAGRANT="yes" + shift + ;; --) shift break @@ -84,9 +100,49 @@ arguments() { done } -HOST_SSL_PORT=443 +CONFIG_FILE="${SCRIPT_DIR}/local.params" +CONFIG_DIR="local_config_dir" +LOG_LEVEL="info" +CONTROLLER_EXT_SSL_PORT=443 +TESTS_DIR="tests" + +CLUSTER="" +DOMAIN="" + +# Hostnames/IPs used for single-host deploys +HOSTNAME_EXT="" +HOSTNAME_INT="127.0.1.1" + +# Initial user setup +INITIAL_USER="" +INITIAL_USER_EMAIL="" +INITIAL_USER_PASSWORD="" + +CONTROLLER_EXT_SSL_PORT=8000 +KEEP_EXT_SSL_PORT=25101 +# Both for collections and downloads +KEEPWEB_EXT_SSL_PORT=9002 +WEBSHELL_EXT_SSL_PORT=4202 +WEBSOCKET_EXT_SSL_PORT=8002 +WORKBENCH1_EXT_SSL_PORT=443 +WORKBENCH2_EXT_SSL_PORT=3001 + +# For a stable release, change RELEASE "production" and VERSION to the +# package version (including the iteration, e.g. X.Y.Z-1) of the +# release. +RELEASE="development" +VERSION="latest" + +# The arvados-formula version. For a stable release, this should be a +# branch name (e.g. X.Y-dev) or tag for the release. +ARVADOS_TAG="main" -arguments $@ +# Other formula versions we depend on +POSTGRES_TAG="v0.41.6" +NGINX_TAG="v2.7.4" +DOCKER_TAG="v1.0.0" +LOCALE_TAG="v0.3.4" +LETSENCRYPT_TAG="v2.1.0" # Salt's dir ## states @@ -96,15 +152,50 @@ F_DIR="/srv/formulas" ##pillars P_DIR="/srv/pillars" +arguments ${@} + +if [ -s ${CONFIG_FILE} ]; then + source ${CONFIG_FILE} +else + echo >&2 "Please create a '${CONFIG_FILE}' file with initial values, as described in" + echo >&2 " * https://doc.arvados.org/install/salt-single-host.html#single_host, or" + echo >&2 " * https://doc.arvados.org/install/salt-multi-host.html#multi_host_multi_hostnames" + exit 1 +fi + +if [ ! -d ${CONFIG_DIR} ]; then + echo >&2 "Please create a '${CONFIG_DIR}' with initial values, as described in" + echo >&2 " * https://doc.arvados.org/install/salt-single-host.html#single_host, or" + echo >&2 " * https://doc.arvados.org/install/salt-multi-host.html#multi_host_multi_hostnames" + exit 1 +fi + +if grep -q 'fixme_or_this_wont_work' ${CONFIG_FILE} ; then + echo >&2 "The config file ${CONFIG_FILE} has some parameters that need to be modified." + echo >&2 "Please, fix them and re-run the provision script." + exit 1 +fi + +if ! grep -E '^[[:alnum:]]{5}$' <<<${CLUSTER} ; then + echo >&2 "ERROR: must be exactly 5 alphanumeric characters long" + echo >&2 "Fix the cluster name in the 'local.params' file and re-run the provision script" + exit 1 +fi + +# Only used in single_host/single_name deploys +if [ "x${HOSTNAME_EXT}" = "x" ] ; then + HOSTNAME_EXT="${CLUSTER}.${DOMAIN}" +fi + apt-get update -apt-get install -y curl git +apt-get install -y curl git jq -dpkg -l |grep salt-minion -if [ ${?} -eq 0 ]; then +if which salt-call; then echo "Salt already installed" else curl -L https://bootstrap.saltstack.com -o /tmp/bootstrap_salt.sh - sh /tmp/bootstrap_salt.sh -XUdfP -x python3 + sh /tmp/bootstrap_salt.sh -XdfP -x python3 + /bin/systemctl stop salt-minion.service /bin/systemctl disable salt-minion.service fi @@ -115,85 +206,298 @@ file_roots: base: - ${S_DIR} - ${F_DIR}/* - - ${F_DIR}/*/test/salt/states pillar_roots: base: - ${P_DIR} EOFSM -mkdir -p ${S_DIR} -mkdir -p ${F_DIR} -mkdir -p ${P_DIR} +mkdir -p ${S_DIR} ${F_DIR} ${P_DIR} + +# Get the formula and dependencies +cd ${F_DIR} || exit 1 +git clone --branch "${ARVADOS_TAG}" https://git.arvados.org/arvados-formula.git +git clone --branch "${DOCKER_TAG}" https://github.com/saltstack-formulas/docker-formula.git +git clone --branch "${LOCALE_TAG}" https://github.com/saltstack-formulas/locale-formula.git +git clone --branch "${NGINX_TAG}" https://github.com/saltstack-formulas/nginx-formula.git +git clone --branch "${POSTGRES_TAG}" https://github.com/saltstack-formulas/postgres-formula.git +git clone --branch "${LETSENCRYPT_TAG}" https://github.com/saltstack-formulas/letsencrypt-formula.git + +# If we want to try a specific branch of the formula +if [ "x${BRANCH}" != "x" ]; then + cd ${F_DIR}/arvados-formula || exit 1 + git checkout -t origin/"${BRANCH}" -b "${BRANCH}" + cd - +fi + +if [ "x${VAGRANT}" = "xyes" ]; then + EXTRA_STATES_DIR="/home/vagrant/${CONFIG_DIR}/states" + SOURCE_PILLARS_DIR="/home/vagrant/${CONFIG_DIR}/pillars" + SOURCE_TESTS_DIR="/home/vagrant/${TESTS_DIR}" +else + EXTRA_STATES_DIR="${SCRIPT_DIR}/${CONFIG_DIR}/states" + SOURCE_PILLARS_DIR="${SCRIPT_DIR}/${CONFIG_DIR}/pillars" + SOURCE_TESTS_DIR="${SCRIPT_DIR}/${TESTS_DIR}" +fi + +SOURCE_STATES_DIR="${EXTRA_STATES_DIR}" + +# Replace variables (cluster, domain, etc) in the pillars, states and tests +# to ease deployment for newcomers +if [ ! -d "${SOURCE_PILLARS_DIR}" ]; then + echo "${SOURCE_PILLARS_DIR} does not exist or is not a directory. Exiting." + exit 1 +fi +for f in $(ls "${SOURCE_PILLARS_DIR}"/*); do + sed "s#__ANONYMOUS_USER_TOKEN__#${ANONYMOUS_USER_TOKEN}#g; + s#__BLOB_SIGNING_KEY__#${BLOB_SIGNING_KEY}#g; + s#__CONTROLLER_EXT_SSL_PORT__#${CONTROLLER_EXT_SSL_PORT}#g; + s#__CLUSTER__#${CLUSTER}#g; + s#__DOMAIN__#${DOMAIN}#g; + s#__HOSTNAME_EXT__#${HOSTNAME_EXT}#g; + s#__HOSTNAME_INT__#${HOSTNAME_INT}#g; + s#__INITIAL_USER_EMAIL__#${INITIAL_USER_EMAIL}#g; + s#__INITIAL_USER_PASSWORD__#${INITIAL_USER_PASSWORD}#g; + s#__INITIAL_USER__#${INITIAL_USER}#g; + s#__LE_AWS_REGION__#${LE_AWS_REGION}#g; + s#__LE_AWS_SECRET_ACCESS_KEY__#${LE_AWS_SECRET_ACCESS_KEY}#g; + s#__LE_AWS_ACCESS_KEY_ID__#${LE_AWS_ACCESS_KEY_ID}#g; + s#__DATABASE_PASSWORD__#${DATABASE_PASSWORD}#g; + s#__KEEPWEB_EXT_SSL_PORT__#${KEEPWEB_EXT_SSL_PORT}#g; + s#__KEEP_EXT_SSL_PORT__#${KEEP_EXT_SSL_PORT}#g; + s#__MANAGEMENT_TOKEN__#${MANAGEMENT_TOKEN}#g; + s#__RELEASE__#${RELEASE}#g; + s#__SYSTEM_ROOT_TOKEN__#${SYSTEM_ROOT_TOKEN}#g; + s#__VERSION__#${VERSION}#g; + s#__WEBSHELL_EXT_SSL_PORT__#${WEBSHELL_EXT_SSL_PORT}#g; + s#__WEBSOCKET_EXT_SSL_PORT__#${WEBSOCKET_EXT_SSL_PORT}#g; + s#__WORKBENCH1_EXT_SSL_PORT__#${WORKBENCH1_EXT_SSL_PORT}#g; + s#__WORKBENCH2_EXT_SSL_PORT__#${WORKBENCH2_EXT_SSL_PORT}#g; + s#__CLUSTER_INT_CIDR__#${CLUSTER_INT_CIDR}#g; + s#__CONTROLLER_INT_IP__#${CONTROLLER_INT_IP}#g; + s#__WEBSOCKET_INT_IP__#${WEBSOCKET_INT_IP}#g; + s#__KEEP_INT_IP__#${KEEP_INT_IP}#g; + s#__KEEPSTORE0_INT_IP__#${KEEPSTORE0_INT_IP}#g; + s#__KEEPSTORE1_INT_IP__#${KEEPSTORE1_INT_IP}#g; + s#__KEEPWEB_INT_IP__#${KEEPWEB_INT_IP}#g; + s#__WEBSHELL_INT_IP__#${WEBSHELL_INT_IP}#g; + s#__SHELL_INT_IP__#${SHELL_INT_IP}#g; + s#__WORKBENCH1_INT_IP__#${WORKBENCH1_INT_IP}#g; + s#__WORKBENCH2_INT_IP__#${WORKBENCH2_INT_IP}#g; + s#__DATABASE_INT_IP__#${DATABASE_INT_IP}#g; + s#__WORKBENCH_SECRET_KEY__#${WORKBENCH_SECRET_KEY}#g" \ + "${f}" > "${P_DIR}"/$(basename "${f}") +done + +if [ "x${TEST}" = "xyes" ] && [ ! -d "${SOURCE_TESTS_DIR}" ]; then + echo "You requested to run tests, but ${SOURCE_TESTS_DIR} does not exist or is not a directory. Exiting." + exit 1 +fi +mkdir -p /tmp/cluster_tests +# Replace cluster and domain name in the test files +for f in $(ls "${SOURCE_TESTS_DIR}"/*); do + sed "s#__CLUSTER__#${CLUSTER}#g; + s#__CONTROLLER_EXT_SSL_PORT__#${CONTROLLER_EXT_SSL_PORT}#g; + s#__DOMAIN__#${DOMAIN}#g; + s#__HOSTNAME_INT__#${HOSTNAME_INT}#g; + s#__INITIAL_USER_EMAIL__#${INITIAL_USER_EMAIL}#g; + s#__INITIAL_USER_PASSWORD__#${INITIAL_USER_PASSWORD}#g + s#__INITIAL_USER__#${INITIAL_USER}#g; + s#__DATABASE_PASSWORD__#${DATABASE_PASSWORD}#g; + s#__SYSTEM_ROOT_TOKEN__#${SYSTEM_ROOT_TOKEN}#g" \ + "${f}" > "/tmp/cluster_tests"/$(basename "${f}") +done +chmod 755 /tmp/cluster_tests/run-test.sh + +# Replace helper state files that differ from the formula's examples +if [ -d "${SOURCE_STATES_DIR}" ]; then + mkdir -p "${F_DIR}"/extra/extra + + for f in $(ls "${SOURCE_STATES_DIR}"/*); do + sed "s#__ANONYMOUS_USER_TOKEN__#${ANONYMOUS_USER_TOKEN}#g; + s#__CLUSTER__#${CLUSTER}#g; + s#__BLOB_SIGNING_KEY__#${BLOB_SIGNING_KEY}#g; + s#__CONTROLLER_EXT_SSL_PORT__#${CONTROLLER_EXT_SSL_PORT}#g; + s#__DOMAIN__#${DOMAIN}#g; + s#__HOSTNAME_EXT__#${HOSTNAME_EXT}#g; + s#__HOSTNAME_INT__#${HOSTNAME_INT}#g; + s#__INITIAL_USER_EMAIL__#${INITIAL_USER_EMAIL}#g; + s#__INITIAL_USER_PASSWORD__#${INITIAL_USER_PASSWORD}#g; + s#__INITIAL_USER__#${INITIAL_USER}#g; + s#__DATABASE_PASSWORD__#${DATABASE_PASSWORD}#g; + s#__KEEPWEB_EXT_SSL_PORT__#${KEEPWEB_EXT_SSL_PORT}#g; + s#__KEEP_EXT_SSL_PORT__#${KEEP_EXT_SSL_PORT}#g; + s#__MANAGEMENT_TOKEN__#${MANAGEMENT_TOKEN}#g; + s#__RELEASE__#${RELEASE}#g; + s#__SYSTEM_ROOT_TOKEN__#${SYSTEM_ROOT_TOKEN}#g; + s#__VERSION__#${VERSION}#g; + s#__CLUSTER_INT_CIDR__#${CLUSTER_INT_CIDR}#g; + s#__CONTROLLER_INT_IP__#${CONTROLLER_INT_IP}#g; + s#__WEBSOCKET_INT_IP__#${WEBSOCKET_INT_IP}#g; + s#__KEEP_INT_IP__#${KEEP_INT_IP}#g; + s#__KEEPSTORE0_INT_IP__#${KEEPSTORE0_INT_IP}#g; + s#__KEEPSTORE1_INT_IP__#${KEEPSTORE1_INT_IP}#g; + s#__KEEPWEB_INT_IP__#${KEEPWEB_INT_IP}#g; + s#__WEBSHELL_INT_IP__#${WEBSHELL_INT_IP}#g; + s#__WORKBENCH1_INT_IP__#${WORKBENCH1_INT_IP}#g; + s#__WORKBENCH2_INT_IP__#${WORKBENCH2_INT_IP}#g; + s#__DATABASE_INT_IP__#${DATABASE_INT_IP}#g; + s#__WEBSHELL_EXT_SSL_PORT__#${WEBSHELL_EXT_SSL_PORT}#g; + s#__WEBSOCKET_EXT_SSL_PORT__#${WEBSOCKET_EXT_SSL_PORT}#g; + s#__WORKBENCH1_EXT_SSL_PORT__#${WORKBENCH1_EXT_SSL_PORT}#g; + s#__WORKBENCH2_EXT_SSL_PORT__#${WORKBENCH2_EXT_SSL_PORT}#g; + s#__WORKBENCH_SECRET_KEY__#${WORKBENCH_SECRET_KEY}#g" \ + "${f}" > "${F_DIR}/extra/extra"/$(basename "${f}") + done +fi + +# Now, we build the SALT states/pillars trees +# As we need to separate both states and pillars in case we want specific +# roles, we iterate on both at the same time # States cat > ${S_DIR}/top.sls << EOFTSLS base: '*': - - example_add_snakeoil_certs - locale - - nginx.passenger - - postgres - - docker - - arvados EOFTSLS # Pillars cat > ${P_DIR}/top.sls << EOFPSLS base: '*': - - arvados - locale - - nginx_api_configuration - - nginx_controller_configuration - - nginx_keepproxy_configuration - - nginx_keepweb_configuration - - nginx_passenger - - nginx_websocket_configuration - - nginx_webshell_configuration - - nginx_workbench2_configuration - - nginx_workbench_configuration - - postgresql + - arvados EOFPSLS - -# Get the formula and dependencies -cd ${F_DIR} || exit 1 -for f in postgres arvados nginx docker locale; do - # git clone https://github.com/saltstack-formulas/${f}-formula.git - git clone https://github.com/netmanagers/${f}-formula.git -done - -if [ "x${BRANCH}" != "x" ]; then - cd ${F_DIR}/arvados-formula - git checkout -t origin/${BRANCH} - cd - +# States, extra states +if [ -d "${F_DIR}"/extra/extra ]; then + for f in $(ls "${F_DIR}"/extra/extra/*.sls); do + echo " - extra.$(basename ${f} | sed 's/.sls$//g')" >> ${S_DIR}/top.sls + done fi -# sed "s/__DOMAIN__/${DOMAIN}/g; s/__CLUSTER__/${CLUSTER}/g; s/__RELEASE__/${RELEASE}/g; s/__VERSION__/${VERSION}/g" \ -# ${CONFIG_DIR}/arvados_dev.sls > ${P_DIR}/arvados.sls - -if [ "x${VAGRANT}" = "xyes" ]; then - SOURCE_PILLARS_DIR="/vagrant/${CONFIG_DIR}" +# If we want specific roles for a node, just add the desired states +# and its dependencies +if [ -z "${ROLES}" ]; then + # States + echo " - nginx.passenger" >> ${S_DIR}/top.sls + # Currently, only available on config_examples/multi_host/aws + if [ "x${USE_LETSENCRYPT}" = "xyes" ]; then + if [ "x${USE_LETSENCRYPT_IAM_USER}" = "xyes" ]; then + grep -q "aws_credentials" ${S_DIR}/top.sls || echo " - aws_credentials" >> ${S_DIR}/top.sls + fi + grep -q "letsencrypt" ${S_DIR}/top.sls || echo " - letsencrypt" >> ${S_DIR}/top.sls + fi + echo " - postgres" >> ${S_DIR}/top.sls + echo " - docker.software" >> ${S_DIR}/top.sls + echo " - arvados" >> ${S_DIR}/top.sls + + # Pillars + echo " - docker" >> ${P_DIR}/top.sls + echo " - nginx_api_configuration" >> ${P_DIR}/top.sls + echo " - nginx_controller_configuration" >> ${P_DIR}/top.sls + echo " - nginx_keepproxy_configuration" >> ${P_DIR}/top.sls + echo " - nginx_keepweb_configuration" >> ${P_DIR}/top.sls + echo " - nginx_passenger" >> ${P_DIR}/top.sls + echo " - nginx_websocket_configuration" >> ${P_DIR}/top.sls + echo " - nginx_webshell_configuration" >> ${P_DIR}/top.sls + echo " - nginx_workbench2_configuration" >> ${P_DIR}/top.sls + echo " - nginx_workbench_configuration" >> ${P_DIR}/top.sls + echo " - postgresql" >> ${P_DIR}/top.sls + # Currently, only available on config_examples/multi_host/aws + if [ "x${USE_LETSENCRYPT}" = "xyes" ]; then + if [ "x${USE_LETSENCRYPT_IAM_USER}" = "xyes" ]; then + grep -q "aws_credentials" ${P_DIR}/top.sls || echo " - aws_credentials" >> ${P_DIR}/top.sls + fi + grep -q "letsencrypt" ${P_DIR}/top.sls || echo " - letsencrypt" >> ${P_DIR}/top.sls + fi else - SOURCE_PILLARS_DIR="./${CONFIG_DIR}" -fi - -# Replace cluster and domain name in the example pillars -for f in ${SOURCE_PILLARS_DIR}/*; do - # sed "s/example.net/${DOMAIN}/g; s/fixme/${CLUSTER}/g" \ - sed "s/__DOMAIN__/${DOMAIN}/g; - s/__CLUSTER__/${CLUSTER}/g; - s/__RELEASE__/${RELEASE}/g; - s/__HOST_SSL_PORT__/${HOST_SSL_PORT}/g; - s/__GUEST_SSL_PORT__/${GUEST_SSL_PORT}/g; - s/__VERSION__/${VERSION}/g" \ - ${f} > ${P_DIR}/$(basename ${f}) -done - -# Let's write an /etc/hosts file that points all the hosts to localhost - -echo "127.0.0.2 api keep keep0 collections download ws workbench workbench2 ${CLUSTER}.${DOMAIN} api.${CLUSTER}.${DOMAIN} keep.${CLUSTER}.${DOMAIN} keep0.${CLUSTER}.${DOMAIN} collections.${CLUSTER}.${DOMAIN} download.${CLUSTER}.${DOMAIN} ws.${CLUSTER}.${DOMAIN} workbench.${CLUSTER}.${DOMAIN} workbench2.${CLUSTER}.${DOMAIN}" >> /etc/hosts + # If we add individual roles, make sure we add the repo first + echo " - arvados.repo" >> ${S_DIR}/top.sls + for R in ${ROLES}; do + case "${R}" in + "database") + # States + echo " - postgres" >> ${S_DIR}/top.sls + # Pillars + echo ' - postgresql' >> ${P_DIR}/top.sls + ;; + "api") + # States + # FIXME: https://dev.arvados.org/issues/17352 + grep -q "postgres.client" ${S_DIR}/top.sls || echo " - postgres.client" >> ${S_DIR}/top.sls + grep -q "nginx.passenger" ${S_DIR}/top.sls || echo " - nginx.passenger" >> ${S_DIR}/top.sls + ### If we don't install and run LE before arvados-api-server, it fails and breaks everything + ### after it so we add this here, as we are, after all, sharing the host for api and controller + # Currently, only available on config_examples/multi_host/aws + if [ "x${USE_LETSENCRYPT}" = "xyes" ]; then + if [ "x${USE_LETSENCRYPT_IAM_USER}" = "xyes" ]; then + grep -q "aws_credentials" ${S_DIR}/top.sls || echo " - aws_credentials" >> ${S_DIR}/top.sls + fi + grep -q "letsencrypt" ${S_DIR}/top.sls || echo " - letsencrypt" >> ${S_DIR}/top.sls + fi + grep -q "arvados.${R}" ${S_DIR}/top.sls || echo " - arvados.${R}" >> ${S_DIR}/top.sls + # Pillars + grep -q "aws_credentials" ${P_DIR}/top.sls || echo " - aws_credentials" >> ${P_DIR}/top.sls + grep -q "docker" ${P_DIR}/top.sls || echo " - docker" >> ${P_DIR}/top.sls + grep -q "postgresql" ${P_DIR}/top.sls || echo " - postgresql" >> ${P_DIR}/top.sls + grep -q "nginx_passenger" ${P_DIR}/top.sls || echo " - nginx_passenger" >> ${P_DIR}/top.sls + grep -q "nginx_${R}_configuration" ${P_DIR}/top.sls || echo " - nginx_${R}_configuration" >> ${P_DIR}/top.sls + ;; + "controller" | "websocket" | "workbench" | "workbench2" | "webshell" | "keepweb" | "keepproxy") + # States + grep -q "nginx.passenger" ${S_DIR}/top.sls || echo " - nginx.passenger" >> ${S_DIR}/top.sls + # Currently, only available on config_examples/multi_host/aws + if [ "x${USE_LETSENCRYPT}" = "xyes" ]; then + if [ "x${USE_LETSENCRYPT_IAM_USER}" = "xyes" ]; then + grep -q "aws_credentials" ${S_DIR}/top.sls || echo " - aws_credentials" >> ${S_DIR}/top.sls + fi + grep -q "letsencrypt" ${S_DIR}/top.sls || echo " - letsencrypt" >> ${S_DIR}/top.sls + fi + # webshell role is just a nginx vhost, so it has no state + if [ "${R}" != "webshell" ]; then + grep -q "arvados.${R}" ${S_DIR}/top.sls || echo " - arvados.${R}" >> ${S_DIR}/top.sls + fi + # Pillars + grep -q "nginx_passenger" ${P_DIR}/top.sls || echo " - nginx_passenger" >> ${P_DIR}/top.sls + grep -q "nginx_${R}_configuration" ${P_DIR}/top.sls || echo " - nginx_${R}_configuration" >> ${P_DIR}/top.sls + # Currently, only available on config_examples/multi_host/aws + if [ "x${USE_LETSENCRYPT}" = "xyes" ]; then + if [ "x${USE_LETSENCRYPT_IAM_USER}" = "xyes" ]; then + grep -q "aws_credentials" ${P_DIR}/top.sls || echo " - aws_credentials" >> ${P_DIR}/top.sls + fi + grep -q "letsencrypt" ${P_DIR}/top.sls || echo " - letsencrypt" >> ${P_DIR}/top.sls + grep -q "letsencrypt_${R}_configuration" ${P_DIR}/top.sls || echo " - letsencrypt_${R}_configuration" >> ${P_DIR}/top.sls + fi + ;; + "shell") + # States + grep -q "docker" ${S_DIR}/top.sls || echo " - docker.software" >> ${S_DIR}/top.sls + grep -q "arvados.${R}" ${S_DIR}/top.sls || echo " - arvados.${R}" >> ${S_DIR}/top.sls + # Pillars + grep -q "" ${P_DIR}/top.sls || echo " - docker" >> ${P_DIR}/top.sls + ;; + "dispatcher") + # States + grep -q "docker" ${S_DIR}/top.sls || echo " - docker.software" >> ${S_DIR}/top.sls + grep -q "arvados.${R}" ${S_DIR}/top.sls || echo " - arvados.${R}" >> ${S_DIR}/top.sls + # Pillars + # ATM, no specific pillar needed + ;; + "keepstore") + # States + grep -q "arvados.${R}" ${S_DIR}/top.sls || echo " - arvados.${R}" >> ${S_DIR}/top.sls + # Pillars + # ATM, no specific pillar needed + ;; + *) + echo "Unknown role ${R}" + exit 1 + ;; + esac + done +fi # FIXME! #16992 Temporary fix for psql call in arvados-api-server if [ -e /root/.psqlrc ]; then @@ -209,7 +513,7 @@ echo '\pset pager off' >> /root/.psqlrc # END FIXME! #16992 Temporary fix for psql call in arvados-api-server # Now run the install -salt-call --local state.apply -l debug +salt-call --local state.apply -l ${LOG_LEVEL} # FIXME! #16992 Temporary fix for psql call in arvados-api-server if [ "x${DELETE_PSQL}" = "xyes" ]; then @@ -218,7 +522,25 @@ if [ "x${DELETE_PSQL}" = "xyes" ]; then fi if [ "x${RESTORE_PSQL}" = "xyes" ]; then - echo "Restroting .psql file" + echo "Restoring .psql file" mv -v /root/.psqlrc.provision.backup /root/.psqlrc fi # END FIXME! #16992 Temporary fix for psql call in arvados-api-server + +# Leave a copy of the Arvados CA so the user can copy it where it's required +echo "Copying the Arvados CA certificate to the installer dir, so you can import it" +# If running in a vagrant VM, also add default user to docker group +if [ "x${VAGRANT}" = "xyes" ]; then + cp /etc/ssl/certs/arvados-snakeoil-ca.pem /vagrant/${CLUSTER}.${DOMAIN}-arvados-snakeoil-ca.pem + + echo "Adding the vagrant user to the docker group" + usermod -a -G docker vagrant +else + cp /etc/ssl/certs/arvados-snakeoil-ca.pem ${SCRIPT_DIR}/${CLUSTER}.${DOMAIN}-arvados-snakeoil-ca.pem +fi + +# Test that the installation finished correctly +if [ "x${TEST}" = "xyes" ]; then + cd /tmp/cluster_tests + ./run-test.sh +fi