X-Git-Url: https://git.arvados.org/arvados.git/blobdiff_plain/4823a1b88754ef8dc3a4fe3fcb549cb4e6f34246..c342e37e4fd4925926b6dd03d116da96c1c5fba3:/tools/salt-install/local.params.example.single_host_single_hostname diff --git a/tools/salt-install/local.params.example.single_host_single_hostname b/tools/salt-install/local.params.example.single_host_single_hostname index eaa71e7a20..0c4f5c3567 100644 --- a/tools/salt-install/local.params.example.single_host_single_hostname +++ b/tools/salt-install/local.params.example.single_host_single_hostname @@ -13,39 +13,14 @@ DOMAIN="domain_fixme_or_this_wont_work" # For multi-node installs, the ssh log in for each node # must be root or able to sudo -DEPLOY_USER=root +DEPLOY_USER=admin -# The mapping of nodes to roles -# installer.sh will log in to each of these nodes and then provision -# it for the specified roles. -NODES=( - [localhost]='' -) - -# HOSTNAME_EXT must be set to the address that users will use to -# connect to the instance (e.g. what they will type into the URL bar -# of the browser to get to workbench). If you haven't given the -# instance a working DNS name, you might need to use an IP address -# here. -HOSTNAME_EXT="hostname_ext_fixme_or_this_wont_work" +INITIAL_USER=admin -# The internal IP address for the host. -IP_INT="ip_int_fixme_or_this_wont_work" - -# External ports used by the Arvados services -CONTROLLER_EXT_SSL_PORT=8800 -KEEP_EXT_SSL_PORT=8801 -KEEPWEB_EXT_SSL_PORT=8802 -WEBSHELL_EXT_SSL_PORT=8803 -WEBSOCKET_EXT_SSL_PORT=8804 -WORKBENCH1_EXT_SSL_PORT=8805 -WORKBENCH2_EXT_SSL_PORT=443 - -INITIAL_USER="admin" # If not specified, the initial user email will be composed as # INITIAL_USER@CLUSTER.DOMAIN INITIAL_USER_EMAIL="admin@cluster_fixme_or_this_wont_work.domain_fixme_or_this_wont_work" -INITIAL_USER_PASSWORD="password" +INITIAL_USER_PASSWORD="fixmepassword" # Populate these values with random strings BLOB_SIGNING_KEY=fixmeblobsigningkeymushaveatleast32characters @@ -68,6 +43,44 @@ SSL_MODE="self-signed" # See https://doc.arvados.org/intall/salt-single-host.html#bring-your-own for more information. # CUSTOM_CERTS_DIR="${SCRIPT_DIR}/local_config_dir/certs" +# Set the following to "yes" if the key files are encrypted and optionally set +# a custom AWS secret name for each node to retrieve the password. +SSL_KEY_ENCRYPTED="no" +SSL_KEY_AWS_SECRET_NAME="${CLUSTER}-arvados-ssl-privkey-password" + +# Customize Prometheus & Grafana web UI access credentials +MONITORING_USERNAME=${INITIAL_USER} +MONITORING_PASSWORD=${INITIAL_USER_PASSWORD} +MONITORING_EMAIL=${INITIAL_USER_EMAIL} +# Sets the directory for Grafana dashboards +# GRAFANA_DASHBOARDS_DIR="${SCRIPT_DIR}/local_config_dir/dashboards" + +# The mapping of nodes to roles +# installer.sh will log in to each of these nodes and then provision +# it for the specified roles. +NODES=( + [localhost]='' +) + +# HOSTNAME_EXT must be set to the address that users will use to +# connect to the instance (e.g. what they will type into the URL bar +# of the browser to get to workbench). If you haven't given the +# instance a working DNS name, you might need to use an IP address +# here. +HOSTNAME_EXT="hostname_ext_fixme_or_this_wont_work" + +# The internal IP address for the host. +IP_INT="ip_int_fixme_or_this_wont_work" + +# External ports used by the Arvados services +CONTROLLER_EXT_SSL_PORT=8800 +KEEP_EXT_SSL_PORT=8801 +KEEPWEB_EXT_SSL_PORT=8802 +WEBSHELL_EXT_SSL_PORT=8803 +WEBSOCKET_EXT_SSL_PORT=8804 +WORKBENCH1_EXT_SSL_PORT=8805 +WORKBENCH2_EXT_SSL_PORT=443 + # The directory to check for the config files (pillars, states) you want to use. # There are a few examples under 'config_examples'. # CONFIG_DIR="local_config_dir" @@ -96,3 +109,5 @@ RELEASE="production" # DOCKER_TAG="v2.4.2" # LOCALE_TAG="v0.3.4" # LETSENCRYPT_TAG="v2.1.0" +# PROMETHEUS_TAG="v5.6.5" +# GRAFANA_TAG="v3.1.3"