X-Git-Url: https://git.arvados.org/arvados.git/blobdiff_plain/675a683db2d878c5336f9f8c21d91e79dfe7a57e..8a27fe370239ecb8e50d53f46b45ed61203a35ca:/doc/install/salt-multi-host.html.textile.liquid diff --git a/doc/install/salt-multi-host.html.textile.liquid b/doc/install/salt-multi-host.html.textile.liquid index bb8cd0bdf4..640883baab 100644 --- a/doc/install/salt-multi-host.html.textile.liquid +++ b/doc/install/salt-multi-host.html.textile.liquid @@ -9,23 +9,39 @@ Copyright (C) The Arvados Authors. All rights reserved. SPDX-License-Identifier: CC-BY-SA-3.0 {% endcomment %} -# "Hosts preparation":#hosts_preparation -## "Hosts setup using terraform (experimental)":#hosts_setup_using_terraform -## "Create a compute image":#create_a_compute_image -# "Multi host install using the provision.sh script":#multi_host -# "Choose the desired configuration":#choose_configuration -## "Multiple hosts / multiple hostnames":#multi_host_multi_hostnames -## "Further customization of the installation (modifying the salt pillars and states)":#further_customization +# "Introduction":#introduction +# "Prerequisites and planning":#prerequisites +# "Download the installer":#download +# "Copy the configuration files":#copy_config +# "Choose the SSL configuration":#certificates +## "Using a self-signed certificates":#self-signed +## "Using a Let's Encrypt certificates":#lets-encrypt +## "Bring your own certificates":#bring-your-own +# "Create a compute image":#create_a_compute_image +# "Further customization of the installation (modifying the salt pillars and states)":#further_customization # "Installation order":#installation_order # "Run the provision.sh script":#run_provision_script +# "Install the CA root certificate":#ca_root_certificate # "Initial user and login":#initial_user # "Test the installed cluster running a simple workflow":#test_install +# "After the installation":#post_install -h2(#hosts_preparation). Hosts preparation +h2(#introduction). Introduction -In order to run Arvados on a multi-host installation, there are a few requirements that your infrastructure has to fulfill. +This multi host installer is an AWS specific example that is generally useful, but will likely need to be adapted for your environment. The installer is highly configurable. -These instructions explain how to setup a multi-host environment that is suitable for production use of Arvados. +h2(#prerequisites). Prerequisites and planning + +Prerequisites: + +* git +* a number of (virtual) machines for your Arvados cluster with at least 2 cores and 8 GiB of RAM, running a "supported Arvados distribution":{{site.baseurl}}/install/install-manual-prerequisites.html#supportedlinux +* a number of DNS hostnames that resolve to the IP addresses of your Arvados hosts +* ports 443 need to be reachable from your client (configurable in @local.params@, see below) +* port 80 needs to be reachable from everywhere on the internet (only when using "Let's Encrypt":#lets-encrypt without Route53 integration) +* SSL certificatse matching the hostnames in use (only when using "bring your own certificates":#bring-your-own) + +Planning: We suggest distributing the Arvados components in the following way, creating at least 6 hosts: @@ -36,58 +52,44 @@ We suggest distributing the Arvados components in the following way, creating at ## arvados controller ## arvados websocket ## arvados cloud dispatcher +## arvados keepbalance # WORKBENCH node: ## arvados workbench ## arvados workbench2 +## arvados webshell # KEEPPROXY node: ## arvados keepproxy ## arvados keepweb -# KEEPSTOREs (at least 2) +# KEEPSTORE nodes (at least 2) ## arvados keepstore # SHELL node (optional): ## arvados shell -Note that these hosts can be virtual machines in your infrastructure and they don't need to be physical machines. - -h3(#hosts_setup_using_terraform). Hosts setup using terraform (experimental) - -We added a few "terraform":https://terraform.io/ scripts (https://github.com/arvados/arvados/tree/master/tools/terraform) to let you create these instances easier. -Check "the Arvados terraform documentation":/doc/install/terraform.html for more details. - -h2(#multi_host). Multi host install using the provision.sh script - -This is a package-based installation method. The Salt scripts are available from the "tools/salt-install":https://github.com/arvados/arvados/tree/master/tools/salt-install directory in the Arvados git repository. - -This procedure will install all the main Arvados components to get you up and running in a multi host environment. - -We suggest you to use the @provision.sh@ script to deploy Arvados, which is implemented with the @arvados-formula@ in a Saltstack master-less setup. After setting up a few variables in a config file (next step), you'll be ready to run it and get Arvados deployed. - -h3(#create_a_compute_image). Create a compute image +If your infrastructure differs from the setup proposed above (ie, using RDS or an existing DB server), remember that you will need to edit the configuration files for the scripts so they work with your infrastructure. -In a multi-host installation, containers are dispatched in docker daemons running in the compute instances, which need some special setup. We provide a "compute image builder script":https://github.com/arvados/arvados/tree/master/tools/compute-images that you can use to build a template image following "these instructions":https://doc.arvados.org/main/install/crunch2-cloud/install-compute-node.html . Once you have that image created, you can use the image reference in the Arvados configuration in the next steps. +h2(#download). Download the installer -h2(#choose_configuration). Choose the desired configuration +{% include 'download_installer' %} -For documentation's sake, we will use the cluster name arva2 and the domain arv.local. If you don't change them as required in the next steps, installation won't proceed. +h2(#copy_config). Copy the configuration files -We will try to provide a few Arvados' multi host installation configurations examples for different infrastructure providers. Currently only AWS is available but they can be used with almost any provider with little changes. - -You need to copy one of the example configuration files and directory, and edit them to suit your needs. - -h3(#multi_host_multi_hostnames). Multiple hosts / multiple hostnames
cp local.params.example.multiple_hosts local.params
 cp -r config_examples/multi_host/aws local_config_dir
 
-Edit the variables in the local.params file. Pay attention to the *_INT_IP, *_TOKEN and *KEY variables. Those variables will be used to do a search and replace on the pillars/* in place of any matching __VARIABLE__. +Edit the variables in the local.params file. Pay attention to the *_INT_IP, *_TOKEN and *_KEY variables. The *SSL_MODE* variable is discussed in the next section. + +{% include 'ssl_config_multi' %} + +h3(#create_a_compute_image). Create a compute image -The multi_host include LetsEncrypt salt code to automatically request and install the certificates for the public-facing hosts (API, Workbench) so it will need the hostnames to be reachable from the Internet. If this cluster will not be the case, please set the variable USE_LETSENCRYPT=no. +In a multi-host installation, containers are dispatched in docker daemons running in the compute instances, which need some special setup. We provide a "compute image builder script":https://github.com/arvados/arvados/tree/main/tools/compute-images that you can use to build a template image following "these instructions":https://doc.arvados.org/main/install/crunch2-cloud/install-compute-node.html. Once you have that image created, you will need to update the pillars/arvados.sls file with the AMI ID and the private ssh key for the dispatcher. -## "Further customization of the installation (modifying the salt pillars and states)":#further_customization +h3(#further_customization). Further customization of the installation (modifying the salt pillars and states) -You will need further customization to suit your environment, which can be done editing the Saltstack pillars and states files. Pay particular attention to the pillars/arvados.sls file, where you will need to provide some information that can be retrieved as output of the terraform run. +You will need further customization to suit your environment, which can be done editing the Saltstack pillars and states files. Pay particular attention to the pillars/arvados.sls file, where you will need to provide some information that describes your environment. Any extra state file you add under local_config_dir/states will be added to the salt run and applied to the hosts. @@ -95,9 +97,9 @@ h2(#installation_order). Installation order A few Arvados nodes need to be installed in certain order. The required order is -#. Database -#. API server -#. The other nodes can be installed in any order after the two above +* Database +* API server +* The other nodes can be installed in any order after the two above h2(#run_provision_script). Run the provision.sh script @@ -121,49 +123,51 @@ arvados: Failed: 0 The distribution of role as described above can be applied running these commands: -#. Database +h4. Database
scp -r provision.sh local* user@host:
 ssh user@host sudo ./provision.sh --config local.params --roles database
 
-#. API +h4. API
scp -r provision.sh local* user@host:
-ssh user@host sudo ./provision.sh --config local.params --roles api,controller,websocket,dispatcher
+ssh user@host sudo ./provision.sh --config local.params --roles api,controller,websocket,dispatcher,keepbalance
 
-#. Keepstore/s +h4. Keepstore(s)
scp -r provision.sh local* user@host:
 ssh user@host sudo ./provision.sh --config local.params --roles keepstore
 
-#. Workbench +h4. Workbench
scp -r provision.sh local* user@host:
-ssh user@host sudo ./provision.sh --config local.params --roles workbench,workbench2
+ssh user@host sudo ./provision.sh --config local.params --roles workbench,workbench2,webshell
 
-#. Keepproxy / Keepweb +h4. Keepproxy / Keepweb
scp -r provision.sh local* user@host:
 ssh user@host sudo ./provision.sh --config local.params --roles keepproxy,keepweb
 
-#. Shell +h4. Shell (here we copy the CLI test workflow too) -
scp -r provision.sh local* user@host:
+
scp -r provision.sh local* tests user@host:
 ssh user@host sudo ./provision.sh --config local.params --roles shell
 
-h2(#initial_user). Initial user and login +{% include 'install_ca_cert' %} + +h2(#initial_user). Initial user and login At this point you should be able to log into the Arvados cluster. The initial URL will be: @@ -183,11 +187,11 @@ Assuming you didn't change these values in the @local.params@ file, the initial h2(#test_install). Test the installed cluster running a simple workflow -The @provision.sh@ script saves a simple example test workflow in the @/tmp/cluster_tests@ directory in the node. If you want to run it, just ssh to the node, change to that directory and run: +If you followed the instructions above, the @provision.sh@ script saves a simple example test workflow in the @/tmp/cluster_tests@ directory in the @shell@ node. If you want to run it, just ssh to the node, change to that directory and run:
cd /tmp/cluster_tests
-./run-test.sh
+sudo /run-test.sh
 
@@ -272,3 +276,9 @@ INFO Final output collection d6c69a88147dde9d52a418d50ef788df+123 INFO Final process status is success
+ +h2(#post_install). After the installation + +Once the installation is complete, it is recommended to keep a copy of your local configuration files. Committing them to version control is a good idea. + +Re-running the Salt-based installer is not recommended for maintaining and upgrading Arvados, please see "Maintenance and upgrading":{{site.baseurl}}/admin/maintenance-and-upgrading.html for more information.