---
layout: default
navsection: installguide
title: Install the Crunch dispatcher
...
{% comment %}
Copyright (C) The Arvados Authors. All rights reserved.
SPDX-License-Identifier: CC-BY-SA-3.0
{% endcomment %}
The dispatcher normally runs on the same host/VM as the API server.
h2. Test the Arvados job queue
Crunch dispatches work from the job queue on the Arvados API server. Before you start installing the Crunch dispatcher, now's a good time to check that the API server and Git server can coordinate to create job records. Run these commands *on your shell server* to create a collection, and a job to calculate the MD5 checksum of every file in it:
~$ echo 'Hello, Crunch!' | arv-put --portable-data-hash -
…
d40c7f35d80da669afb9db1896e760ad+49
~$ read -rd $'\000' newjob <<EOF; arv job create --job "$newjob"
{"script_parameters":{"input":"d40c7f35d80da669afb9db1896e760ad+49"},
"script_version":"0988acb472849dc0",
"script":"hash",
"repository":"arvados"}
EOF
ArgumentError: Specified script_version does not resolve to a commitit often means that the API server can't read the specified repository—either because it doesn't exist, or because the user running the API server doesn't have permission to read the repository files. Check the API server's log (@/var/www/arvados-api/current/log/production.log@) for details, and double-check the instructions in the "Git server installation guide":install-arv-git-httpd.html. If everything goes well, the API server should create a job record, and your @arv@ command will output the JSON for that record. It should have state @Queued@ and script_version @0988acb472849dc08d576ee40493e70bde2132ca@. If the job JSON includes those fields, you can proceed to install the Crunch dispatcher and a compute node. This job will remain queued until you install those services. h2. Perl SDK dependencies Install the Perl SDK on the controller. * See "Perl SDK":{{site.baseurl}}/sdk/perl/index.html page for details. h2. Python SDK dependencies Install the Python SDK and CLI tools on controller and all compute nodes. * See "Python SDK":{{site.baseurl}}/sdk/python/sdk-python.html page for details. h2(#slurm). Set up SLURM On the API server, install SLURM and munge, and generate a munge key. On Debian-based systems:
~$ sudo /usr/bin/apt-get install slurm-llnl munge
~$ sudo /usr/sbin/create-munge-key
~$ sudo yum install slurm munge slurm-munge
ControlMachine=uuid_prefix.your.domain SlurmctldPort=6817 SlurmdPort=6818 AuthType=auth/munge StateSaveLocation=/tmp SlurmdSpoolDir=/tmp/slurmd SwitchType=switch/none MpiDefault=none SlurmctldPidFile=/var/run/slurmctld.pid SlurmdPidFile=/var/run/slurmd.pid ProctrackType=proctrack/pgid CacheGroups=0 ReturnToService=2 TaskPlugin=task/affinity # # TIMERS SlurmctldTimeout=300 SlurmdTimeout=300 InactiveLimit=0 MinJobAge=300 KillWait=30 Waittime=0 # # SCHEDULING SchedulerType=sched/backfill SchedulerPort=7321 SelectType=select/linear FastSchedule=0 # # LOGGING SlurmctldDebug=3 #SlurmctldLogFile= SlurmdDebug=3 #SlurmdLogFile= JobCompType=jobcomp/none #JobCompLoc= JobAcctGatherType=jobacct_gather/none # # COMPUTE NODES NodeName=DEFAULT PartitionName=DEFAULT MaxTime=INFINITE State=UP NodeName=compute[0-255] PartitionName=compute Nodes=compute[0-255] Default=YES Shared=YES
assign_node_hostname: worker1-%04d
* In @slurm.conf@: NodeName=worker1-[0000-0255]
If your worker hostnames are already assigned by other means, and the full set of names is known in advance, have your worker node bootstrapping script (see "Installing a compute node":install-compute-node.html) send its current hostname, rather than expect Arvados to assign one.
* In @application.yml@: assign_node_hostname: false
* In @slurm.conf@: NodeName=alice,bob,clay,darlene
If your worker hostnames are already assigned by other means, but the full set of names is _not_ known in advance, you can use the @slurm.conf@ and @application.yml@ settings in the previous example, but you must also update @slurm.conf@ (both on the controller and on all worker nodes) and run @sudo scontrol reconfigure@ whenever a new node comes online.
h2. Enable SLURM job dispatch
In your API server's @application.yml@ configuration file, add the line @crunch_job_wrapper: :slurm_immediate@ under the appropriate section. (The second colon is not a typo. It denotes a Ruby symbol.)
h2. Crunch user account
Run @sudo adduser crunch@. The crunch user should have the same UID, GID, and home directory on all compute nodes and on the dispatcher (API server).
h2. Run the Crunch dispatcher service
To dispatch Arvados jobs:
* The API server script @crunch-dispatch.rb@ must be running.
* @crunch-job@ needs the installation path of the Perl SDK in its @PERLLIB@.
* @crunch-job@ needs the @ARVADOS_API_HOST@ (and, if necessary, @ARVADOS_API_HOST_INSECURE@) environment variable set.
Install runit to monitor the Crunch dispatch daemon. {% include 'install_runit' %}
Install the script below as the run script for the Crunch dispatch service, modifying it as directed by the comments.
#!/bin/sh
set -e
rvmexec=""
## Uncomment this line if you use RVM:
#rvmexec="/usr/local/rvm/bin/rvm-exec default"
export ARVADOS_API_HOST=uuid_prefix.your.domain
export CRUNCH_DISPATCH_LOCKFILE=/var/lock/crunch-dispatch
export HOME=$(pwd)
export RAILS_ENV=production
## Uncomment and edit this line if your compute nodes have cgroup info
## somewhere other than /sys/fs/cgroup (e.g., "/cgroup" for CentOS 7)
#export CRUNCH_CGROUP_ROOT="/sys/fs/cgroup"
## Uncomment this line if your cluster uses self-signed SSL certificates:
#export ARVADOS_API_HOST_INSECURE=yes
# This is the path to docker on your compute nodes. You might need to
# change it to "docker", "/opt/bin/docker", etc.
export CRUNCH_JOB_DOCKER_BIN=docker.io
fuser -TERM -k $CRUNCH_DISPATCH_LOCKFILE || true
cd /var/www/arvados-api/current
exec $rvmexec bundle exec ./script/crunch-dispatch.rb 2>&1