20462: Refactor "common prefix" function & fix fencepost errors
[arvados.git] / sdk / cwl / arvados_cwl / arvcontainer.py
1 # Copyright (C) The Arvados Authors. All rights reserved.
2 #
3 # SPDX-License-Identifier: Apache-2.0
4
5 from future import standard_library
6 standard_library.install_aliases()
7 from builtins import str
8
9 import logging
10 import json
11 import os
12 import urllib.request, urllib.parse, urllib.error
13 import time
14 import datetime
15 import ciso8601
16 import uuid
17 import math
18 import re
19
20 import arvados_cwl.util
21 import ruamel.yaml
22
23 from cwltool.errors import WorkflowException
24 from cwltool.process import UnsupportedRequirement, shortname
25 from cwltool.utils import aslist, adjustFileObjs, adjustDirObjs, visit_class
26 from cwltool.job import JobBase
27
28 import arvados.collection
29
30 from .arvdocker import arv_docker_get_image
31 from . import done
32 from .runner import Runner, arvados_jobs_image, packed_workflow, trim_anonymous_location, remove_redundant_fields, make_builder
33 from .fsaccess import CollectionFetcher
34 from .pathmapper import NoFollowPathMapper, trim_listing
35 from .perf import Perf
36 from ._version import __version__
37
38 logger = logging.getLogger('arvados.cwl-runner')
39 metrics = logging.getLogger('arvados.cwl-runner.metrics')
40
41 def cleanup_name_for_collection(name):
42     return name.replace("/", " ")
43
44 class ArvadosContainer(JobBase):
45     """Submit and manage a Crunch container request for executing a CWL CommandLineTool."""
46
47     def __init__(self, runner, job_runtime,
48                  builder,   # type: Builder
49                  joborder,  # type: Dict[Text, Union[Dict[Text, Any], List, Text]]
50                  make_path_mapper,  # type: Callable[..., PathMapper]
51                  requirements,      # type: List[Dict[Text, Text]]
52                  hints,     # type: List[Dict[Text, Text]]
53                  name       # type: Text
54     ):
55         super(ArvadosContainer, self).__init__(builder, joborder, make_path_mapper, requirements, hints, name)
56         self.arvrunner = runner
57         self.job_runtime = job_runtime
58         self.running = False
59         self.uuid = None
60         self.attempt_count = 0
61
62     def update_pipeline_component(self, r):
63         pass
64
65     def _required_env(self):
66         env = {}
67         env["HOME"] = self.outdir
68         env["TMPDIR"] = self.tmpdir
69         return env
70
71     def run(self, toplevelRuntimeContext):
72         # ArvadosCommandTool subclasses from cwltool.CommandLineTool,
73         # which calls makeJobRunner() to get a new ArvadosContainer
74         # object.  The fields that define execution such as
75         # command_line, environment, etc are set on the
76         # ArvadosContainer object by CommandLineTool.job() before
77         # run() is called.
78
79         runtimeContext = self.job_runtime
80
81         if runtimeContext.submit_request_uuid:
82             container_request = self.arvrunner.api.container_requests().get(
83                 uuid=runtimeContext.submit_request_uuid
84             ).execute(num_retries=self.arvrunner.num_retries)
85         else:
86             container_request = {}
87
88         container_request["command"] = self.command_line
89         container_request["name"] = self.name
90         container_request["output_path"] = self.outdir
91         container_request["cwd"] = self.outdir
92         container_request["priority"] = runtimeContext.priority
93         container_request["state"] = "Uncommitted"
94         container_request.setdefault("properties", {})
95
96         container_request["properties"]["cwl_input"] = self.joborder
97
98         runtime_constraints = {}
99
100         if runtimeContext.project_uuid:
101             container_request["owner_uuid"] = runtimeContext.project_uuid
102
103         if self.arvrunner.secret_store.has_secret(self.command_line):
104             raise WorkflowException("Secret material leaked on command line, only file literals may contain secrets")
105
106         if self.arvrunner.secret_store.has_secret(self.environment):
107             raise WorkflowException("Secret material leaked in environment, only file literals may contain secrets")
108
109         resources = self.builder.resources
110         if resources is not None:
111             runtime_constraints["vcpus"] = math.ceil(resources.get("cores", 1))
112             runtime_constraints["ram"] = math.ceil(resources.get("ram") * 2**20)
113
114         mounts = {
115             self.outdir: {
116                 "kind": "tmp",
117                 "capacity": math.ceil(resources.get("outdirSize", 0) * 2**20)
118             },
119             self.tmpdir: {
120                 "kind": "tmp",
121                 "capacity": math.ceil(resources.get("tmpdirSize", 0) * 2**20)
122             }
123         }
124         secret_mounts = {}
125         scheduling_parameters = {}
126
127         rf = [self.pathmapper.mapper(f) for f in self.pathmapper.referenced_files]
128         rf.sort(key=lambda k: k.resolved)
129         prevdir = None
130         for resolved, target, tp, stg in rf:
131             if not stg:
132                 continue
133             if prevdir and target.startswith(prevdir):
134                 continue
135             if tp == "Directory":
136                 targetdir = target
137             else:
138                 targetdir = os.path.dirname(target)
139             sp = resolved.split("/", 1)
140             pdh = sp[0][5:]   # remove "keep:"
141             mounts[targetdir] = {
142                 "kind": "collection",
143                 "portable_data_hash": pdh
144             }
145             if pdh in self.pathmapper.pdh_to_uuid:
146                 mounts[targetdir]["uuid"] = self.pathmapper.pdh_to_uuid[pdh]
147             if len(sp) == 2:
148                 if tp == "Directory":
149                     path = sp[1]
150                 else:
151                     path = os.path.dirname(sp[1])
152                 if path and path != "/":
153                     mounts[targetdir]["path"] = path
154             prevdir = targetdir + "/"
155
156         intermediate_collection_info = arvados_cwl.util.get_intermediate_collection_info(self.name, runtimeContext.current_container, runtimeContext.intermediate_output_ttl)
157
158         with Perf(metrics, "generatefiles %s" % self.name):
159             if self.generatefiles["listing"]:
160                 vwd = arvados.collection.Collection(api_client=self.arvrunner.api,
161                                                     keep_client=self.arvrunner.keep_client,
162                                                     num_retries=self.arvrunner.num_retries)
163                 generatemapper = NoFollowPathMapper(self.generatefiles["listing"], "", "",
164                                                     separateDirs=False)
165
166                 sorteditems = sorted(generatemapper.items(), key=lambda n: n[1].target)
167
168                 logger.debug("generatemapper is %s", sorteditems)
169
170                 with Perf(metrics, "createfiles %s" % self.name):
171                     for f, p in sorteditems:
172                         if not p.target:
173                             continue
174
175                         if p.target.startswith("/"):
176                             dst = p.target[len(self.outdir)+1:] if p.target.startswith(self.outdir+"/") else p.target[1:]
177                         else:
178                             dst = p.target
179
180                         if p.type in ("File", "Directory", "WritableFile", "WritableDirectory"):
181                             if p.resolved.startswith("_:"):
182                                 vwd.mkdirs(dst)
183                             else:
184                                 source, path = self.arvrunner.fs_access.get_collection(p.resolved)
185                                 vwd.copy(path or ".", dst, source_collection=source)
186                         elif p.type == "CreateFile":
187                             if self.arvrunner.secret_store.has_secret(p.resolved):
188                                 mountpoint = p.target if p.target.startswith("/") else os.path.join(self.outdir, p.target)
189                                 secret_mounts[mountpoint] = {
190                                     "kind": "text",
191                                     "content": self.arvrunner.secret_store.retrieve(p.resolved)
192                                 }
193                             else:
194                                 with vwd.open(dst, "w") as n:
195                                     n.write(p.resolved)
196
197                 def keepemptydirs(p):
198                     if isinstance(p, arvados.collection.RichCollectionBase):
199                         if len(p) == 0:
200                             p.open(".keep", "w").close()
201                         else:
202                             for c in p:
203                                 keepemptydirs(p[c])
204
205                 keepemptydirs(vwd)
206
207                 if not runtimeContext.current_container:
208                     runtimeContext.current_container = arvados_cwl.util.get_current_container(self.arvrunner.api, self.arvrunner.num_retries, logger)
209                 vwd.save_new(name=intermediate_collection_info["name"],
210                              owner_uuid=runtimeContext.project_uuid,
211                              ensure_unique_name=True,
212                              trash_at=intermediate_collection_info["trash_at"],
213                              properties=intermediate_collection_info["properties"])
214
215                 prev = None
216                 for f, p in sorteditems:
217                     if (not p.target or self.arvrunner.secret_store.has_secret(p.resolved) or
218                         (prev is not None and p.target.startswith(prev))):
219                         continue
220                     if p.target.startswith("/"):
221                         dst = p.target[len(self.outdir)+1:] if p.target.startswith(self.outdir+"/") else p.target[1:]
222                     else:
223                         dst = p.target
224                     mountpoint = p.target if p.target.startswith("/") else os.path.join(self.outdir, p.target)
225                     mounts[mountpoint] = {"kind": "collection",
226                                           "portable_data_hash": vwd.portable_data_hash(),
227                                           "path": dst}
228                     if p.type.startswith("Writable"):
229                         mounts[mountpoint]["writable"] = True
230                     prev = p.target + "/"
231
232         container_request["environment"] = {"TMPDIR": self.tmpdir, "HOME": self.outdir}
233         if self.environment:
234             container_request["environment"].update(self.environment)
235
236         if self.stdin:
237             sp = self.stdin[6:].split("/", 1)
238             mounts["stdin"] = {"kind": "collection",
239                                 "portable_data_hash": sp[0],
240                                 "path": sp[1]}
241
242         if self.stderr:
243             mounts["stderr"] = {"kind": "file",
244                                 "path": "%s/%s" % (self.outdir, self.stderr)}
245
246         if self.stdout:
247             mounts["stdout"] = {"kind": "file",
248                                 "path": "%s/%s" % (self.outdir, self.stdout)}
249
250         (docker_req, docker_is_req) = self.get_requirement("DockerRequirement")
251
252         container_request["container_image"] = arv_docker_get_image(self.arvrunner.api,
253                                                                     docker_req,
254                                                                     runtimeContext.pull_image,
255                                                                     runtimeContext)
256
257         network_req, _ = self.get_requirement("NetworkAccess")
258         if network_req:
259             runtime_constraints["API"] = network_req["networkAccess"]
260
261         api_req, _ = self.get_requirement("http://arvados.org/cwl#APIRequirement")
262         if api_req:
263             runtime_constraints["API"] = True
264
265         use_disk_cache = (self.arvrunner.api.config()["Containers"].get("DefaultKeepCacheRAM", 0) == 0)
266
267         keep_cache_type_req, _ = self.get_requirement("http://arvados.org/cwl#KeepCacheTypeRequirement")
268         if keep_cache_type_req:
269             if "keepCacheType" in keep_cache_type_req:
270                 if keep_cache_type_req["keepCacheType"] == "ram_cache":
271                     use_disk_cache = False
272
273         runtime_req, _ = self.get_requirement("http://arvados.org/cwl#RuntimeConstraints")
274         if runtime_req:
275             if "keep_cache" in runtime_req:
276                 if use_disk_cache:
277                     # If DefaultKeepCacheRAM is zero it means we should use disk cache.
278                     runtime_constraints["keep_cache_disk"] = math.ceil(runtime_req["keep_cache"] * 2**20)
279                 else:
280                     runtime_constraints["keep_cache_ram"] = math.ceil(runtime_req["keep_cache"] * 2**20)
281             if "outputDirType" in runtime_req:
282                 if runtime_req["outputDirType"] == "local_output_dir":
283                     # Currently the default behavior.
284                     pass
285                 elif runtime_req["outputDirType"] == "keep_output_dir":
286                     mounts[self.outdir]= {
287                         "kind": "collection",
288                         "writable": True
289                     }
290
291         partition_req, _ = self.get_requirement("http://arvados.org/cwl#PartitionRequirement")
292         if partition_req:
293             scheduling_parameters["partitions"] = aslist(partition_req["partition"])
294
295         intermediate_output_req, _ = self.get_requirement("http://arvados.org/cwl#IntermediateOutput")
296         if intermediate_output_req:
297             self.output_ttl = intermediate_output_req["outputTTL"]
298         else:
299             self.output_ttl = self.arvrunner.intermediate_output_ttl
300
301         if self.output_ttl < 0:
302             raise WorkflowException("Invalid value %d for output_ttl, cannot be less than zero" % container_request["output_ttl"])
303
304
305         if self.arvrunner.api._rootDesc["revision"] >= "20210628":
306             storage_class_req, _ = self.get_requirement("http://arvados.org/cwl#OutputStorageClass")
307             if storage_class_req and storage_class_req.get("intermediateStorageClass"):
308                 container_request["output_storage_classes"] = aslist(storage_class_req["intermediateStorageClass"])
309             else:
310                 container_request["output_storage_classes"] = runtimeContext.intermediate_storage_classes.strip().split(",")
311
312         cuda_req, _ = self.get_requirement("http://commonwl.org/cwltool#CUDARequirement")
313         if cuda_req:
314             runtime_constraints["cuda"] = {
315                 "device_count": resources.get("cudaDeviceCount", 1),
316                 "driver_version": cuda_req["cudaVersionMin"],
317                 "hardware_capability": aslist(cuda_req["cudaComputeCapability"])[0]
318             }
319
320         if runtimeContext.enable_preemptible is False:
321             scheduling_parameters["preemptible"] = False
322         else:
323             preemptible_req, _ = self.get_requirement("http://arvados.org/cwl#UsePreemptible")
324             if preemptible_req:
325                 scheduling_parameters["preemptible"] = preemptible_req["usePreemptible"]
326             elif runtimeContext.enable_preemptible is True:
327                 scheduling_parameters["preemptible"] = True
328             elif runtimeContext.enable_preemptible is None:
329                 pass
330
331         if self.timelimit is not None and self.timelimit > 0:
332             scheduling_parameters["max_run_time"] = self.timelimit
333
334         extra_submit_params = {}
335         if runtimeContext.submit_runner_cluster:
336             extra_submit_params["cluster_id"] = runtimeContext.submit_runner_cluster
337
338         container_request["output_name"] = cleanup_name_for_collection("Output from step %s" % (self.name))
339         container_request["output_ttl"] = self.output_ttl
340         container_request["mounts"] = mounts
341         container_request["secret_mounts"] = secret_mounts
342         container_request["runtime_constraints"] = runtime_constraints
343         container_request["scheduling_parameters"] = scheduling_parameters
344
345         enable_reuse = runtimeContext.enable_reuse
346         if enable_reuse:
347             reuse_req, _ = self.get_requirement("WorkReuse")
348             if reuse_req:
349                 enable_reuse = reuse_req["enableReuse"]
350             reuse_req, _ = self.get_requirement("http://arvados.org/cwl#ReuseRequirement")
351             if reuse_req:
352                 enable_reuse = reuse_req["enableReuse"]
353         container_request["use_existing"] = enable_reuse
354
355         properties_req, _ = self.get_requirement("http://arvados.org/cwl#ProcessProperties")
356         if properties_req:
357             for pr in properties_req["processProperties"]:
358                 container_request["properties"][pr["propertyName"]] = self.builder.do_eval(pr["propertyValue"])
359
360         output_properties_req, _ = self.get_requirement("http://arvados.org/cwl#OutputCollectionProperties")
361         if output_properties_req:
362             if self.arvrunner.api._rootDesc["revision"] >= "20220510":
363                 container_request["output_properties"] = {}
364                 for pr in output_properties_req["outputProperties"]:
365                     container_request["output_properties"][pr["propertyName"]] = self.builder.do_eval(pr["propertyValue"])
366             else:
367                 logger.warning("%s API revision is %s, revision %s is required to support setting properties on output collections.",
368                                self.arvrunner.label(self), self.arvrunner.api._rootDesc["revision"], "20220510")
369
370         ram_multiplier = [1]
371
372         oom_retry_req, _ = self.get_requirement("http://arvados.org/cwl#OutOfMemoryRetry")
373         if oom_retry_req and oom_retry_req.get('memoryRetryMultipler'):
374             ram_multiplier.append(oom_retry_req.get('memoryRetryMultipler'))
375
376         if runtimeContext.runnerjob.startswith("arvwf:"):
377             wfuuid = runtimeContext.runnerjob[6:runtimeContext.runnerjob.index("#")]
378             wfrecord = self.arvrunner.api.workflows().get(uuid=wfuuid).execute(num_retries=self.arvrunner.num_retries)
379             if container_request["name"] == "main":
380                 container_request["name"] = wfrecord["name"]
381             container_request["properties"]["template_uuid"] = wfuuid
382
383         if self.attempt_count == 0:
384             self.output_callback = self.arvrunner.get_wrapped_callback(self.output_callback)
385
386         try:
387             ram = runtime_constraints["ram"]
388
389             self.uuid = runtimeContext.submit_request_uuid
390
391             for i in ram_multiplier:
392                 runtime_constraints["ram"] = ram * i
393
394                 if self.uuid:
395                     response = self.arvrunner.api.container_requests().update(
396                         uuid=self.uuid,
397                         body=container_request,
398                         **extra_submit_params
399                     ).execute(num_retries=self.arvrunner.num_retries)
400                 else:
401                     response = self.arvrunner.api.container_requests().create(
402                         body=container_request,
403                         **extra_submit_params
404                     ).execute(num_retries=self.arvrunner.num_retries)
405                     self.uuid = response["uuid"]
406
407                 if response["container_uuid"] is not None:
408                     break
409
410             if response["container_uuid"] is None:
411                 runtime_constraints["ram"] = ram * ram_multiplier[self.attempt_count]
412
413             container_request["state"] = "Committed"
414             response = self.arvrunner.api.container_requests().update(
415                 uuid=self.uuid,
416                 body=container_request,
417                 **extra_submit_params
418             ).execute(num_retries=self.arvrunner.num_retries)
419
420             self.arvrunner.process_submitted(self)
421             self.attempt_count += 1
422
423             if response["state"] == "Final":
424                 logger.info("%s reused container %s", self.arvrunner.label(self), response["container_uuid"])
425             else:
426                 logger.info("%s %s state is %s", self.arvrunner.label(self), response["uuid"], response["state"])
427         except Exception as e:
428             logger.exception("%s error submitting container\n%s", self.arvrunner.label(self), e)
429             logger.debug("Container request was %s", container_request)
430             self.output_callback({}, "permanentFail")
431
432     def out_of_memory_retry(self, record, container):
433         oom_retry_req, _ = self.get_requirement("http://arvados.org/cwl#OutOfMemoryRetry")
434         if oom_retry_req is None:
435             return False
436
437         # Sometimes it gets killed with no warning
438         if container["exit_code"] == 137:
439             return True
440
441         logc = arvados.collection.CollectionReader(record["log_uuid"],
442                                                    api_client=self.arvrunner.api,
443                                                    keep_client=self.arvrunner.keep_client,
444                                                    num_retries=self.arvrunner.num_retries)
445
446         loglines = [""]
447         def callback(v1, v2, v3):
448             loglines[0] = v3
449
450         done.logtail(logc, callback, "", maxlen=1000)
451
452         # Check allocation failure
453         oom_matches = oom_retry_req.get('memoryErrorRegex') or r'(bad_alloc|out ?of ?memory|memory ?error|container using over 9.% of memory)'
454         if re.search(oom_matches, loglines[0], re.IGNORECASE | re.MULTILINE):
455             return True
456
457         return False
458
459     def done(self, record):
460         outputs = {}
461         retried = False
462         try:
463             container = self.arvrunner.api.containers().get(
464                 uuid=record["container_uuid"]
465             ).execute(num_retries=self.arvrunner.num_retries)
466             if container["state"] == "Complete":
467                 rcode = container["exit_code"]
468                 if self.successCodes and rcode in self.successCodes:
469                     processStatus = "success"
470                 elif self.temporaryFailCodes and rcode in self.temporaryFailCodes:
471                     processStatus = "temporaryFail"
472                 elif self.permanentFailCodes and rcode in self.permanentFailCodes:
473                     processStatus = "permanentFail"
474                 elif rcode == 0:
475                     processStatus = "success"
476                 else:
477                     processStatus = "permanentFail"
478
479                 if processStatus == "permanentFail" and self.attempt_count == 1 and self.out_of_memory_retry(record, container):
480                     logger.warning("%s Container failed with out of memory error, retrying with more RAM.",
481                                  self.arvrunner.label(self))
482                     self.job_runtime.submit_request_uuid = None
483                     self.uuid = None
484                     self.run(None)
485                     retried = True
486                     return
487
488                 if rcode == 137:
489                     logger.warning("%s Container may have been killed for using too much RAM.  Try resubmitting with a higher 'ramMin' or use the arv:OutOfMemoryRetry feature.",
490                                  self.arvrunner.label(self))
491             else:
492                 processStatus = "permanentFail"
493
494             if processStatus == "permanentFail" and record["log_uuid"]:
495                 logc = arvados.collection.CollectionReader(record["log_uuid"],
496                                                            api_client=self.arvrunner.api,
497                                                            keep_client=self.arvrunner.keep_client,
498                                                            num_retries=self.arvrunner.num_retries)
499                 label = self.arvrunner.label(self)
500                 done.logtail(
501                     logc, logger.error,
502                     "%s (%s) error log:" % (label, record["uuid"]), maxlen=40)
503
504             if record["output_uuid"]:
505                 if self.arvrunner.trash_intermediate or self.arvrunner.intermediate_output_ttl:
506                     # Compute the trash time to avoid requesting the collection record.
507                     trash_at = ciso8601.parse_datetime_as_naive(record["modified_at"]) + datetime.timedelta(0, self.arvrunner.intermediate_output_ttl)
508                     aftertime = " at %s" % trash_at.strftime("%Y-%m-%d %H:%M:%S UTC") if self.arvrunner.intermediate_output_ttl else ""
509                     orpart = ", or" if self.arvrunner.trash_intermediate and self.arvrunner.intermediate_output_ttl else ""
510                     oncomplete = " upon successful completion of the workflow" if self.arvrunner.trash_intermediate else ""
511                     logger.info("%s Intermediate output %s (%s) will be trashed%s%s%s." % (
512                         self.arvrunner.label(self), record["output_uuid"], container["output"], aftertime, orpart, oncomplete))
513                 self.arvrunner.add_intermediate_output(record["output_uuid"])
514
515             if container["output"]:
516                 outputs = done.done_outputs(self, container, "/tmp", self.outdir, "/keep")
517
518             properties = record["properties"].copy()
519             properties["cwl_output"] = outputs
520             self.arvrunner.api.container_requests().update(
521                 uuid=self.uuid,
522                 body={"container_request": {"properties": properties}}
523             ).execute(num_retries=self.arvrunner.num_retries)
524         except WorkflowException as e:
525             # Only include a stack trace if in debug mode.
526             # A stack trace may obfuscate more useful output about the workflow.
527             logger.error("%s unable to collect output from %s:\n%s",
528                          self.arvrunner.label(self), container["output"], e, exc_info=(e if self.arvrunner.debug else False))
529             processStatus = "permanentFail"
530         except Exception:
531             logger.exception("%s while getting output object:", self.arvrunner.label(self))
532             processStatus = "permanentFail"
533         finally:
534             if not retried:
535                 self.output_callback(outputs, processStatus)
536
537
538 class RunnerContainer(Runner):
539     """Submit and manage a container that runs arvados-cwl-runner."""
540
541     def arvados_job_spec(self, runtimeContext, git_info):
542         """Create an Arvados container request for this workflow.
543
544         The returned dict can be used to create a container passed as
545         the +body+ argument to container_requests().create().
546         """
547
548         adjustDirObjs(self.job_order, trim_listing)
549         visit_class(self.job_order, ("File", "Directory"), trim_anonymous_location)
550         visit_class(self.job_order, ("File", "Directory"), remove_redundant_fields)
551
552         secret_mounts = {}
553         for param in sorted(self.job_order.keys()):
554             if self.secret_store.has_secret(self.job_order[param]):
555                 mnt = "/secrets/s%d" % len(secret_mounts)
556                 secret_mounts[mnt] = {
557                     "kind": "text",
558                     "content": self.secret_store.retrieve(self.job_order[param])
559                 }
560                 self.job_order[param] = {"$include": mnt}
561
562         container_req = {
563             "name": self.name,
564             "output_path": "/var/spool/cwl",
565             "cwd": "/var/spool/cwl",
566             "priority": self.priority,
567             "state": "Committed",
568             "container_image": arvados_jobs_image(self.arvrunner, self.jobs_image, runtimeContext),
569             "mounts": {
570                 "/var/lib/cwl/cwl.input.json": {
571                     "kind": "json",
572                     "content": self.job_order
573                 },
574                 "stdout": {
575                     "kind": "file",
576                     "path": "/var/spool/cwl/cwl.output.json"
577                 },
578                 "/var/spool/cwl": {
579                     "kind": "collection",
580                     "writable": True
581                 }
582             },
583             "secret_mounts": secret_mounts,
584             "runtime_constraints": {
585                 "vcpus": math.ceil(self.submit_runner_cores),
586                 "ram": 1024*1024 * (math.ceil(self.submit_runner_ram) + math.ceil(self.collection_cache_size)),
587                 "API": True
588             },
589             "use_existing": False, # Never reuse the runner container - see #15497.
590             "properties": {}
591         }
592
593         if self.embedded_tool.tool.get("id", "").startswith("keep:"):
594             sp = self.embedded_tool.tool["id"].split('/')
595             workflowcollection = sp[0][5:]
596             workflowname = "/".join(sp[1:])
597             workflowpath = "/var/lib/cwl/workflow/%s" % workflowname
598             container_req["mounts"]["/var/lib/cwl/workflow"] = {
599                 "kind": "collection",
600                 "portable_data_hash": "%s" % workflowcollection
601             }
602         elif self.embedded_tool.tool.get("id", "").startswith("arvwf:"):
603             uuid, frg = urllib.parse.urldefrag(self.embedded_tool.tool["id"])
604             workflowpath = "/var/lib/cwl/workflow.json#" + frg
605             packedtxt = self.loadingContext.loader.fetch_text(uuid)
606             yaml = ruamel.yaml.YAML(typ='safe', pure=True)
607             packed = yaml.load(packedtxt)
608             container_req["mounts"]["/var/lib/cwl/workflow.json"] = {
609                 "kind": "json",
610                 "content": packed
611             }
612             container_req["properties"]["template_uuid"] = self.embedded_tool.tool["id"][6:33]
613         else:
614             main = self.loadingContext.loader.idx["_:main"]
615             if main.get("id") == "_:main":
616                 del main["id"]
617             workflowpath = "/var/lib/cwl/workflow.json#main"
618             container_req["mounts"]["/var/lib/cwl/workflow.json"] = {
619                 "kind": "json",
620                 "content": main
621             }
622
623         container_req["properties"].update({k.replace("http://arvados.org/cwl#", "arv:"): v for k, v in git_info.items()})
624
625         properties_req, _ = self.embedded_tool.get_requirement("http://arvados.org/cwl#ProcessProperties")
626         if properties_req:
627             builder = make_builder(self.job_order, self.embedded_tool.hints, self.embedded_tool.requirements, runtimeContext, self.embedded_tool.metadata)
628             for pr in properties_req["processProperties"]:
629                 container_req["properties"][pr["propertyName"]] = builder.do_eval(pr["propertyValue"])
630
631         # --local means execute the workflow instead of submitting a container request
632         # --api=containers means use the containers API
633         # --no-log-timestamps means don't add timestamps (the logging infrastructure does this)
634         # --disable-validate because we already validated so don't need to do it again
635         # --eval-timeout is the timeout for javascript invocation
636         # --parallel-task-count is the number of threads to use for job submission
637         # --enable/disable-reuse sets desired job reuse
638         # --collection-cache-size sets aside memory to store collections
639         command = ["arvados-cwl-runner",
640                    "--local",
641                    "--api=containers",
642                    "--no-log-timestamps",
643                    "--disable-validate",
644                    "--disable-color",
645                    "--eval-timeout=%s" % self.arvrunner.eval_timeout,
646                    "--thread-count=%s" % self.arvrunner.thread_count,
647                    "--enable-reuse" if self.enable_reuse else "--disable-reuse",
648                    "--collection-cache-size=%s" % self.collection_cache_size]
649
650         if self.output_name:
651             command.append("--output-name=" + self.output_name)
652             container_req["output_name"] = self.output_name
653
654         if self.output_tags:
655             command.append("--output-tags=" + self.output_tags)
656
657         if runtimeContext.debug:
658             command.append("--debug")
659
660         if runtimeContext.storage_classes != "default" and runtimeContext.storage_classes:
661             command.append("--storage-classes=" + runtimeContext.storage_classes)
662
663         if runtimeContext.intermediate_storage_classes != "default" and runtimeContext.intermediate_storage_classes:
664             command.append("--intermediate-storage-classes=" + runtimeContext.intermediate_storage_classes)
665
666         if runtimeContext.on_error:
667             command.append("--on-error=" + self.on_error)
668
669         if runtimeContext.intermediate_output_ttl:
670             command.append("--intermediate-output-ttl=%d" % runtimeContext.intermediate_output_ttl)
671
672         if runtimeContext.trash_intermediate:
673             command.append("--trash-intermediate")
674
675         if runtimeContext.project_uuid:
676             command.append("--project-uuid="+runtimeContext.project_uuid)
677
678         if self.enable_dev:
679             command.append("--enable-dev")
680
681         if runtimeContext.enable_preemptible is True:
682             command.append("--enable-preemptible")
683
684         if runtimeContext.enable_preemptible is False:
685             command.append("--disable-preemptible")
686
687         if runtimeContext.varying_url_params:
688             command.append("--varying-url-params="+runtimeContext.varying_url_params)
689
690         if runtimeContext.prefer_cached_downloads:
691             command.append("--prefer-cached-downloads")
692
693         if self.fast_parser:
694             command.append("--fast-parser")
695
696         command.extend([workflowpath, "/var/lib/cwl/cwl.input.json"])
697
698         container_req["command"] = command
699
700         return container_req
701
702
703     def run(self, runtimeContext):
704         runtimeContext.keepprefix = "keep:"
705         job_spec = self.arvados_job_spec(runtimeContext, self.git_info)
706         if runtimeContext.project_uuid:
707             job_spec["owner_uuid"] = runtimeContext.project_uuid
708
709         extra_submit_params = {}
710         if runtimeContext.submit_runner_cluster:
711             extra_submit_params["cluster_id"] = runtimeContext.submit_runner_cluster
712
713         if runtimeContext.submit_request_uuid:
714             if "cluster_id" in extra_submit_params:
715                 # Doesn't make sense for "update" and actually fails
716                 del extra_submit_params["cluster_id"]
717             response = self.arvrunner.api.container_requests().update(
718                 uuid=runtimeContext.submit_request_uuid,
719                 body=job_spec,
720                 **extra_submit_params
721             ).execute(num_retries=self.arvrunner.num_retries)
722         else:
723             response = self.arvrunner.api.container_requests().create(
724                 body=job_spec,
725                 **extra_submit_params
726             ).execute(num_retries=self.arvrunner.num_retries)
727
728         self.uuid = response["uuid"]
729         self.arvrunner.process_submitted(self)
730
731         logger.info("%s submitted container_request %s", self.arvrunner.label(self), response["uuid"])
732
733         workbench1 = self.arvrunner.api.config()["Services"]["Workbench1"]["ExternalURL"]
734         workbench2 = self.arvrunner.api.config()["Services"]["Workbench2"]["ExternalURL"]
735         url = ""
736         if workbench2:
737             url = "{}processes/{}".format(workbench2, response["uuid"])
738         elif workbench1:
739             url = "{}container_requests/{}".format(workbench1, response["uuid"])
740         if url:
741             logger.info("Monitor workflow progress at %s", url)
742
743
744     def done(self, record):
745         try:
746             container = self.arvrunner.api.containers().get(
747                 uuid=record["container_uuid"]
748             ).execute(num_retries=self.arvrunner.num_retries)
749             container["log"] = record["log_uuid"]
750         except Exception:
751             logger.exception("%s while getting runner container", self.arvrunner.label(self))
752             self.arvrunner.output_callback({}, "permanentFail")
753         else:
754             super(RunnerContainer, self).done(container)