X-Git-Url: https://git.arvados.org/arvados.git/blobdiff_plain/97403e08475b328115373a2c6a23e82116199aad..5e0dd2c6b8f080c81ff6077e629f5ec9f377802f:/services/api/config/application.default.yml diff --git a/services/api/config/application.default.yml b/services/api/config/application.default.yml index e7dbf29c59..d46c3978a6 100644 --- a/services/api/config/application.default.yml +++ b/services/api/config/application.default.yml @@ -220,9 +220,23 @@ common: # a site secret. It should be at least 50 characters. blob_signing_key: ~ - # Amount of time (in seconds) for which a blob permission signature - # remains valid. Default: 2 weeks (1209600 seconds) - blob_signing_ttl: 1209600 + # Lifetime (in seconds) of blob permission signatures generated by + # the API server. This determines how long a client can take (after + # retrieving a collection record) to retrieve the collection data + # from Keep. If the client needs more time than that (assuming the + # collection still has the same content and the relevant user/token + # still has permission) the client can retrieve the collection again + # to get fresh signatures. + # + # Datamanager considers an unreferenced block older than this to be + # eligible for garbage collection. Therefore, it should never be + # smaller than the corresponding value used by any local keepstore + # service (see keepstore -blob-signature-ttl flag). This rule + # prevents datamanager from trying to garbage-collect recently + # written blocks while clients are still holding valid signatures. + # + # The default is 2 weeks. + blob_signature_ttl: 1209600 # Allow clients to create collections by providing a manifest with # unsigned data blob locators. IMPORTANT: This effectively disables @@ -276,7 +290,27 @@ common: # actually enforce the desired maximum request size on the server side. max_request_size: 134217728 + # Stop collecting records for an index request after we read this much + # data (in bytes) from large database columns. + # Currently only `GET /collections` respects this parameter, when the + # user requests an index that includes manifest_text. Once the API + # server collects records with a total manifest_text size at or above + # this amount, it returns those results immediately. + # Note this is a threshold, not a limit. Record collection stops + # *after* reading this much data. + max_index_database_read: 134217728 + # When you run the db:delete_old_job_logs task, it will find jobs that # have been finished for at least this many seconds, and delete their # stderr logs from the logs table. clean_job_log_rows_after: <%= 30.days %> + + # The maximum number of compute nodes that can be in use simultaneously + # If this limit is reduced, any existing nodes with slot number >= new limit + # will not be counted against the new limit. In other words, the new limit + # won't be strictly enforced until those nodes with higher slot numbers + # go down. + max_compute_nodes: 64 + + # Docker image to be used when none found in runtime_constraints of a job + default_docker_image_for_jobs: false