X-Git-Url: https://git.arvados.org/arvados.git/blobdiff_plain/697e73b0605b6c182f1051e97ed370d5afa7d954..c3d4f8a585202ec58df5506934b698039c200b68:/sdk/python/arvados/commands/put.py diff --git a/sdk/python/arvados/commands/put.py b/sdk/python/arvados/commands/put.py index 714281cc95..5b46ba75b7 100644 --- a/sdk/python/arvados/commands/put.py +++ b/sdk/python/arvados/commands/put.py @@ -133,6 +133,15 @@ physical storage devices (e.g., disks) should have a copy of each data block. Default is to use the server-provided default (if any) or 2. """) +upload_opts.add_argument('--threads', type=int, metavar='N', default=None, + help=""" +Set the number of upload threads to be used. Take into account that +using lots of threads will increase the RAM requirements. Default is +to use 2 threads. +On high latency installations, using a greater number will improve +overall throughput. +""") + run_opts = argparse.ArgumentParser(add_help=False) run_opts.add_argument('--project-uuid', metavar='UUID', help=""" @@ -347,8 +356,9 @@ class ArvPutUploadJob(object): def __init__(self, paths, resume=True, use_cache=True, reporter=None, bytes_expected=None, name=None, owner_uuid=None, - ensure_unique_name=False, num_retries=None, replication_desired=None, - filename=None, update_time=20.0, update_collection=None, + ensure_unique_name=False, num_retries=None, + put_threads=None, replication_desired=None, + filename=None, update_time=60.0, update_collection=None, logger=logging.getLogger('arvados.arv_put'), dry_run=False): self.paths = paths self.resume = resume @@ -363,6 +373,7 @@ class ArvPutUploadJob(object): self.ensure_unique_name = ensure_unique_name self.num_retries = num_retries self.replication_desired = replication_desired + self.put_threads = put_threads self.filename = filename self._state_lock = threading.Lock() self._state = None # Previous run state (file list & manifest) @@ -505,14 +516,18 @@ class ArvPutUploadJob(object): with self._collection_lock: self.bytes_written = self._collection_size(self._local_collection) if self.use_cache: + if final: + manifest = self._local_collection.manifest_text() + else: + # Get the manifest text without comitting pending blocks + manifest = self._local_collection.manifest_text(strip=False, + normalize=False, + only_committed=True) # Update cache with self._state_lock: - if final: - self._state['manifest'] = self._local_collection.manifest_text() - else: - # Get the manifest text without comitting pending blocks - self._state['manifest'] = self._local_collection._get_manifest_text(".", strip=False, normalize=False, only_committed=True) - self._save_state() + self._state['manifest'] = manifest + if self.use_cache: + self._save_state() # Call the reporter, if any self.report_progress() @@ -663,7 +678,7 @@ class ArvPutUploadJob(object): # No cache file, set empty state self._state = copy.deepcopy(self.EMPTY_STATE) # Load the previous manifest so we can check if files were modified remotely. - self._local_collection = arvados.collection.Collection(self._state['manifest'], replication_desired=self.replication_desired) + self._local_collection = arvados.collection.Collection(self._state['manifest'], replication_desired=self.replication_desired, put_threads=self.put_threads) def collection_file_paths(self, col, path_prefix='.'): """Return a list of file paths by recursively go through the entire collection `col`""" @@ -851,6 +866,7 @@ def main(arguments=None, stdout=sys.stdout, stderr=sys.stderr): bytes_expected = bytes_expected, num_retries = args.retries, replication_desired = args.replication, + put_threads = args.threads, name = collection_name, owner_uuid = project_uuid, ensure_unique_name = True,