X-Git-Url: https://git.arvados.org/arvados.git/blobdiff_plain/0e57453d2b637a3d105d4e3d67031f3915f9d302..12bfd7a65c6635a882cb2e5e419321db100b9d56:/sdk/python/arvados/commands/put.py diff --git a/sdk/python/arvados/commands/put.py b/sdk/python/arvados/commands/put.py index 496db86ab4..5cb699f49f 100644 --- a/sdk/python/arvados/commands/put.py +++ b/sdk/python/arvados/commands/put.py @@ -167,7 +167,9 @@ def parse_arguments(arguments): args = arg_parser.parse_args(arguments) if len(args.paths) == 0: - args.paths += ['/dev/stdin'] + args.paths = ['-'] + + args.paths = map(lambda x: "-" if x == "/dev/stdin" else x, args.paths) if len(args.paths) != 1 or os.path.isdir(args.paths[0]): if args.filename: @@ -182,9 +184,9 @@ def parse_arguments(arguments): args.progress = True if args.paths == ['-']: - args.paths = ['/dev/stdin'] + args.resume = False if not args.filename: - args.filename = '-' + args.filename = 'stdin' return args @@ -224,6 +226,23 @@ class ResumeCache(object): self.cache_file.seek(0) return json.load(self.cache_file) + def check_cache(self, api_client=None, num_retries=0): + try: + state = self.load() + locator = None + try: + if "_finished_streams" in state and len(state["_finished_streams"]) > 0: + locator = state["_finished_streams"][0][1][0] + elif "_current_stream_locators" in state and len(state["_current_stream_locators"]) > 0: + locator = state["_current_stream_locators"][0] + if locator is not None: + kc = arvados.keep.KeepClient(api_client=api_client) + kc.head(locator, num_retries=num_retries) + except Exception as e: + self.restart() + except (ValueError): + pass + def save(self, data): try: new_cache_fd, new_cache_name = tempfile.mkstemp( @@ -280,7 +299,9 @@ class ArvPutCollectionWriter(arvados.ResumableCollectionWriter): replication=replication) except (TypeError, ValueError, arvados.errors.StaleWriterStateError) as error: - return cls(cache, reporter, bytes_expected, num_retries=num_retries) + return cls(cache, reporter, bytes_expected, + num_retries=num_retries, + replication=replication) else: return writer @@ -434,6 +455,7 @@ def main(arguments=None, stdout=sys.stdout, stderr=sys.stderr): if args.resume: try: resume_cache = ResumeCache(ResumeCache.make_path(args)) + resume_cache.check_cache(api_client=api_client, num_retries=args.retries) except (IOError, OSError, ValueError): pass # Couldn't open cache directory/file. Continue without it. except ResumeCacheConflict: @@ -466,7 +488,16 @@ def main(arguments=None, stdout=sys.stdout, stderr=sys.stderr): writer.report_progress() writer.do_queued_work() # Do work resumed from cache. for path in args.paths: # Copy file data to Keep. - if os.path.isdir(path): + if path == '-': + writer.start_new_stream() + writer.start_new_file(args.filename) + r = sys.stdin.read(64*1024) + while r: + # Need to bypass _queued_file check in ResumableCollectionWriter.write() to get + # CollectionWriter.write(). + super(arvados.collection.ResumableCollectionWriter, writer).write(r) + r = sys.stdin.read(64*1024) + elif os.path.isdir(path): writer.write_directory_tree( path, max_manifest_depth=args.max_manifest_depth) else: @@ -477,6 +508,7 @@ def main(arguments=None, stdout=sys.stdout, stderr=sys.stderr): if args.progress: # Print newline to split stderr from stdout for humans. print >>stderr + output = None if args.stream: output = writer.manifest_text() if args.normalize: @@ -517,9 +549,12 @@ def main(arguments=None, stdout=sys.stdout, stderr=sys.stderr): status = 1 # Print the locator (uuid) of the new collection. - stdout.write(output) - if not output.endswith('\n'): - stdout.write('\n') + if output is None: + status = status or 1 + else: + stdout.write(output) + if not output.endswith('\n'): + stdout.write('\n') for sigcode, orig_handler in orig_signal_handlers.items(): signal.signal(sigcode, orig_handler)