11684: Reverted easy fix to expose the bug: when there's a delay writing a block...
[arvados.git] / sdk / python / arvados / commands / keepdocker.py
index 7e1b170674c5be0054d09c6a5678fe83426a4d9a..22ea4760c94cfb421014085b4ad453bab0be8d0a 100644 (file)
@@ -1,5 +1,4 @@
-#!/usr/bin/env python
-
+from builtins import next
 import argparse
 import collections
 import datetime
@@ -98,7 +97,7 @@ def docker_image_format(image_hash):
     cmd = popen_docker(['inspect', '--format={{.Id}}', image_hash],
                         stdout=subprocess.PIPE)
     try:
-        image_id = next(cmd.stdout).strip()
+        image_id = next(cmd.stdout).decode().strip()
         if image_id.startswith('sha256:'):
             return 'v2'
         elif ':' not in image_id:
@@ -111,8 +110,8 @@ def docker_image_format(image_hash):
 def docker_image_compatible(api, image_hash):
     supported = api._rootDesc.get('dockerImageFormats', [])
     if not supported:
-        logger.warn("server does not specify supported image formats (see docker_image_formats in server config). Continuing.")
-        return True
+        logger.warning("server does not specify supported image formats (see docker_image_formats in server config).")
+        return False
 
     fmt = docker_image_format(image_hash)
     if fmt in supported:
@@ -315,7 +314,7 @@ def list_images_in_arv(api_client, num_retries, image_name=None, image_tag=None)
     # and add image listings for them, retaining the API server preference
     # sorting.
     images_start_size = len(images)
-    for collection_uuid, link in hash_link_map.iteritems():
+    for collection_uuid, link in hash_link_map.items():
         if not seen_image_names[collection_uuid]:
             images.append(_new_image_listing(link, link['name']))
     if len(images) > images_start_size:
@@ -338,85 +337,6 @@ def _uuid2pdh(api, uuid):
         select=['portable_data_hash'],
     ).execute()['items'][0]['portable_data_hash']
 
-_migration_link_class = 'docker_image_migration'
-_migration_link_name = 'migrate_1.9_1.10'
-
-def migrate19():
-    api_client  = arvados.api()
-
-    images = arvados.commands.keepdocker.list_images_in_arv(api_client, 3)
-
-    is_new = lambda img: img['dockerhash'].startswith('sha256:')
-
-    count_new = 0
-    old_images = []
-    for uuid, img in images:
-        if img["dockerhash"].startswith("sha256:"):
-            continue
-        key = (img["repo"], img["tag"], img["timestamp"])
-        old_images.append(img)
-
-    migration_links = arvados.util.list_all(api_client.links().list, filters=[
-        ['link_class', '=', _migration_link_class],
-        ['name', '=', _migration_link_name],
-    ])
-
-    already_migrated = set()
-    for m in migration_links:
-        already_migrated.add(m["tail_uuid"])
-
-    need_migrate = [img for img in old_images if img["collection"] not in already_migrated]
-
-    logger.info("Already migrated %i images", len(already_migrated))
-    logger.info("Need to migrate %i images", len(need_migrate))
-
-    for old_image in need_migrate:
-        logger.info("Migrating %s", old_image["collection"])
-
-        col = CollectionReader(old_image["collection"])
-        tarfile = col.keys()[0]
-
-        try:
-            varlibdocker = tempfile.mkdtemp()
-            with tempfile.NamedTemporaryFile() as envfile:
-                envfile.write("ARVADOS_API_HOST=%s\n" % (os.environ["ARVADOS_API_HOST"]))
-                envfile.write("ARVADOS_API_TOKEN=%s\n" % (os.environ["ARVADOS_API_TOKEN"]))
-                envfile.write("ARVADOS_API_HOST_INSECURE=%s\n" % (os.environ["ARVADOS_API_HOST_INSECURE"]))
-                envfile.flush()
-
-                dockercmd = ["docker", "run",
-                             "--privileged",
-                             "--rm",
-                             "--env-file", envfile.name,
-                             "--volume", "%s:/var/lib/docker" % varlibdocker,
-                             "arvados/docker19-migrate",
-                             "/root/migrate.sh",
-                             "%s/%s" % (old_image["collection"], tarfile),
-                             tarfile[0:40],
-                             old_image["repo"],
-                             old_image["tag"],
-                             col.api_response()["owner_uuid"]]
-
-                out = subprocess.check_output(dockercmd)
-
-            new_collection = re.search(r"Migrated uuid is ([a-z0-9]{5}-[a-z0-9]{5}-[a-z0-9]{15})", out)
-            api_client.links().create(body={"link": {
-                'owner_uuid': col.api_response()["owner_uuid"],
-                'link_class': arvados.commands.keepdocker._migration_link_class,
-                'name': arvados.commands.keepdocker._migration_link_name,
-                'tail_uuid': old_image["collection"],
-                'head_uuid': new_collection.group(1)
-                }}).execute(num_retries=3)
-
-            logger.info("Migrated '%s' to '%s'", old_image["collection"], new_collection.group(1))
-        except Exception as e:
-            logger.exception("Migration failed")
-        finally:
-            shutil.rmtree(varlibdocker)
-
-    logger.info("All done")
-
-
 def main(arguments=None, stdout=sys.stdout):
     args = arg_parser.parse_args(arguments)
     api = arvados.api('v1')
@@ -424,8 +344,14 @@ def main(arguments=None, stdout=sys.stdout):
     if args.image is None or args.image == 'images':
         fmt = "{:30}  {:10}  {:12}  {:29}  {:20}\n"
         stdout.write(fmt.format("REPOSITORY", "TAG", "IMAGE ID", "COLLECTION", "CREATED"))
-        for i, j in list_images_in_arv(api, args.retries):
-            stdout.write(fmt.format(j["repo"], j["tag"], j["dockerhash"][0:12], i, j["timestamp"].strftime("%c")))
+        try:
+            for i, j in list_images_in_arv(api, args.retries):
+                stdout.write(fmt.format(j["repo"], j["tag"], j["dockerhash"][0:12], i, j["timestamp"].strftime("%c")))
+        except IOError as e:
+            if e.errno == errno.EPIPE:
+                pass
+            else:
+                raise
         sys.exit(0)
 
     # Pull the image if requested, unless the image is specified as a hash
@@ -441,7 +367,7 @@ def main(arguments=None, stdout=sys.stdout):
 
     if not docker_image_compatible(api, image_hash):
         if args.force_image_format:
-            logger.warn("forcing incompatible image")
+            logger.warning("forcing incompatible image")
         else:
             logger.error("refusing to store " \
                 "incompatible format (use --force-image-format to override)")
@@ -486,7 +412,7 @@ def main(arguments=None, stdout=sys.stdout):
                         api, args.retries,
                         filters=[['link_class', '=', 'docker_image_repo+tag'],
                                  ['name', '=', image_repo_tag],
-                                 ['head_uuid', 'in', collections]])
+                                 ['head_uuid', 'in', [c["uuid"] for c in collections]]])
                 else:
                     existing_repo_tag = []