11684: Reverted easy fix to expose the bug: when there's a delay writing a block...
[arvados.git] / sdk / python / arvados / commands / keepdocker.py
index 135edd4810b2205518363bcbb34c515459d9df02..22ea4760c94cfb421014085b4ad453bab0be8d0a 100644 (file)
@@ -1,5 +1,4 @@
-#!/usr/bin/env python
-
+from builtins import next
 import argparse
 import collections
 import datetime
@@ -98,7 +97,7 @@ def docker_image_format(image_hash):
     cmd = popen_docker(['inspect', '--format={{.Id}}', image_hash],
                         stdout=subprocess.PIPE)
     try:
-        image_id = next(cmd.stdout).strip()
+        image_id = next(cmd.stdout).decode().strip()
         if image_id.startswith('sha256:'):
             return 'v2'
         elif ':' not in image_id:
@@ -111,8 +110,8 @@ def docker_image_format(image_hash):
 def docker_image_compatible(api, image_hash):
     supported = api._rootDesc.get('dockerImageFormats', [])
     if not supported:
-        logger.warn("server does not specify supported image formats (see docker_image_formats in server config). Continuing.")
-        return True
+        logger.warning("server does not specify supported image formats (see docker_image_formats in server config).")
+        return False
 
     fmt = docker_image_format(image_hash)
     if fmt in supported:
@@ -315,7 +314,7 @@ def list_images_in_arv(api_client, num_retries, image_name=None, image_tag=None)
     # and add image listings for them, retaining the API server preference
     # sorting.
     images_start_size = len(images)
-    for collection_uuid, link in hash_link_map.iteritems():
+    for collection_uuid, link in hash_link_map.items():
         if not seen_image_names[collection_uuid]:
             images.append(_new_image_listing(link, link['name']))
     if len(images) > images_start_size:
@@ -338,135 +337,6 @@ def _uuid2pdh(api, uuid):
         select=['portable_data_hash'],
     ).execute()['items'][0]['portable_data_hash']
 
-_migration_link_class = 'docker_image_migration'
-_migration_link_name = 'migrate_1.9_1.10'
-
-def migrate19():
-    """Docker image format migration tool for Arvados.
-
-    This converts Docker images stored in Arvados from image format v1
-    (Docker <= 1.9) to image format v2 (Docker >= 1.10).
-
-    Requires Docker running on the local host.
-
-    Usage:
-
-    1) Run arvados/docker/migrate-docker19/build.sh to create
-    arvados/migrate-docker19 Docker image.
-
-    2) Set ARVADOS_API_HOST and ARVADOS_API_TOKEN to the cluster you want to migrate.
-
-    3) Run arv-migrate-docker19 from the Arvados Python SDK on the host (not in a container).
-
-    This will query Arvados for v1 format Docker images.  For each image that
-    does not already have a corresponding v2 format image (as indicated by a
-    docker_image_migration tag) it will perform the following process:
-
-    i) download the image from Arvados
-    ii) load it into Docker
-    iii) update the Docker version, which updates the image
-    iv) save the v2 format image and upload to Arvados
-    v) create a migration link
-
-    """
-
-    api_client  = arvados.api()
-
-    images = arvados.commands.keepdocker.list_images_in_arv(api_client, 3)
-
-    is_new = lambda img: img['dockerhash'].startswith('sha256:')
-
-    count_new = 0
-    old_images = []
-    for uuid, img in images:
-        if img["dockerhash"].startswith("sha256:"):
-            continue
-        key = (img["repo"], img["tag"], img["timestamp"])
-        old_images.append(img)
-
-    migration_links = arvados.util.list_all(api_client.links().list, filters=[
-        ['link_class', '=', _migration_link_class],
-        ['name', '=', _migration_link_name],
-    ])
-
-    already_migrated = set()
-    for m in migration_links:
-        already_migrated.add(m["tail_uuid"])
-
-    items = arvados.util.list_all(api_client.collections().list,
-                                  filters=[["uuid", "in", [img["collection"] for img in old_images]]],
-                                  select=["uuid", "portable_data_hash"])
-    uuid_to_pdh = {i["uuid"]: i["portable_data_hash"] for i in items}
-    need_migrate = [img for img in old_images
-                    if uuid_to_pdh[img["collection"]] not in already_migrated]
-
-    logger.info("Already migrated %i images", len(already_migrated))
-    logger.info("Need to migrate %i images", len(need_migrate))
-
-    success = []
-    failures = []
-    for old_image in need_migrate:
-        if uuid_to_pdh[old_image["collection"]] in already_migrated:
-            continue
-
-        logger.info("Migrating %s:%s (%s)", old_image["repo"], old_image["tag"], old_image["collection"])
-
-        oldcol = CollectionReader(old_image["collection"])
-        tarfile = oldcol.keys()[0]
-
-        try:
-            varlibdocker = tempfile.mkdtemp()
-            with tempfile.NamedTemporaryFile() as envfile:
-                envfile.write("ARVADOS_API_HOST=%s\n" % (os.environ["ARVADOS_API_HOST"]))
-                envfile.write("ARVADOS_API_TOKEN=%s\n" % (os.environ["ARVADOS_API_TOKEN"]))
-                if "ARVADOS_API_HOST_INSECURE" in os.environ:
-                    envfile.write("ARVADOS_API_HOST_INSECURE=%s\n" % (os.environ["ARVADOS_API_HOST_INSECURE"]))
-                envfile.flush()
-
-                dockercmd = ["docker", "run",
-                             "--privileged",
-                             "--rm",
-                             "--env-file", envfile.name,
-                             "--volume", "%s:/var/lib/docker" % varlibdocker,
-                             "arvados/migrate-docker19",
-                             "/root/migrate.sh",
-                             "%s/%s" % (old_image["collection"], tarfile),
-                             tarfile[0:40],
-                             old_image["repo"],
-                             old_image["tag"],
-                             oldcol.api_response()["owner_uuid"]]
-
-                out = subprocess.check_output(dockercmd)
-
-            migrated = re.search(r"Migrated uuid is ([a-z0-9]{5}-[a-z0-9]{5}-[a-z0-9]{15})", out)
-            if migrated:
-                newcol = CollectionReader(migrated.group(1))
-
-                api_client.links().create(body={"link": {
-                    'owner_uuid': oldcol.api_response()["owner_uuid"],
-                    'link_class': arvados.commands.keepdocker._migration_link_class,
-                    'name': arvados.commands.keepdocker._migration_link_name,
-                    'tail_uuid': oldcol.portable_data_hash(),
-                    'head_uuid': newcol.portable_data_hash()
-                    }}).execute(num_retries=3)
-
-                logger.info("Migrated '%s' to '%s'", oldcol.portable_data_hash(), newcol.portable_data_hash())
-                already_migrated.add(oldcol.portable_data_hash())
-                success.append(old_image["collection"])
-            else:
-                logger.error("Error migrating '%s'", old_image["collection"])
-                failures.append(old_image["collection"])
-        except Exception as e:
-            logger.exception("Migration failed")
-            failures.append(old_image["collection"])
-        finally:
-            shutil.rmtree(varlibdocker)
-
-    logger.info("Successfully migrated %i images", len(success))
-    if failures:
-        logger.error("Failure migrating images: %s", failures)
-
-
 def main(arguments=None, stdout=sys.stdout):
     args = arg_parser.parse_args(arguments)
     api = arvados.api('v1')
@@ -474,8 +344,14 @@ def main(arguments=None, stdout=sys.stdout):
     if args.image is None or args.image == 'images':
         fmt = "{:30}  {:10}  {:12}  {:29}  {:20}\n"
         stdout.write(fmt.format("REPOSITORY", "TAG", "IMAGE ID", "COLLECTION", "CREATED"))
-        for i, j in list_images_in_arv(api, args.retries):
-            stdout.write(fmt.format(j["repo"], j["tag"], j["dockerhash"][0:12], i, j["timestamp"].strftime("%c")))
+        try:
+            for i, j in list_images_in_arv(api, args.retries):
+                stdout.write(fmt.format(j["repo"], j["tag"], j["dockerhash"][0:12], i, j["timestamp"].strftime("%c")))
+        except IOError as e:
+            if e.errno == errno.EPIPE:
+                pass
+            else:
+                raise
         sys.exit(0)
 
     # Pull the image if requested, unless the image is specified as a hash
@@ -491,7 +367,7 @@ def main(arguments=None, stdout=sys.stdout):
 
     if not docker_image_compatible(api, image_hash):
         if args.force_image_format:
-            logger.warn("forcing incompatible image")
+            logger.warning("forcing incompatible image")
         else:
             logger.error("refusing to store " \
                 "incompatible format (use --force-image-format to override)")
@@ -536,7 +412,7 @@ def main(arguments=None, stdout=sys.stdout):
                         api, args.retries,
                         filters=[['link_class', '=', 'docker_image_repo+tag'],
                                  ['name', '=', image_repo_tag],
-                                 ['head_uuid', 'in', collections]])
+                                 ['head_uuid', 'in', [c["uuid"] for c in collections]]])
                 else:
                     existing_repo_tag = []