Merge branch '21666-provision-test-improvement'
[arvados.git] / services / keep-web / s3_test.go
index a6aab357e301e4b4703f2c72ef4a1a490ab65766..79b3712c6b7a766efeb0b0c66ff2c8024e9771b0 100644 (file)
@@ -2,18 +2,22 @@
 //
 // SPDX-License-Identifier: AGPL-3.0
 
-package main
+package keepweb
 
 import (
        "bytes"
+       "context"
        "crypto/rand"
+       "crypto/sha256"
        "fmt"
        "io/ioutil"
+       "mime"
        "net/http"
        "net/http/httptest"
        "net/url"
        "os"
        "os/exec"
+       "sort"
        "strings"
        "sync"
        "time"
@@ -24,6 +28,10 @@ import (
        "git.arvados.org/arvados.git/sdk/go/keepclient"
        "github.com/AdRoll/goamz/aws"
        "github.com/AdRoll/goamz/s3"
+       aws_aws "github.com/aws/aws-sdk-go/aws"
+       aws_credentials "github.com/aws/aws-sdk-go/aws/credentials"
+       aws_session "github.com/aws/aws-sdk-go/aws/session"
+       aws_s3 "github.com/aws/aws-sdk-go/service/s3"
        check "gopkg.in/check.v1"
 )
 
@@ -33,12 +41,13 @@ type s3stage struct {
        kc         *keepclient.KeepClient
        proj       arvados.Group
        projbucket *s3.Bucket
+       subproj    arvados.Group
        coll       arvados.Collection
        collbucket *s3.Bucket
 }
 
 func (s *IntegrationSuite) s3setup(c *check.C) s3stage {
-       var proj arvados.Group
+       var proj, subproj arvados.Group
        var coll arvados.Collection
        arv := arvados.NewClientFromEnv()
        arv.AuthToken = arvadostest.ActiveToken
@@ -46,14 +55,40 @@ func (s *IntegrationSuite) s3setup(c *check.C) s3stage {
                "group": map[string]interface{}{
                        "group_class": "project",
                        "name":        "keep-web s3 test",
+                       "properties": map[string]interface{}{
+                               "project-properties-key": "project properties value",
+                       },
                },
                "ensure_unique_name": true,
        })
        c.Assert(err, check.IsNil)
+       err = arv.RequestAndDecode(&subproj, "POST", "arvados/v1/groups", nil, map[string]interface{}{
+               "group": map[string]interface{}{
+                       "owner_uuid":  proj.UUID,
+                       "group_class": "project",
+                       "name":        "keep-web s3 test subproject",
+                       "properties": map[string]interface{}{
+                               "subproject_properties_key": "subproject properties value",
+                               "invalid header key":        "this value will not be returned because key contains spaces",
+                       },
+               },
+       })
+       c.Assert(err, check.IsNil)
        err = arv.RequestAndDecode(&coll, "POST", "arvados/v1/collections", nil, map[string]interface{}{"collection": map[string]interface{}{
                "owner_uuid":    proj.UUID,
                "name":          "keep-web s3 test collection",
                "manifest_text": ". d41d8cd98f00b204e9800998ecf8427e+0 0:0:emptyfile\n./emptydir d41d8cd98f00b204e9800998ecf8427e+0 0:0:.\n",
+               "properties": map[string]interface{}{
+                       "string":   "string value",
+                       "array":    []string{"element1", "element2"},
+                       "object":   map[string]interface{}{"key": map[string]interface{}{"key2": "value⛵"}},
+                       "nonascii": "⛵",
+                       "newline":  "foo\r\nX-Bad: header",
+                       // This key cannot be expressed as a MIME
+                       // header key, so it will be silently skipped
+                       // (see "Inject" in PropertiesAsMetadata test)
+                       "a: a\r\nInject": "bogus",
+               },
        }})
        c.Assert(err, check.IsNil)
        ac, err := arvadosclient.New(arv)
@@ -75,8 +110,8 @@ func (s *IntegrationSuite) s3setup(c *check.C) s3stage {
 
        auth := aws.NewAuth(arvadostest.ActiveTokenUUID, arvadostest.ActiveToken, "", time.Now().Add(time.Hour))
        region := aws.Region{
-               Name:       s.testServer.Addr,
-               S3Endpoint: "http://" + s.testServer.Addr,
+               Name:       "zzzzz",
+               S3Endpoint: s.testServer.URL,
        }
        client := s3.New(*auth, region)
        client.Signature = aws.V4Signature
@@ -89,7 +124,8 @@ func (s *IntegrationSuite) s3setup(c *check.C) s3stage {
                        S3:   client,
                        Name: proj.UUID,
                },
-               coll: coll,
+               subproj: subproj,
+               coll:    coll,
                collbucket: &s3.Bucket{
                        S3:   client,
                        Name: coll.UUID,
@@ -202,19 +238,98 @@ func (s *IntegrationSuite) testS3GetObject(c *check.C, bucket *s3.Bucket, prefix
        c.Check(err, check.IsNil)
        c.Check(resp.StatusCode, check.Equals, http.StatusOK)
        c.Check(resp.ContentLength, check.Equals, int64(4))
+
+       // HeadObject with superfluous leading slashes
+       exists, err = bucket.Exists(prefix + "//sailboat.txt")
+       c.Check(err, check.IsNil)
+       c.Check(exists, check.Equals, true)
+}
+
+func (s *IntegrationSuite) checkMetaEquals(c *check.C, hdr http.Header, expect map[string]string) {
+       got := map[string]string{}
+       for hk, hv := range hdr {
+               if k := strings.TrimPrefix(hk, "X-Amz-Meta-"); k != hk && len(hv) == 1 {
+                       got[k] = hv[0]
+               }
+       }
+       c.Check(got, check.DeepEquals, expect)
+}
+
+func (s *IntegrationSuite) TestS3PropertiesAsMetadata(c *check.C) {
+       stage := s.s3setup(c)
+       defer stage.teardown(c)
+
+       expectCollectionTags := map[string]string{
+               "String":   "string value",
+               "Array":    `["element1","element2"]`,
+               "Object":   mime.BEncoding.Encode("UTF-8", `{"key":{"key2":"value⛵"}}`),
+               "Nonascii": "=?UTF-8?b?4pu1?=",
+               "Newline":  mime.BEncoding.Encode("UTF-8", "foo\r\nX-Bad: header"),
+       }
+       expectSubprojectTags := map[string]string{
+               "Subproject_properties_key": "subproject properties value",
+       }
+       expectProjectTags := map[string]string{
+               "Project-Properties-Key": "project properties value",
+       }
+
+       c.Log("HEAD object with metadata from collection")
+       resp, err := stage.collbucket.Head("sailboat.txt", nil)
+       c.Assert(err, check.IsNil)
+       s.checkMetaEquals(c, resp.Header, expectCollectionTags)
+
+       c.Log("GET object with metadata from collection")
+       rdr, hdr, err := stage.collbucket.GetReaderWithHeaders("sailboat.txt")
+       c.Assert(err, check.IsNil)
+       content, err := ioutil.ReadAll(rdr)
+       c.Check(err, check.IsNil)
+       rdr.Close()
+       c.Check(content, check.HasLen, 4)
+       s.checkMetaEquals(c, hdr, expectCollectionTags)
+       c.Check(hdr["Inject"], check.IsNil)
+
+       c.Log("HEAD bucket with metadata from collection")
+       resp, err = stage.collbucket.Head("/", nil)
+       c.Assert(err, check.IsNil)
+       s.checkMetaEquals(c, resp.Header, expectCollectionTags)
+
+       c.Log("HEAD directory placeholder with metadata from collection")
+       resp, err = stage.projbucket.Head("keep-web s3 test collection/", nil)
+       c.Assert(err, check.IsNil)
+       s.checkMetaEquals(c, resp.Header, expectCollectionTags)
+
+       c.Log("HEAD file with metadata from collection")
+       resp, err = stage.projbucket.Head("keep-web s3 test collection/sailboat.txt", nil)
+       c.Assert(err, check.IsNil)
+       s.checkMetaEquals(c, resp.Header, expectCollectionTags)
+
+       c.Log("HEAD directory placeholder with metadata from subproject")
+       resp, err = stage.projbucket.Head("keep-web s3 test subproject/", nil)
+       c.Assert(err, check.IsNil)
+       s.checkMetaEquals(c, resp.Header, expectSubprojectTags)
+
+       c.Log("HEAD bucket with metadata from project")
+       resp, err = stage.projbucket.Head("/", nil)
+       c.Assert(err, check.IsNil)
+       s.checkMetaEquals(c, resp.Header, expectProjectTags)
 }
 
 func (s *IntegrationSuite) TestS3CollectionPutObjectSuccess(c *check.C) {
        stage := s.s3setup(c)
        defer stage.teardown(c)
-       s.testS3PutObjectSuccess(c, stage.collbucket, "")
+       s.testS3PutObjectSuccess(c, stage.collbucket, "", stage.coll.UUID)
 }
 func (s *IntegrationSuite) TestS3ProjectPutObjectSuccess(c *check.C) {
        stage := s.s3setup(c)
        defer stage.teardown(c)
-       s.testS3PutObjectSuccess(c, stage.projbucket, stage.coll.Name+"/")
+       s.testS3PutObjectSuccess(c, stage.projbucket, stage.coll.Name+"/", stage.coll.UUID)
 }
-func (s *IntegrationSuite) testS3PutObjectSuccess(c *check.C, bucket *s3.Bucket, prefix string) {
+func (s *IntegrationSuite) testS3PutObjectSuccess(c *check.C, bucket *s3.Bucket, prefix string, collUUID string) {
+       // We insert a delay between test cases to ensure we exercise
+       // rollover of expired sessions.
+       sleep := time.Second / 100
+       s.handler.Cluster.Collections.WebDAVCache.TTL = arvados.Duration(sleep * 3)
+
        for _, trial := range []struct {
                path        string
                size        int
@@ -228,6 +343,18 @@ func (s *IntegrationSuite) testS3PutObjectSuccess(c *check.C, bucket *s3.Bucket,
                        path:        "newdir/newfile",
                        size:        1 << 26,
                        contentType: "application/octet-stream",
+               }, {
+                       path:        "/aaa",
+                       size:        2,
+                       contentType: "application/octet-stream",
+               }, {
+                       path:        "//bbb",
+                       size:        2,
+                       contentType: "application/octet-stream",
+               }, {
+                       path:        "ccc//",
+                       size:        0,
+                       contentType: "application/x-directory",
                }, {
                        path:        "newdir1/newdir2/newfile",
                        size:        0,
@@ -238,14 +365,20 @@ func (s *IntegrationSuite) testS3PutObjectSuccess(c *check.C, bucket *s3.Bucket,
                        contentType: "application/x-directory",
                },
        } {
+               time.Sleep(sleep)
                c.Logf("=== %v", trial)
 
                objname := prefix + trial.path
 
                _, err := bucket.GetReader(objname)
-               c.Check(err.(*s3.Error).StatusCode, check.Equals, 404)
+               if !c.Check(err, check.NotNil) {
+                       continue
+               }
+               c.Check(err.(*s3.Error).StatusCode, check.Equals, http.StatusNotFound)
                c.Check(err.(*s3.Error).Code, check.Equals, `NoSuchKey`)
-               c.Assert(err, check.ErrorMatches, `The specified key does not exist.`)
+               if !c.Check(err, check.ErrorMatches, `The specified key does not exist.`) {
+                       continue
+               }
 
                buf := make([]byte, trial.size)
                rand.Read(buf)
@@ -254,7 +387,7 @@ func (s *IntegrationSuite) testS3PutObjectSuccess(c *check.C, bucket *s3.Bucket,
                c.Check(err, check.IsNil)
 
                rdr, err := bucket.GetReader(objname)
-               if strings.HasSuffix(trial.path, "/") && !s.testServer.Config.cluster.Collections.S3FolderObjects {
+               if strings.HasSuffix(trial.path, "/") && !s.handler.Cluster.Collections.S3FolderObjects {
                        c.Check(err, check.NotNil)
                        continue
                } else if !c.Check(err, check.IsNil) {
@@ -264,6 +397,14 @@ func (s *IntegrationSuite) testS3PutObjectSuccess(c *check.C, bucket *s3.Bucket,
                c.Check(err, check.IsNil)
                c.Check(buf2, check.HasLen, len(buf))
                c.Check(bytes.Equal(buf, buf2), check.Equals, true)
+
+               // Check that the change is immediately visible via
+               // (non-S3) webdav request.
+               _, resp := s.do("GET", "http://"+collUUID+".keep-web.example/"+trial.path, arvadostest.ActiveTokenV2, nil)
+               c.Check(resp.Code, check.Equals, http.StatusOK)
+               if !strings.HasSuffix(trial.path, "/") {
+                       c.Check(resp.Body.Len(), check.Equals, trial.size)
+               }
        }
 }
 
@@ -304,7 +445,7 @@ func (s *IntegrationSuite) TestS3ProjectPutObjectNotSupported(c *check.C) {
                err = bucket.PutReader(trial.path, bytes.NewReader(buf), int64(len(buf)), trial.contentType, s3.Private, s3.Options{})
                c.Check(err.(*s3.Error).StatusCode, check.Equals, 400)
                c.Check(err.(*s3.Error).Code, check.Equals, `InvalidArgument`)
-               c.Check(err, check.ErrorMatches, `(mkdir "/by_id/zzzzz-j7d0g-[a-z0-9]{15}/newdir2?"|open "/zzzzz-j7d0g-[a-z0-9]{15}/newfile") failed: invalid argument`)
+               c.Check(err, check.ErrorMatches, `(mkdir "/by_id/zzzzz-j7d0g-[a-z0-9]{15}/newdir2?"|open "/zzzzz-j7d0g-[a-z0-9]{15}/newfile") failed: invalid (argument|operation)`)
 
                _, err = bucket.GetReader(trial.path)
                c.Check(err.(*s3.Error).StatusCode, check.Equals, 404)
@@ -324,7 +465,7 @@ func (s *IntegrationSuite) TestS3ProjectDeleteObject(c *check.C) {
        s.testS3DeleteObject(c, stage.projbucket, stage.coll.Name+"/")
 }
 func (s *IntegrationSuite) testS3DeleteObject(c *check.C, bucket *s3.Bucket, prefix string) {
-       s.testServer.Config.cluster.Collections.S3FolderObjects = true
+       s.handler.Cluster.Collections.S3FolderObjects = true
        for _, trial := range []struct {
                path string
        }{
@@ -361,15 +502,7 @@ func (s *IntegrationSuite) TestS3ProjectPutObjectFailure(c *check.C) {
        s.testS3PutObjectFailure(c, stage.projbucket, stage.coll.Name+"/")
 }
 func (s *IntegrationSuite) testS3PutObjectFailure(c *check.C, bucket *s3.Bucket, prefix string) {
-       s.testServer.Config.cluster.Collections.S3FolderObjects = false
-
-       // Can't use V4 signature for these tests, because
-       // double-slash is incorrectly cleaned by the aws.V4Signature,
-       // resulting in a "bad signature" error. (Cleaning the path is
-       // appropriate for other services, but not in S3 where object
-       // names "foo//bar" and "foo/bar" are semantically different.)
-       bucket.S3.Auth = *(aws.NewAuth(arvadostest.ActiveToken, "none", "", time.Now().Add(time.Hour)))
-       bucket.S3.Signature = aws.V2Signature
+       s.handler.Cluster.Collections.S3FolderObjects = false
 
        var wg sync.WaitGroup
        for _, trial := range []struct {
@@ -393,8 +526,6 @@ func (s *IntegrationSuite) testS3PutObjectFailure(c *check.C, bucket *s3.Bucket,
                        path: "/",
                }, {
                        path: "//",
-               }, {
-                       path: "foo//bar",
                }, {
                        path: "",
                },
@@ -441,6 +572,17 @@ func (stage *s3stage) writeBigDirs(c *check.C, dirs int, filesPerDir int) {
        c.Assert(fs.Sync(), check.IsNil)
 }
 
+func (s *IntegrationSuite) sign(c *check.C, req *http.Request, key, secret string) {
+       scope := "20200202/zzzzz/service/aws4_request"
+       signedHeaders := "date"
+       req.Header.Set("Date", time.Now().UTC().Format(time.RFC1123))
+       stringToSign, err := s3stringToSign(s3SignAlgorithm, scope, signedHeaders, req)
+       c.Assert(err, check.IsNil)
+       sig, err := s3signature(secret, scope, signedHeaders, stringToSign)
+       c.Assert(err, check.IsNil)
+       req.Header.Set("Authorization", s3SignAlgorithm+" Credential="+key+"/"+scope+", SignedHeaders="+signedHeaders+", Signature="+sig)
+}
+
 func (s *IntegrationSuite) TestS3VirtualHostStyleRequests(c *check.C) {
        stage := s.s3setup(c)
        defer stage.teardown(c)
@@ -487,14 +629,31 @@ func (s *IntegrationSuite) TestS3VirtualHostStyleRequests(c *check.C) {
                        responseCode:   http.StatusOK,
                        responseRegexp: []string{`boop`},
                },
+               {
+                       url:          "https://" + stage.projbucket.Name + ".example.com/" + stage.coll.Name + "//boop",
+                       method:       "GET",
+                       responseCode: http.StatusNotFound,
+               },
+               {
+                       url:          "https://" + stage.projbucket.Name + ".example.com/" + stage.coll.Name + "//boop",
+                       method:       "PUT",
+                       body:         "boop",
+                       responseCode: http.StatusOK,
+               },
+               {
+                       url:            "https://" + stage.projbucket.Name + ".example.com/" + stage.coll.Name + "//boop",
+                       method:         "GET",
+                       responseCode:   http.StatusOK,
+                       responseRegexp: []string{`boop`},
+               },
        } {
                url, err := url.Parse(trial.url)
                c.Assert(err, check.IsNil)
                req, err := http.NewRequest(trial.method, url.String(), bytes.NewReader([]byte(trial.body)))
                c.Assert(err, check.IsNil)
-               req.Header.Set("Authorization", "AWS "+arvadostest.ActiveTokenV2+":none")
+               s.sign(c, req, arvadostest.ActiveTokenUUID, arvadostest.ActiveToken)
                rr := httptest.NewRecorder()
-               s.testServer.Server.Handler.ServeHTTP(rr, req)
+               s.handler.ServeHTTP(rr, req)
                resp := rr.Result()
                c.Check(resp.StatusCode, check.Equals, trial.responseCode)
                body, err := ioutil.ReadAll(resp.Body)
@@ -505,6 +664,59 @@ func (s *IntegrationSuite) TestS3VirtualHostStyleRequests(c *check.C) {
        }
 }
 
+func (s *IntegrationSuite) TestS3NormalizeURIForSignature(c *check.C) {
+       stage := s.s3setup(c)
+       defer stage.teardown(c)
+       for _, trial := range []struct {
+               rawPath        string
+               normalizedPath string
+       }{
+               {"/foo", "/foo"},                           // boring case
+               {"/foo%5fbar", "/foo_bar"},                 // _ must not be escaped
+               {"/foo%2fbar", "/foo/bar"},                 // / must not be escaped
+               {"/(foo)/[];,", "/%28foo%29/%5B%5D%3B%2C"}, // ()[];, must be escaped
+               {"/foo%5bbar", "/foo%5Bbar"},               // %XX must be uppercase
+               {"//foo///.bar", "/foo/.bar"},              // "//" and "///" must be squashed to "/"
+       } {
+               c.Logf("trial %q", trial)
+
+               date := time.Now().UTC().Format("20060102T150405Z")
+               scope := "20200202/zzzzz/S3/aws4_request"
+               canonicalRequest := fmt.Sprintf("%s\n%s\n%s\n%s\n%s\n%s", "GET", trial.normalizedPath, "", "host:host.example.com\n", "host", "")
+               c.Logf("canonicalRequest %q", canonicalRequest)
+               expect := fmt.Sprintf("%s\n%s\n%s\n%s", s3SignAlgorithm, date, scope, hashdigest(sha256.New(), canonicalRequest))
+               c.Logf("expected stringToSign %q", expect)
+
+               req, err := http.NewRequest("GET", "https://host.example.com"+trial.rawPath, nil)
+               req.Header.Set("X-Amz-Date", date)
+               req.Host = "host.example.com"
+               c.Assert(err, check.IsNil)
+
+               obtained, err := s3stringToSign(s3SignAlgorithm, scope, "host", req)
+               if !c.Check(err, check.IsNil) {
+                       continue
+               }
+               c.Check(obtained, check.Equals, expect)
+       }
+}
+
+func (s *IntegrationSuite) TestS3GetBucketLocation(c *check.C) {
+       stage := s.s3setup(c)
+       defer stage.teardown(c)
+       for _, bucket := range []*s3.Bucket{stage.collbucket, stage.projbucket} {
+               req, err := http.NewRequest("GET", bucket.URL("/"), nil)
+               c.Check(err, check.IsNil)
+               req.Header.Set("Authorization", "AWS "+arvadostest.ActiveTokenV2+":none")
+               req.URL.RawQuery = "location"
+               resp, err := http.DefaultClient.Do(req)
+               c.Assert(err, check.IsNil)
+               c.Check(resp.Header.Get("Content-Type"), check.Equals, "application/xml")
+               buf, err := ioutil.ReadAll(resp.Body)
+               c.Assert(err, check.IsNil)
+               c.Check(string(buf), check.Equals, "<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n<LocationConstraint><LocationConstraint xmlns=\"http://s3.amazonaws.com/doc/2006-03-01/\">zzzzz</LocationConstraint></LocationConstraint>\n")
+       }
+}
+
 func (s *IntegrationSuite) TestS3GetBucketVersioning(c *check.C) {
        stage := s.s3setup(c)
        defer stage.teardown(c)
@@ -522,6 +734,37 @@ func (s *IntegrationSuite) TestS3GetBucketVersioning(c *check.C) {
        }
 }
 
+func (s *IntegrationSuite) TestS3UnsupportedAPIs(c *check.C) {
+       stage := s.s3setup(c)
+       defer stage.teardown(c)
+       for _, trial := range []struct {
+               method   string
+               path     string
+               rawquery string
+       }{
+               {"GET", "/", "acl&versionId=1234"},    // GetBucketAcl
+               {"GET", "/foo", "acl&versionId=1234"}, // GetObjectAcl
+               {"PUT", "/", "acl"},                   // PutBucketAcl
+               {"PUT", "/foo", "acl"},                // PutObjectAcl
+               {"DELETE", "/", "tagging"},            // DeleteBucketTagging
+               {"DELETE", "/foo", "tagging"},         // DeleteObjectTagging
+       } {
+               for _, bucket := range []*s3.Bucket{stage.collbucket, stage.projbucket} {
+                       c.Logf("trial %v bucket %v", trial, bucket)
+                       req, err := http.NewRequest(trial.method, bucket.URL(trial.path), nil)
+                       c.Check(err, check.IsNil)
+                       req.Header.Set("Authorization", "AWS "+arvadostest.ActiveTokenV2+":none")
+                       req.URL.RawQuery = trial.rawquery
+                       resp, err := http.DefaultClient.Do(req)
+                       c.Assert(err, check.IsNil)
+                       c.Check(resp.Header.Get("Content-Type"), check.Equals, "application/xml")
+                       buf, err := ioutil.ReadAll(resp.Body)
+                       c.Assert(err, check.IsNil)
+                       c.Check(string(buf), check.Matches, "(?ms).*InvalidRequest.*API not supported.*")
+               }
+       }
+}
+
 // If there are no CommonPrefixes entries, the CommonPrefixes XML tag
 // should not appear at all.
 func (s *IntegrationSuite) TestS3ListNoCommonPrefixes(c *check.C) {
@@ -580,9 +823,9 @@ func (s *IntegrationSuite) TestS3CollectionList(c *check.C) {
        defer stage.teardown(c)
 
        var markers int
-       for markers, s.testServer.Config.cluster.Collections.S3FolderObjects = range []bool{false, true} {
-               dirs := 2
-               filesPerDir := 1001
+       for markers, s.handler.Cluster.Collections.S3FolderObjects = range []bool{false, true} {
+               dirs := 2000
+               filesPerDir := 2
                stage.writeBigDirs(c, dirs, filesPerDir)
                // Total # objects is:
                //                 2 file entries from s3setup (emptyfile and sailboat.txt)
@@ -591,11 +834,12 @@ func (s *IntegrationSuite) TestS3CollectionList(c *check.C) {
                // +filesPerDir*dirs file entries from writeBigDirs (dir0/file0.txt, etc.)
                s.testS3List(c, stage.collbucket, "", 4000, markers+2+(filesPerDir+markers)*dirs)
                s.testS3List(c, stage.collbucket, "", 131, markers+2+(filesPerDir+markers)*dirs)
+               s.testS3List(c, stage.collbucket, "", 51, markers+2+(filesPerDir+markers)*dirs)
                s.testS3List(c, stage.collbucket, "dir0/", 71, filesPerDir+markers)
        }
 }
 func (s *IntegrationSuite) testS3List(c *check.C, bucket *s3.Bucket, prefix string, pageSize, expectFiles int) {
-       c.Logf("testS3List: prefix=%q pageSize=%d S3FolderObjects=%v", prefix, pageSize, s.testServer.Config.cluster.Collections.S3FolderObjects)
+       c.Logf("testS3List: prefix=%q pageSize=%d S3FolderObjects=%v", prefix, pageSize, s.handler.Cluster.Collections.S3FolderObjects)
        expectPageSize := pageSize
        if expectPageSize > 1000 {
                expectPageSize = 1000
@@ -613,6 +857,9 @@ func (s *IntegrationSuite) testS3List(c *check.C, bucket *s3.Bucket, prefix stri
                        break
                }
                for _, key := range resp.Contents {
+                       if _, dup := gotKeys[key.Key]; dup {
+                               c.Errorf("got duplicate key %q on page %d", key.Key, pages)
+                       }
                        gotKeys[key.Key] = key
                        if strings.Contains(key.Key, "sailboat.txt") {
                                c.Check(key.Size, check.Equals, int64(4))
@@ -627,11 +874,20 @@ func (s *IntegrationSuite) testS3List(c *check.C, bucket *s3.Bucket, prefix stri
                }
                nextMarker = resp.NextMarker
        }
-       c.Check(len(gotKeys), check.Equals, expectFiles)
+       if !c.Check(len(gotKeys), check.Equals, expectFiles) {
+               var sorted []string
+               for k := range gotKeys {
+                       sorted = append(sorted, k)
+               }
+               sort.Strings(sorted)
+               for _, k := range sorted {
+                       c.Logf("got %s", k)
+               }
+       }
 }
 
 func (s *IntegrationSuite) TestS3CollectionListRollup(c *check.C) {
-       for _, s.testServer.Config.cluster.Collections.S3FolderObjects = range []bool{false, true} {
+       for _, s.handler.Cluster.Collections.S3FolderObjects = range []bool{false, true} {
                s.testS3CollectionListRollup(c)
        }
 }
@@ -660,7 +916,7 @@ func (s *IntegrationSuite) testS3CollectionListRollup(c *check.C) {
                }
        }
        markers := 0
-       if s.testServer.Config.cluster.Collections.S3FolderObjects {
+       if s.handler.Cluster.Collections.S3FolderObjects {
                markers = 1
        }
        c.Check(allfiles, check.HasLen, dirs*(filesPerDir+markers)+3+markers)
@@ -693,7 +949,8 @@ func (s *IntegrationSuite) testS3CollectionListRollup(c *check.C) {
                {"dir0", "", ""},
                {"dir0/", "", ""},
                {"dir0/f", "", ""},
-               {"dir0", "/", "dir0/file14.txt"},       // no commonprefixes
+               {"dir0", "/", "dir0/file14.txt"},       // one commonprefix, "dir0/"
+               {"dir0", "/", "dir0/zzzzfile.txt"},     // no commonprefixes
                {"", "", "dir0/file14.txt"},            // middle page, skip walking dir1
                {"", "", "dir1/file14.txt"},            // middle page, skip walking dir0
                {"", "", "dir1/file498.txt"},           // last page of results
@@ -724,28 +981,31 @@ func (s *IntegrationSuite) testS3CollectionListRollup(c *check.C) {
                var expectTruncated bool
                for _, key := range allfiles {
                        full := len(expectKeys)+len(expectPrefixes) >= maxKeys
-                       if !strings.HasPrefix(key, trial.prefix) || key < trial.marker {
+                       if !strings.HasPrefix(key, trial.prefix) || key <= trial.marker {
                                continue
                        } else if idx := strings.Index(key[len(trial.prefix):], trial.delimiter); trial.delimiter != "" && idx >= 0 {
                                prefix := key[:len(trial.prefix)+idx+1]
                                if len(expectPrefixes) > 0 && expectPrefixes[len(expectPrefixes)-1] == prefix {
                                        // same prefix as previous key
                                } else if full {
-                                       expectNextMarker = key
                                        expectTruncated = true
                                } else {
                                        expectPrefixes = append(expectPrefixes, prefix)
+                                       expectNextMarker = prefix
                                }
                        } else if full {
-                               if trial.delimiter != "" {
-                                       expectNextMarker = key
-                               }
                                expectTruncated = true
                                break
                        } else {
                                expectKeys = append(expectKeys, key)
+                               if trial.delimiter != "" {
+                                       expectNextMarker = key
+                               }
                        }
                }
+               if !expectTruncated {
+                       expectNextMarker = ""
+               }
 
                var gotKeys []string
                for _, key := range resp.Contents {
@@ -764,6 +1024,251 @@ func (s *IntegrationSuite) testS3CollectionListRollup(c *check.C) {
        }
 }
 
+func (s *IntegrationSuite) TestS3ListObjectsV2ManySubprojects(c *check.C) {
+       stage := s.s3setup(c)
+       defer stage.teardown(c)
+       projects := 50
+       collectionsPerProject := 2
+       for i := 0; i < projects; i++ {
+               var subproj arvados.Group
+               err := stage.arv.RequestAndDecode(&subproj, "POST", "arvados/v1/groups", nil, map[string]interface{}{
+                       "group": map[string]interface{}{
+                               "owner_uuid":  stage.subproj.UUID,
+                               "group_class": "project",
+                               "name":        fmt.Sprintf("keep-web s3 test subproject %d", i),
+                       },
+               })
+               c.Assert(err, check.IsNil)
+               for j := 0; j < collectionsPerProject; j++ {
+                       err = stage.arv.RequestAndDecode(nil, "POST", "arvados/v1/collections", nil, map[string]interface{}{"collection": map[string]interface{}{
+                               "owner_uuid":    subproj.UUID,
+                               "name":          fmt.Sprintf("keep-web s3 test collection %d", j),
+                               "manifest_text": ". d41d8cd98f00b204e9800998ecf8427e+0 0:0:emptyfile\n./emptydir d41d8cd98f00b204e9800998ecf8427e+0 0:0:.\n",
+                       }})
+                       c.Assert(err, check.IsNil)
+               }
+       }
+       c.Logf("setup complete")
+
+       sess := aws_session.Must(aws_session.NewSession(&aws_aws.Config{
+               Region:           aws_aws.String("auto"),
+               Endpoint:         aws_aws.String(s.testServer.URL),
+               Credentials:      aws_credentials.NewStaticCredentials(url.QueryEscape(arvadostest.ActiveTokenV2), url.QueryEscape(arvadostest.ActiveTokenV2), ""),
+               S3ForcePathStyle: aws_aws.Bool(true),
+       }))
+       client := aws_s3.New(sess)
+       ctx := context.Background()
+       params := aws_s3.ListObjectsV2Input{
+               Bucket:    aws_aws.String(stage.proj.UUID),
+               Delimiter: aws_aws.String("/"),
+               Prefix:    aws_aws.String("keep-web s3 test subproject/"),
+               MaxKeys:   aws_aws.Int64(int64(projects / 2)),
+       }
+       for page := 1; ; page++ {
+               t0 := time.Now()
+               result, err := client.ListObjectsV2WithContext(ctx, &params)
+               if !c.Check(err, check.IsNil) {
+                       break
+               }
+               c.Logf("got page %d in %v with len(Contents) == %d, len(CommonPrefixes) == %d", page, time.Since(t0), len(result.Contents), len(result.CommonPrefixes))
+               if !*result.IsTruncated {
+                       break
+               }
+               params.ContinuationToken = result.NextContinuationToken
+               *params.MaxKeys = *params.MaxKeys/2 + 1
+       }
+}
+
+func (s *IntegrationSuite) TestS3ListObjectsV2(c *check.C) {
+       stage := s.s3setup(c)
+       defer stage.teardown(c)
+       dirs := 2
+       filesPerDir := 40
+       stage.writeBigDirs(c, dirs, filesPerDir)
+
+       sess := aws_session.Must(aws_session.NewSession(&aws_aws.Config{
+               Region:           aws_aws.String("auto"),
+               Endpoint:         aws_aws.String(s.testServer.URL),
+               Credentials:      aws_credentials.NewStaticCredentials(url.QueryEscape(arvadostest.ActiveTokenV2), url.QueryEscape(arvadostest.ActiveTokenV2), ""),
+               S3ForcePathStyle: aws_aws.Bool(true),
+       }))
+
+       stringOrNil := func(s string) *string {
+               if s == "" {
+                       return nil
+               } else {
+                       return &s
+               }
+       }
+
+       client := aws_s3.New(sess)
+       ctx := context.Background()
+
+       for _, trial := range []struct {
+               prefix               string
+               delimiter            string
+               startAfter           string
+               maxKeys              int
+               expectKeys           int
+               expectCommonPrefixes map[string]bool
+       }{
+               {
+                       // Expect {filesPerDir plus the dir itself}
+                       // for each dir, plus emptydir, emptyfile, and
+                       // sailboat.txt.
+                       expectKeys: (filesPerDir+1)*dirs + 3,
+               },
+               {
+                       maxKeys:    15,
+                       expectKeys: (filesPerDir+1)*dirs + 3,
+               },
+               {
+                       startAfter: "dir0/z",
+                       maxKeys:    15,
+                       // Expect {filesPerDir plus the dir itself}
+                       // for each dir except dir0, plus emptydir,
+                       // emptyfile, and sailboat.txt.
+                       expectKeys: (filesPerDir+1)*(dirs-1) + 3,
+               },
+               {
+                       maxKeys:              1,
+                       delimiter:            "/",
+                       expectKeys:           2, // emptyfile, sailboat.txt
+                       expectCommonPrefixes: map[string]bool{"dir0/": true, "dir1/": true, "emptydir/": true},
+               },
+               {
+                       startAfter:           "dir0/z",
+                       maxKeys:              15,
+                       delimiter:            "/",
+                       expectKeys:           2, // emptyfile, sailboat.txt
+                       expectCommonPrefixes: map[string]bool{"dir1/": true, "emptydir/": true},
+               },
+               {
+                       startAfter:           "dir0/file10.txt",
+                       maxKeys:              15,
+                       delimiter:            "/",
+                       expectKeys:           2,
+                       expectCommonPrefixes: map[string]bool{"dir0/": true, "dir1/": true, "emptydir/": true},
+               },
+               {
+                       startAfter:           "dir0/file10.txt",
+                       maxKeys:              15,
+                       prefix:               "d",
+                       delimiter:            "/",
+                       expectKeys:           0,
+                       expectCommonPrefixes: map[string]bool{"dir0/": true, "dir1/": true},
+               },
+       } {
+               c.Logf("[trial %+v]", trial)
+               params := aws_s3.ListObjectsV2Input{
+                       Bucket:     aws_aws.String(stage.collbucket.Name),
+                       Prefix:     stringOrNil(trial.prefix),
+                       Delimiter:  stringOrNil(trial.delimiter),
+                       StartAfter: stringOrNil(trial.startAfter),
+                       MaxKeys:    aws_aws.Int64(int64(trial.maxKeys)),
+               }
+               keySeen := map[string]bool{}
+               prefixSeen := map[string]bool{}
+               for {
+                       result, err := client.ListObjectsV2WithContext(ctx, &params)
+                       if !c.Check(err, check.IsNil) {
+                               break
+                       }
+                       c.Check(result.Name, check.DeepEquals, aws_aws.String(stage.collbucket.Name))
+                       c.Check(result.Prefix, check.DeepEquals, aws_aws.String(trial.prefix))
+                       c.Check(result.Delimiter, check.DeepEquals, aws_aws.String(trial.delimiter))
+                       // The following two fields are expected to be
+                       // nil (i.e., no tag in XML response) rather
+                       // than "" when the corresponding request
+                       // field was empty or nil.
+                       c.Check(result.StartAfter, check.DeepEquals, stringOrNil(trial.startAfter))
+                       c.Check(result.ContinuationToken, check.DeepEquals, params.ContinuationToken)
+
+                       if trial.maxKeys > 0 {
+                               c.Check(result.MaxKeys, check.DeepEquals, aws_aws.Int64(int64(trial.maxKeys)))
+                               c.Check(len(result.Contents)+len(result.CommonPrefixes) <= trial.maxKeys, check.Equals, true)
+                       } else {
+                               c.Check(result.MaxKeys, check.DeepEquals, aws_aws.Int64(int64(s3MaxKeys)))
+                       }
+
+                       for _, ent := range result.Contents {
+                               c.Assert(ent.Key, check.NotNil)
+                               c.Check(*ent.Key > trial.startAfter, check.Equals, true)
+                               c.Check(keySeen[*ent.Key], check.Equals, false, check.Commentf("dup key %q", *ent.Key))
+                               keySeen[*ent.Key] = true
+                       }
+                       for _, ent := range result.CommonPrefixes {
+                               c.Assert(ent.Prefix, check.NotNil)
+                               c.Check(strings.HasSuffix(*ent.Prefix, trial.delimiter), check.Equals, true, check.Commentf("bad CommonPrefix %q", *ent.Prefix))
+                               if strings.HasPrefix(trial.startAfter, *ent.Prefix) {
+                                       // If we asked for
+                                       // startAfter=dir0/file10.txt,
+                                       // we expect dir0/ to be
+                                       // returned as a common prefix
+                               } else {
+                                       c.Check(*ent.Prefix > trial.startAfter, check.Equals, true)
+                               }
+                               c.Check(prefixSeen[*ent.Prefix], check.Equals, false, check.Commentf("dup common prefix %q", *ent.Prefix))
+                               prefixSeen[*ent.Prefix] = true
+                       }
+                       if *result.IsTruncated && c.Check(result.NextContinuationToken, check.Not(check.Equals), "") {
+                               params.ContinuationToken = aws_aws.String(*result.NextContinuationToken)
+                       } else {
+                               break
+                       }
+               }
+               c.Check(keySeen, check.HasLen, trial.expectKeys)
+               c.Check(prefixSeen, check.HasLen, len(trial.expectCommonPrefixes))
+               if len(trial.expectCommonPrefixes) > 0 {
+                       c.Check(prefixSeen, check.DeepEquals, trial.expectCommonPrefixes)
+               }
+       }
+}
+
+func (s *IntegrationSuite) TestS3ListObjectsV2EncodingTypeURL(c *check.C) {
+       stage := s.s3setup(c)
+       defer stage.teardown(c)
+       dirs := 2
+       filesPerDir := 40
+       stage.writeBigDirs(c, dirs, filesPerDir)
+
+       sess := aws_session.Must(aws_session.NewSession(&aws_aws.Config{
+               Region:           aws_aws.String("auto"),
+               Endpoint:         aws_aws.String(s.testServer.URL),
+               Credentials:      aws_credentials.NewStaticCredentials(url.QueryEscape(arvadostest.ActiveTokenV2), url.QueryEscape(arvadostest.ActiveTokenV2), ""),
+               S3ForcePathStyle: aws_aws.Bool(true),
+       }))
+
+       client := aws_s3.New(sess)
+       ctx := context.Background()
+
+       result, err := client.ListObjectsV2WithContext(ctx, &aws_s3.ListObjectsV2Input{
+               Bucket:       aws_aws.String(stage.collbucket.Name),
+               Prefix:       aws_aws.String("dir0/"),
+               Delimiter:    aws_aws.String("/"),
+               StartAfter:   aws_aws.String("dir0/"),
+               EncodingType: aws_aws.String("url"),
+       })
+       c.Assert(err, check.IsNil)
+       c.Check(*result.Prefix, check.Equals, "dir0%2F")
+       c.Check(*result.Delimiter, check.Equals, "%2F")
+       c.Check(*result.StartAfter, check.Equals, "dir0%2F")
+       for _, ent := range result.Contents {
+               c.Check(*ent.Key, check.Matches, "dir0%2F.*")
+       }
+       result, err = client.ListObjectsV2WithContext(ctx, &aws_s3.ListObjectsV2Input{
+               Bucket:       aws_aws.String(stage.collbucket.Name),
+               Delimiter:    aws_aws.String("/"),
+               EncodingType: aws_aws.String("url"),
+       })
+       c.Assert(err, check.IsNil)
+       c.Check(*result.Delimiter, check.Equals, "%2F")
+       c.Check(result.CommonPrefixes, check.HasLen, dirs+1)
+       for _, ent := range result.CommonPrefixes {
+               c.Check(*ent.Prefix, check.Matches, ".*%2F")
+       }
+}
+
 // TestS3cmd checks compatibility with the s3cmd command line tool, if
 // it's installed. As of Debian buster, s3cmd is only in backports, so
 // `arvados-server install` don't install it, and this test skips if
@@ -777,10 +1282,24 @@ func (s *IntegrationSuite) TestS3cmd(c *check.C) {
        stage := s.s3setup(c)
        defer stage.teardown(c)
 
-       cmd := exec.Command("s3cmd", "--no-ssl", "--host="+s.testServer.Addr, "--host-bucket="+s.testServer.Addr, "--access_key="+arvadostest.ActiveTokenUUID, "--secret_key="+arvadostest.ActiveToken, "ls", "s3://"+arvadostest.FooCollection)
+       cmd := exec.Command("s3cmd", "--no-ssl", "--host="+s.testServer.URL[7:], "--host-bucket="+s.testServer.URL[7:], "--access_key="+arvadostest.ActiveTokenUUID, "--secret_key="+arvadostest.ActiveToken, "ls", "s3://"+arvadostest.FooCollection)
        buf, err := cmd.CombinedOutput()
        c.Check(err, check.IsNil)
        c.Check(string(buf), check.Matches, `.* 3 +s3://`+arvadostest.FooCollection+`/foo\n`)
+
+       // This tests whether s3cmd's path normalization agrees with
+       // keep-web's signature verification wrt chars like "|"
+       // (neither reserved nor unreserved) and "," (not normally
+       // percent-encoded in a path).
+       tmpfile := c.MkDir() + "/dstfile"
+       cmd = exec.Command("s3cmd", "--no-ssl", "--host="+s.testServer.URL[7:], "--host-bucket="+s.testServer.URL[7:], "--access_key="+arvadostest.ActiveTokenUUID, "--secret_key="+arvadostest.ActiveToken, "get", "s3://"+arvadostest.FooCollection+"/foo,;$[|]bar", tmpfile)
+       buf, err = cmd.CombinedOutput()
+       c.Check(err, check.NotNil)
+       // As of commit b7520e5c25e1bf25c1a8bf5aa2eadb299be8f606
+       // (between debian bullseye and bookworm versions), s3cmd
+       // started catching the NoSuchKey error code and replacing it
+       // with "Source object '%s' does not exist.".
+       c.Check(string(buf), check.Matches, `(?ms).*(NoSuchKey|Source object.*does not exist).*\n`)
 }
 
 func (s *IntegrationSuite) TestS3BucketInHost(c *check.C) {