1 // Copyright (C) The Arvados Authors. All rights reserved.
3 // SPDX-License-Identifier: AGPL-3.0
17 "git.arvados.org/arvados.git/sdk/go/arvados"
18 "git.arvados.org/arvados.git/sdk/go/arvadosclient"
19 "git.arvados.org/arvados.git/sdk/go/arvadostest"
20 "git.arvados.org/arvados.git/sdk/go/keepclient"
21 "github.com/AdRoll/goamz/aws"
22 "github.com/AdRoll/goamz/s3"
23 check "gopkg.in/check.v1"
28 ac *arvadosclient.ArvadosClient
29 kc *keepclient.KeepClient
32 coll arvados.Collection
36 func (s *IntegrationSuite) s3setup(c *check.C) s3stage {
37 var proj arvados.Group
38 var coll arvados.Collection
39 arv := arvados.NewClientFromEnv()
40 arv.AuthToken = arvadostest.ActiveToken
41 err := arv.RequestAndDecode(&proj, "POST", "arvados/v1/groups", nil, map[string]interface{}{
42 "group": map[string]interface{}{
43 "group_class": "project",
44 "name": "keep-web s3 test",
46 "ensure_unique_name": true,
48 c.Assert(err, check.IsNil)
49 err = arv.RequestAndDecode(&coll, "POST", "arvados/v1/collections", nil, map[string]interface{}{"collection": map[string]interface{}{
50 "owner_uuid": proj.UUID,
51 "name": "keep-web s3 test collection",
52 "manifest_text": ". d41d8cd98f00b204e9800998ecf8427e+0 0:0:emptyfile\n./emptydir d41d8cd98f00b204e9800998ecf8427e+0 0:0:.\n",
54 c.Assert(err, check.IsNil)
55 ac, err := arvadosclient.New(arv)
56 c.Assert(err, check.IsNil)
57 kc, err := keepclient.MakeKeepClient(ac)
58 c.Assert(err, check.IsNil)
59 fs, err := coll.FileSystem(arv, kc)
60 c.Assert(err, check.IsNil)
61 f, err := fs.OpenFile("sailboat.txt", os.O_CREATE|os.O_WRONLY, 0644)
62 c.Assert(err, check.IsNil)
63 _, err = f.Write([]byte("⛵\n"))
64 c.Assert(err, check.IsNil)
66 c.Assert(err, check.IsNil)
68 c.Assert(err, check.IsNil)
69 err = arv.RequestAndDecode(&coll, "GET", "arvados/v1/collections/"+coll.UUID, nil, nil)
70 c.Assert(err, check.IsNil)
72 auth := aws.NewAuth(arvadostest.ActiveTokenV2, arvadostest.ActiveTokenV2, "", time.Now().Add(time.Hour))
74 Name: s.testServer.Addr,
75 S3Endpoint: "http://" + s.testServer.Addr,
77 client := s3.New(*auth, region)
83 projbucket: &s3.Bucket{
88 collbucket: &s3.Bucket{
95 func (stage s3stage) teardown(c *check.C) {
96 if stage.coll.UUID != "" {
97 err := stage.arv.RequestAndDecode(&stage.coll, "DELETE", "arvados/v1/collections/"+stage.coll.UUID, nil, nil)
98 c.Check(err, check.IsNil)
100 if stage.proj.UUID != "" {
101 err := stage.arv.RequestAndDecode(&stage.proj, "DELETE", "arvados/v1/groups/"+stage.proj.UUID, nil, nil)
102 c.Check(err, check.IsNil)
106 func (s *IntegrationSuite) TestS3CollectionGetObject(c *check.C) {
107 stage := s.s3setup(c)
108 defer stage.teardown(c)
109 s.testS3GetObject(c, stage.collbucket, "")
111 func (s *IntegrationSuite) TestS3ProjectGetObject(c *check.C) {
112 stage := s.s3setup(c)
113 defer stage.teardown(c)
114 s.testS3GetObject(c, stage.projbucket, stage.coll.Name+"/")
116 func (s *IntegrationSuite) testS3GetObject(c *check.C, bucket *s3.Bucket, prefix string) {
117 rdr, err := bucket.GetReader(prefix + "emptyfile")
118 c.Assert(err, check.IsNil)
119 buf, err := ioutil.ReadAll(rdr)
120 c.Check(err, check.IsNil)
121 c.Check(len(buf), check.Equals, 0)
123 c.Check(err, check.IsNil)
125 rdr, err = bucket.GetReader(prefix + "missingfile")
126 c.Check(err, check.ErrorMatches, `404 Not Found`)
128 rdr, err = bucket.GetReader(prefix + "sailboat.txt")
129 c.Assert(err, check.IsNil)
130 buf, err = ioutil.ReadAll(rdr)
131 c.Check(err, check.IsNil)
132 c.Check(buf, check.DeepEquals, []byte("⛵\n"))
134 c.Check(err, check.IsNil)
137 func (s *IntegrationSuite) TestS3CollectionPutObjectSuccess(c *check.C) {
138 stage := s.s3setup(c)
139 defer stage.teardown(c)
140 s.testS3PutObjectSuccess(c, stage.collbucket, "")
142 func (s *IntegrationSuite) TestS3ProjectPutObjectSuccess(c *check.C) {
143 stage := s.s3setup(c)
144 defer stage.teardown(c)
145 s.testS3PutObjectSuccess(c, stage.projbucket, stage.coll.Name+"/")
147 func (s *IntegrationSuite) testS3PutObjectSuccess(c *check.C, bucket *s3.Bucket, prefix string) {
148 for _, trial := range []struct {
156 path: "newdir/newfile",
159 path: "newdir1/newdir2/newfile",
163 c.Logf("=== %v", trial)
165 objname := prefix + trial.path
167 _, err := bucket.GetReader(objname)
168 c.Assert(err, check.ErrorMatches, `404 Not Found`)
170 buf := make([]byte, trial.size)
173 err = bucket.PutReader(objname, bytes.NewReader(buf), int64(len(buf)), "application/octet-stream", s3.Private, s3.Options{})
174 c.Check(err, check.IsNil)
176 rdr, err := bucket.GetReader(objname)
177 if !c.Check(err, check.IsNil) {
180 buf2, err := ioutil.ReadAll(rdr)
181 c.Check(err, check.IsNil)
182 c.Check(buf2, check.HasLen, len(buf))
183 c.Check(bytes.Equal(buf, buf2), check.Equals, true)
187 func (s *IntegrationSuite) TestS3CollectionPutObjectFailure(c *check.C) {
188 stage := s.s3setup(c)
189 defer stage.teardown(c)
190 s.testS3PutObjectFailure(c, stage.collbucket, "")
192 func (s *IntegrationSuite) TestS3ProjectPutObjectFailure(c *check.C) {
193 stage := s.s3setup(c)
194 defer stage.teardown(c)
195 s.testS3PutObjectFailure(c, stage.projbucket, stage.coll.Name+"/")
197 func (s *IntegrationSuite) testS3PutObjectFailure(c *check.C, bucket *s3.Bucket, prefix string) {
198 var wg sync.WaitGroup
199 for _, trial := range []struct {
203 path: "emptyfile/newname", // emptyfile exists, see s3setup()
205 path: "emptyfile/", // emptyfile exists, see s3setup()
207 path: "emptydir", // dir already exists, see s3setup()
230 c.Logf("=== %v", trial)
232 objname := prefix + trial.path
234 buf := make([]byte, 1234)
237 err := bucket.PutReader(objname, bytes.NewReader(buf), int64(len(buf)), "application/octet-stream", s3.Private, s3.Options{})
238 if !c.Check(err, check.ErrorMatches, `400 Bad.*`, check.Commentf("PUT %q should fail", objname)) {
242 if objname != "" && objname != "/" {
243 _, err = bucket.GetReader(objname)
244 c.Check(err, check.ErrorMatches, `404 Not Found`, check.Commentf("GET %q should return 404", objname))
251 func (stage *s3stage) writeBigDirs(c *check.C, dirs int, filesPerDir int) {
252 fs, err := stage.coll.FileSystem(stage.arv, stage.kc)
253 c.Assert(err, check.IsNil)
254 for d := 0; d < dirs; d++ {
255 dir := fmt.Sprintf("dir%d", d)
256 c.Assert(fs.Mkdir(dir, 0755), check.IsNil)
257 for i := 0; i < filesPerDir; i++ {
258 f, err := fs.OpenFile(fmt.Sprintf("%s/file%d.txt", dir, i), os.O_CREATE|os.O_WRONLY, 0644)
259 c.Assert(err, check.IsNil)
260 c.Assert(f.Close(), check.IsNil)
263 c.Assert(fs.Sync(), check.IsNil)
266 func (s *IntegrationSuite) TestS3CollectionList(c *check.C) {
267 stage := s.s3setup(c)
268 defer stage.teardown(c)
271 stage.writeBigDirs(c, 2, filesPerDir)
272 s.testS3List(c, stage.collbucket, "", 4000, 2+filesPerDir*2)
273 s.testS3List(c, stage.collbucket, "", 131, 2+filesPerDir*2)
274 s.testS3List(c, stage.collbucket, "dir0/", 71, filesPerDir)
276 func (s *IntegrationSuite) testS3List(c *check.C, bucket *s3.Bucket, prefix string, pageSize, expectFiles int) {
277 expectPageSize := pageSize
278 if expectPageSize > 1000 {
279 expectPageSize = 1000
281 gotKeys := map[string]s3.Key{}
285 resp, err := bucket.List(prefix, "", nextMarker, pageSize)
286 if !c.Check(err, check.IsNil) {
289 c.Check(len(resp.Contents) <= expectPageSize, check.Equals, true)
290 if pages++; !c.Check(pages <= (expectFiles/expectPageSize)+1, check.Equals, true) {
293 for _, key := range resp.Contents {
294 gotKeys[key.Key] = key
296 if !resp.IsTruncated {
297 c.Check(resp.NextMarker, check.Equals, "")
300 if !c.Check(resp.NextMarker, check.Not(check.Equals), "") {
303 nextMarker = resp.NextMarker
305 c.Check(len(gotKeys), check.Equals, expectFiles)
308 func (s *IntegrationSuite) TestS3CollectionListRollup(c *check.C) {
309 stage := s.s3setup(c)
310 defer stage.teardown(c)
314 stage.writeBigDirs(c, dirs, filesPerDir)
315 err := stage.collbucket.PutReader("dingbats", &bytes.Buffer{}, 0, "application/octet-stream", s3.Private, s3.Options{})
316 c.Assert(err, check.IsNil)
317 var allfiles []string
318 for marker := ""; ; {
319 resp, err := stage.collbucket.List("", "", marker, 20000)
320 c.Check(err, check.IsNil)
321 for _, key := range resp.Contents {
322 if len(allfiles) == 0 || allfiles[len(allfiles)-1] != key.Key {
323 allfiles = append(allfiles, key.Key)
326 marker = resp.NextMarker
331 c.Check(allfiles, check.HasLen, dirs*filesPerDir+3)
333 for _, trial := range []struct {
342 {"dir0", "/", "dir0/file14.txt"}, // no commonprefixes
343 {"", "", "dir0/file14.txt"}, // middle page, skip walking dir1
344 {"", "", "dir1/file14.txt"}, // middle page, skip walking dir0
345 {"", "", "dir1/file498.txt"}, // last page of results
346 {"dir1/file", "", "dir1/file498.txt"}, // last page of results, with prefix
347 {"dir1/file", "/", "dir1/file498.txt"}, // last page of results, with prefix + delimiter
348 {"dir1", "Z", "dir1/file498.txt"}, // delimiter "Z" never appears
349 {"dir2", "/", ""}, // prefix "dir2" does not exist
352 c.Logf("\n\n=== trial %+v", trial)
355 resp, err := stage.collbucket.List(trial.prefix, trial.delimiter, trial.marker, maxKeys)
356 c.Check(err, check.IsNil)
357 if resp.IsTruncated && trial.delimiter == "" {
358 // goamz List method fills in the missing
359 // NextMarker field if resp.IsTruncated, so
360 // now we can't really tell whether it was
361 // sent by the server or by goamz. In cases
362 // where it should be empty but isn't, assume
363 // it's goamz's fault.
367 var expectKeys []string
368 var expectPrefixes []string
369 var expectNextMarker string
370 var expectTruncated bool
371 for _, key := range allfiles {
372 full := len(expectKeys)+len(expectPrefixes) >= maxKeys
373 if !strings.HasPrefix(key, trial.prefix) || key < trial.marker {
375 } else if idx := strings.Index(key[len(trial.prefix):], trial.delimiter); trial.delimiter != "" && idx >= 0 {
376 prefix := key[:len(trial.prefix)+idx+1]
377 if len(expectPrefixes) > 0 && expectPrefixes[len(expectPrefixes)-1] == prefix {
378 // same prefix as previous key
380 expectNextMarker = key
381 expectTruncated = true
383 expectPrefixes = append(expectPrefixes, prefix)
386 if trial.delimiter != "" {
387 expectNextMarker = key
389 expectTruncated = true
392 expectKeys = append(expectKeys, key)
397 for _, key := range resp.Contents {
398 gotKeys = append(gotKeys, key.Key)
400 var gotPrefixes []string
401 for _, prefix := range resp.CommonPrefixes {
402 gotPrefixes = append(gotPrefixes, prefix)
404 c.Check(gotKeys, check.DeepEquals, expectKeys)
405 c.Check(gotPrefixes, check.DeepEquals, expectPrefixes)
406 c.Check(resp.NextMarker, check.Equals, expectNextMarker)
407 c.Check(resp.IsTruncated, check.Equals, expectTruncated)
408 c.Logf("=== trial %+v keys %q prefixes %q nextMarker %q", trial, gotKeys, gotPrefixes, resp.NextMarker)