// Make sure LoginCluster is properly configured
for cls := range s.testClusters {
- c.Check(
- s.testClusters[cls].Config.Clusters[cls].Login.LoginCluster,
- check.Equals, "z1111",
- check.Commentf("incorrect LoginCluster config on cluster %q", cls))
+ if cls == "z1111" || cls == "z3333" {
+ c.Check(
+ s.testClusters[cls].Config.Clusters[cls].Login.LoginCluster,
+ check.Equals, "z1111",
+ check.Commentf("incorrect LoginCluster config on cluster %q", cls))
+ }
}
// Create some users, request them on the federated cluster so they're cached.
var users []arvados.User
- for userNr := range []int{0, 1} {
+ for userNr := 0; userNr < 2; userNr++ {
_, _, _, user := s.testClusters["z1111"].UserClients(
rootctx1,
c,
c.Assert(user.Username, check.Not(check.Equals), "")
users = append(users, user)
- lst, err := conn3.UserList(rootctx1, arvados.ListOptions{Limit: math.MaxInt64})
+ lst, err := conn3.UserList(rootctx1, arvados.ListOptions{Limit: -1})
c.Assert(err, check.Equals, nil)
userFound := false
for _, fedUser := range lst.Items {
c.Assert(err, check.Equals, nil)
// Re-request the list on the federated cluster & check for updates
- lst, err := conn3.UserList(rootctx1, arvados.ListOptions{Limit: math.MaxInt64})
+ lst, err := conn3.UserList(rootctx1, arvados.ListOptions{Limit: -1})
c.Assert(err, check.Equals, nil)
var user0Found, user1Found bool
for _, user := range lst.Items {