X-Git-Url: https://git.arvados.org/arvados.git/blobdiff_plain/f451703c7c38cb006f468525ea02196b659f3e27..469ea187586ea8017e26874c2d80414ce7571fae:/lib/controller/federation.go diff --git a/lib/controller/federation.go b/lib/controller/federation.go index 1d4844486c..e7d6e29b88 100644 --- a/lib/controller/federation.go +++ b/lib/controller/federation.go @@ -5,708 +5,103 @@ package controller import ( - "bufio" "bytes" - "context" - "crypto/md5" "database/sql" "encoding/json" "fmt" "io" "io/ioutil" + "mime" "net/http" "net/url" "regexp" "strings" - "sync" - "git.curoverse.com/arvados.git/sdk/go/arvados" - "git.curoverse.com/arvados.git/sdk/go/auth" - "git.curoverse.com/arvados.git/sdk/go/httpserver" - "git.curoverse.com/arvados.git/sdk/go/keepclient" + "git.arvados.org/arvados.git/sdk/go/arvados" + "git.arvados.org/arvados.git/sdk/go/auth" + "git.arvados.org/arvados.git/sdk/go/ctxlog" + "github.com/jmcvetta/randutil" ) var pathPattern = `^/arvados/v1/%s(/([0-9a-z]{5})-%s-[0-9a-z]{15})?(.*)$` var wfRe = regexp.MustCompile(fmt.Sprintf(pathPattern, "workflows", "7fd4e")) var containersRe = regexp.MustCompile(fmt.Sprintf(pathPattern, "containers", "dz642")) var containerRequestsRe = regexp.MustCompile(fmt.Sprintf(pathPattern, "container_requests", "xvhdp")) -var collectionRe = regexp.MustCompile(fmt.Sprintf(pathPattern, "collections", "4zz18")) -var collectionByPDHRe = regexp.MustCompile(`^/arvados/v1/collections/([0-9a-fA-F]{32}\+[0-9]+)+$`) +var collectionsRe = regexp.MustCompile(fmt.Sprintf(pathPattern, "collections", "4zz18")) +var collectionsByPDHRe = regexp.MustCompile(`^/arvados/v1/collections/([0-9a-fA-F]{32}\+[0-9]+)+$`) +var linksRe = regexp.MustCompile(fmt.Sprintf(pathPattern, "links", "o0j2j")) -type genericFederatedRequestHandler struct { - next http.Handler - handler *Handler - matcher *regexp.Regexp -} - -type collectionFederatedRequestHandler struct { - next http.Handler - handler *Handler -} - -func (h *Handler) remoteClusterRequest(remoteID string, w http.ResponseWriter, req *http.Request, filter ResponseFilter) { +func (h *Handler) remoteClusterRequest(remoteID string, req *http.Request) (*http.Response, error) { remote, ok := h.Cluster.RemoteClusters[remoteID] if !ok { - httpserver.Error(w, "no proxy available for cluster "+remoteID, http.StatusNotFound) - return + return nil, HTTPError{fmt.Sprintf("no proxy available for cluster %v", remoteID), http.StatusNotFound} } scheme := remote.Scheme if scheme == "" { scheme = "https" } - err := h.saltAuthToken(req, remoteID) + saltedReq, err := h.saltAuthToken(req, remoteID) if err != nil { - httpserver.Error(w, err.Error(), http.StatusBadRequest) - return + return nil, err } urlOut := &url.URL{ Scheme: scheme, Host: remote.Host, - Path: req.URL.Path, - RawPath: req.URL.RawPath, - RawQuery: req.URL.RawQuery, + Path: saltedReq.URL.Path, + RawPath: saltedReq.URL.RawPath, + RawQuery: saltedReq.URL.RawQuery, } client := h.secureClient if remote.Insecure { client = h.insecureClient } - h.proxy.Do(w, req, urlOut, client, filter) + return h.proxy.Do(saltedReq, urlOut, client) } -// loadParamsFromForm expects a request with -// application/x-www-form-urlencoded body. It parses the query, adds -// the query parameters to "params", and replaces the request body -// with a buffer holding the original body contents so it can be -// re-read by downstream proxy steps. -func loadParamsFromForm(req *http.Request, params url.Values) error { - body, err := ioutil.ReadAll(req.Body) - if err != nil { - return err - } - req.Body = ioutil.NopCloser(bytes.NewBuffer(body)) - var v2 url.Values - if v2, err = url.ParseQuery(string(body)); err != nil { - return err - } - for k, v := range v2 { - params[k] = append(params[k], v...) - } - return nil -} - -// loadParamsFromForm expects a request with application/json body. -// It parses the body, populates "loadInto", and replaces the request -// body with a buffer holding the original body contents so it can be -// re-read by downstream proxy steps. -func loadParamsFromJson(req *http.Request, loadInto interface{}) error { - var cl int64 - if req.ContentLength > 0 { - cl = req.ContentLength - } - postBody := bytes.NewBuffer(make([]byte, 0, cl)) - defer req.Body.Close() - - rdr := io.TeeReader(req.Body, postBody) - - err := json.NewDecoder(rdr).Decode(loadInto) - if err != nil { +// Buffer request body, parse form parameters in request, and then +// replace original body with the buffer so it can be re-read by +// downstream proxy steps. +func loadParamsFromForm(req *http.Request) error { + var postBody *bytes.Buffer + if ct := req.Header.Get("Content-Type"); ct == "" { + // Assume application/octet-stream, i.e., no form to parse. + } else if ct, _, err := mime.ParseMediaType(ct); err != nil { return err - } - req.Body = ioutil.NopCloser(postBody) - return nil -} - -type multiClusterQueryResponseCollector struct { - responses []interface{} - error error - kind string - clusterID string -} - -func (c *multiClusterQueryResponseCollector) collectResponse(resp *http.Response, - requestError error) (newResponse *http.Response, err error) { - if requestError != nil { - c.error = requestError - return nil, nil - } - - defer resp.Body.Close() - loadInto := make(map[string]interface{}) - err = json.NewDecoder(resp.Body).Decode(&loadInto) - - if err == nil { - if resp.StatusCode != http.StatusOK { - c.error = fmt.Errorf("error fetching from %v (%v): %v", c.clusterID, resp.Status, loadInto["errors"]) - } else { - c.responses = loadInto["items"].([]interface{}) - c.kind, _ = loadInto["kind"].(string) - } - } else { - c.error = fmt.Errorf("error fetching from %v (%v): %v", c.clusterID, resp.Status, err) - } - - return nil, nil -} - -func (h *genericFederatedRequestHandler) remoteQueryUUIDs(w http.ResponseWriter, - req *http.Request, params url.Values, - clusterID string, uuids []string) (rp []interface{}, kind string, err error) { - - found := make(map[string]bool) - for len(uuids) > 0 { - var remoteReq http.Request - remoteReq.Header = req.Header - remoteReq.Method = "POST" - remoteReq.URL = &url.URL{Path: req.URL.Path} - remoteParams := make(url.Values) - remoteParams["_method"] = []string{"GET"} - remoteParams["count"] = []string{"none"} - if len(params["select"]) != 0 { - remoteParams["select"] = params["select"] - } - content, err := json.Marshal(uuids) - if err != nil { - return nil, "", err - } - remoteParams["filters"] = []string{fmt.Sprintf(`[["uuid", "in", %s]]`, content)} - enc := remoteParams.Encode() - remoteReq.Body = ioutil.NopCloser(bytes.NewBufferString(enc)) - - rc := multiClusterQueryResponseCollector{clusterID: clusterID} - - if clusterID == h.handler.Cluster.ClusterID { - h.handler.localClusterRequest(w, &remoteReq, - rc.collectResponse) - } else { - h.handler.remoteClusterRequest(clusterID, w, &remoteReq, - rc.collectResponse) - } - if rc.error != nil { - return nil, "", rc.error - } - - kind = rc.kind - - if len(rc.responses) == 0 { - // We got zero responses, no point in doing - // another query. - return rp, kind, nil - } - - rp = append(rp, rc.responses...) - - // Go through the responses and determine what was - // returned. If there are remaining items, loop - // around and do another request with just the - // stragglers. - for _, i := range rc.responses { - m, ok := i.(map[string]interface{}) - if ok { - uuid, ok := m["uuid"].(string) - if ok { - found[uuid] = true - } - } - } - - l := []string{} - for _, u := range uuids { - if !found[u] { - l = append(l, u) - } - } - uuids = l - } - - return rp, kind, nil -} - -func (h *genericFederatedRequestHandler) handleMultiClusterQuery(w http.ResponseWriter, req *http.Request, - params url.Values, clusterId *string) bool { - - var filters [][]interface{} - err := json.Unmarshal([]byte(params["filters"][0]), &filters) - if err != nil { - httpserver.Error(w, err.Error(), http.StatusBadRequest) - return true - } - - // Split the list of uuids by prefix - queryClusters := make(map[string][]string) - expectCount := 0 - for _, f1 := range filters { - if len(f1) != 3 { - return false - } - lhs, ok := f1[0].(string) - if ok && lhs == "uuid" { - op, ok := f1[1].(string) - if !ok { - return false - } - - if op == "in" { - rhs, ok := f1[2].([]interface{}) - if ok { - for _, i := range rhs { - u := i.(string) - *clusterId = u[0:5] - queryClusters[u[0:5]] = append(queryClusters[u[0:5]], u) - } - expectCount += len(rhs) - } - } else if op == "=" { - u, ok := f1[2].(string) - if ok { - *clusterId = u[0:5] - queryClusters[u[0:5]] = append(queryClusters[u[0:5]], u) - expectCount += 1 - } - } else { - return false - } - } else { - return false - } - } - - if len(queryClusters) <= 1 { - // Query does not search for uuids across multiple - // clusters. - return false - } - - // Validations - if !(len(params["count"]) == 1 && (params["count"][0] == `none` || - params["count"][0] == `"none"`)) { - httpserver.Error(w, "Federated multi-object query must have 'count=none'", http.StatusBadRequest) - return true - } - if len(params["limit"]) != 0 || len(params["offset"]) != 0 || len(params["order"]) != 0 { - httpserver.Error(w, "Federated multi-object may not provide 'limit', 'offset' or 'order'.", http.StatusBadRequest) - return true - } - if expectCount > h.handler.Cluster.MaxItemsPerResponse { - httpserver.Error(w, fmt.Sprintf("Federated multi-object request for %v objects which is more than max page size %v.", - expectCount, h.handler.Cluster.MaxItemsPerResponse), http.StatusBadRequest) - return true - } - if len(params["select"]) == 1 { - foundUUID := false - var selects []interface{} - err := json.Unmarshal([]byte(params["select"][0]), &selects) - if err != nil { - httpserver.Error(w, err.Error(), http.StatusBadRequest) - return true - } - - for _, r := range selects { - if r.(string) == "uuid" { - foundUUID = true - break - } - } - if !foundUUID { - httpserver.Error(w, "Federated multi-object request must include 'uuid' in 'select'", http.StatusBadRequest) - return true - } - } - - // Perform parallel requests to each cluster - - // use channel as a semaphore to limit the number of parallel - // requests at a time - sem := make(chan bool, h.handler.Cluster.ParallelRemoteRequests) - defer close(sem) - wg := sync.WaitGroup{} - - req.Header.Set("Content-Type", "application/x-www-form-urlencoded") - mtx := sync.Mutex{} - errors := []error{} - var completeResponses []interface{} - var kind string - - for k, v := range queryClusters { - if len(v) == 0 { - // Nothing to query - continue - } - - // blocks until it can put a value into the - // channel (which has a max queue capacity) - sem <- true - wg.Add(1) - go func(k string, v []string) { - rp, kn, err := h.remoteQueryUUIDs(w, req, params, k, v) - mtx.Lock() - if err == nil { - completeResponses = append(completeResponses, rp...) - kind = kn - } else { - errors = append(errors, err) - } - mtx.Unlock() - wg.Done() - <-sem - }(k, v) - } - wg.Wait() - - if len(errors) > 0 { - var strerr []string - for _, e := range errors { - strerr = append(strerr, e.Error()) - } - httpserver.Errors(w, strerr, http.StatusBadGateway) - return true - } - - w.Header().Set("Content-Type", "application/json") - w.WriteHeader(http.StatusOK) - itemList := make(map[string]interface{}) - itemList["items"] = completeResponses - itemList["kind"] = kind - json.NewEncoder(w).Encode(itemList) - - return true -} - -func (h *genericFederatedRequestHandler) ServeHTTP(w http.ResponseWriter, req *http.Request) { - m := h.matcher.FindStringSubmatch(req.URL.Path) - clusterId := "" - - if len(m) > 0 && m[2] != "" { - clusterId = m[2] - } - - // First, parse the query portion of the URL. - var params url.Values - var err error - if params, err = url.ParseQuery(req.URL.RawQuery); err != nil { - httpserver.Error(w, err.Error(), http.StatusBadRequest) - return - } - - // Next, if appropriate, merge in parameters from the form POST body. - if req.Method == "POST" && req.Header.Get("Content-Type") == "application/x-www-form-urlencoded" { - if err = loadParamsFromForm(req, params); err != nil { - httpserver.Error(w, err.Error(), http.StatusBadRequest) - return - } - } - - // Check if the parameters have an explicit cluster_id - if len(params["cluster_id"]) == 1 { - clusterId = params["cluster_id"][0] - } - - // Handle the POST-as-GET special case (workaround for large - // GET requests that potentially exceed maximum URL length, - // like multi-object queries where the filter has 100s of - // items) - effectiveMethod := req.Method - if req.Method == "POST" && len(params["_method"]) == 1 { - effectiveMethod = params["_method"][0] - } - - if effectiveMethod == "GET" && clusterId == "" && len(params["filters"]) == 1 { - if h.handleMultiClusterQuery(w, req, params, &clusterId) { - return - } - } - - if clusterId == "" || clusterId == h.handler.Cluster.ClusterID { - h.next.ServeHTTP(w, req) - } else { - h.handler.remoteClusterRequest(clusterId, w, req, nil) - } -} - -type rewriteSignaturesClusterId struct { - clusterID string - expectHash string -} - -func (rw rewriteSignaturesClusterId) rewriteSignatures(resp *http.Response, requestError error) (newResponse *http.Response, err error) { - if requestError != nil { - return resp, requestError - } - - if resp.StatusCode != 200 { - return resp, nil - } - - originalBody := resp.Body - defer originalBody.Close() - - var col arvados.Collection - err = json.NewDecoder(resp.Body).Decode(&col) - if err != nil { - return nil, err - } - - // rewriting signatures will make manifest text 5-10% bigger so calculate - // capacity accordingly - updatedManifest := bytes.NewBuffer(make([]byte, 0, int(float64(len(col.ManifestText))*1.1))) - - hasher := md5.New() - mw := io.MultiWriter(hasher, updatedManifest) - sz := 0 - - scanner := bufio.NewScanner(strings.NewReader(col.ManifestText)) - scanner.Buffer(make([]byte, 1048576), len(col.ManifestText)) - for scanner.Scan() { - line := scanner.Text() - tokens := strings.Split(line, " ") - if len(tokens) < 3 { - return nil, fmt.Errorf("Invalid stream (<3 tokens): %q", line) - } - - n, err := mw.Write([]byte(tokens[0])) - if err != nil { - return nil, fmt.Errorf("Error updating manifest: %v", err) - } - sz += n - for _, token := range tokens[1:] { - n, err = mw.Write([]byte(" ")) - if err != nil { - return nil, fmt.Errorf("Error updating manifest: %v", err) - } - sz += n - - m := keepclient.SignedLocatorRe.FindStringSubmatch(token) - if m != nil { - // Rewrite the block signature to be a remote signature - _, err = fmt.Fprintf(updatedManifest, "%s%s%s+R%s-%s%s", m[1], m[2], m[3], rw.clusterID, m[5][2:], m[8]) - if err != nil { - return nil, fmt.Errorf("Error updating manifest: %v", err) - } - - // for hash checking, ignore signatures - n, err = fmt.Fprintf(hasher, "%s%s", m[1], m[2]) - if err != nil { - return nil, fmt.Errorf("Error updating manifest: %v", err) - } - sz += n - } else { - n, err = mw.Write([]byte(token)) - if err != nil { - return nil, fmt.Errorf("Error updating manifest: %v", err) - } - sz += n - } - } - n, err = mw.Write([]byte("\n")) - if err != nil { - return nil, fmt.Errorf("Error updating manifest: %v", err) - } - sz += n - } - - // Check that expected hash is consistent with - // portable_data_hash field of the returned record - if rw.expectHash == "" { - rw.expectHash = col.PortableDataHash - } else if rw.expectHash != col.PortableDataHash { - return nil, fmt.Errorf("portable_data_hash %q on returned record did not match expected hash %q ", rw.expectHash, col.PortableDataHash) - } - - // Certify that the computed hash of the manifest_text matches our expectation - sum := hasher.Sum(nil) - computedHash := fmt.Sprintf("%x+%v", sum, sz) - if computedHash != rw.expectHash { - return nil, fmt.Errorf("Computed manifest_text hash %q did not match expected hash %q", computedHash, rw.expectHash) - } - - col.ManifestText = updatedManifest.String() - - newbody, err := json.Marshal(col) - if err != nil { - return nil, err - } - - buf := bytes.NewBuffer(newbody) - resp.Body = ioutil.NopCloser(buf) - resp.ContentLength = int64(buf.Len()) - resp.Header.Set("Content-Length", fmt.Sprintf("%v", buf.Len())) - - return resp, nil -} - -func filterLocalClusterResponse(resp *http.Response, requestError error) (newResponse *http.Response, err error) { - if requestError != nil { - return resp, requestError - } - - if resp.StatusCode == 404 { - // Suppress returning this result, because we want to - // search the federation. - return nil, nil - } - return resp, nil -} - -type searchRemoteClusterForPDH struct { - pdh string - remoteID string - mtx *sync.Mutex - sentResponse *bool - sharedContext *context.Context - cancelFunc func() - errors *[]string - statusCode *int -} - -func (s *searchRemoteClusterForPDH) filterRemoteClusterResponse(resp *http.Response, requestError error) (newResponse *http.Response, err error) { - s.mtx.Lock() - defer s.mtx.Unlock() - - if *s.sentResponse { - // Another request already returned a response - return nil, nil - } - - if requestError != nil { - *s.errors = append(*s.errors, fmt.Sprintf("Request error contacting %q: %v", s.remoteID, requestError)) - // Record the error and suppress response - return nil, nil - } - - if resp.StatusCode != 200 { - // Suppress returning unsuccessful result. Maybe - // another request will find it. - // TODO collect and return error responses. - *s.errors = append(*s.errors, fmt.Sprintf("Response from %q: %v", s.remoteID, resp.Status)) - if resp.StatusCode != 404 { - // Got a non-404 error response, convert into BadGateway - *s.statusCode = http.StatusBadGateway + } else if ct == "application/x-www-form-urlencoded" && req.Body != nil { + var cl int64 + if req.ContentLength > 0 { + cl = req.ContentLength } - return nil, nil - } - - s.mtx.Unlock() - - // This reads the response body. We don't want to hold the - // lock while doing this because other remote requests could - // also have made it to this point, and we don't want a - // slow response holding the lock to block a faster response - // that is waiting on the lock. - newResponse, err = rewriteSignaturesClusterId{s.remoteID, s.pdh}.rewriteSignatures(resp, nil) - - s.mtx.Lock() - - if *s.sentResponse { - // Another request already returned a response - return nil, nil + postBody = bytes.NewBuffer(make([]byte, 0, cl)) + originalBody := req.Body + defer originalBody.Close() + req.Body = ioutil.NopCloser(io.TeeReader(req.Body, postBody)) } + err := req.ParseForm() if err != nil { - // Suppress returning unsuccessful result. Maybe - // another request will be successful. - *s.errors = append(*s.errors, fmt.Sprintf("Error parsing response from %q: %v", s.remoteID, err)) - return nil, nil - } - - // We have a successful response. Suppress/cancel all the - // other requests/responses. - *s.sentResponse = true - s.cancelFunc() - - return newResponse, nil -} - -func (h *collectionFederatedRequestHandler) ServeHTTP(w http.ResponseWriter, req *http.Request) { - if req.Method != "GET" { - // Only handle GET requests right now - h.next.ServeHTTP(w, req) - return - } - - m := collectionByPDHRe.FindStringSubmatch(req.URL.Path) - if len(m) != 2 { - // Not a collection PDH GET request - m = collectionRe.FindStringSubmatch(req.URL.Path) - clusterId := "" - - if len(m) > 0 { - clusterId = m[2] - } - - if clusterId != "" && clusterId != h.handler.Cluster.ClusterID { - // request for remote collection by uuid - h.handler.remoteClusterRequest(clusterId, w, req, - rewriteSignaturesClusterId{clusterId, ""}.rewriteSignatures) - return - } - // not a collection UUID request, or it is a request - // for a local UUID, either way, continue down the - // handler stack. - h.next.ServeHTTP(w, req) - return - } - - // Request for collection by PDH. Search the federation. - - // First, query the local cluster. - if h.handler.localClusterRequest(w, req, filterLocalClusterResponse) { - return - } - - sharedContext, cancelFunc := context.WithCancel(req.Context()) - defer cancelFunc() - req = req.WithContext(sharedContext) - - // Create a goroutine for each cluster in the - // RemoteClusters map. The first valid result gets - // returned to the client. When that happens, all - // other outstanding requests are cancelled or - // suppressed. - sentResponse := false - mtx := sync.Mutex{} - wg := sync.WaitGroup{} - var errors []string - var errorCode int = 404 - - // use channel as a semaphore to limit the number of parallel - // requests at a time - sem := make(chan bool, h.handler.Cluster.ParallelRemoteRequests) - defer close(sem) - for remoteID := range h.handler.Cluster.RemoteClusters { - // blocks until it can put a value into the - // channel (which has a max queue capacity) - sem <- true - if sentResponse { - break - } - search := &searchRemoteClusterForPDH{m[1], remoteID, &mtx, &sentResponse, - &sharedContext, cancelFunc, &errors, &errorCode} - wg.Add(1) - go func() { - h.handler.remoteClusterRequest(search.remoteID, w, req, search.filterRemoteClusterResponse) - wg.Done() - <-sem - }() + return err } - wg.Wait() - if sentResponse { - return + if req.Body != nil && postBody != nil { + req.Body = ioutil.NopCloser(postBody) } - - // No successful responses, so return the error - httpserver.Errors(w, errors, errorCode) + return nil } func (h *Handler) setupProxyRemoteCluster(next http.Handler) http.Handler { mux := http.NewServeMux() - mux.Handle("/arvados/v1/workflows", &genericFederatedRequestHandler{next, h, wfRe}) - mux.Handle("/arvados/v1/workflows/", &genericFederatedRequestHandler{next, h, wfRe}) - mux.Handle("/arvados/v1/containers", &genericFederatedRequestHandler{next, h, containersRe}) - mux.Handle("/arvados/v1/containers/", &genericFederatedRequestHandler{next, h, containersRe}) - mux.Handle("/arvados/v1/container_requests", &genericFederatedRequestHandler{next, h, containerRequestsRe}) - mux.Handle("/arvados/v1/container_requests/", &genericFederatedRequestHandler{next, h, containerRequestsRe}) - mux.Handle("/arvados/v1/collections", next) - mux.Handle("/arvados/v1/collections/", &collectionFederatedRequestHandler{next, h}) + + wfHandler := &genericFederatedRequestHandler{next, h, wfRe, nil} + containersHandler := &genericFederatedRequestHandler{next, h, containersRe, nil} + linksRequestsHandler := &genericFederatedRequestHandler{next, h, linksRe, nil} + + mux.Handle("/arvados/v1/workflows", wfHandler) + mux.Handle("/arvados/v1/workflows/", wfHandler) + mux.Handle("/arvados/v1/containers", containersHandler) + mux.Handle("/arvados/v1/containers/", containersHandler) + mux.Handle("/arvados/v1/links", linksRequestsHandler) + mux.Handle("/arvados/v1/links/", linksRequestsHandler) mux.Handle("/", next) return http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) { @@ -726,8 +121,6 @@ func (h *Handler) setupProxyRemoteCluster(next http.Handler) http.Handler { mux.ServeHTTP(w, req) }) - - return mux } type CurrentUser struct { @@ -735,68 +128,176 @@ type CurrentUser struct { UUID string } -func (h *Handler) validateAPItoken(req *http.Request, user *CurrentUser) error { - db, err := h.db(req) +// validateAPItoken extracts the token from the provided http request, +// checks it again api_client_authorizations table in the database, +// and fills in the token scope and user UUID. Does not handle remote +// tokens unless they are already in the database and not expired. +// +// Return values are: +// +// nil, false, non-nil -- if there was an internal error +// +// nil, false, nil -- if the token is invalid +// +// non-nil, true, nil -- if the token is valid +func (h *Handler) validateAPItoken(req *http.Request, token string) (*CurrentUser, bool, error) { + user := CurrentUser{Authorization: arvados.APIClientAuthorization{APIToken: token}} + db, err := h.db(req.Context()) if err != nil { - return err + ctxlog.FromContext(req.Context()).WithError(err).Debugf("validateAPItoken(%s): database error", token) + return nil, false, err + } + + var uuid string + if strings.HasPrefix(token, "v2/") { + sp := strings.Split(token, "/") + uuid = sp[1] + token = sp[2] + } + user.Authorization.APIToken = token + var scopes string + err = db.QueryRowContext(req.Context(), `SELECT api_client_authorizations.uuid, api_client_authorizations.scopes, users.uuid FROM api_client_authorizations JOIN users on api_client_authorizations.user_id=users.id WHERE api_token=$1 AND (expires_at IS NULL OR expires_at > current_timestamp AT TIME ZONE 'UTC') LIMIT 1`, token).Scan(&user.Authorization.UUID, &scopes, &user.UUID) + if err == sql.ErrNoRows { + ctxlog.FromContext(req.Context()).Debugf("validateAPItoken(%s): not found in database", token) + return nil, false, nil + } else if err != nil { + ctxlog.FromContext(req.Context()).WithError(err).Debugf("validateAPItoken(%s): database error", token) + return nil, false, err + } + if uuid != "" && user.Authorization.UUID != uuid { + // secret part matches, but UUID doesn't -- somewhat surprising + ctxlog.FromContext(req.Context()).Debugf("validateAPItoken(%s): secret part found, but with different UUID: %s", token, user.Authorization.UUID) + return nil, false, nil + } + err = json.Unmarshal([]byte(scopes), &user.Authorization.Scopes) + if err != nil { + ctxlog.FromContext(req.Context()).WithError(err).Debugf("validateAPItoken(%s): error parsing scopes from db", token) + return nil, false, err } - return db.QueryRowContext(req.Context(), `SELECT api_client_authorizations.uuid, users.uuid FROM api_client_authorizations JOIN users on api_client_authorizations.user_id=users.id WHERE api_token=$1 AND (expires_at IS NULL OR expires_at > current_timestamp) LIMIT 1`, user.Authorization.APIToken).Scan(&user.Authorization.UUID, &user.UUID) + ctxlog.FromContext(req.Context()).Debugf("validateAPItoken(%s): ok", token) + return &user, true, nil +} + +func (h *Handler) createAPItoken(req *http.Request, userUUID string, scopes []string) (*arvados.APIClientAuthorization, error) { + db, err := h.db(req.Context()) + if err != nil { + return nil, err + } + rd, err := randutil.String(15, "abcdefghijklmnopqrstuvwxyz0123456789") + if err != nil { + return nil, err + } + uuid := fmt.Sprintf("%v-gj3su-%v", h.Cluster.ClusterID, rd) + token, err := randutil.String(50, "abcdefghijklmnopqrstuvwxyz0123456789") + if err != nil { + return nil, err + } + if len(scopes) == 0 { + scopes = append(scopes, "all") + } + scopesjson, err := json.Marshal(scopes) + if err != nil { + return nil, err + } + _, err = db.ExecContext(req.Context(), + `INSERT INTO api_client_authorizations +(uuid, api_token, expires_at, scopes, +user_id, +api_client_id, created_at, updated_at) +VALUES ($1, $2, CURRENT_TIMESTAMP AT TIME ZONE 'UTC' + INTERVAL '2 weeks', $3, +(SELECT id FROM users WHERE users.uuid=$4 LIMIT 1), +0, CURRENT_TIMESTAMP AT TIME ZONE 'UTC', CURRENT_TIMESTAMP AT TIME ZONE 'UTC')`, + uuid, token, string(scopesjson), userUUID) + + if err != nil { + return nil, err + } + + return &arvados.APIClientAuthorization{ + UUID: uuid, + APIToken: token, + Scopes: scopes}, nil } // Extract the auth token supplied in req, and replace it with a // salted token for the remote cluster. -func (h *Handler) saltAuthToken(req *http.Request, remote string) error { +func (h *Handler) saltAuthToken(req *http.Request, remote string) (updatedReq *http.Request, err error) { + updatedReq = (&http.Request{ + Method: req.Method, + URL: req.URL, + Header: req.Header, + Body: req.Body, + ContentLength: req.ContentLength, + Host: req.Host, + }).WithContext(req.Context()) + creds := auth.NewCredentials() - creds.LoadTokensFromHTTPRequest(req) - if len(creds.Tokens) == 0 && req.Header.Get("Content-Type") == "application/x-www-form-encoded" { + creds.LoadTokensFromHTTPRequest(updatedReq) + if len(creds.Tokens) == 0 && updatedReq.Header.Get("Content-Type") == "application/x-www-form-encoded" { // Override ParseForm's 10MiB limit by ensuring // req.Body is a *http.maxBytesReader. - req.Body = http.MaxBytesReader(nil, req.Body, 1<<28) // 256MiB. TODO: use MaxRequestSize from discovery doc or config. - if err := creds.LoadTokensFromHTTPRequestBody(req); err != nil { - return err + updatedReq.Body = http.MaxBytesReader(nil, updatedReq.Body, 1<<28) // 256MiB. TODO: use MaxRequestSize from discovery doc or config. + if err := creds.LoadTokensFromHTTPRequestBody(updatedReq); err != nil { + return nil, err } // Replace req.Body with a buffer that re-encodes the // form without api_token, in case we end up // forwarding the request. - if req.PostForm != nil { - req.PostForm.Del("api_token") + if updatedReq.PostForm != nil { + updatedReq.PostForm.Del("api_token") } - req.Body = ioutil.NopCloser(bytes.NewBufferString(req.PostForm.Encode())) + updatedReq.Body = ioutil.NopCloser(bytes.NewBufferString(updatedReq.PostForm.Encode())) } if len(creds.Tokens) == 0 { - return nil + return updatedReq, nil } + + ctxlog.FromContext(req.Context()).Debugf("saltAuthToken: cluster %s token %s remote %s", h.Cluster.ClusterID, creds.Tokens[0], remote) token, err := auth.SaltToken(creds.Tokens[0], remote) - if err == auth.ErrObsoleteToken { - // If the token exists in our own database, salt it - // for the remote. Otherwise, assume it was issued by - // the remote, and pass it through unmodified. - currentUser := CurrentUser{Authorization: arvados.APIClientAuthorization{APIToken: creds.Tokens[0]}} - err = h.validateAPItoken(req, ¤tUser) - if err == sql.ErrNoRows { - // Not ours; pass through unmodified. - token = currentUser.Authorization.APIToken - } else if err != nil { - return err + + if err == auth.ErrObsoleteToken || err == auth.ErrTokenFormat { + // If the token exists in our own database for our own + // user, salt it for the remote. Otherwise, assume it + // was issued by the remote, and pass it through + // unmodified. + currentUser, ok, err := h.validateAPItoken(req, creds.Tokens[0]) + if err != nil { + return nil, err + } else if !ok || strings.HasPrefix(currentUser.UUID, remote) { + // Unknown, or cached + belongs to remote; + // pass through unmodified. + token = creds.Tokens[0] } else { // Found; make V2 version and salt it. token, err = auth.SaltToken(currentUser.Authorization.TokenV2(), remote) if err != nil { - return err + return nil, err } } } else if err != nil { - return err + return nil, err + } + updatedReq.Header = http.Header{} + for k, v := range req.Header { + if k != "Authorization" { + updatedReq.Header[k] = v + } } - req.Header.Set("Authorization", "Bearer "+token) + updatedReq.Header.Set("Authorization", "Bearer "+token) - // Remove api_token=... from the the query string, in case we + // Remove api_token=... from the query string, in case we // end up forwarding the request. - if values, err := url.ParseQuery(req.URL.RawQuery); err != nil { - return err + if values, err := url.ParseQuery(updatedReq.URL.RawQuery); err != nil { + return nil, err } else if _, ok := values["api_token"]; ok { delete(values, "api_token") - req.URL.RawQuery = values.Encode() + updatedReq.URL = &url.URL{ + Scheme: req.URL.Scheme, + Host: req.URL.Host, + Path: req.URL.Path, + RawPath: req.URL.RawPath, + RawQuery: values.Encode(), + } } - return nil + return updatedReq, nil }