X-Git-Url: https://git.arvados.org/arvados.git/blobdiff_plain/0446c0a3a433936985d6f46b0eab9b253ed98e80..b1ed7c643f311605092991e01bcc3437130d6072:/lib/boot/cmd.go diff --git a/lib/boot/cmd.go b/lib/boot/cmd.go index 4d2c01f2c5..6a32ab142d 100644 --- a/lib/boot/cmd.go +++ b/lib/boot/cmd.go @@ -5,609 +5,120 @@ package boot import ( - "bytes" "context" - "crypto/rand" - "encoding/json" + "errors" "flag" "fmt" "io" - "io/ioutil" - "net" - "os" - "os/exec" - "os/signal" - "path/filepath" - "strings" - "sync" - "syscall" "time" "git.arvados.org/arvados.git/lib/cmd" - "git.arvados.org/arvados.git/lib/config" - "git.arvados.org/arvados.git/lib/controller" - "git.arvados.org/arvados.git/lib/dispatchcloud" - "git.arvados.org/arvados.git/sdk/go/arvados" "git.arvados.org/arvados.git/sdk/go/ctxlog" - "git.arvados.org/arvados.git/sdk/go/health" - "github.com/sirupsen/logrus" ) var Command cmd.Handler = bootCommand{} -type bootCommand struct{} - -func (bootCommand) RunCommand(prog string, args []string, stdin io.Reader, stdout, stderr io.Writer) int { - boot := &Booter{ - Stderr: stderr, - logger: ctxlog.New(stderr, "json", "info"), - } - - ctx := ctxlog.Context(context.Background(), boot.logger) - ctx, cancel := context.WithCancel(ctx) - defer cancel() - - ch := make(chan os.Signal) - signal.Notify(ch, syscall.SIGINT, syscall.SIGTERM) - go func() { - for sig := range ch { - boot.logger.WithField("signal", sig).Info("caught signal") - cancel() - } - }() - - var err error - defer func() { - if err != nil { - boot.logger.WithError(err).Info("exiting") - } - }() - - flags := flag.NewFlagSet(prog, flag.ContinueOnError) - flags.SetOutput(stderr) - loader := config.NewLoader(stdin, boot.logger) - loader.SetupFlags(flags) - versionFlag := flags.Bool("version", false, "Write version information to stdout and exit 0") - flags.StringVar(&boot.SourcePath, "source", ".", "arvados source tree `directory`") - flags.StringVar(&boot.LibPath, "lib", "/var/lib/arvados", "`directory` to install dependencies and library files") - flags.StringVar(&boot.ClusterType, "type", "production", "cluster `type`: development, test, or production") - err = flags.Parse(args) - if err == flag.ErrHelp { - err = nil - return 0 - } else if err != nil { - return 2 - } else if *versionFlag { - return cmd.Version.RunCommand(prog, args, stdin, stdout, stderr) - } else if boot.ClusterType != "development" && boot.ClusterType != "test" && boot.ClusterType != "production" { - err = fmt.Errorf("cluster type must be 'development', 'test', or 'production'") - return 2 - } - - boot.Start(ctx, loader) - defer boot.Stop() - if boot.WaitReady() { - fmt.Fprintln(stdout, boot.cluster.Services.Controller.ExternalURL) - <-ctx.Done() // wait for signal - return 0 - } else { - return 1 - } +type supervisedTask interface { + // Execute the task. Run should return nil when the task is + // done enough to satisfy a dependency relationship (e.g., the + // service is running and ready). If the task starts a + // goroutine that fails after Run returns (e.g., the service + // shuts down), it should call fail(). + Run(ctx context.Context, fail func(error), super *Supervisor) error + String() string } -type Booter struct { - SourcePath string // e.g., /home/username/src/arvados - LibPath string // e.g., /var/lib/arvados - ClusterType string // e.g., production - Stderr io.Writer - - logger logrus.FieldLogger - cluster *arvados.Cluster - - ctx context.Context - cancel context.CancelFunc - done chan struct{} - healthChecker *health.Aggregator - - tempdir string - configfile string - environ []string // for child processes - - setupRubyOnce sync.Once - setupRubyErr error - goMutex sync.Mutex -} - -func (boot *Booter) Start(ctx context.Context, loader *config.Loader) { - boot.ctx, boot.cancel = context.WithCancel(ctx) - boot.done = make(chan struct{}) - go func() { - err := boot.run(loader) - if err != nil { - fmt.Fprintln(boot.Stderr, err) - } - close(boot.done) - }() -} - -func (boot *Booter) run(loader *config.Loader) error { - cwd, err := os.Getwd() - if err != nil { - return err - } - if !strings.HasPrefix(boot.SourcePath, "/") { - boot.SourcePath = filepath.Join(cwd, boot.SourcePath) - } - boot.SourcePath, err = filepath.EvalSymlinks(boot.SourcePath) - if err != nil { - return err - } - - boot.tempdir, err = ioutil.TempDir("", "arvados-server-boot-") - if err != nil { - return err - } - defer os.RemoveAll(boot.tempdir) - - loader.SkipAPICalls = true - cfg, err := loader.Load() - if err != nil { - return err - } - - // Fill in any missing config keys, and write the resulting - // config in the temp dir for child services to use. - err = boot.autofillConfig(cfg, boot.logger) - if err != nil { - return err - } - conffile, err := os.OpenFile(filepath.Join(boot.tempdir, "config.yml"), os.O_CREATE|os.O_WRONLY, 0777) - if err != nil { - return err - } - defer conffile.Close() - err = json.NewEncoder(conffile).Encode(cfg) - if err != nil { - return err - } - err = conffile.Close() - if err != nil { - return err - } - boot.configfile = conffile.Name() - - boot.environ = os.Environ() - boot.setEnv("ARVADOS_CONFIG", boot.configfile) - boot.setEnv("RAILS_ENV", boot.ClusterType) - boot.prependEnv("PATH", filepath.Join(boot.LibPath, "bin")+":") - - boot.cluster, err = cfg.GetCluster("") - if err != nil { - return err - } - // Now that we have the config, replace the bootstrap logger - // with a new one according to the logging config. - boot.logger = ctxlog.New(boot.Stderr, boot.cluster.SystemLogs.Format, boot.cluster.SystemLogs.LogLevel).WithFields(logrus.Fields{ - "PID": os.Getpid(), - }) - boot.healthChecker = &health.Aggregator{Cluster: boot.cluster} - - for _, dir := range []string{boot.LibPath, filepath.Join(boot.LibPath, "bin")} { - if _, err = os.Stat(filepath.Join(dir, ".")); os.IsNotExist(err) { - err = os.Mkdir(dir, 0755) - if err != nil { - return err - } - } else if err != nil { - return err - } - } - err = boot.installGoProgram(boot.ctx, "cmd/arvados-server") - if err != nil { - return err - } - err = boot.setupRubyEnv() - if err != nil { - return err - } - - var wg sync.WaitGroup - for _, cmpt := range []component{ - {name: "nginx", runFunc: runNginx}, - {name: "controller", cmdHandler: controller.Command}, - {name: "dispatchcloud", cmdHandler: dispatchcloud.Command, notIfTest: true}, - {name: "git-httpd", goProg: "services/arv-git-httpd"}, - {name: "health", goProg: "services/health"}, - {name: "keep-balance", goProg: "services/keep-balance", notIfTest: true}, - {name: "keepproxy", goProg: "services/keepproxy"}, - {name: "keepstore", goProg: "services/keepstore", svc: boot.cluster.Services.Keepstore}, - {name: "keep-web", goProg: "services/keep-web"}, - {name: "railsAPI", svc: boot.cluster.Services.RailsAPI, railsApp: "services/api"}, - {name: "workbench1", svc: boot.cluster.Services.Workbench1, railsApp: "apps/workbench"}, - {name: "ws", goProg: "services/ws"}, - } { - cmpt := cmpt - wg.Add(1) - go func() { - defer wg.Done() - defer boot.cancel() - boot.logger.WithField("component", cmpt.name).Info("starting") - err := cmpt.Run(boot.ctx, boot) - if err != nil && err != context.Canceled { - boot.logger.WithError(err).WithField("component", cmpt.name).Error("exited") - } - }() - } - wg.Wait() - return nil -} +var errNeedConfigReload = errors.New("config changed, restart needed") +var errParseFlags = errors.New("error parsing command line arguments") -func (boot *Booter) Stop() { - boot.cancel() - <-boot.done -} +type bootCommand struct{} -func (boot *Booter) WaitReady() bool { - for waiting := true; waiting; { - time.Sleep(time.Second) - if boot.ctx.Err() != nil { - return false - } - if boot.healthChecker == nil { - // not set up yet +func (bcmd bootCommand) RunCommand(prog string, args []string, stdin io.Reader, stdout, stderr io.Writer) int { + logger := ctxlog.New(stderr, "json", "info") + ctx := ctxlog.Context(context.Background(), logger) + for { + err := bcmd.run(ctx, prog, args, stdin, stdout, stderr) + if err == errNeedConfigReload { continue - } - resp := boot.healthChecker.ClusterHealth() - // The overall health check (resp.Health=="OK") might - // never pass due to missing components (like - // arvados-dispatch-cloud in a test cluster), so - // instead we wait for all configured components to - // pass. - waiting = false - for _, check := range resp.Checks { - if check.Health != "OK" { - waiting = true - } - } - } - return true -} - -func (boot *Booter) prependEnv(key, prepend string) { - for i, s := range boot.environ { - if strings.HasPrefix(s, key+"=") { - boot.environ[i] = key + "=" + prepend + s[len(key)+1:] - return - } - } - boot.environ = append(boot.environ, key+"="+prepend) -} - -func (boot *Booter) setEnv(key, val string) { - for i, s := range boot.environ { - if strings.HasPrefix(s, key+"=") { - boot.environ[i] = key + "=" + val - return - } - } - boot.environ = append(boot.environ, key+"="+val) -} - -func (boot *Booter) installGoProgram(ctx context.Context, srcpath string) error { - boot.goMutex.Lock() - defer boot.goMutex.Unlock() - return boot.RunProgram(ctx, filepath.Join(boot.SourcePath, srcpath), nil, []string{"GOPATH=" + boot.LibPath}, "go", "install") -} - -func (boot *Booter) setupRubyEnv() error { - buf, err := exec.Command("gem", "env", "gempath").Output() // /var/lib/arvados/.gem/ruby/2.5.0/bin:... - if err != nil || len(buf) == 0 { - return fmt.Errorf("gem env gempath: %v", err) - } - gempath := string(bytes.Split(buf, []byte{':'})[0]) - boot.prependEnv("PATH", gempath+"/bin:") - boot.setEnv("GEM_HOME", gempath) - boot.setEnv("GEM_PATH", gempath) - return nil -} - -func (boot *Booter) lookPath(prog string) string { - for _, val := range boot.environ { - if strings.HasPrefix(val, "PATH=") { - for _, dir := range filepath.SplitList(val[5:]) { - path := filepath.Join(dir, prog) - if fi, err := os.Stat(path); err == nil && fi.Mode()&0111 != 0 { - return path - } - } + } else if err == errParseFlags { + return 2 + } else if err != nil { + logger.WithError(err).Info("exiting") + return 1 + } else { + return 0 } } - return prog } -// Run prog with args, using dir as working directory. If ctx is -// cancelled while the child is running, RunProgram terminates the -// child, waits for it to exit, then returns. -// -// Child's environment will have our env vars, plus any given in env. -// -// Child's stdout will be written to output if non-nil, otherwise the -// boot command's stderr. -func (boot *Booter) RunProgram(ctx context.Context, dir string, output io.Writer, env []string, prog string, args ...string) error { - cmdline := fmt.Sprintf("%s", append([]string{prog}, args...)) - fmt.Fprintf(boot.Stderr, "%s executing in %s\n", cmdline, dir) - cmd := exec.Command(boot.lookPath(prog), args...) - if output == nil { - cmd.Stdout = boot.Stderr - } else { - cmd.Stdout = output - } - cmd.Stderr = boot.Stderr - if strings.HasPrefix(dir, "/") { - cmd.Dir = dir - } else { - cmd.Dir = filepath.Join(boot.SourcePath, dir) - } - cmd.Env = append(env, boot.environ...) - - exited := false - defer func() { exited = true }() - go func() { - <-ctx.Done() - log := ctxlog.FromContext(ctx).WithFields(logrus.Fields{"dir": dir, "cmdline": cmdline}) - for !exited { - if cmd.Process == nil { - log.Debug("waiting for child process to start") - time.Sleep(time.Second / 2) - } else { - log.WithField("PID", cmd.Process.Pid).Debug("sending SIGTERM") - cmd.Process.Signal(syscall.SIGTERM) - time.Sleep(5 * time.Second) - if !exited { - log.WithField("PID", cmd.Process.Pid).Warn("still waiting for child process to exit 5s after SIGTERM") - } - } - } - }() - - err := cmd.Run() - if err != nil && ctx.Err() == nil { - // Only report errors that happen before the context ends. - return fmt.Errorf("%s: error: %v", cmdline, err) +func (bcmd bootCommand) run(ctx context.Context, prog string, args []string, stdin io.Reader, stdout, stderr io.Writer) error { + ctx, cancel := context.WithCancel(ctx) + defer cancel() + super := &Supervisor{ + Stdin: stdin, + Stderr: stderr, + logger: ctxlog.FromContext(ctx), } - return nil -} - -type component struct { - name string - svc arvados.Service - cmdHandler cmd.Handler - runFunc func(ctx context.Context, boot *Booter) error - railsApp string // source dir in arvados tree, e.g., "services/api" - goProg string // source dir in arvados tree, e.g., "services/keepstore" - notIfTest bool // don't run this component on a test cluster -} -func (cmpt *component) Run(ctx context.Context, boot *Booter) error { - if cmpt.notIfTest && boot.ClusterType == "test" { - fmt.Fprintf(boot.Stderr, "skipping component %q in %s mode\n", cmpt.name, boot.ClusterType) - <-ctx.Done() - return nil - } - fmt.Fprintf(boot.Stderr, "starting component %q\n", cmpt.name) - if cmpt.cmdHandler != nil { - errs := make(chan error, 1) - go func() { - defer close(errs) - exitcode := cmpt.cmdHandler.RunCommand(cmpt.name, []string{"-config", boot.configfile}, bytes.NewBuffer(nil), boot.Stderr, boot.Stderr) - if exitcode != 0 { - errs <- fmt.Errorf("exit code %d", exitcode) - } - }() - select { - case err := <-errs: - return err - case <-ctx.Done(): - // cmpt.cmdHandler.RunCommand() doesn't have - // access to our context, so it won't shut - // down by itself. We just abandon it. - return nil - } - } - if cmpt.goProg != "" { - boot.RunProgram(ctx, cmpt.goProg, nil, nil, "go", "install") - if ctx.Err() != nil { + flags := flag.NewFlagSet(prog, flag.ContinueOnError) + versionFlag := flags.Bool("version", false, "Write version information to stdout and exit 0") + flags.StringVar(&super.ConfigPath, "config", "/etc/arvados/config.yml", "arvados config file `path`") + flags.StringVar(&super.SourcePath, "source", ".", "arvados source tree `directory`") + flags.StringVar(&super.ClusterType, "type", "production", "cluster `type`: development, test, or production") + flags.StringVar(&super.ListenHost, "listen-host", "localhost", "host name or interface address for external services, and internal services whose InternalURLs are not configured") + flags.StringVar(&super.ControllerAddr, "controller-address", ":0", "desired controller address, `host:port` or `:port`") + flags.StringVar(&super.Workbench2Source, "workbench2-source", "../arvados-workbench2", "path to arvados-workbench2 source tree") + flags.BoolVar(&super.NoWorkbench1, "no-workbench1", false, "do not run workbench1") + flags.BoolVar(&super.NoWorkbench2, "no-workbench2", true, "do not run workbench2") + flags.BoolVar(&super.OwnTemporaryDatabase, "own-temporary-database", false, "bring up a postgres server and create a temporary database") + timeout := flags.Duration("timeout", 0, "maximum time to wait for cluster to be ready") + shutdown := flags.Bool("shutdown", false, "shut down when the cluster becomes ready") + if ok, code := cmd.ParseFlags(flags, prog, args, "", stderr); !ok { + if code == 0 { return nil - } - _, basename := filepath.Split(cmpt.goProg) - if len(cmpt.svc.InternalURLs) > 0 { - // Run one for each URL - var wg sync.WaitGroup - for u := range cmpt.svc.InternalURLs { - u := u - wg.Add(1) - go func() { - defer wg.Done() - boot.RunProgram(ctx, boot.tempdir, nil, []string{"ARVADOS_SERVICE_INTERNAL_URL=" + u.String()}, basename) - }() - } - wg.Wait() } else { - // Just run one - boot.RunProgram(ctx, boot.tempdir, nil, nil, basename) - } - return nil - } - if cmpt.runFunc != nil { - return cmpt.runFunc(ctx, boot) - } - if cmpt.railsApp != "" { - port, err := internalPort(cmpt.svc) - if err != nil { - return fmt.Errorf("bug: no InternalURLs for component %q: %v", cmpt.name, cmpt.svc.InternalURLs) - } - var buf bytes.Buffer - err = boot.RunProgram(ctx, cmpt.railsApp, &buf, nil, "gem", "list", "--details", "bundler") - if err != nil { - return err - } - for _, version := range []string{"1.11.0", "1.17.3", "2.0.2"} { - if !strings.Contains(buf.String(), "("+version+")") { - err = boot.RunProgram(ctx, cmpt.railsApp, nil, nil, "gem", "install", "--user", "bundler:1.11", "bundler:1.17.3", "bundler:2.0.2") - if err != nil { - return err - } - break - } - } - err = boot.RunProgram(ctx, cmpt.railsApp, nil, nil, "bundle", "install", "--jobs", "4", "--path", filepath.Join(os.Getenv("HOME"), ".gem")) - if err != nil { - return err - } - err = boot.RunProgram(ctx, cmpt.railsApp, nil, nil, "bundle", "exec", "passenger-config", "build-native-support") - if err != nil { - return err - } - err = boot.RunProgram(ctx, cmpt.railsApp, nil, nil, "bundle", "exec", "passenger-config", "install-standalone-runtime") - if err != nil { - return err - } - err = boot.RunProgram(ctx, cmpt.railsApp, nil, nil, "bundle", "exec", "passenger-config", "validate-install") - if err != nil { - return err - } - err = boot.RunProgram(ctx, cmpt.railsApp, nil, nil, "bundle", "exec", "passenger", "start", "-p", port) - if err != nil { - return err + return errParseFlags } + } else if *versionFlag { + cmd.Version.RunCommand(prog, args, stdin, stdout, stderr) return nil + } else if super.ClusterType != "development" && super.ClusterType != "test" && super.ClusterType != "production" { + return fmt.Errorf("cluster type must be 'development', 'test', or 'production'") } - return fmt.Errorf("bug: component %q has nothing to run", cmpt.name) -} - -func (boot *Booter) autofillConfig(cfg *arvados.Config, log logrus.FieldLogger) error { - cluster, err := cfg.GetCluster("") - if err != nil { - return err - } - port := 9000 - for _, svc := range []*arvados.Service{ - &cluster.Services.Controller, - &cluster.Services.DispatchCloud, - &cluster.Services.GitHTTP, - &cluster.Services.Health, - &cluster.Services.Keepproxy, - &cluster.Services.Keepstore, - &cluster.Services.RailsAPI, - &cluster.Services.WebDAV, - &cluster.Services.WebDAVDownload, - &cluster.Services.Websocket, - &cluster.Services.Workbench1, - } { - if svc == &cluster.Services.DispatchCloud && boot.ClusterType == "test" { - continue - } - if len(svc.InternalURLs) == 0 { - port++ - svc.InternalURLs = map[arvados.URL]arvados.ServiceInstance{ - arvados.URL{Scheme: "http", Host: fmt.Sprintf("localhost:%d", port)}: arvados.ServiceInstance{}, - } - } - if svc.ExternalURL.Host == "" && (svc == &cluster.Services.Controller || - svc == &cluster.Services.GitHTTP || - svc == &cluster.Services.Keepproxy || - svc == &cluster.Services.WebDAV || - svc == &cluster.Services.WebDAVDownload || - svc == &cluster.Services.Websocket || - svc == &cluster.Services.Workbench1) { - port++ - svc.ExternalURL = arvados.URL{Scheme: "https", Host: fmt.Sprintf("localhost:%d", port)} - } - } - if cluster.SystemRootToken == "" { - cluster.SystemRootToken = randomHexString(64) - } - if cluster.ManagementToken == "" { - cluster.ManagementToken = randomHexString(64) - } - if cluster.API.RailsSessionSecretToken == "" { - cluster.API.RailsSessionSecretToken = randomHexString(64) - } - if cluster.Collections.BlobSigningKey == "" { - cluster.Collections.BlobSigningKey = randomHexString(64) - } - if boot.ClusterType != "production" && cluster.Containers.DispatchPrivateKey == "" { - buf, err := ioutil.ReadFile(filepath.Join(boot.SourcePath, "lib", "dispatchcloud", "test", "sshkey_dispatch")) - if err != nil { - return err - } - cluster.Containers.DispatchPrivateKey = string(buf) - } - if boot.ClusterType != "production" { - cluster.TLS.Insecure = true - } - if boot.ClusterType == "test" { - // Add a second keepstore process. - port++ - cluster.Services.Keepstore.InternalURLs[arvados.URL{Scheme: "http", Host: fmt.Sprintf("localhost:%d", port)}] = arvados.ServiceInstance{} - // Create a directory-backed volume for each keepstore - // process. - cluster.Volumes = map[string]arvados.Volume{} - for url := range cluster.Services.Keepstore.InternalURLs { - volnum := len(cluster.Volumes) - datadir := fmt.Sprintf("%s/keep%d.data", boot.tempdir, volnum) - if _, err = os.Stat(datadir + "/."); err == nil { - } else if !os.IsNotExist(err) { - return err - } else if err = os.Mkdir(datadir, 0777); err != nil { - return err - } - cluster.Volumes[fmt.Sprintf("zzzzz-nyw5e-%015d", volnum)] = arvados.Volume{ - Driver: "Directory", - DriverParameters: json.RawMessage(fmt.Sprintf(`{"Root":%q}`, datadir)), - AccessViaHosts: map[arvados.URL]arvados.VolumeAccess{ - url: {}, - }, - } - } - } - cfg.Clusters[cluster.ClusterID] = *cluster - return nil -} + super.Start(ctx) + defer super.Stop() -func randomHexString(chars int) string { - b := make([]byte, chars/2) - _, err := rand.Read(b) - if err != nil { - panic(err) + var timer *time.Timer + if *timeout > 0 { + timer = time.AfterFunc(*timeout, super.Stop) } - return fmt.Sprintf("%x", b) -} - -func internalPort(svc arvados.Service) (string, error) { - for u := range svc.InternalURLs { - if _, p, err := net.SplitHostPort(u.Host); err != nil { - return "", err - } else if p != "" { - return p, nil - } else if u.Scheme == "https" { - return "443", nil - } else { - return "80", nil - } - } - return "", fmt.Errorf("service has no InternalURLs") -} -func externalPort(svc arvados.Service) (string, error) { - if _, p, err := net.SplitHostPort(svc.ExternalURL.Host); err != nil { - return "", err - } else if p != "" { - return p, nil - } else if svc.ExternalURL.Scheme == "https" { - return "443", nil + ok := super.WaitReady() + if timer != nil && !timer.Stop() { + return errors.New("boot timed out") + } else if !ok { + super.logger.Error("boot failed") } else { - return "80", nil + // Write each cluster's controller URL to stdout. + // Nothing else goes to stdout, so this allows a + // calling script to determine when the cluster is + // ready to use, and the controller's host:port (which + // may have been dynamically assigned depending on + // config/options). + for _, cc := range super.Clusters() { + fmt.Fprintln(stdout, cc.Services.Controller.ExternalURL) + } + if *shutdown { + super.Stop() + // Wait for children to exit. Don't report the + // ensuing "context cancelled" error, though: + // return nil to indicate successful startup. + _ = super.Wait() + fmt.Fprintln(stderr, "PASS - all services booted successfully") + return nil + } } + // Wait for signal/crash + orderly shutdown + return super.Wait() }