X-Git-Url: https://git.arvados.org/arvados.git/blobdiff_plain/d2d20f664ec207d1c9492edea6d893b1e333b055..9d1ff3299a57d0e820bf7975f0f3e6080b22f0a5:/lib/cloud/ec2/ec2.go diff --git a/lib/cloud/ec2/ec2.go b/lib/cloud/ec2/ec2.go index 071c95006c..6251f18df0 100644 --- a/lib/cloud/ec2/ec2.go +++ b/lib/cloud/ec2/ec2.go @@ -13,6 +13,9 @@ import ( "encoding/json" "fmt" "math/big" + "regexp" + "strconv" + "strings" "sync" "sync/atomic" "time" @@ -20,12 +23,14 @@ import ( "git.arvados.org/arvados.git/lib/cloud" "git.arvados.org/arvados.git/sdk/go/arvados" "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/awserr" "github.com/aws/aws-sdk-go/aws/credentials" "github.com/aws/aws-sdk-go/aws/credentials/ec2rolecreds" "github.com/aws/aws-sdk-go/aws/ec2metadata" "github.com/aws/aws-sdk-go/aws/request" "github.com/aws/aws-sdk-go/aws/session" "github.com/aws/aws-sdk-go/service/ec2" + "github.com/prometheus/client_golang/prometheus" "github.com/sirupsen/logrus" "golang.org/x/crypto/ssh" ) @@ -39,13 +44,49 @@ const ( ) type ec2InstanceSetConfig struct { - AccessKeyID string - SecretAccessKey string - Region string - SecurityGroupIDs arvados.StringSet - SubnetID string - AdminUsername string - EBSVolumeType string + AccessKeyID string + SecretAccessKey string + Region string + SecurityGroupIDs arvados.StringSet + SubnetID sliceOrSingleString + AdminUsername string + EBSVolumeType string + EBSPrice float64 + IAMInstanceProfile string + SpotPriceUpdateInterval arvados.Duration +} + +type sliceOrSingleString []string + +// UnmarshalJSON unmarshals an array of strings, and also accepts "" +// as [], and "foo" as ["foo"]. +func (ss *sliceOrSingleString) UnmarshalJSON(data []byte) error { + if len(data) == 0 { + *ss = nil + } else if data[0] == '[' { + var slice []string + err := json.Unmarshal(data, &slice) + if err != nil { + return err + } + if len(slice) == 0 { + *ss = nil + } else { + *ss = slice + } + } else { + var str string + err := json.Unmarshal(data, &str) + if err != nil { + return err + } + if str == "" { + *ss = nil + } else { + *ss = []string{str} + } + } + return nil } type ec2Interface interface { @@ -53,12 +94,15 @@ type ec2Interface interface { ImportKeyPair(input *ec2.ImportKeyPairInput) (*ec2.ImportKeyPairOutput, error) RunInstances(input *ec2.RunInstancesInput) (*ec2.Reservation, error) DescribeInstances(input *ec2.DescribeInstancesInput) (*ec2.DescribeInstancesOutput, error) + DescribeInstanceStatusPages(input *ec2.DescribeInstanceStatusInput, fn func(*ec2.DescribeInstanceStatusOutput, bool) bool) error + DescribeSpotPriceHistoryPages(input *ec2.DescribeSpotPriceHistoryInput, fn func(*ec2.DescribeSpotPriceHistoryOutput, bool) bool) error CreateTags(input *ec2.CreateTagsInput) (*ec2.CreateTagsOutput, error) TerminateInstances(input *ec2.TerminateInstancesInput) (*ec2.TerminateInstancesOutput, error) } type ec2InstanceSet struct { ec2config ec2InstanceSetConfig + currentSubnetIDIndex int32 instanceSetID cloud.InstanceSetID logger logrus.FieldLogger client ec2Interface @@ -66,9 +110,16 @@ type ec2InstanceSet struct { keys map[string]string throttleDelayCreate atomic.Value throttleDelayInstances atomic.Value + + prices map[priceKey][]cloud.InstancePrice + pricesLock sync.Mutex + pricesUpdated map[priceKey]time.Time + + mInstances *prometheus.GaugeVec + mInstanceStarts *prometheus.CounterVec } -func newEC2InstanceSet(config json.RawMessage, instanceSetID cloud.InstanceSetID, _ cloud.SharedResourceTags, logger logrus.FieldLogger) (prv cloud.InstanceSet, err error) { +func newEC2InstanceSet(config json.RawMessage, instanceSetID cloud.InstanceSetID, _ cloud.SharedResourceTags, logger logrus.FieldLogger, reg *prometheus.Registry) (prv cloud.InstanceSet, err error) { instanceSet := &ec2InstanceSet{ instanceSetID: instanceSetID, logger: logger, @@ -95,6 +146,36 @@ func newEC2InstanceSet(config json.RawMessage, instanceSetID cloud.InstanceSetID if instanceSet.ec2config.EBSVolumeType == "" { instanceSet.ec2config.EBSVolumeType = "gp2" } + + // Set up metrics + instanceSet.mInstances = prometheus.NewGaugeVec(prometheus.GaugeOpts{ + Namespace: "arvados", + Subsystem: "dispatchcloud", + Name: "ec2_instances", + Help: "Number of instances running", + }, []string{"subnet_id"}) + instanceSet.mInstanceStarts = prometheus.NewCounterVec(prometheus.CounterOpts{ + Namespace: "arvados", + Subsystem: "dispatchcloud", + Name: "ec2_instance_starts_total", + Help: "Number of attempts to start a new instance", + }, []string{"subnet_id", "success"}) + // Initialize all of the series we'll be reporting. Otherwise + // the {subnet=A, success=0} series doesn't appear in metrics + // at all until there's a failure in subnet A. + for _, subnet := range instanceSet.ec2config.SubnetID { + instanceSet.mInstanceStarts.WithLabelValues(subnet, "0").Add(0) + instanceSet.mInstanceStarts.WithLabelValues(subnet, "1").Add(0) + } + if len(instanceSet.ec2config.SubnetID) == 0 { + instanceSet.mInstanceStarts.WithLabelValues("", "0").Add(0) + instanceSet.mInstanceStarts.WithLabelValues("", "1").Add(0) + } + if reg != nil { + reg.MustRegister(instanceSet.mInstances) + reg.MustRegister(instanceSet.mInstanceStarts) + } + return instanceSet, nil } @@ -138,40 +219,6 @@ func (instanceSet *ec2InstanceSet) Create( initCommand cloud.InitCommand, publicKey ssh.PublicKey) (cloud.Instance, error) { - md5keyFingerprint, sha1keyFingerprint, err := awsKeyFingerprint(publicKey) - if err != nil { - return nil, fmt.Errorf("Could not make key fingerprint: %v", err) - } - instanceSet.keysMtx.Lock() - var keyname string - var ok bool - if keyname, ok = instanceSet.keys[md5keyFingerprint]; !ok { - keyout, err := instanceSet.client.DescribeKeyPairs(&ec2.DescribeKeyPairsInput{ - Filters: []*ec2.Filter{{ - Name: aws.String("fingerprint"), - Values: []*string{&md5keyFingerprint, &sha1keyFingerprint}, - }}, - }) - if err != nil { - return nil, fmt.Errorf("Could not search for keypair: %v", err) - } - - if len(keyout.KeyPairs) > 0 { - keyname = *(keyout.KeyPairs[0].KeyName) - } else { - keyname = "arvados-dispatch-keypair-" + md5keyFingerprint - _, err := instanceSet.client.ImportKeyPair(&ec2.ImportKeyPairInput{ - KeyName: &keyname, - PublicKeyMaterial: ssh.MarshalAuthorizedKey(publicKey), - }) - if err != nil { - return nil, fmt.Errorf("Could not import keypair: %v", err) - } - } - instanceSet.keys[md5keyFingerprint] = keyname - } - instanceSet.keysMtx.Unlock() - ec2tags := []*ec2.Tag{} for k, v := range newTags { ec2tags = append(ec2tags, &ec2.Tag{ @@ -190,7 +237,6 @@ func (instanceSet *ec2InstanceSet) Create( InstanceType: &instanceType.ProviderType, MaxCount: aws.Int64(1), MinCount: aws.Int64(1), - KeyName: &keyname, NetworkInterfaces: []*ec2.InstanceNetworkInterfaceSpecification{ { @@ -198,7 +244,6 @@ func (instanceSet *ec2InstanceSet) Create( DeleteOnTermination: aws.Bool(true), DeviceIndex: aws.Int64(0), Groups: aws.StringSlice(groups), - SubnetId: &instanceSet.ec2config.SubnetID, }}, DisableApiTermination: aws.Bool(false), InstanceInitiatedShutdownBehavior: aws.String("terminate"), @@ -207,9 +252,23 @@ func (instanceSet *ec2InstanceSet) Create( ResourceType: aws.String("instance"), Tags: ec2tags, }}, + MetadataOptions: &ec2.InstanceMetadataOptionsRequest{ + // Require IMDSv2, as described at + // https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/configuring-IMDS-new-instances.html + HttpEndpoint: aws.String(ec2.InstanceMetadataEndpointStateEnabled), + HttpTokens: aws.String(ec2.HttpTokensStateRequired), + }, UserData: aws.String(base64.StdEncoding.EncodeToString([]byte("#!/bin/sh\n" + initCommand + "\n"))), } + if publicKey != nil { + keyname, err := instanceSet.getKeyName(publicKey) + if err != nil { + return nil, err + } + rii.KeyName = &keyname + } + if instanceType.AddedScratch > 0 { rii.BlockDeviceMappings = []*ec2.BlockDeviceMapping{{ DeviceName: aws.String("/dev/xvdt"), @@ -229,18 +288,95 @@ func (instanceSet *ec2InstanceSet) Create( }} } - rsv, err := instanceSet.client.RunInstances(&rii) - err = wrapError(err, &instanceSet.throttleDelayCreate) - if err != nil { - return nil, err + if instanceSet.ec2config.IAMInstanceProfile != "" { + rii.IamInstanceProfile = &ec2.IamInstanceProfileSpecification{ + Name: aws.String(instanceSet.ec2config.IAMInstanceProfile), + } } + var rsv *ec2.Reservation + var errToReturn error + subnets := instanceSet.ec2config.SubnetID + currentSubnetIDIndex := int(atomic.LoadInt32(&instanceSet.currentSubnetIDIndex)) + for tryOffset := 0; ; tryOffset++ { + tryIndex := 0 + trySubnet := "" + if len(subnets) > 0 { + tryIndex = (currentSubnetIDIndex + tryOffset) % len(subnets) + trySubnet = subnets[tryIndex] + rii.NetworkInterfaces[0].SubnetId = aws.String(trySubnet) + } + var err error + rsv, err = instanceSet.client.RunInstances(&rii) + instanceSet.mInstanceStarts.WithLabelValues(trySubnet, boolLabelValue[err == nil]).Add(1) + if !isErrorCapacity(errToReturn) || isErrorCapacity(err) { + // We want to return the last capacity error, + // if any; otherwise the last non-capacity + // error. + errToReturn = err + } + if isErrorSubnetSpecific(err) && + tryOffset < len(subnets)-1 { + instanceSet.logger.WithError(err).WithField("SubnetID", subnets[tryIndex]). + Warn("RunInstances failed, trying next subnet") + continue + } + // Succeeded, or exhausted all subnets, or got a + // non-subnet-related error. + // + // We intentionally update currentSubnetIDIndex even + // in the non-retryable-failure case here to avoid a + // situation where successive calls to Create() keep + // returning errors for the same subnet (perhaps + // "subnet full") and never reveal the errors for the + // other configured subnets (perhaps "subnet ID + // invalid"). + atomic.StoreInt32(&instanceSet.currentSubnetIDIndex, int32(tryIndex)) + break + } + if rsv == nil || len(rsv.Instances) == 0 { + return nil, wrapError(errToReturn, &instanceSet.throttleDelayCreate) + } return &ec2Instance{ provider: instanceSet, instance: rsv.Instances[0], }, nil } +func (instanceSet *ec2InstanceSet) getKeyName(publicKey ssh.PublicKey) (string, error) { + instanceSet.keysMtx.Lock() + defer instanceSet.keysMtx.Unlock() + md5keyFingerprint, sha1keyFingerprint, err := awsKeyFingerprint(publicKey) + if err != nil { + return "", fmt.Errorf("Could not make key fingerprint: %v", err) + } + if keyname, ok := instanceSet.keys[md5keyFingerprint]; ok { + return keyname, nil + } + keyout, err := instanceSet.client.DescribeKeyPairs(&ec2.DescribeKeyPairsInput{ + Filters: []*ec2.Filter{{ + Name: aws.String("fingerprint"), + Values: []*string{&md5keyFingerprint, &sha1keyFingerprint}, + }}, + }) + if err != nil { + return "", fmt.Errorf("Could not search for keypair: %v", err) + } + if len(keyout.KeyPairs) > 0 { + return *(keyout.KeyPairs[0].KeyName), nil + } + keyname := "arvados-dispatch-keypair-" + md5keyFingerprint + _, err = instanceSet.client.ImportKeyPair(&ec2.ImportKeyPairInput{ + KeyName: &keyname, + PublicKeyMaterial: ssh.MarshalAuthorizedKey(publicKey), + }) + if err != nil { + return "", fmt.Errorf("Could not import keypair: %v", err) + } + instanceSet.keys[md5keyFingerprint] = keyname + return keyname, nil +} + func (instanceSet *ec2InstanceSet) Instances(tags cloud.InstanceTags) (instances []cloud.Instance, err error) { var filters []*ec2.Filter for k, v := range tags { @@ -249,6 +385,7 @@ func (instanceSet *ec2InstanceSet) Instances(tags cloud.InstanceTags) (instances Values: []*string{aws.String(v)}, }) } + needAZs := false dii := &ec2.DescribeInstancesInput{Filters: filters} for { dio, err := instanceSet.client.DescribeInstances(dii) @@ -260,23 +397,168 @@ func (instanceSet *ec2InstanceSet) Instances(tags cloud.InstanceTags) (instances for _, rsv := range dio.Reservations { for _, inst := range rsv.Instances { if *inst.State.Name != "shutting-down" && *inst.State.Name != "terminated" { - instances = append(instances, &ec2Instance{instanceSet, inst}) + instances = append(instances, &ec2Instance{ + provider: instanceSet, + instance: inst, + }) + if aws.StringValue(inst.InstanceLifecycle) == "spot" { + needAZs = true + } } } } if dio.NextToken == nil { - return instances, err + break } dii.NextToken = dio.NextToken } + if needAZs && instanceSet.ec2config.SpotPriceUpdateInterval > 0 { + az := map[string]string{} + err := instanceSet.client.DescribeInstanceStatusPages(&ec2.DescribeInstanceStatusInput{ + IncludeAllInstances: aws.Bool(true), + }, func(page *ec2.DescribeInstanceStatusOutput, lastPage bool) bool { + for _, ent := range page.InstanceStatuses { + az[*ent.InstanceId] = *ent.AvailabilityZone + } + return true + }) + if err != nil { + instanceSet.logger.Warnf("error getting instance statuses: %s", err) + } + for _, inst := range instances { + inst := inst.(*ec2Instance) + inst.availabilityZone = az[*inst.instance.InstanceId] + } + instanceSet.updateSpotPrices(instances) + } + + // Count instances in each subnet, and report in metrics. + subnetInstances := map[string]int{"": 0} + for _, subnet := range instanceSet.ec2config.SubnetID { + subnetInstances[subnet] = 0 + } + for _, inst := range instances { + subnet := inst.(*ec2Instance).instance.SubnetId + if subnet != nil { + subnetInstances[*subnet]++ + } else { + subnetInstances[""]++ + } + } + for subnet, count := range subnetInstances { + instanceSet.mInstances.WithLabelValues(subnet).Set(float64(count)) + } + + return instances, err +} + +type priceKey struct { + instanceType string + spot bool + availabilityZone string +} + +// Refresh recent spot instance pricing data for the given instances, +// unless we already have recent pricing data for all relevant types. +func (instanceSet *ec2InstanceSet) updateSpotPrices(instances []cloud.Instance) { + if len(instances) == 0 { + return + } + + instanceSet.pricesLock.Lock() + defer instanceSet.pricesLock.Unlock() + if instanceSet.prices == nil { + instanceSet.prices = map[priceKey][]cloud.InstancePrice{} + instanceSet.pricesUpdated = map[priceKey]time.Time{} + } + + updateTime := time.Now() + staleTime := updateTime.Add(-instanceSet.ec2config.SpotPriceUpdateInterval.Duration()) + needUpdate := false + allTypes := map[string]bool{} + + for _, inst := range instances { + ec2inst := inst.(*ec2Instance).instance + if aws.StringValue(ec2inst.InstanceLifecycle) == "spot" { + pk := priceKey{ + instanceType: *ec2inst.InstanceType, + spot: true, + availabilityZone: inst.(*ec2Instance).availabilityZone, + } + if instanceSet.pricesUpdated[pk].Before(staleTime) { + needUpdate = true + } + allTypes[*ec2inst.InstanceType] = true + } + } + if !needUpdate { + return + } + var typeFilterValues []*string + for instanceType := range allTypes { + typeFilterValues = append(typeFilterValues, aws.String(instanceType)) + } + // Get 3x update interval worth of pricing data. (Ideally the + // AWS API would tell us "we have shown you all of the price + // changes up to time T", but it doesn't, so we'll just ask + // for 3 intervals worth of data on each update, de-duplicate + // the data points, and not worry too much about occasionally + // missing some data points when our lookups fail twice in a + // row. + dsphi := &ec2.DescribeSpotPriceHistoryInput{ + StartTime: aws.Time(updateTime.Add(-3 * instanceSet.ec2config.SpotPriceUpdateInterval.Duration())), + Filters: []*ec2.Filter{ + &ec2.Filter{Name: aws.String("instance-type"), Values: typeFilterValues}, + &ec2.Filter{Name: aws.String("product-description"), Values: []*string{aws.String("Linux/UNIX")}}, + }, + } + err := instanceSet.client.DescribeSpotPriceHistoryPages(dsphi, func(page *ec2.DescribeSpotPriceHistoryOutput, lastPage bool) bool { + for _, ent := range page.SpotPriceHistory { + if ent.InstanceType == nil || ent.SpotPrice == nil || ent.Timestamp == nil { + // bogus record? + continue + } + price, err := strconv.ParseFloat(*ent.SpotPrice, 64) + if err != nil { + // bogus record? + continue + } + pk := priceKey{ + instanceType: *ent.InstanceType, + spot: true, + availabilityZone: *ent.AvailabilityZone, + } + instanceSet.prices[pk] = append(instanceSet.prices[pk], cloud.InstancePrice{ + StartTime: *ent.Timestamp, + Price: price, + }) + instanceSet.pricesUpdated[pk] = updateTime + } + return true + }) + if err != nil { + instanceSet.logger.Warnf("error retrieving spot instance prices: %s", err) + } + + expiredTime := updateTime.Add(-64 * instanceSet.ec2config.SpotPriceUpdateInterval.Duration()) + for pk, last := range instanceSet.pricesUpdated { + if last.Before(expiredTime) { + delete(instanceSet.pricesUpdated, pk) + delete(instanceSet.prices, pk) + } + } + for pk, prices := range instanceSet.prices { + instanceSet.prices[pk] = cloud.NormalizePriceHistory(prices) + } } func (instanceSet *ec2InstanceSet) Stop() { } type ec2Instance struct { - provider *ec2InstanceSet - instance *ec2.Instance + provider *ec2InstanceSet + instance *ec2.Instance + availabilityZone string // sometimes available for spot instances } func (inst *ec2Instance) ID() cloud.InstanceID { @@ -340,6 +622,53 @@ func (inst *ec2Instance) VerifyHostKey(ssh.PublicKey, *ssh.Client) error { return cloud.ErrNotImplemented } +// PriceHistory returns the price history for this specific instance. +// +// AWS documentation is elusive about whether the hourly cost of a +// given spot instance changes as the current spot price changes for +// the corresponding instance type and availability zone. Our +// implementation assumes the answer is yes, based on the following +// hints. +// +// https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/spot-requests.html +// says: "After your Spot Instance is running, if the Spot price rises +// above your maximum price, Amazon EC2 interrupts your Spot +// Instance." (This doesn't address what happens when the spot price +// rises *without* exceeding your maximum price.) +// +// https://docs.aws.amazon.com/whitepapers/latest/cost-optimization-leveraging-ec2-spot-instances/how-spot-instances-work.html +// says: "You pay the Spot price that's in effect, billed to the +// nearest second." (But it's not explicitly stated whether "the price +// in effect" changes over time for a given instance.) +// +// The same page also says, in a discussion about the effect of +// specifying a maximum price: "Note that you never pay more than the +// Spot price that is in effect when your Spot Instance is running." +// (The use of the phrase "is running", as opposed to "was launched", +// hints that pricing is dynamic.) +func (inst *ec2Instance) PriceHistory(instType arvados.InstanceType) []cloud.InstancePrice { + inst.provider.pricesLock.Lock() + defer inst.provider.pricesLock.Unlock() + // Note updateSpotPrices currently populates + // inst.provider.prices only for spot instances, so if + // spot==false here, we will return no data. + pk := priceKey{ + instanceType: *inst.instance.InstanceType, + spot: aws.StringValue(inst.instance.InstanceLifecycle) == "spot", + availabilityZone: inst.availabilityZone, + } + var prices []cloud.InstancePrice + for _, price := range inst.provider.prices[pk] { + // ceil(added scratch space in GiB) + gib := (instType.AddedScratch + 1<<30 - 1) >> 30 + monthly := inst.provider.ec2config.EBSPrice * float64(gib) + hourly := monthly / 30 / 24 + price.Price += hourly + prices = append(prices, price) + } + return prices +} + type rateLimitError struct { error earliestRetry time.Time @@ -349,6 +678,85 @@ func (err rateLimitError) EarliestRetry() time.Time { return err.earliestRetry } +type capacityError struct { + error + isInstanceTypeSpecific bool +} + +func (er *capacityError) IsCapacityError() bool { + return true +} + +func (er *capacityError) IsInstanceTypeSpecific() bool { + return er.isInstanceTypeSpecific +} + +var isCodeQuota = map[string]bool{ + "InstanceLimitExceeded": true, + "InsufficientAddressCapacity": true, + "InsufficientFreeAddressesInSubnet": true, + "InsufficientVolumeCapacity": true, + "MaxSpotInstanceCountExceeded": true, + "VcpuLimitExceeded": true, +} + +// isErrorQuota returns whether the error indicates we have reached +// some usage quota/limit -- i.e., immediately retrying with an equal +// or larger instance type will probably not work. +// +// Returns false if error is nil. +func isErrorQuota(err error) bool { + if aerr, ok := err.(awserr.Error); ok && aerr != nil { + if _, ok := isCodeQuota[aerr.Code()]; ok { + return true + } + } + return false +} + +var reSubnetSpecificInvalidParameterMessage = regexp.MustCompile(`(?ms).*( subnet |sufficient free [Ii]pv[46] addresses).*`) + +// isErrorSubnetSpecific returns true if the problem encountered by +// RunInstances might be avoided by trying a different subnet. +func isErrorSubnetSpecific(err error) bool { + aerr, ok := err.(awserr.Error) + if !ok { + return false + } + code := aerr.Code() + return strings.Contains(code, "Subnet") || + code == "InsufficientInstanceCapacity" || + code == "InsufficientVolumeCapacity" || + code == "Unsupported" || + // See TestIsErrorSubnetSpecific for examples of why + // we look for substrings in code/message instead of + // only using specific codes here. + (strings.Contains(code, "InvalidParameter") && + reSubnetSpecificInvalidParameterMessage.MatchString(aerr.Message())) +} + +// isErrorCapacity returns true if the error indicates lack of +// capacity (either temporary or permanent) to run a specific instance +// type -- i.e., retrying with a different instance type might +// succeed. +func isErrorCapacity(err error) bool { + aerr, ok := err.(awserr.Error) + if !ok { + return false + } + code := aerr.Code() + return code == "InsufficientInstanceCapacity" || + (code == "Unsupported" && strings.Contains(aerr.Message(), "requested instance type")) +} + +type ec2QuotaError struct { + error +} + +func (er *ec2QuotaError) IsQuotaError() bool { + return true +} + func wrapError(err error, throttleValue *atomic.Value) error { if request.IsErrorThrottle(err) { // Back off exponentially until an upstream call @@ -362,6 +770,10 @@ func wrapError(err error, throttleValue *atomic.Value) error { } throttleValue.Store(d) return rateLimitError{error: err, earliestRetry: time.Now().Add(d)} + } else if isErrorQuota(err) { + return &ec2QuotaError{err} + } else if isErrorCapacity(err) { + return &capacityError{err, true} } else if err != nil { throttleValue.Store(time.Duration(0)) return err @@ -369,3 +781,5 @@ func wrapError(err error, throttleValue *atomic.Value) error { throttleValue.Store(time.Duration(0)) return nil } + +var boolLabelValue = map[bool]string{false: "0", true: "1"}