Dataset Viewer
Auto-converted to Parquet Duplicate
hash
stringlengths
40
40
date
stringdate
2018-12-11 14:31:19
2025-03-22 02:45:31
author
stringclasses
280 values
commit_message
stringlengths
14
176
is_merge
bool
1 class
git_diff
stringlengths
198
25.8M
type
stringclasses
83 values
masked_commit_message
stringlengths
8
170
bba6c14459bea3e5ef6907c30e9f4341f6365d7d
2022-08-31 02:33:19
Dylan Guedes
loki: Modify ingesters to return rate-limited streams to distributors (#6977)
false
diff --git a/go.mod b/go.mod index cd0948025d588..63da30a9cc8c9 100644 --- a/go.mod +++ b/go.mod @@ -113,6 +113,7 @@ require ( ) require ( + github.com/gogo/googleapis v1.4.0 github.com/grafana/groupcache_exporter v0.0.0-20220629095919-59a8c6428a43 github.com/heroku/x v0.0.50 github.com/mailgun/groupcache/v2 v2.3.2 @@ -179,7 +180,6 @@ require ( github.com/go-stack/stack v1.8.1 // indirect github.com/go-zookeeper/zk v1.0.2 // indirect github.com/gofrs/flock v0.7.1 // indirect - github.com/gogo/googleapis v1.4.0 // indirect github.com/golang-jwt/jwt/v4 v4.2.0 // indirect github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da // indirect github.com/google/btree v1.0.1 // indirect diff --git a/pkg/ingester/ingester_test.go b/pkg/ingester/ingester_test.go index 6549891426ed8..92a16f35da661 100644 --- a/pkg/ingester/ingester_test.go +++ b/pkg/ingester/ingester_test.go @@ -10,6 +10,8 @@ import ( "testing" "time" + "github.com/gogo/protobuf/types" + "github.com/gogo/status" "github.com/grafana/dskit/flagext" "github.com/grafana/dskit/services" "github.com/prometheus/common/model" @@ -19,6 +21,7 @@ import ( "github.com/weaveworks/common/middleware" "github.com/weaveworks/common/user" "golang.org/x/net/context" + "golang.org/x/time/rate" "google.golang.org/grpc" "google.golang.org/grpc/credentials/insecure" "google.golang.org/grpc/metadata" @@ -40,6 +43,68 @@ import ( "github.com/grafana/loki/pkg/validation" ) +func TestRateLimitedStreamsReturn(t *testing.T) { + // setup. + ingesterConfig := defaultIngesterTestConfig(t) + limits := defaultLimitsTestConfig() + limits.PerStreamRateLimit.Set("1") //nolint:errcheck + limits.IngestionBurstSizeMB = 1 + limits.PerStreamRateLimitBurst.Set("1") //nolint:errcheck + overrides, err := validation.NewOverrides(limits, nil) + limit := overrides.PerStreamRateLimit("test") + require.Equal(t, limit.Limit, rate.Limit(1)) + require.NoError(t, err) + + store := &mockStore{ + chunks: map[string][]chunk.Chunk{}, + } + + i, err := New(ingesterConfig, client.Config{}, store, overrides, runtime.DefaultTenantConfigs(), nil) + require.NoError(t, err) + defer services.StopAndAwaitTerminated(context.Background(), i) //nolint:errcheck + + req := logproto.PushRequest{ + Streams: []logproto.Stream{ + { + Labels: `{bar="baz1", foo="bar"}`, + }, + { + Labels: `{bar="baz2", foo="bar"}`, + }, + }, + } + for i := 0; i < 10; i++ { + req.Streams[0].Entries = append(req.Streams[0].Entries, logproto.Entry{ + Timestamp: time.Unix(0, 0), + Line: fmt.Sprintf("line %d", i), + }) + req.Streams[1].Entries = append(req.Streams[1].Entries, logproto.Entry{ + Timestamp: time.Unix(0, 0), + Line: fmt.Sprintf("line %d", i), + }) + } + + ctx := user.InjectOrgID(context.Background(), "test") + _, err = i.Push(ctx, &req) + require.Error(t, err) + + s, ok := status.FromError(err) + require.True(t, ok) + details := s.Proto().Details + + var rateLimitedLabels []string + for _, detail := range details { + rls := &logproto.RateLimitedStream{} + err := types.UnmarshalAny(detail, rls) + require.NoError(t, err) + rateLimitedLabels = append(rateLimitedLabels, rls.Labels) + } + + // note that streams[0], although rate-limited, isn't present in the details. + // that's because push as of now is only returning the last errored stream. + require.EqualValues(t, []string{req.Streams[1].Labels}, rateLimitedLabels) +} + func TestIngester(t *testing.T) { ingesterConfig := defaultIngesterTestConfig(t) limits, err := validation.NewOverrides(defaultLimitsTestConfig(), nil) diff --git a/pkg/ingester/instance.go b/pkg/ingester/instance.go index 1c8e9d8d0ae68..05bd3d6ec3911 100644 --- a/pkg/ingester/instance.go +++ b/pkg/ingester/instance.go @@ -1,7 +1,9 @@ package ingester import ( + "bytes" "context" + fmt "fmt" "net/http" "os" "sync" @@ -9,6 +11,9 @@ import ( "time" "github.com/go-kit/log/level" + spb "github.com/gogo/googleapis/google/rpc" + "github.com/gogo/protobuf/types" + "github.com/gogo/status" "github.com/pkg/errors" "github.com/prometheus/client_golang/prometheus" "github.com/prometheus/client_golang/prometheus/promauto" @@ -19,6 +24,7 @@ import ( "github.com/weaveworks/common/httpgrpc" "go.uber.org/atomic" + "github.com/grafana/loki/pkg/chunkenc" "github.com/grafana/loki/pkg/ingester/index" "github.com/grafana/loki/pkg/iter" "github.com/grafana/loki/pkg/logproto" @@ -158,6 +164,11 @@ func (i *instance) consumeChunk(ctx context.Context, ls labels.Labels, chunk *lo return err } +// Push will iterate over the given streams present in the PushRequest and attempt to store them. +// +// Although multiple streams are part of the PushRequest, the returned error only reflects what +// happened to *the last stream in the request*. Ex: if three streams are part of the PushRequest +// and all three failed, the returned error only describes what happened to the last processed stream. func (i *instance) Push(ctx context.Context, req *logproto.PushRequest) error { record := recordPool.GetRecord() record.UserID = i.instanceID @@ -185,10 +196,9 @@ func (i *instance) Push(ctx context.Context, req *logproto.PushRequest) error { continue } - _, err = s.Push(ctx, reqStream.Entries, record, 0, false) - if err != nil { - appendErr = err - } + _, failedEntriesWithError := s.Push(ctx, reqStream.Entries, record, 0, false) + appendErr = errorForFailedEntries(s, failedEntriesWithError, len(reqStream.Entries)) + s.chunkMtx.Unlock() } @@ -211,6 +221,73 @@ func (i *instance) Push(ctx context.Context, req *logproto.PushRequest) error { return appendErr } +// errorForFailedEntries mounts an error to be returned in a GRPC call based on entries that couldn't be ingested. +// +// The returned error is enriched with the gRPC error status and with a list of details. +// As of now, the list of details is a list of all streams that were rate-limited. This list can be used +// by the distributor to fine tune streams that were limited. +func errorForFailedEntries(s *stream, failedEntriesWithError []entryWithError, totalEntries int) error { + if len(failedEntriesWithError) == 0 { + return nil + } + + lastEntryWithErr := failedEntriesWithError[len(failedEntriesWithError)-1] + _, ok := lastEntryWithErr.e.(*validation.ErrStreamRateLimit) + outOfOrder := chunkenc.IsOutOfOrderErr(lastEntryWithErr.e) + if !outOfOrder && !ok { + return lastEntryWithErr.e + } + + var statusCode int + if outOfOrder { + statusCode = http.StatusBadRequest + } + if ok { + // per-stream or ingestion limited. + statusCode = http.StatusTooManyRequests + } + + // Return a http status 4xx request response with all failed entries. + buf := bytes.Buffer{} + streamName := s.labelsString + + limitedFailedEntries := failedEntriesWithError + if maxIgnore := s.cfg.MaxReturnedErrors; maxIgnore > 0 && len(limitedFailedEntries) > maxIgnore { + limitedFailedEntries = limitedFailedEntries[:maxIgnore] + } + + for _, entryWithError := range limitedFailedEntries { + fmt.Fprintf(&buf, + "entry with timestamp %s ignored, reason: '%s' for stream: %s,\n", + entryWithError.entry.Timestamp.String(), entryWithError.e.Error(), streamName) + } + + fmt.Fprintf(&buf, "total ignored: %d out of %d", len(failedEntriesWithError), totalEntries) + + var details []*types.Any + + if statusCode == http.StatusTooManyRequests { + details = append(details, mountPerStreamDetails(streamName)...) + } + + return status.ErrorProto(&spb.Status{ + Code: int32(statusCode), + Message: buf.String(), + Details: details, + }) +} + +func mountPerStreamDetails(streamLabels string) []*types.Any { + rls := logproto.RateLimitedStream{Labels: streamLabels} + marshalledStream, err := types.MarshalAny(&rls) + if err == nil { + return []*types.Any{marshalledStream} + } + + level.Error(util_log.Logger).Log("msg", "error marshalling rate-limited stream", "err", err, "labels", streamLabels) + return []*types.Any{} +} + func (i *instance) createStream(pushReqStream logproto.Stream, record *WALRecord) (*stream, error) { // record is only nil when replaying WAL. We don't want to drop data when replaying a WAL after // reducing the stream limits, for instance. diff --git a/pkg/ingester/recovery.go b/pkg/ingester/recovery.go index c3bbc9a8af442..1a0c18786c8e4 100644 --- a/pkg/ingester/recovery.go +++ b/pkg/ingester/recovery.go @@ -165,10 +165,13 @@ func (r *ingesterRecoverer) Push(userID string, entries RefEntries) error { } // ignore out of order errors here (it's possible for a checkpoint to already have data from the wal segments) - bytesAdded, err := s.(*stream).Push(context.Background(), entries.Entries, nil, entries.Counter, true) + bytesAdded, entriesWithErrors := s.(*stream).Push(context.Background(), entries.Entries, nil, entries.Counter, true) r.ing.replayController.Add(int64(bytesAdded)) - if err != nil && err == ErrEntriesExist { - r.ing.metrics.duplicateEntriesTotal.Add(float64(len(entries.Entries))) + if len(entriesWithErrors) > 0 { + lastEntryWithError := entriesWithErrors[len(entriesWithErrors)-1] + if errors.Is(lastEntryWithError.e, ErrEntriesExist) { + r.ing.metrics.duplicateEntriesTotal.Add(float64(len(entries.Entries))) + } } return nil }) diff --git a/pkg/ingester/stream.go b/pkg/ingester/stream.go index 23a9d3d674573..3d824bd592529 100644 --- a/pkg/ingester/stream.go +++ b/pkg/ingester/stream.go @@ -1,18 +1,15 @@ package ingester import ( - "bytes" "context" - "fmt" - "net/http" "sync" "time" "github.com/go-kit/log/level" + "github.com/pkg/errors" "github.com/prometheus/common/model" "github.com/prometheus/prometheus/model/labels" - "github.com/weaveworks/common/httpgrpc" "github.com/grafana/loki/pkg/chunkenc" "github.com/grafana/loki/pkg/iter" @@ -148,7 +145,7 @@ func (s *stream) Push( // Lock chunkMtx while pushing. // If this is false, chunkMtx must be held outside Push. lockChunk bool, -) (int, error) { +) (int, []entryWithError) { if lockChunk { s.chunkMtx.Lock() defer s.chunkMtx.Unlock() @@ -163,7 +160,7 @@ func (s *stream) Push( s.metrics.walReplaySamplesDropped.WithLabelValues(duplicateReason).Add(float64(len(entries))) s.metrics.walReplayBytesDropped.WithLabelValues(duplicateReason).Add(float64(byteCt)) - return 0, ErrEntriesExist + return 0, []entryWithError{{entry: &logproto.Entry{}, e: ErrEntriesExist}} } var bytesAdded int @@ -300,41 +297,8 @@ func (s *stream) Push( s.metrics.memoryChunks.Add(float64(len(s.chunks) - prevNumChunks)) } - if len(failedEntriesWithError) > 0 { - lastEntryWithErr := failedEntriesWithError[len(failedEntriesWithError)-1] - _, ok := lastEntryWithErr.e.(*validation.ErrStreamRateLimit) - outOfOrder := chunkenc.IsOutOfOrderErr(lastEntryWithErr.e) - if !outOfOrder && !ok { - return bytesAdded, lastEntryWithErr.e - } - var statusCode int - if outOfOrder { - statusCode = http.StatusBadRequest - } - if ok { - statusCode = http.StatusTooManyRequests - } - // Return a http status 4xx request response with all failed entries. - buf := bytes.Buffer{} - streamName := s.labelsString - - limitedFailedEntries := failedEntriesWithError - if maxIgnore := s.cfg.MaxReturnedErrors; maxIgnore > 0 && len(limitedFailedEntries) > maxIgnore { - limitedFailedEntries = limitedFailedEntries[:maxIgnore] - } - - for _, entryWithError := range limitedFailedEntries { - fmt.Fprintf(&buf, - "entry with timestamp %s ignored, reason: '%s' for stream: %s,\n", - entryWithError.entry.Timestamp.String(), entryWithError.e.Error(), streamName) - } - - fmt.Fprintf(&buf, "total ignored: %d out of %d", len(failedEntriesWithError), len(entries)) - - return bytesAdded, httpgrpc.Errorf(statusCode, buf.String()) - } + return bytesAdded, failedEntriesWithError - return bytesAdded, nil } func (s *stream) cutChunk(ctx context.Context) *chunkDesc { diff --git a/pkg/ingester/stream_test.go b/pkg/ingester/stream_test.go index 2b5e3b8711efc..780317a98ed42 100644 --- a/pkg/ingester/stream_test.go +++ b/pkg/ingester/stream_test.go @@ -64,10 +64,10 @@ func TestMaxReturnedStreamsErrors(t *testing.T) { NilMetrics, ) - _, err := s.Push(context.Background(), []logproto.Entry{ + _, entriesWithErrors := s.Push(context.Background(), []logproto.Entry{ {Timestamp: time.Unix(int64(numLogs), 0), Line: "log"}, }, recordPool.GetRecord(), 0, true) - require.NoError(t, err) + require.Empty(t, entriesWithErrors) newLines := make([]logproto.Entry, numLogs) for i := 0; i < numLogs; i++ { @@ -86,9 +86,9 @@ func TestMaxReturnedStreamsErrors(t *testing.T) { fmt.Fprintf(&expected, "total ignored: %d out of %d", numLogs, numLogs) expectErr := httpgrpc.Errorf(http.StatusBadRequest, expected.String()) - _, err = s.Push(context.Background(), newLines, recordPool.GetRecord(), 0, true) - require.Error(t, err) - require.Equal(t, expectErr.Error(), err.Error()) + _, entriesWithErrors = s.Push(context.Background(), newLines, recordPool.GetRecord(), 0, true) + finalErr := errorForFailedEntries(s, entriesWithErrors, len(newLines)) + require.Equal(t, expectErr.Error(), finalErr.Error()) }) } } @@ -110,12 +110,12 @@ func TestPushDeduplication(t *testing.T) { NilMetrics, ) - written, err := s.Push(context.Background(), []logproto.Entry{ + written, entriesWithErrors := s.Push(context.Background(), []logproto.Entry{ {Timestamp: time.Unix(1, 0), Line: "test"}, {Timestamp: time.Unix(1, 0), Line: "test"}, {Timestamp: time.Unix(1, 0), Line: "newer, better test"}, }, recordPool.GetRecord(), 0, true) - require.NoError(t, err) + require.Empty(t, entriesWithErrors) require.Len(t, s.chunks, 1) require.Equal(t, s.chunks[0].chunk.Size(), 2, "expected exact duplicate to be dropped and newer content with same timestamp to be appended") @@ -140,27 +140,28 @@ func TestPushRejectOldCounter(t *testing.T) { ) // counter should be 2 now since the first line will be deduped - _, err = s.Push(context.Background(), []logproto.Entry{ + _, entriesWithErrors := s.Push(context.Background(), []logproto.Entry{ {Timestamp: time.Unix(1, 0), Line: "test"}, {Timestamp: time.Unix(1, 0), Line: "test"}, {Timestamp: time.Unix(1, 0), Line: "newer, better test"}, }, recordPool.GetRecord(), 0, true) - require.NoError(t, err) + require.Empty(t, entriesWithErrors) require.Len(t, s.chunks, 1) require.Equal(t, s.chunks[0].chunk.Size(), 2, "expected exact duplicate to be dropped and newer content with same timestamp to be appended") // fail to push with a counter <= the streams internal counter - _, err = s.Push(context.Background(), []logproto.Entry{ + _, entriesWithErrors = s.Push(context.Background(), []logproto.Entry{ {Timestamp: time.Unix(1, 0), Line: "test"}, }, recordPool.GetRecord(), 2, true) - require.Equal(t, ErrEntriesExist, err) + require.Len(t, entriesWithErrors, 1) + require.Equal(t, entriesWithErrors[0].e, ErrEntriesExist) // succeed with a greater counter - _, err = s.Push(context.Background(), []logproto.Entry{ + _, entriesWithErrors = s.Push(context.Background(), []logproto.Entry{ {Timestamp: time.Unix(1, 0), Line: "test"}, }, recordPool.GetRecord(), 3, true) - require.Nil(t, err) + require.Empty(t, entriesWithErrors) } @@ -273,11 +274,11 @@ func TestUnorderedPush(t *testing.T) { if x.cutBefore { _ = s.cutChunk(context.Background()) } - written, err := s.Push(context.Background(), x.entries, recordPool.GetRecord(), 0, true) + written, entriesWithErrors := s.Push(context.Background(), x.entries, recordPool.GetRecord(), 0, true) if x.err { - require.NotNil(t, err) + require.NotEmpty(t, entriesWithErrors) } else { - require.Nil(t, err) + require.Empty(t, entriesWithErrors) } require.Equal(t, x.written, written) } @@ -334,8 +335,9 @@ func TestPushRateLimit(t *testing.T) { {Timestamp: time.Unix(1, 0), Line: "aaaaaaaaab"}, } // Counter should be 2 now since the first line will be deduped. - _, err = s.Push(context.Background(), entries, recordPool.GetRecord(), 0, true) - require.Contains(t, err.Error(), (&validation.ErrStreamRateLimit{RateLimit: l.PerStreamRateLimit, Labels: s.labelsString, Bytes: flagext.ByteSize(len(entries[1].Line))}).Error()) + _, entriesWithErrors := s.Push(context.Background(), entries, recordPool.GetRecord(), 0, true) + require.Len(t, entriesWithErrors, 1) + require.Contains(t, entriesWithErrors[0].e.Error(), (&validation.ErrStreamRateLimit{RateLimit: l.PerStreamRateLimit, Labels: s.labelsString, Bytes: flagext.ByteSize(len(entries[1].Line))}).Error()) } func TestReplayAppendIgnoresValidityWindow(t *testing.T) { @@ -365,8 +367,8 @@ func TestReplayAppendIgnoresValidityWindow(t *testing.T) { } // Push a first entry (it doesn't matter if we look like we're replaying or not) - _, err = s.Push(context.Background(), entries, nil, 1, true) - require.Nil(t, err) + _, entriesWithErrors := s.Push(context.Background(), entries, nil, 1, true) + require.Empty(t, entriesWithErrors) // Create a sample outside the validity window entries = []logproto.Entry{ @@ -374,12 +376,12 @@ func TestReplayAppendIgnoresValidityWindow(t *testing.T) { } // Pretend it's not a replay, ensure we error - _, err = s.Push(context.Background(), entries, recordPool.GetRecord(), 0, true) - require.NotNil(t, err) + _, entriesWithErrors = s.Push(context.Background(), entries, recordPool.GetRecord(), 0, true) + require.NotEmpty(t, entriesWithErrors) // Now pretend it's a replay. The same write should succeed. - _, err = s.Push(context.Background(), entries, nil, 2, true) - require.Nil(t, err) + _, entriesWithErrors = s.Push(context.Background(), entries, nil, 2, true) + require.Empty(t, entriesWithErrors) } @@ -420,8 +422,8 @@ func Benchmark_PushStream(b *testing.B) { for n := 0; n < b.N; n++ { rec := recordPool.GetRecord() - _, err := s.Push(ctx, e, rec, 0, true) - require.NoError(b, err) + _, entriesWithErrors := s.Push(ctx, e, rec, 0, true) + require.Empty(b, entriesWithErrors) recordPool.PutRecord(rec) } } diff --git a/pkg/logproto/logproto.pb.go b/pkg/logproto/logproto.pb.go index cb3e65e2f8ab2..ead1e41c3ccdd 100644 --- a/pkg/logproto/logproto.pb.go +++ b/pkg/logproto/logproto.pb.go @@ -511,6 +511,49 @@ func (m *LabelRequest) GetEnd() *time.Time { return nil } +type RateLimitedStream struct { + Labels string `protobuf:"bytes,1,opt,name=labels,proto3" json:"labels,omitempty"` +} + +func (m *RateLimitedStream) Reset() { *m = RateLimitedStream{} } +func (*RateLimitedStream) ProtoMessage() {} +func (*RateLimitedStream) Descriptor() ([]byte, []int) { + return fileDescriptor_c28a5f14f1f4c79a, []int{8} +} +func (m *RateLimitedStream) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *RateLimitedStream) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + if deterministic { + return xxx_messageInfo_RateLimitedStream.Marshal(b, m, deterministic) + } else { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil + } +} +func (m *RateLimitedStream) XXX_Merge(src proto.Message) { + xxx_messageInfo_RateLimitedStream.Merge(m, src) +} +func (m *RateLimitedStream) XXX_Size() int { + return m.Size() +} +func (m *RateLimitedStream) XXX_DiscardUnknown() { + xxx_messageInfo_RateLimitedStream.DiscardUnknown(m) +} + +var xxx_messageInfo_RateLimitedStream proto.InternalMessageInfo + +func (m *RateLimitedStream) GetLabels() string { + if m != nil { + return m.Labels + } + return "" +} + type LabelResponse struct { Values []string `protobuf:"bytes,1,rep,name=values,proto3" json:"values,omitempty"` } @@ -518,7 +561,7 @@ type LabelResponse struct { func (m *LabelResponse) Reset() { *m = LabelResponse{} } func (*LabelResponse) ProtoMessage() {} func (*LabelResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{8} + return fileDescriptor_c28a5f14f1f4c79a, []int{9} } func (m *LabelResponse) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -564,7 +607,7 @@ type StreamAdapter struct { func (m *StreamAdapter) Reset() { *m = StreamAdapter{} } func (*StreamAdapter) ProtoMessage() {} func (*StreamAdapter) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{9} + return fileDescriptor_c28a5f14f1f4c79a, []int{10} } func (m *StreamAdapter) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -622,7 +665,7 @@ type EntryAdapter struct { func (m *EntryAdapter) Reset() { *m = EntryAdapter{} } func (*EntryAdapter) ProtoMessage() {} func (*EntryAdapter) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{10} + return fileDescriptor_c28a5f14f1f4c79a, []int{11} } func (m *EntryAdapter) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -674,7 +717,7 @@ type Sample struct { func (m *Sample) Reset() { *m = Sample{} } func (*Sample) ProtoMessage() {} func (*Sample) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{11} + return fileDescriptor_c28a5f14f1f4c79a, []int{12} } func (m *Sample) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -733,7 +776,7 @@ type LegacySample struct { func (m *LegacySample) Reset() { *m = LegacySample{} } func (*LegacySample) ProtoMessage() {} func (*LegacySample) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{12} + return fileDescriptor_c28a5f14f1f4c79a, []int{13} } func (m *LegacySample) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -785,7 +828,7 @@ type Series struct { func (m *Series) Reset() { *m = Series{} } func (*Series) ProtoMessage() {} func (*Series) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{13} + return fileDescriptor_c28a5f14f1f4c79a, []int{14} } func (m *Series) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -845,7 +888,7 @@ type TailRequest struct { func (m *TailRequest) Reset() { *m = TailRequest{} } func (*TailRequest) ProtoMessage() {} func (*TailRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{14} + return fileDescriptor_c28a5f14f1f4c79a, []int{15} } func (m *TailRequest) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -910,7 +953,7 @@ type TailResponse struct { func (m *TailResponse) Reset() { *m = TailResponse{} } func (*TailResponse) ProtoMessage() {} func (*TailResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{15} + return fileDescriptor_c28a5f14f1f4c79a, []int{16} } func (m *TailResponse) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -956,7 +999,7 @@ type SeriesRequest struct { func (m *SeriesRequest) Reset() { *m = SeriesRequest{} } func (*SeriesRequest) ProtoMessage() {} func (*SeriesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{16} + return fileDescriptor_c28a5f14f1f4c79a, []int{17} } func (m *SeriesRequest) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1020,7 +1063,7 @@ type SeriesResponse struct { func (m *SeriesResponse) Reset() { *m = SeriesResponse{} } func (*SeriesResponse) ProtoMessage() {} func (*SeriesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{17} + return fileDescriptor_c28a5f14f1f4c79a, []int{18} } func (m *SeriesResponse) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1063,7 +1106,7 @@ type SeriesIdentifier struct { func (m *SeriesIdentifier) Reset() { *m = SeriesIdentifier{} } func (*SeriesIdentifier) ProtoMessage() {} func (*SeriesIdentifier) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{18} + return fileDescriptor_c28a5f14f1f4c79a, []int{19} } func (m *SeriesIdentifier) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1108,7 +1151,7 @@ type DroppedStream struct { func (m *DroppedStream) Reset() { *m = DroppedStream{} } func (*DroppedStream) ProtoMessage() {} func (*DroppedStream) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{19} + return fileDescriptor_c28a5f14f1f4c79a, []int{20} } func (m *DroppedStream) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1168,7 +1211,7 @@ type TimeSeriesChunk struct { func (m *TimeSeriesChunk) Reset() { *m = TimeSeriesChunk{} } func (*TimeSeriesChunk) ProtoMessage() {} func (*TimeSeriesChunk) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{20} + return fileDescriptor_c28a5f14f1f4c79a, []int{21} } func (m *TimeSeriesChunk) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1233,7 +1276,7 @@ type LabelPair struct { func (m *LabelPair) Reset() { *m = LabelPair{} } func (*LabelPair) ProtoMessage() {} func (*LabelPair) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{21} + return fileDescriptor_c28a5f14f1f4c79a, []int{22} } func (m *LabelPair) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1285,7 +1328,7 @@ type LegacyLabelPair struct { func (m *LegacyLabelPair) Reset() { *m = LegacyLabelPair{} } func (*LegacyLabelPair) ProtoMessage() {} func (*LegacyLabelPair) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{22} + return fileDescriptor_c28a5f14f1f4c79a, []int{23} } func (m *LegacyLabelPair) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1335,7 +1378,7 @@ type Chunk struct { func (m *Chunk) Reset() { *m = Chunk{} } func (*Chunk) ProtoMessage() {} func (*Chunk) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{23} + return fileDescriptor_c28a5f14f1f4c79a, []int{24} } func (m *Chunk) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1377,7 +1420,7 @@ type TransferChunksResponse struct { func (m *TransferChunksResponse) Reset() { *m = TransferChunksResponse{} } func (*TransferChunksResponse) ProtoMessage() {} func (*TransferChunksResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{24} + return fileDescriptor_c28a5f14f1f4c79a, []int{25} } func (m *TransferChunksResponse) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1412,7 +1455,7 @@ type TailersCountRequest struct { func (m *TailersCountRequest) Reset() { *m = TailersCountRequest{} } func (*TailersCountRequest) ProtoMessage() {} func (*TailersCountRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{25} + return fileDescriptor_c28a5f14f1f4c79a, []int{26} } func (m *TailersCountRequest) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1448,7 +1491,7 @@ type TailersCountResponse struct { func (m *TailersCountResponse) Reset() { *m = TailersCountResponse{} } func (*TailersCountResponse) ProtoMessage() {} func (*TailersCountResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{26} + return fileDescriptor_c28a5f14f1f4c79a, []int{27} } func (m *TailersCountResponse) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1493,7 +1536,7 @@ type GetChunkIDsRequest struct { func (m *GetChunkIDsRequest) Reset() { *m = GetChunkIDsRequest{} } func (*GetChunkIDsRequest) ProtoMessage() {} func (*GetChunkIDsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{27} + return fileDescriptor_c28a5f14f1f4c79a, []int{28} } func (m *GetChunkIDsRequest) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1550,7 +1593,7 @@ type GetChunkIDsResponse struct { func (m *GetChunkIDsResponse) Reset() { *m = GetChunkIDsResponse{} } func (*GetChunkIDsResponse) ProtoMessage() {} func (*GetChunkIDsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{28} + return fileDescriptor_c28a5f14f1f4c79a, []int{29} } func (m *GetChunkIDsResponse) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1603,7 +1646,7 @@ type ChunkRef struct { func (m *ChunkRef) Reset() { *m = ChunkRef{} } func (*ChunkRef) ProtoMessage() {} func (*ChunkRef) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{29} + return fileDescriptor_c28a5f14f1f4c79a, []int{30} } func (m *ChunkRef) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1664,7 +1707,7 @@ type LabelValuesForMetricNameRequest struct { func (m *LabelValuesForMetricNameRequest) Reset() { *m = LabelValuesForMetricNameRequest{} } func (*LabelValuesForMetricNameRequest) ProtoMessage() {} func (*LabelValuesForMetricNameRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{30} + return fileDescriptor_c28a5f14f1f4c79a, []int{31} } func (m *LabelValuesForMetricNameRequest) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1723,7 +1766,7 @@ type LabelNamesForMetricNameRequest struct { func (m *LabelNamesForMetricNameRequest) Reset() { *m = LabelNamesForMetricNameRequest{} } func (*LabelNamesForMetricNameRequest) ProtoMessage() {} func (*LabelNamesForMetricNameRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{31} + return fileDescriptor_c28a5f14f1f4c79a, []int{32} } func (m *LabelNamesForMetricNameRequest) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1768,7 +1811,7 @@ type GetChunkRefRequest struct { func (m *GetChunkRefRequest) Reset() { *m = GetChunkRefRequest{} } func (*GetChunkRefRequest) ProtoMessage() {} func (*GetChunkRefRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{32} + return fileDescriptor_c28a5f14f1f4c79a, []int{33} } func (m *GetChunkRefRequest) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1811,7 +1854,7 @@ type GetChunkRefResponse struct { func (m *GetChunkRefResponse) Reset() { *m = GetChunkRefResponse{} } func (*GetChunkRefResponse) ProtoMessage() {} func (*GetChunkRefResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{33} + return fileDescriptor_c28a5f14f1f4c79a, []int{34} } func (m *GetChunkRefResponse) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1856,7 +1899,7 @@ type GetSeriesRequest struct { func (m *GetSeriesRequest) Reset() { *m = GetSeriesRequest{} } func (*GetSeriesRequest) ProtoMessage() {} func (*GetSeriesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{34} + return fileDescriptor_c28a5f14f1f4c79a, []int{35} } func (m *GetSeriesRequest) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1899,7 +1942,7 @@ type GetSeriesResponse struct { func (m *GetSeriesResponse) Reset() { *m = GetSeriesResponse{} } func (*GetSeriesResponse) ProtoMessage() {} func (*GetSeriesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{35} + return fileDescriptor_c28a5f14f1f4c79a, []int{36} } func (m *GetSeriesResponse) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1943,7 +1986,7 @@ type IndexSeries struct { func (m *IndexSeries) Reset() { *m = IndexSeries{} } func (*IndexSeries) ProtoMessage() {} func (*IndexSeries) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{36} + return fileDescriptor_c28a5f14f1f4c79a, []int{37} } func (m *IndexSeries) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1980,7 +2023,7 @@ type QueryIndexResponse struct { func (m *QueryIndexResponse) Reset() { *m = QueryIndexResponse{} } func (*QueryIndexResponse) ProtoMessage() {} func (*QueryIndexResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{37} + return fileDescriptor_c28a5f14f1f4c79a, []int{38} } func (m *QueryIndexResponse) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2031,7 +2074,7 @@ type Row struct { func (m *Row) Reset() { *m = Row{} } func (*Row) ProtoMessage() {} func (*Row) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{38} + return fileDescriptor_c28a5f14f1f4c79a, []int{39} } func (m *Row) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2081,7 +2124,7 @@ type QueryIndexRequest struct { func (m *QueryIndexRequest) Reset() { *m = QueryIndexRequest{} } func (*QueryIndexRequest) ProtoMessage() {} func (*QueryIndexRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{39} + return fileDescriptor_c28a5f14f1f4c79a, []int{40} } func (m *QueryIndexRequest) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2128,7 +2171,7 @@ type IndexQuery struct { func (m *IndexQuery) Reset() { *m = IndexQuery{} } func (*IndexQuery) ProtoMessage() {} func (*IndexQuery) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{40} + return fileDescriptor_c28a5f14f1f4c79a, []int{41} } func (m *IndexQuery) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2201,7 +2244,7 @@ type IndexStatsRequest struct { func (m *IndexStatsRequest) Reset() { *m = IndexStatsRequest{} } func (*IndexStatsRequest) ProtoMessage() {} func (*IndexStatsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{41} + return fileDescriptor_c28a5f14f1f4c79a, []int{42} } func (m *IndexStatsRequest) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2247,7 +2290,7 @@ type IndexStatsResponse struct { func (m *IndexStatsResponse) Reset() { *m = IndexStatsResponse{} } func (*IndexStatsResponse) ProtoMessage() {} func (*IndexStatsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_c28a5f14f1f4c79a, []int{42} + return fileDescriptor_c28a5f14f1f4c79a, []int{43} } func (m *IndexStatsResponse) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2314,6 +2357,7 @@ func init() { proto.RegisterType((*QueryResponse)(nil), "logproto.QueryResponse") proto.RegisterType((*SampleQueryResponse)(nil), "logproto.SampleQueryResponse") proto.RegisterType((*LabelRequest)(nil), "logproto.LabelRequest") + proto.RegisterType((*RateLimitedStream)(nil), "logproto.RateLimitedStream") proto.RegisterType((*LabelResponse)(nil), "logproto.LabelResponse") proto.RegisterType((*StreamAdapter)(nil), "logproto.StreamAdapter") proto.RegisterType((*EntryAdapter)(nil), "logproto.EntryAdapter") @@ -2355,138 +2399,139 @@ func init() { func init() { proto.RegisterFile("pkg/logproto/logproto.proto", fileDescriptor_c28a5f14f1f4c79a) } var fileDescriptor_c28a5f14f1f4c79a = []byte{ - // 2088 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xd4, 0x59, 0x4f, 0x6f, 0x1b, 0xc7, - 0x15, 0xe7, 0x90, 0x4b, 0x8a, 0x7c, 0xa4, 0x24, 0x66, 0x2c, 0xcb, 0x0c, 0x6d, 0x73, 0xe5, 0x45, - 0x6a, 0x0b, 0x8e, 0x4d, 0xd6, 0x4a, 0xdb, 0x38, 0x76, 0xd3, 0x42, 0x94, 0x62, 0x5b, 0xb6, 0xe2, - 0x28, 0x2b, 0xd7, 0x01, 0x02, 0x14, 0xc6, 0x8a, 0x1c, 0x92, 0x0b, 0x71, 0xb9, 0xf4, 0xee, 0x30, - 0x8e, 0x80, 0x02, 0xed, 0x07, 0x68, 0x80, 0xf4, 0x54, 0xf4, 0x5e, 0xa0, 0x45, 0x0f, 0x3d, 0x14, - 0xe8, 0xb5, 0xed, 0xad, 0xee, 0xcd, 0xbd, 0x05, 0x39, 0xb0, 0xb5, 0x7c, 0x29, 0x74, 0xca, 0x27, - 0x28, 0x8a, 0xf9, 0xb7, 0x3b, 0x5c, 0x53, 0x8d, 0xe9, 0x1a, 0x28, 0x72, 0x11, 0x67, 0xde, 0xcc, - 0xbc, 0x37, 0xef, 0x37, 0xef, 0xef, 0x0a, 0x4e, 0x0f, 0xf7, 0xbb, 0x8d, 0xbe, 0xdf, 0x1d, 0x06, - 0x3e, 0xf5, 0xa3, 0x41, 0x9d, 0xff, 0xc5, 0x79, 0x35, 0xaf, 0x5e, 0xee, 0xba, 0xb4, 0x37, 0xda, - 0xab, 0xb7, 0x7c, 0xaf, 0xd1, 0xf5, 0xbb, 0x7e, 0x83, 0x93, 0xf7, 0x46, 0x1d, 0x3e, 0x13, 0x87, - 0xd9, 0x48, 0x1c, 0xac, 0x9a, 0x5d, 0xdf, 0xef, 0xf6, 0x49, 0xbc, 0x8b, 0xba, 0x1e, 0x09, 0xa9, - 0xe3, 0x0d, 0xe5, 0x86, 0x15, 0x29, 0xf6, 0x61, 0xdf, 0xf3, 0xdb, 0xa4, 0xdf, 0x08, 0xa9, 0x43, - 0x43, 0xf1, 0x57, 0xec, 0xb0, 0x3e, 0x82, 0xe2, 0xce, 0x28, 0xec, 0xd9, 0xe4, 0xe1, 0x88, 0x84, - 0x14, 0xdf, 0x82, 0xb9, 0x90, 0x06, 0xc4, 0xf1, 0xc2, 0x0a, 0x5a, 0xc9, 0xac, 0x16, 0xd7, 0x4e, - 0xd5, 0xa3, 0xcb, 0xee, 0xf2, 0x85, 0xf5, 0xb6, 0x33, 0xa4, 0x24, 0x68, 0x9e, 0xfc, 0x72, 0x6c, - 0xe6, 0x04, 0xe9, 0x68, 0x6c, 0xaa, 0x53, 0xb6, 0x1a, 0x58, 0x0b, 0x50, 0x12, 0x8c, 0xc3, 0xa1, - 0x3f, 0x08, 0x89, 0xf5, 0xd7, 0x34, 0x94, 0x3e, 0x1c, 0x91, 0xe0, 0x40, 0x89, 0xaa, 0x42, 0x3e, - 0x24, 0x7d, 0xd2, 0xa2, 0x7e, 0x50, 0x41, 0x2b, 0x68, 0xb5, 0x60, 0x47, 0x73, 0xbc, 0x04, 0xd9, - 0xbe, 0xeb, 0xb9, 0xb4, 0x92, 0x5e, 0x41, 0xab, 0xf3, 0xb6, 0x98, 0xe0, 0x6b, 0x90, 0x0d, 0xa9, - 0x13, 0xd0, 0x4a, 0x66, 0x05, 0xad, 0x16, 0xd7, 0xaa, 0x75, 0xa1, 0x7e, 0x5d, 0xa9, 0x5f, 0xbf, - 0xa7, 0xd4, 0x6f, 0xe6, 0x1f, 0x8f, 0xcd, 0xd4, 0xe7, 0xff, 0x30, 0x91, 0x2d, 0x8e, 0xe0, 0xef, - 0x41, 0x86, 0x0c, 0xda, 0x15, 0x63, 0x86, 0x93, 0xec, 0x00, 0xbe, 0x02, 0x85, 0xb6, 0x1b, 0x90, - 0x16, 0x75, 0xfd, 0x41, 0x25, 0xbb, 0x82, 0x56, 0x17, 0xd6, 0x4e, 0xc4, 0x90, 0x6c, 0xaa, 0x25, - 0x3b, 0xde, 0x85, 0x2f, 0x41, 0x2e, 0xec, 0x39, 0x41, 0x3b, 0xac, 0xcc, 0xad, 0x64, 0x56, 0x0b, - 0xcd, 0xa5, 0xa3, 0xb1, 0x59, 0x16, 0x94, 0x4b, 0xbe, 0xe7, 0x52, 0xe2, 0x0d, 0xe9, 0x81, 0x2d, - 0xf7, 0xe0, 0x8b, 0x30, 0xd7, 0x26, 0x7d, 0x42, 0x49, 0x58, 0xc9, 0x73, 0xc4, 0xcb, 0x1a, 0x7b, - 0xbe, 0x60, 0xab, 0x0d, 0xb7, 0x8d, 0x7c, 0xae, 0x3c, 0x67, 0xfd, 0x1b, 0x01, 0xde, 0x75, 0xbc, - 0x61, 0x9f, 0xbc, 0x30, 0x9e, 0x11, 0x72, 0xe9, 0x97, 0x46, 0x2e, 0x33, 0x2b, 0x72, 0x31, 0x0c, - 0xc6, 0x6c, 0x30, 0x64, 0xbf, 0x06, 0x06, 0x6b, 0x1b, 0x72, 0x82, 0xf4, 0x75, 0x36, 0x14, 0xeb, - 0x9c, 0x51, 0xda, 0x94, 0x63, 0x6d, 0x32, 0xfc, 0x9e, 0xd6, 0x4f, 0x61, 0x5e, 0xe2, 0x28, 0x2c, - 0x15, 0xaf, 0xbf, 0xb0, 0x0f, 0x2c, 0x3c, 0x1e, 0x9b, 0x28, 0xf6, 0x83, 0xc8, 0xf8, 0xf1, 0x9b, - 0x5c, 0x36, 0x0d, 0x25, 0xde, 0x8b, 0x75, 0xe1, 0x72, 0x5b, 0x83, 0x2e, 0x09, 0xd9, 0x41, 0x83, - 0x41, 0x65, 0x8b, 0x3d, 0xd6, 0x4f, 0xe0, 0xc4, 0xc4, 0x73, 0xca, 0x6b, 0x5c, 0x85, 0x5c, 0x48, - 0x02, 0x97, 0xa8, 0x5b, 0x68, 0x80, 0xec, 0x72, 0xba, 0x26, 0x9e, 0xcf, 0x6d, 0xb9, 0x7f, 0x36, - 0xe9, 0xbf, 0x47, 0x50, 0xda, 0x76, 0xf6, 0x48, 0x5f, 0xd9, 0x11, 0x06, 0x63, 0xe0, 0x78, 0x44, - 0xe2, 0xc9, 0xc7, 0x78, 0x19, 0x72, 0x9f, 0x38, 0xfd, 0x11, 0x11, 0x2c, 0xf3, 0xb6, 0x9c, 0xcd, - 0xea, 0x91, 0xe8, 0xa5, 0x3d, 0x12, 0x45, 0x76, 0x65, 0x5d, 0x80, 0x79, 0x79, 0x5f, 0x09, 0x54, - 0x7c, 0x39, 0x06, 0x54, 0x41, 0x5d, 0xce, 0xfa, 0x05, 0x82, 0xf9, 0x89, 0xf7, 0xc2, 0x16, 0xe4, - 0xfa, 0xec, 0x68, 0x28, 0x94, 0x6b, 0xc2, 0xd1, 0xd8, 0x94, 0x14, 0x5b, 0xfe, 0xb2, 0xd7, 0x27, - 0x03, 0xca, 0x71, 0x4f, 0x73, 0xdc, 0x97, 0x63, 0xdc, 0xdf, 0x1b, 0xd0, 0xe0, 0x40, 0x3d, 0xfe, - 0x22, 0x43, 0x91, 0x85, 0x3e, 0xb9, 0xdd, 0x56, 0x03, 0xfc, 0x3a, 0x18, 0x3d, 0x27, 0xec, 0x71, - 0x50, 0x8c, 0x66, 0xf6, 0x68, 0x6c, 0xa2, 0xcb, 0x36, 0x27, 0x59, 0x9f, 0x40, 0x49, 0x67, 0x82, - 0x6f, 0x41, 0x21, 0x8a, 0xd9, 0xfc, 0x52, 0xff, 0x1d, 0x8a, 0x05, 0x29, 0x33, 0x4d, 0x43, 0x0e, - 0x48, 0x7c, 0x18, 0x9f, 0x01, 0xa3, 0xef, 0x0e, 0x08, 0x7f, 0xa0, 0x42, 0x33, 0x7f, 0x34, 0x36, - 0xf9, 0xdc, 0xe6, 0x7f, 0x2d, 0x0f, 0x72, 0xc2, 0xc6, 0xf0, 0x1b, 0x49, 0x89, 0x99, 0x66, 0x4e, - 0x70, 0xd4, 0xb9, 0x99, 0x90, 0xe5, 0x28, 0x72, 0x76, 0xa8, 0x59, 0x38, 0x1a, 0x9b, 0x82, 0x60, - 0x8b, 0x1f, 0x26, 0x4e, 0xd3, 0x91, 0x8b, 0x63, 0x73, 0xa9, 0xe6, 0x4d, 0x28, 0x6d, 0x93, 0xae, - 0xd3, 0x3a, 0x90, 0x42, 0x97, 0x14, 0x3b, 0x26, 0x10, 0x29, 0x1e, 0xe7, 0xa0, 0x14, 0x49, 0x7c, - 0xe0, 0x85, 0xd2, 0x51, 0x8b, 0x11, 0xed, 0xfd, 0xd0, 0xfa, 0x15, 0x02, 0x69, 0xdd, 0x2f, 0xf4, - 0x78, 0xd7, 0x61, 0x2e, 0xe4, 0x12, 0xd5, 0xe3, 0xe9, 0x4e, 0xc3, 0x17, 0xe2, 0x67, 0x93, 0x1b, - 0x6d, 0x35, 0xc0, 0x75, 0x00, 0xe1, 0xbf, 0xb7, 0x62, 0xc5, 0x16, 0x8e, 0xc6, 0xa6, 0x46, 0xb5, - 0xb5, 0xb1, 0xf5, 0x4b, 0x04, 0xc5, 0x7b, 0x8e, 0x1b, 0x39, 0xce, 0x12, 0x64, 0x1f, 0x32, 0x0f, - 0x96, 0x9e, 0x23, 0x26, 0x2c, 0x44, 0xb5, 0x49, 0xdf, 0x39, 0xb8, 0xe1, 0x07, 0x9c, 0xe7, 0xbc, - 0x1d, 0xcd, 0xe3, 0x34, 0x67, 0x4c, 0x4d, 0x73, 0xd9, 0x99, 0x83, 0xf5, 0x6d, 0x23, 0x9f, 0x2e, - 0x67, 0xac, 0x9f, 0x23, 0x28, 0x89, 0x9b, 0x49, 0x17, 0xb9, 0x0e, 0x39, 0x71, 0x71, 0x69, 0x63, - 0xc7, 0x46, 0x34, 0xd0, 0xa2, 0x99, 0x3c, 0x82, 0x7f, 0x08, 0x0b, 0xed, 0xc0, 0x1f, 0x0e, 0x49, - 0x7b, 0x57, 0x86, 0xc5, 0x74, 0x32, 0x2c, 0x6e, 0xea, 0xeb, 0x76, 0x62, 0xbb, 0xf5, 0x37, 0xe6, - 0x88, 0x22, 0x44, 0x49, 0xa8, 0x22, 0x15, 0xd1, 0x4b, 0xe7, 0xa3, 0xf4, 0xac, 0xf9, 0x68, 0x19, - 0x72, 0xdd, 0xc0, 0x1f, 0x0d, 0xc3, 0x4a, 0x46, 0x84, 0x09, 0x31, 0x9b, 0x2d, 0x4f, 0x59, 0xb7, - 0x61, 0x41, 0xa9, 0x72, 0x4c, 0x9c, 0xae, 0x26, 0xe3, 0xf4, 0x56, 0x9b, 0x0c, 0xa8, 0xdb, 0x71, - 0xa3, 0xc8, 0x2b, 0xf7, 0x5b, 0x9f, 0x21, 0x28, 0x27, 0xb7, 0xe0, 0x1f, 0x68, 0x66, 0xce, 0xd8, - 0x9d, 0x3f, 0x9e, 0x5d, 0x9d, 0xc7, 0xc1, 0x90, 0x07, 0x14, 0xe5, 0x02, 0xd5, 0x77, 0xa0, 0xa8, - 0x91, 0x59, 0xbe, 0xdb, 0x27, 0xca, 0x24, 0xd9, 0x30, 0xf6, 0xc5, 0xb4, 0x30, 0x53, 0x3e, 0xb9, - 0x96, 0xbe, 0x8a, 0x98, 0x41, 0xcf, 0x4f, 0xbc, 0x24, 0xbe, 0x0a, 0x46, 0x27, 0xf0, 0xbd, 0x99, - 0x9e, 0x89, 0x9f, 0xc0, 0xdf, 0x81, 0x34, 0xf5, 0x67, 0x7a, 0xa4, 0x34, 0xf5, 0xd9, 0x1b, 0x49, - 0xe5, 0x33, 0xfc, 0x72, 0x72, 0x66, 0xfd, 0x0e, 0xc1, 0x22, 0x3b, 0x23, 0x10, 0xd8, 0xe8, 0x8d, - 0x06, 0xfb, 0x78, 0x15, 0xca, 0x4c, 0xd2, 0x03, 0x57, 0xa6, 0xb5, 0x07, 0x6e, 0x5b, 0xaa, 0xb9, - 0xc0, 0xe8, 0x2a, 0xdb, 0x6d, 0xb5, 0xf1, 0x29, 0x98, 0x1b, 0x85, 0x62, 0x83, 0xd0, 0x39, 0xc7, - 0xa6, 0x5b, 0x6d, 0xfc, 0xa6, 0x26, 0x8e, 0x61, 0xad, 0x55, 0x76, 0x1c, 0xc3, 0x1d, 0xc7, 0x0d, - 0xa2, 0xd8, 0x72, 0x01, 0x72, 0x2d, 0x26, 0x58, 0xd8, 0x09, 0x4b, 0xab, 0xd1, 0x66, 0x7e, 0x21, - 0x5b, 0x2e, 0x5b, 0xdf, 0x85, 0x42, 0x74, 0x7a, 0x6a, 0x36, 0x9d, 0xfa, 0x02, 0xd6, 0x75, 0x58, - 0x14, 0x31, 0x73, 0xfa, 0xe1, 0xd2, 0xb4, 0xc3, 0x25, 0x75, 0xf8, 0x34, 0x64, 0x05, 0x2a, 0x18, - 0x8c, 0xb6, 0x43, 0x1d, 0x75, 0x84, 0x8d, 0xad, 0x0a, 0x2c, 0xdf, 0x0b, 0x9c, 0x41, 0xd8, 0x21, - 0x01, 0xdf, 0x14, 0xd9, 0xae, 0x75, 0x12, 0x4e, 0xb0, 0x38, 0x41, 0x82, 0x70, 0xc3, 0x1f, 0x0d, - 0xa8, 0x74, 0x4f, 0xeb, 0x12, 0x2c, 0x4d, 0x92, 0xa5, 0xa9, 0x2f, 0x41, 0xb6, 0xc5, 0x08, 0x9c, - 0xfb, 0xbc, 0x2d, 0x26, 0xd6, 0xaf, 0x11, 0xe0, 0x9b, 0x84, 0x72, 0xd6, 0x5b, 0x9b, 0xa1, 0x56, - 0x8f, 0x7a, 0x0e, 0x6d, 0xf5, 0x48, 0x10, 0xaa, 0xda, 0x4c, 0xcd, 0xff, 0x1f, 0xf5, 0xa8, 0x75, - 0x05, 0x4e, 0x4c, 0xdc, 0x52, 0xea, 0x54, 0x85, 0x7c, 0x4b, 0xd2, 0x64, 0xfd, 0x10, 0xcd, 0xad, - 0x3f, 0xa4, 0x21, 0x2f, 0xde, 0x96, 0x74, 0xf0, 0x15, 0x28, 0x76, 0x98, 0xad, 0x05, 0xc3, 0xc0, - 0x95, 0x10, 0x18, 0xcd, 0xc5, 0xa3, 0xb1, 0xa9, 0x93, 0x6d, 0x7d, 0x82, 0x2f, 0x27, 0x0c, 0xaf, - 0xb9, 0x74, 0x38, 0x36, 0x73, 0x3f, 0x62, 0xc6, 0xb7, 0xc9, 0xb2, 0x17, 0x37, 0xc3, 0xcd, 0xc8, - 0x1c, 0xef, 0x48, 0x6f, 0xe3, 0xc5, 0x69, 0xf3, 0x6d, 0x76, 0xfd, 0x2f, 0xc7, 0xe6, 0x05, 0xad, - 0x27, 0x1c, 0x06, 0xbe, 0x47, 0x68, 0x8f, 0x8c, 0xc2, 0x46, 0xcb, 0xf7, 0x3c, 0x7f, 0xd0, 0xe0, - 0x7d, 0x1d, 0x57, 0x9a, 0xa5, 0x60, 0x76, 0x5c, 0x3a, 0xe0, 0x3d, 0x98, 0xa3, 0xbd, 0xc0, 0x1f, - 0x75, 0x7b, 0x3c, 0xbb, 0x64, 0x9a, 0xd7, 0x66, 0xe7, 0xa7, 0x38, 0xd8, 0x6a, 0x80, 0xcf, 0x31, - 0xb4, 0x48, 0x6b, 0x3f, 0x1c, 0x79, 0x3c, 0x3d, 0xcd, 0xab, 0xf2, 0x26, 0x22, 0x5b, 0x9f, 0xa5, - 0xc1, 0xe4, 0x26, 0x7c, 0x9f, 0x97, 0x61, 0x37, 0xfc, 0xe0, 0x7d, 0x42, 0x03, 0xb7, 0x75, 0xd7, - 0xf1, 0x88, 0xb2, 0x0d, 0x13, 0x8a, 0x1e, 0x27, 0x3e, 0xd0, 0x9c, 0x03, 0xbc, 0x68, 0x1f, 0x3e, - 0x0b, 0xc0, 0xdd, 0x4e, 0xac, 0x0b, 0x3f, 0x29, 0x70, 0x0a, 0x5f, 0xde, 0x98, 0x40, 0xaa, 0x31, - 0xa3, 0x66, 0x12, 0xa1, 0xad, 0x24, 0x42, 0x33, 0xf3, 0x89, 0x60, 0xd1, 0x6d, 0x3d, 0x3b, 0x69, - 0xeb, 0xd6, 0xdf, 0x11, 0xd4, 0xb6, 0xd5, 0xcd, 0x5f, 0x12, 0x0e, 0xa5, 0x6f, 0xfa, 0x15, 0xe9, - 0x9b, 0xf9, 0xdf, 0xf4, 0xb5, 0xfe, 0xa2, 0xb9, 0xbc, 0x4d, 0x3a, 0x4a, 0x8f, 0x0d, 0x2d, 0x5d, - 0xbc, 0x8a, 0x6b, 0xa6, 0x5f, 0xe1, 0xb3, 0x64, 0x12, 0xcf, 0xf2, 0x6e, 0x1c, 0x0e, 0xb8, 0x06, - 0x32, 0x1c, 0x9c, 0x07, 0x23, 0x20, 0x1d, 0x95, 0x7c, 0x71, 0x32, 0xc6, 0x93, 0x8e, 0xcd, 0xd7, - 0xad, 0x3f, 0x21, 0x28, 0xdf, 0x24, 0x74, 0xb2, 0xac, 0xf9, 0x26, 0xe9, 0x7f, 0x0b, 0x5e, 0xd3, - 0xee, 0x2f, 0xb5, 0x7f, 0x2b, 0x51, 0xcb, 0x9c, 0x8c, 0xf5, 0xdf, 0x1a, 0xb4, 0xc9, 0xa7, 0xb2, - 0xf1, 0x9c, 0x2c, 0x63, 0x76, 0xa0, 0xa8, 0x2d, 0xe2, 0xf5, 0x44, 0x01, 0x33, 0x2d, 0xa9, 0x36, - 0x97, 0xa4, 0x4e, 0xa2, 0xf5, 0x94, 0xd5, 0x67, 0x94, 0xee, 0x77, 0x01, 0xf3, 0x5e, 0x98, 0xb3, - 0xd5, 0x23, 0x35, 0xa7, 0xde, 0x89, 0xea, 0x99, 0x68, 0x8e, 0xcf, 0x81, 0x11, 0xf8, 0x8f, 0x54, - 0x65, 0x3a, 0x1f, 0x8b, 0xb4, 0xfd, 0x47, 0x36, 0x5f, 0xb2, 0xae, 0x43, 0xc6, 0xf6, 0x1f, 0xe1, - 0x1a, 0x40, 0xe0, 0x0c, 0xba, 0xe4, 0x7e, 0xd4, 0x8f, 0x94, 0x6c, 0x8d, 0x72, 0x4c, 0x7e, 0xdd, - 0x80, 0xd7, 0xf4, 0x1b, 0x89, 0xe7, 0xae, 0xc3, 0x1c, 0x23, 0xc6, 0x70, 0x2d, 0x25, 0xe0, 0x12, - 0x0d, 0xbd, 0xda, 0xc4, 0x6c, 0x06, 0x62, 0x3a, 0x3e, 0x03, 0x05, 0xea, 0xec, 0xf5, 0xc9, 0xdd, - 0xd8, 0xe7, 0x63, 0x02, 0x5b, 0x65, 0xad, 0xd4, 0x7d, 0xad, 0x50, 0x88, 0x09, 0xf8, 0x22, 0x94, - 0xe3, 0x3b, 0xef, 0x04, 0xa4, 0xe3, 0x7e, 0xca, 0x5f, 0xb8, 0x64, 0x3f, 0x47, 0xc7, 0xab, 0xb0, - 0x18, 0xd3, 0x76, 0x79, 0xda, 0x35, 0xf8, 0xd6, 0x24, 0x99, 0x61, 0xc3, 0xd5, 0x7d, 0xef, 0xe1, - 0xc8, 0xe9, 0xf3, 0x40, 0x56, 0xb2, 0x35, 0x8a, 0xf5, 0x67, 0x04, 0xaf, 0x89, 0xa7, 0xa6, 0x0e, - 0xfd, 0x46, 0x5a, 0xfd, 0x6f, 0x10, 0x60, 0x5d, 0x03, 0x69, 0x5a, 0xdf, 0xd2, 0x3f, 0xf9, 0xb0, - 0xbc, 0x5e, 0x9c, 0xf6, 0x4d, 0x93, 0xb5, 0xa0, 0xb2, 0x04, 0x4c, 0xf3, 0x5d, 0xbc, 0x05, 0x15, - 0x14, 0x55, 0xfd, 0xb1, 0xce, 0x79, 0xef, 0x80, 0x92, 0x50, 0x36, 0x90, 0xbc, 0x73, 0xe6, 0x04, - 0x5b, 0xfc, 0x30, 0x59, 0xea, 0x03, 0x83, 0x11, 0xcb, 0x4a, 0x7e, 0x44, 0xb8, 0x78, 0x1e, 0x0a, - 0xd1, 0xd7, 0x45, 0x5c, 0x84, 0xb9, 0x1b, 0x1f, 0xd8, 0x1f, 0xad, 0xdb, 0x9b, 0xe5, 0x14, 0x2e, - 0x41, 0xbe, 0xb9, 0xbe, 0x71, 0x87, 0xcf, 0xd0, 0xda, 0x3a, 0xe4, 0x76, 0x46, 0x61, 0x8f, 0x04, - 0xf8, 0x6d, 0x30, 0xd8, 0x08, 0x6b, 0x4e, 0xab, 0x7d, 0xda, 0xad, 0x2e, 0x27, 0xc9, 0xb2, 0x06, - 0x4c, 0xad, 0xfd, 0xd1, 0x50, 0x86, 0x1c, 0xe0, 0xef, 0x43, 0x56, 0x58, 0xa7, 0xb6, 0x5d, 0xff, - 0xcc, 0x58, 0x3d, 0xf5, 0x1c, 0x5d, 0xf1, 0xf9, 0x36, 0xc2, 0x77, 0xa1, 0xc8, 0x89, 0xb2, 0xed, - 0x3f, 0x93, 0xec, 0xbe, 0x27, 0x38, 0x9d, 0x3d, 0x66, 0x55, 0xe3, 0x77, 0x0d, 0xb2, 0x3c, 0x40, - 0xe8, 0xb7, 0xd1, 0x3f, 0x56, 0xe9, 0xb7, 0x99, 0xf8, 0x28, 0x64, 0xa5, 0xf0, 0x3b, 0x60, 0xb0, - 0x22, 0x56, 0x87, 0x43, 0xeb, 0xd6, 0x75, 0x38, 0xf4, 0x56, 0x99, 0x8b, 0x7d, 0x37, 0xfa, 0xe8, - 0x70, 0x2a, 0xd9, 0x7d, 0xa9, 0xe3, 0x95, 0xe7, 0x17, 0x22, 0xc9, 0x1f, 0x88, 0xee, 0x5b, 0x95, - 0xcf, 0xf8, 0xec, 0xa4, 0xa8, 0x44, 0xb5, 0x5d, 0xad, 0x1d, 0xb7, 0x1c, 0x31, 0xdc, 0x86, 0xa2, - 0x56, 0xba, 0xea, 0xb0, 0x3e, 0x5f, 0x77, 0xeb, 0xb0, 0x4e, 0xa9, 0x77, 0xad, 0x14, 0xbe, 0x09, - 0x79, 0x16, 0xf9, 0x99, 0x03, 0xe0, 0xd3, 0xc9, 0x00, 0xaf, 0x39, 0x76, 0xf5, 0xcc, 0xf4, 0xc5, - 0xc8, 0x6e, 0x7e, 0x0c, 0x79, 0xd5, 0x65, 0xe1, 0x0f, 0x61, 0x61, 0xb2, 0xc7, 0xc0, 0xaf, 0x6b, - 0x6a, 0x4d, 0xb6, 0x6e, 0xd5, 0x15, 0x6d, 0x69, 0x7a, 0x63, 0x92, 0x5a, 0x45, 0xcd, 0x8f, 0x9f, - 0x3c, 0xad, 0xa5, 0xbe, 0x78, 0x5a, 0x4b, 0x7d, 0xf5, 0xb4, 0x86, 0x7e, 0x76, 0x58, 0x43, 0xbf, - 0x3d, 0xac, 0xa1, 0xc7, 0x87, 0x35, 0xf4, 0xe4, 0xb0, 0x86, 0xfe, 0x79, 0x58, 0x43, 0xff, 0x3a, - 0xac, 0xa5, 0xbe, 0x3a, 0xac, 0xa1, 0xcf, 0x9f, 0xd5, 0x52, 0x4f, 0x9e, 0xd5, 0x52, 0x5f, 0x3c, - 0xab, 0xa5, 0x3e, 0x7e, 0x43, 0xff, 0x17, 0x4a, 0xe0, 0x74, 0x9c, 0x81, 0xd3, 0xe8, 0xfb, 0xfb, - 0x6e, 0x43, 0xff, 0x0f, 0xcc, 0x5e, 0x8e, 0xff, 0xbc, 0xf5, 0x9f, 0x00, 0x00, 0x00, 0xff, 0xff, - 0xdc, 0x73, 0x3e, 0x8e, 0x98, 0x19, 0x00, 0x00, + // 2098 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xd4, 0x59, 0xcd, 0x6f, 0x1b, 0xc7, + 0x15, 0xe7, 0x90, 0x4b, 0x8a, 0x7c, 0xa4, 0x24, 0x7a, 0x2c, 0xdb, 0x0c, 0x6d, 0x73, 0xe5, 0x45, + 0x6a, 0x0b, 0xfe, 0x20, 0x6b, 0xa5, 0x6d, 0x1c, 0xbb, 0x69, 0x21, 0x4a, 0xb1, 0x2d, 0x5b, 0x71, + 0x94, 0x95, 0xeb, 0x00, 0x01, 0x0a, 0x63, 0x45, 0x0e, 0xc9, 0x85, 0xb8, 0x5c, 0x7a, 0x77, 0x18, + 0x47, 0x40, 0x81, 0xf6, 0x0f, 0x68, 0x80, 0xf4, 0x54, 0xf4, 0x5e, 0xa0, 0x45, 0x0f, 0x3d, 0x14, + 0xe8, 0xb5, 0xed, 0xad, 0xee, 0xcd, 0xbd, 0x05, 0x39, 0xb0, 0xb5, 0x7c, 0x29, 0x74, 0xca, 0x5f, + 0x50, 0x14, 0xf3, 0xb5, 0x3b, 0x5c, 0x53, 0x8d, 0xe9, 0x1a, 0x28, 0x72, 0x11, 0x67, 0xde, 0xcc, + 0xbc, 0x37, 0xef, 0x37, 0xef, 0x73, 0x05, 0xa7, 0x87, 0x7b, 0xdd, 0x46, 0xdf, 0xef, 0x0e, 0x03, + 0x9f, 0xfa, 0xd1, 0xa0, 0xce, 0xff, 0xe2, 0xbc, 0x9a, 0x57, 0xaf, 0x74, 0x5d, 0xda, 0x1b, 0xed, + 0xd6, 0x5b, 0xbe, 0xd7, 0xe8, 0xfa, 0x5d, 0xbf, 0xc1, 0xc9, 0xbb, 0xa3, 0x0e, 0x9f, 0x89, 0xc3, + 0x6c, 0x24, 0x0e, 0x56, 0xcd, 0xae, 0xef, 0x77, 0xfb, 0x24, 0xde, 0x45, 0x5d, 0x8f, 0x84, 0xd4, + 0xf1, 0x86, 0x72, 0xc3, 0xb2, 0x14, 0xfb, 0xa8, 0xef, 0xf9, 0x6d, 0xd2, 0x6f, 0x84, 0xd4, 0xa1, + 0xa1, 0xf8, 0x2b, 0x76, 0x58, 0x1f, 0x41, 0x71, 0x7b, 0x14, 0xf6, 0x6c, 0xf2, 0x68, 0x44, 0x42, + 0x8a, 0x6f, 0xc3, 0x5c, 0x48, 0x03, 0xe2, 0x78, 0x61, 0x05, 0x2d, 0x67, 0x56, 0x8a, 0xab, 0xa7, + 0xea, 0xd1, 0x65, 0x77, 0xf8, 0xc2, 0x5a, 0xdb, 0x19, 0x52, 0x12, 0x34, 0x4f, 0x7c, 0x39, 0x36, + 0x73, 0x82, 0x74, 0x38, 0x36, 0xd5, 0x29, 0x5b, 0x0d, 0xac, 0x05, 0x28, 0x09, 0xc6, 0xe1, 0xd0, + 0x1f, 0x84, 0xc4, 0xfa, 0x6b, 0x1a, 0x4a, 0x1f, 0x8e, 0x48, 0xb0, 0xaf, 0x44, 0x55, 0x21, 0x1f, + 0x92, 0x3e, 0x69, 0x51, 0x3f, 0xa8, 0xa0, 0x65, 0xb4, 0x52, 0xb0, 0xa3, 0x39, 0x5e, 0x82, 0x6c, + 0xdf, 0xf5, 0x5c, 0x5a, 0x49, 0x2f, 0xa3, 0x95, 0x79, 0x5b, 0x4c, 0xf0, 0x75, 0xc8, 0x86, 0xd4, + 0x09, 0x68, 0x25, 0xb3, 0x8c, 0x56, 0x8a, 0xab, 0xd5, 0xba, 0x50, 0xbf, 0xae, 0xd4, 0xaf, 0xdf, + 0x57, 0xea, 0x37, 0xf3, 0x4f, 0xc6, 0x66, 0xea, 0xf3, 0x7f, 0x98, 0xc8, 0x16, 0x47, 0xf0, 0xf7, + 0x20, 0x43, 0x06, 0xed, 0x8a, 0x31, 0xc3, 0x49, 0x76, 0x00, 0x5f, 0x85, 0x42, 0xdb, 0x0d, 0x48, + 0x8b, 0xba, 0xfe, 0xa0, 0x92, 0x5d, 0x46, 0x2b, 0x0b, 0xab, 0xc7, 0x63, 0x48, 0x36, 0xd4, 0x92, + 0x1d, 0xef, 0xc2, 0x97, 0x21, 0x17, 0xf6, 0x9c, 0xa0, 0x1d, 0x56, 0xe6, 0x96, 0x33, 0x2b, 0x85, + 0xe6, 0xd2, 0xe1, 0xd8, 0x2c, 0x0b, 0xca, 0x65, 0xdf, 0x73, 0x29, 0xf1, 0x86, 0x74, 0xdf, 0x96, + 0x7b, 0xf0, 0x45, 0x98, 0x6b, 0x93, 0x3e, 0xa1, 0x24, 0xac, 0xe4, 0x39, 0xe2, 0x65, 0x8d, 0x3d, + 0x5f, 0xb0, 0xd5, 0x86, 0x3b, 0x46, 0x3e, 0x57, 0x9e, 0xb3, 0xfe, 0x8d, 0x00, 0xef, 0x38, 0xde, + 0xb0, 0x4f, 0x5e, 0x1a, 0xcf, 0x08, 0xb9, 0xf4, 0x2b, 0x23, 0x97, 0x99, 0x15, 0xb9, 0x18, 0x06, + 0x63, 0x36, 0x18, 0xb2, 0x5f, 0x03, 0x83, 0xb5, 0x05, 0x39, 0x41, 0xfa, 0x3a, 0x1b, 0x8a, 0x75, + 0xce, 0x28, 0x6d, 0xca, 0xb1, 0x36, 0x19, 0x7e, 0x4f, 0xeb, 0xa7, 0x30, 0x2f, 0x71, 0x14, 0x96, + 0x8a, 0xd7, 0x5e, 0xda, 0x07, 0x16, 0x9e, 0x8c, 0x4d, 0x14, 0xfb, 0x41, 0x64, 0xfc, 0xf8, 0x12, + 0x97, 0x4d, 0x43, 0x89, 0xf7, 0x62, 0x5d, 0xb8, 0xdc, 0xe6, 0xa0, 0x4b, 0x42, 0x76, 0xd0, 0x60, + 0x50, 0xd9, 0x62, 0x8f, 0xf5, 0x13, 0x38, 0x3e, 0xf1, 0x9c, 0xf2, 0x1a, 0xd7, 0x20, 0x17, 0x92, + 0xc0, 0x25, 0xea, 0x16, 0x1a, 0x20, 0x3b, 0x9c, 0xae, 0x89, 0xe7, 0x73, 0x5b, 0xee, 0x9f, 0x4d, + 0xfa, 0xef, 0x11, 0x94, 0xb6, 0x9c, 0x5d, 0xd2, 0x57, 0x76, 0x84, 0xc1, 0x18, 0x38, 0x1e, 0x91, + 0x78, 0xf2, 0x31, 0x3e, 0x09, 0xb9, 0x4f, 0x9c, 0xfe, 0x88, 0x08, 0x96, 0x79, 0x5b, 0xce, 0x66, + 0xf5, 0x48, 0xf4, 0xca, 0x1e, 0x89, 0x22, 0xbb, 0xb2, 0x2e, 0xc1, 0x31, 0xdb, 0xa1, 0x64, 0x8b, + 0x85, 0x04, 0xd2, 0x16, 0xc8, 0xb3, 0x0b, 0xf6, 0x99, 0x12, 0xa1, 0xbc, 0xb6, 0x9c, 0x59, 0x17, + 0x60, 0x5e, 0x2a, 0x27, 0x51, 0x8d, 0x35, 0x61, 0xa8, 0x16, 0x94, 0x26, 0xd6, 0x2f, 0x10, 0xcc, + 0x4f, 0x3c, 0x2e, 0xb6, 0x26, 0x59, 0x36, 0xe1, 0x70, 0x6c, 0x4a, 0x8a, 0x62, 0xcf, 0x4c, 0x85, + 0x0c, 0x28, 0x7f, 0xa4, 0x34, 0x7f, 0xa4, 0x93, 0xf1, 0x23, 0xbd, 0x37, 0xa0, 0xc1, 0xbe, 0xb2, + 0x94, 0x45, 0x06, 0x39, 0x8b, 0x93, 0x72, 0xbb, 0xad, 0x06, 0xf8, 0x0d, 0x30, 0x7a, 0x4e, 0xd8, + 0xe3, 0x08, 0x1a, 0xcd, 0xec, 0xe1, 0xd8, 0x44, 0x57, 0x6c, 0x4e, 0xb2, 0x3e, 0x81, 0x92, 0xce, + 0x04, 0xdf, 0x86, 0x42, 0x14, 0xe0, 0xf9, 0xa5, 0xfe, 0x3b, 0x6e, 0x0b, 0x52, 0x66, 0x9a, 0x86, + 0x1c, 0xbd, 0xf8, 0x30, 0x3e, 0x03, 0x46, 0xdf, 0x1d, 0x10, 0xfe, 0x9a, 0x85, 0x66, 0xfe, 0x70, + 0x6c, 0xf2, 0xb9, 0xcd, 0xff, 0x5a, 0x1e, 0xe4, 0x84, 0x41, 0xe2, 0x37, 0x93, 0x12, 0x33, 0xcd, + 0x9c, 0xe0, 0xa8, 0x73, 0x33, 0x21, 0xcb, 0x51, 0xe4, 0xec, 0x50, 0xb3, 0x70, 0x38, 0x36, 0x05, + 0xc1, 0x16, 0x3f, 0x4c, 0x9c, 0xa6, 0x23, 0x17, 0xc7, 0xe6, 0x52, 0xcd, 0x5b, 0x50, 0xda, 0x22, + 0x5d, 0xa7, 0xb5, 0x2f, 0x85, 0x2e, 0x29, 0x76, 0x4c, 0x20, 0x52, 0x3c, 0xce, 0x41, 0x29, 0x92, + 0xf8, 0xd0, 0x0b, 0xa5, 0x57, 0x17, 0x23, 0xda, 0xfb, 0xa1, 0xf5, 0x2b, 0x04, 0xd2, 0x15, 0x5e, + 0xea, 0xf1, 0x6e, 0xc0, 0x5c, 0xc8, 0x25, 0xaa, 0xc7, 0xd3, 0x3d, 0x8c, 0x2f, 0xc4, 0xcf, 0x26, + 0x37, 0xda, 0x6a, 0x80, 0xeb, 0x00, 0xc2, 0xd9, 0x6f, 0xc7, 0x8a, 0x2d, 0x1c, 0x8e, 0x4d, 0x8d, + 0x6a, 0x6b, 0x63, 0xeb, 0x97, 0x08, 0x8a, 0xf7, 0x1d, 0x37, 0xf2, 0xb2, 0x25, 0xc8, 0x3e, 0x62, + 0xee, 0x2e, 0xed, 0x55, 0x4c, 0x58, 0x3c, 0x6b, 0x93, 0xbe, 0xb3, 0x7f, 0xd3, 0x0f, 0x38, 0xcf, + 0x79, 0x3b, 0x9a, 0xc7, 0x39, 0xd1, 0x98, 0x9a, 0x13, 0xb3, 0x33, 0x47, 0xf6, 0x3b, 0x46, 0x3e, + 0x5d, 0xce, 0x58, 0x3f, 0x47, 0x50, 0x12, 0x37, 0x93, 0x2e, 0x72, 0x03, 0x72, 0xe2, 0xe2, 0xd2, + 0xc6, 0x8e, 0x0c, 0x7f, 0xa0, 0x85, 0x3e, 0x79, 0x04, 0xff, 0x10, 0x16, 0xda, 0x81, 0x3f, 0x1c, + 0x2a, 0xcf, 0x54, 0xd8, 0x6a, 0x4c, 0x36, 0xf4, 0x75, 0x3b, 0xb1, 0xdd, 0xfa, 0x1b, 0x73, 0x44, + 0x11, 0xcf, 0x24, 0x54, 0x91, 0x8a, 0xe8, 0x95, 0x93, 0x57, 0x7a, 0xd6, 0xe4, 0x75, 0x12, 0x72, + 0xdd, 0xc0, 0x1f, 0x0d, 0xc3, 0x4a, 0x46, 0x84, 0x09, 0x31, 0x9b, 0x2d, 0xa9, 0x59, 0x77, 0x60, + 0x41, 0xa9, 0x72, 0x44, 0x50, 0xaf, 0x26, 0x83, 0xfa, 0x66, 0x9b, 0x0c, 0xa8, 0xdb, 0x71, 0xa3, + 0x30, 0x2d, 0xf7, 0x5b, 0x9f, 0x21, 0x28, 0x27, 0xb7, 0xe0, 0x1f, 0x68, 0x66, 0xce, 0xd8, 0x9d, + 0x3f, 0x9a, 0x5d, 0x9d, 0xc7, 0xc1, 0x90, 0x07, 0x14, 0xe5, 0x02, 0xd5, 0x77, 0xa0, 0xa8, 0x91, + 0x59, 0x72, 0xdc, 0x23, 0xca, 0x24, 0xd9, 0x30, 0xf6, 0xc5, 0xb4, 0x30, 0x53, 0x3e, 0xb9, 0x9e, + 0xbe, 0x86, 0x98, 0x41, 0xcf, 0x4f, 0xbc, 0x24, 0xbe, 0x06, 0x46, 0x27, 0xf0, 0xbd, 0x99, 0x9e, + 0x89, 0x9f, 0xc0, 0xdf, 0x81, 0x34, 0xf5, 0x67, 0x7a, 0xa4, 0x34, 0xf5, 0xb5, 0x98, 0x9f, 0x99, + 0x88, 0xf9, 0xbf, 0x43, 0xb0, 0xc8, 0xce, 0x08, 0x04, 0xd6, 0x7b, 0xa3, 0xc1, 0x1e, 0x5e, 0x81, + 0x32, 0x93, 0xf4, 0xd0, 0x95, 0x39, 0xf0, 0xa1, 0xdb, 0x96, 0x6a, 0x2e, 0x30, 0xba, 0x4a, 0x8d, + 0x9b, 0x6d, 0x7c, 0x0a, 0xe6, 0x46, 0xa1, 0xd8, 0x20, 0x74, 0xce, 0xb1, 0xe9, 0x66, 0x1b, 0x5f, + 0xd2, 0xc4, 0x31, 0xac, 0xb5, 0x32, 0x90, 0x63, 0xb8, 0xed, 0xb8, 0x41, 0x14, 0x5b, 0x2e, 0x40, + 0xae, 0xc5, 0x04, 0x0b, 0x3b, 0x61, 0x39, 0x38, 0xda, 0xcc, 0x2f, 0x64, 0xcb, 0x65, 0xeb, 0xbb, + 0x50, 0x88, 0x4e, 0x4f, 0x4d, 0xbd, 0x53, 0x5f, 0xc0, 0xba, 0x01, 0x8b, 0x22, 0x66, 0x4e, 0x3f, + 0x5c, 0x9a, 0x76, 0xb8, 0xa4, 0x0e, 0x9f, 0x86, 0xac, 0x40, 0x05, 0x83, 0xd1, 0x76, 0xa8, 0xa3, + 0x8e, 0xb0, 0xb1, 0x55, 0x81, 0x93, 0xf7, 0x03, 0x67, 0x10, 0x76, 0x48, 0xc0, 0x37, 0x45, 0xb6, + 0x6b, 0x9d, 0x80, 0xe3, 0x2c, 0x4e, 0x90, 0x20, 0x5c, 0xf7, 0x47, 0x03, 0x2a, 0xdd, 0xd3, 0xba, + 0x0c, 0x4b, 0x93, 0x64, 0x69, 0xea, 0x4b, 0x90, 0x6d, 0x31, 0x02, 0xe7, 0x3e, 0x6f, 0x8b, 0x89, + 0xf5, 0x6b, 0x04, 0xf8, 0x16, 0xa1, 0x9c, 0xf5, 0xe6, 0x46, 0xa8, 0x15, 0xaf, 0x9e, 0x43, 0x5b, + 0x3d, 0x12, 0xa8, 0x0c, 0x1e, 0xcd, 0xff, 0x1f, 0xc5, 0xab, 0x75, 0x15, 0x8e, 0x4f, 0xdc, 0x52, + 0xea, 0x54, 0x85, 0x7c, 0x4b, 0xd2, 0x64, 0xfd, 0x10, 0xcd, 0xad, 0x3f, 0xa4, 0x21, 0x2f, 0xde, + 0x96, 0x74, 0xf0, 0x55, 0x28, 0x76, 0x98, 0xad, 0x05, 0xc3, 0xc0, 0x95, 0x10, 0x18, 0xcd, 0xc5, + 0xc3, 0xb1, 0xa9, 0x93, 0x6d, 0x7d, 0x82, 0xaf, 0x24, 0x0c, 0xaf, 0xb9, 0x74, 0x30, 0x36, 0x73, + 0x3f, 0x62, 0xc6, 0xb7, 0xc1, 0xb2, 0x17, 0x37, 0xc3, 0x8d, 0xc8, 0x1c, 0xef, 0x4a, 0x6f, 0xe3, + 0x95, 0x6c, 0xf3, 0x6d, 0x76, 0xfd, 0x2f, 0xc7, 0xe6, 0x05, 0xad, 0x81, 0x1c, 0x06, 0xbe, 0x47, + 0x68, 0x8f, 0x8c, 0xc2, 0x46, 0xcb, 0xf7, 0x3c, 0x7f, 0xd0, 0xe0, 0x4d, 0x20, 0x57, 0x9a, 0xa5, + 0x60, 0x76, 0x5c, 0x3a, 0xe0, 0x7d, 0x98, 0xa3, 0xbd, 0xc0, 0x1f, 0x75, 0x7b, 0x3c, 0xbb, 0x64, + 0x9a, 0xd7, 0x67, 0xe7, 0xa7, 0x38, 0xd8, 0x6a, 0x80, 0xcf, 0x31, 0xb4, 0x48, 0x6b, 0x2f, 0x1c, + 0x79, 0x3c, 0x3d, 0xcd, 0xab, 0xf2, 0x26, 0x22, 0x5b, 0x9f, 0xa5, 0xc1, 0xe4, 0x26, 0xfc, 0x80, + 0x97, 0x61, 0x37, 0xfd, 0xe0, 0x7d, 0x42, 0x03, 0xb7, 0x75, 0xcf, 0xf1, 0x88, 0xb2, 0x0d, 0x13, + 0x8a, 0x1e, 0x27, 0x3e, 0xd4, 0x9c, 0x03, 0xbc, 0x68, 0x1f, 0x3e, 0x0b, 0xc0, 0xdd, 0x4e, 0xac, + 0x0b, 0x3f, 0x29, 0x70, 0x0a, 0x5f, 0x5e, 0x9f, 0x40, 0xaa, 0x31, 0xa3, 0x66, 0x12, 0xa1, 0xcd, + 0x24, 0x42, 0x33, 0xf3, 0x89, 0x60, 0xd1, 0x6d, 0x3d, 0x3b, 0x69, 0xeb, 0xd6, 0xdf, 0x11, 0xd4, + 0xb6, 0xd4, 0xcd, 0x5f, 0x11, 0x0e, 0xa5, 0x6f, 0xfa, 0x35, 0xe9, 0x9b, 0xf9, 0xdf, 0xf4, 0xb5, + 0xfe, 0xa2, 0xb9, 0xbc, 0x4d, 0x3a, 0x4a, 0x8f, 0x75, 0x2d, 0x5d, 0xbc, 0x8e, 0x6b, 0xa6, 0x5f, + 0xe3, 0xb3, 0x64, 0x12, 0xcf, 0xf2, 0x6e, 0x1c, 0x0e, 0xb8, 0x06, 0x32, 0x1c, 0x9c, 0x07, 0x23, + 0x20, 0x1d, 0x95, 0x7c, 0x71, 0x32, 0xc6, 0x93, 0x8e, 0xcd, 0xd7, 0xad, 0x3f, 0x21, 0x28, 0xdf, + 0x22, 0x74, 0xb2, 0xac, 0xf9, 0x26, 0xe9, 0x7f, 0x1b, 0x8e, 0x69, 0xf7, 0x97, 0xda, 0xbf, 0x95, + 0xa8, 0x65, 0x4e, 0xc4, 0xfa, 0x6f, 0x0e, 0xda, 0xe4, 0x53, 0xd9, 0xa5, 0x4e, 0x96, 0x31, 0xdb, + 0x50, 0xd4, 0x16, 0xf1, 0x5a, 0xa2, 0x80, 0x99, 0x96, 0x54, 0x9b, 0x4b, 0x52, 0x27, 0xd1, 0xa7, + 0xca, 0xea, 0x33, 0x4a, 0xf7, 0x3b, 0x80, 0x79, 0xe3, 0xcc, 0xd9, 0xea, 0x91, 0x9a, 0x53, 0xef, + 0x46, 0xf5, 0x4c, 0x34, 0xc7, 0xe7, 0xc0, 0x08, 0xfc, 0xc7, 0xaa, 0x32, 0x9d, 0x8f, 0x45, 0xda, + 0xfe, 0x63, 0x9b, 0x2f, 0x59, 0x37, 0x20, 0x63, 0xfb, 0x8f, 0x71, 0x0d, 0x20, 0x70, 0x06, 0x5d, + 0xf2, 0x20, 0xea, 0x47, 0x4a, 0xb6, 0x46, 0x39, 0x22, 0xbf, 0xae, 0xc3, 0x31, 0xfd, 0x46, 0xe2, + 0xb9, 0xeb, 0x30, 0xc7, 0x88, 0x31, 0x5c, 0x4b, 0x09, 0xb8, 0x44, 0xf7, 0xaf, 0x36, 0x31, 0x9b, + 0x81, 0x98, 0x8e, 0xcf, 0x40, 0x81, 0x3a, 0xbb, 0x7d, 0x72, 0x2f, 0xf6, 0xf9, 0x98, 0xc0, 0x56, + 0x59, 0x2b, 0xf5, 0x40, 0x2b, 0x14, 0x62, 0x02, 0xbe, 0x08, 0xe5, 0xf8, 0xce, 0xdb, 0x01, 0xe9, + 0xb8, 0x9f, 0xf2, 0x17, 0x2e, 0xd9, 0x2f, 0xd0, 0xf1, 0x0a, 0x2c, 0xc6, 0xb4, 0x1d, 0x9e, 0x76, + 0x0d, 0xbe, 0x35, 0x49, 0x66, 0xd8, 0x70, 0x75, 0xdf, 0x7b, 0x34, 0x72, 0xfa, 0x3c, 0x90, 0x95, + 0x6c, 0x8d, 0x62, 0xfd, 0x19, 0xc1, 0x31, 0xf1, 0xd4, 0xd4, 0xa1, 0xdf, 0x48, 0xab, 0xff, 0x0d, + 0x02, 0xac, 0x6b, 0x20, 0x4d, 0xeb, 0x5b, 0xfa, 0xf7, 0x21, 0x96, 0xd7, 0x8b, 0xd3, 0x3e, 0x80, + 0xb2, 0x16, 0x54, 0x96, 0x80, 0x69, 0xbe, 0x8b, 0xb7, 0xa0, 0x82, 0xa2, 0xaa, 0x3f, 0xd6, 0x39, + 0xef, 0xee, 0x53, 0x12, 0xca, 0x06, 0x92, 0x77, 0xce, 0x9c, 0x60, 0x8b, 0x1f, 0x26, 0x4b, 0x7d, + 0x60, 0x30, 0x62, 0x59, 0xc9, 0x8f, 0x08, 0x17, 0xcf, 0x43, 0x21, 0xfa, 0x14, 0x89, 0x8b, 0x30, + 0x77, 0xf3, 0x03, 0xfb, 0xa3, 0x35, 0x7b, 0xa3, 0x9c, 0xc2, 0x25, 0xc8, 0x37, 0xd7, 0xd6, 0xef, + 0xf2, 0x19, 0x5a, 0x5d, 0x83, 0xdc, 0xf6, 0x28, 0xec, 0x91, 0x00, 0xbf, 0x0d, 0x06, 0x1b, 0x61, + 0xcd, 0x69, 0xb5, 0xef, 0xc0, 0xd5, 0x93, 0x49, 0xb2, 0xac, 0x01, 0x53, 0xab, 0x7f, 0x34, 0x94, + 0x21, 0x07, 0xf8, 0xfb, 0x90, 0x15, 0xd6, 0xa9, 0x6d, 0xd7, 0xbf, 0x49, 0x56, 0x4f, 0xbd, 0x40, + 0x57, 0x7c, 0xbe, 0x8d, 0xf0, 0x3d, 0x28, 0x72, 0xa2, 0x6c, 0xfb, 0xcf, 0x24, 0xbb, 0xef, 0x09, + 0x4e, 0x67, 0x8f, 0x58, 0xd5, 0xf8, 0x5d, 0x87, 0x2c, 0x0f, 0x10, 0xfa, 0x6d, 0xf4, 0x2f, 0x5b, + 0xfa, 0x6d, 0x26, 0x3e, 0x0a, 0x59, 0x29, 0xfc, 0x0e, 0x18, 0xac, 0x88, 0xd5, 0xe1, 0xd0, 0xba, + 0x75, 0x1d, 0x0e, 0xbd, 0x55, 0xe6, 0x62, 0xdf, 0x8d, 0x3e, 0x3a, 0x9c, 0x4a, 0x76, 0x5f, 0xea, + 0x78, 0xe5, 0xc5, 0x85, 0x48, 0xf2, 0x07, 0xa2, 0xfb, 0x56, 0xe5, 0x33, 0x3e, 0x3b, 0x29, 0x2a, + 0x51, 0x6d, 0x57, 0x6b, 0x47, 0x2d, 0x47, 0x0c, 0xb7, 0xa0, 0xa8, 0x95, 0xae, 0x3a, 0xac, 0x2f, + 0xd6, 0xdd, 0x3a, 0xac, 0x53, 0xea, 0x5d, 0x2b, 0x85, 0x6f, 0x41, 0x9e, 0x45, 0x7e, 0xe6, 0x00, + 0xf8, 0x74, 0x32, 0xc0, 0x6b, 0x8e, 0x5d, 0x3d, 0x33, 0x7d, 0x31, 0xb2, 0x9b, 0x1f, 0x43, 0x5e, + 0x75, 0x59, 0xf8, 0x43, 0x58, 0x98, 0xec, 0x31, 0xf0, 0x1b, 0x9a, 0x5a, 0x93, 0xad, 0x5b, 0x75, + 0x59, 0x5b, 0x9a, 0xde, 0x98, 0xa4, 0x56, 0x50, 0xf3, 0xe3, 0xa7, 0xcf, 0x6a, 0xa9, 0x2f, 0x9e, + 0xd5, 0x52, 0x5f, 0x3d, 0xab, 0xa1, 0x9f, 0x1d, 0xd4, 0xd0, 0x6f, 0x0f, 0x6a, 0xe8, 0xc9, 0x41, + 0x0d, 0x3d, 0x3d, 0xa8, 0xa1, 0x7f, 0x1e, 0xd4, 0xd0, 0xbf, 0x0e, 0x6a, 0xa9, 0xaf, 0x0e, 0x6a, + 0xe8, 0xf3, 0xe7, 0xb5, 0xd4, 0xd3, 0xe7, 0xb5, 0xd4, 0x17, 0xcf, 0x6b, 0xa9, 0x8f, 0xdf, 0xd4, + 0xff, 0xdf, 0x12, 0x38, 0x1d, 0x67, 0xe0, 0x34, 0xfa, 0xfe, 0x9e, 0xdb, 0xd0, 0xff, 0x5d, 0xb3, + 0x9b, 0xe3, 0x3f, 0x6f, 0xfd, 0x27, 0x00, 0x00, 0xff, 0xff, 0x20, 0xae, 0x71, 0xe2, 0xc5, 0x19, + 0x00, 0x00, } func (x Direction) String() string { @@ -2779,6 +2824,30 @@ func (this *LabelRequest) Equal(that interface{}) bool { } return true } +func (this *RateLimitedStream) Equal(that interface{}) bool { + if that == nil { + return this == nil + } + + that1, ok := that.(*RateLimitedStream) + if !ok { + that2, ok := that.(RateLimitedStream) + if ok { + that1 = &that2 + } else { + return false + } + } + if that1 == nil { + return this == nil + } else if this == nil { + return false + } + if this.Labels != that1.Labels { + return false + } + return true +} func (this *LabelResponse) Equal(that interface{}) bool { if that == nil { return this == nil @@ -3944,6 +4013,16 @@ func (this *LabelRequest) GoString() string { s = append(s, "}") return strings.Join(s, "") } +func (this *RateLimitedStream) GoString() string { + if this == nil { + return "nil" + } + s := make([]string, 0, 5) + s = append(s, "&logproto.RateLimitedStream{") + s = append(s, "Labels: "+fmt.Sprintf("%#v", this.Labels)+",\n") + s = append(s, "}") + return strings.Join(s, "") +} func (this *LabelResponse) GoString() string { if this == nil { return "nil" @@ -5396,6 +5475,36 @@ func (m *LabelRequest) MarshalToSizedBuffer(dAtA []byte) (int, error) { return len(dAtA) - i, nil } +func (m *RateLimitedStream) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *RateLimitedStream) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *RateLimitedStream) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if len(m.Labels) > 0 { + i -= len(m.Labels) + copy(dAtA[i:], m.Labels) + i = encodeVarintLogproto(dAtA, i, uint64(len(m.Labels))) + i-- + dAtA[i] = 0xa + } + return len(dAtA) - i, nil +} + func (m *LabelResponse) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) @@ -7007,6 +7116,19 @@ func (m *LabelRequest) Size() (n int) { return n } +func (m *RateLimitedStream) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = len(m.Labels) + if l > 0 { + n += 1 + l + sovLogproto(uint64(l)) + } + return n +} + func (m *LabelResponse) Size() (n int) { if m == nil { return 0 @@ -7752,6 +7874,16 @@ func (this *LabelRequest) String() string { }, "") return s } +func (this *RateLimitedStream) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&RateLimitedStream{`, + `Labels:` + fmt.Sprintf("%v", this.Labels) + `,`, + `}`, + }, "") + return s +} func (this *LabelResponse) String() string { if this == nil { return "nil" @@ -9369,6 +9501,91 @@ func (m *LabelRequest) Unmarshal(dAtA []byte) error { } return nil } +func (m *RateLimitedStream) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowLogproto + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: RateLimitedStream: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: RateLimitedStream: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Labels", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowLogproto + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthLogproto + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthLogproto + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Labels = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipLogproto(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthLogproto + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthLogproto + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} func (m *LabelResponse) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 diff --git a/pkg/logproto/logproto.proto b/pkg/logproto/logproto.proto index ad9631c1e09d9..ca11a736f928d 100644 --- a/pkg/logproto/logproto.proto +++ b/pkg/logproto/logproto.proto @@ -116,6 +116,10 @@ message LabelRequest { ]; } +message RateLimitedStream { + string labels = 1; +} + message LabelResponse { repeated string values = 1; }
loki
Modify ingesters to return rate-limited streams to distributors (#6977)
c89df1a795b8eaf6f3c765db85e86f1a20f16deb
2019-11-27 00:03:15
Putra Sattvika, I Gusti Ngurah
storage: fix missing logs with batched chunk iterator (#1299)
false
diff --git a/pkg/storage/iterator.go b/pkg/storage/iterator.go index 229a89d5b3055..5acf5ea25d8e8 100644 --- a/pkg/storage/iterator.go +++ b/pkg/storage/iterator.go @@ -77,6 +77,16 @@ type batchChunkIterator struct { // newBatchChunkIterator creates a new batch iterator with the given batchSize. func newBatchChunkIterator(ctx context.Context, chunks []*chunkenc.LazyChunk, batchSize int, matchers []*labels.Matcher, filter logql.Filter, req *logproto.QueryRequest) *batchChunkIterator { + + // __name__ is not something we filter by because it's a constant in loki and only used for upstream compatibility. + // Therefore remove it + for i := range matchers { + if matchers[i].Name == labels.MetricName { + matchers = append(matchers[:i], matchers[i+1:]...) + break + } + } + res := &batchChunkIterator{ batchSize: batchSize, matchers: matchers, @@ -112,6 +122,9 @@ func (it *batchChunkIterator) Next() bool { } func (it *batchChunkIterator) nextBatch() (iter.EntryIterator, error) { + // the first chunk of the batch + headChunk := it.chunks.Peek() + // pop the next batch of chunks and append/preprend previous overlapping chunks // so we can merge/de-dupe overlapping entries. batch := make([]*chunkenc.LazyChunk, 0, it.batchSize+len(it.lastOverlapping)) @@ -130,6 +143,14 @@ func (it *batchChunkIterator) nextBatch() (iter.EntryIterator, error) { // so that overlapping chunks are together if it.req.Direction == logproto.BACKWARD { from = time.Unix(0, nextChunk.Chunk.Through.UnixNano()) + + // we have to reverse the inclusivity of the chunk iterator from + // [from, through) to (from, through] for backward queries, except when + // the batch's `from` is equal to the query's Start. This can be achieved + // by shifting `from` by one nanosecond. + if !from.Equal(it.req.Start) { + from = from.Add(time.Nanosecond) + } } else { through = time.Unix(0, nextChunk.Chunk.From.UnixNano()) } @@ -149,7 +170,7 @@ func (it *batchChunkIterator) nextBatch() (iter.EntryIterator, error) { // └────────────────────┘ // // And nextChunk is # 49, we need to keep references to #47 and #48 as they won't be - // iterated over completely (we're clipping through to #49's from) and then add them to the next batch. + // iterated over completely (we're clipping through to #49's from) and then add them to the next batch. it.lastOverlapping = it.lastOverlapping[:0] for _, c := range batch { if it.req.Direction == logproto.BACKWARD { @@ -162,13 +183,27 @@ func (it *batchChunkIterator) nextBatch() (iter.EntryIterator, error) { } } } + } + + if it.req.Direction == logproto.BACKWARD { + through = time.Unix(0, headChunk.Chunk.Through.UnixNano()) + + if through.After(it.req.End) { + through = it.req.End + } + + // we have to reverse the inclusivity of the chunk iterator from + // [from, through) to (from, through] for backward queries, except when + // the batch's `through` is equal to the query's End. This can be achieved + // by shifting `through` by one nanosecond. + if !through.Equal(it.req.End) { + through = through.Add(time.Nanosecond) + } } else { - if len(it.lastOverlapping) > 0 { - if it.req.Direction == logproto.BACKWARD { - through = time.Unix(0, it.lastOverlapping[0].Chunk.From.UnixNano()) - } else { - from = time.Unix(0, it.lastOverlapping[0].Chunk.Through.UnixNano()) - } + from = time.Unix(0, headChunk.Chunk.From.UnixNano()) + + if from.Before(it.req.Start) { + from = it.req.Start } } @@ -250,12 +285,9 @@ func buildIterators(ctx context.Context, chks map[model.Fingerprint][][]*chunken func buildHeapIterator(ctx context.Context, chks [][]*chunkenc.LazyChunk, filter logql.Filter, direction logproto.Direction, from, through time.Time) (iter.EntryIterator, error) { result := make([]iter.EntryIterator, 0, len(chks)) - if chks[0][0].Chunk.Metric.Has("__name__") { - labelsBuilder := labels.NewBuilder(chks[0][0].Chunk.Metric) - labelsBuilder.Del("__name__") - chks[0][0].Chunk.Metric = labelsBuilder.Labels() - } - labels := chks[0][0].Chunk.Metric.String() + + // __name__ is only used for upstream compatibility and is hardcoded within loki. Strip it from the return label set. + labels := dropLabels(chks[0][0].Chunk.Metric, labels.MetricName).String() for i := range chks { iterators := make([]iter.EntryIterator, 0, len(chks[i])) @@ -400,3 +432,20 @@ outer: return css } + +// dropLabels returns a new label set with certain labels dropped +func dropLabels(ls labels.Labels, removals ...string) (dst labels.Labels) { + toDel := make(map[string]struct{}) + for _, r := range removals { + toDel[r] = struct{}{} + } + + for _, l := range ls { + _, remove := toDel[l.Name] + if !remove { + dst = append(dst, l) + } + } + + return dst +} diff --git a/pkg/storage/iterator_test.go b/pkg/storage/iterator_test.go index 8f04348148dcc..fac34f5fe77aa 100644 --- a/pkg/storage/iterator_test.go +++ b/pkg/storage/iterator_test.go @@ -2,12 +2,15 @@ package storage import ( "context" + "fmt" "testing" "time" "github.com/grafana/loki/pkg/chunkenc" "github.com/grafana/loki/pkg/iter" "github.com/grafana/loki/pkg/logproto" + "github.com/prometheus/prometheus/pkg/labels" + "github.com/stretchr/testify/require" ) func Test_newBatchChunkIterator(t *testing.T) { @@ -18,11 +21,12 @@ func Test_newBatchChunkIterator(t *testing.T) { matchers string start, end time.Time direction logproto.Direction + batchSize int }{ "forward with overlap": { []*chunkenc.LazyChunk{ newLazyChunk(logproto.Stream{ - Labels: fooLabels, + Labels: fooLabelsWithName, Entries: []logproto.Entry{ { Timestamp: from, @@ -35,7 +39,7 @@ func Test_newBatchChunkIterator(t *testing.T) { }, }), newLazyChunk(logproto.Stream{ - Labels: fooLabels, + Labels: fooLabelsWithName, Entries: []logproto.Entry{ { Timestamp: from.Add(time.Millisecond), @@ -48,7 +52,7 @@ func Test_newBatchChunkIterator(t *testing.T) { }, }), newLazyChunk(logproto.Stream{ - Labels: fooLabels, + Labels: fooLabelsWithName, Entries: []logproto.Entry{ { Timestamp: from.Add(time.Millisecond), @@ -61,7 +65,120 @@ func Test_newBatchChunkIterator(t *testing.T) { }, }), newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, + Entries: []logproto.Entry{ + { + Timestamp: from.Add(2 * time.Millisecond), + Line: "3", + }, + { + Timestamp: from.Add(3 * time.Millisecond), + Line: "4", + }, + }, + }), + newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, + Entries: []logproto.Entry{ + { + Timestamp: from.Add(2 * time.Millisecond), + Line: "3", + }, + { + Timestamp: from.Add(3 * time.Millisecond), + Line: "4", + }, + }, + }), + newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, + Entries: []logproto.Entry{ + { + Timestamp: from.Add(3 * time.Millisecond), + Line: "4", + }, + { + Timestamp: from.Add(4 * time.Millisecond), + Line: "5", + }, + }, + }), + }, + []*logproto.Stream{ + { Labels: fooLabels, + Entries: []logproto.Entry{ + { + Timestamp: from, + Line: "1", + }, + { + Timestamp: from.Add(time.Millisecond), + Line: "2", + }, + { + Timestamp: from.Add(2 * time.Millisecond), + Line: "3", + }, + { + Timestamp: from.Add(3 * time.Millisecond), + Line: "4", + }, + }, + }, + }, + fooLabelsWithName, + from, from.Add(4 * time.Millisecond), + logproto.FORWARD, + 2, + }, + "forward with overlapping non-continuous entries": { + []*chunkenc.LazyChunk{ + newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, + Entries: []logproto.Entry{ + { + Timestamp: from, + Line: "1", + }, + { + Timestamp: from.Add(time.Millisecond), + Line: "2", + }, + { + Timestamp: from.Add(3 * time.Millisecond), + Line: "4", + }, + }, + }), + newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, + Entries: []logproto.Entry{ + { + Timestamp: from.Add(time.Millisecond), + Line: "2", + }, + { + Timestamp: from.Add(2 * time.Millisecond), + Line: "3", + }, + }, + }), + newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, + Entries: []logproto.Entry{ + { + Timestamp: from.Add(time.Millisecond), + Line: "2", + }, + { + Timestamp: from.Add(3 * time.Millisecond), + Line: "4", + }, + }, + }), + newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, Entries: []logproto.Entry{ { Timestamp: from.Add(2 * time.Millisecond), @@ -93,14 +210,15 @@ func Test_newBatchChunkIterator(t *testing.T) { }, }, }, - fooLabels, + fooLabelsWithName, from, from.Add(3 * time.Millisecond), logproto.FORWARD, + 2, }, "backward with overlap": { []*chunkenc.LazyChunk{ newLazyChunk(logproto.Stream{ - Labels: fooLabels, + Labels: fooLabelsWithName, Entries: []logproto.Entry{ { Timestamp: from, @@ -113,7 +231,7 @@ func Test_newBatchChunkIterator(t *testing.T) { }, }), newLazyChunk(logproto.Stream{ - Labels: fooLabels, + Labels: fooLabelsWithName, Entries: []logproto.Entry{ { Timestamp: from.Add(time.Millisecond), @@ -126,7 +244,7 @@ func Test_newBatchChunkIterator(t *testing.T) { }, }), newLazyChunk(logproto.Stream{ - Labels: fooLabels, + Labels: fooLabelsWithName, Entries: []logproto.Entry{ { Timestamp: from.Add(time.Millisecond), @@ -139,7 +257,7 @@ func Test_newBatchChunkIterator(t *testing.T) { }, }), newLazyChunk(logproto.Stream{ - Labels: fooLabels, + Labels: fooLabelsWithName, Entries: []logproto.Entry{ { Timestamp: from.Add(2 * time.Millisecond), @@ -151,11 +269,41 @@ func Test_newBatchChunkIterator(t *testing.T) { }, }, }), + newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, + Entries: []logproto.Entry{ + { + Timestamp: from.Add(2 * time.Millisecond), + Line: "3", + }, + { + Timestamp: from.Add(3 * time.Millisecond), + Line: "4", + }, + }, + }), + newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, + Entries: []logproto.Entry{ + { + Timestamp: from.Add(3 * time.Millisecond), + Line: "4", + }, + { + Timestamp: from.Add(4 * time.Millisecond), + Line: "5", + }, + }, + }), }, []*logproto.Stream{ { Labels: fooLabels, Entries: []logproto.Entry{ + { + Timestamp: from.Add(3 * time.Millisecond), + Line: "4", + }, { Timestamp: from.Add(2 * time.Millisecond), Line: "3", @@ -171,14 +319,114 @@ func Test_newBatchChunkIterator(t *testing.T) { }, }, }, - fooLabels, - from, from.Add(3 * time.Millisecond), + fooLabelsWithName, + from, from.Add(4 * time.Millisecond), logproto.BACKWARD, + 2, }, - "forward without overlap": { + "backward with overlapping non-continuous entries": { []*chunkenc.LazyChunk{ newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, + Entries: []logproto.Entry{ + { + Timestamp: from.Add(0 * time.Millisecond), + Line: "0", + }, + { + Timestamp: from.Add(3 * time.Millisecond), + Line: "3", + }, + }, + }), + newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, + Entries: []logproto.Entry{ + { + Timestamp: from.Add(1 * time.Millisecond), + Line: "1", + }, + { + Timestamp: from.Add(6 * time.Millisecond), + Line: "6", + }, + }, + }), + newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, + Entries: []logproto.Entry{ + { + Timestamp: from.Add(2 * time.Millisecond), + Line: "2", + }, + { + Timestamp: from.Add(5 * time.Millisecond), + Line: "5", + }, + }, + }), + newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, + Entries: []logproto.Entry{ + { + Timestamp: from.Add(4 * time.Millisecond), + Line: "4", + }, + { + Timestamp: from.Add(7 * time.Millisecond), + Line: "7", + }, + }, + }), + }, + []*logproto.Stream{ + { Labels: fooLabels, + Entries: []logproto.Entry{ + { + Timestamp: from.Add(7 * time.Millisecond), + Line: "7", + }, + { + Timestamp: from.Add(6 * time.Millisecond), + Line: "6", + }, + { + Timestamp: from.Add(5 * time.Millisecond), + Line: "5", + }, + { + Timestamp: from.Add(4 * time.Millisecond), + Line: "4", + }, + { + Timestamp: from.Add(3 * time.Millisecond), + Line: "3", + }, + { + Timestamp: from.Add(2 * time.Millisecond), + Line: "2", + }, + { + Timestamp: from.Add(1 * time.Millisecond), + Line: "1", + }, + { + Timestamp: from.Add(0 * time.Millisecond), + Line: "0", + }, + }, + }, + }, + fooLabelsWithName, + from, from.Add(8 * time.Millisecond), + logproto.BACKWARD, + 2, + }, + "forward without overlap": { + []*chunkenc.LazyChunk{ + newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, Entries: []logproto.Entry{ { Timestamp: from, @@ -191,7 +439,7 @@ func Test_newBatchChunkIterator(t *testing.T) { }, }), newLazyChunk(logproto.Stream{ - Labels: fooLabels, + Labels: fooLabelsWithName, Entries: []logproto.Entry{ { Timestamp: from.Add(2 * time.Millisecond), @@ -200,7 +448,7 @@ func Test_newBatchChunkIterator(t *testing.T) { }, }), newLazyChunk(logproto.Stream{ - Labels: fooLabels, + Labels: fooLabelsWithName, Entries: []logproto.Entry{ { Timestamp: from.Add(3 * time.Millisecond), @@ -228,14 +476,15 @@ func Test_newBatchChunkIterator(t *testing.T) { }, }, }, - fooLabels, + fooLabelsWithName, from, from.Add(3 * time.Millisecond), logproto.FORWARD, + 2, }, "backward without overlap": { []*chunkenc.LazyChunk{ newLazyChunk(logproto.Stream{ - Labels: fooLabels, + Labels: fooLabelsWithName, Entries: []logproto.Entry{ { Timestamp: from, @@ -248,7 +497,7 @@ func Test_newBatchChunkIterator(t *testing.T) { }, }), newLazyChunk(logproto.Stream{ - Labels: fooLabels, + Labels: fooLabelsWithName, Entries: []logproto.Entry{ { Timestamp: from.Add(2 * time.Millisecond), @@ -257,7 +506,7 @@ func Test_newBatchChunkIterator(t *testing.T) { }, }), newLazyChunk(logproto.Stream{ - Labels: fooLabels, + Labels: fooLabelsWithName, Entries: []logproto.Entry{ { Timestamp: from.Add(3 * time.Millisecond), @@ -285,16 +534,17 @@ func Test_newBatchChunkIterator(t *testing.T) { }, }, }, - fooLabels, + fooLabelsWithName, from, from.Add(3 * time.Millisecond), logproto.BACKWARD, + 2, }, } for name, tt := range tests { tt := tt t.Run(name, func(t *testing.T) { - it := newBatchChunkIterator(context.Background(), tt.chunks, 2, newMatchers(tt.matchers), nil, newQuery("", tt.start, tt.end, tt.direction)) + it := newBatchChunkIterator(context.Background(), tt.chunks, tt.batchSize, newMatchers(tt.matchers), nil, newQuery("", tt.start, tt.end, tt.direction)) streams, _, err := iter.ReadBatch(it, 1000) _ = it.Close() if err != nil { @@ -305,5 +555,127 @@ func Test_newBatchChunkIterator(t *testing.T) { }) } +} +func TestPartitionOverlappingchunks(t *testing.T) { + var ( + oneThroughFour = newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, + Entries: []logproto.Entry{ + { + Timestamp: from, + Line: "1", + }, + { + Timestamp: from.Add(3 * time.Millisecond), + Line: "4", + }, + }, + }) + two = newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, + Entries: []logproto.Entry{ + { + Timestamp: from.Add(1 * time.Millisecond), + Line: "2", + }, + }, + }) + three = newLazyChunk(logproto.Stream{ + Labels: fooLabelsWithName, + Entries: []logproto.Entry{ + { + Timestamp: from.Add(2 * time.Millisecond), + Line: "3", + }, + }, + }) + ) + + for i, tc := range []struct { + input []*chunkenc.LazyChunk + expected [][]*chunkenc.LazyChunk + }{ + { + input: []*chunkenc.LazyChunk{ + oneThroughFour, + two, + three, + }, + expected: [][]*chunkenc.LazyChunk{ + []*chunkenc.LazyChunk{oneThroughFour}, + []*chunkenc.LazyChunk{two, three}, + }, + }, + { + input: []*chunkenc.LazyChunk{ + two, + oneThroughFour, + three, + }, + expected: [][]*chunkenc.LazyChunk{ + []*chunkenc.LazyChunk{oneThroughFour}, + []*chunkenc.LazyChunk{two, three}, + }, + }, + { + input: []*chunkenc.LazyChunk{ + two, + two, + three, + three, + }, + expected: [][]*chunkenc.LazyChunk{ + []*chunkenc.LazyChunk{two, three}, + []*chunkenc.LazyChunk{two, three}, + }, + }, + } { + t.Run(fmt.Sprintf("%d", i), func(t *testing.T) { + out := partitionOverlappingChunks(tc.input) + require.Equal(t, tc.expected, out) + }) + } +} + +func TestDropLabels(t *testing.T) { + + for i, tc := range []struct { + ls labels.Labels + drop []string + expected labels.Labels + }{ + { + ls: labels.Labels{ + labels.Label{ + Name: "a", + Value: "1", + }, + labels.Label{ + Name: "b", + Value: "2", + }, + labels.Label{ + Name: "c", + Value: "3", + }, + }, + drop: []string{"b"}, + expected: labels.Labels{ + labels.Label{ + Name: "a", + Value: "1", + }, + labels.Label{ + Name: "c", + Value: "3", + }, + }, + }, + } { + t.Run(fmt.Sprintf("%d", i), func(t *testing.T) { + dropped := dropLabels(tc.ls, tc.drop...) + require.Equal(t, tc.expected, dropped) + }) + } } diff --git a/pkg/storage/util_test.go b/pkg/storage/util_test.go index 9107438b8f145..57d9f974488de 100644 --- a/pkg/storage/util_test.go +++ b/pkg/storage/util_test.go @@ -19,6 +19,7 @@ import ( "github.com/stretchr/testify/assert" ) +var fooLabelsWithName = "{foo=\"bar\", __name__=\"log\"}" var fooLabels = "{foo=\"bar\"}" var from = time.Unix(0, time.Millisecond.Nanoseconds())
storage
fix missing logs with batched chunk iterator (#1299)
2edbe1280f3c32e1d02794343d666edb67296811
2024-03-21 21:58:15
Christian Haudum
fix(blooms): Correctly skip block page in case it exceeds the max page size for querying (#12297)
false
diff --git a/pkg/storage/bloom/v1/block.go b/pkg/storage/bloom/v1/block.go index 84bc71a6b203c..c9eef5fa33027 100644 --- a/pkg/storage/bloom/v1/block.go +++ b/pkg/storage/bloom/v1/block.go @@ -148,25 +148,26 @@ func (bq *BlockQuerier) Seek(fp model.Fingerprint) error { } func (bq *BlockQuerier) Next() bool { - if !bq.series.Next() { - return false - } - - series := bq.series.At() - - bq.blooms.Seek(series.Offset) - if !bq.blooms.Next() { - return false - } - - bloom := bq.blooms.At() - - bq.cur = &SeriesWithBloom{ - Series: &series.Series, - Bloom: bloom, + for bq.series.Next() { + series := bq.series.At() + bq.blooms.Seek(series.Offset) + if !bq.blooms.Next() { + // skip blocks that are too large + if errors.Is(bq.blooms.Err(), ErrPageTooLarge) { + // fmt.Printf("skipping bloom page: %s (%d)\n", series.Fingerprint, series.Chunks.Len()) + bq.blooms.err = nil + continue + } + return false + } + bloom := bq.blooms.At() + bq.cur = &SeriesWithBloom{ + Series: &series.Series, + Bloom: bloom, + } + return true } - return true - + return false } func (bq *BlockQuerier) At() *SeriesWithBloom { diff --git a/pkg/storage/bloom/v1/bloom.go b/pkg/storage/bloom/v1/bloom.go index 6a6c2610e82e2..da0a770fb2579 100644 --- a/pkg/storage/bloom/v1/bloom.go +++ b/pkg/storage/bloom/v1/bloom.go @@ -18,7 +18,7 @@ import ( // Figure out a decent maximum page size that we can process. // TODO(chaudum): Make max page size configurable var maxPageSize = 32 << 20 // 32MB -var errPageTooLarge = "bloom page too large to process: N=%d Offset=%d Len=%d DecompressedLen=%d" +var ErrPageTooLarge = errors.Errorf("bloom page too large: size limit is %.1fMiB", float64(maxPageSize)/float64(1<<20)) type Bloom struct { filter.ScalableBloomFilter @@ -253,9 +253,10 @@ func (b *BloomBlock) BloomPageDecoder(r io.ReadSeeker, pageIdx int) (*BloomPageD } page := b.pageHeaders[pageIdx] + // fmt.Printf("pageIdx=%d page=%+v size=%.2fMiB\n", pageIdx, page, float64(page.Len)/float64(1<<20)) if page.Len > maxPageSize { - return nil, fmt.Errorf(errPageTooLarge, page.N, page.Offset, page.Len, page.DecompressedLen) + return nil, ErrPageTooLarge } if _, err := r.Seek(int64(page.Offset), io.SeekStart); err != nil { diff --git a/tools/bloom/inspector/main.go b/tools/bloom/inspector/main.go new file mode 100644 index 0000000000000..bb81d02b260b1 --- /dev/null +++ b/tools/bloom/inspector/main.go @@ -0,0 +1,37 @@ +package main + +import ( + "fmt" + "os" + + v1 "github.com/grafana/loki/pkg/storage/bloom/v1" +) + +func main() { + if len(os.Args) < 2 { + fmt.Println("Usage: go run main.go BLOCK_DIRECTORY") + os.Exit(2) + } + + path := os.Args[1] + fmt.Printf("Block directory: %s\n", path) + + r := v1.NewDirectoryBlockReader(path) + b := v1.NewBlock(r) + q := v1.NewBlockQuerier(b) + + md, err := q.Metadata() + if err != nil { + panic(err) + } + + fmt.Printf("Metadata: %+v\n", md) + + for q.Next() { + swb := q.At() + fmt.Printf("%s (%d)\n", swb.Series.Fingerprint, swb.Series.Chunks.Len()) + } + if q.Err() != nil { + fmt.Printf("error: %s\n", q.Err()) + } +}
fix
Correctly skip block page in case it exceeds the max page size for querying (#12297)
f0ec743ba09d38c4cdb01e88fb8b0a1198b3c25f
2023-12-13 16:21:36
Zirko
helm: add cilium networkpolicies (#11425)
false
diff --git a/docs/sources/setup/install/helm/reference.md b/docs/sources/setup/install/helm/reference.md index 8252a6fd103a3..e650d0fca2fc2 100644 --- a/docs/sources/setup/install/helm/reference.md +++ b/docs/sources/setup/install/helm/reference.md @@ -3110,6 +3110,15 @@ false <td><pre lang="json"> [] </pre> +</td> + </tr> + <tr> + <td>networkPolicy.flavor</td> + <td>string</td> + <td>Specifies whether the policies created will be standard Network Policies (flavor: kubernetes) or Cilium Network Policies (flavor: cilium)</td> + <td><pre lang="json"> +"kubernetes" +</pre> </td> </tr> <tr> diff --git a/production/helm/loki/CHANGELOG.md b/production/helm/loki/CHANGELOG.md index 78571b3de600d..96bebdf5aebc9 100644 --- a/production/helm/loki/CHANGELOG.md +++ b/production/helm/loki/CHANGELOG.md @@ -13,6 +13,10 @@ Entries should include a reference to the pull request that introduced the chang [//]: # (<AUTOMATED_UPDATES_LOCATOR> : do not remove this line. This locator is used by the CI pipeline to automatically create a changelog entry for each new Loki release. Add other chart versions and respective changelog entries bellow this line.) +## 5.41.2 + +- [FEATURE] Add ciliumnetworkpolicies. + ## 5.41.1 - [FEATURE] Allow topology spread constraints for Loki read deployment component. diff --git a/production/helm/loki/Chart.yaml b/production/helm/loki/Chart.yaml index d9cf011e4f23e..fc7e0fbacbc6e 100644 --- a/production/helm/loki/Chart.yaml +++ b/production/helm/loki/Chart.yaml @@ -3,7 +3,7 @@ name: loki description: Helm chart for Grafana Loki in simple, scalable mode type: application appVersion: 2.9.3 -version: 5.41.1 +version: 5.41.2 home: https://grafana.github.io/helm-charts sources: - https://github.com/grafana/loki diff --git a/production/helm/loki/README.md b/production/helm/loki/README.md index 3caad398ada44..e1da365b5bf92 100644 --- a/production/helm/loki/README.md +++ b/production/helm/loki/README.md @@ -1,6 +1,6 @@ # loki -![Version: 5.41.1](https://img.shields.io/badge/Version-5.41.1-informational?style=flat-square) ![Type: application](https://img.shields.io/badge/Type-application-informational?style=flat-square) ![AppVersion: 2.9.3](https://img.shields.io/badge/AppVersion-2.9.3-informational?style=flat-square) +![Version: 5.41.2](https://img.shields.io/badge/Version-5.41.2-informational?style=flat-square) ![Type: application](https://img.shields.io/badge/Type-application-informational?style=flat-square) ![AppVersion: 2.9.3](https://img.shields.io/badge/AppVersion-2.9.3-informational?style=flat-square) Helm chart for Grafana Loki in simple, scalable mode diff --git a/production/helm/loki/templates/ciliumnetworkpolicy.yaml b/production/helm/loki/templates/ciliumnetworkpolicy.yaml new file mode 100644 index 0000000000000..5633ae1945206 --- /dev/null +++ b/production/helm/loki/templates/ciliumnetworkpolicy.yaml @@ -0,0 +1,184 @@ +{{- if and (.Values.networkPolicy.enabled) (eq .Values.networkPolicy.flavor "cilium") }} +--- +apiVersion: cilium.io/v2 +kind: CiliumNetworkPolicy +metadata: + name: {{ include "loki.name" . }}-namespace-only + namespace: {{ $.Release.Namespace }} + labels: + {{- include "loki.labels" . | nindent 4 }} +spec: + endpointSelector: {} + egress: + - toEndpoints: + - {} + ingress: + - fromEndpoints: + - {} + +--- +apiVersion: cilium.io/v2 +kind: CiliumNetworkPolicy +metadata: + name: {{ include "loki.name" . }}-egress-dns + namespace: {{ $.Release.Namespace }} + labels: + {{- include "loki.labels" . | nindent 4 }} +spec: + endpointSelector: + matchLabels: + {{- include "loki.selectorLabels" . | nindent 6 }} + egress: + - toPorts: + - ports: + - port: dns + protocol: UDP + toEndpoints: + - namespaceSelector: {} + +--- +apiVersion: cilium.io/v2 +kind: CiliumNetworkPolicy +metadata: + name: {{ include "loki.name" . }}-ingress + namespace: {{ $.Release.Namespace }} + labels: + {{- include "loki.labels" . | nindent 4 }} +spec: + endpointSelector: + matchExpressions: + - key: app.kubernetes.io/component + operator: In + values: + {{- if .Values.gateway.enabled }} + - gateway + {{- else }} + - read + - write + {{- end }} + matchLabels: + {{- include "loki.selectorLabels" . | nindent 6 }} + ingress: + - toPorts: + - port: http + protocol: TCP + {{- if .Values.networkPolicy.ingress.namespaceSelector }} + fromEndpoints: + - matchLabels: + {{- toYaml .Values.networkPolicy.ingress.namespaceSelector | nindent 8 }} + {{- if .Values.networkPolicy.ingress.podSelector }} + {{- toYaml .Values.networkPolicy.ingress.podSelector | nindent 8 }} + {{- end }} + {{- end }} + +--- +apiVersion: cilium.io/v2 +kind: CiliumNetworkPolicy +metadata: + name: {{ include "loki.name" . }}-ingress-metrics + namespace: {{ $.Release.Namespace }} + labels: + {{- include "loki.labels" . | nindent 4 }} +spec: + endpointSelector: + matchLabels: + {{- include "loki.selectorLabels" . | nindent 6 }} + ingress: + - toPorts: + - port: http-metrics + protocol: TCP + {{- if .Values.networkPolicy.metrics.cidrs }} + {{- range $cidr := .Values.networkPolicy.metrics.cidrs }} + toCIDR: + - {{ $cidr }} + {{- end }} + {{- if .Values.networkPolicy.metrics.namespaceSelector }} + fromEndpoints: + - matchLabels: + {{- toYaml .Values.networkPolicy.metrics.namespaceSelector | nindent 8 }} + {{- if .Values.networkPolicy.metrics.podSelector }} + {{- toYaml .Values.networkPolicy.metrics.podSelector | nindent 8 }} + {{- end }} + {{- end }} + {{- end }} + +--- +apiVersion: cilium.io/v2 +kind: CiliumNetworkPolicy +metadata: + name: {{ include "loki.name" . }}-egress-alertmanager + namespace: {{ $.Release.Namespace }} + labels: + {{- include "loki.labels" . | nindent 4 }} +spec: + endpointSelector: + matchLabels: + {{- include "loki.backendSelectorLabels" . | nindent 6 }} + egress: + - toPorts: + - port: {{ .Values.networkPolicy.alertmanager.port }} + protocol: TCP + {{- if .Values.networkPolicy.alertmanager.namespaceSelector }} + toEndpoints: + - matchLabels: + {{- toYaml .Values.networkPolicy.alertmanager.namespaceSelector | nindent 8 }} + {{- if .Values.networkPolicy.alertmanager.podSelector }} + {{- toYaml .Values.networkPolicy.alertmanager.podSelector | nindent 8 }} + {{- end }} + {{- end }} + +{{- if .Values.networkPolicy.externalStorage.ports }} +--- +apiVersion: cilium.io/v2 +kind: CiliumNetworkPolicy +metadata: + name: {{ include "loki.name" . }}-egress-external-storage + namespace: {{ $.Release.Namespace }} + labels: + {{- include "loki.labels" . | nindent 4 }} +spec: + endpointSelector: + matchLabels: + {{- include "loki.selectorLabels" . | nindent 6 }} + egress: + - toPorts: + {{- range $port := .Values.networkPolicy.externalStorage.ports }} + - port: {{ $port }} + protocol: TCP + {{- end }} + {{- if .Values.networkPolicy.externalStorage.cidrs }} + {{- range $cidr := .Values.networkPolicy.externalStorage.cidrs }} + toCIDR: + - {{ $cidr }} + {{- end }} + {{- end }} +{{- end }} + +{{- end }} + +{{- if .Values.networkPolicy.discovery.port }} +--- +apiVersion: cilium.io/v2 +kind: CiliumNetworkPolicy +metadata: + name: {{ include "loki.name" . }}-egress-discovery + namespace: {{ $.Release.Namespace }} + labels: + {{- include "loki.labels" . | nindent 4 }} +spec: + endpointSelector: + matchLabels: + {{- include "loki.selectorLabels" . | nindent 6 }} + egress: + - toPorts: + - port: {{ .Values.networkPolicy.discovery.port }} + protocol: TCP + {{- if .Values.networkPolicy.discovery.namespaceSelector }} + toEndpoints: + - matchLabels: + {{- toYaml .Values.networkPolicy.discovery.namespaceSelector | nindent 8 }} + {{- if .Values.networkPolicy.discovery.podSelector }} + {{- toYaml .Values.networkPolicy.discovery.podSelector | nindent 8 }} + {{- end }} + {{- end }} +{{- end }} diff --git a/production/helm/loki/templates/networkpolicy.yaml b/production/helm/loki/templates/networkpolicy.yaml index 4424d90db08d4..27c85280eb08c 100644 --- a/production/helm/loki/templates/networkpolicy.yaml +++ b/production/helm/loki/templates/networkpolicy.yaml @@ -1,4 +1,4 @@ -{{- if .Values.networkPolicy.enabled }} +{{- if and (.Values.networkPolicy.enabled) (eq .Values.networkPolicy.flavor "kubernetes") }} --- apiVersion: networking.k8s.io/v1 kind: NetworkPolicy diff --git a/production/helm/loki/values.yaml b/production/helm/loki/values.yaml index 738cf6ea25ae7..e82967a4efb3b 100644 --- a/production/helm/loki/values.yaml +++ b/production/helm/loki/values.yaml @@ -1465,6 +1465,9 @@ gateway: networkPolicy: # -- Specifies whether Network Policies should be created enabled: false + # -- Specifies whether the policies created will be standard Network Policies (flavor: kubernetes) + # or Cilium Network Policies (flavor: cilium) + flavor: kubernetes metrics: # -- Specifies the Pods which are allowed to access the metrics port. # As this is cross-namespace communication, you also need the namespaceSelector.
helm
add cilium networkpolicies (#11425)
86f6c2eb8d549438df6628be4d1efacae5444831
2022-09-01 23:08:04
Gerard Vanloo
operator: Adding Lokistack Gateway Request Errors Alert (#6999)
false
diff --git a/operator/CHANGELOG.md b/operator/CHANGELOG.md index dd721d52f0c1a..d36dd519d4071 100644 --- a/operator/CHANGELOG.md +++ b/operator/CHANGELOG.md @@ -1,5 +1,6 @@ ## Main +- [6999](https://github.com/grafana/loki/pull/6999) **Red-GV**: Adding LokiStack Gateway alerts - [7000](https://github.com/grafana/loki/pull/7000) **xperimental**: Configure default node affinity for all pods - [6923](https://github.com/grafana/loki/pull/6923) **xperimental**: Reconcile owner reference for existing objects - [6907](https://github.com/grafana/loki/pull/6907) **Red-GV**: Adding valid subscription annotation to operator metadata diff --git a/operator/docs/lokistack/sop.md b/operator/docs/lokistack/sop.md index 3e1ff566c0035..9f2cdaa8b2e7b 100644 --- a/operator/docs/lokistack/sop.md +++ b/operator/docs/lokistack/sop.md @@ -34,8 +34,8 @@ A service(s) is failing to process at least 10% of all incoming requests. - Console access to the cluster - Edit access to the deployed operator and Loki namespace: - OpenShift - - `openshift-logging` - - `openshift-operators-redhat` + - `openshift-logging` (LokiStack) + - `openshift-operators-redhat` (Loki Operator) ### Steps @@ -46,6 +46,69 @@ A service(s) is failing to process at least 10% of all incoming requests. - `loki_ingester_wal_disk_full_failures_total` - `loki_ingester_wal_corruptions_total` +## LokiStack Write Request Errors + +### Impact + +The LokiStack Gateway component is unable to perform its duties for a number of write requests, resulting in potential loss of data. + +### Summary + +The LokiStack Gateway is failing to process at least 10% of all incoming write requests. + +### Severity + +`Critical` + +### Access Required + +- Console access to the cluster +- Edit access to the deployed operator and Loki namespace: + - OpenShift + - `openshift-logging` (LokiStack) + - `openshift-operators-redhat` (Loki Operator) + +### Steps + +- Ensure that the LokiStack Gateway component is ready and available +- Ensure that the `distributor`, `ingester`, and `index-gateway` components are ready and available +- Ensure that store services (`ingester`, `querier`, `index-gateway`, `compactor`) can communicate with backend storage +- Examine metrics for signs of failure + - WAL Complications + - `loki_ingester_wal_disk_full_failures_total` + - `loki_ingester_wal_corruptions_total` + +## LokiStack Read Request Errors + +### Impact + +The LokiStack Gateway component is unable to perform its duties for a number of query requests, resulting in a potential disruption. + +### Summary + +The LokiStack Gateway is failing to process at least 10% of all incoming query requests. + +### Severity + +`Critical` + +### Access Required + +- Console access to the cluster +- Edit access to the deployed operator and Loki namespace: + - OpenShift + - `openshift-logging` (LokiStack) + - `openshift-operators-redhat` (Loki Operator) + +### Steps + +- Ensure that the LokiStack Gateway component is ready and available +- Ensure that the `query-frontend`, `querier`, `ingester`, and `index-gateway` components are ready and available +- Ensure that store services (`ingester`, `querier`, `index-gateway`, `compactor`) can communicate with backend storage +- Examine metrics for signs of failure + - WAL Complications + - `loki_ingester_wal_disk_full_failures_total` + - `loki_ingester_wal_corruptions_total` ## Loki Request Panics @@ -66,8 +129,8 @@ A service(s) has crashed. - Console access to the cluster - Edit access to the deployed operator and Loki namespace: - OpenShift - - `openshift-logging` - - `openshift-operators-redhat` + - `openshift-logging` (LokiStack) + - `openshift-operators-redhat` (Loki Operator) ### Steps diff --git a/operator/internal/manifests/internal/alerts/build.go b/operator/internal/manifests/internal/alerts/build.go index b1bf76558af16..4e860664f2266 100644 --- a/operator/internal/manifests/internal/alerts/build.go +++ b/operator/internal/manifests/internal/alerts/build.go @@ -13,7 +13,7 @@ import ( const ( // RunbookDefaultURL is the default url for the documentation of the Prometheus alerts - RunbookDefaultURL = "https://github.com/grafana/loki/tree/main/operator/docs/alerts.md" + RunbookDefaultURL = "https://github.com/grafana/loki/blob/main/operator/docs/lokistack/sop.md" ) var ( diff --git a/operator/internal/manifests/internal/alerts/prometheus-alerts.yaml b/operator/internal/manifests/internal/alerts/prometheus-alerts.yaml index 8635c347e9e3e..0e408200b16a0 100644 --- a/operator/internal/manifests/internal/alerts/prometheus-alerts.yaml +++ b/operator/internal/manifests/internal/alerts/prometheus-alerts.yaml @@ -10,21 +10,67 @@ groups: runbook_url: "[[ .RunbookURL ]]#Loki-Request-Errors" expr: | sum( - rate( - loki_request_duration_seconds_count{status_code=~"5.."}[1m] - ) + rate( + loki_request_duration_seconds_count{status_code=~"5.."}[1m] + ) ) by (namespace, job, route) / sum( - rate( - loki_request_duration_seconds_count[1m] - ) + rate( + loki_request_duration_seconds_count[1m] + ) ) by (namespace, job, route) * 100 > 10 for: 15m labels: severity: critical + - alert: LokiStackWriteRequestErrors + annotations: + message: |- + {{ printf "%.2f" $value }}% of write requests from {{ $labels.job }} are returned with server errors. + summary: "At least 10% of write requests to the lokistack-gateway are responded with 5xx server errors." + runbook_url: "[[ .RunbookURL ]]#LokiStack-Write-Request-Errors" + expr: | + sum( + rate( + http_requests_total{code=~"5..", group="logsv1", handler="push"}[1m] + ) + ) by (namespace, job, tenant) + / + sum( + rate( + http_requests_total{group="logsv1", handler="push"}[1m] + ) + ) by (namespace, job, tenant) + * 100 + > 10 + for: 15m + labels: + severity: critical + - alert: LokiStackReadRequestErrors + annotations: + message: |- + {{ printf "%.2f" $value }}% of query requests from {{ $labels.job }} are returned with server errors. + summary: "At least 10% of query requests to the lokistack-gateway are responded with 5xx server errors." + runbook_url: "[[ .RunbookURL ]]#LokiStack-Read-Request-Errors" + expr: | + sum( + rate( + http_requests_total{code=~"5..", group="logsv1", handler=~"query|query_range|label|labels|label_values"}[1m] + ) + ) by (namespace, job, tenant) + / + sum( + rate( + http_requests_total{group="logsv1", handler=~"query|query_range|label|labels|label_values"}[1m] + ) + ) by (namespace, job, tenant) + * 100 + > 10 + for: 15m + labels: + severity: critical - alert: LokiRequestPanics annotations: message: |- @@ -33,9 +79,9 @@ groups: runbook_url: "[[ .RunbookURL ]]#Loki-Request-Panics" expr: | sum( - increase( - loki_panic_total[10m] - ) + increase( + loki_panic_total[10m] + ) ) by (namespace, job) > 0 labels: diff --git a/operator/internal/manifests/internal/alerts/testdata/test.yaml b/operator/internal/manifests/internal/alerts/testdata/test.yaml index 295c97fe2c11f..6de3c7266fba8 100644 --- a/operator/internal/manifests/internal/alerts/testdata/test.yaml +++ b/operator/internal/manifests/internal/alerts/testdata/test.yaml @@ -11,13 +11,20 @@ tests: values: '1+1x20' - series: 'loki_request_duration_seconds_count{status_code="200", namespace="my-ns", job="ingester", route="my-route"}' values: '1+3x20' + - series: 'http_requests_total{code="500", namespace="my-ns", job="gateway", handler="push", group="logsv1"}' + values: '1+1x20' + - series: 'http_requests_total{code="200", namespace="my-ns", job="gateway", handler="push", group="logsv1"}' + values: '1+3x20' + - series: 'http_requests_total{code="500", namespace="my-ns", job="gateway", handler="query", group="logsv1"}' + values: '1+1x20' + - series: 'http_requests_total{code="200", namespace="my-ns", job="gateway", handler="query", group="logsv1"}' + values: '1+3x20' - series: 'loki_panic_total{namespace="my-ns", job="ingester"}' values: '0 1 1 2+0x10' # Unit test for alerting rules. alert_rule_test: - # --------- LokiRequestErrors --------- - eval_time: 16m alertname: LokiRequestErrors exp_alerts: @@ -30,8 +37,28 @@ tests: summary: "At least 10% of requests are responded by 5xx server errors." message: "ingester my-route is experiencing 25.00% errors." runbook_url: "[[ .RunbookURL ]]#Loki-Request-Errors" - - # --------- LokiRequestPanics --------- + - eval_time: 16m + alertname: LokiStackWriteRequestErrors + exp_alerts: + - exp_labels: + namespace: my-ns + job: gateway + severity: critical + exp_annotations: + summary: "At least 10% of write requests to the lokistack-gateway are responded with 5xx server errors." + message: "25.00% of write requests from gateway are returned with server errors." + runbook_url: "[[ .RunbookURL ]]#LokiStack-Write-Request-Errors" + - eval_time: 16m + alertname: LokiStackReadRequestErrors + exp_alerts: + - exp_labels: + namespace: my-ns + job: gateway + severity: critical + exp_annotations: + summary: "At least 10% of query requests to the lokistack-gateway are responded with 5xx server errors." + message: "25.00% of query requests from gateway are returned with server errors." + runbook_url: "[[ .RunbookURL ]]#LokiStack-Read-Request-Errors" - eval_time: 10m alertname: LokiRequestPanics exp_alerts:
operator
Adding Lokistack Gateway Request Errors Alert (#6999)
d0a285926b7257d54cf948ba644c619a4b49a871
2024-05-29 15:24:59
benclive
feat: Increase drain max depth from 8 -> 30 (#13063)
false
"diff --git a/pkg/pattern/drain/drain.go b/pkg/pattern/drain/drain.go\nindex 31932832f7010..c3076386(...TRUNCATED)
feat
Increase drain max depth from 8 -> 30 (#13063)
4e04d07168a8c5cb7086ced8486c6d584faa1045
2024-04-19 13:08:48
Paul Rogers
fix: promtail race fixes (#12656)
false
"diff --git a/clients/pkg/promtail/client/client_writeto_test.go b/clients/pkg/promtail/client/clien(...TRUNCATED)
fix
promtail race fixes (#12656)
d2e1992366bf4dd34e6401cc6734ce532b38e5b9
2024-11-22 23:32:20
George Robinson
chore: separate usage of partition owner gauge (#15079)
false
"diff --git a/pkg/kafka/partition/reader_service.go b/pkg/kafka/partition/reader_service.go\nindex d(...TRUNCATED)
chore
separate usage of partition owner gauge (#15079)
aeaefe6aab3f611684dcf8c05756947d02c6e1e5
2024-02-21 23:44:49
Robert Jacob
feat(operator): Extend Azure secret validation (#12007)
false
"diff --git a/operator/CHANGELOG.md b/operator/CHANGELOG.md\nindex 8dae6eced0bfa..e6aaec29b99c7 1006(...TRUNCATED)
feat
Extend Azure secret validation (#12007)
3a0be4897512ec16a0e43cd5e0a49774cda87aed
2022-11-23 22:20:56
Kaviraj Kanagaraj
config: Improve error message loading config with ENV variables. (#7759)
false
"diff --git a/CHANGELOG.md b/CHANGELOG.md\nindex ff4b1f64c72ef..ba79c01ed7a84 100644\n--- a/CHANGELO(...TRUNCATED)
config
Improve error message loading config with ENV variables. (#7759)
End of preview. Expand in Data Studio

No dataset card yet

Downloads last month
8