hash
stringlengths 40
40
| date
stringdate 2018-12-11 14:31:19
2025-03-22 02:45:31
| author
stringclasses 280
values | commit_message
stringlengths 14
176
| is_merge
bool 1
class | git_diff
stringlengths 198
25.8M
⌀ | type
stringclasses 83
values | masked_commit_message
stringlengths 8
170
|
|---|---|---|---|---|---|---|---|
bba6c14459bea3e5ef6907c30e9f4341f6365d7d
|
2022-08-31 02:33:19
|
Dylan Guedes
|
loki: Modify ingesters to return rate-limited streams to distributors (#6977)
| false
|
diff --git a/go.mod b/go.mod
index cd0948025d588..63da30a9cc8c9 100644
--- a/go.mod
+++ b/go.mod
@@ -113,6 +113,7 @@ require (
)
require (
+ github.com/gogo/googleapis v1.4.0
github.com/grafana/groupcache_exporter v0.0.0-20220629095919-59a8c6428a43
github.com/heroku/x v0.0.50
github.com/mailgun/groupcache/v2 v2.3.2
@@ -179,7 +180,6 @@ require (
github.com/go-stack/stack v1.8.1 // indirect
github.com/go-zookeeper/zk v1.0.2 // indirect
github.com/gofrs/flock v0.7.1 // indirect
- github.com/gogo/googleapis v1.4.0 // indirect
github.com/golang-jwt/jwt/v4 v4.2.0 // indirect
github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da // indirect
github.com/google/btree v1.0.1 // indirect
diff --git a/pkg/ingester/ingester_test.go b/pkg/ingester/ingester_test.go
index 6549891426ed8..92a16f35da661 100644
--- a/pkg/ingester/ingester_test.go
+++ b/pkg/ingester/ingester_test.go
@@ -10,6 +10,8 @@ import (
"testing"
"time"
+ "github.com/gogo/protobuf/types"
+ "github.com/gogo/status"
"github.com/grafana/dskit/flagext"
"github.com/grafana/dskit/services"
"github.com/prometheus/common/model"
@@ -19,6 +21,7 @@ import (
"github.com/weaveworks/common/middleware"
"github.com/weaveworks/common/user"
"golang.org/x/net/context"
+ "golang.org/x/time/rate"
"google.golang.org/grpc"
"google.golang.org/grpc/credentials/insecure"
"google.golang.org/grpc/metadata"
@@ -40,6 +43,68 @@ import (
"github.com/grafana/loki/pkg/validation"
)
+func TestRateLimitedStreamsReturn(t *testing.T) {
+ // setup.
+ ingesterConfig := defaultIngesterTestConfig(t)
+ limits := defaultLimitsTestConfig()
+ limits.PerStreamRateLimit.Set("1") //nolint:errcheck
+ limits.IngestionBurstSizeMB = 1
+ limits.PerStreamRateLimitBurst.Set("1") //nolint:errcheck
+ overrides, err := validation.NewOverrides(limits, nil)
+ limit := overrides.PerStreamRateLimit("test")
+ require.Equal(t, limit.Limit, rate.Limit(1))
+ require.NoError(t, err)
+
+ store := &mockStore{
+ chunks: map[string][]chunk.Chunk{},
+ }
+
+ i, err := New(ingesterConfig, client.Config{}, store, overrides, runtime.DefaultTenantConfigs(), nil)
+ require.NoError(t, err)
+ defer services.StopAndAwaitTerminated(context.Background(), i) //nolint:errcheck
+
+ req := logproto.PushRequest{
+ Streams: []logproto.Stream{
+ {
+ Labels: `{bar="baz1", foo="bar"}`,
+ },
+ {
+ Labels: `{bar="baz2", foo="bar"}`,
+ },
+ },
+ }
+ for i := 0; i < 10; i++ {
+ req.Streams[0].Entries = append(req.Streams[0].Entries, logproto.Entry{
+ Timestamp: time.Unix(0, 0),
+ Line: fmt.Sprintf("line %d", i),
+ })
+ req.Streams[1].Entries = append(req.Streams[1].Entries, logproto.Entry{
+ Timestamp: time.Unix(0, 0),
+ Line: fmt.Sprintf("line %d", i),
+ })
+ }
+
+ ctx := user.InjectOrgID(context.Background(), "test")
+ _, err = i.Push(ctx, &req)
+ require.Error(t, err)
+
+ s, ok := status.FromError(err)
+ require.True(t, ok)
+ details := s.Proto().Details
+
+ var rateLimitedLabels []string
+ for _, detail := range details {
+ rls := &logproto.RateLimitedStream{}
+ err := types.UnmarshalAny(detail, rls)
+ require.NoError(t, err)
+ rateLimitedLabels = append(rateLimitedLabels, rls.Labels)
+ }
+
+ // note that streams[0], although rate-limited, isn't present in the details.
+ // that's because push as of now is only returning the last errored stream.
+ require.EqualValues(t, []string{req.Streams[1].Labels}, rateLimitedLabels)
+}
+
func TestIngester(t *testing.T) {
ingesterConfig := defaultIngesterTestConfig(t)
limits, err := validation.NewOverrides(defaultLimitsTestConfig(), nil)
diff --git a/pkg/ingester/instance.go b/pkg/ingester/instance.go
index 1c8e9d8d0ae68..05bd3d6ec3911 100644
--- a/pkg/ingester/instance.go
+++ b/pkg/ingester/instance.go
@@ -1,7 +1,9 @@
package ingester
import (
+ "bytes"
"context"
+ fmt "fmt"
"net/http"
"os"
"sync"
@@ -9,6 +11,9 @@ import (
"time"
"github.com/go-kit/log/level"
+ spb "github.com/gogo/googleapis/google/rpc"
+ "github.com/gogo/protobuf/types"
+ "github.com/gogo/status"
"github.com/pkg/errors"
"github.com/prometheus/client_golang/prometheus"
"github.com/prometheus/client_golang/prometheus/promauto"
@@ -19,6 +24,7 @@ import (
"github.com/weaveworks/common/httpgrpc"
"go.uber.org/atomic"
+ "github.com/grafana/loki/pkg/chunkenc"
"github.com/grafana/loki/pkg/ingester/index"
"github.com/grafana/loki/pkg/iter"
"github.com/grafana/loki/pkg/logproto"
@@ -158,6 +164,11 @@ func (i *instance) consumeChunk(ctx context.Context, ls labels.Labels, chunk *lo
return err
}
+// Push will iterate over the given streams present in the PushRequest and attempt to store them.
+//
+// Although multiple streams are part of the PushRequest, the returned error only reflects what
+// happened to *the last stream in the request*. Ex: if three streams are part of the PushRequest
+// and all three failed, the returned error only describes what happened to the last processed stream.
func (i *instance) Push(ctx context.Context, req *logproto.PushRequest) error {
record := recordPool.GetRecord()
record.UserID = i.instanceID
@@ -185,10 +196,9 @@ func (i *instance) Push(ctx context.Context, req *logproto.PushRequest) error {
continue
}
- _, err = s.Push(ctx, reqStream.Entries, record, 0, false)
- if err != nil {
- appendErr = err
- }
+ _, failedEntriesWithError := s.Push(ctx, reqStream.Entries, record, 0, false)
+ appendErr = errorForFailedEntries(s, failedEntriesWithError, len(reqStream.Entries))
+
s.chunkMtx.Unlock()
}
@@ -211,6 +221,73 @@ func (i *instance) Push(ctx context.Context, req *logproto.PushRequest) error {
return appendErr
}
+// errorForFailedEntries mounts an error to be returned in a GRPC call based on entries that couldn't be ingested.
+//
+// The returned error is enriched with the gRPC error status and with a list of details.
+// As of now, the list of details is a list of all streams that were rate-limited. This list can be used
+// by the distributor to fine tune streams that were limited.
+func errorForFailedEntries(s *stream, failedEntriesWithError []entryWithError, totalEntries int) error {
+ if len(failedEntriesWithError) == 0 {
+ return nil
+ }
+
+ lastEntryWithErr := failedEntriesWithError[len(failedEntriesWithError)-1]
+ _, ok := lastEntryWithErr.e.(*validation.ErrStreamRateLimit)
+ outOfOrder := chunkenc.IsOutOfOrderErr(lastEntryWithErr.e)
+ if !outOfOrder && !ok {
+ return lastEntryWithErr.e
+ }
+
+ var statusCode int
+ if outOfOrder {
+ statusCode = http.StatusBadRequest
+ }
+ if ok {
+ // per-stream or ingestion limited.
+ statusCode = http.StatusTooManyRequests
+ }
+
+ // Return a http status 4xx request response with all failed entries.
+ buf := bytes.Buffer{}
+ streamName := s.labelsString
+
+ limitedFailedEntries := failedEntriesWithError
+ if maxIgnore := s.cfg.MaxReturnedErrors; maxIgnore > 0 && len(limitedFailedEntries) > maxIgnore {
+ limitedFailedEntries = limitedFailedEntries[:maxIgnore]
+ }
+
+ for _, entryWithError := range limitedFailedEntries {
+ fmt.Fprintf(&buf,
+ "entry with timestamp %s ignored, reason: '%s' for stream: %s,\n",
+ entryWithError.entry.Timestamp.String(), entryWithError.e.Error(), streamName)
+ }
+
+ fmt.Fprintf(&buf, "total ignored: %d out of %d", len(failedEntriesWithError), totalEntries)
+
+ var details []*types.Any
+
+ if statusCode == http.StatusTooManyRequests {
+ details = append(details, mountPerStreamDetails(streamName)...)
+ }
+
+ return status.ErrorProto(&spb.Status{
+ Code: int32(statusCode),
+ Message: buf.String(),
+ Details: details,
+ })
+}
+
+func mountPerStreamDetails(streamLabels string) []*types.Any {
+ rls := logproto.RateLimitedStream{Labels: streamLabels}
+ marshalledStream, err := types.MarshalAny(&rls)
+ if err == nil {
+ return []*types.Any{marshalledStream}
+ }
+
+ level.Error(util_log.Logger).Log("msg", "error marshalling rate-limited stream", "err", err, "labels", streamLabels)
+ return []*types.Any{}
+}
+
func (i *instance) createStream(pushReqStream logproto.Stream, record *WALRecord) (*stream, error) {
// record is only nil when replaying WAL. We don't want to drop data when replaying a WAL after
// reducing the stream limits, for instance.
diff --git a/pkg/ingester/recovery.go b/pkg/ingester/recovery.go
index c3bbc9a8af442..1a0c18786c8e4 100644
--- a/pkg/ingester/recovery.go
+++ b/pkg/ingester/recovery.go
@@ -165,10 +165,13 @@ func (r *ingesterRecoverer) Push(userID string, entries RefEntries) error {
}
// ignore out of order errors here (it's possible for a checkpoint to already have data from the wal segments)
- bytesAdded, err := s.(*stream).Push(context.Background(), entries.Entries, nil, entries.Counter, true)
+ bytesAdded, entriesWithErrors := s.(*stream).Push(context.Background(), entries.Entries, nil, entries.Counter, true)
r.ing.replayController.Add(int64(bytesAdded))
- if err != nil && err == ErrEntriesExist {
- r.ing.metrics.duplicateEntriesTotal.Add(float64(len(entries.Entries)))
+ if len(entriesWithErrors) > 0 {
+ lastEntryWithError := entriesWithErrors[len(entriesWithErrors)-1]
+ if errors.Is(lastEntryWithError.e, ErrEntriesExist) {
+ r.ing.metrics.duplicateEntriesTotal.Add(float64(len(entries.Entries)))
+ }
}
return nil
})
diff --git a/pkg/ingester/stream.go b/pkg/ingester/stream.go
index 23a9d3d674573..3d824bd592529 100644
--- a/pkg/ingester/stream.go
+++ b/pkg/ingester/stream.go
@@ -1,18 +1,15 @@
package ingester
import (
- "bytes"
"context"
- "fmt"
- "net/http"
"sync"
"time"
"github.com/go-kit/log/level"
+
"github.com/pkg/errors"
"github.com/prometheus/common/model"
"github.com/prometheus/prometheus/model/labels"
- "github.com/weaveworks/common/httpgrpc"
"github.com/grafana/loki/pkg/chunkenc"
"github.com/grafana/loki/pkg/iter"
@@ -148,7 +145,7 @@ func (s *stream) Push(
// Lock chunkMtx while pushing.
// If this is false, chunkMtx must be held outside Push.
lockChunk bool,
-) (int, error) {
+) (int, []entryWithError) {
if lockChunk {
s.chunkMtx.Lock()
defer s.chunkMtx.Unlock()
@@ -163,7 +160,7 @@ func (s *stream) Push(
s.metrics.walReplaySamplesDropped.WithLabelValues(duplicateReason).Add(float64(len(entries)))
s.metrics.walReplayBytesDropped.WithLabelValues(duplicateReason).Add(float64(byteCt))
- return 0, ErrEntriesExist
+ return 0, []entryWithError{{entry: &logproto.Entry{}, e: ErrEntriesExist}}
}
var bytesAdded int
@@ -300,41 +297,8 @@ func (s *stream) Push(
s.metrics.memoryChunks.Add(float64(len(s.chunks) - prevNumChunks))
}
- if len(failedEntriesWithError) > 0 {
- lastEntryWithErr := failedEntriesWithError[len(failedEntriesWithError)-1]
- _, ok := lastEntryWithErr.e.(*validation.ErrStreamRateLimit)
- outOfOrder := chunkenc.IsOutOfOrderErr(lastEntryWithErr.e)
- if !outOfOrder && !ok {
- return bytesAdded, lastEntryWithErr.e
- }
- var statusCode int
- if outOfOrder {
- statusCode = http.StatusBadRequest
- }
- if ok {
- statusCode = http.StatusTooManyRequests
- }
- // Return a http status 4xx request response with all failed entries.
- buf := bytes.Buffer{}
- streamName := s.labelsString
-
- limitedFailedEntries := failedEntriesWithError
- if maxIgnore := s.cfg.MaxReturnedErrors; maxIgnore > 0 && len(limitedFailedEntries) > maxIgnore {
- limitedFailedEntries = limitedFailedEntries[:maxIgnore]
- }
-
- for _, entryWithError := range limitedFailedEntries {
- fmt.Fprintf(&buf,
- "entry with timestamp %s ignored, reason: '%s' for stream: %s,\n",
- entryWithError.entry.Timestamp.String(), entryWithError.e.Error(), streamName)
- }
-
- fmt.Fprintf(&buf, "total ignored: %d out of %d", len(failedEntriesWithError), len(entries))
-
- return bytesAdded, httpgrpc.Errorf(statusCode, buf.String())
- }
+ return bytesAdded, failedEntriesWithError
- return bytesAdded, nil
}
func (s *stream) cutChunk(ctx context.Context) *chunkDesc {
diff --git a/pkg/ingester/stream_test.go b/pkg/ingester/stream_test.go
index 2b5e3b8711efc..780317a98ed42 100644
--- a/pkg/ingester/stream_test.go
+++ b/pkg/ingester/stream_test.go
@@ -64,10 +64,10 @@ func TestMaxReturnedStreamsErrors(t *testing.T) {
NilMetrics,
)
- _, err := s.Push(context.Background(), []logproto.Entry{
+ _, entriesWithErrors := s.Push(context.Background(), []logproto.Entry{
{Timestamp: time.Unix(int64(numLogs), 0), Line: "log"},
}, recordPool.GetRecord(), 0, true)
- require.NoError(t, err)
+ require.Empty(t, entriesWithErrors)
newLines := make([]logproto.Entry, numLogs)
for i := 0; i < numLogs; i++ {
@@ -86,9 +86,9 @@ func TestMaxReturnedStreamsErrors(t *testing.T) {
fmt.Fprintf(&expected, "total ignored: %d out of %d", numLogs, numLogs)
expectErr := httpgrpc.Errorf(http.StatusBadRequest, expected.String())
- _, err = s.Push(context.Background(), newLines, recordPool.GetRecord(), 0, true)
- require.Error(t, err)
- require.Equal(t, expectErr.Error(), err.Error())
+ _, entriesWithErrors = s.Push(context.Background(), newLines, recordPool.GetRecord(), 0, true)
+ finalErr := errorForFailedEntries(s, entriesWithErrors, len(newLines))
+ require.Equal(t, expectErr.Error(), finalErr.Error())
})
}
}
@@ -110,12 +110,12 @@ func TestPushDeduplication(t *testing.T) {
NilMetrics,
)
- written, err := s.Push(context.Background(), []logproto.Entry{
+ written, entriesWithErrors := s.Push(context.Background(), []logproto.Entry{
{Timestamp: time.Unix(1, 0), Line: "test"},
{Timestamp: time.Unix(1, 0), Line: "test"},
{Timestamp: time.Unix(1, 0), Line: "newer, better test"},
}, recordPool.GetRecord(), 0, true)
- require.NoError(t, err)
+ require.Empty(t, entriesWithErrors)
require.Len(t, s.chunks, 1)
require.Equal(t, s.chunks[0].chunk.Size(), 2,
"expected exact duplicate to be dropped and newer content with same timestamp to be appended")
@@ -140,27 +140,28 @@ func TestPushRejectOldCounter(t *testing.T) {
)
// counter should be 2 now since the first line will be deduped
- _, err = s.Push(context.Background(), []logproto.Entry{
+ _, entriesWithErrors := s.Push(context.Background(), []logproto.Entry{
{Timestamp: time.Unix(1, 0), Line: "test"},
{Timestamp: time.Unix(1, 0), Line: "test"},
{Timestamp: time.Unix(1, 0), Line: "newer, better test"},
}, recordPool.GetRecord(), 0, true)
- require.NoError(t, err)
+ require.Empty(t, entriesWithErrors)
require.Len(t, s.chunks, 1)
require.Equal(t, s.chunks[0].chunk.Size(), 2,
"expected exact duplicate to be dropped and newer content with same timestamp to be appended")
// fail to push with a counter <= the streams internal counter
- _, err = s.Push(context.Background(), []logproto.Entry{
+ _, entriesWithErrors = s.Push(context.Background(), []logproto.Entry{
{Timestamp: time.Unix(1, 0), Line: "test"},
}, recordPool.GetRecord(), 2, true)
- require.Equal(t, ErrEntriesExist, err)
+ require.Len(t, entriesWithErrors, 1)
+ require.Equal(t, entriesWithErrors[0].e, ErrEntriesExist)
// succeed with a greater counter
- _, err = s.Push(context.Background(), []logproto.Entry{
+ _, entriesWithErrors = s.Push(context.Background(), []logproto.Entry{
{Timestamp: time.Unix(1, 0), Line: "test"},
}, recordPool.GetRecord(), 3, true)
- require.Nil(t, err)
+ require.Empty(t, entriesWithErrors)
}
@@ -273,11 +274,11 @@ func TestUnorderedPush(t *testing.T) {
if x.cutBefore {
_ = s.cutChunk(context.Background())
}
- written, err := s.Push(context.Background(), x.entries, recordPool.GetRecord(), 0, true)
+ written, entriesWithErrors := s.Push(context.Background(), x.entries, recordPool.GetRecord(), 0, true)
if x.err {
- require.NotNil(t, err)
+ require.NotEmpty(t, entriesWithErrors)
} else {
- require.Nil(t, err)
+ require.Empty(t, entriesWithErrors)
}
require.Equal(t, x.written, written)
}
@@ -334,8 +335,9 @@ func TestPushRateLimit(t *testing.T) {
{Timestamp: time.Unix(1, 0), Line: "aaaaaaaaab"},
}
// Counter should be 2 now since the first line will be deduped.
- _, err = s.Push(context.Background(), entries, recordPool.GetRecord(), 0, true)
- require.Contains(t, err.Error(), (&validation.ErrStreamRateLimit{RateLimit: l.PerStreamRateLimit, Labels: s.labelsString, Bytes: flagext.ByteSize(len(entries[1].Line))}).Error())
+ _, entriesWithErrors := s.Push(context.Background(), entries, recordPool.GetRecord(), 0, true)
+ require.Len(t, entriesWithErrors, 1)
+ require.Contains(t, entriesWithErrors[0].e.Error(), (&validation.ErrStreamRateLimit{RateLimit: l.PerStreamRateLimit, Labels: s.labelsString, Bytes: flagext.ByteSize(len(entries[1].Line))}).Error())
}
func TestReplayAppendIgnoresValidityWindow(t *testing.T) {
@@ -365,8 +367,8 @@ func TestReplayAppendIgnoresValidityWindow(t *testing.T) {
}
// Push a first entry (it doesn't matter if we look like we're replaying or not)
- _, err = s.Push(context.Background(), entries, nil, 1, true)
- require.Nil(t, err)
+ _, entriesWithErrors := s.Push(context.Background(), entries, nil, 1, true)
+ require.Empty(t, entriesWithErrors)
// Create a sample outside the validity window
entries = []logproto.Entry{
@@ -374,12 +376,12 @@ func TestReplayAppendIgnoresValidityWindow(t *testing.T) {
}
// Pretend it's not a replay, ensure we error
- _, err = s.Push(context.Background(), entries, recordPool.GetRecord(), 0, true)
- require.NotNil(t, err)
+ _, entriesWithErrors = s.Push(context.Background(), entries, recordPool.GetRecord(), 0, true)
+ require.NotEmpty(t, entriesWithErrors)
// Now pretend it's a replay. The same write should succeed.
- _, err = s.Push(context.Background(), entries, nil, 2, true)
- require.Nil(t, err)
+ _, entriesWithErrors = s.Push(context.Background(), entries, nil, 2, true)
+ require.Empty(t, entriesWithErrors)
}
@@ -420,8 +422,8 @@ func Benchmark_PushStream(b *testing.B) {
for n := 0; n < b.N; n++ {
rec := recordPool.GetRecord()
- _, err := s.Push(ctx, e, rec, 0, true)
- require.NoError(b, err)
+ _, entriesWithErrors := s.Push(ctx, e, rec, 0, true)
+ require.Empty(b, entriesWithErrors)
recordPool.PutRecord(rec)
}
}
diff --git a/pkg/logproto/logproto.pb.go b/pkg/logproto/logproto.pb.go
index cb3e65e2f8ab2..ead1e41c3ccdd 100644
--- a/pkg/logproto/logproto.pb.go
+++ b/pkg/logproto/logproto.pb.go
@@ -511,6 +511,49 @@ func (m *LabelRequest) GetEnd() *time.Time {
return nil
}
+type RateLimitedStream struct {
+ Labels string `protobuf:"bytes,1,opt,name=labels,proto3" json:"labels,omitempty"`
+}
+
+func (m *RateLimitedStream) Reset() { *m = RateLimitedStream{} }
+func (*RateLimitedStream) ProtoMessage() {}
+func (*RateLimitedStream) Descriptor() ([]byte, []int) {
+ return fileDescriptor_c28a5f14f1f4c79a, []int{8}
+}
+func (m *RateLimitedStream) XXX_Unmarshal(b []byte) error {
+ return m.Unmarshal(b)
+}
+func (m *RateLimitedStream) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
+ if deterministic {
+ return xxx_messageInfo_RateLimitedStream.Marshal(b, m, deterministic)
+ } else {
+ b = b[:cap(b)]
+ n, err := m.MarshalToSizedBuffer(b)
+ if err != nil {
+ return nil, err
+ }
+ return b[:n], nil
+ }
+}
+func (m *RateLimitedStream) XXX_Merge(src proto.Message) {
+ xxx_messageInfo_RateLimitedStream.Merge(m, src)
+}
+func (m *RateLimitedStream) XXX_Size() int {
+ return m.Size()
+}
+func (m *RateLimitedStream) XXX_DiscardUnknown() {
+ xxx_messageInfo_RateLimitedStream.DiscardUnknown(m)
+}
+
+var xxx_messageInfo_RateLimitedStream proto.InternalMessageInfo
+
+func (m *RateLimitedStream) GetLabels() string {
+ if m != nil {
+ return m.Labels
+ }
+ return ""
+}
+
type LabelResponse struct {
Values []string `protobuf:"bytes,1,rep,name=values,proto3" json:"values,omitempty"`
}
@@ -518,7 +561,7 @@ type LabelResponse struct {
func (m *LabelResponse) Reset() { *m = LabelResponse{} }
func (*LabelResponse) ProtoMessage() {}
func (*LabelResponse) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{8}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{9}
}
func (m *LabelResponse) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -564,7 +607,7 @@ type StreamAdapter struct {
func (m *StreamAdapter) Reset() { *m = StreamAdapter{} }
func (*StreamAdapter) ProtoMessage() {}
func (*StreamAdapter) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{9}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{10}
}
func (m *StreamAdapter) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -622,7 +665,7 @@ type EntryAdapter struct {
func (m *EntryAdapter) Reset() { *m = EntryAdapter{} }
func (*EntryAdapter) ProtoMessage() {}
func (*EntryAdapter) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{10}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{11}
}
func (m *EntryAdapter) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -674,7 +717,7 @@ type Sample struct {
func (m *Sample) Reset() { *m = Sample{} }
func (*Sample) ProtoMessage() {}
func (*Sample) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{11}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{12}
}
func (m *Sample) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -733,7 +776,7 @@ type LegacySample struct {
func (m *LegacySample) Reset() { *m = LegacySample{} }
func (*LegacySample) ProtoMessage() {}
func (*LegacySample) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{12}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{13}
}
func (m *LegacySample) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -785,7 +828,7 @@ type Series struct {
func (m *Series) Reset() { *m = Series{} }
func (*Series) ProtoMessage() {}
func (*Series) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{13}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{14}
}
func (m *Series) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -845,7 +888,7 @@ type TailRequest struct {
func (m *TailRequest) Reset() { *m = TailRequest{} }
func (*TailRequest) ProtoMessage() {}
func (*TailRequest) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{14}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{15}
}
func (m *TailRequest) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -910,7 +953,7 @@ type TailResponse struct {
func (m *TailResponse) Reset() { *m = TailResponse{} }
func (*TailResponse) ProtoMessage() {}
func (*TailResponse) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{15}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{16}
}
func (m *TailResponse) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -956,7 +999,7 @@ type SeriesRequest struct {
func (m *SeriesRequest) Reset() { *m = SeriesRequest{} }
func (*SeriesRequest) ProtoMessage() {}
func (*SeriesRequest) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{16}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{17}
}
func (m *SeriesRequest) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1020,7 +1063,7 @@ type SeriesResponse struct {
func (m *SeriesResponse) Reset() { *m = SeriesResponse{} }
func (*SeriesResponse) ProtoMessage() {}
func (*SeriesResponse) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{17}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{18}
}
func (m *SeriesResponse) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1063,7 +1106,7 @@ type SeriesIdentifier struct {
func (m *SeriesIdentifier) Reset() { *m = SeriesIdentifier{} }
func (*SeriesIdentifier) ProtoMessage() {}
func (*SeriesIdentifier) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{18}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{19}
}
func (m *SeriesIdentifier) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1108,7 +1151,7 @@ type DroppedStream struct {
func (m *DroppedStream) Reset() { *m = DroppedStream{} }
func (*DroppedStream) ProtoMessage() {}
func (*DroppedStream) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{19}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{20}
}
func (m *DroppedStream) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1168,7 +1211,7 @@ type TimeSeriesChunk struct {
func (m *TimeSeriesChunk) Reset() { *m = TimeSeriesChunk{} }
func (*TimeSeriesChunk) ProtoMessage() {}
func (*TimeSeriesChunk) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{20}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{21}
}
func (m *TimeSeriesChunk) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1233,7 +1276,7 @@ type LabelPair struct {
func (m *LabelPair) Reset() { *m = LabelPair{} }
func (*LabelPair) ProtoMessage() {}
func (*LabelPair) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{21}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{22}
}
func (m *LabelPair) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1285,7 +1328,7 @@ type LegacyLabelPair struct {
func (m *LegacyLabelPair) Reset() { *m = LegacyLabelPair{} }
func (*LegacyLabelPair) ProtoMessage() {}
func (*LegacyLabelPair) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{22}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{23}
}
func (m *LegacyLabelPair) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1335,7 +1378,7 @@ type Chunk struct {
func (m *Chunk) Reset() { *m = Chunk{} }
func (*Chunk) ProtoMessage() {}
func (*Chunk) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{23}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{24}
}
func (m *Chunk) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1377,7 +1420,7 @@ type TransferChunksResponse struct {
func (m *TransferChunksResponse) Reset() { *m = TransferChunksResponse{} }
func (*TransferChunksResponse) ProtoMessage() {}
func (*TransferChunksResponse) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{24}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{25}
}
func (m *TransferChunksResponse) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1412,7 +1455,7 @@ type TailersCountRequest struct {
func (m *TailersCountRequest) Reset() { *m = TailersCountRequest{} }
func (*TailersCountRequest) ProtoMessage() {}
func (*TailersCountRequest) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{25}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{26}
}
func (m *TailersCountRequest) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1448,7 +1491,7 @@ type TailersCountResponse struct {
func (m *TailersCountResponse) Reset() { *m = TailersCountResponse{} }
func (*TailersCountResponse) ProtoMessage() {}
func (*TailersCountResponse) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{26}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{27}
}
func (m *TailersCountResponse) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1493,7 +1536,7 @@ type GetChunkIDsRequest struct {
func (m *GetChunkIDsRequest) Reset() { *m = GetChunkIDsRequest{} }
func (*GetChunkIDsRequest) ProtoMessage() {}
func (*GetChunkIDsRequest) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{27}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{28}
}
func (m *GetChunkIDsRequest) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1550,7 +1593,7 @@ type GetChunkIDsResponse struct {
func (m *GetChunkIDsResponse) Reset() { *m = GetChunkIDsResponse{} }
func (*GetChunkIDsResponse) ProtoMessage() {}
func (*GetChunkIDsResponse) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{28}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{29}
}
func (m *GetChunkIDsResponse) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1603,7 +1646,7 @@ type ChunkRef struct {
func (m *ChunkRef) Reset() { *m = ChunkRef{} }
func (*ChunkRef) ProtoMessage() {}
func (*ChunkRef) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{29}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{30}
}
func (m *ChunkRef) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1664,7 +1707,7 @@ type LabelValuesForMetricNameRequest struct {
func (m *LabelValuesForMetricNameRequest) Reset() { *m = LabelValuesForMetricNameRequest{} }
func (*LabelValuesForMetricNameRequest) ProtoMessage() {}
func (*LabelValuesForMetricNameRequest) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{30}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{31}
}
func (m *LabelValuesForMetricNameRequest) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1723,7 +1766,7 @@ type LabelNamesForMetricNameRequest struct {
func (m *LabelNamesForMetricNameRequest) Reset() { *m = LabelNamesForMetricNameRequest{} }
func (*LabelNamesForMetricNameRequest) ProtoMessage() {}
func (*LabelNamesForMetricNameRequest) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{31}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{32}
}
func (m *LabelNamesForMetricNameRequest) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1768,7 +1811,7 @@ type GetChunkRefRequest struct {
func (m *GetChunkRefRequest) Reset() { *m = GetChunkRefRequest{} }
func (*GetChunkRefRequest) ProtoMessage() {}
func (*GetChunkRefRequest) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{32}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{33}
}
func (m *GetChunkRefRequest) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1811,7 +1854,7 @@ type GetChunkRefResponse struct {
func (m *GetChunkRefResponse) Reset() { *m = GetChunkRefResponse{} }
func (*GetChunkRefResponse) ProtoMessage() {}
func (*GetChunkRefResponse) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{33}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{34}
}
func (m *GetChunkRefResponse) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1856,7 +1899,7 @@ type GetSeriesRequest struct {
func (m *GetSeriesRequest) Reset() { *m = GetSeriesRequest{} }
func (*GetSeriesRequest) ProtoMessage() {}
func (*GetSeriesRequest) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{34}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{35}
}
func (m *GetSeriesRequest) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1899,7 +1942,7 @@ type GetSeriesResponse struct {
func (m *GetSeriesResponse) Reset() { *m = GetSeriesResponse{} }
func (*GetSeriesResponse) ProtoMessage() {}
func (*GetSeriesResponse) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{35}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{36}
}
func (m *GetSeriesResponse) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1943,7 +1986,7 @@ type IndexSeries struct {
func (m *IndexSeries) Reset() { *m = IndexSeries{} }
func (*IndexSeries) ProtoMessage() {}
func (*IndexSeries) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{36}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{37}
}
func (m *IndexSeries) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -1980,7 +2023,7 @@ type QueryIndexResponse struct {
func (m *QueryIndexResponse) Reset() { *m = QueryIndexResponse{} }
func (*QueryIndexResponse) ProtoMessage() {}
func (*QueryIndexResponse) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{37}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{38}
}
func (m *QueryIndexResponse) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -2031,7 +2074,7 @@ type Row struct {
func (m *Row) Reset() { *m = Row{} }
func (*Row) ProtoMessage() {}
func (*Row) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{38}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{39}
}
func (m *Row) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -2081,7 +2124,7 @@ type QueryIndexRequest struct {
func (m *QueryIndexRequest) Reset() { *m = QueryIndexRequest{} }
func (*QueryIndexRequest) ProtoMessage() {}
func (*QueryIndexRequest) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{39}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{40}
}
func (m *QueryIndexRequest) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -2128,7 +2171,7 @@ type IndexQuery struct {
func (m *IndexQuery) Reset() { *m = IndexQuery{} }
func (*IndexQuery) ProtoMessage() {}
func (*IndexQuery) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{40}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{41}
}
func (m *IndexQuery) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -2201,7 +2244,7 @@ type IndexStatsRequest struct {
func (m *IndexStatsRequest) Reset() { *m = IndexStatsRequest{} }
func (*IndexStatsRequest) ProtoMessage() {}
func (*IndexStatsRequest) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{41}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{42}
}
func (m *IndexStatsRequest) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -2247,7 +2290,7 @@ type IndexStatsResponse struct {
func (m *IndexStatsResponse) Reset() { *m = IndexStatsResponse{} }
func (*IndexStatsResponse) ProtoMessage() {}
func (*IndexStatsResponse) Descriptor() ([]byte, []int) {
- return fileDescriptor_c28a5f14f1f4c79a, []int{42}
+ return fileDescriptor_c28a5f14f1f4c79a, []int{43}
}
func (m *IndexStatsResponse) XXX_Unmarshal(b []byte) error {
return m.Unmarshal(b)
@@ -2314,6 +2357,7 @@ func init() {
proto.RegisterType((*QueryResponse)(nil), "logproto.QueryResponse")
proto.RegisterType((*SampleQueryResponse)(nil), "logproto.SampleQueryResponse")
proto.RegisterType((*LabelRequest)(nil), "logproto.LabelRequest")
+ proto.RegisterType((*RateLimitedStream)(nil), "logproto.RateLimitedStream")
proto.RegisterType((*LabelResponse)(nil), "logproto.LabelResponse")
proto.RegisterType((*StreamAdapter)(nil), "logproto.StreamAdapter")
proto.RegisterType((*EntryAdapter)(nil), "logproto.EntryAdapter")
@@ -2355,138 +2399,139 @@ func init() {
func init() { proto.RegisterFile("pkg/logproto/logproto.proto", fileDescriptor_c28a5f14f1f4c79a) }
var fileDescriptor_c28a5f14f1f4c79a = []byte{
- // 2088 bytes of a gzipped FileDescriptorProto
- 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xd4, 0x59, 0x4f, 0x6f, 0x1b, 0xc7,
- 0x15, 0xe7, 0x90, 0x4b, 0x8a, 0x7c, 0xa4, 0x24, 0x66, 0x2c, 0xcb, 0x0c, 0x6d, 0x73, 0xe5, 0x45,
- 0x6a, 0x0b, 0x8e, 0x4d, 0xd6, 0x4a, 0xdb, 0x38, 0x76, 0xd3, 0x42, 0x94, 0x62, 0x5b, 0xb6, 0xe2,
- 0x28, 0x2b, 0xd7, 0x01, 0x02, 0x14, 0xc6, 0x8a, 0x1c, 0x92, 0x0b, 0x71, 0xb9, 0xf4, 0xee, 0x30,
- 0x8e, 0x80, 0x02, 0xed, 0x07, 0x68, 0x80, 0xf4, 0x54, 0xf4, 0x5e, 0xa0, 0x45, 0x0f, 0x3d, 0x14,
- 0xe8, 0xb5, 0xed, 0xad, 0xee, 0xcd, 0xbd, 0x05, 0x39, 0xb0, 0xb5, 0x7c, 0x29, 0x74, 0xca, 0x27,
- 0x28, 0x8a, 0xf9, 0xb7, 0x3b, 0x5c, 0x53, 0x8d, 0xe9, 0x1a, 0x28, 0x72, 0x11, 0x67, 0xde, 0xcc,
- 0xbc, 0x37, 0xef, 0x37, 0xef, 0xef, 0x0a, 0x4e, 0x0f, 0xf7, 0xbb, 0x8d, 0xbe, 0xdf, 0x1d, 0x06,
- 0x3e, 0xf5, 0xa3, 0x41, 0x9d, 0xff, 0xc5, 0x79, 0x35, 0xaf, 0x5e, 0xee, 0xba, 0xb4, 0x37, 0xda,
- 0xab, 0xb7, 0x7c, 0xaf, 0xd1, 0xf5, 0xbb, 0x7e, 0x83, 0x93, 0xf7, 0x46, 0x1d, 0x3e, 0x13, 0x87,
- 0xd9, 0x48, 0x1c, 0xac, 0x9a, 0x5d, 0xdf, 0xef, 0xf6, 0x49, 0xbc, 0x8b, 0xba, 0x1e, 0x09, 0xa9,
- 0xe3, 0x0d, 0xe5, 0x86, 0x15, 0x29, 0xf6, 0x61, 0xdf, 0xf3, 0xdb, 0xa4, 0xdf, 0x08, 0xa9, 0x43,
- 0x43, 0xf1, 0x57, 0xec, 0xb0, 0x3e, 0x82, 0xe2, 0xce, 0x28, 0xec, 0xd9, 0xe4, 0xe1, 0x88, 0x84,
- 0x14, 0xdf, 0x82, 0xb9, 0x90, 0x06, 0xc4, 0xf1, 0xc2, 0x0a, 0x5a, 0xc9, 0xac, 0x16, 0xd7, 0x4e,
- 0xd5, 0xa3, 0xcb, 0xee, 0xf2, 0x85, 0xf5, 0xb6, 0x33, 0xa4, 0x24, 0x68, 0x9e, 0xfc, 0x72, 0x6c,
- 0xe6, 0x04, 0xe9, 0x68, 0x6c, 0xaa, 0x53, 0xb6, 0x1a, 0x58, 0x0b, 0x50, 0x12, 0x8c, 0xc3, 0xa1,
- 0x3f, 0x08, 0x89, 0xf5, 0xd7, 0x34, 0x94, 0x3e, 0x1c, 0x91, 0xe0, 0x40, 0x89, 0xaa, 0x42, 0x3e,
- 0x24, 0x7d, 0xd2, 0xa2, 0x7e, 0x50, 0x41, 0x2b, 0x68, 0xb5, 0x60, 0x47, 0x73, 0xbc, 0x04, 0xd9,
- 0xbe, 0xeb, 0xb9, 0xb4, 0x92, 0x5e, 0x41, 0xab, 0xf3, 0xb6, 0x98, 0xe0, 0x6b, 0x90, 0x0d, 0xa9,
- 0x13, 0xd0, 0x4a, 0x66, 0x05, 0xad, 0x16, 0xd7, 0xaa, 0x75, 0xa1, 0x7e, 0x5d, 0xa9, 0x5f, 0xbf,
- 0xa7, 0xd4, 0x6f, 0xe6, 0x1f, 0x8f, 0xcd, 0xd4, 0xe7, 0xff, 0x30, 0x91, 0x2d, 0x8e, 0xe0, 0xef,
- 0x41, 0x86, 0x0c, 0xda, 0x15, 0x63, 0x86, 0x93, 0xec, 0x00, 0xbe, 0x02, 0x85, 0xb6, 0x1b, 0x90,
- 0x16, 0x75, 0xfd, 0x41, 0x25, 0xbb, 0x82, 0x56, 0x17, 0xd6, 0x4e, 0xc4, 0x90, 0x6c, 0xaa, 0x25,
- 0x3b, 0xde, 0x85, 0x2f, 0x41, 0x2e, 0xec, 0x39, 0x41, 0x3b, 0xac, 0xcc, 0xad, 0x64, 0x56, 0x0b,
- 0xcd, 0xa5, 0xa3, 0xb1, 0x59, 0x16, 0x94, 0x4b, 0xbe, 0xe7, 0x52, 0xe2, 0x0d, 0xe9, 0x81, 0x2d,
- 0xf7, 0xe0, 0x8b, 0x30, 0xd7, 0x26, 0x7d, 0x42, 0x49, 0x58, 0xc9, 0x73, 0xc4, 0xcb, 0x1a, 0x7b,
- 0xbe, 0x60, 0xab, 0x0d, 0xb7, 0x8d, 0x7c, 0xae, 0x3c, 0x67, 0xfd, 0x1b, 0x01, 0xde, 0x75, 0xbc,
- 0x61, 0x9f, 0xbc, 0x30, 0x9e, 0x11, 0x72, 0xe9, 0x97, 0x46, 0x2e, 0x33, 0x2b, 0x72, 0x31, 0x0c,
- 0xc6, 0x6c, 0x30, 0x64, 0xbf, 0x06, 0x06, 0x6b, 0x1b, 0x72, 0x82, 0xf4, 0x75, 0x36, 0x14, 0xeb,
- 0x9c, 0x51, 0xda, 0x94, 0x63, 0x6d, 0x32, 0xfc, 0x9e, 0xd6, 0x4f, 0x61, 0x5e, 0xe2, 0x28, 0x2c,
- 0x15, 0xaf, 0xbf, 0xb0, 0x0f, 0x2c, 0x3c, 0x1e, 0x9b, 0x28, 0xf6, 0x83, 0xc8, 0xf8, 0xf1, 0x9b,
- 0x5c, 0x36, 0x0d, 0x25, 0xde, 0x8b, 0x75, 0xe1, 0x72, 0x5b, 0x83, 0x2e, 0x09, 0xd9, 0x41, 0x83,
- 0x41, 0x65, 0x8b, 0x3d, 0xd6, 0x4f, 0xe0, 0xc4, 0xc4, 0x73, 0xca, 0x6b, 0x5c, 0x85, 0x5c, 0x48,
- 0x02, 0x97, 0xa8, 0x5b, 0x68, 0x80, 0xec, 0x72, 0xba, 0x26, 0x9e, 0xcf, 0x6d, 0xb9, 0x7f, 0x36,
- 0xe9, 0xbf, 0x47, 0x50, 0xda, 0x76, 0xf6, 0x48, 0x5f, 0xd9, 0x11, 0x06, 0x63, 0xe0, 0x78, 0x44,
- 0xe2, 0xc9, 0xc7, 0x78, 0x19, 0x72, 0x9f, 0x38, 0xfd, 0x11, 0x11, 0x2c, 0xf3, 0xb6, 0x9c, 0xcd,
- 0xea, 0x91, 0xe8, 0xa5, 0x3d, 0x12, 0x45, 0x76, 0x65, 0x5d, 0x80, 0x79, 0x79, 0x5f, 0x09, 0x54,
- 0x7c, 0x39, 0x06, 0x54, 0x41, 0x5d, 0xce, 0xfa, 0x05, 0x82, 0xf9, 0x89, 0xf7, 0xc2, 0x16, 0xe4,
- 0xfa, 0xec, 0x68, 0x28, 0x94, 0x6b, 0xc2, 0xd1, 0xd8, 0x94, 0x14, 0x5b, 0xfe, 0xb2, 0xd7, 0x27,
- 0x03, 0xca, 0x71, 0x4f, 0x73, 0xdc, 0x97, 0x63, 0xdc, 0xdf, 0x1b, 0xd0, 0xe0, 0x40, 0x3d, 0xfe,
- 0x22, 0x43, 0x91, 0x85, 0x3e, 0xb9, 0xdd, 0x56, 0x03, 0xfc, 0x3a, 0x18, 0x3d, 0x27, 0xec, 0x71,
- 0x50, 0x8c, 0x66, 0xf6, 0x68, 0x6c, 0xa2, 0xcb, 0x36, 0x27, 0x59, 0x9f, 0x40, 0x49, 0x67, 0x82,
- 0x6f, 0x41, 0x21, 0x8a, 0xd9, 0xfc, 0x52, 0xff, 0x1d, 0x8a, 0x05, 0x29, 0x33, 0x4d, 0x43, 0x0e,
- 0x48, 0x7c, 0x18, 0x9f, 0x01, 0xa3, 0xef, 0x0e, 0x08, 0x7f, 0xa0, 0x42, 0x33, 0x7f, 0x34, 0x36,
- 0xf9, 0xdc, 0xe6, 0x7f, 0x2d, 0x0f, 0x72, 0xc2, 0xc6, 0xf0, 0x1b, 0x49, 0x89, 0x99, 0x66, 0x4e,
- 0x70, 0xd4, 0xb9, 0x99, 0x90, 0xe5, 0x28, 0x72, 0x76, 0xa8, 0x59, 0x38, 0x1a, 0x9b, 0x82, 0x60,
- 0x8b, 0x1f, 0x26, 0x4e, 0xd3, 0x91, 0x8b, 0x63, 0x73, 0xa9, 0xe6, 0x4d, 0x28, 0x6d, 0x93, 0xae,
- 0xd3, 0x3a, 0x90, 0x42, 0x97, 0x14, 0x3b, 0x26, 0x10, 0x29, 0x1e, 0xe7, 0xa0, 0x14, 0x49, 0x7c,
- 0xe0, 0x85, 0xd2, 0x51, 0x8b, 0x11, 0xed, 0xfd, 0xd0, 0xfa, 0x15, 0x02, 0x69, 0xdd, 0x2f, 0xf4,
- 0x78, 0xd7, 0x61, 0x2e, 0xe4, 0x12, 0xd5, 0xe3, 0xe9, 0x4e, 0xc3, 0x17, 0xe2, 0x67, 0x93, 0x1b,
- 0x6d, 0x35, 0xc0, 0x75, 0x00, 0xe1, 0xbf, 0xb7, 0x62, 0xc5, 0x16, 0x8e, 0xc6, 0xa6, 0x46, 0xb5,
- 0xb5, 0xb1, 0xf5, 0x4b, 0x04, 0xc5, 0x7b, 0x8e, 0x1b, 0x39, 0xce, 0x12, 0x64, 0x1f, 0x32, 0x0f,
- 0x96, 0x9e, 0x23, 0x26, 0x2c, 0x44, 0xb5, 0x49, 0xdf, 0x39, 0xb8, 0xe1, 0x07, 0x9c, 0xe7, 0xbc,
- 0x1d, 0xcd, 0xe3, 0x34, 0x67, 0x4c, 0x4d, 0x73, 0xd9, 0x99, 0x83, 0xf5, 0x6d, 0x23, 0x9f, 0x2e,
- 0x67, 0xac, 0x9f, 0x23, 0x28, 0x89, 0x9b, 0x49, 0x17, 0xb9, 0x0e, 0x39, 0x71, 0x71, 0x69, 0x63,
- 0xc7, 0x46, 0x34, 0xd0, 0xa2, 0x99, 0x3c, 0x82, 0x7f, 0x08, 0x0b, 0xed, 0xc0, 0x1f, 0x0e, 0x49,
- 0x7b, 0x57, 0x86, 0xc5, 0x74, 0x32, 0x2c, 0x6e, 0xea, 0xeb, 0x76, 0x62, 0xbb, 0xf5, 0x37, 0xe6,
- 0x88, 0x22, 0x44, 0x49, 0xa8, 0x22, 0x15, 0xd1, 0x4b, 0xe7, 0xa3, 0xf4, 0xac, 0xf9, 0x68, 0x19,
- 0x72, 0xdd, 0xc0, 0x1f, 0x0d, 0xc3, 0x4a, 0x46, 0x84, 0x09, 0x31, 0x9b, 0x2d, 0x4f, 0x59, 0xb7,
- 0x61, 0x41, 0xa9, 0x72, 0x4c, 0x9c, 0xae, 0x26, 0xe3, 0xf4, 0x56, 0x9b, 0x0c, 0xa8, 0xdb, 0x71,
- 0xa3, 0xc8, 0x2b, 0xf7, 0x5b, 0x9f, 0x21, 0x28, 0x27, 0xb7, 0xe0, 0x1f, 0x68, 0x66, 0xce, 0xd8,
- 0x9d, 0x3f, 0x9e, 0x5d, 0x9d, 0xc7, 0xc1, 0x90, 0x07, 0x14, 0xe5, 0x02, 0xd5, 0x77, 0xa0, 0xa8,
- 0x91, 0x59, 0xbe, 0xdb, 0x27, 0xca, 0x24, 0xd9, 0x30, 0xf6, 0xc5, 0xb4, 0x30, 0x53, 0x3e, 0xb9,
- 0x96, 0xbe, 0x8a, 0x98, 0x41, 0xcf, 0x4f, 0xbc, 0x24, 0xbe, 0x0a, 0x46, 0x27, 0xf0, 0xbd, 0x99,
- 0x9e, 0x89, 0x9f, 0xc0, 0xdf, 0x81, 0x34, 0xf5, 0x67, 0x7a, 0xa4, 0x34, 0xf5, 0xd9, 0x1b, 0x49,
- 0xe5, 0x33, 0xfc, 0x72, 0x72, 0x66, 0xfd, 0x0e, 0xc1, 0x22, 0x3b, 0x23, 0x10, 0xd8, 0xe8, 0x8d,
- 0x06, 0xfb, 0x78, 0x15, 0xca, 0x4c, 0xd2, 0x03, 0x57, 0xa6, 0xb5, 0x07, 0x6e, 0x5b, 0xaa, 0xb9,
- 0xc0, 0xe8, 0x2a, 0xdb, 0x6d, 0xb5, 0xf1, 0x29, 0x98, 0x1b, 0x85, 0x62, 0x83, 0xd0, 0x39, 0xc7,
- 0xa6, 0x5b, 0x6d, 0xfc, 0xa6, 0x26, 0x8e, 0x61, 0xad, 0x55, 0x76, 0x1c, 0xc3, 0x1d, 0xc7, 0x0d,
- 0xa2, 0xd8, 0x72, 0x01, 0x72, 0x2d, 0x26, 0x58, 0xd8, 0x09, 0x4b, 0xab, 0xd1, 0x66, 0x7e, 0x21,
- 0x5b, 0x2e, 0x5b, 0xdf, 0x85, 0x42, 0x74, 0x7a, 0x6a, 0x36, 0x9d, 0xfa, 0x02, 0xd6, 0x75, 0x58,
- 0x14, 0x31, 0x73, 0xfa, 0xe1, 0xd2, 0xb4, 0xc3, 0x25, 0x75, 0xf8, 0x34, 0x64, 0x05, 0x2a, 0x18,
- 0x8c, 0xb6, 0x43, 0x1d, 0x75, 0x84, 0x8d, 0xad, 0x0a, 0x2c, 0xdf, 0x0b, 0x9c, 0x41, 0xd8, 0x21,
- 0x01, 0xdf, 0x14, 0xd9, 0xae, 0x75, 0x12, 0x4e, 0xb0, 0x38, 0x41, 0x82, 0x70, 0xc3, 0x1f, 0x0d,
- 0xa8, 0x74, 0x4f, 0xeb, 0x12, 0x2c, 0x4d, 0x92, 0xa5, 0xa9, 0x2f, 0x41, 0xb6, 0xc5, 0x08, 0x9c,
- 0xfb, 0xbc, 0x2d, 0x26, 0xd6, 0xaf, 0x11, 0xe0, 0x9b, 0x84, 0x72, 0xd6, 0x5b, 0x9b, 0xa1, 0x56,
- 0x8f, 0x7a, 0x0e, 0x6d, 0xf5, 0x48, 0x10, 0xaa, 0xda, 0x4c, 0xcd, 0xff, 0x1f, 0xf5, 0xa8, 0x75,
- 0x05, 0x4e, 0x4c, 0xdc, 0x52, 0xea, 0x54, 0x85, 0x7c, 0x4b, 0xd2, 0x64, 0xfd, 0x10, 0xcd, 0xad,
- 0x3f, 0xa4, 0x21, 0x2f, 0xde, 0x96, 0x74, 0xf0, 0x15, 0x28, 0x76, 0x98, 0xad, 0x05, 0xc3, 0xc0,
- 0x95, 0x10, 0x18, 0xcd, 0xc5, 0xa3, 0xb1, 0xa9, 0x93, 0x6d, 0x7d, 0x82, 0x2f, 0x27, 0x0c, 0xaf,
- 0xb9, 0x74, 0x38, 0x36, 0x73, 0x3f, 0x62, 0xc6, 0xb7, 0xc9, 0xb2, 0x17, 0x37, 0xc3, 0xcd, 0xc8,
- 0x1c, 0xef, 0x48, 0x6f, 0xe3, 0xc5, 0x69, 0xf3, 0x6d, 0x76, 0xfd, 0x2f, 0xc7, 0xe6, 0x05, 0xad,
- 0x27, 0x1c, 0x06, 0xbe, 0x47, 0x68, 0x8f, 0x8c, 0xc2, 0x46, 0xcb, 0xf7, 0x3c, 0x7f, 0xd0, 0xe0,
- 0x7d, 0x1d, 0x57, 0x9a, 0xa5, 0x60, 0x76, 0x5c, 0x3a, 0xe0, 0x3d, 0x98, 0xa3, 0xbd, 0xc0, 0x1f,
- 0x75, 0x7b, 0x3c, 0xbb, 0x64, 0x9a, 0xd7, 0x66, 0xe7, 0xa7, 0x38, 0xd8, 0x6a, 0x80, 0xcf, 0x31,
- 0xb4, 0x48, 0x6b, 0x3f, 0x1c, 0x79, 0x3c, 0x3d, 0xcd, 0xab, 0xf2, 0x26, 0x22, 0x5b, 0x9f, 0xa5,
- 0xc1, 0xe4, 0x26, 0x7c, 0x9f, 0x97, 0x61, 0x37, 0xfc, 0xe0, 0x7d, 0x42, 0x03, 0xb7, 0x75, 0xd7,
- 0xf1, 0x88, 0xb2, 0x0d, 0x13, 0x8a, 0x1e, 0x27, 0x3e, 0xd0, 0x9c, 0x03, 0xbc, 0x68, 0x1f, 0x3e,
- 0x0b, 0xc0, 0xdd, 0x4e, 0xac, 0x0b, 0x3f, 0x29, 0x70, 0x0a, 0x5f, 0xde, 0x98, 0x40, 0xaa, 0x31,
- 0xa3, 0x66, 0x12, 0xa1, 0xad, 0x24, 0x42, 0x33, 0xf3, 0x89, 0x60, 0xd1, 0x6d, 0x3d, 0x3b, 0x69,
- 0xeb, 0xd6, 0xdf, 0x11, 0xd4, 0xb6, 0xd5, 0xcd, 0x5f, 0x12, 0x0e, 0xa5, 0x6f, 0xfa, 0x15, 0xe9,
- 0x9b, 0xf9, 0xdf, 0xf4, 0xb5, 0xfe, 0xa2, 0xb9, 0xbc, 0x4d, 0x3a, 0x4a, 0x8f, 0x0d, 0x2d, 0x5d,
- 0xbc, 0x8a, 0x6b, 0xa6, 0x5f, 0xe1, 0xb3, 0x64, 0x12, 0xcf, 0xf2, 0x6e, 0x1c, 0x0e, 0xb8, 0x06,
- 0x32, 0x1c, 0x9c, 0x07, 0x23, 0x20, 0x1d, 0x95, 0x7c, 0x71, 0x32, 0xc6, 0x93, 0x8e, 0xcd, 0xd7,
- 0xad, 0x3f, 0x21, 0x28, 0xdf, 0x24, 0x74, 0xb2, 0xac, 0xf9, 0x26, 0xe9, 0x7f, 0x0b, 0x5e, 0xd3,
- 0xee, 0x2f, 0xb5, 0x7f, 0x2b, 0x51, 0xcb, 0x9c, 0x8c, 0xf5, 0xdf, 0x1a, 0xb4, 0xc9, 0xa7, 0xb2,
- 0xf1, 0x9c, 0x2c, 0x63, 0x76, 0xa0, 0xa8, 0x2d, 0xe2, 0xf5, 0x44, 0x01, 0x33, 0x2d, 0xa9, 0x36,
- 0x97, 0xa4, 0x4e, 0xa2, 0xf5, 0x94, 0xd5, 0x67, 0x94, 0xee, 0x77, 0x01, 0xf3, 0x5e, 0x98, 0xb3,
- 0xd5, 0x23, 0x35, 0xa7, 0xde, 0x89, 0xea, 0x99, 0x68, 0x8e, 0xcf, 0x81, 0x11, 0xf8, 0x8f, 0x54,
- 0x65, 0x3a, 0x1f, 0x8b, 0xb4, 0xfd, 0x47, 0x36, 0x5f, 0xb2, 0xae, 0x43, 0xc6, 0xf6, 0x1f, 0xe1,
- 0x1a, 0x40, 0xe0, 0x0c, 0xba, 0xe4, 0x7e, 0xd4, 0x8f, 0x94, 0x6c, 0x8d, 0x72, 0x4c, 0x7e, 0xdd,
- 0x80, 0xd7, 0xf4, 0x1b, 0x89, 0xe7, 0xae, 0xc3, 0x1c, 0x23, 0xc6, 0x70, 0x2d, 0x25, 0xe0, 0x12,
- 0x0d, 0xbd, 0xda, 0xc4, 0x6c, 0x06, 0x62, 0x3a, 0x3e, 0x03, 0x05, 0xea, 0xec, 0xf5, 0xc9, 0xdd,
- 0xd8, 0xe7, 0x63, 0x02, 0x5b, 0x65, 0xad, 0xd4, 0x7d, 0xad, 0x50, 0x88, 0x09, 0xf8, 0x22, 0x94,
- 0xe3, 0x3b, 0xef, 0x04, 0xa4, 0xe3, 0x7e, 0xca, 0x5f, 0xb8, 0x64, 0x3f, 0x47, 0xc7, 0xab, 0xb0,
- 0x18, 0xd3, 0x76, 0x79, 0xda, 0x35, 0xf8, 0xd6, 0x24, 0x99, 0x61, 0xc3, 0xd5, 0x7d, 0xef, 0xe1,
- 0xc8, 0xe9, 0xf3, 0x40, 0x56, 0xb2, 0x35, 0x8a, 0xf5, 0x67, 0x04, 0xaf, 0x89, 0xa7, 0xa6, 0x0e,
- 0xfd, 0x46, 0x5a, 0xfd, 0x6f, 0x10, 0x60, 0x5d, 0x03, 0x69, 0x5a, 0xdf, 0xd2, 0x3f, 0xf9, 0xb0,
- 0xbc, 0x5e, 0x9c, 0xf6, 0x4d, 0x93, 0xb5, 0xa0, 0xb2, 0x04, 0x4c, 0xf3, 0x5d, 0xbc, 0x05, 0x15,
- 0x14, 0x55, 0xfd, 0xb1, 0xce, 0x79, 0xef, 0x80, 0x92, 0x50, 0x36, 0x90, 0xbc, 0x73, 0xe6, 0x04,
- 0x5b, 0xfc, 0x30, 0x59, 0xea, 0x03, 0x83, 0x11, 0xcb, 0x4a, 0x7e, 0x44, 0xb8, 0x78, 0x1e, 0x0a,
- 0xd1, 0xd7, 0x45, 0x5c, 0x84, 0xb9, 0x1b, 0x1f, 0xd8, 0x1f, 0xad, 0xdb, 0x9b, 0xe5, 0x14, 0x2e,
- 0x41, 0xbe, 0xb9, 0xbe, 0x71, 0x87, 0xcf, 0xd0, 0xda, 0x3a, 0xe4, 0x76, 0x46, 0x61, 0x8f, 0x04,
- 0xf8, 0x6d, 0x30, 0xd8, 0x08, 0x6b, 0x4e, 0xab, 0x7d, 0xda, 0xad, 0x2e, 0x27, 0xc9, 0xb2, 0x06,
- 0x4c, 0xad, 0xfd, 0xd1, 0x50, 0x86, 0x1c, 0xe0, 0xef, 0x43, 0x56, 0x58, 0xa7, 0xb6, 0x5d, 0xff,
- 0xcc, 0x58, 0x3d, 0xf5, 0x1c, 0x5d, 0xf1, 0xf9, 0x36, 0xc2, 0x77, 0xa1, 0xc8, 0x89, 0xb2, 0xed,
- 0x3f, 0x93, 0xec, 0xbe, 0x27, 0x38, 0x9d, 0x3d, 0x66, 0x55, 0xe3, 0x77, 0x0d, 0xb2, 0x3c, 0x40,
- 0xe8, 0xb7, 0xd1, 0x3f, 0x56, 0xe9, 0xb7, 0x99, 0xf8, 0x28, 0x64, 0xa5, 0xf0, 0x3b, 0x60, 0xb0,
- 0x22, 0x56, 0x87, 0x43, 0xeb, 0xd6, 0x75, 0x38, 0xf4, 0x56, 0x99, 0x8b, 0x7d, 0x37, 0xfa, 0xe8,
- 0x70, 0x2a, 0xd9, 0x7d, 0xa9, 0xe3, 0x95, 0xe7, 0x17, 0x22, 0xc9, 0x1f, 0x88, 0xee, 0x5b, 0x95,
- 0xcf, 0xf8, 0xec, 0xa4, 0xa8, 0x44, 0xb5, 0x5d, 0xad, 0x1d, 0xb7, 0x1c, 0x31, 0xdc, 0x86, 0xa2,
- 0x56, 0xba, 0xea, 0xb0, 0x3e, 0x5f, 0x77, 0xeb, 0xb0, 0x4e, 0xa9, 0x77, 0xad, 0x14, 0xbe, 0x09,
- 0x79, 0x16, 0xf9, 0x99, 0x03, 0xe0, 0xd3, 0xc9, 0x00, 0xaf, 0x39, 0x76, 0xf5, 0xcc, 0xf4, 0xc5,
- 0xc8, 0x6e, 0x7e, 0x0c, 0x79, 0xd5, 0x65, 0xe1, 0x0f, 0x61, 0x61, 0xb2, 0xc7, 0xc0, 0xaf, 0x6b,
- 0x6a, 0x4d, 0xb6, 0x6e, 0xd5, 0x15, 0x6d, 0x69, 0x7a, 0x63, 0x92, 0x5a, 0x45, 0xcd, 0x8f, 0x9f,
- 0x3c, 0xad, 0xa5, 0xbe, 0x78, 0x5a, 0x4b, 0x7d, 0xf5, 0xb4, 0x86, 0x7e, 0x76, 0x58, 0x43, 0xbf,
- 0x3d, 0xac, 0xa1, 0xc7, 0x87, 0x35, 0xf4, 0xe4, 0xb0, 0x86, 0xfe, 0x79, 0x58, 0x43, 0xff, 0x3a,
- 0xac, 0xa5, 0xbe, 0x3a, 0xac, 0xa1, 0xcf, 0x9f, 0xd5, 0x52, 0x4f, 0x9e, 0xd5, 0x52, 0x5f, 0x3c,
- 0xab, 0xa5, 0x3e, 0x7e, 0x43, 0xff, 0x17, 0x4a, 0xe0, 0x74, 0x9c, 0x81, 0xd3, 0xe8, 0xfb, 0xfb,
- 0x6e, 0x43, 0xff, 0x0f, 0xcc, 0x5e, 0x8e, 0xff, 0xbc, 0xf5, 0x9f, 0x00, 0x00, 0x00, 0xff, 0xff,
- 0xdc, 0x73, 0x3e, 0x8e, 0x98, 0x19, 0x00, 0x00,
+ // 2098 bytes of a gzipped FileDescriptorProto
+ 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xd4, 0x59, 0xcd, 0x6f, 0x1b, 0xc7,
+ 0x15, 0xe7, 0x90, 0x4b, 0x8a, 0x7c, 0xa4, 0x24, 0x7a, 0x2c, 0xdb, 0x0c, 0x6d, 0x73, 0xe5, 0x45,
+ 0x6a, 0x0b, 0xfe, 0x20, 0x6b, 0xa5, 0x6d, 0x1c, 0xbb, 0x69, 0x21, 0x4a, 0xb1, 0x2d, 0x5b, 0x71,
+ 0x94, 0x95, 0xeb, 0x00, 0x01, 0x0a, 0x63, 0x45, 0x0e, 0xc9, 0x85, 0xb8, 0x5c, 0x7a, 0x77, 0x18,
+ 0x47, 0x40, 0x81, 0xf6, 0x0f, 0x68, 0x80, 0xf4, 0x54, 0xf4, 0x5e, 0xa0, 0x45, 0x0f, 0x3d, 0x14,
+ 0xe8, 0xb5, 0xed, 0xad, 0xee, 0xcd, 0xbd, 0x05, 0x39, 0xb0, 0xb5, 0x7c, 0x29, 0x74, 0xca, 0x5f,
+ 0x50, 0x14, 0xf3, 0xb5, 0x3b, 0x5c, 0x53, 0x8d, 0xe9, 0x1a, 0x28, 0x72, 0x11, 0x67, 0xde, 0xcc,
+ 0xbc, 0x37, 0xef, 0x37, 0xef, 0x73, 0x05, 0xa7, 0x87, 0x7b, 0xdd, 0x46, 0xdf, 0xef, 0x0e, 0x03,
+ 0x9f, 0xfa, 0xd1, 0xa0, 0xce, 0xff, 0xe2, 0xbc, 0x9a, 0x57, 0xaf, 0x74, 0x5d, 0xda, 0x1b, 0xed,
+ 0xd6, 0x5b, 0xbe, 0xd7, 0xe8, 0xfa, 0x5d, 0xbf, 0xc1, 0xc9, 0xbb, 0xa3, 0x0e, 0x9f, 0x89, 0xc3,
+ 0x6c, 0x24, 0x0e, 0x56, 0xcd, 0xae, 0xef, 0x77, 0xfb, 0x24, 0xde, 0x45, 0x5d, 0x8f, 0x84, 0xd4,
+ 0xf1, 0x86, 0x72, 0xc3, 0xb2, 0x14, 0xfb, 0xa8, 0xef, 0xf9, 0x6d, 0xd2, 0x6f, 0x84, 0xd4, 0xa1,
+ 0xa1, 0xf8, 0x2b, 0x76, 0x58, 0x1f, 0x41, 0x71, 0x7b, 0x14, 0xf6, 0x6c, 0xf2, 0x68, 0x44, 0x42,
+ 0x8a, 0x6f, 0xc3, 0x5c, 0x48, 0x03, 0xe2, 0x78, 0x61, 0x05, 0x2d, 0x67, 0x56, 0x8a, 0xab, 0xa7,
+ 0xea, 0xd1, 0x65, 0x77, 0xf8, 0xc2, 0x5a, 0xdb, 0x19, 0x52, 0x12, 0x34, 0x4f, 0x7c, 0x39, 0x36,
+ 0x73, 0x82, 0x74, 0x38, 0x36, 0xd5, 0x29, 0x5b, 0x0d, 0xac, 0x05, 0x28, 0x09, 0xc6, 0xe1, 0xd0,
+ 0x1f, 0x84, 0xc4, 0xfa, 0x6b, 0x1a, 0x4a, 0x1f, 0x8e, 0x48, 0xb0, 0xaf, 0x44, 0x55, 0x21, 0x1f,
+ 0x92, 0x3e, 0x69, 0x51, 0x3f, 0xa8, 0xa0, 0x65, 0xb4, 0x52, 0xb0, 0xa3, 0x39, 0x5e, 0x82, 0x6c,
+ 0xdf, 0xf5, 0x5c, 0x5a, 0x49, 0x2f, 0xa3, 0x95, 0x79, 0x5b, 0x4c, 0xf0, 0x75, 0xc8, 0x86, 0xd4,
+ 0x09, 0x68, 0x25, 0xb3, 0x8c, 0x56, 0x8a, 0xab, 0xd5, 0xba, 0x50, 0xbf, 0xae, 0xd4, 0xaf, 0xdf,
+ 0x57, 0xea, 0x37, 0xf3, 0x4f, 0xc6, 0x66, 0xea, 0xf3, 0x7f, 0x98, 0xc8, 0x16, 0x47, 0xf0, 0xf7,
+ 0x20, 0x43, 0x06, 0xed, 0x8a, 0x31, 0xc3, 0x49, 0x76, 0x00, 0x5f, 0x85, 0x42, 0xdb, 0x0d, 0x48,
+ 0x8b, 0xba, 0xfe, 0xa0, 0x92, 0x5d, 0x46, 0x2b, 0x0b, 0xab, 0xc7, 0x63, 0x48, 0x36, 0xd4, 0x92,
+ 0x1d, 0xef, 0xc2, 0x97, 0x21, 0x17, 0xf6, 0x9c, 0xa0, 0x1d, 0x56, 0xe6, 0x96, 0x33, 0x2b, 0x85,
+ 0xe6, 0xd2, 0xe1, 0xd8, 0x2c, 0x0b, 0xca, 0x65, 0xdf, 0x73, 0x29, 0xf1, 0x86, 0x74, 0xdf, 0x96,
+ 0x7b, 0xf0, 0x45, 0x98, 0x6b, 0x93, 0x3e, 0xa1, 0x24, 0xac, 0xe4, 0x39, 0xe2, 0x65, 0x8d, 0x3d,
+ 0x5f, 0xb0, 0xd5, 0x86, 0x3b, 0x46, 0x3e, 0x57, 0x9e, 0xb3, 0xfe, 0x8d, 0x00, 0xef, 0x38, 0xde,
+ 0xb0, 0x4f, 0x5e, 0x1a, 0xcf, 0x08, 0xb9, 0xf4, 0x2b, 0x23, 0x97, 0x99, 0x15, 0xb9, 0x18, 0x06,
+ 0x63, 0x36, 0x18, 0xb2, 0x5f, 0x03, 0x83, 0xb5, 0x05, 0x39, 0x41, 0xfa, 0x3a, 0x1b, 0x8a, 0x75,
+ 0xce, 0x28, 0x6d, 0xca, 0xb1, 0x36, 0x19, 0x7e, 0x4f, 0xeb, 0xa7, 0x30, 0x2f, 0x71, 0x14, 0x96,
+ 0x8a, 0xd7, 0x5e, 0xda, 0x07, 0x16, 0x9e, 0x8c, 0x4d, 0x14, 0xfb, 0x41, 0x64, 0xfc, 0xf8, 0x12,
+ 0x97, 0x4d, 0x43, 0x89, 0xf7, 0x62, 0x5d, 0xb8, 0xdc, 0xe6, 0xa0, 0x4b, 0x42, 0x76, 0xd0, 0x60,
+ 0x50, 0xd9, 0x62, 0x8f, 0xf5, 0x13, 0x38, 0x3e, 0xf1, 0x9c, 0xf2, 0x1a, 0xd7, 0x20, 0x17, 0x92,
+ 0xc0, 0x25, 0xea, 0x16, 0x1a, 0x20, 0x3b, 0x9c, 0xae, 0x89, 0xe7, 0x73, 0x5b, 0xee, 0x9f, 0x4d,
+ 0xfa, 0xef, 0x11, 0x94, 0xb6, 0x9c, 0x5d, 0xd2, 0x57, 0x76, 0x84, 0xc1, 0x18, 0x38, 0x1e, 0x91,
+ 0x78, 0xf2, 0x31, 0x3e, 0x09, 0xb9, 0x4f, 0x9c, 0xfe, 0x88, 0x08, 0x96, 0x79, 0x5b, 0xce, 0x66,
+ 0xf5, 0x48, 0xf4, 0xca, 0x1e, 0x89, 0x22, 0xbb, 0xb2, 0x2e, 0xc1, 0x31, 0xdb, 0xa1, 0x64, 0x8b,
+ 0x85, 0x04, 0xd2, 0x16, 0xc8, 0xb3, 0x0b, 0xf6, 0x99, 0x12, 0xa1, 0xbc, 0xb6, 0x9c, 0x59, 0x17,
+ 0x60, 0x5e, 0x2a, 0x27, 0x51, 0x8d, 0x35, 0x61, 0xa8, 0x16, 0x94, 0x26, 0xd6, 0x2f, 0x10, 0xcc,
+ 0x4f, 0x3c, 0x2e, 0xb6, 0x26, 0x59, 0x36, 0xe1, 0x70, 0x6c, 0x4a, 0x8a, 0x62, 0xcf, 0x4c, 0x85,
+ 0x0c, 0x28, 0x7f, 0xa4, 0x34, 0x7f, 0xa4, 0x93, 0xf1, 0x23, 0xbd, 0x37, 0xa0, 0xc1, 0xbe, 0xb2,
+ 0x94, 0x45, 0x06, 0x39, 0x8b, 0x93, 0x72, 0xbb, 0xad, 0x06, 0xf8, 0x0d, 0x30, 0x7a, 0x4e, 0xd8,
+ 0xe3, 0x08, 0x1a, 0xcd, 0xec, 0xe1, 0xd8, 0x44, 0x57, 0x6c, 0x4e, 0xb2, 0x3e, 0x81, 0x92, 0xce,
+ 0x04, 0xdf, 0x86, 0x42, 0x14, 0xe0, 0xf9, 0xa5, 0xfe, 0x3b, 0x6e, 0x0b, 0x52, 0x66, 0x9a, 0x86,
+ 0x1c, 0xbd, 0xf8, 0x30, 0x3e, 0x03, 0x46, 0xdf, 0x1d, 0x10, 0xfe, 0x9a, 0x85, 0x66, 0xfe, 0x70,
+ 0x6c, 0xf2, 0xb9, 0xcd, 0xff, 0x5a, 0x1e, 0xe4, 0x84, 0x41, 0xe2, 0x37, 0x93, 0x12, 0x33, 0xcd,
+ 0x9c, 0xe0, 0xa8, 0x73, 0x33, 0x21, 0xcb, 0x51, 0xe4, 0xec, 0x50, 0xb3, 0x70, 0x38, 0x36, 0x05,
+ 0xc1, 0x16, 0x3f, 0x4c, 0x9c, 0xa6, 0x23, 0x17, 0xc7, 0xe6, 0x52, 0xcd, 0x5b, 0x50, 0xda, 0x22,
+ 0x5d, 0xa7, 0xb5, 0x2f, 0x85, 0x2e, 0x29, 0x76, 0x4c, 0x20, 0x52, 0x3c, 0xce, 0x41, 0x29, 0x92,
+ 0xf8, 0xd0, 0x0b, 0xa5, 0x57, 0x17, 0x23, 0xda, 0xfb, 0xa1, 0xf5, 0x2b, 0x04, 0xd2, 0x15, 0x5e,
+ 0xea, 0xf1, 0x6e, 0xc0, 0x5c, 0xc8, 0x25, 0xaa, 0xc7, 0xd3, 0x3d, 0x8c, 0x2f, 0xc4, 0xcf, 0x26,
+ 0x37, 0xda, 0x6a, 0x80, 0xeb, 0x00, 0xc2, 0xd9, 0x6f, 0xc7, 0x8a, 0x2d, 0x1c, 0x8e, 0x4d, 0x8d,
+ 0x6a, 0x6b, 0x63, 0xeb, 0x97, 0x08, 0x8a, 0xf7, 0x1d, 0x37, 0xf2, 0xb2, 0x25, 0xc8, 0x3e, 0x62,
+ 0xee, 0x2e, 0xed, 0x55, 0x4c, 0x58, 0x3c, 0x6b, 0x93, 0xbe, 0xb3, 0x7f, 0xd3, 0x0f, 0x38, 0xcf,
+ 0x79, 0x3b, 0x9a, 0xc7, 0x39, 0xd1, 0x98, 0x9a, 0x13, 0xb3, 0x33, 0x47, 0xf6, 0x3b, 0x46, 0x3e,
+ 0x5d, 0xce, 0x58, 0x3f, 0x47, 0x50, 0x12, 0x37, 0x93, 0x2e, 0x72, 0x03, 0x72, 0xe2, 0xe2, 0xd2,
+ 0xc6, 0x8e, 0x0c, 0x7f, 0xa0, 0x85, 0x3e, 0x79, 0x04, 0xff, 0x10, 0x16, 0xda, 0x81, 0x3f, 0x1c,
+ 0x2a, 0xcf, 0x54, 0xd8, 0x6a, 0x4c, 0x36, 0xf4, 0x75, 0x3b, 0xb1, 0xdd, 0xfa, 0x1b, 0x73, 0x44,
+ 0x11, 0xcf, 0x24, 0x54, 0x91, 0x8a, 0xe8, 0x95, 0x93, 0x57, 0x7a, 0xd6, 0xe4, 0x75, 0x12, 0x72,
+ 0xdd, 0xc0, 0x1f, 0x0d, 0xc3, 0x4a, 0x46, 0x84, 0x09, 0x31, 0x9b, 0x2d, 0xa9, 0x59, 0x77, 0x60,
+ 0x41, 0xa9, 0x72, 0x44, 0x50, 0xaf, 0x26, 0x83, 0xfa, 0x66, 0x9b, 0x0c, 0xa8, 0xdb, 0x71, 0xa3,
+ 0x30, 0x2d, 0xf7, 0x5b, 0x9f, 0x21, 0x28, 0x27, 0xb7, 0xe0, 0x1f, 0x68, 0x66, 0xce, 0xd8, 0x9d,
+ 0x3f, 0x9a, 0x5d, 0x9d, 0xc7, 0xc1, 0x90, 0x07, 0x14, 0xe5, 0x02, 0xd5, 0x77, 0xa0, 0xa8, 0x91,
+ 0x59, 0x72, 0xdc, 0x23, 0xca, 0x24, 0xd9, 0x30, 0xf6, 0xc5, 0xb4, 0x30, 0x53, 0x3e, 0xb9, 0x9e,
+ 0xbe, 0x86, 0x98, 0x41, 0xcf, 0x4f, 0xbc, 0x24, 0xbe, 0x06, 0x46, 0x27, 0xf0, 0xbd, 0x99, 0x9e,
+ 0x89, 0x9f, 0xc0, 0xdf, 0x81, 0x34, 0xf5, 0x67, 0x7a, 0xa4, 0x34, 0xf5, 0xb5, 0x98, 0x9f, 0x99,
+ 0x88, 0xf9, 0xbf, 0x43, 0xb0, 0xc8, 0xce, 0x08, 0x04, 0xd6, 0x7b, 0xa3, 0xc1, 0x1e, 0x5e, 0x81,
+ 0x32, 0x93, 0xf4, 0xd0, 0x95, 0x39, 0xf0, 0xa1, 0xdb, 0x96, 0x6a, 0x2e, 0x30, 0xba, 0x4a, 0x8d,
+ 0x9b, 0x6d, 0x7c, 0x0a, 0xe6, 0x46, 0xa1, 0xd8, 0x20, 0x74, 0xce, 0xb1, 0xe9, 0x66, 0x1b, 0x5f,
+ 0xd2, 0xc4, 0x31, 0xac, 0xb5, 0x32, 0x90, 0x63, 0xb8, 0xed, 0xb8, 0x41, 0x14, 0x5b, 0x2e, 0x40,
+ 0xae, 0xc5, 0x04, 0x0b, 0x3b, 0x61, 0x39, 0x38, 0xda, 0xcc, 0x2f, 0x64, 0xcb, 0x65, 0xeb, 0xbb,
+ 0x50, 0x88, 0x4e, 0x4f, 0x4d, 0xbd, 0x53, 0x5f, 0xc0, 0xba, 0x01, 0x8b, 0x22, 0x66, 0x4e, 0x3f,
+ 0x5c, 0x9a, 0x76, 0xb8, 0xa4, 0x0e, 0x9f, 0x86, 0xac, 0x40, 0x05, 0x83, 0xd1, 0x76, 0xa8, 0xa3,
+ 0x8e, 0xb0, 0xb1, 0x55, 0x81, 0x93, 0xf7, 0x03, 0x67, 0x10, 0x76, 0x48, 0xc0, 0x37, 0x45, 0xb6,
+ 0x6b, 0x9d, 0x80, 0xe3, 0x2c, 0x4e, 0x90, 0x20, 0x5c, 0xf7, 0x47, 0x03, 0x2a, 0xdd, 0xd3, 0xba,
+ 0x0c, 0x4b, 0x93, 0x64, 0x69, 0xea, 0x4b, 0x90, 0x6d, 0x31, 0x02, 0xe7, 0x3e, 0x6f, 0x8b, 0x89,
+ 0xf5, 0x6b, 0x04, 0xf8, 0x16, 0xa1, 0x9c, 0xf5, 0xe6, 0x46, 0xa8, 0x15, 0xaf, 0x9e, 0x43, 0x5b,
+ 0x3d, 0x12, 0xa8, 0x0c, 0x1e, 0xcd, 0xff, 0x1f, 0xc5, 0xab, 0x75, 0x15, 0x8e, 0x4f, 0xdc, 0x52,
+ 0xea, 0x54, 0x85, 0x7c, 0x4b, 0xd2, 0x64, 0xfd, 0x10, 0xcd, 0xad, 0x3f, 0xa4, 0x21, 0x2f, 0xde,
+ 0x96, 0x74, 0xf0, 0x55, 0x28, 0x76, 0x98, 0xad, 0x05, 0xc3, 0xc0, 0x95, 0x10, 0x18, 0xcd, 0xc5,
+ 0xc3, 0xb1, 0xa9, 0x93, 0x6d, 0x7d, 0x82, 0xaf, 0x24, 0x0c, 0xaf, 0xb9, 0x74, 0x30, 0x36, 0x73,
+ 0x3f, 0x62, 0xc6, 0xb7, 0xc1, 0xb2, 0x17, 0x37, 0xc3, 0x8d, 0xc8, 0x1c, 0xef, 0x4a, 0x6f, 0xe3,
+ 0x95, 0x6c, 0xf3, 0x6d, 0x76, 0xfd, 0x2f, 0xc7, 0xe6, 0x05, 0xad, 0x81, 0x1c, 0x06, 0xbe, 0x47,
+ 0x68, 0x8f, 0x8c, 0xc2, 0x46, 0xcb, 0xf7, 0x3c, 0x7f, 0xd0, 0xe0, 0x4d, 0x20, 0x57, 0x9a, 0xa5,
+ 0x60, 0x76, 0x5c, 0x3a, 0xe0, 0x7d, 0x98, 0xa3, 0xbd, 0xc0, 0x1f, 0x75, 0x7b, 0x3c, 0xbb, 0x64,
+ 0x9a, 0xd7, 0x67, 0xe7, 0xa7, 0x38, 0xd8, 0x6a, 0x80, 0xcf, 0x31, 0xb4, 0x48, 0x6b, 0x2f, 0x1c,
+ 0x79, 0x3c, 0x3d, 0xcd, 0xab, 0xf2, 0x26, 0x22, 0x5b, 0x9f, 0xa5, 0xc1, 0xe4, 0x26, 0xfc, 0x80,
+ 0x97, 0x61, 0x37, 0xfd, 0xe0, 0x7d, 0x42, 0x03, 0xb7, 0x75, 0xcf, 0xf1, 0x88, 0xb2, 0x0d, 0x13,
+ 0x8a, 0x1e, 0x27, 0x3e, 0xd4, 0x9c, 0x03, 0xbc, 0x68, 0x1f, 0x3e, 0x0b, 0xc0, 0xdd, 0x4e, 0xac,
+ 0x0b, 0x3f, 0x29, 0x70, 0x0a, 0x5f, 0x5e, 0x9f, 0x40, 0xaa, 0x31, 0xa3, 0x66, 0x12, 0xa1, 0xcd,
+ 0x24, 0x42, 0x33, 0xf3, 0x89, 0x60, 0xd1, 0x6d, 0x3d, 0x3b, 0x69, 0xeb, 0xd6, 0xdf, 0x11, 0xd4,
+ 0xb6, 0xd4, 0xcd, 0x5f, 0x11, 0x0e, 0xa5, 0x6f, 0xfa, 0x35, 0xe9, 0x9b, 0xf9, 0xdf, 0xf4, 0xb5,
+ 0xfe, 0xa2, 0xb9, 0xbc, 0x4d, 0x3a, 0x4a, 0x8f, 0x75, 0x2d, 0x5d, 0xbc, 0x8e, 0x6b, 0xa6, 0x5f,
+ 0xe3, 0xb3, 0x64, 0x12, 0xcf, 0xf2, 0x6e, 0x1c, 0x0e, 0xb8, 0x06, 0x32, 0x1c, 0x9c, 0x07, 0x23,
+ 0x20, 0x1d, 0x95, 0x7c, 0x71, 0x32, 0xc6, 0x93, 0x8e, 0xcd, 0xd7, 0xad, 0x3f, 0x21, 0x28, 0xdf,
+ 0x22, 0x74, 0xb2, 0xac, 0xf9, 0x26, 0xe9, 0x7f, 0x1b, 0x8e, 0x69, 0xf7, 0x97, 0xda, 0xbf, 0x95,
+ 0xa8, 0x65, 0x4e, 0xc4, 0xfa, 0x6f, 0x0e, 0xda, 0xe4, 0x53, 0xd9, 0xa5, 0x4e, 0x96, 0x31, 0xdb,
+ 0x50, 0xd4, 0x16, 0xf1, 0x5a, 0xa2, 0x80, 0x99, 0x96, 0x54, 0x9b, 0x4b, 0x52, 0x27, 0xd1, 0xa7,
+ 0xca, 0xea, 0x33, 0x4a, 0xf7, 0x3b, 0x80, 0x79, 0xe3, 0xcc, 0xd9, 0xea, 0x91, 0x9a, 0x53, 0xef,
+ 0x46, 0xf5, 0x4c, 0x34, 0xc7, 0xe7, 0xc0, 0x08, 0xfc, 0xc7, 0xaa, 0x32, 0x9d, 0x8f, 0x45, 0xda,
+ 0xfe, 0x63, 0x9b, 0x2f, 0x59, 0x37, 0x20, 0x63, 0xfb, 0x8f, 0x71, 0x0d, 0x20, 0x70, 0x06, 0x5d,
+ 0xf2, 0x20, 0xea, 0x47, 0x4a, 0xb6, 0x46, 0x39, 0x22, 0xbf, 0xae, 0xc3, 0x31, 0xfd, 0x46, 0xe2,
+ 0xb9, 0xeb, 0x30, 0xc7, 0x88, 0x31, 0x5c, 0x4b, 0x09, 0xb8, 0x44, 0xf7, 0xaf, 0x36, 0x31, 0x9b,
+ 0x81, 0x98, 0x8e, 0xcf, 0x40, 0x81, 0x3a, 0xbb, 0x7d, 0x72, 0x2f, 0xf6, 0xf9, 0x98, 0xc0, 0x56,
+ 0x59, 0x2b, 0xf5, 0x40, 0x2b, 0x14, 0x62, 0x02, 0xbe, 0x08, 0xe5, 0xf8, 0xce, 0xdb, 0x01, 0xe9,
+ 0xb8, 0x9f, 0xf2, 0x17, 0x2e, 0xd9, 0x2f, 0xd0, 0xf1, 0x0a, 0x2c, 0xc6, 0xb4, 0x1d, 0x9e, 0x76,
+ 0x0d, 0xbe, 0x35, 0x49, 0x66, 0xd8, 0x70, 0x75, 0xdf, 0x7b, 0x34, 0x72, 0xfa, 0x3c, 0x90, 0x95,
+ 0x6c, 0x8d, 0x62, 0xfd, 0x19, 0xc1, 0x31, 0xf1, 0xd4, 0xd4, 0xa1, 0xdf, 0x48, 0xab, 0xff, 0x0d,
+ 0x02, 0xac, 0x6b, 0x20, 0x4d, 0xeb, 0x5b, 0xfa, 0xf7, 0x21, 0x96, 0xd7, 0x8b, 0xd3, 0x3e, 0x80,
+ 0xb2, 0x16, 0x54, 0x96, 0x80, 0x69, 0xbe, 0x8b, 0xb7, 0xa0, 0x82, 0xa2, 0xaa, 0x3f, 0xd6, 0x39,
+ 0xef, 0xee, 0x53, 0x12, 0xca, 0x06, 0x92, 0x77, 0xce, 0x9c, 0x60, 0x8b, 0x1f, 0x26, 0x4b, 0x7d,
+ 0x60, 0x30, 0x62, 0x59, 0xc9, 0x8f, 0x08, 0x17, 0xcf, 0x43, 0x21, 0xfa, 0x14, 0x89, 0x8b, 0x30,
+ 0x77, 0xf3, 0x03, 0xfb, 0xa3, 0x35, 0x7b, 0xa3, 0x9c, 0xc2, 0x25, 0xc8, 0x37, 0xd7, 0xd6, 0xef,
+ 0xf2, 0x19, 0x5a, 0x5d, 0x83, 0xdc, 0xf6, 0x28, 0xec, 0x91, 0x00, 0xbf, 0x0d, 0x06, 0x1b, 0x61,
+ 0xcd, 0x69, 0xb5, 0xef, 0xc0, 0xd5, 0x93, 0x49, 0xb2, 0xac, 0x01, 0x53, 0xab, 0x7f, 0x34, 0x94,
+ 0x21, 0x07, 0xf8, 0xfb, 0x90, 0x15, 0xd6, 0xa9, 0x6d, 0xd7, 0xbf, 0x49, 0x56, 0x4f, 0xbd, 0x40,
+ 0x57, 0x7c, 0xbe, 0x8d, 0xf0, 0x3d, 0x28, 0x72, 0xa2, 0x6c, 0xfb, 0xcf, 0x24, 0xbb, 0xef, 0x09,
+ 0x4e, 0x67, 0x8f, 0x58, 0xd5, 0xf8, 0x5d, 0x87, 0x2c, 0x0f, 0x10, 0xfa, 0x6d, 0xf4, 0x2f, 0x5b,
+ 0xfa, 0x6d, 0x26, 0x3e, 0x0a, 0x59, 0x29, 0xfc, 0x0e, 0x18, 0xac, 0x88, 0xd5, 0xe1, 0xd0, 0xba,
+ 0x75, 0x1d, 0x0e, 0xbd, 0x55, 0xe6, 0x62, 0xdf, 0x8d, 0x3e, 0x3a, 0x9c, 0x4a, 0x76, 0x5f, 0xea,
+ 0x78, 0xe5, 0xc5, 0x85, 0x48, 0xf2, 0x07, 0xa2, 0xfb, 0x56, 0xe5, 0x33, 0x3e, 0x3b, 0x29, 0x2a,
+ 0x51, 0x6d, 0x57, 0x6b, 0x47, 0x2d, 0x47, 0x0c, 0xb7, 0xa0, 0xa8, 0x95, 0xae, 0x3a, 0xac, 0x2f,
+ 0xd6, 0xdd, 0x3a, 0xac, 0x53, 0xea, 0x5d, 0x2b, 0x85, 0x6f, 0x41, 0x9e, 0x45, 0x7e, 0xe6, 0x00,
+ 0xf8, 0x74, 0x32, 0xc0, 0x6b, 0x8e, 0x5d, 0x3d, 0x33, 0x7d, 0x31, 0xb2, 0x9b, 0x1f, 0x43, 0x5e,
+ 0x75, 0x59, 0xf8, 0x43, 0x58, 0x98, 0xec, 0x31, 0xf0, 0x1b, 0x9a, 0x5a, 0x93, 0xad, 0x5b, 0x75,
+ 0x59, 0x5b, 0x9a, 0xde, 0x98, 0xa4, 0x56, 0x50, 0xf3, 0xe3, 0xa7, 0xcf, 0x6a, 0xa9, 0x2f, 0x9e,
+ 0xd5, 0x52, 0x5f, 0x3d, 0xab, 0xa1, 0x9f, 0x1d, 0xd4, 0xd0, 0x6f, 0x0f, 0x6a, 0xe8, 0xc9, 0x41,
+ 0x0d, 0x3d, 0x3d, 0xa8, 0xa1, 0x7f, 0x1e, 0xd4, 0xd0, 0xbf, 0x0e, 0x6a, 0xa9, 0xaf, 0x0e, 0x6a,
+ 0xe8, 0xf3, 0xe7, 0xb5, 0xd4, 0xd3, 0xe7, 0xb5, 0xd4, 0x17, 0xcf, 0x6b, 0xa9, 0x8f, 0xdf, 0xd4,
+ 0xff, 0xdf, 0x12, 0x38, 0x1d, 0x67, 0xe0, 0x34, 0xfa, 0xfe, 0x9e, 0xdb, 0xd0, 0xff, 0x5d, 0xb3,
+ 0x9b, 0xe3, 0x3f, 0x6f, 0xfd, 0x27, 0x00, 0x00, 0xff, 0xff, 0x20, 0xae, 0x71, 0xe2, 0xc5, 0x19,
+ 0x00, 0x00,
}
func (x Direction) String() string {
@@ -2779,6 +2824,30 @@ func (this *LabelRequest) Equal(that interface{}) bool {
}
return true
}
+func (this *RateLimitedStream) Equal(that interface{}) bool {
+ if that == nil {
+ return this == nil
+ }
+
+ that1, ok := that.(*RateLimitedStream)
+ if !ok {
+ that2, ok := that.(RateLimitedStream)
+ if ok {
+ that1 = &that2
+ } else {
+ return false
+ }
+ }
+ if that1 == nil {
+ return this == nil
+ } else if this == nil {
+ return false
+ }
+ if this.Labels != that1.Labels {
+ return false
+ }
+ return true
+}
func (this *LabelResponse) Equal(that interface{}) bool {
if that == nil {
return this == nil
@@ -3944,6 +4013,16 @@ func (this *LabelRequest) GoString() string {
s = append(s, "}")
return strings.Join(s, "")
}
+func (this *RateLimitedStream) GoString() string {
+ if this == nil {
+ return "nil"
+ }
+ s := make([]string, 0, 5)
+ s = append(s, "&logproto.RateLimitedStream{")
+ s = append(s, "Labels: "+fmt.Sprintf("%#v", this.Labels)+",\n")
+ s = append(s, "}")
+ return strings.Join(s, "")
+}
func (this *LabelResponse) GoString() string {
if this == nil {
return "nil"
@@ -5396,6 +5475,36 @@ func (m *LabelRequest) MarshalToSizedBuffer(dAtA []byte) (int, error) {
return len(dAtA) - i, nil
}
+func (m *RateLimitedStream) Marshal() (dAtA []byte, err error) {
+ size := m.Size()
+ dAtA = make([]byte, size)
+ n, err := m.MarshalToSizedBuffer(dAtA[:size])
+ if err != nil {
+ return nil, err
+ }
+ return dAtA[:n], nil
+}
+
+func (m *RateLimitedStream) MarshalTo(dAtA []byte) (int, error) {
+ size := m.Size()
+ return m.MarshalToSizedBuffer(dAtA[:size])
+}
+
+func (m *RateLimitedStream) MarshalToSizedBuffer(dAtA []byte) (int, error) {
+ i := len(dAtA)
+ _ = i
+ var l int
+ _ = l
+ if len(m.Labels) > 0 {
+ i -= len(m.Labels)
+ copy(dAtA[i:], m.Labels)
+ i = encodeVarintLogproto(dAtA, i, uint64(len(m.Labels)))
+ i--
+ dAtA[i] = 0xa
+ }
+ return len(dAtA) - i, nil
+}
+
func (m *LabelResponse) Marshal() (dAtA []byte, err error) {
size := m.Size()
dAtA = make([]byte, size)
@@ -7007,6 +7116,19 @@ func (m *LabelRequest) Size() (n int) {
return n
}
+func (m *RateLimitedStream) Size() (n int) {
+ if m == nil {
+ return 0
+ }
+ var l int
+ _ = l
+ l = len(m.Labels)
+ if l > 0 {
+ n += 1 + l + sovLogproto(uint64(l))
+ }
+ return n
+}
+
func (m *LabelResponse) Size() (n int) {
if m == nil {
return 0
@@ -7752,6 +7874,16 @@ func (this *LabelRequest) String() string {
}, "")
return s
}
+func (this *RateLimitedStream) String() string {
+ if this == nil {
+ return "nil"
+ }
+ s := strings.Join([]string{`&RateLimitedStream{`,
+ `Labels:` + fmt.Sprintf("%v", this.Labels) + `,`,
+ `}`,
+ }, "")
+ return s
+}
func (this *LabelResponse) String() string {
if this == nil {
return "nil"
@@ -9369,6 +9501,91 @@ func (m *LabelRequest) Unmarshal(dAtA []byte) error {
}
return nil
}
+func (m *RateLimitedStream) Unmarshal(dAtA []byte) error {
+ l := len(dAtA)
+ iNdEx := 0
+ for iNdEx < l {
+ preIndex := iNdEx
+ var wire uint64
+ for shift := uint(0); ; shift += 7 {
+ if shift >= 64 {
+ return ErrIntOverflowLogproto
+ }
+ if iNdEx >= l {
+ return io.ErrUnexpectedEOF
+ }
+ b := dAtA[iNdEx]
+ iNdEx++
+ wire |= uint64(b&0x7F) << shift
+ if b < 0x80 {
+ break
+ }
+ }
+ fieldNum := int32(wire >> 3)
+ wireType := int(wire & 0x7)
+ if wireType == 4 {
+ return fmt.Errorf("proto: RateLimitedStream: wiretype end group for non-group")
+ }
+ if fieldNum <= 0 {
+ return fmt.Errorf("proto: RateLimitedStream: illegal tag %d (wire type %d)", fieldNum, wire)
+ }
+ switch fieldNum {
+ case 1:
+ if wireType != 2 {
+ return fmt.Errorf("proto: wrong wireType = %d for field Labels", wireType)
+ }
+ var stringLen uint64
+ for shift := uint(0); ; shift += 7 {
+ if shift >= 64 {
+ return ErrIntOverflowLogproto
+ }
+ if iNdEx >= l {
+ return io.ErrUnexpectedEOF
+ }
+ b := dAtA[iNdEx]
+ iNdEx++
+ stringLen |= uint64(b&0x7F) << shift
+ if b < 0x80 {
+ break
+ }
+ }
+ intStringLen := int(stringLen)
+ if intStringLen < 0 {
+ return ErrInvalidLengthLogproto
+ }
+ postIndex := iNdEx + intStringLen
+ if postIndex < 0 {
+ return ErrInvalidLengthLogproto
+ }
+ if postIndex > l {
+ return io.ErrUnexpectedEOF
+ }
+ m.Labels = string(dAtA[iNdEx:postIndex])
+ iNdEx = postIndex
+ default:
+ iNdEx = preIndex
+ skippy, err := skipLogproto(dAtA[iNdEx:])
+ if err != nil {
+ return err
+ }
+ if skippy < 0 {
+ return ErrInvalidLengthLogproto
+ }
+ if (iNdEx + skippy) < 0 {
+ return ErrInvalidLengthLogproto
+ }
+ if (iNdEx + skippy) > l {
+ return io.ErrUnexpectedEOF
+ }
+ iNdEx += skippy
+ }
+ }
+
+ if iNdEx > l {
+ return io.ErrUnexpectedEOF
+ }
+ return nil
+}
func (m *LabelResponse) Unmarshal(dAtA []byte) error {
l := len(dAtA)
iNdEx := 0
diff --git a/pkg/logproto/logproto.proto b/pkg/logproto/logproto.proto
index ad9631c1e09d9..ca11a736f928d 100644
--- a/pkg/logproto/logproto.proto
+++ b/pkg/logproto/logproto.proto
@@ -116,6 +116,10 @@ message LabelRequest {
];
}
+message RateLimitedStream {
+ string labels = 1;
+}
+
message LabelResponse {
repeated string values = 1;
}
|
loki
|
Modify ingesters to return rate-limited streams to distributors (#6977)
|
c89df1a795b8eaf6f3c765db85e86f1a20f16deb
|
2019-11-27 00:03:15
|
Putra Sattvika, I Gusti Ngurah
|
storage: fix missing logs with batched chunk iterator (#1299)
| false
|
diff --git a/pkg/storage/iterator.go b/pkg/storage/iterator.go
index 229a89d5b3055..5acf5ea25d8e8 100644
--- a/pkg/storage/iterator.go
+++ b/pkg/storage/iterator.go
@@ -77,6 +77,16 @@ type batchChunkIterator struct {
// newBatchChunkIterator creates a new batch iterator with the given batchSize.
func newBatchChunkIterator(ctx context.Context, chunks []*chunkenc.LazyChunk, batchSize int, matchers []*labels.Matcher, filter logql.Filter, req *logproto.QueryRequest) *batchChunkIterator {
+
+ // __name__ is not something we filter by because it's a constant in loki and only used for upstream compatibility.
+ // Therefore remove it
+ for i := range matchers {
+ if matchers[i].Name == labels.MetricName {
+ matchers = append(matchers[:i], matchers[i+1:]...)
+ break
+ }
+ }
+
res := &batchChunkIterator{
batchSize: batchSize,
matchers: matchers,
@@ -112,6 +122,9 @@ func (it *batchChunkIterator) Next() bool {
}
func (it *batchChunkIterator) nextBatch() (iter.EntryIterator, error) {
+ // the first chunk of the batch
+ headChunk := it.chunks.Peek()
+
// pop the next batch of chunks and append/preprend previous overlapping chunks
// so we can merge/de-dupe overlapping entries.
batch := make([]*chunkenc.LazyChunk, 0, it.batchSize+len(it.lastOverlapping))
@@ -130,6 +143,14 @@ func (it *batchChunkIterator) nextBatch() (iter.EntryIterator, error) {
// so that overlapping chunks are together
if it.req.Direction == logproto.BACKWARD {
from = time.Unix(0, nextChunk.Chunk.Through.UnixNano())
+
+ // we have to reverse the inclusivity of the chunk iterator from
+ // [from, through) to (from, through] for backward queries, except when
+ // the batch's `from` is equal to the query's Start. This can be achieved
+ // by shifting `from` by one nanosecond.
+ if !from.Equal(it.req.Start) {
+ from = from.Add(time.Nanosecond)
+ }
} else {
through = time.Unix(0, nextChunk.Chunk.From.UnixNano())
}
@@ -149,7 +170,7 @@ func (it *batchChunkIterator) nextBatch() (iter.EntryIterator, error) {
// └────────────────────┘
//
// And nextChunk is # 49, we need to keep references to #47 and #48 as they won't be
- // iterated over completely (we're clipping through to #49's from) and then add them to the next batch.
+ // iterated over completely (we're clipping through to #49's from) and then add them to the next batch.
it.lastOverlapping = it.lastOverlapping[:0]
for _, c := range batch {
if it.req.Direction == logproto.BACKWARD {
@@ -162,13 +183,27 @@ func (it *batchChunkIterator) nextBatch() (iter.EntryIterator, error) {
}
}
}
+ }
+
+ if it.req.Direction == logproto.BACKWARD {
+ through = time.Unix(0, headChunk.Chunk.Through.UnixNano())
+
+ if through.After(it.req.End) {
+ through = it.req.End
+ }
+
+ // we have to reverse the inclusivity of the chunk iterator from
+ // [from, through) to (from, through] for backward queries, except when
+ // the batch's `through` is equal to the query's End. This can be achieved
+ // by shifting `through` by one nanosecond.
+ if !through.Equal(it.req.End) {
+ through = through.Add(time.Nanosecond)
+ }
} else {
- if len(it.lastOverlapping) > 0 {
- if it.req.Direction == logproto.BACKWARD {
- through = time.Unix(0, it.lastOverlapping[0].Chunk.From.UnixNano())
- } else {
- from = time.Unix(0, it.lastOverlapping[0].Chunk.Through.UnixNano())
- }
+ from = time.Unix(0, headChunk.Chunk.From.UnixNano())
+
+ if from.Before(it.req.Start) {
+ from = it.req.Start
}
}
@@ -250,12 +285,9 @@ func buildIterators(ctx context.Context, chks map[model.Fingerprint][][]*chunken
func buildHeapIterator(ctx context.Context, chks [][]*chunkenc.LazyChunk, filter logql.Filter, direction logproto.Direction, from, through time.Time) (iter.EntryIterator, error) {
result := make([]iter.EntryIterator, 0, len(chks))
- if chks[0][0].Chunk.Metric.Has("__name__") {
- labelsBuilder := labels.NewBuilder(chks[0][0].Chunk.Metric)
- labelsBuilder.Del("__name__")
- chks[0][0].Chunk.Metric = labelsBuilder.Labels()
- }
- labels := chks[0][0].Chunk.Metric.String()
+
+ // __name__ is only used for upstream compatibility and is hardcoded within loki. Strip it from the return label set.
+ labels := dropLabels(chks[0][0].Chunk.Metric, labels.MetricName).String()
for i := range chks {
iterators := make([]iter.EntryIterator, 0, len(chks[i]))
@@ -400,3 +432,20 @@ outer:
return css
}
+
+// dropLabels returns a new label set with certain labels dropped
+func dropLabels(ls labels.Labels, removals ...string) (dst labels.Labels) {
+ toDel := make(map[string]struct{})
+ for _, r := range removals {
+ toDel[r] = struct{}{}
+ }
+
+ for _, l := range ls {
+ _, remove := toDel[l.Name]
+ if !remove {
+ dst = append(dst, l)
+ }
+ }
+
+ return dst
+}
diff --git a/pkg/storage/iterator_test.go b/pkg/storage/iterator_test.go
index 8f04348148dcc..fac34f5fe77aa 100644
--- a/pkg/storage/iterator_test.go
+++ b/pkg/storage/iterator_test.go
@@ -2,12 +2,15 @@ package storage
import (
"context"
+ "fmt"
"testing"
"time"
"github.com/grafana/loki/pkg/chunkenc"
"github.com/grafana/loki/pkg/iter"
"github.com/grafana/loki/pkg/logproto"
+ "github.com/prometheus/prometheus/pkg/labels"
+ "github.com/stretchr/testify/require"
)
func Test_newBatchChunkIterator(t *testing.T) {
@@ -18,11 +21,12 @@ func Test_newBatchChunkIterator(t *testing.T) {
matchers string
start, end time.Time
direction logproto.Direction
+ batchSize int
}{
"forward with overlap": {
[]*chunkenc.LazyChunk{
newLazyChunk(logproto.Stream{
- Labels: fooLabels,
+ Labels: fooLabelsWithName,
Entries: []logproto.Entry{
{
Timestamp: from,
@@ -35,7 +39,7 @@ func Test_newBatchChunkIterator(t *testing.T) {
},
}),
newLazyChunk(logproto.Stream{
- Labels: fooLabels,
+ Labels: fooLabelsWithName,
Entries: []logproto.Entry{
{
Timestamp: from.Add(time.Millisecond),
@@ -48,7 +52,7 @@ func Test_newBatchChunkIterator(t *testing.T) {
},
}),
newLazyChunk(logproto.Stream{
- Labels: fooLabels,
+ Labels: fooLabelsWithName,
Entries: []logproto.Entry{
{
Timestamp: from.Add(time.Millisecond),
@@ -61,7 +65,120 @@ func Test_newBatchChunkIterator(t *testing.T) {
},
}),
newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from.Add(2 * time.Millisecond),
+ Line: "3",
+ },
+ {
+ Timestamp: from.Add(3 * time.Millisecond),
+ Line: "4",
+ },
+ },
+ }),
+ newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from.Add(2 * time.Millisecond),
+ Line: "3",
+ },
+ {
+ Timestamp: from.Add(3 * time.Millisecond),
+ Line: "4",
+ },
+ },
+ }),
+ newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from.Add(3 * time.Millisecond),
+ Line: "4",
+ },
+ {
+ Timestamp: from.Add(4 * time.Millisecond),
+ Line: "5",
+ },
+ },
+ }),
+ },
+ []*logproto.Stream{
+ {
Labels: fooLabels,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from,
+ Line: "1",
+ },
+ {
+ Timestamp: from.Add(time.Millisecond),
+ Line: "2",
+ },
+ {
+ Timestamp: from.Add(2 * time.Millisecond),
+ Line: "3",
+ },
+ {
+ Timestamp: from.Add(3 * time.Millisecond),
+ Line: "4",
+ },
+ },
+ },
+ },
+ fooLabelsWithName,
+ from, from.Add(4 * time.Millisecond),
+ logproto.FORWARD,
+ 2,
+ },
+ "forward with overlapping non-continuous entries": {
+ []*chunkenc.LazyChunk{
+ newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from,
+ Line: "1",
+ },
+ {
+ Timestamp: from.Add(time.Millisecond),
+ Line: "2",
+ },
+ {
+ Timestamp: from.Add(3 * time.Millisecond),
+ Line: "4",
+ },
+ },
+ }),
+ newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from.Add(time.Millisecond),
+ Line: "2",
+ },
+ {
+ Timestamp: from.Add(2 * time.Millisecond),
+ Line: "3",
+ },
+ },
+ }),
+ newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from.Add(time.Millisecond),
+ Line: "2",
+ },
+ {
+ Timestamp: from.Add(3 * time.Millisecond),
+ Line: "4",
+ },
+ },
+ }),
+ newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
Entries: []logproto.Entry{
{
Timestamp: from.Add(2 * time.Millisecond),
@@ -93,14 +210,15 @@ func Test_newBatchChunkIterator(t *testing.T) {
},
},
},
- fooLabels,
+ fooLabelsWithName,
from, from.Add(3 * time.Millisecond),
logproto.FORWARD,
+ 2,
},
"backward with overlap": {
[]*chunkenc.LazyChunk{
newLazyChunk(logproto.Stream{
- Labels: fooLabels,
+ Labels: fooLabelsWithName,
Entries: []logproto.Entry{
{
Timestamp: from,
@@ -113,7 +231,7 @@ func Test_newBatchChunkIterator(t *testing.T) {
},
}),
newLazyChunk(logproto.Stream{
- Labels: fooLabels,
+ Labels: fooLabelsWithName,
Entries: []logproto.Entry{
{
Timestamp: from.Add(time.Millisecond),
@@ -126,7 +244,7 @@ func Test_newBatchChunkIterator(t *testing.T) {
},
}),
newLazyChunk(logproto.Stream{
- Labels: fooLabels,
+ Labels: fooLabelsWithName,
Entries: []logproto.Entry{
{
Timestamp: from.Add(time.Millisecond),
@@ -139,7 +257,7 @@ func Test_newBatchChunkIterator(t *testing.T) {
},
}),
newLazyChunk(logproto.Stream{
- Labels: fooLabels,
+ Labels: fooLabelsWithName,
Entries: []logproto.Entry{
{
Timestamp: from.Add(2 * time.Millisecond),
@@ -151,11 +269,41 @@ func Test_newBatchChunkIterator(t *testing.T) {
},
},
}),
+ newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from.Add(2 * time.Millisecond),
+ Line: "3",
+ },
+ {
+ Timestamp: from.Add(3 * time.Millisecond),
+ Line: "4",
+ },
+ },
+ }),
+ newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from.Add(3 * time.Millisecond),
+ Line: "4",
+ },
+ {
+ Timestamp: from.Add(4 * time.Millisecond),
+ Line: "5",
+ },
+ },
+ }),
},
[]*logproto.Stream{
{
Labels: fooLabels,
Entries: []logproto.Entry{
+ {
+ Timestamp: from.Add(3 * time.Millisecond),
+ Line: "4",
+ },
{
Timestamp: from.Add(2 * time.Millisecond),
Line: "3",
@@ -171,14 +319,114 @@ func Test_newBatchChunkIterator(t *testing.T) {
},
},
},
- fooLabels,
- from, from.Add(3 * time.Millisecond),
+ fooLabelsWithName,
+ from, from.Add(4 * time.Millisecond),
logproto.BACKWARD,
+ 2,
},
- "forward without overlap": {
+ "backward with overlapping non-continuous entries": {
[]*chunkenc.LazyChunk{
newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from.Add(0 * time.Millisecond),
+ Line: "0",
+ },
+ {
+ Timestamp: from.Add(3 * time.Millisecond),
+ Line: "3",
+ },
+ },
+ }),
+ newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from.Add(1 * time.Millisecond),
+ Line: "1",
+ },
+ {
+ Timestamp: from.Add(6 * time.Millisecond),
+ Line: "6",
+ },
+ },
+ }),
+ newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from.Add(2 * time.Millisecond),
+ Line: "2",
+ },
+ {
+ Timestamp: from.Add(5 * time.Millisecond),
+ Line: "5",
+ },
+ },
+ }),
+ newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from.Add(4 * time.Millisecond),
+ Line: "4",
+ },
+ {
+ Timestamp: from.Add(7 * time.Millisecond),
+ Line: "7",
+ },
+ },
+ }),
+ },
+ []*logproto.Stream{
+ {
Labels: fooLabels,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from.Add(7 * time.Millisecond),
+ Line: "7",
+ },
+ {
+ Timestamp: from.Add(6 * time.Millisecond),
+ Line: "6",
+ },
+ {
+ Timestamp: from.Add(5 * time.Millisecond),
+ Line: "5",
+ },
+ {
+ Timestamp: from.Add(4 * time.Millisecond),
+ Line: "4",
+ },
+ {
+ Timestamp: from.Add(3 * time.Millisecond),
+ Line: "3",
+ },
+ {
+ Timestamp: from.Add(2 * time.Millisecond),
+ Line: "2",
+ },
+ {
+ Timestamp: from.Add(1 * time.Millisecond),
+ Line: "1",
+ },
+ {
+ Timestamp: from.Add(0 * time.Millisecond),
+ Line: "0",
+ },
+ },
+ },
+ },
+ fooLabelsWithName,
+ from, from.Add(8 * time.Millisecond),
+ logproto.BACKWARD,
+ 2,
+ },
+ "forward without overlap": {
+ []*chunkenc.LazyChunk{
+ newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
Entries: []logproto.Entry{
{
Timestamp: from,
@@ -191,7 +439,7 @@ func Test_newBatchChunkIterator(t *testing.T) {
},
}),
newLazyChunk(logproto.Stream{
- Labels: fooLabels,
+ Labels: fooLabelsWithName,
Entries: []logproto.Entry{
{
Timestamp: from.Add(2 * time.Millisecond),
@@ -200,7 +448,7 @@ func Test_newBatchChunkIterator(t *testing.T) {
},
}),
newLazyChunk(logproto.Stream{
- Labels: fooLabels,
+ Labels: fooLabelsWithName,
Entries: []logproto.Entry{
{
Timestamp: from.Add(3 * time.Millisecond),
@@ -228,14 +476,15 @@ func Test_newBatchChunkIterator(t *testing.T) {
},
},
},
- fooLabels,
+ fooLabelsWithName,
from, from.Add(3 * time.Millisecond),
logproto.FORWARD,
+ 2,
},
"backward without overlap": {
[]*chunkenc.LazyChunk{
newLazyChunk(logproto.Stream{
- Labels: fooLabels,
+ Labels: fooLabelsWithName,
Entries: []logproto.Entry{
{
Timestamp: from,
@@ -248,7 +497,7 @@ func Test_newBatchChunkIterator(t *testing.T) {
},
}),
newLazyChunk(logproto.Stream{
- Labels: fooLabels,
+ Labels: fooLabelsWithName,
Entries: []logproto.Entry{
{
Timestamp: from.Add(2 * time.Millisecond),
@@ -257,7 +506,7 @@ func Test_newBatchChunkIterator(t *testing.T) {
},
}),
newLazyChunk(logproto.Stream{
- Labels: fooLabels,
+ Labels: fooLabelsWithName,
Entries: []logproto.Entry{
{
Timestamp: from.Add(3 * time.Millisecond),
@@ -285,16 +534,17 @@ func Test_newBatchChunkIterator(t *testing.T) {
},
},
},
- fooLabels,
+ fooLabelsWithName,
from, from.Add(3 * time.Millisecond),
logproto.BACKWARD,
+ 2,
},
}
for name, tt := range tests {
tt := tt
t.Run(name, func(t *testing.T) {
- it := newBatchChunkIterator(context.Background(), tt.chunks, 2, newMatchers(tt.matchers), nil, newQuery("", tt.start, tt.end, tt.direction))
+ it := newBatchChunkIterator(context.Background(), tt.chunks, tt.batchSize, newMatchers(tt.matchers), nil, newQuery("", tt.start, tt.end, tt.direction))
streams, _, err := iter.ReadBatch(it, 1000)
_ = it.Close()
if err != nil {
@@ -305,5 +555,127 @@ func Test_newBatchChunkIterator(t *testing.T) {
})
}
+}
+func TestPartitionOverlappingchunks(t *testing.T) {
+ var (
+ oneThroughFour = newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from,
+ Line: "1",
+ },
+ {
+ Timestamp: from.Add(3 * time.Millisecond),
+ Line: "4",
+ },
+ },
+ })
+ two = newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from.Add(1 * time.Millisecond),
+ Line: "2",
+ },
+ },
+ })
+ three = newLazyChunk(logproto.Stream{
+ Labels: fooLabelsWithName,
+ Entries: []logproto.Entry{
+ {
+ Timestamp: from.Add(2 * time.Millisecond),
+ Line: "3",
+ },
+ },
+ })
+ )
+
+ for i, tc := range []struct {
+ input []*chunkenc.LazyChunk
+ expected [][]*chunkenc.LazyChunk
+ }{
+ {
+ input: []*chunkenc.LazyChunk{
+ oneThroughFour,
+ two,
+ three,
+ },
+ expected: [][]*chunkenc.LazyChunk{
+ []*chunkenc.LazyChunk{oneThroughFour},
+ []*chunkenc.LazyChunk{two, three},
+ },
+ },
+ {
+ input: []*chunkenc.LazyChunk{
+ two,
+ oneThroughFour,
+ three,
+ },
+ expected: [][]*chunkenc.LazyChunk{
+ []*chunkenc.LazyChunk{oneThroughFour},
+ []*chunkenc.LazyChunk{two, three},
+ },
+ },
+ {
+ input: []*chunkenc.LazyChunk{
+ two,
+ two,
+ three,
+ three,
+ },
+ expected: [][]*chunkenc.LazyChunk{
+ []*chunkenc.LazyChunk{two, three},
+ []*chunkenc.LazyChunk{two, three},
+ },
+ },
+ } {
+ t.Run(fmt.Sprintf("%d", i), func(t *testing.T) {
+ out := partitionOverlappingChunks(tc.input)
+ require.Equal(t, tc.expected, out)
+ })
+ }
+}
+
+func TestDropLabels(t *testing.T) {
+
+ for i, tc := range []struct {
+ ls labels.Labels
+ drop []string
+ expected labels.Labels
+ }{
+ {
+ ls: labels.Labels{
+ labels.Label{
+ Name: "a",
+ Value: "1",
+ },
+ labels.Label{
+ Name: "b",
+ Value: "2",
+ },
+ labels.Label{
+ Name: "c",
+ Value: "3",
+ },
+ },
+ drop: []string{"b"},
+ expected: labels.Labels{
+ labels.Label{
+ Name: "a",
+ Value: "1",
+ },
+ labels.Label{
+ Name: "c",
+ Value: "3",
+ },
+ },
+ },
+ } {
+ t.Run(fmt.Sprintf("%d", i), func(t *testing.T) {
+ dropped := dropLabels(tc.ls, tc.drop...)
+ require.Equal(t, tc.expected, dropped)
+ })
+ }
}
diff --git a/pkg/storage/util_test.go b/pkg/storage/util_test.go
index 9107438b8f145..57d9f974488de 100644
--- a/pkg/storage/util_test.go
+++ b/pkg/storage/util_test.go
@@ -19,6 +19,7 @@ import (
"github.com/stretchr/testify/assert"
)
+var fooLabelsWithName = "{foo=\"bar\", __name__=\"log\"}"
var fooLabels = "{foo=\"bar\"}"
var from = time.Unix(0, time.Millisecond.Nanoseconds())
|
storage
|
fix missing logs with batched chunk iterator (#1299)
|
2edbe1280f3c32e1d02794343d666edb67296811
|
2024-03-21 21:58:15
|
Christian Haudum
|
fix(blooms): Correctly skip block page in case it exceeds the max page size for querying (#12297)
| false
|
diff --git a/pkg/storage/bloom/v1/block.go b/pkg/storage/bloom/v1/block.go
index 84bc71a6b203c..c9eef5fa33027 100644
--- a/pkg/storage/bloom/v1/block.go
+++ b/pkg/storage/bloom/v1/block.go
@@ -148,25 +148,26 @@ func (bq *BlockQuerier) Seek(fp model.Fingerprint) error {
}
func (bq *BlockQuerier) Next() bool {
- if !bq.series.Next() {
- return false
- }
-
- series := bq.series.At()
-
- bq.blooms.Seek(series.Offset)
- if !bq.blooms.Next() {
- return false
- }
-
- bloom := bq.blooms.At()
-
- bq.cur = &SeriesWithBloom{
- Series: &series.Series,
- Bloom: bloom,
+ for bq.series.Next() {
+ series := bq.series.At()
+ bq.blooms.Seek(series.Offset)
+ if !bq.blooms.Next() {
+ // skip blocks that are too large
+ if errors.Is(bq.blooms.Err(), ErrPageTooLarge) {
+ // fmt.Printf("skipping bloom page: %s (%d)\n", series.Fingerprint, series.Chunks.Len())
+ bq.blooms.err = nil
+ continue
+ }
+ return false
+ }
+ bloom := bq.blooms.At()
+ bq.cur = &SeriesWithBloom{
+ Series: &series.Series,
+ Bloom: bloom,
+ }
+ return true
}
- return true
-
+ return false
}
func (bq *BlockQuerier) At() *SeriesWithBloom {
diff --git a/pkg/storage/bloom/v1/bloom.go b/pkg/storage/bloom/v1/bloom.go
index 6a6c2610e82e2..da0a770fb2579 100644
--- a/pkg/storage/bloom/v1/bloom.go
+++ b/pkg/storage/bloom/v1/bloom.go
@@ -18,7 +18,7 @@ import (
// Figure out a decent maximum page size that we can process.
// TODO(chaudum): Make max page size configurable
var maxPageSize = 32 << 20 // 32MB
-var errPageTooLarge = "bloom page too large to process: N=%d Offset=%d Len=%d DecompressedLen=%d"
+var ErrPageTooLarge = errors.Errorf("bloom page too large: size limit is %.1fMiB", float64(maxPageSize)/float64(1<<20))
type Bloom struct {
filter.ScalableBloomFilter
@@ -253,9 +253,10 @@ func (b *BloomBlock) BloomPageDecoder(r io.ReadSeeker, pageIdx int) (*BloomPageD
}
page := b.pageHeaders[pageIdx]
+ // fmt.Printf("pageIdx=%d page=%+v size=%.2fMiB\n", pageIdx, page, float64(page.Len)/float64(1<<20))
if page.Len > maxPageSize {
- return nil, fmt.Errorf(errPageTooLarge, page.N, page.Offset, page.Len, page.DecompressedLen)
+ return nil, ErrPageTooLarge
}
if _, err := r.Seek(int64(page.Offset), io.SeekStart); err != nil {
diff --git a/tools/bloom/inspector/main.go b/tools/bloom/inspector/main.go
new file mode 100644
index 0000000000000..bb81d02b260b1
--- /dev/null
+++ b/tools/bloom/inspector/main.go
@@ -0,0 +1,37 @@
+package main
+
+import (
+ "fmt"
+ "os"
+
+ v1 "github.com/grafana/loki/pkg/storage/bloom/v1"
+)
+
+func main() {
+ if len(os.Args) < 2 {
+ fmt.Println("Usage: go run main.go BLOCK_DIRECTORY")
+ os.Exit(2)
+ }
+
+ path := os.Args[1]
+ fmt.Printf("Block directory: %s\n", path)
+
+ r := v1.NewDirectoryBlockReader(path)
+ b := v1.NewBlock(r)
+ q := v1.NewBlockQuerier(b)
+
+ md, err := q.Metadata()
+ if err != nil {
+ panic(err)
+ }
+
+ fmt.Printf("Metadata: %+v\n", md)
+
+ for q.Next() {
+ swb := q.At()
+ fmt.Printf("%s (%d)\n", swb.Series.Fingerprint, swb.Series.Chunks.Len())
+ }
+ if q.Err() != nil {
+ fmt.Printf("error: %s\n", q.Err())
+ }
+}
|
fix
|
Correctly skip block page in case it exceeds the max page size for querying (#12297)
|
f0ec743ba09d38c4cdb01e88fb8b0a1198b3c25f
|
2023-12-13 16:21:36
|
Zirko
|
helm: add cilium networkpolicies (#11425)
| false
|
diff --git a/docs/sources/setup/install/helm/reference.md b/docs/sources/setup/install/helm/reference.md
index 8252a6fd103a3..e650d0fca2fc2 100644
--- a/docs/sources/setup/install/helm/reference.md
+++ b/docs/sources/setup/install/helm/reference.md
@@ -3110,6 +3110,15 @@ false
<td><pre lang="json">
[]
</pre>
+</td>
+ </tr>
+ <tr>
+ <td>networkPolicy.flavor</td>
+ <td>string</td>
+ <td>Specifies whether the policies created will be standard Network Policies (flavor: kubernetes) or Cilium Network Policies (flavor: cilium)</td>
+ <td><pre lang="json">
+"kubernetes"
+</pre>
</td>
</tr>
<tr>
diff --git a/production/helm/loki/CHANGELOG.md b/production/helm/loki/CHANGELOG.md
index 78571b3de600d..96bebdf5aebc9 100644
--- a/production/helm/loki/CHANGELOG.md
+++ b/production/helm/loki/CHANGELOG.md
@@ -13,6 +13,10 @@ Entries should include a reference to the pull request that introduced the chang
[//]: # (<AUTOMATED_UPDATES_LOCATOR> : do not remove this line. This locator is used by the CI pipeline to automatically create a changelog entry for each new Loki release. Add other chart versions and respective changelog entries bellow this line.)
+## 5.41.2
+
+- [FEATURE] Add ciliumnetworkpolicies.
+
## 5.41.1
- [FEATURE] Allow topology spread constraints for Loki read deployment component.
diff --git a/production/helm/loki/Chart.yaml b/production/helm/loki/Chart.yaml
index d9cf011e4f23e..fc7e0fbacbc6e 100644
--- a/production/helm/loki/Chart.yaml
+++ b/production/helm/loki/Chart.yaml
@@ -3,7 +3,7 @@ name: loki
description: Helm chart for Grafana Loki in simple, scalable mode
type: application
appVersion: 2.9.3
-version: 5.41.1
+version: 5.41.2
home: https://grafana.github.io/helm-charts
sources:
- https://github.com/grafana/loki
diff --git a/production/helm/loki/README.md b/production/helm/loki/README.md
index 3caad398ada44..e1da365b5bf92 100644
--- a/production/helm/loki/README.md
+++ b/production/helm/loki/README.md
@@ -1,6 +1,6 @@
# loki
-  
+  
Helm chart for Grafana Loki in simple, scalable mode
diff --git a/production/helm/loki/templates/ciliumnetworkpolicy.yaml b/production/helm/loki/templates/ciliumnetworkpolicy.yaml
new file mode 100644
index 0000000000000..5633ae1945206
--- /dev/null
+++ b/production/helm/loki/templates/ciliumnetworkpolicy.yaml
@@ -0,0 +1,184 @@
+{{- if and (.Values.networkPolicy.enabled) (eq .Values.networkPolicy.flavor "cilium") }}
+---
+apiVersion: cilium.io/v2
+kind: CiliumNetworkPolicy
+metadata:
+ name: {{ include "loki.name" . }}-namespace-only
+ namespace: {{ $.Release.Namespace }}
+ labels:
+ {{- include "loki.labels" . | nindent 4 }}
+spec:
+ endpointSelector: {}
+ egress:
+ - toEndpoints:
+ - {}
+ ingress:
+ - fromEndpoints:
+ - {}
+
+---
+apiVersion: cilium.io/v2
+kind: CiliumNetworkPolicy
+metadata:
+ name: {{ include "loki.name" . }}-egress-dns
+ namespace: {{ $.Release.Namespace }}
+ labels:
+ {{- include "loki.labels" . | nindent 4 }}
+spec:
+ endpointSelector:
+ matchLabels:
+ {{- include "loki.selectorLabels" . | nindent 6 }}
+ egress:
+ - toPorts:
+ - ports:
+ - port: dns
+ protocol: UDP
+ toEndpoints:
+ - namespaceSelector: {}
+
+---
+apiVersion: cilium.io/v2
+kind: CiliumNetworkPolicy
+metadata:
+ name: {{ include "loki.name" . }}-ingress
+ namespace: {{ $.Release.Namespace }}
+ labels:
+ {{- include "loki.labels" . | nindent 4 }}
+spec:
+ endpointSelector:
+ matchExpressions:
+ - key: app.kubernetes.io/component
+ operator: In
+ values:
+ {{- if .Values.gateway.enabled }}
+ - gateway
+ {{- else }}
+ - read
+ - write
+ {{- end }}
+ matchLabels:
+ {{- include "loki.selectorLabels" . | nindent 6 }}
+ ingress:
+ - toPorts:
+ - port: http
+ protocol: TCP
+ {{- if .Values.networkPolicy.ingress.namespaceSelector }}
+ fromEndpoints:
+ - matchLabels:
+ {{- toYaml .Values.networkPolicy.ingress.namespaceSelector | nindent 8 }}
+ {{- if .Values.networkPolicy.ingress.podSelector }}
+ {{- toYaml .Values.networkPolicy.ingress.podSelector | nindent 8 }}
+ {{- end }}
+ {{- end }}
+
+---
+apiVersion: cilium.io/v2
+kind: CiliumNetworkPolicy
+metadata:
+ name: {{ include "loki.name" . }}-ingress-metrics
+ namespace: {{ $.Release.Namespace }}
+ labels:
+ {{- include "loki.labels" . | nindent 4 }}
+spec:
+ endpointSelector:
+ matchLabels:
+ {{- include "loki.selectorLabels" . | nindent 6 }}
+ ingress:
+ - toPorts:
+ - port: http-metrics
+ protocol: TCP
+ {{- if .Values.networkPolicy.metrics.cidrs }}
+ {{- range $cidr := .Values.networkPolicy.metrics.cidrs }}
+ toCIDR:
+ - {{ $cidr }}
+ {{- end }}
+ {{- if .Values.networkPolicy.metrics.namespaceSelector }}
+ fromEndpoints:
+ - matchLabels:
+ {{- toYaml .Values.networkPolicy.metrics.namespaceSelector | nindent 8 }}
+ {{- if .Values.networkPolicy.metrics.podSelector }}
+ {{- toYaml .Values.networkPolicy.metrics.podSelector | nindent 8 }}
+ {{- end }}
+ {{- end }}
+ {{- end }}
+
+---
+apiVersion: cilium.io/v2
+kind: CiliumNetworkPolicy
+metadata:
+ name: {{ include "loki.name" . }}-egress-alertmanager
+ namespace: {{ $.Release.Namespace }}
+ labels:
+ {{- include "loki.labels" . | nindent 4 }}
+spec:
+ endpointSelector:
+ matchLabels:
+ {{- include "loki.backendSelectorLabels" . | nindent 6 }}
+ egress:
+ - toPorts:
+ - port: {{ .Values.networkPolicy.alertmanager.port }}
+ protocol: TCP
+ {{- if .Values.networkPolicy.alertmanager.namespaceSelector }}
+ toEndpoints:
+ - matchLabels:
+ {{- toYaml .Values.networkPolicy.alertmanager.namespaceSelector | nindent 8 }}
+ {{- if .Values.networkPolicy.alertmanager.podSelector }}
+ {{- toYaml .Values.networkPolicy.alertmanager.podSelector | nindent 8 }}
+ {{- end }}
+ {{- end }}
+
+{{- if .Values.networkPolicy.externalStorage.ports }}
+---
+apiVersion: cilium.io/v2
+kind: CiliumNetworkPolicy
+metadata:
+ name: {{ include "loki.name" . }}-egress-external-storage
+ namespace: {{ $.Release.Namespace }}
+ labels:
+ {{- include "loki.labels" . | nindent 4 }}
+spec:
+ endpointSelector:
+ matchLabels:
+ {{- include "loki.selectorLabels" . | nindent 6 }}
+ egress:
+ - toPorts:
+ {{- range $port := .Values.networkPolicy.externalStorage.ports }}
+ - port: {{ $port }}
+ protocol: TCP
+ {{- end }}
+ {{- if .Values.networkPolicy.externalStorage.cidrs }}
+ {{- range $cidr := .Values.networkPolicy.externalStorage.cidrs }}
+ toCIDR:
+ - {{ $cidr }}
+ {{- end }}
+ {{- end }}
+{{- end }}
+
+{{- end }}
+
+{{- if .Values.networkPolicy.discovery.port }}
+---
+apiVersion: cilium.io/v2
+kind: CiliumNetworkPolicy
+metadata:
+ name: {{ include "loki.name" . }}-egress-discovery
+ namespace: {{ $.Release.Namespace }}
+ labels:
+ {{- include "loki.labels" . | nindent 4 }}
+spec:
+ endpointSelector:
+ matchLabels:
+ {{- include "loki.selectorLabels" . | nindent 6 }}
+ egress:
+ - toPorts:
+ - port: {{ .Values.networkPolicy.discovery.port }}
+ protocol: TCP
+ {{- if .Values.networkPolicy.discovery.namespaceSelector }}
+ toEndpoints:
+ - matchLabels:
+ {{- toYaml .Values.networkPolicy.discovery.namespaceSelector | nindent 8 }}
+ {{- if .Values.networkPolicy.discovery.podSelector }}
+ {{- toYaml .Values.networkPolicy.discovery.podSelector | nindent 8 }}
+ {{- end }}
+ {{- end }}
+{{- end }}
diff --git a/production/helm/loki/templates/networkpolicy.yaml b/production/helm/loki/templates/networkpolicy.yaml
index 4424d90db08d4..27c85280eb08c 100644
--- a/production/helm/loki/templates/networkpolicy.yaml
+++ b/production/helm/loki/templates/networkpolicy.yaml
@@ -1,4 +1,4 @@
-{{- if .Values.networkPolicy.enabled }}
+{{- if and (.Values.networkPolicy.enabled) (eq .Values.networkPolicy.flavor "kubernetes") }}
---
apiVersion: networking.k8s.io/v1
kind: NetworkPolicy
diff --git a/production/helm/loki/values.yaml b/production/helm/loki/values.yaml
index 738cf6ea25ae7..e82967a4efb3b 100644
--- a/production/helm/loki/values.yaml
+++ b/production/helm/loki/values.yaml
@@ -1465,6 +1465,9 @@ gateway:
networkPolicy:
# -- Specifies whether Network Policies should be created
enabled: false
+ # -- Specifies whether the policies created will be standard Network Policies (flavor: kubernetes)
+ # or Cilium Network Policies (flavor: cilium)
+ flavor: kubernetes
metrics:
# -- Specifies the Pods which are allowed to access the metrics port.
# As this is cross-namespace communication, you also need the namespaceSelector.
|
helm
|
add cilium networkpolicies (#11425)
|
86f6c2eb8d549438df6628be4d1efacae5444831
|
2022-09-01 23:08:04
|
Gerard Vanloo
|
operator: Adding Lokistack Gateway Request Errors Alert (#6999)
| false
|
diff --git a/operator/CHANGELOG.md b/operator/CHANGELOG.md
index dd721d52f0c1a..d36dd519d4071 100644
--- a/operator/CHANGELOG.md
+++ b/operator/CHANGELOG.md
@@ -1,5 +1,6 @@
## Main
+- [6999](https://github.com/grafana/loki/pull/6999) **Red-GV**: Adding LokiStack Gateway alerts
- [7000](https://github.com/grafana/loki/pull/7000) **xperimental**: Configure default node affinity for all pods
- [6923](https://github.com/grafana/loki/pull/6923) **xperimental**: Reconcile owner reference for existing objects
- [6907](https://github.com/grafana/loki/pull/6907) **Red-GV**: Adding valid subscription annotation to operator metadata
diff --git a/operator/docs/lokistack/sop.md b/operator/docs/lokistack/sop.md
index 3e1ff566c0035..9f2cdaa8b2e7b 100644
--- a/operator/docs/lokistack/sop.md
+++ b/operator/docs/lokistack/sop.md
@@ -34,8 +34,8 @@ A service(s) is failing to process at least 10% of all incoming requests.
- Console access to the cluster
- Edit access to the deployed operator and Loki namespace:
- OpenShift
- - `openshift-logging`
- - `openshift-operators-redhat`
+ - `openshift-logging` (LokiStack)
+ - `openshift-operators-redhat` (Loki Operator)
### Steps
@@ -46,6 +46,69 @@ A service(s) is failing to process at least 10% of all incoming requests.
- `loki_ingester_wal_disk_full_failures_total`
- `loki_ingester_wal_corruptions_total`
+## LokiStack Write Request Errors
+
+### Impact
+
+The LokiStack Gateway component is unable to perform its duties for a number of write requests, resulting in potential loss of data.
+
+### Summary
+
+The LokiStack Gateway is failing to process at least 10% of all incoming write requests.
+
+### Severity
+
+`Critical`
+
+### Access Required
+
+- Console access to the cluster
+- Edit access to the deployed operator and Loki namespace:
+ - OpenShift
+ - `openshift-logging` (LokiStack)
+ - `openshift-operators-redhat` (Loki Operator)
+
+### Steps
+
+- Ensure that the LokiStack Gateway component is ready and available
+- Ensure that the `distributor`, `ingester`, and `index-gateway` components are ready and available
+- Ensure that store services (`ingester`, `querier`, `index-gateway`, `compactor`) can communicate with backend storage
+- Examine metrics for signs of failure
+ - WAL Complications
+ - `loki_ingester_wal_disk_full_failures_total`
+ - `loki_ingester_wal_corruptions_total`
+
+## LokiStack Read Request Errors
+
+### Impact
+
+The LokiStack Gateway component is unable to perform its duties for a number of query requests, resulting in a potential disruption.
+
+### Summary
+
+The LokiStack Gateway is failing to process at least 10% of all incoming query requests.
+
+### Severity
+
+`Critical`
+
+### Access Required
+
+- Console access to the cluster
+- Edit access to the deployed operator and Loki namespace:
+ - OpenShift
+ - `openshift-logging` (LokiStack)
+ - `openshift-operators-redhat` (Loki Operator)
+
+### Steps
+
+- Ensure that the LokiStack Gateway component is ready and available
+- Ensure that the `query-frontend`, `querier`, `ingester`, and `index-gateway` components are ready and available
+- Ensure that store services (`ingester`, `querier`, `index-gateway`, `compactor`) can communicate with backend storage
+- Examine metrics for signs of failure
+ - WAL Complications
+ - `loki_ingester_wal_disk_full_failures_total`
+ - `loki_ingester_wal_corruptions_total`
## Loki Request Panics
@@ -66,8 +129,8 @@ A service(s) has crashed.
- Console access to the cluster
- Edit access to the deployed operator and Loki namespace:
- OpenShift
- - `openshift-logging`
- - `openshift-operators-redhat`
+ - `openshift-logging` (LokiStack)
+ - `openshift-operators-redhat` (Loki Operator)
### Steps
diff --git a/operator/internal/manifests/internal/alerts/build.go b/operator/internal/manifests/internal/alerts/build.go
index b1bf76558af16..4e860664f2266 100644
--- a/operator/internal/manifests/internal/alerts/build.go
+++ b/operator/internal/manifests/internal/alerts/build.go
@@ -13,7 +13,7 @@ import (
const (
// RunbookDefaultURL is the default url for the documentation of the Prometheus alerts
- RunbookDefaultURL = "https://github.com/grafana/loki/tree/main/operator/docs/alerts.md"
+ RunbookDefaultURL = "https://github.com/grafana/loki/blob/main/operator/docs/lokistack/sop.md"
)
var (
diff --git a/operator/internal/manifests/internal/alerts/prometheus-alerts.yaml b/operator/internal/manifests/internal/alerts/prometheus-alerts.yaml
index 8635c347e9e3e..0e408200b16a0 100644
--- a/operator/internal/manifests/internal/alerts/prometheus-alerts.yaml
+++ b/operator/internal/manifests/internal/alerts/prometheus-alerts.yaml
@@ -10,21 +10,67 @@ groups:
runbook_url: "[[ .RunbookURL ]]#Loki-Request-Errors"
expr: |
sum(
- rate(
- loki_request_duration_seconds_count{status_code=~"5.."}[1m]
- )
+ rate(
+ loki_request_duration_seconds_count{status_code=~"5.."}[1m]
+ )
) by (namespace, job, route)
/
sum(
- rate(
- loki_request_duration_seconds_count[1m]
- )
+ rate(
+ loki_request_duration_seconds_count[1m]
+ )
) by (namespace, job, route)
* 100
> 10
for: 15m
labels:
severity: critical
+ - alert: LokiStackWriteRequestErrors
+ annotations:
+ message: |-
+ {{ printf "%.2f" $value }}% of write requests from {{ $labels.job }} are returned with server errors.
+ summary: "At least 10% of write requests to the lokistack-gateway are responded with 5xx server errors."
+ runbook_url: "[[ .RunbookURL ]]#LokiStack-Write-Request-Errors"
+ expr: |
+ sum(
+ rate(
+ http_requests_total{code=~"5..", group="logsv1", handler="push"}[1m]
+ )
+ ) by (namespace, job, tenant)
+ /
+ sum(
+ rate(
+ http_requests_total{group="logsv1", handler="push"}[1m]
+ )
+ ) by (namespace, job, tenant)
+ * 100
+ > 10
+ for: 15m
+ labels:
+ severity: critical
+ - alert: LokiStackReadRequestErrors
+ annotations:
+ message: |-
+ {{ printf "%.2f" $value }}% of query requests from {{ $labels.job }} are returned with server errors.
+ summary: "At least 10% of query requests to the lokistack-gateway are responded with 5xx server errors."
+ runbook_url: "[[ .RunbookURL ]]#LokiStack-Read-Request-Errors"
+ expr: |
+ sum(
+ rate(
+ http_requests_total{code=~"5..", group="logsv1", handler=~"query|query_range|label|labels|label_values"}[1m]
+ )
+ ) by (namespace, job, tenant)
+ /
+ sum(
+ rate(
+ http_requests_total{group="logsv1", handler=~"query|query_range|label|labels|label_values"}[1m]
+ )
+ ) by (namespace, job, tenant)
+ * 100
+ > 10
+ for: 15m
+ labels:
+ severity: critical
- alert: LokiRequestPanics
annotations:
message: |-
@@ -33,9 +79,9 @@ groups:
runbook_url: "[[ .RunbookURL ]]#Loki-Request-Panics"
expr: |
sum(
- increase(
- loki_panic_total[10m]
- )
+ increase(
+ loki_panic_total[10m]
+ )
) by (namespace, job)
> 0
labels:
diff --git a/operator/internal/manifests/internal/alerts/testdata/test.yaml b/operator/internal/manifests/internal/alerts/testdata/test.yaml
index 295c97fe2c11f..6de3c7266fba8 100644
--- a/operator/internal/manifests/internal/alerts/testdata/test.yaml
+++ b/operator/internal/manifests/internal/alerts/testdata/test.yaml
@@ -11,13 +11,20 @@ tests:
values: '1+1x20'
- series: 'loki_request_duration_seconds_count{status_code="200", namespace="my-ns", job="ingester", route="my-route"}'
values: '1+3x20'
+ - series: 'http_requests_total{code="500", namespace="my-ns", job="gateway", handler="push", group="logsv1"}'
+ values: '1+1x20'
+ - series: 'http_requests_total{code="200", namespace="my-ns", job="gateway", handler="push", group="logsv1"}'
+ values: '1+3x20'
+ - series: 'http_requests_total{code="500", namespace="my-ns", job="gateway", handler="query", group="logsv1"}'
+ values: '1+1x20'
+ - series: 'http_requests_total{code="200", namespace="my-ns", job="gateway", handler="query", group="logsv1"}'
+ values: '1+3x20'
- series: 'loki_panic_total{namespace="my-ns", job="ingester"}'
values: '0 1 1 2+0x10'
# Unit test for alerting rules.
alert_rule_test:
- # --------- LokiRequestErrors ---------
- eval_time: 16m
alertname: LokiRequestErrors
exp_alerts:
@@ -30,8 +37,28 @@ tests:
summary: "At least 10% of requests are responded by 5xx server errors."
message: "ingester my-route is experiencing 25.00% errors."
runbook_url: "[[ .RunbookURL ]]#Loki-Request-Errors"
-
- # --------- LokiRequestPanics ---------
+ - eval_time: 16m
+ alertname: LokiStackWriteRequestErrors
+ exp_alerts:
+ - exp_labels:
+ namespace: my-ns
+ job: gateway
+ severity: critical
+ exp_annotations:
+ summary: "At least 10% of write requests to the lokistack-gateway are responded with 5xx server errors."
+ message: "25.00% of write requests from gateway are returned with server errors."
+ runbook_url: "[[ .RunbookURL ]]#LokiStack-Write-Request-Errors"
+ - eval_time: 16m
+ alertname: LokiStackReadRequestErrors
+ exp_alerts:
+ - exp_labels:
+ namespace: my-ns
+ job: gateway
+ severity: critical
+ exp_annotations:
+ summary: "At least 10% of query requests to the lokistack-gateway are responded with 5xx server errors."
+ message: "25.00% of query requests from gateway are returned with server errors."
+ runbook_url: "[[ .RunbookURL ]]#LokiStack-Read-Request-Errors"
- eval_time: 10m
alertname: LokiRequestPanics
exp_alerts:
|
operator
|
Adding Lokistack Gateway Request Errors Alert (#6999)
|
d0a285926b7257d54cf948ba644c619a4b49a871
|
2024-05-29 15:24:59
|
benclive
|
feat: Increase drain max depth from 8 -> 30 (#13063)
| false
| "diff --git a/pkg/pattern/drain/drain.go b/pkg/pattern/drain/drain.go\nindex 31932832f7010..c3076386(...TRUNCATED)
|
feat
|
Increase drain max depth from 8 -> 30 (#13063)
|
4e04d07168a8c5cb7086ced8486c6d584faa1045
|
2024-04-19 13:08:48
|
Paul Rogers
|
fix: promtail race fixes (#12656)
| false
| "diff --git a/clients/pkg/promtail/client/client_writeto_test.go b/clients/pkg/promtail/client/clien(...TRUNCATED)
|
fix
|
promtail race fixes (#12656)
|
d2e1992366bf4dd34e6401cc6734ce532b38e5b9
|
2024-11-22 23:32:20
|
George Robinson
|
chore: separate usage of partition owner gauge (#15079)
| false
| "diff --git a/pkg/kafka/partition/reader_service.go b/pkg/kafka/partition/reader_service.go\nindex d(...TRUNCATED)
|
chore
|
separate usage of partition owner gauge (#15079)
|
aeaefe6aab3f611684dcf8c05756947d02c6e1e5
|
2024-02-21 23:44:49
|
Robert Jacob
|
feat(operator): Extend Azure secret validation (#12007)
| false
| "diff --git a/operator/CHANGELOG.md b/operator/CHANGELOG.md\nindex 8dae6eced0bfa..e6aaec29b99c7 1006(...TRUNCATED)
|
feat
|
Extend Azure secret validation (#12007)
|
3a0be4897512ec16a0e43cd5e0a49774cda87aed
|
2022-11-23 22:20:56
|
Kaviraj Kanagaraj
|
config: Improve error message loading config with ENV variables. (#7759)
| false
| "diff --git a/CHANGELOG.md b/CHANGELOG.md\nindex ff4b1f64c72ef..ba79c01ed7a84 100644\n--- a/CHANGELO(...TRUNCATED)
|
config
|
Improve error message loading config with ENV variables. (#7759)
|
End of preview. Expand
in Data Studio
No dataset card yet
- Downloads last month
- 8