mirror of https://github.com/prometheus/prometheus
444 lines
17 KiB
Go
444 lines
17 KiB
Go
// Copyright 2021 The Prometheus Authors
|
||
// Licensed under the Apache License, Version 2.0 (the "License");
|
||
// you may not use this file except in compliance with the License.
|
||
// You may obtain a copy of the License at
|
||
//
|
||
// http://www.apache.org/licenses/LICENSE-2.0
|
||
//
|
||
// Unless required by applicable law or agreed to in writing, software
|
||
// distributed under the License is distributed on an "AS IS" BASIS,
|
||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||
// See the License for the specific language governing permissions and
|
||
// limitations under the License.
|
||
|
||
package remote
|
||
|
||
import (
|
||
"bytes"
|
||
"context"
|
||
"errors"
|
||
"io"
|
||
"net/http"
|
||
"net/http/httptest"
|
||
"testing"
|
||
"time"
|
||
|
||
"github.com/gogo/protobuf/proto"
|
||
"github.com/golang/snappy"
|
||
"github.com/stretchr/testify/require"
|
||
|
||
"github.com/prometheus/prometheus/config"
|
||
"github.com/prometheus/prometheus/model/labels"
|
||
"github.com/prometheus/prometheus/prompb"
|
||
"github.com/prometheus/prometheus/promql/promqltest"
|
||
"github.com/prometheus/prometheus/storage"
|
||
"github.com/prometheus/prometheus/tsdb/tsdbutil"
|
||
"github.com/prometheus/prometheus/util/teststorage"
|
||
)
|
||
|
||
func TestSampledReadEndpoint(t *testing.T) {
|
||
store := promqltest.LoadedStorage(t, `
|
||
load 1m
|
||
test_metric1{foo="bar",baz="qux"} 1
|
||
`)
|
||
defer store.Close()
|
||
|
||
addNativeHistogramsToTestSuite(t, store, 1)
|
||
|
||
h := NewReadHandler(nil, nil, store, func() config.Config {
|
||
return config.Config{
|
||
GlobalConfig: config.GlobalConfig{
|
||
// We expect external labels to be added, with the source labels honored.
|
||
ExternalLabels: labels.FromStrings("b", "c", "baz", "a", "d", "e"),
|
||
},
|
||
}
|
||
}, 1e6, 1, 0)
|
||
|
||
// Encode the request.
|
||
matcher1, err := labels.NewMatcher(labels.MatchEqual, "__name__", "test_metric1")
|
||
require.NoError(t, err)
|
||
|
||
matcher2, err := labels.NewMatcher(labels.MatchEqual, "d", "e")
|
||
require.NoError(t, err)
|
||
|
||
matcher3, err := labels.NewMatcher(labels.MatchEqual, "__name__", "test_histogram_metric1")
|
||
require.NoError(t, err)
|
||
|
||
query1, err := ToQuery(0, 1, []*labels.Matcher{matcher1, matcher2}, &storage.SelectHints{Step: 0, Func: "avg"})
|
||
require.NoError(t, err)
|
||
|
||
query2, err := ToQuery(0, 1, []*labels.Matcher{matcher3, matcher2}, &storage.SelectHints{Step: 0, Func: "avg"})
|
||
require.NoError(t, err)
|
||
|
||
req := &prompb.ReadRequest{Queries: []*prompb.Query{query1, query2}}
|
||
data, err := proto.Marshal(req)
|
||
require.NoError(t, err)
|
||
|
||
compressed := snappy.Encode(nil, data)
|
||
request, err := http.NewRequest(http.MethodPost, "", bytes.NewBuffer(compressed))
|
||
require.NoError(t, err)
|
||
|
||
recorder := httptest.NewRecorder()
|
||
h.ServeHTTP(recorder, request)
|
||
|
||
require.Equal(t, 2, recorder.Code/100)
|
||
|
||
require.Equal(t, "application/x-protobuf", recorder.Result().Header.Get("Content-Type"))
|
||
require.Equal(t, "snappy", recorder.Result().Header.Get("Content-Encoding"))
|
||
|
||
// Decode the response.
|
||
compressed, err = io.ReadAll(recorder.Result().Body)
|
||
require.NoError(t, err)
|
||
|
||
uncompressed, err := snappy.Decode(nil, compressed)
|
||
require.NoError(t, err)
|
||
|
||
var resp prompb.ReadResponse
|
||
err = proto.Unmarshal(uncompressed, &resp)
|
||
require.NoError(t, err)
|
||
|
||
require.Len(t, resp.Results, 2, "Expected 2 results.")
|
||
|
||
require.Equal(t, &prompb.QueryResult{
|
||
Timeseries: []*prompb.TimeSeries{
|
||
{
|
||
Labels: []prompb.Label{
|
||
{Name: "__name__", Value: "test_metric1"},
|
||
{Name: "b", Value: "c"},
|
||
{Name: "baz", Value: "qux"},
|
||
{Name: "d", Value: "e"},
|
||
{Name: "foo", Value: "bar"},
|
||
},
|
||
Samples: []prompb.Sample{{Value: 1, Timestamp: 0}},
|
||
},
|
||
},
|
||
}, resp.Results[0])
|
||
|
||
require.Equal(t, &prompb.QueryResult{
|
||
Timeseries: []*prompb.TimeSeries{
|
||
{
|
||
Labels: []prompb.Label{
|
||
{Name: "__name__", Value: "test_histogram_metric1"},
|
||
{Name: "b", Value: "c"},
|
||
{Name: "baz", Value: "qux"},
|
||
{Name: "d", Value: "e"},
|
||
},
|
||
Histograms: []prompb.Histogram{
|
||
prompb.FromFloatHistogram(0, tsdbutil.GenerateTestFloatHistogram(0)),
|
||
},
|
||
},
|
||
},
|
||
}, resp.Results[1])
|
||
}
|
||
|
||
func BenchmarkStreamReadEndpoint(b *testing.B) {
|
||
store := promqltest.LoadedStorage(b, `
|
||
load 1m
|
||
test_metric1{foo="bar1",baz="qux"} 0+100x119
|
||
test_metric1{foo="bar2",baz="qux"} 0+100x120
|
||
test_metric1{foo="bar3",baz="qux"} 0+100x240
|
||
`)
|
||
|
||
b.Cleanup(func() { store.Close() })
|
||
|
||
api := NewReadHandler(nil, nil, store, func() config.Config {
|
||
return config.Config{}
|
||
},
|
||
0, 1, 0,
|
||
)
|
||
|
||
matcher, err := labels.NewMatcher(labels.MatchEqual, "__name__", "test_metric1")
|
||
require.NoError(b, err)
|
||
|
||
query, err := ToQuery(0, 14400001, []*labels.Matcher{matcher}, &storage.SelectHints{
|
||
Step: 1,
|
||
Func: "sum",
|
||
Start: 0,
|
||
End: 14400001,
|
||
})
|
||
require.NoError(b, err)
|
||
|
||
req := &prompb.ReadRequest{
|
||
Queries: []*prompb.Query{query},
|
||
AcceptedResponseTypes: []prompb.ReadRequest_ResponseType{prompb.ReadRequest_STREAMED_XOR_CHUNKS},
|
||
}
|
||
data, err := proto.Marshal(req)
|
||
require.NoError(b, err)
|
||
|
||
b.ResetTimer()
|
||
b.ReportAllocs()
|
||
|
||
for i := 0; i < b.N; i++ {
|
||
compressed := snappy.Encode(nil, data)
|
||
request, err := http.NewRequest(http.MethodPost, "", bytes.NewBuffer(compressed))
|
||
require.NoError(b, err)
|
||
|
||
recorder := httptest.NewRecorder()
|
||
api.ServeHTTP(recorder, request)
|
||
|
||
require.Equal(b, 2, recorder.Code/100)
|
||
|
||
var results []*prompb.ChunkedReadResponse
|
||
stream := NewChunkedReader(recorder.Result().Body, DefaultChunkedReadLimit, nil)
|
||
|
||
for {
|
||
res := &prompb.ChunkedReadResponse{}
|
||
err := stream.NextProto(res)
|
||
if errors.Is(err, io.EOF) {
|
||
break
|
||
}
|
||
require.NoError(b, err)
|
||
results = append(results, res)
|
||
}
|
||
|
||
require.Len(b, results, 6, "Expected 6 results.")
|
||
}
|
||
}
|
||
|
||
func TestStreamReadEndpoint(t *testing.T) {
|
||
// First with 120 float samples. We expect 1 frame with 1 chunk.
|
||
// Second with 121 float samples, We expect 1 frame with 2 chunks.
|
||
// Third with 241 float samples. We expect 1 frame with 2 chunks, and 1 frame with 1 chunk for the same series due to bytes limit.
|
||
// Fourth with 25 histogram samples. We expect 1 frame with 1 chunk.
|
||
store := promqltest.LoadedStorage(t, `
|
||
load 1m
|
||
test_metric1{foo="bar1",baz="qux"} 0+100x119
|
||
test_metric1{foo="bar2",baz="qux"} 0+100x120
|
||
test_metric1{foo="bar3",baz="qux"} 0+100x240
|
||
`)
|
||
defer store.Close()
|
||
|
||
addNativeHistogramsToTestSuite(t, store, 25)
|
||
|
||
api := NewReadHandler(nil, nil, store, func() config.Config {
|
||
return config.Config{
|
||
GlobalConfig: config.GlobalConfig{
|
||
// We expect external labels to be added, with the source labels honored.
|
||
ExternalLabels: labels.FromStrings("baz", "a", "b", "c", "d", "e"),
|
||
},
|
||
}
|
||
},
|
||
1e6, 1,
|
||
// Labelset has 57 bytes. Full chunk in test data has roughly 240 bytes. This allows us to have at max 2 chunks in this test.
|
||
57+480,
|
||
)
|
||
|
||
// Encode the request.
|
||
matcher1, err := labels.NewMatcher(labels.MatchEqual, "__name__", "test_metric1")
|
||
require.NoError(t, err)
|
||
|
||
matcher2, err := labels.NewMatcher(labels.MatchEqual, "d", "e")
|
||
require.NoError(t, err)
|
||
|
||
matcher3, err := labels.NewMatcher(labels.MatchEqual, "foo", "bar1")
|
||
require.NoError(t, err)
|
||
|
||
matcher4, err := labels.NewMatcher(labels.MatchEqual, "__name__", "test_histogram_metric1")
|
||
require.NoError(t, err)
|
||
|
||
query1, err := ToQuery(0, 14400001, []*labels.Matcher{matcher1, matcher2}, &storage.SelectHints{
|
||
Step: 1,
|
||
Func: "avg",
|
||
Start: 0,
|
||
End: 14400001,
|
||
})
|
||
require.NoError(t, err)
|
||
|
||
query2, err := ToQuery(0, 14400001, []*labels.Matcher{matcher1, matcher3}, &storage.SelectHints{
|
||
Step: 1,
|
||
Func: "avg",
|
||
Start: 0,
|
||
End: 14400001,
|
||
})
|
||
require.NoError(t, err)
|
||
|
||
query3, err := ToQuery(0, 14400001, []*labels.Matcher{matcher4}, &storage.SelectHints{
|
||
Step: 1,
|
||
Func: "avg",
|
||
Start: 0,
|
||
End: 14400001,
|
||
})
|
||
require.NoError(t, err)
|
||
|
||
req := &prompb.ReadRequest{
|
||
Queries: []*prompb.Query{query1, query2, query3},
|
||
AcceptedResponseTypes: []prompb.ReadRequest_ResponseType{prompb.ReadRequest_STREAMED_XOR_CHUNKS},
|
||
}
|
||
data, err := proto.Marshal(req)
|
||
require.NoError(t, err)
|
||
|
||
compressed := snappy.Encode(nil, data)
|
||
request, err := http.NewRequest(http.MethodPost, "", bytes.NewBuffer(compressed))
|
||
require.NoError(t, err)
|
||
|
||
recorder := httptest.NewRecorder()
|
||
api.ServeHTTP(recorder, request)
|
||
|
||
require.Equal(t, 2, recorder.Code/100)
|
||
|
||
require.Equal(t, "application/x-streamed-protobuf; proto=prometheus.ChunkedReadResponse", recorder.Result().Header.Get("Content-Type"))
|
||
require.Equal(t, "", recorder.Result().Header.Get("Content-Encoding"))
|
||
|
||
var results []*prompb.ChunkedReadResponse
|
||
stream := NewChunkedReader(recorder.Result().Body, DefaultChunkedReadLimit, nil)
|
||
for {
|
||
res := &prompb.ChunkedReadResponse{}
|
||
err := stream.NextProto(res)
|
||
if errors.Is(err, io.EOF) {
|
||
break
|
||
}
|
||
require.NoError(t, err)
|
||
results = append(results, res)
|
||
}
|
||
|
||
require.Len(t, results, 6, "Expected 6 results.")
|
||
|
||
require.Equal(t, []*prompb.ChunkedReadResponse{
|
||
{
|
||
ChunkedSeries: []*prompb.ChunkedSeries{
|
||
{
|
||
Labels: []prompb.Label{
|
||
{Name: "__name__", Value: "test_metric1"},
|
||
{Name: "b", Value: "c"},
|
||
{Name: "baz", Value: "qux"},
|
||
{Name: "d", Value: "e"},
|
||
{Name: "foo", Value: "bar1"},
|
||
},
|
||
Chunks: []prompb.Chunk{
|
||
{
|
||
Type: prompb.Chunk_XOR,
|
||
MaxTimeMs: 7140000,
|
||
Data: []byte("\000x\000\000\000\000\000\000\000\000\000\340\324\003\302|\005\224\000\301\254}\351z2\320O\355\264n[\007\316\224\243md\371\320\375\032Pm\nS\235\016Q\255\006P\275\250\277\312\201Z\003(3\240R\207\332\005(\017\240\322\201\332=(\023\2402\203Z\007(w\2402\201Z\017(\023\265\227\364P\033@\245\007\364\nP\033C\245\002t\036P+@e\036\364\016Pk@e\002t:P;A\245\001\364\nS\373@\245\006t\006P+C\345\002\364\006Pk@\345\036t\nP\033A\245\003\364:P\033@\245\006t\016ZJ\377\\\205\313\210\327\270\017\345+F[\310\347E)\355\024\241\366\342}(v\215(N\203)\326\207(\336\203(V\332W\362\202t4\240m\005(\377AJ\006\320\322\202t\374\240\255\003(oA\312:\3202"),
|
||
},
|
||
},
|
||
},
|
||
},
|
||
},
|
||
{
|
||
ChunkedSeries: []*prompb.ChunkedSeries{
|
||
{
|
||
Labels: []prompb.Label{
|
||
{Name: "__name__", Value: "test_metric1"},
|
||
{Name: "b", Value: "c"},
|
||
{Name: "baz", Value: "qux"},
|
||
{Name: "d", Value: "e"},
|
||
{Name: "foo", Value: "bar2"},
|
||
},
|
||
Chunks: []prompb.Chunk{
|
||
{
|
||
Type: prompb.Chunk_XOR,
|
||
MaxTimeMs: 7140000,
|
||
Data: []byte("\000x\000\000\000\000\000\000\000\000\000\340\324\003\302|\005\224\000\301\254}\351z2\320O\355\264n[\007\316\224\243md\371\320\375\032Pm\nS\235\016Q\255\006P\275\250\277\312\201Z\003(3\240R\207\332\005(\017\240\322\201\332=(\023\2402\203Z\007(w\2402\201Z\017(\023\265\227\364P\033@\245\007\364\nP\033C\245\002t\036P+@e\036\364\016Pk@e\002t:P;A\245\001\364\nS\373@\245\006t\006P+C\345\002\364\006Pk@\345\036t\nP\033A\245\003\364:P\033@\245\006t\016ZJ\377\\\205\313\210\327\270\017\345+F[\310\347E)\355\024\241\366\342}(v\215(N\203)\326\207(\336\203(V\332W\362\202t4\240m\005(\377AJ\006\320\322\202t\374\240\255\003(oA\312:\3202"),
|
||
},
|
||
{
|
||
Type: prompb.Chunk_XOR,
|
||
MinTimeMs: 7200000,
|
||
MaxTimeMs: 7200000,
|
||
Data: []byte("\000\001\200\364\356\006@\307p\000\000\000\000\000\000"),
|
||
},
|
||
},
|
||
},
|
||
},
|
||
},
|
||
{
|
||
ChunkedSeries: []*prompb.ChunkedSeries{
|
||
{
|
||
Labels: []prompb.Label{
|
||
{Name: "__name__", Value: "test_metric1"},
|
||
{Name: "b", Value: "c"},
|
||
{Name: "baz", Value: "qux"},
|
||
{Name: "d", Value: "e"},
|
||
{Name: "foo", Value: "bar3"},
|
||
},
|
||
Chunks: []prompb.Chunk{
|
||
{
|
||
Type: prompb.Chunk_XOR,
|
||
MaxTimeMs: 7140000,
|
||
Data: []byte("\000x\000\000\000\000\000\000\000\000\000\340\324\003\302|\005\224\000\301\254}\351z2\320O\355\264n[\007\316\224\243md\371\320\375\032Pm\nS\235\016Q\255\006P\275\250\277\312\201Z\003(3\240R\207\332\005(\017\240\322\201\332=(\023\2402\203Z\007(w\2402\201Z\017(\023\265\227\364P\033@\245\007\364\nP\033C\245\002t\036P+@e\036\364\016Pk@e\002t:P;A\245\001\364\nS\373@\245\006t\006P+C\345\002\364\006Pk@\345\036t\nP\033A\245\003\364:P\033@\245\006t\016ZJ\377\\\205\313\210\327\270\017\345+F[\310\347E)\355\024\241\366\342}(v\215(N\203)\326\207(\336\203(V\332W\362\202t4\240m\005(\377AJ\006\320\322\202t\374\240\255\003(oA\312:\3202"),
|
||
},
|
||
{
|
||
Type: prompb.Chunk_XOR,
|
||
MinTimeMs: 7200000,
|
||
MaxTimeMs: 14340000,
|
||
Data: []byte("\000x\200\364\356\006@\307p\000\000\000\000\000\340\324\003\340>\224\355\260\277\322\200\372\005(=\240R\207:\003(\025\240\362\201z\003(\365\240r\203:\005(\r\241\322\201\372\r(\r\240R\237:\007(5\2402\201z\037(\025\2402\203:\005(\375\240R\200\372\r(\035\241\322\201:\003(5\240r\326g\364\271\213\227!\253q\037\312N\340GJ\033E)\375\024\241\266\362}(N\217(V\203)\336\207(\326\203(N\334W\322\203\2644\240}\005(\373AJ\031\3202\202\264\374\240\275\003(kA\3129\320R\201\2644\240\375\264\277\322\200\332\005(3\240r\207Z\003(\027\240\362\201Z\003(\363\240R\203\332\005(\017\241\322\201\332\r(\023\2402\237Z\007(7\2402\201Z\037(\023\240\322\200\332\005(\377\240R\200\332\r "),
|
||
},
|
||
},
|
||
},
|
||
},
|
||
},
|
||
{
|
||
ChunkedSeries: []*prompb.ChunkedSeries{
|
||
{
|
||
Labels: []prompb.Label{
|
||
{Name: "__name__", Value: "test_metric1"},
|
||
{Name: "b", Value: "c"},
|
||
{Name: "baz", Value: "qux"},
|
||
{Name: "d", Value: "e"},
|
||
{Name: "foo", Value: "bar3"},
|
||
},
|
||
Chunks: []prompb.Chunk{
|
||
{
|
||
Type: prompb.Chunk_XOR,
|
||
MinTimeMs: 14400000,
|
||
MaxTimeMs: 14400000,
|
||
Data: []byte("\000\001\200\350\335\r@\327p\000\000\000\000\000\000"),
|
||
},
|
||
},
|
||
},
|
||
},
|
||
},
|
||
{
|
||
ChunkedSeries: []*prompb.ChunkedSeries{
|
||
{
|
||
Labels: []prompb.Label{
|
||
{Name: "__name__", Value: "test_metric1"},
|
||
{Name: "b", Value: "c"},
|
||
{Name: "baz", Value: "qux"},
|
||
{Name: "d", Value: "e"},
|
||
{Name: "foo", Value: "bar1"},
|
||
},
|
||
Chunks: []prompb.Chunk{
|
||
{
|
||
Type: prompb.Chunk_XOR,
|
||
MaxTimeMs: 7140000,
|
||
Data: []byte("\000x\000\000\000\000\000\000\000\000\000\340\324\003\302|\005\224\000\301\254}\351z2\320O\355\264n[\007\316\224\243md\371\320\375\032Pm\nS\235\016Q\255\006P\275\250\277\312\201Z\003(3\240R\207\332\005(\017\240\322\201\332=(\023\2402\203Z\007(w\2402\201Z\017(\023\265\227\364P\033@\245\007\364\nP\033C\245\002t\036P+@e\036\364\016Pk@e\002t:P;A\245\001\364\nS\373@\245\006t\006P+C\345\002\364\006Pk@\345\036t\nP\033A\245\003\364:P\033@\245\006t\016ZJ\377\\\205\313\210\327\270\017\345+F[\310\347E)\355\024\241\366\342}(v\215(N\203)\326\207(\336\203(V\332W\362\202t4\240m\005(\377AJ\006\320\322\202t\374\240\255\003(oA\312:\3202"),
|
||
},
|
||
},
|
||
},
|
||
},
|
||
QueryIndex: 1,
|
||
},
|
||
{
|
||
ChunkedSeries: []*prompb.ChunkedSeries{
|
||
{
|
||
Labels: []prompb.Label{
|
||
{Name: "__name__", Value: "test_histogram_metric1"},
|
||
{Name: "b", Value: "c"},
|
||
{Name: "baz", Value: "qux"},
|
||
{Name: "d", Value: "e"},
|
||
},
|
||
Chunks: []prompb.Chunk{
|
||
{
|
||
Type: prompb.Chunk_FLOAT_HISTOGRAM,
|
||
MaxTimeMs: 1440000,
|
||
Data: []byte("\x00\x19\x00\xff?PbM\xd2\xf1\xa9\xfc\x8c\xa4\x94e$\xa2@(\x00\x00\x00\x00\x00\x00@\x00\x00\x00\x00\x00\x00\x00@2ffffff?\xf0\x00\x00\x00\x00\x00\x00@\x00\x00\x00\x00\x00\x00\x00?\xf0\x00\x00\x00\x00\x00\x00?\xf0\x00\x00\x00\x00\x00\x00?\xf0\x00\x00\x00\x00\x00\x00@\x00\x00\x00\x00\x00\x00\x00?\xf0\x00\x00\x00\x00\x00\x00?\xf0\x00\x00\x00\x00\x00\x00\xf8\xea`\xd6/v\x03\xd2\x1f\xc2_\xff\xd8\x0f\t\x7f\xff\t\x7f\xff\t\x7f\xff`<%\xff\xfc%\xff\xf4\xbda{4\x9f\xff\xff\xff\xff\xff\xfd\x80\xf5\x85\xec\a\xb0\x1e\xc0z\xc2\xf6\x03\xd8\r\xa4\x8f\xbd\xa0\xf5\xeb\x9f\xff\xff\xff\xff\xff\xfda{A\xeb\v\xd6\x17\xac/h=az\xc2о\xc0\xb8\xac\xcc\xcc\xcc\xcc\xcc\xdbA\xec\v\xda\x0fh=\xa0\xf6\x05\xed\a\xb4\x1a\t\x99\x9333333;\x02\xe7`^\xc0\xbd\x81s\xb0/`Z8\xd4'\xea\x1f\xbc\xea\x13\xe6gP\x9f2\xdc$\xee\a\xbb)\xff\xff\xff\xff\xff\xffP\x9f\xb8\x1e\xa1?P\x9f\xa8O\xdc\x0fP\x9f\xa8Om\x17\xfbB\xf6\xe7\xb5UUUUUw\x03\xda\x17\xb8\x1e\xe0{\x81\xed\v\xdc\x0fp4\x99\x9d\x99\x99\x99\x99\x99\x9eй\xda\x17\xb4/h\\\xed\v\xda\x16\xd87{\x03\xfb2\xe4\xcc\xcc\xcc\xcc\xcc\xe7`~fv\a\xe6Q1ݕ\xaa\xaa\xaa\xaa\xaa\xab\xb0?\x1d\x81\xfd\x81\xfd\x81\xf8\xec\x0f\xec\x0f\xa5\xe7\xb7<\xff\xff\xff\xff\xff\xff\x19\xc61\x9cb\xd4O\xc6:\xf5\xef\xff\xff\xff\xff\xff\xfc\xe39\xce3\x9a\x05\xeb\x13\xe0\xac\xcc\xcc\xcc\xcc\xcc\xc7X\x9f\x18\xc7X\x9f\x18\xa0\xce\xf0pə\x99\x99\x99\x99\xb5\x89\xfb\xc1\xeb\x13\xf5\x89\xfa\xc4\xfd\xe0\xf5\x89\xfa\xc4\xf4\x0f\xdc\x17\a\xaa\xaa\xaa\xaa\xaa\xabx=\xc1{\xc1\xef\a\xbc\x1e\xe0\xbd\xe0\xf7\x83C\x99\x8e\x7f\xff\xff\xff\xff\xff\xb8.w\x05\xee\v\xdc\x17;\x82\xf7\x05\xa0^\xd0\xfc\x16\xaa\xaa\xaa\xaa\xaa\xa9\xda\x1f\x99\x9d\xa1\xf9\x94\x19\x8c-\x99\x99\x99\x99\x99\x9d\xa1\xf8\xed\x0f\xed\x0f\xed\x0f\xc7h\x7fh}\x1d\xe7<\x99\x99\x99\x99\x99\x9a3\x8cc8\xc5\x02c\x05\xaa\xaa\xaa\xaa\xaa\xaaq\x9c\xe7\x19\xcd\x06\xf6\t\xf0\xcf\xff\xff\xff\xff\xff\xe3\xb0O\x8cc\xb0O\x8cP&\x18=UUUUU[\x04\xf8v\t\xfb\x04\xfd\x82|;\x04\xfd\x82z\x1f\xc7\x1c\x99\x99\x99\x99\x99\x9a\x18\xe1\x86\x18\xe1\x84"),
|
||
},
|
||
},
|
||
},
|
||
},
|
||
QueryIndex: 2,
|
||
},
|
||
}, results)
|
||
}
|
||
|
||
func addNativeHistogramsToTestSuite(t *testing.T, storage *teststorage.TestStorage, n int) {
|
||
lbls := labels.FromStrings("__name__", "test_histogram_metric1", "baz", "qux")
|
||
|
||
app := storage.Appender(context.TODO())
|
||
for i, fh := range tsdbutil.GenerateTestFloatHistograms(n) {
|
||
_, err := app.AppendHistogram(0, lbls, int64(i)*int64(60*time.Second/time.Millisecond), nil, fh)
|
||
require.NoError(t, err)
|
||
}
|
||
require.NoError(t, app.Commit())
|
||
}
|