mirror of https://github.com/prometheus/prometheus
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
3662 lines
92 KiB
3662 lines
92 KiB
// Copyright 2016 The Prometheus Authors |
|
// Licensed under the Apache License, Version 2.0 (the "License"); |
|
// you may not use this file except in compliance with the License. |
|
// You may obtain a copy of the License at |
|
// |
|
// http://www.apache.org/licenses/LICENSE-2.0 |
|
// |
|
// Unless required by applicable law or agreed to in writing, software |
|
// distributed under the License is distributed on an "AS IS" BASIS, |
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
|
// See the License for the specific language governing permissions and |
|
// limitations under the License. |
|
|
|
package v1 |
|
|
|
import ( |
|
"context" |
|
"encoding/json" |
|
"fmt" |
|
"io" |
|
"net/http" |
|
"net/http/httptest" |
|
"net/url" |
|
"os" |
|
"reflect" |
|
"runtime" |
|
"sort" |
|
"strings" |
|
"testing" |
|
"time" |
|
|
|
"github.com/prometheus/prometheus/prompb" |
|
"github.com/prometheus/prometheus/util/stats" |
|
|
|
"github.com/go-kit/log" |
|
"github.com/pkg/errors" |
|
"github.com/prometheus/client_golang/prometheus" |
|
config_util "github.com/prometheus/common/config" |
|
"github.com/prometheus/common/model" |
|
"github.com/prometheus/common/promlog" |
|
"github.com/prometheus/common/route" |
|
"github.com/stretchr/testify/require" |
|
|
|
"github.com/prometheus/prometheus/config" |
|
"github.com/prometheus/prometheus/model/exemplar" |
|
"github.com/prometheus/prometheus/model/labels" |
|
"github.com/prometheus/prometheus/model/textparse" |
|
"github.com/prometheus/prometheus/model/timestamp" |
|
"github.com/prometheus/prometheus/promql" |
|
"github.com/prometheus/prometheus/promql/parser" |
|
"github.com/prometheus/prometheus/rules" |
|
"github.com/prometheus/prometheus/scrape" |
|
"github.com/prometheus/prometheus/storage" |
|
"github.com/prometheus/prometheus/storage/remote" |
|
"github.com/prometheus/prometheus/tsdb" |
|
"github.com/prometheus/prometheus/util/teststorage" |
|
) |
|
|
|
var testEngine = promql.NewEngine(promql.EngineOpts{ |
|
Logger: nil, |
|
Reg: nil, |
|
MaxSamples: 10000, |
|
Timeout: 100 * time.Second, |
|
NoStepSubqueryIntervalFn: func(int64) int64 { return 60 * 1000 }, |
|
EnableAtModifier: true, |
|
EnableNegativeOffset: true, |
|
EnablePerStepStats: true, |
|
}) |
|
|
|
// testMetaStore satisfies the scrape.MetricMetadataStore interface. |
|
// It is used to inject specific metadata as part of a test case. |
|
type testMetaStore struct { |
|
Metadata []scrape.MetricMetadata |
|
} |
|
|
|
func (s *testMetaStore) ListMetadata() []scrape.MetricMetadata { |
|
return s.Metadata |
|
} |
|
|
|
func (s *testMetaStore) GetMetadata(metric string) (scrape.MetricMetadata, bool) { |
|
for _, m := range s.Metadata { |
|
if metric == m.Metric { |
|
return m, true |
|
} |
|
} |
|
|
|
return scrape.MetricMetadata{}, false |
|
} |
|
|
|
func (s *testMetaStore) SizeMetadata() int { return 0 } |
|
func (s *testMetaStore) LengthMetadata() int { return 0 } |
|
|
|
// testTargetRetriever represents a list of targets to scrape. |
|
// It is used to represent targets as part of test cases. |
|
type testTargetRetriever struct { |
|
activeTargets map[string][]*scrape.Target |
|
droppedTargets map[string][]*scrape.Target |
|
} |
|
|
|
type testTargetParams struct { |
|
Identifier string |
|
Labels labels.Labels |
|
DiscoveredLabels labels.Labels |
|
Params url.Values |
|
Reports []*testReport |
|
Active bool |
|
} |
|
|
|
type testReport struct { |
|
Start time.Time |
|
Duration time.Duration |
|
Error error |
|
} |
|
|
|
func newTestTargetRetriever(targetsInfo []*testTargetParams) *testTargetRetriever { |
|
var activeTargets map[string][]*scrape.Target |
|
var droppedTargets map[string][]*scrape.Target |
|
activeTargets = make(map[string][]*scrape.Target) |
|
droppedTargets = make(map[string][]*scrape.Target) |
|
|
|
for _, t := range targetsInfo { |
|
nt := scrape.NewTarget(t.Labels, t.DiscoveredLabels, t.Params) |
|
|
|
for _, r := range t.Reports { |
|
nt.Report(r.Start, r.Duration, r.Error) |
|
} |
|
|
|
if t.Active { |
|
activeTargets[t.Identifier] = []*scrape.Target{nt} |
|
} else { |
|
droppedTargets[t.Identifier] = []*scrape.Target{nt} |
|
} |
|
} |
|
|
|
return &testTargetRetriever{ |
|
activeTargets: activeTargets, |
|
droppedTargets: droppedTargets, |
|
} |
|
} |
|
|
|
var scrapeStart = time.Now().Add(-11 * time.Second) |
|
|
|
func (t testTargetRetriever) TargetsActive() map[string][]*scrape.Target { |
|
return t.activeTargets |
|
} |
|
|
|
func (t testTargetRetriever) TargetsDropped() map[string][]*scrape.Target { |
|
return t.droppedTargets |
|
} |
|
|
|
func (t testTargetRetriever) TargetsDroppedCounts() map[string]int { |
|
r := make(map[string]int) |
|
for k, v := range t.droppedTargets { |
|
r[k] = len(v) |
|
} |
|
return r |
|
} |
|
|
|
func (t *testTargetRetriever) SetMetadataStoreForTargets(identifier string, metadata scrape.MetricMetadataStore) error { |
|
targets, ok := t.activeTargets[identifier] |
|
|
|
if !ok { |
|
return errors.New("targets not found") |
|
} |
|
|
|
for _, at := range targets { |
|
at.SetMetadataStore(metadata) |
|
} |
|
|
|
return nil |
|
} |
|
|
|
func (t *testTargetRetriever) ResetMetadataStore() { |
|
for _, at := range t.activeTargets { |
|
for _, tt := range at { |
|
tt.SetMetadataStore(&testMetaStore{}) |
|
} |
|
} |
|
} |
|
|
|
func (t *testTargetRetriever) toFactory() func(context.Context) TargetRetriever { |
|
return func(context.Context) TargetRetriever { return t } |
|
} |
|
|
|
type testAlertmanagerRetriever struct{} |
|
|
|
func (t testAlertmanagerRetriever) Alertmanagers() []*url.URL { |
|
return []*url.URL{ |
|
{ |
|
Scheme: "http", |
|
Host: "alertmanager.example.com:8080", |
|
Path: "/api/v1/alerts", |
|
}, |
|
} |
|
} |
|
|
|
func (t testAlertmanagerRetriever) DroppedAlertmanagers() []*url.URL { |
|
return []*url.URL{ |
|
{ |
|
Scheme: "http", |
|
Host: "dropped.alertmanager.example.com:8080", |
|
Path: "/api/v1/alerts", |
|
}, |
|
} |
|
} |
|
|
|
func (t testAlertmanagerRetriever) toFactory() func(context.Context) AlertmanagerRetriever { |
|
return func(context.Context) AlertmanagerRetriever { return t } |
|
} |
|
|
|
type rulesRetrieverMock struct { |
|
testing *testing.T |
|
} |
|
|
|
func (m rulesRetrieverMock) AlertingRules() []*rules.AlertingRule { |
|
expr1, err := parser.ParseExpr(`absent(test_metric3) != 1`) |
|
if err != nil { |
|
m.testing.Fatalf("unable to parse alert expression: %s", err) |
|
} |
|
expr2, err := parser.ParseExpr(`up == 1`) |
|
if err != nil { |
|
m.testing.Fatalf("Unable to parse alert expression: %s", err) |
|
} |
|
|
|
rule1 := rules.NewAlertingRule( |
|
"test_metric3", |
|
expr1, |
|
time.Second, |
|
0, |
|
labels.Labels{}, |
|
labels.Labels{}, |
|
labels.Labels{}, |
|
"", |
|
true, |
|
log.NewNopLogger(), |
|
) |
|
rule2 := rules.NewAlertingRule( |
|
"test_metric4", |
|
expr2, |
|
time.Second, |
|
0, |
|
labels.Labels{}, |
|
labels.Labels{}, |
|
labels.Labels{}, |
|
"", |
|
true, |
|
log.NewNopLogger(), |
|
) |
|
var r []*rules.AlertingRule |
|
r = append(r, rule1) |
|
r = append(r, rule2) |
|
return r |
|
} |
|
|
|
func (m rulesRetrieverMock) RuleGroups() []*rules.Group { |
|
var ar rulesRetrieverMock |
|
arules := ar.AlertingRules() |
|
storage := teststorage.New(m.testing) |
|
defer storage.Close() |
|
|
|
engineOpts := promql.EngineOpts{ |
|
Logger: nil, |
|
Reg: nil, |
|
MaxSamples: 10, |
|
Timeout: 100 * time.Second, |
|
} |
|
|
|
engine := promql.NewEngine(engineOpts) |
|
opts := &rules.ManagerOptions{ |
|
QueryFunc: rules.EngineQueryFunc(engine, storage), |
|
Appendable: storage, |
|
Context: context.Background(), |
|
Logger: log.NewNopLogger(), |
|
} |
|
|
|
var r []rules.Rule |
|
|
|
for _, alertrule := range arules { |
|
r = append(r, alertrule) |
|
} |
|
|
|
recordingExpr, err := parser.ParseExpr(`vector(1)`) |
|
if err != nil { |
|
m.testing.Fatalf("unable to parse alert expression: %s", err) |
|
} |
|
recordingRule := rules.NewRecordingRule("recording-rule-1", recordingExpr, labels.Labels{}) |
|
r = append(r, recordingRule) |
|
|
|
group := rules.NewGroup(rules.GroupOptions{ |
|
Name: "grp", |
|
File: "/path/to/file", |
|
Interval: time.Second, |
|
Rules: r, |
|
ShouldRestore: false, |
|
Opts: opts, |
|
}) |
|
return []*rules.Group{group} |
|
} |
|
|
|
func (m rulesRetrieverMock) toFactory() func(context.Context) RulesRetriever { |
|
return func(context.Context) RulesRetriever { return m } |
|
} |
|
|
|
var samplePrometheusCfg = config.Config{ |
|
GlobalConfig: config.GlobalConfig{}, |
|
AlertingConfig: config.AlertingConfig{}, |
|
RuleFiles: []string{}, |
|
ScrapeConfigs: []*config.ScrapeConfig{}, |
|
RemoteWriteConfigs: []*config.RemoteWriteConfig{}, |
|
RemoteReadConfigs: []*config.RemoteReadConfig{}, |
|
} |
|
|
|
var sampleFlagMap = map[string]string{ |
|
"flag1": "value1", |
|
"flag2": "value2", |
|
} |
|
|
|
func TestEndpoints(t *testing.T) { |
|
storage := promql.LoadedStorage(t, ` |
|
load 1m |
|
test_metric1{foo="bar"} 0+100x100 |
|
test_metric1{foo="boo"} 1+0x100 |
|
test_metric2{foo="boo"} 1+0x100 |
|
test_metric3{foo="bar", dup="1"} 1+0x100 |
|
test_metric3{foo="boo", dup="1"} 1+0x100 |
|
test_metric4{foo="bar", dup="1"} 1+0x100 |
|
test_metric4{foo="boo", dup="1"} 1+0x100 |
|
test_metric4{foo="boo"} 1+0x100 |
|
`) |
|
t.Cleanup(func() { storage.Close() }) |
|
|
|
start := time.Unix(0, 0) |
|
exemplars := []exemplar.QueryResult{ |
|
{ |
|
SeriesLabels: labels.FromStrings("__name__", "test_metric3", "foo", "boo", "dup", "1"), |
|
Exemplars: []exemplar.Exemplar{ |
|
{ |
|
Labels: labels.FromStrings("id", "abc"), |
|
Value: 10, |
|
Ts: timestamp.FromTime(start.Add(2 * time.Second)), |
|
}, |
|
}, |
|
}, |
|
{ |
|
SeriesLabels: labels.FromStrings("__name__", "test_metric4", "foo", "bar", "dup", "1"), |
|
Exemplars: []exemplar.Exemplar{ |
|
{ |
|
Labels: labels.FromStrings("id", "lul"), |
|
Value: 10, |
|
Ts: timestamp.FromTime(start.Add(4 * time.Second)), |
|
}, |
|
}, |
|
}, |
|
{ |
|
SeriesLabels: labels.FromStrings("__name__", "test_metric3", "foo", "boo", "dup", "1"), |
|
Exemplars: []exemplar.Exemplar{ |
|
{ |
|
Labels: labels.FromStrings("id", "abc2"), |
|
Value: 10, |
|
Ts: timestamp.FromTime(start.Add(4053 * time.Millisecond)), |
|
}, |
|
}, |
|
}, |
|
{ |
|
SeriesLabels: labels.FromStrings("__name__", "test_metric4", "foo", "bar", "dup", "1"), |
|
Exemplars: []exemplar.Exemplar{ |
|
{ |
|
Labels: labels.FromStrings("id", "lul2"), |
|
Value: 10, |
|
Ts: timestamp.FromTime(start.Add(4153 * time.Millisecond)), |
|
}, |
|
}, |
|
}, |
|
} |
|
for _, ed := range exemplars { |
|
_, err := storage.AppendExemplar(0, ed.SeriesLabels, ed.Exemplars[0]) |
|
require.NoError(t, err, "failed to add exemplar: %+v", ed.Exemplars[0]) |
|
} |
|
|
|
now := time.Now() |
|
|
|
t.Run("local", func(t *testing.T) { |
|
var algr rulesRetrieverMock |
|
algr.testing = t |
|
|
|
algr.AlertingRules() |
|
|
|
algr.RuleGroups() |
|
|
|
testTargetRetriever := setupTestTargetRetriever(t) |
|
|
|
api := &API{ |
|
Queryable: storage, |
|
QueryEngine: testEngine, |
|
ExemplarQueryable: storage.ExemplarQueryable(), |
|
targetRetriever: testTargetRetriever.toFactory(), |
|
alertmanagerRetriever: testAlertmanagerRetriever{}.toFactory(), |
|
flagsMap: sampleFlagMap, |
|
now: func() time.Time { return now }, |
|
config: func() config.Config { return samplePrometheusCfg }, |
|
ready: func(f http.HandlerFunc) http.HandlerFunc { return f }, |
|
rulesRetriever: algr.toFactory(), |
|
} |
|
testEndpoints(t, api, testTargetRetriever, storage, true) |
|
}) |
|
|
|
// Run all the API tests against a API that is wired to forward queries via |
|
// the remote read client to a test server, which in turn sends them to the |
|
// data from the test storage. |
|
t.Run("remote", func(t *testing.T) { |
|
server := setupRemote(storage) |
|
defer server.Close() |
|
|
|
u, err := url.Parse(server.URL) |
|
require.NoError(t, err) |
|
|
|
al := promlog.AllowedLevel{} |
|
require.NoError(t, al.Set("debug")) |
|
|
|
af := promlog.AllowedFormat{} |
|
require.NoError(t, af.Set("logfmt")) |
|
|
|
promlogConfig := promlog.Config{ |
|
Level: &al, |
|
Format: &af, |
|
} |
|
|
|
dbDir := t.TempDir() |
|
|
|
remote := remote.NewStorage(promlog.New(&promlogConfig), prometheus.DefaultRegisterer, func() (int64, error) { |
|
return 0, nil |
|
}, dbDir, 1*time.Second, nil) |
|
|
|
err = remote.ApplyConfig(&config.Config{ |
|
RemoteReadConfigs: []*config.RemoteReadConfig{ |
|
{ |
|
URL: &config_util.URL{URL: u}, |
|
RemoteTimeout: model.Duration(1 * time.Second), |
|
ReadRecent: true, |
|
}, |
|
}, |
|
}) |
|
require.NoError(t, err) |
|
|
|
var algr rulesRetrieverMock |
|
algr.testing = t |
|
|
|
algr.AlertingRules() |
|
|
|
algr.RuleGroups() |
|
|
|
testTargetRetriever := setupTestTargetRetriever(t) |
|
|
|
api := &API{ |
|
Queryable: remote, |
|
QueryEngine: testEngine, |
|
ExemplarQueryable: storage.ExemplarQueryable(), |
|
targetRetriever: testTargetRetriever.toFactory(), |
|
alertmanagerRetriever: testAlertmanagerRetriever{}.toFactory(), |
|
flagsMap: sampleFlagMap, |
|
now: func() time.Time { return now }, |
|
config: func() config.Config { return samplePrometheusCfg }, |
|
ready: func(f http.HandlerFunc) http.HandlerFunc { return f }, |
|
rulesRetriever: algr.toFactory(), |
|
} |
|
testEndpoints(t, api, testTargetRetriever, storage, false) |
|
}) |
|
} |
|
|
|
type byLabels []labels.Labels |
|
|
|
func (b byLabels) Len() int { return len(b) } |
|
func (b byLabels) Swap(i, j int) { b[i], b[j] = b[j], b[i] } |
|
func (b byLabels) Less(i, j int) bool { return labels.Compare(b[i], b[j]) < 0 } |
|
|
|
func TestGetSeries(t *testing.T) { |
|
// TestEndpoints doesn't have enough label names to test api.labelNames |
|
// endpoint properly. Hence we test it separately. |
|
storage := promql.LoadedStorage(t, ` |
|
load 1m |
|
test_metric1{foo1="bar", baz="abc"} 0+100x100 |
|
test_metric1{foo2="boo"} 1+0x100 |
|
test_metric2{foo="boo"} 1+0x100 |
|
test_metric2{foo="boo", xyz="qwerty"} 1+0x100 |
|
test_metric2{foo="baz", abc="qwerty"} 1+0x100 |
|
`) |
|
t.Cleanup(func() { storage.Close() }) |
|
api := &API{ |
|
Queryable: storage, |
|
} |
|
request := func(method string, matchers ...string) (*http.Request, error) { |
|
u, err := url.Parse("http://example.com") |
|
require.NoError(t, err) |
|
q := u.Query() |
|
for _, matcher := range matchers { |
|
q.Add("match[]", matcher) |
|
} |
|
u.RawQuery = q.Encode() |
|
|
|
r, err := http.NewRequest(method, u.String(), nil) |
|
if method == http.MethodPost { |
|
r.Header.Set("Content-Type", "application/x-www-form-urlencoded") |
|
} |
|
return r, err |
|
} |
|
|
|
for _, tc := range []struct { |
|
name string |
|
api *API |
|
matchers []string |
|
expected []labels.Labels |
|
expectedErrorType errorType |
|
}{ |
|
{ |
|
name: "no matchers", |
|
expectedErrorType: errorBadData, |
|
api: api, |
|
}, |
|
{ |
|
name: "non empty label matcher", |
|
matchers: []string{`{foo=~".+"}`}, |
|
expected: []labels.Labels{ |
|
labels.FromStrings("__name__", "test_metric2", "abc", "qwerty", "foo", "baz"), |
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"), |
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo", "xyz", "qwerty"), |
|
}, |
|
api: api, |
|
}, |
|
{ |
|
name: "exact label matcher", |
|
matchers: []string{`{foo="boo"}`}, |
|
expected: []labels.Labels{ |
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"), |
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo", "xyz", "qwerty"), |
|
}, |
|
api: api, |
|
}, |
|
{ |
|
name: "two matchers", |
|
matchers: []string{`{foo="boo"}`, `{foo="baz"}`}, |
|
expected: []labels.Labels{ |
|
labels.FromStrings("__name__", "test_metric2", "abc", "qwerty", "foo", "baz"), |
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"), |
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo", "xyz", "qwerty"), |
|
}, |
|
api: api, |
|
}, |
|
{ |
|
name: "exec error type", |
|
matchers: []string{`{foo="boo"}`, `{foo="baz"}`}, |
|
expectedErrorType: errorExec, |
|
api: &API{ |
|
Queryable: errorTestQueryable{err: fmt.Errorf("generic")}, |
|
}, |
|
}, |
|
{ |
|
name: "storage error type", |
|
matchers: []string{`{foo="boo"}`, `{foo="baz"}`}, |
|
expectedErrorType: errorInternal, |
|
api: &API{ |
|
Queryable: errorTestQueryable{err: promql.ErrStorage{Err: fmt.Errorf("generic")}}, |
|
}, |
|
}, |
|
} { |
|
t.Run(tc.name, func(t *testing.T) { |
|
ctx := context.Background() |
|
req, err := request(http.MethodGet, tc.matchers...) |
|
require.NoError(t, err) |
|
res := tc.api.series(req.WithContext(ctx)) |
|
assertAPIError(t, res.err, tc.expectedErrorType) |
|
if tc.expectedErrorType == errorNone { |
|
r := res.data.([]labels.Labels) |
|
sort.Sort(byLabels(tc.expected)) |
|
sort.Sort(byLabels(r)) |
|
require.Equal(t, tc.expected, r) |
|
} |
|
}) |
|
} |
|
} |
|
|
|
func TestQueryExemplars(t *testing.T) { |
|
start := time.Unix(0, 0) |
|
storage := promql.LoadedStorage(t, ` |
|
load 1m |
|
test_metric1{foo="bar"} 0+100x100 |
|
test_metric1{foo="boo"} 1+0x100 |
|
test_metric2{foo="boo"} 1+0x100 |
|
test_metric3{foo="bar", dup="1"} 1+0x100 |
|
test_metric3{foo="boo", dup="1"} 1+0x100 |
|
test_metric4{foo="bar", dup="1"} 1+0x100 |
|
test_metric4{foo="boo", dup="1"} 1+0x100 |
|
test_metric4{foo="boo"} 1+0x100 |
|
`) |
|
t.Cleanup(func() { storage.Close() }) |
|
|
|
api := &API{ |
|
Queryable: storage, |
|
QueryEngine: testEngine, |
|
ExemplarQueryable: storage.ExemplarQueryable(), |
|
} |
|
|
|
request := func(method string, qs url.Values) (*http.Request, error) { |
|
u, err := url.Parse("http://example.com") |
|
require.NoError(t, err) |
|
u.RawQuery = qs.Encode() |
|
r, err := http.NewRequest(method, u.String(), nil) |
|
if method == http.MethodPost { |
|
r.Header.Set("Content-Type", "application/x-www-form-urlencoded") |
|
} |
|
return r, err |
|
} |
|
|
|
for _, tc := range []struct { |
|
name string |
|
query url.Values |
|
exemplars []exemplar.QueryResult |
|
api *API |
|
expectedErrorType errorType |
|
}{ |
|
{ |
|
name: "no error", |
|
api: api, |
|
query: url.Values{ |
|
"query": []string{`test_metric3{foo="boo"} - test_metric4{foo="bar"}`}, |
|
"start": []string{"0"}, |
|
"end": []string{"4"}, |
|
}, |
|
exemplars: []exemplar.QueryResult{ |
|
{ |
|
SeriesLabels: labels.FromStrings("__name__", "test_metric3", "foo", "boo", "dup", "1"), |
|
Exemplars: []exemplar.Exemplar{ |
|
{ |
|
Labels: labels.FromStrings("id", "abc"), |
|
Value: 10, |
|
Ts: timestamp.FromTime(start.Add(0 * time.Second)), |
|
}, |
|
}, |
|
}, |
|
{ |
|
SeriesLabels: labels.FromStrings("__name__", "test_metric4", "foo", "bar", "dup", "1"), |
|
Exemplars: []exemplar.Exemplar{ |
|
{ |
|
Labels: labels.FromStrings("id", "lul"), |
|
Value: 10, |
|
Ts: timestamp.FromTime(start.Add(3 * time.Second)), |
|
}, |
|
}, |
|
}, |
|
}, |
|
}, |
|
{ |
|
name: "should return errorExec upon genetic error", |
|
expectedErrorType: errorExec, |
|
api: &API{ |
|
ExemplarQueryable: errorTestQueryable{err: fmt.Errorf("generic")}, |
|
}, |
|
query: url.Values{ |
|
"query": []string{`test_metric3{foo="boo"} - test_metric4{foo="bar"}`}, |
|
"start": []string{"0"}, |
|
"end": []string{"4"}, |
|
}, |
|
}, |
|
{ |
|
name: "should return errorInternal err type is ErrStorage", |
|
expectedErrorType: errorInternal, |
|
api: &API{ |
|
ExemplarQueryable: errorTestQueryable{err: promql.ErrStorage{Err: fmt.Errorf("generic")}}, |
|
}, |
|
query: url.Values{ |
|
"query": []string{`test_metric3{foo="boo"} - test_metric4{foo="bar"}`}, |
|
"start": []string{"0"}, |
|
"end": []string{"4"}, |
|
}, |
|
}, |
|
} { |
|
t.Run(tc.name, func(t *testing.T) { |
|
es := storage |
|
ctx := context.Background() |
|
|
|
for _, te := range tc.exemplars { |
|
for _, e := range te.Exemplars { |
|
_, err := es.AppendExemplar(0, te.SeriesLabels, e) |
|
if err != nil { |
|
t.Fatal(err) |
|
} |
|
} |
|
} |
|
|
|
req, err := request(http.MethodGet, tc.query) |
|
require.NoError(t, err) |
|
res := tc.api.queryExemplars(req.WithContext(ctx)) |
|
assertAPIError(t, res.err, tc.expectedErrorType) |
|
|
|
if tc.expectedErrorType == errorNone { |
|
assertAPIResponse(t, res.data, tc.exemplars) |
|
} |
|
}) |
|
} |
|
} |
|
|
|
func TestLabelNames(t *testing.T) { |
|
// TestEndpoints doesn't have enough label names to test api.labelNames |
|
// endpoint properly. Hence we test it separately. |
|
storage := promql.LoadedStorage(t, ` |
|
load 1m |
|
test_metric1{foo1="bar", baz="abc"} 0+100x100 |
|
test_metric1{foo2="boo"} 1+0x100 |
|
test_metric2{foo="boo"} 1+0x100 |
|
test_metric2{foo="boo", xyz="qwerty"} 1+0x100 |
|
test_metric2{foo="baz", abc="qwerty"} 1+0x100 |
|
`) |
|
t.Cleanup(func() { storage.Close() }) |
|
api := &API{ |
|
Queryable: storage, |
|
} |
|
request := func(method string, matchers ...string) (*http.Request, error) { |
|
u, err := url.Parse("http://example.com") |
|
require.NoError(t, err) |
|
q := u.Query() |
|
for _, matcher := range matchers { |
|
q.Add("match[]", matcher) |
|
} |
|
u.RawQuery = q.Encode() |
|
|
|
r, err := http.NewRequest(method, u.String(), nil) |
|
if method == http.MethodPost { |
|
r.Header.Set("Content-Type", "application/x-www-form-urlencoded") |
|
} |
|
return r, err |
|
} |
|
|
|
for _, tc := range []struct { |
|
name string |
|
api *API |
|
matchers []string |
|
expected []string |
|
expectedErrorType errorType |
|
}{ |
|
{ |
|
name: "no matchers", |
|
expected: []string{"__name__", "abc", "baz", "foo", "foo1", "foo2", "xyz"}, |
|
api: api, |
|
}, |
|
{ |
|
name: "non empty label matcher", |
|
matchers: []string{`{foo=~".+"}`}, |
|
expected: []string{"__name__", "abc", "foo", "xyz"}, |
|
api: api, |
|
}, |
|
{ |
|
name: "exact label matcher", |
|
matchers: []string{`{foo="boo"}`}, |
|
expected: []string{"__name__", "foo", "xyz"}, |
|
api: api, |
|
}, |
|
{ |
|
name: "two matchers", |
|
matchers: []string{`{foo="boo"}`, `{foo="baz"}`}, |
|
expected: []string{"__name__", "abc", "foo", "xyz"}, |
|
api: api, |
|
}, |
|
{ |
|
name: "exec error type", |
|
matchers: []string{`{foo="boo"}`, `{foo="baz"}`}, |
|
expectedErrorType: errorExec, |
|
api: &API{ |
|
Queryable: errorTestQueryable{err: fmt.Errorf("generic")}, |
|
}, |
|
}, |
|
{ |
|
name: "storage error type", |
|
matchers: []string{`{foo="boo"}`, `{foo="baz"}`}, |
|
expectedErrorType: errorInternal, |
|
api: &API{ |
|
Queryable: errorTestQueryable{err: promql.ErrStorage{Err: fmt.Errorf("generic")}}, |
|
}, |
|
}, |
|
} { |
|
t.Run(tc.name, func(t *testing.T) { |
|
for _, method := range []string{http.MethodGet, http.MethodPost} { |
|
ctx := context.Background() |
|
req, err := request(method, tc.matchers...) |
|
require.NoError(t, err) |
|
res := tc.api.labelNames(req.WithContext(ctx)) |
|
assertAPIError(t, res.err, tc.expectedErrorType) |
|
if tc.expectedErrorType == errorNone { |
|
assertAPIResponse(t, res.data, tc.expected) |
|
} |
|
} |
|
}) |
|
} |
|
} |
|
|
|
type testStats struct { |
|
Custom string `json:"custom"` |
|
} |
|
|
|
func (testStats) Builtin() (_ stats.BuiltinStats) { |
|
return |
|
} |
|
|
|
func TestStats(t *testing.T) { |
|
storage := teststorage.New(t) |
|
t.Cleanup(func() { storage.Close() }) |
|
|
|
api := &API{ |
|
Queryable: storage, |
|
QueryEngine: testEngine, |
|
now: func() time.Time { |
|
return time.Unix(123, 0) |
|
}, |
|
} |
|
request := func(method, param string) (*http.Request, error) { |
|
u, err := url.Parse("http://example.com") |
|
require.NoError(t, err) |
|
q := u.Query() |
|
q.Add("stats", param) |
|
q.Add("query", "up") |
|
q.Add("start", "0") |
|
q.Add("end", "100") |
|
q.Add("step", "10") |
|
u.RawQuery = q.Encode() |
|
|
|
r, err := http.NewRequest(method, u.String(), nil) |
|
if method == http.MethodPost { |
|
r.Header.Set("Content-Type", "application/x-www-form-urlencoded") |
|
} |
|
return r, err |
|
} |
|
|
|
for _, tc := range []struct { |
|
name string |
|
renderer StatsRenderer |
|
param string |
|
expected func(*testing.T, interface{}) |
|
}{ |
|
{ |
|
name: "stats is blank", |
|
param: "", |
|
expected: func(t *testing.T, i interface{}) { |
|
require.IsType(t, i, &QueryData{}) |
|
qd := i.(*QueryData) |
|
require.Nil(t, qd.Stats) |
|
}, |
|
}, |
|
{ |
|
name: "stats is true", |
|
param: "true", |
|
expected: func(t *testing.T, i interface{}) { |
|
require.IsType(t, i, &QueryData{}) |
|
qd := i.(*QueryData) |
|
require.NotNil(t, qd.Stats) |
|
qs := qd.Stats.Builtin() |
|
require.NotNil(t, qs.Timings) |
|
require.Greater(t, qs.Timings.EvalTotalTime, float64(0)) |
|
require.NotNil(t, qs.Samples) |
|
require.NotNil(t, qs.Samples.TotalQueryableSamples) |
|
require.Nil(t, qs.Samples.TotalQueryableSamplesPerStep) |
|
}, |
|
}, |
|
{ |
|
name: "stats is all", |
|
param: "all", |
|
expected: func(t *testing.T, i interface{}) { |
|
require.IsType(t, i, &QueryData{}) |
|
qd := i.(*QueryData) |
|
require.NotNil(t, qd.Stats) |
|
qs := qd.Stats.Builtin() |
|
require.NotNil(t, qs.Timings) |
|
require.Greater(t, qs.Timings.EvalTotalTime, float64(0)) |
|
require.NotNil(t, qs.Samples) |
|
require.NotNil(t, qs.Samples.TotalQueryableSamples) |
|
require.NotNil(t, qs.Samples.TotalQueryableSamplesPerStep) |
|
}, |
|
}, |
|
{ |
|
name: "custom handler with known value", |
|
renderer: func(ctx context.Context, s *stats.Statistics, p string) stats.QueryStats { |
|
if p == "known" { |
|
return testStats{"Custom Value"} |
|
} |
|
return nil |
|
}, |
|
param: "known", |
|
expected: func(t *testing.T, i interface{}) { |
|
require.IsType(t, i, &QueryData{}) |
|
qd := i.(*QueryData) |
|
require.NotNil(t, qd.Stats) |
|
j, err := json.Marshal(qd.Stats) |
|
require.NoError(t, err) |
|
require.JSONEq(t, string(j), `{"custom":"Custom Value"}`) |
|
}, |
|
}, |
|
} { |
|
t.Run(tc.name, func(t *testing.T) { |
|
before := api.statsRenderer |
|
defer func() { api.statsRenderer = before }() |
|
api.statsRenderer = tc.renderer |
|
|
|
for _, method := range []string{http.MethodGet, http.MethodPost} { |
|
ctx := context.Background() |
|
req, err := request(method, tc.param) |
|
require.NoError(t, err) |
|
res := api.query(req.WithContext(ctx)) |
|
assertAPIError(t, res.err, "") |
|
tc.expected(t, res.data) |
|
|
|
res = api.queryRange(req.WithContext(ctx)) |
|
assertAPIError(t, res.err, "") |
|
tc.expected(t, res.data) |
|
} |
|
}) |
|
} |
|
} |
|
|
|
func setupTestTargetRetriever(t *testing.T) *testTargetRetriever { |
|
t.Helper() |
|
|
|
targets := []*testTargetParams{ |
|
{ |
|
Identifier: "test", |
|
Labels: labels.FromMap(map[string]string{ |
|
model.SchemeLabel: "http", |
|
model.AddressLabel: "example.com:8080", |
|
model.MetricsPathLabel: "/metrics", |
|
model.JobLabel: "test", |
|
model.ScrapeIntervalLabel: "15s", |
|
model.ScrapeTimeoutLabel: "5s", |
|
}), |
|
DiscoveredLabels: labels.EmptyLabels(), |
|
Params: url.Values{}, |
|
Reports: []*testReport{{scrapeStart, 70 * time.Millisecond, nil}}, |
|
Active: true, |
|
}, |
|
{ |
|
Identifier: "blackbox", |
|
Labels: labels.FromMap(map[string]string{ |
|
model.SchemeLabel: "http", |
|
model.AddressLabel: "localhost:9115", |
|
model.MetricsPathLabel: "/probe", |
|
model.JobLabel: "blackbox", |
|
model.ScrapeIntervalLabel: "20s", |
|
model.ScrapeTimeoutLabel: "10s", |
|
}), |
|
DiscoveredLabels: labels.EmptyLabels(), |
|
Params: url.Values{"target": []string{"example.com"}}, |
|
Reports: []*testReport{{scrapeStart, 100 * time.Millisecond, errors.New("failed")}}, |
|
Active: true, |
|
}, |
|
{ |
|
Identifier: "blackbox", |
|
Labels: labels.EmptyLabels(), |
|
DiscoveredLabels: labels.FromMap(map[string]string{ |
|
model.SchemeLabel: "http", |
|
model.AddressLabel: "http://dropped.example.com:9115", |
|
model.MetricsPathLabel: "/probe", |
|
model.JobLabel: "blackbox", |
|
model.ScrapeIntervalLabel: "30s", |
|
model.ScrapeTimeoutLabel: "15s", |
|
}), |
|
Params: url.Values{}, |
|
Active: false, |
|
}, |
|
} |
|
|
|
return newTestTargetRetriever(targets) |
|
} |
|
|
|
func setupRemote(s storage.Storage) *httptest.Server { |
|
handler := http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { |
|
req, err := remote.DecodeReadRequest(r) |
|
if err != nil { |
|
http.Error(w, err.Error(), http.StatusBadRequest) |
|
return |
|
} |
|
resp := prompb.ReadResponse{ |
|
Results: make([]*prompb.QueryResult, len(req.Queries)), |
|
} |
|
for i, query := range req.Queries { |
|
matchers, err := remote.FromLabelMatchers(query.Matchers) |
|
if err != nil { |
|
http.Error(w, err.Error(), http.StatusBadRequest) |
|
return |
|
} |
|
|
|
var hints *storage.SelectHints |
|
if query.Hints != nil { |
|
hints = &storage.SelectHints{ |
|
Start: query.Hints.StartMs, |
|
End: query.Hints.EndMs, |
|
Step: query.Hints.StepMs, |
|
Func: query.Hints.Func, |
|
} |
|
} |
|
|
|
querier, err := s.Querier(query.StartTimestampMs, query.EndTimestampMs) |
|
if err != nil { |
|
http.Error(w, err.Error(), http.StatusInternalServerError) |
|
return |
|
} |
|
defer querier.Close() |
|
|
|
set := querier.Select(r.Context(), false, hints, matchers...) |
|
resp.Results[i], _, err = remote.ToQueryResult(set, 1e6) |
|
if err != nil { |
|
http.Error(w, err.Error(), http.StatusInternalServerError) |
|
return |
|
} |
|
} |
|
|
|
if err := remote.EncodeReadResponse(&resp, w); err != nil { |
|
http.Error(w, err.Error(), http.StatusInternalServerError) |
|
return |
|
} |
|
}) |
|
|
|
return httptest.NewServer(handler) |
|
} |
|
|
|
func testEndpoints(t *testing.T, api *API, tr *testTargetRetriever, es storage.ExemplarStorage, testLabelAPI bool) { |
|
start := time.Unix(0, 0) |
|
|
|
type targetMetadata struct { |
|
identifier string |
|
metadata []scrape.MetricMetadata |
|
} |
|
|
|
type test struct { |
|
endpoint apiFunc |
|
params map[string]string |
|
query url.Values |
|
response interface{} |
|
responseLen int |
|
responseMetadataTotal int |
|
errType errorType |
|
sorter func(interface{}) |
|
metadata []targetMetadata |
|
exemplars []exemplar.QueryResult |
|
} |
|
|
|
tests := []test{ |
|
{ |
|
endpoint: api.query, |
|
query: url.Values{ |
|
"query": []string{"2"}, |
|
"time": []string{"123.4"}, |
|
}, |
|
response: &QueryData{ |
|
ResultType: parser.ValueTypeScalar, |
|
Result: promql.Scalar{ |
|
V: 2, |
|
T: timestamp.FromTime(start.Add(123*time.Second + 400*time.Millisecond)), |
|
}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.query, |
|
query: url.Values{ |
|
"query": []string{"0.333"}, |
|
"time": []string{"1970-01-01T00:02:03Z"}, |
|
}, |
|
response: &QueryData{ |
|
ResultType: parser.ValueTypeScalar, |
|
Result: promql.Scalar{ |
|
V: 0.333, |
|
T: timestamp.FromTime(start.Add(123 * time.Second)), |
|
}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.query, |
|
query: url.Values{ |
|
"query": []string{"0.333"}, |
|
"time": []string{"1970-01-01T01:02:03+01:00"}, |
|
}, |
|
response: &QueryData{ |
|
ResultType: parser.ValueTypeScalar, |
|
Result: promql.Scalar{ |
|
V: 0.333, |
|
T: timestamp.FromTime(start.Add(123 * time.Second)), |
|
}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.query, |
|
query: url.Values{ |
|
"query": []string{"0.333"}, |
|
}, |
|
response: &QueryData{ |
|
ResultType: parser.ValueTypeScalar, |
|
Result: promql.Scalar{ |
|
V: 0.333, |
|
T: timestamp.FromTime(api.now()), |
|
}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.queryRange, |
|
query: url.Values{ |
|
"query": []string{"time()"}, |
|
"start": []string{"0"}, |
|
"end": []string{"2"}, |
|
"step": []string{"1"}, |
|
}, |
|
response: &QueryData{ |
|
ResultType: parser.ValueTypeMatrix, |
|
Result: promql.Matrix{ |
|
promql.Series{ |
|
Floats: []promql.FPoint{ |
|
{F: 0, T: timestamp.FromTime(start)}, |
|
{F: 1, T: timestamp.FromTime(start.Add(1 * time.Second))}, |
|
{F: 2, T: timestamp.FromTime(start.Add(2 * time.Second))}, |
|
}, |
|
// No Metric returned - use zero value for comparison. |
|
}, |
|
}, |
|
}, |
|
}, |
|
// Missing query params in range queries. |
|
{ |
|
endpoint: api.queryRange, |
|
query: url.Values{ |
|
"query": []string{"time()"}, |
|
"end": []string{"2"}, |
|
"step": []string{"1"}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
{ |
|
endpoint: api.queryRange, |
|
query: url.Values{ |
|
"query": []string{"time()"}, |
|
"start": []string{"0"}, |
|
"step": []string{"1"}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
{ |
|
endpoint: api.queryRange, |
|
query: url.Values{ |
|
"query": []string{"time()"}, |
|
"start": []string{"0"}, |
|
"end": []string{"2"}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
// Bad query expression. |
|
{ |
|
endpoint: api.query, |
|
query: url.Values{ |
|
"query": []string{"invalid][query"}, |
|
"time": []string{"1970-01-01T01:02:03+01:00"}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
{ |
|
endpoint: api.queryRange, |
|
query: url.Values{ |
|
"query": []string{"invalid][query"}, |
|
"start": []string{"0"}, |
|
"end": []string{"100"}, |
|
"step": []string{"1"}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
// Invalid step. |
|
{ |
|
endpoint: api.queryRange, |
|
query: url.Values{ |
|
"query": []string{"time()"}, |
|
"start": []string{"1"}, |
|
"end": []string{"2"}, |
|
"step": []string{"0"}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
// Start after end. |
|
{ |
|
endpoint: api.queryRange, |
|
query: url.Values{ |
|
"query": []string{"time()"}, |
|
"start": []string{"2"}, |
|
"end": []string{"1"}, |
|
"step": []string{"1"}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
// Start overflows int64 internally. |
|
{ |
|
endpoint: api.queryRange, |
|
query: url.Values{ |
|
"query": []string{"time()"}, |
|
"start": []string{"148966367200.372"}, |
|
"end": []string{"1489667272.372"}, |
|
"step": []string{"1"}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
{ |
|
endpoint: api.formatQuery, |
|
query: url.Values{ |
|
"query": []string{"foo+bar"}, |
|
}, |
|
response: "foo + bar", |
|
}, |
|
{ |
|
endpoint: api.formatQuery, |
|
query: url.Values{ |
|
"query": []string{"invalid_expression/"}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
{ |
|
endpoint: api.series, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric2`}, |
|
}, |
|
response: []labels.Labels{ |
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"), |
|
}, |
|
}, |
|
{ |
|
endpoint: api.series, |
|
query: url.Values{ |
|
"match[]": []string{`{foo=""}`}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
{ |
|
endpoint: api.series, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric1{foo=~".+o"}`}, |
|
}, |
|
response: []labels.Labels{ |
|
labels.FromStrings("__name__", "test_metric1", "foo", "boo"), |
|
}, |
|
}, |
|
{ |
|
endpoint: api.series, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric1{foo=~".+o$"}`, `test_metric1{foo=~".+o"}`}, |
|
}, |
|
response: []labels.Labels{ |
|
labels.FromStrings("__name__", "test_metric1", "foo", "boo"), |
|
}, |
|
}, |
|
// Try to overlap the selected series set as much as possible to test the result de-duplication works well. |
|
{ |
|
endpoint: api.series, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric4{foo=~".+o$"}`, `test_metric4{dup=~"^1"}`}, |
|
}, |
|
response: []labels.Labels{ |
|
labels.FromStrings("__name__", "test_metric4", "dup", "1", "foo", "bar"), |
|
labels.FromStrings("__name__", "test_metric4", "dup", "1", "foo", "boo"), |
|
labels.FromStrings("__name__", "test_metric4", "foo", "boo"), |
|
}, |
|
}, |
|
{ |
|
endpoint: api.series, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric1{foo=~".+o"}`, `none`}, |
|
}, |
|
response: []labels.Labels{ |
|
labels.FromStrings("__name__", "test_metric1", "foo", "boo"), |
|
}, |
|
}, |
|
// Start and end before series starts. |
|
{ |
|
endpoint: api.series, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric2`}, |
|
"start": []string{"-2"}, |
|
"end": []string{"-1"}, |
|
}, |
|
response: []labels.Labels{}, |
|
}, |
|
// Start and end after series ends. |
|
{ |
|
endpoint: api.series, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric2`}, |
|
"start": []string{"100000"}, |
|
"end": []string{"100001"}, |
|
}, |
|
response: []labels.Labels{}, |
|
}, |
|
// Start before series starts, end after series ends. |
|
{ |
|
endpoint: api.series, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric2`}, |
|
"start": []string{"-1"}, |
|
"end": []string{"100000"}, |
|
}, |
|
response: []labels.Labels{ |
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"), |
|
}, |
|
}, |
|
// Start and end within series. |
|
{ |
|
endpoint: api.series, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric2`}, |
|
"start": []string{"1"}, |
|
"end": []string{"100"}, |
|
}, |
|
response: []labels.Labels{ |
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"), |
|
}, |
|
}, |
|
// Start within series, end after. |
|
{ |
|
endpoint: api.series, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric2`}, |
|
"start": []string{"1"}, |
|
"end": []string{"100000"}, |
|
}, |
|
response: []labels.Labels{ |
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"), |
|
}, |
|
}, |
|
// Start before series, end within series. |
|
{ |
|
endpoint: api.series, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric2`}, |
|
"start": []string{"-1"}, |
|
"end": []string{"1"}, |
|
}, |
|
response: []labels.Labels{ |
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"), |
|
}, |
|
}, |
|
// Missing match[] query params in series requests. |
|
{ |
|
endpoint: api.series, |
|
errType: errorBadData, |
|
}, |
|
{ |
|
endpoint: api.dropSeries, |
|
errType: errorInternal, |
|
}, |
|
{ |
|
endpoint: api.targets, |
|
response: &TargetDiscovery{ |
|
ActiveTargets: []*Target{ |
|
{ |
|
DiscoveredLabels: map[string]string{}, |
|
Labels: map[string]string{ |
|
"job": "blackbox", |
|
}, |
|
ScrapePool: "blackbox", |
|
ScrapeURL: "http://localhost:9115/probe?target=example.com", |
|
GlobalURL: "http://localhost:9115/probe?target=example.com", |
|
Health: "down", |
|
LastError: "failed: missing port in address", |
|
LastScrape: scrapeStart, |
|
LastScrapeDuration: 0.1, |
|
ScrapeInterval: "20s", |
|
ScrapeTimeout: "10s", |
|
}, |
|
{ |
|
DiscoveredLabels: map[string]string{}, |
|
Labels: map[string]string{ |
|
"job": "test", |
|
}, |
|
ScrapePool: "test", |
|
ScrapeURL: "http://example.com:8080/metrics", |
|
GlobalURL: "http://example.com:8080/metrics", |
|
Health: "up", |
|
LastError: "", |
|
LastScrape: scrapeStart, |
|
LastScrapeDuration: 0.07, |
|
ScrapeInterval: "15s", |
|
ScrapeTimeout: "5s", |
|
}, |
|
}, |
|
DroppedTargets: []*DroppedTarget{ |
|
{ |
|
DiscoveredLabels: map[string]string{ |
|
"__address__": "http://dropped.example.com:9115", |
|
"__metrics_path__": "/probe", |
|
"__scheme__": "http", |
|
"job": "blackbox", |
|
"__scrape_interval__": "30s", |
|
"__scrape_timeout__": "15s", |
|
}, |
|
}, |
|
}, |
|
DroppedTargetCounts: map[string]int{"blackbox": 1}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.targets, |
|
query: url.Values{ |
|
"state": []string{"any"}, |
|
}, |
|
response: &TargetDiscovery{ |
|
ActiveTargets: []*Target{ |
|
{ |
|
DiscoveredLabels: map[string]string{}, |
|
Labels: map[string]string{ |
|
"job": "blackbox", |
|
}, |
|
ScrapePool: "blackbox", |
|
ScrapeURL: "http://localhost:9115/probe?target=example.com", |
|
GlobalURL: "http://localhost:9115/probe?target=example.com", |
|
Health: "down", |
|
LastError: "failed: missing port in address", |
|
LastScrape: scrapeStart, |
|
LastScrapeDuration: 0.1, |
|
ScrapeInterval: "20s", |
|
ScrapeTimeout: "10s", |
|
}, |
|
{ |
|
DiscoveredLabels: map[string]string{}, |
|
Labels: map[string]string{ |
|
"job": "test", |
|
}, |
|
ScrapePool: "test", |
|
ScrapeURL: "http://example.com:8080/metrics", |
|
GlobalURL: "http://example.com:8080/metrics", |
|
Health: "up", |
|
LastError: "", |
|
LastScrape: scrapeStart, |
|
LastScrapeDuration: 0.07, |
|
ScrapeInterval: "15s", |
|
ScrapeTimeout: "5s", |
|
}, |
|
}, |
|
DroppedTargets: []*DroppedTarget{ |
|
{ |
|
DiscoveredLabels: map[string]string{ |
|
"__address__": "http://dropped.example.com:9115", |
|
"__metrics_path__": "/probe", |
|
"__scheme__": "http", |
|
"job": "blackbox", |
|
"__scrape_interval__": "30s", |
|
"__scrape_timeout__": "15s", |
|
}, |
|
}, |
|
}, |
|
DroppedTargetCounts: map[string]int{"blackbox": 1}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.targets, |
|
query: url.Values{ |
|
"state": []string{"active"}, |
|
}, |
|
response: &TargetDiscovery{ |
|
ActiveTargets: []*Target{ |
|
{ |
|
DiscoveredLabels: map[string]string{}, |
|
Labels: map[string]string{ |
|
"job": "blackbox", |
|
}, |
|
ScrapePool: "blackbox", |
|
ScrapeURL: "http://localhost:9115/probe?target=example.com", |
|
GlobalURL: "http://localhost:9115/probe?target=example.com", |
|
Health: "down", |
|
LastError: "failed: missing port in address", |
|
LastScrape: scrapeStart, |
|
LastScrapeDuration: 0.1, |
|
ScrapeInterval: "20s", |
|
ScrapeTimeout: "10s", |
|
}, |
|
{ |
|
DiscoveredLabels: map[string]string{}, |
|
Labels: map[string]string{ |
|
"job": "test", |
|
}, |
|
ScrapePool: "test", |
|
ScrapeURL: "http://example.com:8080/metrics", |
|
GlobalURL: "http://example.com:8080/metrics", |
|
Health: "up", |
|
LastError: "", |
|
LastScrape: scrapeStart, |
|
LastScrapeDuration: 0.07, |
|
ScrapeInterval: "15s", |
|
ScrapeTimeout: "5s", |
|
}, |
|
}, |
|
DroppedTargets: []*DroppedTarget{}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.targets, |
|
query: url.Values{ |
|
"state": []string{"Dropped"}, |
|
}, |
|
response: &TargetDiscovery{ |
|
ActiveTargets: []*Target{}, |
|
DroppedTargets: []*DroppedTarget{ |
|
{ |
|
DiscoveredLabels: map[string]string{ |
|
"__address__": "http://dropped.example.com:9115", |
|
"__metrics_path__": "/probe", |
|
"__scheme__": "http", |
|
"job": "blackbox", |
|
"__scrape_interval__": "30s", |
|
"__scrape_timeout__": "15s", |
|
}, |
|
}, |
|
}, |
|
DroppedTargetCounts: map[string]int{"blackbox": 1}, |
|
}, |
|
}, |
|
// With a matching metric. |
|
{ |
|
endpoint: api.targetMetadata, |
|
query: url.Values{ |
|
"metric": []string{"go_threads"}, |
|
}, |
|
metadata: []targetMetadata{ |
|
{ |
|
identifier: "test", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Number of OS threads created.", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
}, |
|
response: []metricMetadata{ |
|
{ |
|
Target: labels.FromMap(map[string]string{ |
|
"job": "test", |
|
}), |
|
Help: "Number of OS threads created.", |
|
Type: textparse.MetricTypeGauge, |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
// With a matching target. |
|
{ |
|
endpoint: api.targetMetadata, |
|
query: url.Values{ |
|
"match_target": []string{"{job=\"blackbox\"}"}, |
|
}, |
|
metadata: []targetMetadata{ |
|
{ |
|
identifier: "blackbox", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "prometheus_tsdb_storage_blocks_bytes", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "The number of bytes that are currently used for local storage by all blocks.", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
}, |
|
response: []metricMetadata{ |
|
{ |
|
Target: labels.FromMap(map[string]string{ |
|
"job": "blackbox", |
|
}), |
|
Metric: "prometheus_tsdb_storage_blocks_bytes", |
|
Help: "The number of bytes that are currently used for local storage by all blocks.", |
|
Type: textparse.MetricTypeGauge, |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
// Without a target or metric. |
|
{ |
|
endpoint: api.targetMetadata, |
|
metadata: []targetMetadata{ |
|
{ |
|
identifier: "test", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Number of OS threads created.", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
{ |
|
identifier: "blackbox", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "prometheus_tsdb_storage_blocks_bytes", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "The number of bytes that are currently used for local storage by all blocks.", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
}, |
|
response: []metricMetadata{ |
|
{ |
|
Target: labels.FromMap(map[string]string{ |
|
"job": "test", |
|
}), |
|
Metric: "go_threads", |
|
Help: "Number of OS threads created.", |
|
Type: textparse.MetricTypeGauge, |
|
Unit: "", |
|
}, |
|
{ |
|
Target: labels.FromMap(map[string]string{ |
|
"job": "blackbox", |
|
}), |
|
Metric: "prometheus_tsdb_storage_blocks_bytes", |
|
Help: "The number of bytes that are currently used for local storage by all blocks.", |
|
Type: textparse.MetricTypeGauge, |
|
Unit: "", |
|
}, |
|
}, |
|
sorter: func(m interface{}) { |
|
sort.Slice(m.([]metricMetadata), func(i, j int) bool { |
|
s := m.([]metricMetadata) |
|
return s[i].Metric < s[j].Metric |
|
}) |
|
}, |
|
}, |
|
// Without a matching metric. |
|
{ |
|
endpoint: api.targetMetadata, |
|
query: url.Values{ |
|
"match_target": []string{"{job=\"non-existentblackbox\"}"}, |
|
}, |
|
response: []metricMetadata{}, |
|
}, |
|
{ |
|
endpoint: api.alertmanagers, |
|
response: &AlertmanagerDiscovery{ |
|
ActiveAlertmanagers: []*AlertmanagerTarget{ |
|
{ |
|
URL: "http://alertmanager.example.com:8080/api/v1/alerts", |
|
}, |
|
}, |
|
DroppedAlertmanagers: []*AlertmanagerTarget{ |
|
{ |
|
URL: "http://dropped.alertmanager.example.com:8080/api/v1/alerts", |
|
}, |
|
}, |
|
}, |
|
}, |
|
// With metadata available. |
|
{ |
|
endpoint: api.metricMetadata, |
|
metadata: []targetMetadata{ |
|
{ |
|
identifier: "test", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "prometheus_engine_query_duration_seconds", |
|
Type: textparse.MetricTypeSummary, |
|
Help: "Query timings", |
|
Unit: "", |
|
}, |
|
{ |
|
Metric: "go_info", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Information about the Go environment.", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
}, |
|
response: map[string][]metadata{ |
|
"prometheus_engine_query_duration_seconds": {{textparse.MetricTypeSummary, "Query timings", ""}}, |
|
"go_info": {{textparse.MetricTypeGauge, "Information about the Go environment.", ""}}, |
|
}, |
|
}, |
|
// With duplicate metadata for a metric that comes from different targets. |
|
{ |
|
endpoint: api.metricMetadata, |
|
metadata: []targetMetadata{ |
|
{ |
|
identifier: "test", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Number of OS threads created", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
{ |
|
identifier: "blackbox", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Number of OS threads created", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
}, |
|
response: map[string][]metadata{ |
|
"go_threads": {{textparse.MetricTypeGauge, "Number of OS threads created", ""}}, |
|
}, |
|
}, |
|
// With non-duplicate metadata for the same metric from different targets. |
|
{ |
|
endpoint: api.metricMetadata, |
|
metadata: []targetMetadata{ |
|
{ |
|
identifier: "test", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Number of OS threads created", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
{ |
|
identifier: "blackbox", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Number of OS threads that were created.", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
}, |
|
response: map[string][]metadata{ |
|
"go_threads": { |
|
{textparse.MetricTypeGauge, "Number of OS threads created", ""}, |
|
{textparse.MetricTypeGauge, "Number of OS threads that were created.", ""}, |
|
}, |
|
}, |
|
sorter: func(m interface{}) { |
|
v := m.(map[string][]metadata)["go_threads"] |
|
|
|
sort.Slice(v, func(i, j int) bool { |
|
return v[i].Help < v[j].Help |
|
}) |
|
}, |
|
}, |
|
// With a limit for the number of metrics returned. |
|
{ |
|
endpoint: api.metricMetadata, |
|
query: url.Values{ |
|
"limit": []string{"2"}, |
|
}, |
|
metadata: []targetMetadata{ |
|
{ |
|
identifier: "test", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Number of OS threads created", |
|
Unit: "", |
|
}, |
|
{ |
|
Metric: "prometheus_engine_query_duration_seconds", |
|
Type: textparse.MetricTypeSummary, |
|
Help: "Query Timmings.", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
{ |
|
identifier: "blackbox", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "go_gc_duration_seconds", |
|
Type: textparse.MetricTypeSummary, |
|
Help: "A summary of the GC invocation durations.", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
}, |
|
responseLen: 2, |
|
}, |
|
// With a limit for the number of metadata per metric. |
|
{ |
|
endpoint: api.metricMetadata, |
|
query: url.Values{"limit_per_metric": []string{"1"}}, |
|
metadata: []targetMetadata{ |
|
{ |
|
identifier: "test", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Number of OS threads created", |
|
Unit: "", |
|
}, |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Repeated metadata", |
|
Unit: "", |
|
}, |
|
{ |
|
Metric: "go_gc_duration_seconds", |
|
Type: textparse.MetricTypeSummary, |
|
Help: "A summary of the GC invocation durations.", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
}, |
|
response: map[string][]metadata{ |
|
"go_threads": { |
|
{textparse.MetricTypeGauge, "Number of OS threads created", ""}, |
|
}, |
|
"go_gc_duration_seconds": { |
|
{textparse.MetricTypeSummary, "A summary of the GC invocation durations.", ""}, |
|
}, |
|
}, |
|
}, |
|
// With a limit for the number of metadata per metric and per metric. |
|
{ |
|
endpoint: api.metricMetadata, |
|
query: url.Values{"limit_per_metric": []string{"1"}, "limit": []string{"1"}}, |
|
metadata: []targetMetadata{ |
|
{ |
|
identifier: "test", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Number of OS threads created", |
|
Unit: "", |
|
}, |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Repeated metadata", |
|
Unit: "", |
|
}, |
|
{ |
|
Metric: "go_gc_duration_seconds", |
|
Type: textparse.MetricTypeSummary, |
|
Help: "A summary of the GC invocation durations.", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
}, |
|
responseLen: 1, |
|
responseMetadataTotal: 1, |
|
}, |
|
|
|
// With a limit for the number of metadata per metric and per metric, while having multiple targets. |
|
{ |
|
endpoint: api.metricMetadata, |
|
query: url.Values{"limit_per_metric": []string{"1"}, "limit": []string{"1"}}, |
|
metadata: []targetMetadata{ |
|
{ |
|
identifier: "test", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Number of OS threads created", |
|
Unit: "", |
|
}, |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Repeated metadata", |
|
Unit: "", |
|
}, |
|
{ |
|
Metric: "go_gc_duration_seconds", |
|
Type: textparse.MetricTypeSummary, |
|
Help: "A summary of the GC invocation durations.", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
{ |
|
identifier: "secondTarget", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Number of OS threads created, but from a different target", |
|
Unit: "", |
|
}, |
|
{ |
|
Metric: "go_gc_duration_seconds", |
|
Type: textparse.MetricTypeSummary, |
|
Help: "A summary of the GC invocation durations, but from a different target.", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
}, |
|
responseLen: 1, |
|
responseMetadataTotal: 1, |
|
}, |
|
// When requesting a specific metric that is present. |
|
{ |
|
endpoint: api.metricMetadata, |
|
query: url.Values{"metric": []string{"go_threads"}}, |
|
metadata: []targetMetadata{ |
|
{ |
|
identifier: "test", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Number of OS threads created", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
{ |
|
identifier: "blackbox", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "go_gc_duration_seconds", |
|
Type: textparse.MetricTypeSummary, |
|
Help: "A summary of the GC invocation durations.", |
|
Unit: "", |
|
}, |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Number of OS threads that were created.", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
}, |
|
response: map[string][]metadata{ |
|
"go_threads": { |
|
{textparse.MetricTypeGauge, "Number of OS threads created", ""}, |
|
{textparse.MetricTypeGauge, "Number of OS threads that were created.", ""}, |
|
}, |
|
}, |
|
sorter: func(m interface{}) { |
|
v := m.(map[string][]metadata)["go_threads"] |
|
|
|
sort.Slice(v, func(i, j int) bool { |
|
return v[i].Help < v[j].Help |
|
}) |
|
}, |
|
}, |
|
// With a specific metric that is not present. |
|
{ |
|
endpoint: api.metricMetadata, |
|
query: url.Values{"metric": []string{"go_gc_duration_seconds"}}, |
|
metadata: []targetMetadata{ |
|
{ |
|
identifier: "test", |
|
metadata: []scrape.MetricMetadata{ |
|
{ |
|
Metric: "go_threads", |
|
Type: textparse.MetricTypeGauge, |
|
Help: "Number of OS threads created", |
|
Unit: "", |
|
}, |
|
}, |
|
}, |
|
}, |
|
response: map[string][]metadata{}, |
|
}, |
|
// With no available metadata. |
|
{ |
|
endpoint: api.metricMetadata, |
|
response: map[string][]metadata{}, |
|
}, |
|
{ |
|
endpoint: api.serveConfig, |
|
response: &prometheusConfig{ |
|
YAML: samplePrometheusCfg.String(), |
|
}, |
|
}, |
|
{ |
|
endpoint: api.serveFlags, |
|
response: sampleFlagMap, |
|
}, |
|
{ |
|
endpoint: api.alerts, |
|
response: &AlertDiscovery{ |
|
Alerts: []*Alert{}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.rules, |
|
response: &RuleDiscovery{ |
|
RuleGroups: []*RuleGroup{ |
|
{ |
|
Name: "grp", |
|
File: "/path/to/file", |
|
Interval: 1, |
|
Limit: 0, |
|
Rules: []Rule{ |
|
AlertingRule{ |
|
State: "inactive", |
|
Name: "test_metric3", |
|
Query: "absent(test_metric3) != 1", |
|
Duration: 1, |
|
Labels: labels.Labels{}, |
|
Annotations: labels.Labels{}, |
|
Alerts: []*Alert{}, |
|
Health: "unknown", |
|
Type: "alerting", |
|
}, |
|
AlertingRule{ |
|
State: "inactive", |
|
Name: "test_metric4", |
|
Query: "up == 1", |
|
Duration: 1, |
|
Labels: labels.Labels{}, |
|
Annotations: labels.Labels{}, |
|
Alerts: []*Alert{}, |
|
Health: "unknown", |
|
Type: "alerting", |
|
}, |
|
RecordingRule{ |
|
Name: "recording-rule-1", |
|
Query: "vector(1)", |
|
Labels: labels.Labels{}, |
|
Health: "unknown", |
|
Type: "recording", |
|
}, |
|
}, |
|
}, |
|
}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.rules, |
|
query: url.Values{ |
|
"type": []string{"alert"}, |
|
}, |
|
response: &RuleDiscovery{ |
|
RuleGroups: []*RuleGroup{ |
|
{ |
|
Name: "grp", |
|
File: "/path/to/file", |
|
Interval: 1, |
|
Limit: 0, |
|
Rules: []Rule{ |
|
AlertingRule{ |
|
State: "inactive", |
|
Name: "test_metric3", |
|
Query: "absent(test_metric3) != 1", |
|
Duration: 1, |
|
Labels: labels.Labels{}, |
|
Annotations: labels.Labels{}, |
|
Alerts: []*Alert{}, |
|
Health: "unknown", |
|
Type: "alerting", |
|
}, |
|
AlertingRule{ |
|
State: "inactive", |
|
Name: "test_metric4", |
|
Query: "up == 1", |
|
Duration: 1, |
|
Labels: labels.Labels{}, |
|
Annotations: labels.Labels{}, |
|
Alerts: []*Alert{}, |
|
Health: "unknown", |
|
Type: "alerting", |
|
}, |
|
}, |
|
}, |
|
}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.rules, |
|
query: url.Values{ |
|
"type": []string{"record"}, |
|
}, |
|
response: &RuleDiscovery{ |
|
RuleGroups: []*RuleGroup{ |
|
{ |
|
Name: "grp", |
|
File: "/path/to/file", |
|
Interval: 1, |
|
Limit: 0, |
|
Rules: []Rule{ |
|
RecordingRule{ |
|
Name: "recording-rule-1", |
|
Query: "vector(1)", |
|
Labels: labels.Labels{}, |
|
Health: "unknown", |
|
Type: "recording", |
|
}, |
|
}, |
|
}, |
|
}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.rules, |
|
query: url.Values{"rule_name[]": []string{"test_metric4"}}, |
|
response: &RuleDiscovery{ |
|
RuleGroups: []*RuleGroup{ |
|
{ |
|
Name: "grp", |
|
File: "/path/to/file", |
|
Interval: 1, |
|
Limit: 0, |
|
Rules: []Rule{ |
|
AlertingRule{ |
|
State: "inactive", |
|
Name: "test_metric4", |
|
Query: "up == 1", |
|
Duration: 1, |
|
Labels: labels.Labels{}, |
|
Annotations: labels.Labels{}, |
|
Alerts: []*Alert{}, |
|
Health: "unknown", |
|
Type: "alerting", |
|
}, |
|
}, |
|
}, |
|
}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.rules, |
|
query: url.Values{"rule_group[]": []string{"respond-with-nothing"}}, |
|
response: &RuleDiscovery{RuleGroups: []*RuleGroup{}}, |
|
}, |
|
{ |
|
endpoint: api.rules, |
|
query: url.Values{"file[]": []string{"/path/to/file"}, "rule_name[]": []string{"test_metric4"}}, |
|
response: &RuleDiscovery{ |
|
RuleGroups: []*RuleGroup{ |
|
{ |
|
Name: "grp", |
|
File: "/path/to/file", |
|
Interval: 1, |
|
Limit: 0, |
|
Rules: []Rule{ |
|
AlertingRule{ |
|
State: "inactive", |
|
Name: "test_metric4", |
|
Query: "up == 1", |
|
Duration: 1, |
|
Labels: labels.Labels{}, |
|
Annotations: labels.Labels{}, |
|
Alerts: []*Alert{}, |
|
Health: "unknown", |
|
Type: "alerting", |
|
}, |
|
}, |
|
}, |
|
}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.queryExemplars, |
|
query: url.Values{ |
|
"query": []string{`test_metric3{foo="boo"} - test_metric4{foo="bar"}`}, |
|
"start": []string{"0"}, |
|
"end": []string{"4"}, |
|
}, |
|
// Note extra integer length of timestamps for exemplars because of millisecond preservation |
|
// of timestamps within Prometheus (see timestamp package). |
|
|
|
response: []exemplar.QueryResult{ |
|
{ |
|
SeriesLabels: labels.FromStrings("__name__", "test_metric3", "foo", "boo", "dup", "1"), |
|
Exemplars: []exemplar.Exemplar{ |
|
{ |
|
Labels: labels.FromStrings("id", "abc"), |
|
Value: 10, |
|
Ts: timestamp.FromTime(start.Add(2 * time.Second)), |
|
}, |
|
}, |
|
}, |
|
{ |
|
SeriesLabels: labels.FromStrings("__name__", "test_metric4", "foo", "bar", "dup", "1"), |
|
Exemplars: []exemplar.Exemplar{ |
|
{ |
|
Labels: labels.FromStrings("id", "lul"), |
|
Value: 10, |
|
Ts: timestamp.FromTime(start.Add(4 * time.Second)), |
|
}, |
|
}, |
|
}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.queryExemplars, |
|
query: url.Values{ |
|
"query": []string{`{foo="boo"}`}, |
|
"start": []string{"4"}, |
|
"end": []string{"4.1"}, |
|
}, |
|
response: []exemplar.QueryResult{ |
|
{ |
|
SeriesLabels: labels.FromStrings("__name__", "test_metric3", "foo", "boo", "dup", "1"), |
|
Exemplars: []exemplar.Exemplar{ |
|
{ |
|
Labels: labels.FromStrings("id", "abc2"), |
|
Value: 10, |
|
Ts: 4053, |
|
}, |
|
}, |
|
}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.queryExemplars, |
|
query: url.Values{ |
|
"query": []string{`{foo="boo"}`}, |
|
}, |
|
response: []exemplar.QueryResult{ |
|
{ |
|
SeriesLabels: labels.FromStrings("__name__", "test_metric3", "foo", "boo", "dup", "1"), |
|
Exemplars: []exemplar.Exemplar{ |
|
{ |
|
Labels: labels.FromStrings("id", "abc"), |
|
Value: 10, |
|
Ts: 2000, |
|
}, |
|
{ |
|
Labels: labels.FromStrings("id", "abc2"), |
|
Value: 10, |
|
Ts: 4053, |
|
}, |
|
}, |
|
}, |
|
}, |
|
}, |
|
{ |
|
endpoint: api.queryExemplars, |
|
query: url.Values{ |
|
"query": []string{`{__name__="test_metric5"}`}, |
|
}, |
|
response: []exemplar.QueryResult{}, |
|
}, |
|
} |
|
|
|
if testLabelAPI { |
|
tests = append(tests, []test{ |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "__name__", |
|
}, |
|
response: []string{ |
|
"test_metric1", |
|
"test_metric2", |
|
"test_metric3", |
|
"test_metric4", |
|
}, |
|
}, |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
response: []string{ |
|
"bar", |
|
"boo", |
|
}, |
|
}, |
|
// Bad name parameter. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "not!!!allowed", |
|
}, |
|
errType: errorBadData, |
|
}, |
|
// Start and end before LabelValues starts. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"start": []string{"-2"}, |
|
"end": []string{"-1"}, |
|
}, |
|
response: []string{}, |
|
}, |
|
// Start and end within LabelValues. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"start": []string{"1"}, |
|
"end": []string{"100"}, |
|
}, |
|
response: []string{ |
|
"bar", |
|
"boo", |
|
}, |
|
}, |
|
// Start before LabelValues, end within LabelValues. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"start": []string{"-1"}, |
|
"end": []string{"3"}, |
|
}, |
|
response: []string{ |
|
"bar", |
|
"boo", |
|
}, |
|
}, |
|
// Start before LabelValues starts, end after LabelValues ends. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"start": []string{"1969-12-31T00:00:00Z"}, |
|
"end": []string{"1970-02-01T00:02:03Z"}, |
|
}, |
|
response: []string{ |
|
"bar", |
|
"boo", |
|
}, |
|
}, |
|
// Start with bad data, end within LabelValues. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"start": []string{"boop"}, |
|
"end": []string{"1"}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
// Start within LabelValues, end after. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"start": []string{"1"}, |
|
"end": []string{"100000000"}, |
|
}, |
|
response: []string{ |
|
"bar", |
|
"boo", |
|
}, |
|
}, |
|
// Start and end after LabelValues ends. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"start": []string{"148966367200.372"}, |
|
"end": []string{"148966367200.972"}, |
|
}, |
|
response: []string{}, |
|
}, |
|
// Only provide Start within LabelValues, don't provide an end time. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"start": []string{"2"}, |
|
}, |
|
response: []string{ |
|
"bar", |
|
"boo", |
|
}, |
|
}, |
|
// Only provide end within LabelValues, don't provide a start time. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"end": []string{"100"}, |
|
}, |
|
response: []string{ |
|
"bar", |
|
"boo", |
|
}, |
|
}, |
|
// Label values with bad matchers. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"match[]": []string{`{foo=""`, `test_metric2`}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
// Label values with empty matchers. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"match[]": []string{`{foo=""}`}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
// Label values with matcher. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric2`}, |
|
}, |
|
response: []string{ |
|
"boo", |
|
}, |
|
}, |
|
// Label values with matcher. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric1`}, |
|
}, |
|
response: []string{ |
|
"bar", |
|
"boo", |
|
}, |
|
}, |
|
// Label values with matcher using label filter. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric1{foo="bar"}`}, |
|
}, |
|
response: []string{ |
|
"bar", |
|
}, |
|
}, |
|
// Label values with matcher and time range. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric1`}, |
|
"start": []string{"1"}, |
|
"end": []string{"100000000"}, |
|
}, |
|
response: []string{ |
|
"bar", |
|
"boo", |
|
}, |
|
}, |
|
// Try to overlap the selected series set as much as possible to test that the value de-duplication works. |
|
{ |
|
endpoint: api.labelValues, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric4{dup=~"^1"}`, `test_metric4{foo=~".+o$"}`}, |
|
}, |
|
response: []string{ |
|
"bar", |
|
"boo", |
|
}, |
|
}, |
|
// Label names. |
|
{ |
|
endpoint: api.labelNames, |
|
response: []string{"__name__", "dup", "foo"}, |
|
}, |
|
// Start and end before Label names starts. |
|
{ |
|
endpoint: api.labelNames, |
|
query: url.Values{ |
|
"start": []string{"-2"}, |
|
"end": []string{"-1"}, |
|
}, |
|
response: []string{}, |
|
}, |
|
// Start and end within Label names. |
|
{ |
|
endpoint: api.labelNames, |
|
query: url.Values{ |
|
"start": []string{"1"}, |
|
"end": []string{"100"}, |
|
}, |
|
response: []string{"__name__", "dup", "foo"}, |
|
}, |
|
// Start before Label names, end within Label names. |
|
{ |
|
endpoint: api.labelNames, |
|
query: url.Values{ |
|
"start": []string{"-1"}, |
|
"end": []string{"10"}, |
|
}, |
|
response: []string{"__name__", "dup", "foo"}, |
|
}, |
|
|
|
// Start before Label names starts, end after Label names ends. |
|
{ |
|
endpoint: api.labelNames, |
|
query: url.Values{ |
|
"start": []string{"-1"}, |
|
"end": []string{"100000"}, |
|
}, |
|
response: []string{"__name__", "dup", "foo"}, |
|
}, |
|
// Start with bad data for Label names, end within Label names. |
|
{ |
|
endpoint: api.labelNames, |
|
query: url.Values{ |
|
"start": []string{"boop"}, |
|
"end": []string{"1"}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
// Start within Label names, end after. |
|
{ |
|
endpoint: api.labelNames, |
|
query: url.Values{ |
|
"start": []string{"1"}, |
|
"end": []string{"1000000006"}, |
|
}, |
|
response: []string{"__name__", "dup", "foo"}, |
|
}, |
|
// Start and end after Label names ends. |
|
{ |
|
endpoint: api.labelNames, |
|
query: url.Values{ |
|
"start": []string{"148966367200.372"}, |
|
"end": []string{"148966367200.972"}, |
|
}, |
|
response: []string{}, |
|
}, |
|
// Only provide Start within Label names, don't provide an end time. |
|
{ |
|
endpoint: api.labelNames, |
|
query: url.Values{ |
|
"start": []string{"4"}, |
|
}, |
|
response: []string{"__name__", "dup", "foo"}, |
|
}, |
|
// Only provide End within Label names, don't provide a start time. |
|
{ |
|
endpoint: api.labelNames, |
|
query: url.Values{ |
|
"end": []string{"20"}, |
|
}, |
|
response: []string{"__name__", "dup", "foo"}, |
|
}, |
|
// Label names with bad matchers. |
|
{ |
|
endpoint: api.labelNames, |
|
query: url.Values{ |
|
"match[]": []string{`{foo=""`, `test_metric2`}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
// Label values with empty matchers. |
|
{ |
|
endpoint: api.labelNames, |
|
params: map[string]string{ |
|
"name": "foo", |
|
}, |
|
query: url.Values{ |
|
"match[]": []string{`{foo=""}`}, |
|
}, |
|
errType: errorBadData, |
|
}, |
|
// Label names with matcher. |
|
{ |
|
endpoint: api.labelNames, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric2`}, |
|
}, |
|
response: []string{"__name__", "foo"}, |
|
}, |
|
// Label names with matcher. |
|
{ |
|
endpoint: api.labelNames, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric3`}, |
|
}, |
|
response: []string{"__name__", "dup", "foo"}, |
|
}, |
|
// Label names with matcher using label filter. |
|
// There is no matching series. |
|
{ |
|
endpoint: api.labelNames, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric1{foo="test"}`}, |
|
}, |
|
response: []string{}, |
|
}, |
|
// Label names with matcher and time range. |
|
{ |
|
endpoint: api.labelNames, |
|
query: url.Values{ |
|
"match[]": []string{`test_metric2`}, |
|
"start": []string{"1"}, |
|
"end": []string{"100000000"}, |
|
}, |
|
response: []string{"__name__", "foo"}, |
|
}, |
|
}...) |
|
} |
|
|
|
methods := func(f apiFunc) []string { |
|
fp := reflect.ValueOf(f).Pointer() |
|
if fp == reflect.ValueOf(api.query).Pointer() || fp == reflect.ValueOf(api.queryRange).Pointer() || fp == reflect.ValueOf(api.series).Pointer() { |
|
return []string{http.MethodGet, http.MethodPost} |
|
} |
|
return []string{http.MethodGet} |
|
} |
|
|
|
request := func(m string, q url.Values) (*http.Request, error) { |
|
if m == http.MethodPost { |
|
r, err := http.NewRequest(m, "http://example.com", strings.NewReader(q.Encode())) |
|
r.Header.Set("Content-Type", "application/x-www-form-urlencoded") |
|
r.RemoteAddr = "127.0.0.1:20201" |
|
return r, err |
|
} |
|
r, err := http.NewRequest(m, fmt.Sprintf("http://example.com?%s", q.Encode()), nil) |
|
r.RemoteAddr = "127.0.0.1:20201" |
|
return r, err |
|
} |
|
|
|
for i, test := range tests { |
|
t.Run(fmt.Sprintf("run %d %s %q", i, describeAPIFunc(test.endpoint), test.query.Encode()), func(t *testing.T) { |
|
for _, method := range methods(test.endpoint) { |
|
t.Run(method, func(t *testing.T) { |
|
// Build a context with the correct request params. |
|
ctx := context.Background() |
|
for p, v := range test.params { |
|
ctx = route.WithParam(ctx, p, v) |
|
} |
|
|
|
req, err := request(method, test.query) |
|
if err != nil { |
|
t.Fatal(err) |
|
} |
|
|
|
tr.ResetMetadataStore() |
|
for _, tm := range test.metadata { |
|
tr.SetMetadataStoreForTargets(tm.identifier, &testMetaStore{Metadata: tm.metadata}) |
|
} |
|
|
|
for _, te := range test.exemplars { |
|
for _, e := range te.Exemplars { |
|
_, err := es.AppendExemplar(0, te.SeriesLabels, e) |
|
if err != nil { |
|
t.Fatal(err) |
|
} |
|
} |
|
} |
|
|
|
res := test.endpoint(req.WithContext(ctx)) |
|
assertAPIError(t, res.err, test.errType) |
|
|
|
if test.sorter != nil { |
|
test.sorter(res.data) |
|
} |
|
|
|
if test.responseLen != 0 { |
|
assertAPIResponseLength(t, res.data, test.responseLen) |
|
if test.responseMetadataTotal != 0 { |
|
assertAPIResponseMetadataLen(t, res.data, test.responseMetadataTotal) |
|
} |
|
} else { |
|
assertAPIResponse(t, res.data, test.response) |
|
} |
|
}) |
|
} |
|
}) |
|
} |
|
} |
|
|
|
func describeAPIFunc(f apiFunc) string { |
|
name := runtime.FuncForPC(reflect.ValueOf(f).Pointer()).Name() |
|
return strings.Split(name[strings.LastIndex(name, ".")+1:], "-")[0] |
|
} |
|
|
|
func assertAPIError(t *testing.T, got *apiError, exp errorType) { |
|
t.Helper() |
|
|
|
if got != nil { |
|
if exp == errorNone { |
|
t.Fatalf("Unexpected error: %s", got) |
|
} |
|
if exp != got.typ { |
|
t.Fatalf("Expected error of type %q but got type %q (%q)", exp, got.typ, got) |
|
} |
|
return |
|
} |
|
if exp != errorNone { |
|
t.Fatalf("Expected error of type %q but got none", exp) |
|
} |
|
} |
|
|
|
func assertAPIResponse(t *testing.T, got, exp interface{}) { |
|
t.Helper() |
|
|
|
require.Equal(t, exp, got) |
|
} |
|
|
|
func assertAPIResponseLength(t *testing.T, got interface{}, expLen int) { |
|
t.Helper() |
|
|
|
gotLen := reflect.ValueOf(got).Len() |
|
if gotLen != expLen { |
|
t.Fatalf( |
|
"Response length does not match, expected:\n%d\ngot:\n%d", |
|
expLen, |
|
gotLen, |
|
) |
|
} |
|
} |
|
|
|
func assertAPIResponseMetadataLen(t *testing.T, got interface{}, expLen int) { |
|
t.Helper() |
|
|
|
var gotLen int |
|
response := got.(map[string][]metadata) |
|
for _, m := range response { |
|
gotLen += len(m) |
|
} |
|
|
|
if gotLen != expLen { |
|
t.Fatalf( |
|
"Amount of metadata in the response does not match, expected:\n%d\ngot:\n%d", |
|
expLen, |
|
gotLen, |
|
) |
|
} |
|
} |
|
|
|
type fakeDB struct { |
|
err error |
|
} |
|
|
|
func (f *fakeDB) CleanTombstones() error { return f.err } |
|
func (f *fakeDB) Delete(context.Context, int64, int64, ...*labels.Matcher) error { return f.err } |
|
func (f *fakeDB) Snapshot(string, bool) error { return f.err } |
|
func (f *fakeDB) Stats(statsByLabelName string, limit int) (_ *tsdb.Stats, retErr error) { |
|
dbDir, err := os.MkdirTemp("", "tsdb-api-ready") |
|
if err != nil { |
|
return nil, err |
|
} |
|
defer func() { |
|
err := os.RemoveAll(dbDir) |
|
if retErr != nil { |
|
retErr = err |
|
} |
|
}() |
|
opts := tsdb.DefaultHeadOptions() |
|
opts.ChunkRange = 1000 |
|
h, _ := tsdb.NewHead(nil, nil, nil, nil, opts, nil) |
|
return h.Stats(statsByLabelName, limit), nil |
|
} |
|
|
|
func (f *fakeDB) WALReplayStatus() (tsdb.WALReplayStatus, error) { |
|
return tsdb.WALReplayStatus{}, nil |
|
} |
|
|
|
func TestAdminEndpoints(t *testing.T) { |
|
tsdb, tsdbWithError, tsdbNotReady := &fakeDB{}, &fakeDB{err: errors.New("some error")}, &fakeDB{err: errors.Wrap(tsdb.ErrNotReady, "wrap")} |
|
snapshotAPI := func(api *API) apiFunc { return api.snapshot } |
|
cleanAPI := func(api *API) apiFunc { return api.cleanTombstones } |
|
deleteAPI := func(api *API) apiFunc { return api.deleteSeries } |
|
|
|
for _, tc := range []struct { |
|
db *fakeDB |
|
enableAdmin bool |
|
endpoint func(api *API) apiFunc |
|
method string |
|
values url.Values |
|
|
|
errType errorType |
|
}{ |
|
// Tests for the snapshot endpoint. |
|
{ |
|
db: tsdb, |
|
enableAdmin: false, |
|
endpoint: snapshotAPI, |
|
|
|
errType: errorUnavailable, |
|
}, |
|
{ |
|
db: tsdb, |
|
enableAdmin: true, |
|
endpoint: snapshotAPI, |
|
|
|
errType: errorNone, |
|
}, |
|
{ |
|
db: tsdb, |
|
enableAdmin: true, |
|
endpoint: snapshotAPI, |
|
values: map[string][]string{"skip_head": {"true"}}, |
|
|
|
errType: errorNone, |
|
}, |
|
{ |
|
db: tsdb, |
|
enableAdmin: true, |
|
endpoint: snapshotAPI, |
|
values: map[string][]string{"skip_head": {"xxx"}}, |
|
|
|
errType: errorBadData, |
|
}, |
|
{ |
|
db: tsdbWithError, |
|
enableAdmin: true, |
|
endpoint: snapshotAPI, |
|
|
|
errType: errorInternal, |
|
}, |
|
{ |
|
db: tsdbNotReady, |
|
enableAdmin: true, |
|
endpoint: snapshotAPI, |
|
|
|
errType: errorUnavailable, |
|
}, |
|
// Tests for the cleanTombstones endpoint. |
|
{ |
|
db: tsdb, |
|
enableAdmin: false, |
|
endpoint: cleanAPI, |
|
|
|
errType: errorUnavailable, |
|
}, |
|
{ |
|
db: tsdb, |
|
enableAdmin: true, |
|
endpoint: cleanAPI, |
|
|
|
errType: errorNone, |
|
}, |
|
{ |
|
db: tsdbWithError, |
|
enableAdmin: true, |
|
endpoint: cleanAPI, |
|
|
|
errType: errorInternal, |
|
}, |
|
{ |
|
db: tsdbNotReady, |
|
enableAdmin: true, |
|
endpoint: cleanAPI, |
|
|
|
errType: errorUnavailable, |
|
}, |
|
// Tests for the deleteSeries endpoint. |
|
{ |
|
db: tsdb, |
|
enableAdmin: false, |
|
endpoint: deleteAPI, |
|
|
|
errType: errorUnavailable, |
|
}, |
|
{ |
|
db: tsdb, |
|
enableAdmin: true, |
|
endpoint: deleteAPI, |
|
|
|
errType: errorBadData, |
|
}, |
|
{ |
|
db: tsdb, |
|
enableAdmin: true, |
|
endpoint: deleteAPI, |
|
values: map[string][]string{"match[]": {"123"}}, |
|
|
|
errType: errorBadData, |
|
}, |
|
{ |
|
db: tsdb, |
|
enableAdmin: true, |
|
endpoint: deleteAPI, |
|
values: map[string][]string{"match[]": {"up"}, "start": {"xxx"}}, |
|
|
|
errType: errorBadData, |
|
}, |
|
{ |
|
db: tsdb, |
|
enableAdmin: true, |
|
endpoint: deleteAPI, |
|
values: map[string][]string{"match[]": {"up"}, "end": {"xxx"}}, |
|
|
|
errType: errorBadData, |
|
}, |
|
{ |
|
db: tsdb, |
|
enableAdmin: true, |
|
endpoint: deleteAPI, |
|
values: map[string][]string{"match[]": {"up"}}, |
|
|
|
errType: errorNone, |
|
}, |
|
{ |
|
db: tsdb, |
|
enableAdmin: true, |
|
endpoint: deleteAPI, |
|
values: map[string][]string{"match[]": {"up{job!=\"foo\"}", "{job=~\"bar.+\"}", "up{instance!~\"fred.+\"}"}}, |
|
|
|
errType: errorNone, |
|
}, |
|
{ |
|
db: tsdbWithError, |
|
enableAdmin: true, |
|
endpoint: deleteAPI, |
|
values: map[string][]string{"match[]": {"up"}}, |
|
|
|
errType: errorInternal, |
|
}, |
|
{ |
|
db: tsdbNotReady, |
|
enableAdmin: true, |
|
endpoint: deleteAPI, |
|
values: map[string][]string{"match[]": {"up"}}, |
|
|
|
errType: errorUnavailable, |
|
}, |
|
} { |
|
tc := tc |
|
t.Run("", func(t *testing.T) { |
|
dir := t.TempDir() |
|
|
|
api := &API{ |
|
db: tc.db, |
|
dbDir: dir, |
|
ready: func(f http.HandlerFunc) http.HandlerFunc { return f }, |
|
enableAdmin: tc.enableAdmin, |
|
} |
|
|
|
endpoint := tc.endpoint(api) |
|
req, err := http.NewRequest(tc.method, fmt.Sprintf("?%s", tc.values.Encode()), nil) |
|
require.NoError(t, err) |
|
|
|
res := setUnavailStatusOnTSDBNotReady(endpoint(req)) |
|
assertAPIError(t, res.err, tc.errType) |
|
}) |
|
} |
|
} |
|
|
|
func TestRespondSuccess(t *testing.T) { |
|
api := API{ |
|
logger: log.NewNopLogger(), |
|
} |
|
|
|
api.ClearCodecs() |
|
api.InstallCodec(JSONCodec{}) |
|
api.InstallCodec(&testCodec{contentType: MIMEType{"test", "cannot-encode"}, canEncode: false}) |
|
api.InstallCodec(&testCodec{contentType: MIMEType{"test", "can-encode"}, canEncode: true}) |
|
api.InstallCodec(&testCodec{contentType: MIMEType{"test", "can-encode-2"}, canEncode: true}) |
|
|
|
s := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { |
|
api.respond(w, r, "test", nil, "") |
|
})) |
|
defer s.Close() |
|
|
|
for _, tc := range []struct { |
|
name string |
|
acceptHeader string |
|
expectedContentType string |
|
expectedBody string |
|
}{ |
|
{ |
|
name: "no Accept header", |
|
expectedContentType: "application/json", |
|
expectedBody: `{"status":"success","data":"test"}`, |
|
}, |
|
{ |
|
name: "Accept header with single content type which is suitable", |
|
acceptHeader: "test/can-encode", |
|
expectedContentType: "test/can-encode", |
|
expectedBody: `response from test/can-encode codec`, |
|
}, |
|
{ |
|
name: "Accept header with single content type which is not available", |
|
acceptHeader: "test/not-registered", |
|
expectedContentType: "application/json", |
|
expectedBody: `{"status":"success","data":"test"}`, |
|
}, |
|
{ |
|
name: "Accept header with single content type which cannot encode the response payload", |
|
acceptHeader: "test/cannot-encode", |
|
expectedContentType: "application/json", |
|
expectedBody: `{"status":"success","data":"test"}`, |
|
}, |
|
{ |
|
name: "Accept header with multiple content types, all of which are suitable", |
|
acceptHeader: "test/can-encode, test/can-encode-2", |
|
expectedContentType: "test/can-encode", |
|
expectedBody: `response from test/can-encode codec`, |
|
}, |
|
{ |
|
name: "Accept header with multiple content types, only one of which is available", |
|
acceptHeader: "test/not-registered, test/can-encode", |
|
expectedContentType: "test/can-encode", |
|
expectedBody: `response from test/can-encode codec`, |
|
}, |
|
{ |
|
name: "Accept header with multiple content types, only one of which can encode the response payload", |
|
acceptHeader: "test/cannot-encode, test/can-encode", |
|
expectedContentType: "test/can-encode", |
|
expectedBody: `response from test/can-encode codec`, |
|
}, |
|
{ |
|
name: "Accept header with multiple content types, none of which are available", |
|
acceptHeader: "test/not-registered, test/also-not-registered", |
|
expectedContentType: "application/json", |
|
expectedBody: `{"status":"success","data":"test"}`, |
|
}, |
|
} { |
|
t.Run(tc.name, func(t *testing.T) { |
|
req, err := http.NewRequest(http.MethodGet, s.URL, nil) |
|
require.NoError(t, err) |
|
|
|
if tc.acceptHeader != "" { |
|
req.Header.Set("Accept", tc.acceptHeader) |
|
} |
|
|
|
resp, err := http.DefaultClient.Do(req) |
|
require.NoError(t, err) |
|
|
|
body, err := io.ReadAll(resp.Body) |
|
defer resp.Body.Close() |
|
require.NoError(t, err) |
|
|
|
require.Equal(t, http.StatusOK, resp.StatusCode) |
|
require.Equal(t, tc.expectedContentType, resp.Header.Get("Content-Type")) |
|
require.Equal(t, tc.expectedBody, string(body)) |
|
}) |
|
} |
|
} |
|
|
|
func TestRespondSuccess_DefaultCodecCannotEncodeResponse(t *testing.T) { |
|
api := API{ |
|
logger: log.NewNopLogger(), |
|
} |
|
|
|
api.ClearCodecs() |
|
api.InstallCodec(&testCodec{contentType: MIMEType{"application", "default-format"}, canEncode: false}) |
|
|
|
s := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { |
|
api.respond(w, r, "test", nil, "") |
|
})) |
|
defer s.Close() |
|
|
|
req, err := http.NewRequest(http.MethodGet, s.URL, nil) |
|
require.NoError(t, err) |
|
|
|
resp, err := http.DefaultClient.Do(req) |
|
require.NoError(t, err) |
|
|
|
body, err := io.ReadAll(resp.Body) |
|
defer resp.Body.Close() |
|
require.NoError(t, err) |
|
|
|
require.Equal(t, http.StatusNotAcceptable, resp.StatusCode) |
|
require.Equal(t, "application/json", resp.Header.Get("Content-Type")) |
|
require.Equal(t, `{"status":"error","errorType":"not_acceptable","error":"cannot encode response as application/default-format"}`, string(body)) |
|
} |
|
|
|
func TestRespondError(t *testing.T) { |
|
s := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { |
|
api := API{} |
|
api.respondError(w, &apiError{errorTimeout, errors.New("message")}, "test") |
|
})) |
|
defer s.Close() |
|
|
|
resp, err := http.Get(s.URL) |
|
if err != nil { |
|
t.Fatalf("Error on test request: %s", err) |
|
} |
|
body, err := io.ReadAll(resp.Body) |
|
defer resp.Body.Close() |
|
if err != nil { |
|
t.Fatalf("Error reading response body: %s", err) |
|
} |
|
|
|
if want, have := http.StatusServiceUnavailable, resp.StatusCode; want != have { |
|
t.Fatalf("Return code %d expected in error response but got %d", want, have) |
|
} |
|
if h := resp.Header.Get("Content-Type"); h != "application/json" { |
|
t.Fatalf("Expected Content-Type %q but got %q", "application/json", h) |
|
} |
|
|
|
var res Response |
|
if err = json.Unmarshal(body, &res); err != nil { |
|
t.Fatalf("Error unmarshaling JSON body: %s", err) |
|
} |
|
|
|
exp := &Response{ |
|
Status: statusError, |
|
Data: "test", |
|
ErrorType: errorTimeout, |
|
Error: "message", |
|
} |
|
require.Equal(t, exp, &res) |
|
} |
|
|
|
func TestParseTimeParam(t *testing.T) { |
|
type resultType struct { |
|
asTime time.Time |
|
asError func() error |
|
} |
|
|
|
ts, err := parseTime("1582468023986") |
|
require.NoError(t, err) |
|
|
|
tests := []struct { |
|
paramName string |
|
paramValue string |
|
defaultValue time.Time |
|
result resultType |
|
}{ |
|
{ // When data is valid. |
|
paramName: "start", |
|
paramValue: "1582468023986", |
|
defaultValue: MinTime, |
|
result: resultType{ |
|
asTime: ts, |
|
asError: nil, |
|
}, |
|
}, |
|
{ // When data is empty string. |
|
paramName: "end", |
|
paramValue: "", |
|
defaultValue: MaxTime, |
|
result: resultType{ |
|
asTime: MaxTime, |
|
asError: nil, |
|
}, |
|
}, |
|
{ // When data is not valid. |
|
paramName: "foo", |
|
paramValue: "baz", |
|
defaultValue: MaxTime, |
|
result: resultType{ |
|
asTime: time.Time{}, |
|
asError: func() error { |
|
_, err := parseTime("baz") |
|
return errors.Wrapf(err, "Invalid time value for '%s'", "foo") |
|
}, |
|
}, |
|
}, |
|
} |
|
|
|
for _, test := range tests { |
|
req, err := http.NewRequest("GET", "localhost:42/foo?"+test.paramName+"="+test.paramValue, nil) |
|
require.NoError(t, err) |
|
|
|
result := test.result |
|
asTime, err := parseTimeParam(req, test.paramName, test.defaultValue) |
|
|
|
if err != nil { |
|
require.EqualError(t, err, result.asError().Error()) |
|
} else { |
|
require.True(t, asTime.Equal(result.asTime), "time as return value: %s not parsed correctly. Expected %s. Actual %s", test.paramValue, result.asTime, asTime) |
|
} |
|
} |
|
} |
|
|
|
func TestParseTime(t *testing.T) { |
|
ts, err := time.Parse(time.RFC3339Nano, "2015-06-03T13:21:58.555Z") |
|
if err != nil { |
|
panic(err) |
|
} |
|
|
|
tests := []struct { |
|
input string |
|
fail bool |
|
result time.Time |
|
}{ |
|
{ |
|
input: "", |
|
fail: true, |
|
}, |
|
{ |
|
input: "abc", |
|
fail: true, |
|
}, |
|
{ |
|
input: "30s", |
|
fail: true, |
|
}, |
|
{ |
|
input: "123", |
|
result: time.Unix(123, 0), |
|
}, |
|
{ |
|
input: "123.123", |
|
result: time.Unix(123, 123000000), |
|
}, |
|
{ |
|
input: "2015-06-03T13:21:58.555Z", |
|
result: ts, |
|
}, |
|
{ |
|
input: "2015-06-03T14:21:58.555+01:00", |
|
result: ts, |
|
}, |
|
{ |
|
// Test float rounding. |
|
input: "1543578564.705", |
|
result: time.Unix(1543578564, 705*1e6), |
|
}, |
|
{ |
|
input: MinTime.Format(time.RFC3339Nano), |
|
result: MinTime, |
|
}, |
|
{ |
|
input: MaxTime.Format(time.RFC3339Nano), |
|
result: MaxTime, |
|
}, |
|
} |
|
|
|
for _, test := range tests { |
|
ts, err := parseTime(test.input) |
|
if err != nil && !test.fail { |
|
t.Errorf("Unexpected error for %q: %s", test.input, err) |
|
continue |
|
} |
|
if err == nil && test.fail { |
|
t.Errorf("Expected error for %q but got none", test.input) |
|
continue |
|
} |
|
if !test.fail && !ts.Equal(test.result) { |
|
t.Errorf("Expected time %v for input %q but got %v", test.result, test.input, ts) |
|
} |
|
} |
|
} |
|
|
|
func TestParseDuration(t *testing.T) { |
|
tests := []struct { |
|
input string |
|
fail bool |
|
result time.Duration |
|
}{ |
|
{ |
|
input: "", |
|
fail: true, |
|
}, { |
|
input: "abc", |
|
fail: true, |
|
}, { |
|
input: "2015-06-03T13:21:58.555Z", |
|
fail: true, |
|
}, { |
|
// Internal int64 overflow. |
|
input: "-148966367200.372", |
|
fail: true, |
|
}, { |
|
// Internal int64 overflow. |
|
input: "148966367200.372", |
|
fail: true, |
|
}, { |
|
input: "123", |
|
result: 123 * time.Second, |
|
}, { |
|
input: "123.333", |
|
result: 123*time.Second + 333*time.Millisecond, |
|
}, { |
|
input: "15s", |
|
result: 15 * time.Second, |
|
}, { |
|
input: "5m", |
|
result: 5 * time.Minute, |
|
}, |
|
} |
|
|
|
for _, test := range tests { |
|
d, err := parseDuration(test.input) |
|
if err != nil && !test.fail { |
|
t.Errorf("Unexpected error for %q: %s", test.input, err) |
|
continue |
|
} |
|
if err == nil && test.fail { |
|
t.Errorf("Expected error for %q but got none", test.input) |
|
continue |
|
} |
|
if !test.fail && d != test.result { |
|
t.Errorf("Expected duration %v for input %q but got %v", test.result, test.input, d) |
|
} |
|
} |
|
} |
|
|
|
func TestOptionsMethod(t *testing.T) { |
|
r := route.New() |
|
api := &API{ready: func(f http.HandlerFunc) http.HandlerFunc { return f }} |
|
api.Register(r) |
|
|
|
s := httptest.NewServer(r) |
|
defer s.Close() |
|
|
|
req, err := http.NewRequest("OPTIONS", s.URL+"/any_path", nil) |
|
if err != nil { |
|
t.Fatalf("Error creating OPTIONS request: %s", err) |
|
} |
|
client := &http.Client{} |
|
resp, err := client.Do(req) |
|
if err != nil { |
|
t.Fatalf("Error executing OPTIONS request: %s", err) |
|
} |
|
|
|
if resp.StatusCode != http.StatusNoContent { |
|
t.Fatalf("Expected status %d, got %d", http.StatusNoContent, resp.StatusCode) |
|
} |
|
} |
|
|
|
func TestTSDBStatus(t *testing.T) { |
|
tsdb := &fakeDB{} |
|
tsdbStatusAPI := func(api *API) apiFunc { return api.serveTSDBStatus } |
|
|
|
for i, tc := range []struct { |
|
db *fakeDB |
|
endpoint func(api *API) apiFunc |
|
method string |
|
values url.Values |
|
|
|
errType errorType |
|
}{ |
|
// Tests for the TSDB Status endpoint. |
|
{ |
|
db: tsdb, |
|
endpoint: tsdbStatusAPI, |
|
errType: errorNone, |
|
}, |
|
{ |
|
db: tsdb, |
|
endpoint: tsdbStatusAPI, |
|
values: map[string][]string{"limit": {"20"}}, |
|
errType: errorNone, |
|
}, |
|
{ |
|
db: tsdb, |
|
endpoint: tsdbStatusAPI, |
|
values: map[string][]string{"limit": {"0"}}, |
|
errType: errorBadData, |
|
}, |
|
} { |
|
tc := tc |
|
t.Run(fmt.Sprintf("%d", i), func(t *testing.T) { |
|
api := &API{db: tc.db, gatherer: prometheus.DefaultGatherer} |
|
endpoint := tc.endpoint(api) |
|
req, err := http.NewRequest(tc.method, fmt.Sprintf("?%s", tc.values.Encode()), nil) |
|
if err != nil { |
|
t.Fatalf("Error when creating test request: %s", err) |
|
} |
|
res := endpoint(req) |
|
assertAPIError(t, res.err, tc.errType) |
|
}) |
|
} |
|
} |
|
|
|
func TestReturnAPIError(t *testing.T) { |
|
cases := []struct { |
|
err error |
|
expected errorType |
|
}{ |
|
{ |
|
err: promql.ErrStorage{Err: errors.New("storage error")}, |
|
expected: errorInternal, |
|
}, { |
|
err: fmt.Errorf("wrapped: %w", promql.ErrStorage{Err: errors.New("storage error")}), |
|
expected: errorInternal, |
|
}, { |
|
err: promql.ErrQueryTimeout("timeout error"), |
|
expected: errorTimeout, |
|
}, { |
|
err: fmt.Errorf("wrapped: %w", promql.ErrQueryTimeout("timeout error")), |
|
expected: errorTimeout, |
|
}, { |
|
err: promql.ErrQueryCanceled("canceled error"), |
|
expected: errorCanceled, |
|
}, { |
|
err: fmt.Errorf("wrapped: %w", promql.ErrQueryCanceled("canceled error")), |
|
expected: errorCanceled, |
|
}, { |
|
err: errors.New("exec error"), |
|
expected: errorExec, |
|
}, |
|
} |
|
|
|
for ix, c := range cases { |
|
actual := returnAPIError(c.err) |
|
require.Error(t, actual, ix) |
|
require.Equal(t, c.expected, actual.typ, ix) |
|
} |
|
} |
|
|
|
// This is a global to avoid the benchmark being optimized away. |
|
var testResponseWriter = httptest.ResponseRecorder{} |
|
|
|
func BenchmarkRespond(b *testing.B) { |
|
points := []promql.FPoint{} |
|
for i := 0; i < 10000; i++ { |
|
points = append(points, promql.FPoint{F: float64(i * 1000000), T: int64(i)}) |
|
} |
|
matrix := promql.Matrix{} |
|
for i := 0; i < 1000; i++ { |
|
matrix = append(matrix, promql.Series{ |
|
Metric: labels.FromStrings("__name__", fmt.Sprintf("series%v", i), |
|
"label", fmt.Sprintf("series%v", i), |
|
"label2", fmt.Sprintf("series%v", i)), |
|
Floats: points[:10], |
|
}) |
|
} |
|
series := []labels.Labels{} |
|
for i := 0; i < 1000; i++ { |
|
series = append(series, labels.FromStrings("__name__", fmt.Sprintf("series%v", i), |
|
"label", fmt.Sprintf("series%v", i), |
|
"label2", fmt.Sprintf("series%v", i))) |
|
} |
|
|
|
cases := []struct { |
|
name string |
|
response interface{} |
|
}{ |
|
{name: "10000 points no labels", response: &QueryData{ |
|
ResultType: parser.ValueTypeMatrix, |
|
Result: promql.Matrix{ |
|
promql.Series{ |
|
Floats: points, |
|
Metric: labels.EmptyLabels(), |
|
}, |
|
}, |
|
}}, |
|
{name: "1000 labels", response: series}, |
|
{name: "1000 series 10 points", response: &QueryData{ |
|
ResultType: parser.ValueTypeMatrix, |
|
Result: matrix, |
|
}}, |
|
} |
|
for _, c := range cases { |
|
b.Run(c.name, func(b *testing.B) { |
|
b.ReportAllocs() |
|
request, err := http.NewRequest(http.MethodGet, "/does-not-matter", nil) |
|
require.NoError(b, err) |
|
b.ResetTimer() |
|
api := API{} |
|
api.InstallCodec(JSONCodec{}) |
|
for n := 0; n < b.N; n++ { |
|
api.respond(&testResponseWriter, request, c.response, nil, "") |
|
} |
|
}) |
|
} |
|
} |
|
|
|
func TestGetGlobalURL(t *testing.T) { |
|
mustParseURL := func(t *testing.T, u string) *url.URL { |
|
parsed, err := url.Parse(u) |
|
require.NoError(t, err) |
|
return parsed |
|
} |
|
|
|
testcases := []struct { |
|
input *url.URL |
|
opts GlobalURLOptions |
|
expected *url.URL |
|
errorful bool |
|
}{ |
|
{ |
|
mustParseURL(t, "http://127.0.0.1:9090"), |
|
GlobalURLOptions{ |
|
ListenAddress: "127.0.0.1:9090", |
|
Host: "127.0.0.1:9090", |
|
Scheme: "http", |
|
}, |
|
mustParseURL(t, "http://127.0.0.1:9090"), |
|
false, |
|
}, |
|
{ |
|
mustParseURL(t, "http://127.0.0.1:9090"), |
|
GlobalURLOptions{ |
|
ListenAddress: "127.0.0.1:9090", |
|
Host: "prometheus.io", |
|
Scheme: "https", |
|
}, |
|
mustParseURL(t, "https://prometheus.io"), |
|
false, |
|
}, |
|
{ |
|
mustParseURL(t, "http://exemple.com"), |
|
GlobalURLOptions{ |
|
ListenAddress: "127.0.0.1:9090", |
|
Host: "prometheus.io", |
|
Scheme: "https", |
|
}, |
|
mustParseURL(t, "http://exemple.com"), |
|
false, |
|
}, |
|
{ |
|
mustParseURL(t, "http://localhost:8080"), |
|
GlobalURLOptions{ |
|
ListenAddress: "127.0.0.1:9090", |
|
Host: "prometheus.io", |
|
Scheme: "https", |
|
}, |
|
mustParseURL(t, "http://prometheus.io:8080"), |
|
false, |
|
}, |
|
{ |
|
mustParseURL(t, "http://[::1]:8080"), |
|
GlobalURLOptions{ |
|
ListenAddress: "127.0.0.1:9090", |
|
Host: "prometheus.io", |
|
Scheme: "https", |
|
}, |
|
mustParseURL(t, "http://prometheus.io:8080"), |
|
false, |
|
}, |
|
{ |
|
mustParseURL(t, "http://localhost"), |
|
GlobalURLOptions{ |
|
ListenAddress: "127.0.0.1:9090", |
|
Host: "prometheus.io", |
|
Scheme: "https", |
|
}, |
|
mustParseURL(t, "http://prometheus.io"), |
|
false, |
|
}, |
|
{ |
|
mustParseURL(t, "http://localhost:9091"), |
|
GlobalURLOptions{ |
|
ListenAddress: "[::1]:9090", |
|
Host: "[::1]", |
|
Scheme: "https", |
|
}, |
|
mustParseURL(t, "http://[::1]:9091"), |
|
false, |
|
}, |
|
{ |
|
mustParseURL(t, "http://localhost:9091"), |
|
GlobalURLOptions{ |
|
ListenAddress: "[::1]:9090", |
|
Host: "[::1]:9090", |
|
Scheme: "https", |
|
}, |
|
mustParseURL(t, "http://[::1]:9091"), |
|
false, |
|
}, |
|
} |
|
|
|
for i, tc := range testcases { |
|
t.Run(fmt.Sprintf("Test %d", i), func(t *testing.T) { |
|
output, err := getGlobalURL(tc.input, tc.opts) |
|
if tc.errorful { |
|
require.Error(t, err) |
|
return |
|
} |
|
require.NoError(t, err) |
|
require.Equal(t, tc.expected, output) |
|
}) |
|
} |
|
} |
|
|
|
type testCodec struct { |
|
contentType MIMEType |
|
canEncode bool |
|
} |
|
|
|
func (t *testCodec) ContentType() MIMEType { |
|
return t.contentType |
|
} |
|
|
|
func (t *testCodec) CanEncode(_ *Response) bool { |
|
return t.canEncode |
|
} |
|
|
|
func (t *testCodec) Encode(_ *Response) ([]byte, error) { |
|
return []byte(fmt.Sprintf("response from %v codec", t.contentType)), nil |
|
} |
|
|
|
func TestExtractQueryOpts(t *testing.T) { |
|
tests := []struct { |
|
name string |
|
form url.Values |
|
expect promql.QueryOpts |
|
err error |
|
}{ |
|
{ |
|
name: "with stats all", |
|
form: url.Values{ |
|
"stats": []string{"all"}, |
|
}, |
|
expect: promql.NewPrometheusQueryOpts(true, 0), |
|
|
|
err: nil, |
|
}, |
|
{ |
|
name: "with stats none", |
|
form: url.Values{ |
|
"stats": []string{"none"}, |
|
}, |
|
expect: promql.NewPrometheusQueryOpts(false, 0), |
|
err: nil, |
|
}, |
|
{ |
|
name: "with lookback delta", |
|
form: url.Values{ |
|
"stats": []string{"all"}, |
|
"lookback_delta": []string{"30s"}, |
|
}, |
|
expect: promql.NewPrometheusQueryOpts(true, 30*time.Second), |
|
err: nil, |
|
}, |
|
{ |
|
name: "with invalid lookback delta", |
|
form: url.Values{ |
|
"lookback_delta": []string{"invalid"}, |
|
}, |
|
expect: nil, |
|
err: errors.New(`error parsing lookback delta duration: cannot parse "invalid" to a valid duration`), |
|
}, |
|
} |
|
|
|
for _, test := range tests { |
|
t.Run(test.name, func(t *testing.T) { |
|
req := &http.Request{Form: test.form} |
|
opts, err := extractQueryOpts(req) |
|
require.Equal(t, test.expect, opts) |
|
if test.err == nil { |
|
require.NoError(t, err) |
|
} else { |
|
require.Equal(t, test.err.Error(), err.Error()) |
|
} |
|
}) |
|
} |
|
}
|
|
|