mirror of https://github.com/prometheus/prometheus
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
4503 lines
116 KiB
4503 lines
116 KiB
// Copyright 2016 The Prometheus Authors
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package v1
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"net/http"
|
|
"net/http/httptest"
|
|
"net/url"
|
|
"os"
|
|
"reflect"
|
|
"runtime"
|
|
"sort"
|
|
"strconv"
|
|
"strings"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/prometheus/prometheus/prompb"
|
|
"github.com/prometheus/prometheus/util/stats"
|
|
"github.com/prometheus/prometheus/util/testutil"
|
|
|
|
jsoniter "github.com/json-iterator/go"
|
|
"github.com/prometheus/client_golang/prometheus"
|
|
config_util "github.com/prometheus/common/config"
|
|
"github.com/prometheus/common/model"
|
|
"github.com/prometheus/common/promslog"
|
|
"github.com/prometheus/common/route"
|
|
"github.com/stretchr/testify/require"
|
|
|
|
"github.com/prometheus/prometheus/config"
|
|
"github.com/prometheus/prometheus/model/exemplar"
|
|
"github.com/prometheus/prometheus/model/labels"
|
|
"github.com/prometheus/prometheus/model/metadata"
|
|
"github.com/prometheus/prometheus/model/timestamp"
|
|
"github.com/prometheus/prometheus/promql"
|
|
"github.com/prometheus/prometheus/promql/parser"
|
|
"github.com/prometheus/prometheus/promql/promqltest"
|
|
"github.com/prometheus/prometheus/rules"
|
|
"github.com/prometheus/prometheus/scrape"
|
|
"github.com/prometheus/prometheus/storage"
|
|
"github.com/prometheus/prometheus/storage/remote"
|
|
"github.com/prometheus/prometheus/tsdb"
|
|
"github.com/prometheus/prometheus/util/teststorage"
|
|
)
|
|
|
|
func testEngine(t *testing.T) *promql.Engine {
|
|
t.Helper()
|
|
return promqltest.NewTestEngineWithOpts(t, promql.EngineOpts{
|
|
Logger: nil,
|
|
Reg: nil,
|
|
MaxSamples: 10000,
|
|
Timeout: 100 * time.Second,
|
|
NoStepSubqueryIntervalFn: func(int64) int64 { return 60 * 1000 },
|
|
EnableAtModifier: true,
|
|
EnableNegativeOffset: true,
|
|
EnablePerStepStats: true,
|
|
})
|
|
}
|
|
|
|
// testMetaStore satisfies the scrape.MetricMetadataStore interface.
|
|
// It is used to inject specific metadata as part of a test case.
|
|
type testMetaStore struct {
|
|
Metadata []scrape.MetricMetadata
|
|
}
|
|
|
|
func (s *testMetaStore) ListMetadata() []scrape.MetricMetadata {
|
|
return s.Metadata
|
|
}
|
|
|
|
func (s *testMetaStore) GetMetadata(metric string) (scrape.MetricMetadata, bool) {
|
|
for _, m := range s.Metadata {
|
|
if metric == m.Metric {
|
|
return m, true
|
|
}
|
|
}
|
|
|
|
return scrape.MetricMetadata{}, false
|
|
}
|
|
|
|
func (s *testMetaStore) SizeMetadata() int { return 0 }
|
|
func (s *testMetaStore) LengthMetadata() int { return 0 }
|
|
|
|
// testTargetRetriever represents a list of targets to scrape.
|
|
// It is used to represent targets as part of test cases.
|
|
type testTargetRetriever struct {
|
|
activeTargets map[string][]*scrape.Target
|
|
droppedTargets map[string][]*scrape.Target
|
|
}
|
|
|
|
type testTargetParams struct {
|
|
Identifier string
|
|
Labels labels.Labels
|
|
DiscoveredLabels labels.Labels
|
|
Params url.Values
|
|
Reports []*testReport
|
|
Active bool
|
|
}
|
|
|
|
type testReport struct {
|
|
Start time.Time
|
|
Duration time.Duration
|
|
Error error
|
|
}
|
|
|
|
func newTestTargetRetriever(targetsInfo []*testTargetParams) *testTargetRetriever {
|
|
var activeTargets map[string][]*scrape.Target
|
|
var droppedTargets map[string][]*scrape.Target
|
|
activeTargets = make(map[string][]*scrape.Target)
|
|
droppedTargets = make(map[string][]*scrape.Target)
|
|
|
|
for _, t := range targetsInfo {
|
|
nt := scrape.NewTarget(t.Labels, t.DiscoveredLabels, t.Params)
|
|
|
|
for _, r := range t.Reports {
|
|
nt.Report(r.Start, r.Duration, r.Error)
|
|
}
|
|
|
|
if t.Active {
|
|
activeTargets[t.Identifier] = []*scrape.Target{nt}
|
|
} else {
|
|
droppedTargets[t.Identifier] = []*scrape.Target{nt}
|
|
}
|
|
}
|
|
|
|
return &testTargetRetriever{
|
|
activeTargets: activeTargets,
|
|
droppedTargets: droppedTargets,
|
|
}
|
|
}
|
|
|
|
var scrapeStart = time.Now().Add(-11 * time.Second)
|
|
|
|
func (t testTargetRetriever) TargetsActive() map[string][]*scrape.Target {
|
|
return t.activeTargets
|
|
}
|
|
|
|
func (t testTargetRetriever) TargetsDropped() map[string][]*scrape.Target {
|
|
return t.droppedTargets
|
|
}
|
|
|
|
func (t testTargetRetriever) TargetsDroppedCounts() map[string]int {
|
|
r := make(map[string]int)
|
|
for k, v := range t.droppedTargets {
|
|
r[k] = len(v)
|
|
}
|
|
return r
|
|
}
|
|
|
|
func (t *testTargetRetriever) SetMetadataStoreForTargets(identifier string, metadata scrape.MetricMetadataStore) error {
|
|
targets, ok := t.activeTargets[identifier]
|
|
|
|
if !ok {
|
|
return errors.New("targets not found")
|
|
}
|
|
|
|
for _, at := range targets {
|
|
at.SetMetadataStore(metadata)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (t *testTargetRetriever) ResetMetadataStore() {
|
|
for _, at := range t.activeTargets {
|
|
for _, tt := range at {
|
|
tt.SetMetadataStore(&testMetaStore{})
|
|
}
|
|
}
|
|
}
|
|
|
|
func (t *testTargetRetriever) toFactory() func(context.Context) TargetRetriever {
|
|
return func(context.Context) TargetRetriever { return t }
|
|
}
|
|
|
|
type testAlertmanagerRetriever struct{}
|
|
|
|
func (t testAlertmanagerRetriever) Alertmanagers() []*url.URL {
|
|
return []*url.URL{
|
|
{
|
|
Scheme: "http",
|
|
Host: "alertmanager.example.com:8080",
|
|
Path: "/api/v1/alerts",
|
|
},
|
|
}
|
|
}
|
|
|
|
func (t testAlertmanagerRetriever) DroppedAlertmanagers() []*url.URL {
|
|
return []*url.URL{
|
|
{
|
|
Scheme: "http",
|
|
Host: "dropped.alertmanager.example.com:8080",
|
|
Path: "/api/v1/alerts",
|
|
},
|
|
}
|
|
}
|
|
|
|
func (t testAlertmanagerRetriever) toFactory() func(context.Context) AlertmanagerRetriever {
|
|
return func(context.Context) AlertmanagerRetriever { return t }
|
|
}
|
|
|
|
type rulesRetrieverMock struct {
|
|
alertingRules []*rules.AlertingRule
|
|
ruleGroups []*rules.Group
|
|
testing *testing.T
|
|
}
|
|
|
|
func (m *rulesRetrieverMock) CreateAlertingRules() {
|
|
expr1, err := parser.ParseExpr(`absent(test_metric3) != 1`)
|
|
require.NoError(m.testing, err)
|
|
expr2, err := parser.ParseExpr(`up == 1`)
|
|
require.NoError(m.testing, err)
|
|
expr3, err := parser.ParseExpr(`vector(1)`)
|
|
require.NoError(m.testing, err)
|
|
|
|
rule1 := rules.NewAlertingRule(
|
|
"test_metric3",
|
|
expr1,
|
|
time.Second,
|
|
0,
|
|
labels.Labels{},
|
|
labels.Labels{},
|
|
labels.Labels{},
|
|
"",
|
|
true,
|
|
promslog.NewNopLogger(),
|
|
)
|
|
rule2 := rules.NewAlertingRule(
|
|
"test_metric4",
|
|
expr2,
|
|
time.Second,
|
|
0,
|
|
labels.Labels{},
|
|
labels.Labels{},
|
|
labels.Labels{},
|
|
"",
|
|
true,
|
|
promslog.NewNopLogger(),
|
|
)
|
|
rule3 := rules.NewAlertingRule(
|
|
"test_metric5",
|
|
expr3,
|
|
time.Second,
|
|
0,
|
|
labels.FromStrings("name", "tm5"),
|
|
labels.Labels{},
|
|
labels.FromStrings("name", "tm5"),
|
|
"",
|
|
false,
|
|
promslog.NewNopLogger(),
|
|
)
|
|
rule4 := rules.NewAlertingRule(
|
|
"test_metric6",
|
|
expr2,
|
|
time.Second,
|
|
0,
|
|
labels.FromStrings("testlabel", "rule"),
|
|
labels.Labels{},
|
|
labels.Labels{},
|
|
"",
|
|
true,
|
|
promslog.NewNopLogger(),
|
|
)
|
|
rule5 := rules.NewAlertingRule(
|
|
"test_metric7",
|
|
expr2,
|
|
time.Second,
|
|
0,
|
|
labels.FromStrings("templatedlabel", "{{ $externalURL }}"),
|
|
labels.Labels{},
|
|
labels.Labels{},
|
|
"",
|
|
true,
|
|
promslog.NewNopLogger(),
|
|
)
|
|
var r []*rules.AlertingRule
|
|
r = append(r, rule1)
|
|
r = append(r, rule2)
|
|
r = append(r, rule3)
|
|
r = append(r, rule4)
|
|
r = append(r, rule5)
|
|
m.alertingRules = r
|
|
}
|
|
|
|
func (m *rulesRetrieverMock) CreateRuleGroups() {
|
|
m.CreateAlertingRules()
|
|
arules := m.AlertingRules()
|
|
storage := teststorage.New(m.testing)
|
|
defer storage.Close()
|
|
|
|
engineOpts := promql.EngineOpts{
|
|
Logger: nil,
|
|
Reg: nil,
|
|
MaxSamples: 10,
|
|
Timeout: 100 * time.Second,
|
|
}
|
|
engine := promqltest.NewTestEngineWithOpts(m.testing, engineOpts)
|
|
opts := &rules.ManagerOptions{
|
|
QueryFunc: rules.EngineQueryFunc(engine, storage),
|
|
Appendable: storage,
|
|
Context: context.Background(),
|
|
Logger: promslog.NewNopLogger(),
|
|
NotifyFunc: func(ctx context.Context, expr string, alerts ...*rules.Alert) {},
|
|
}
|
|
|
|
var r []rules.Rule
|
|
|
|
for _, alertrule := range arules {
|
|
r = append(r, alertrule)
|
|
}
|
|
|
|
recordingExpr, err := parser.ParseExpr(`vector(1)`)
|
|
require.NoError(m.testing, err, "unable to parse alert expression")
|
|
recordingRule := rules.NewRecordingRule("recording-rule-1", recordingExpr, labels.Labels{})
|
|
recordingRule2 := rules.NewRecordingRule("recording-rule-2", recordingExpr, labels.FromStrings("testlabel", "rule"))
|
|
r = append(r, recordingRule)
|
|
r = append(r, recordingRule2)
|
|
|
|
group := rules.NewGroup(rules.GroupOptions{
|
|
Name: "grp",
|
|
File: "/path/to/file",
|
|
Interval: time.Second,
|
|
Rules: r,
|
|
ShouldRestore: false,
|
|
Opts: opts,
|
|
})
|
|
group2 := rules.NewGroup(rules.GroupOptions{
|
|
Name: "grp2",
|
|
File: "/path/to/file",
|
|
Interval: time.Second,
|
|
Rules: []rules.Rule{r[0]},
|
|
ShouldRestore: false,
|
|
Opts: opts,
|
|
})
|
|
m.ruleGroups = []*rules.Group{group, group2}
|
|
}
|
|
|
|
func (m *rulesRetrieverMock) AlertingRules() []*rules.AlertingRule {
|
|
return m.alertingRules
|
|
}
|
|
|
|
func (m *rulesRetrieverMock) RuleGroups() []*rules.Group {
|
|
return m.ruleGroups
|
|
}
|
|
|
|
func (m *rulesRetrieverMock) toFactory() func(context.Context) RulesRetriever {
|
|
return func(context.Context) RulesRetriever { return m }
|
|
}
|
|
|
|
var samplePrometheusCfg = config.Config{
|
|
GlobalConfig: config.GlobalConfig{},
|
|
AlertingConfig: config.AlertingConfig{},
|
|
RuleFiles: []string{},
|
|
ScrapeConfigs: []*config.ScrapeConfig{},
|
|
RemoteWriteConfigs: []*config.RemoteWriteConfig{},
|
|
RemoteReadConfigs: []*config.RemoteReadConfig{},
|
|
OTLPConfig: config.OTLPConfig{},
|
|
}
|
|
|
|
var sampleFlagMap = map[string]string{
|
|
"flag1": "value1",
|
|
"flag2": "value2",
|
|
}
|
|
|
|
func TestEndpoints(t *testing.T) {
|
|
storage := promqltest.LoadedStorage(t, `
|
|
load 1m
|
|
test_metric1{foo="bar"} 0+100x100
|
|
test_metric1{foo="boo"} 1+0x100
|
|
test_metric2{foo="boo"} 1+0x100
|
|
test_metric3{foo="bar", dup="1"} 1+0x100
|
|
test_metric3{foo="boo", dup="1"} 1+0x100
|
|
test_metric4{foo="bar", dup="1"} 1+0x100
|
|
test_metric4{foo="boo", dup="1"} 1+0x100
|
|
test_metric4{foo="boo"} 1+0x100
|
|
`)
|
|
t.Cleanup(func() { storage.Close() })
|
|
|
|
start := time.Unix(0, 0)
|
|
exemplars := []exemplar.QueryResult{
|
|
{
|
|
SeriesLabels: labels.FromStrings("__name__", "test_metric3", "foo", "boo", "dup", "1"),
|
|
Exemplars: []exemplar.Exemplar{
|
|
{
|
|
Labels: labels.FromStrings("id", "abc"),
|
|
Value: 10,
|
|
Ts: timestamp.FromTime(start.Add(2 * time.Second)),
|
|
},
|
|
},
|
|
},
|
|
{
|
|
SeriesLabels: labels.FromStrings("__name__", "test_metric4", "foo", "bar", "dup", "1"),
|
|
Exemplars: []exemplar.Exemplar{
|
|
{
|
|
Labels: labels.FromStrings("id", "lul"),
|
|
Value: 10,
|
|
Ts: timestamp.FromTime(start.Add(4 * time.Second)),
|
|
},
|
|
},
|
|
},
|
|
{
|
|
SeriesLabels: labels.FromStrings("__name__", "test_metric3", "foo", "boo", "dup", "1"),
|
|
Exemplars: []exemplar.Exemplar{
|
|
{
|
|
Labels: labels.FromStrings("id", "abc2"),
|
|
Value: 10,
|
|
Ts: timestamp.FromTime(start.Add(4053 * time.Millisecond)),
|
|
},
|
|
},
|
|
},
|
|
{
|
|
SeriesLabels: labels.FromStrings("__name__", "test_metric4", "foo", "bar", "dup", "1"),
|
|
Exemplars: []exemplar.Exemplar{
|
|
{
|
|
Labels: labels.FromStrings("id", "lul2"),
|
|
Value: 10,
|
|
Ts: timestamp.FromTime(start.Add(4153 * time.Millisecond)),
|
|
},
|
|
},
|
|
},
|
|
}
|
|
for _, ed := range exemplars {
|
|
_, err := storage.AppendExemplar(0, ed.SeriesLabels, ed.Exemplars[0])
|
|
require.NoError(t, err, "failed to add exemplar: %+v", ed.Exemplars[0])
|
|
}
|
|
|
|
now := time.Now()
|
|
|
|
ng := testEngine(t)
|
|
|
|
t.Run("local", func(t *testing.T) {
|
|
algr := rulesRetrieverMock{testing: t}
|
|
|
|
algr.CreateAlertingRules()
|
|
algr.CreateRuleGroups()
|
|
|
|
g := algr.RuleGroups()
|
|
g[0].Eval(context.Background(), time.Now())
|
|
|
|
testTargetRetriever := setupTestTargetRetriever(t)
|
|
|
|
api := &API{
|
|
Queryable: storage,
|
|
QueryEngine: ng,
|
|
ExemplarQueryable: storage.ExemplarQueryable(),
|
|
targetRetriever: testTargetRetriever.toFactory(),
|
|
alertmanagerRetriever: testAlertmanagerRetriever{}.toFactory(),
|
|
flagsMap: sampleFlagMap,
|
|
now: func() time.Time { return now },
|
|
config: func() config.Config { return samplePrometheusCfg },
|
|
ready: func(f http.HandlerFunc) http.HandlerFunc { return f },
|
|
rulesRetriever: algr.toFactory(),
|
|
}
|
|
testEndpoints(t, api, testTargetRetriever, storage, true)
|
|
})
|
|
|
|
// Run all the API tests against an API that is wired to forward queries via
|
|
// the remote read client to a test server, which in turn sends them to the
|
|
// data from the test storage.
|
|
t.Run("remote", func(t *testing.T) {
|
|
server := setupRemote(storage)
|
|
defer server.Close()
|
|
|
|
u, err := url.Parse(server.URL)
|
|
require.NoError(t, err)
|
|
|
|
al := promslog.AllowedLevel{}
|
|
require.NoError(t, al.Set("debug"))
|
|
|
|
af := promslog.AllowedFormat{}
|
|
require.NoError(t, af.Set("logfmt"))
|
|
|
|
promslogConfig := promslog.Config{
|
|
Level: &al,
|
|
Format: &af,
|
|
}
|
|
|
|
dbDir := t.TempDir()
|
|
|
|
remote := remote.NewStorage(promslog.New(&promslogConfig), prometheus.DefaultRegisterer, func() (int64, error) {
|
|
return 0, nil
|
|
}, dbDir, 1*time.Second, nil, false)
|
|
|
|
err = remote.ApplyConfig(&config.Config{
|
|
RemoteReadConfigs: []*config.RemoteReadConfig{
|
|
{
|
|
URL: &config_util.URL{URL: u},
|
|
RemoteTimeout: model.Duration(1 * time.Second),
|
|
ReadRecent: true,
|
|
},
|
|
},
|
|
})
|
|
require.NoError(t, err)
|
|
|
|
algr := rulesRetrieverMock{testing: t}
|
|
|
|
algr.CreateAlertingRules()
|
|
algr.CreateRuleGroups()
|
|
|
|
g := algr.RuleGroups()
|
|
g[0].Eval(context.Background(), time.Now())
|
|
|
|
testTargetRetriever := setupTestTargetRetriever(t)
|
|
|
|
api := &API{
|
|
Queryable: remote,
|
|
QueryEngine: ng,
|
|
ExemplarQueryable: storage.ExemplarQueryable(),
|
|
targetRetriever: testTargetRetriever.toFactory(),
|
|
alertmanagerRetriever: testAlertmanagerRetriever{}.toFactory(),
|
|
flagsMap: sampleFlagMap,
|
|
now: func() time.Time { return now },
|
|
config: func() config.Config { return samplePrometheusCfg },
|
|
ready: func(f http.HandlerFunc) http.HandlerFunc { return f },
|
|
rulesRetriever: algr.toFactory(),
|
|
}
|
|
testEndpoints(t, api, testTargetRetriever, storage, false)
|
|
})
|
|
}
|
|
|
|
type byLabels []labels.Labels
|
|
|
|
func (b byLabels) Len() int { return len(b) }
|
|
func (b byLabels) Swap(i, j int) { b[i], b[j] = b[j], b[i] }
|
|
func (b byLabels) Less(i, j int) bool { return labels.Compare(b[i], b[j]) < 0 }
|
|
|
|
func TestGetSeries(t *testing.T) {
|
|
// TestEndpoints doesn't have enough label names to test api.labelNames
|
|
// endpoint properly. Hence we test it separately.
|
|
storage := promqltest.LoadedStorage(t, `
|
|
load 1m
|
|
test_metric1{foo1="bar", baz="abc"} 0+100x100
|
|
test_metric1{foo2="boo"} 1+0x100
|
|
test_metric2{foo="boo"} 1+0x100
|
|
test_metric2{foo="boo", xyz="qwerty"} 1+0x100
|
|
test_metric2{foo="baz", abc="qwerty"} 1+0x100
|
|
`)
|
|
t.Cleanup(func() { storage.Close() })
|
|
api := &API{
|
|
Queryable: storage,
|
|
}
|
|
request := func(method string, matchers ...string) (*http.Request, error) {
|
|
u, err := url.Parse("http://example.com")
|
|
require.NoError(t, err)
|
|
q := u.Query()
|
|
for _, matcher := range matchers {
|
|
q.Add("match[]", matcher)
|
|
}
|
|
u.RawQuery = q.Encode()
|
|
|
|
r, err := http.NewRequest(method, u.String(), nil)
|
|
if method == http.MethodPost {
|
|
r.Header.Set("Content-Type", "application/x-www-form-urlencoded")
|
|
}
|
|
return r, err
|
|
}
|
|
|
|
for _, tc := range []struct {
|
|
name string
|
|
api *API
|
|
matchers []string
|
|
expected []labels.Labels
|
|
expectedErrorType errorType
|
|
}{
|
|
{
|
|
name: "no matchers",
|
|
expectedErrorType: errorBadData,
|
|
api: api,
|
|
},
|
|
{
|
|
name: "non empty label matcher",
|
|
matchers: []string{`{foo=~".+"}`},
|
|
expected: []labels.Labels{
|
|
labels.FromStrings("__name__", "test_metric2", "abc", "qwerty", "foo", "baz"),
|
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"),
|
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo", "xyz", "qwerty"),
|
|
},
|
|
api: api,
|
|
},
|
|
{
|
|
name: "exact label matcher",
|
|
matchers: []string{`{foo="boo"}`},
|
|
expected: []labels.Labels{
|
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"),
|
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo", "xyz", "qwerty"),
|
|
},
|
|
api: api,
|
|
},
|
|
{
|
|
name: "two matchers",
|
|
matchers: []string{`{foo="boo"}`, `{foo="baz"}`},
|
|
expected: []labels.Labels{
|
|
labels.FromStrings("__name__", "test_metric2", "abc", "qwerty", "foo", "baz"),
|
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"),
|
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo", "xyz", "qwerty"),
|
|
},
|
|
api: api,
|
|
},
|
|
{
|
|
name: "exec error type",
|
|
matchers: []string{`{foo="boo"}`, `{foo="baz"}`},
|
|
expectedErrorType: errorExec,
|
|
api: &API{
|
|
Queryable: errorTestQueryable{err: errors.New("generic")},
|
|
},
|
|
},
|
|
{
|
|
name: "storage error type",
|
|
matchers: []string{`{foo="boo"}`, `{foo="baz"}`},
|
|
expectedErrorType: errorInternal,
|
|
api: &API{
|
|
Queryable: errorTestQueryable{err: promql.ErrStorage{Err: errors.New("generic")}},
|
|
},
|
|
},
|
|
} {
|
|
t.Run(tc.name, func(t *testing.T) {
|
|
ctx := context.Background()
|
|
req, err := request(http.MethodGet, tc.matchers...)
|
|
require.NoError(t, err)
|
|
res := tc.api.series(req.WithContext(ctx))
|
|
assertAPIError(t, res.err, tc.expectedErrorType)
|
|
if tc.expectedErrorType == errorNone {
|
|
r := res.data.([]labels.Labels)
|
|
sort.Sort(byLabels(tc.expected))
|
|
sort.Sort(byLabels(r))
|
|
testutil.RequireEqual(t, tc.expected, r)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestQueryExemplars(t *testing.T) {
|
|
start := time.Unix(0, 0)
|
|
storage := promqltest.LoadedStorage(t, `
|
|
load 1m
|
|
test_metric1{foo="bar"} 0+100x100
|
|
test_metric1{foo="boo"} 1+0x100
|
|
test_metric2{foo="boo"} 1+0x100
|
|
test_metric3{foo="bar", dup="1"} 1+0x100
|
|
test_metric3{foo="boo", dup="1"} 1+0x100
|
|
test_metric4{foo="bar", dup="1"} 1+0x100
|
|
test_metric4{foo="boo", dup="1"} 1+0x100
|
|
test_metric4{foo="boo"} 1+0x100
|
|
`)
|
|
t.Cleanup(func() { storage.Close() })
|
|
|
|
api := &API{
|
|
Queryable: storage,
|
|
QueryEngine: testEngine(t),
|
|
ExemplarQueryable: storage.ExemplarQueryable(),
|
|
}
|
|
|
|
request := func(method string, qs url.Values) (*http.Request, error) {
|
|
u, err := url.Parse("http://example.com")
|
|
require.NoError(t, err)
|
|
u.RawQuery = qs.Encode()
|
|
r, err := http.NewRequest(method, u.String(), nil)
|
|
if method == http.MethodPost {
|
|
r.Header.Set("Content-Type", "application/x-www-form-urlencoded")
|
|
}
|
|
return r, err
|
|
}
|
|
|
|
for _, tc := range []struct {
|
|
name string
|
|
query url.Values
|
|
exemplars []exemplar.QueryResult
|
|
api *API
|
|
expectedErrorType errorType
|
|
}{
|
|
{
|
|
name: "no error",
|
|
api: api,
|
|
query: url.Values{
|
|
"query": []string{`test_metric3{foo="boo"} - test_metric4{foo="bar"}`},
|
|
"start": []string{"0"},
|
|
"end": []string{"4"},
|
|
},
|
|
exemplars: []exemplar.QueryResult{
|
|
{
|
|
SeriesLabels: labels.FromStrings("__name__", "test_metric3", "foo", "boo", "dup", "1"),
|
|
Exemplars: []exemplar.Exemplar{
|
|
{
|
|
Labels: labels.FromStrings("id", "abc"),
|
|
Value: 10,
|
|
Ts: timestamp.FromTime(start.Add(0 * time.Second)),
|
|
},
|
|
},
|
|
},
|
|
{
|
|
SeriesLabels: labels.FromStrings("__name__", "test_metric4", "foo", "bar", "dup", "1"),
|
|
Exemplars: []exemplar.Exemplar{
|
|
{
|
|
Labels: labels.FromStrings("id", "lul"),
|
|
Value: 10,
|
|
Ts: timestamp.FromTime(start.Add(3 * time.Second)),
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
name: "should return errorExec upon genetic error",
|
|
expectedErrorType: errorExec,
|
|
api: &API{
|
|
ExemplarQueryable: errorTestQueryable{err: errors.New("generic")},
|
|
},
|
|
query: url.Values{
|
|
"query": []string{`test_metric3{foo="boo"} - test_metric4{foo="bar"}`},
|
|
"start": []string{"0"},
|
|
"end": []string{"4"},
|
|
},
|
|
},
|
|
{
|
|
name: "should return errorInternal err type is ErrStorage",
|
|
expectedErrorType: errorInternal,
|
|
api: &API{
|
|
ExemplarQueryable: errorTestQueryable{err: promql.ErrStorage{Err: errors.New("generic")}},
|
|
},
|
|
query: url.Values{
|
|
"query": []string{`test_metric3{foo="boo"} - test_metric4{foo="bar"}`},
|
|
"start": []string{"0"},
|
|
"end": []string{"4"},
|
|
},
|
|
},
|
|
} {
|
|
t.Run(tc.name, func(t *testing.T) {
|
|
es := storage
|
|
ctx := context.Background()
|
|
|
|
for _, te := range tc.exemplars {
|
|
for _, e := range te.Exemplars {
|
|
_, err := es.AppendExemplar(0, te.SeriesLabels, e)
|
|
require.NoError(t, err)
|
|
}
|
|
}
|
|
|
|
req, err := request(http.MethodGet, tc.query)
|
|
require.NoError(t, err)
|
|
res := tc.api.queryExemplars(req.WithContext(ctx))
|
|
assertAPIError(t, res.err, tc.expectedErrorType)
|
|
|
|
if tc.expectedErrorType == errorNone {
|
|
assertAPIResponse(t, res.data, tc.exemplars)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestLabelNames(t *testing.T) {
|
|
// TestEndpoints doesn't have enough label names to test api.labelNames
|
|
// endpoint properly. Hence we test it separately.
|
|
storage := promqltest.LoadedStorage(t, `
|
|
load 1m
|
|
test_metric1{foo1="bar", baz="abc"} 0+100x100
|
|
test_metric1{foo2="boo"} 1+0x100
|
|
test_metric2{foo="boo"} 1+0x100
|
|
test_metric2{foo="boo", xyz="qwerty"} 1+0x100
|
|
test_metric2{foo="baz", abc="qwerty"} 1+0x100
|
|
`)
|
|
t.Cleanup(func() { storage.Close() })
|
|
api := &API{
|
|
Queryable: storage,
|
|
}
|
|
request := func(method, limit string, matchers ...string) (*http.Request, error) {
|
|
u, err := url.Parse("http://example.com")
|
|
require.NoError(t, err)
|
|
q := u.Query()
|
|
for _, matcher := range matchers {
|
|
q.Add("match[]", matcher)
|
|
}
|
|
if limit != "" {
|
|
q.Add("limit", limit)
|
|
}
|
|
u.RawQuery = q.Encode()
|
|
|
|
r, err := http.NewRequest(method, u.String(), nil)
|
|
if method == http.MethodPost {
|
|
r.Header.Set("Content-Type", "application/x-www-form-urlencoded")
|
|
}
|
|
return r, err
|
|
}
|
|
|
|
for _, tc := range []struct {
|
|
name string
|
|
api *API
|
|
matchers []string
|
|
limit string
|
|
expected []string
|
|
expectedErrorType errorType
|
|
}{
|
|
{
|
|
name: "no matchers",
|
|
expected: []string{"__name__", "abc", "baz", "foo", "foo1", "foo2", "xyz"},
|
|
api: api,
|
|
},
|
|
{
|
|
name: "non empty label matcher",
|
|
matchers: []string{`{foo=~".+"}`},
|
|
expected: []string{"__name__", "abc", "foo", "xyz"},
|
|
api: api,
|
|
},
|
|
{
|
|
name: "non empty label matcher with limit",
|
|
matchers: []string{`{foo=~".+"}`},
|
|
expected: []string{"__name__", "abc"},
|
|
limit: "2",
|
|
api: api,
|
|
},
|
|
{
|
|
name: "exact label matcher",
|
|
matchers: []string{`{foo="boo"}`},
|
|
expected: []string{"__name__", "foo", "xyz"},
|
|
api: api,
|
|
},
|
|
{
|
|
name: "two matchers",
|
|
matchers: []string{`{foo="boo"}`, `{foo="baz"}`},
|
|
expected: []string{"__name__", "abc", "foo", "xyz"},
|
|
api: api,
|
|
},
|
|
{
|
|
name: "exec error type",
|
|
matchers: []string{`{foo="boo"}`, `{foo="baz"}`},
|
|
expectedErrorType: errorExec,
|
|
api: &API{
|
|
Queryable: errorTestQueryable{err: errors.New("generic")},
|
|
},
|
|
},
|
|
{
|
|
name: "storage error type",
|
|
matchers: []string{`{foo="boo"}`, `{foo="baz"}`},
|
|
expectedErrorType: errorInternal,
|
|
api: &API{
|
|
Queryable: errorTestQueryable{err: promql.ErrStorage{Err: errors.New("generic")}},
|
|
},
|
|
},
|
|
} {
|
|
t.Run(tc.name, func(t *testing.T) {
|
|
for _, method := range []string{http.MethodGet, http.MethodPost} {
|
|
ctx := context.Background()
|
|
req, err := request(method, tc.limit, tc.matchers...)
|
|
require.NoError(t, err)
|
|
res := tc.api.labelNames(req.WithContext(ctx))
|
|
assertAPIError(t, res.err, tc.expectedErrorType)
|
|
if tc.expectedErrorType == errorNone {
|
|
assertAPIResponse(t, res.data, tc.expected)
|
|
}
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
type testStats struct {
|
|
Custom string `json:"custom"`
|
|
}
|
|
|
|
func (testStats) Builtin() (_ stats.BuiltinStats) {
|
|
return
|
|
}
|
|
|
|
func TestStats(t *testing.T) {
|
|
storage := teststorage.New(t)
|
|
t.Cleanup(func() { storage.Close() })
|
|
|
|
api := &API{
|
|
Queryable: storage,
|
|
QueryEngine: testEngine(t),
|
|
now: func() time.Time {
|
|
return time.Unix(123, 0)
|
|
},
|
|
}
|
|
request := func(method, param string) (*http.Request, error) {
|
|
u, err := url.Parse("http://example.com")
|
|
require.NoError(t, err)
|
|
q := u.Query()
|
|
q.Add("stats", param)
|
|
q.Add("query", "up")
|
|
q.Add("start", "0")
|
|
q.Add("end", "100")
|
|
q.Add("step", "10")
|
|
u.RawQuery = q.Encode()
|
|
|
|
r, err := http.NewRequest(method, u.String(), nil)
|
|
if method == http.MethodPost {
|
|
r.Header.Set("Content-Type", "application/x-www-form-urlencoded")
|
|
}
|
|
return r, err
|
|
}
|
|
|
|
for _, tc := range []struct {
|
|
name string
|
|
renderer StatsRenderer
|
|
param string
|
|
expected func(*testing.T, interface{})
|
|
}{
|
|
{
|
|
name: "stats is blank",
|
|
param: "",
|
|
expected: func(t *testing.T, i interface{}) {
|
|
require.IsType(t, &QueryData{}, i)
|
|
qd := i.(*QueryData)
|
|
require.Nil(t, qd.Stats)
|
|
},
|
|
},
|
|
{
|
|
name: "stats is true",
|
|
param: "true",
|
|
expected: func(t *testing.T, i interface{}) {
|
|
require.IsType(t, &QueryData{}, i)
|
|
qd := i.(*QueryData)
|
|
require.NotNil(t, qd.Stats)
|
|
qs := qd.Stats.Builtin()
|
|
require.NotNil(t, qs.Timings)
|
|
require.Greater(t, qs.Timings.EvalTotalTime, float64(0))
|
|
require.NotNil(t, qs.Samples)
|
|
require.NotNil(t, qs.Samples.TotalQueryableSamples)
|
|
require.Nil(t, qs.Samples.TotalQueryableSamplesPerStep)
|
|
},
|
|
},
|
|
{
|
|
name: "stats is all",
|
|
param: "all",
|
|
expected: func(t *testing.T, i interface{}) {
|
|
require.IsType(t, &QueryData{}, i)
|
|
qd := i.(*QueryData)
|
|
require.NotNil(t, qd.Stats)
|
|
qs := qd.Stats.Builtin()
|
|
require.NotNil(t, qs.Timings)
|
|
require.Greater(t, qs.Timings.EvalTotalTime, float64(0))
|
|
require.NotNil(t, qs.Samples)
|
|
require.NotNil(t, qs.Samples.TotalQueryableSamples)
|
|
require.NotNil(t, qs.Samples.TotalQueryableSamplesPerStep)
|
|
},
|
|
},
|
|
{
|
|
name: "custom handler with known value",
|
|
renderer: func(ctx context.Context, s *stats.Statistics, p string) stats.QueryStats {
|
|
if p == "known" {
|
|
return testStats{"Custom Value"}
|
|
}
|
|
return nil
|
|
},
|
|
param: "known",
|
|
expected: func(t *testing.T, i interface{}) {
|
|
require.IsType(t, &QueryData{}, i)
|
|
qd := i.(*QueryData)
|
|
require.NotNil(t, qd.Stats)
|
|
json := jsoniter.ConfigCompatibleWithStandardLibrary
|
|
j, err := json.Marshal(qd.Stats)
|
|
require.NoError(t, err)
|
|
require.JSONEq(t, `{"custom":"Custom Value"}`, string(j))
|
|
},
|
|
},
|
|
} {
|
|
t.Run(tc.name, func(t *testing.T) {
|
|
before := api.statsRenderer
|
|
defer func() { api.statsRenderer = before }()
|
|
api.statsRenderer = tc.renderer
|
|
|
|
for _, method := range []string{http.MethodGet, http.MethodPost} {
|
|
ctx := context.Background()
|
|
req, err := request(method, tc.param)
|
|
require.NoError(t, err)
|
|
res := api.query(req.WithContext(ctx))
|
|
assertAPIError(t, res.err, "")
|
|
tc.expected(t, res.data)
|
|
|
|
res = api.queryRange(req.WithContext(ctx))
|
|
assertAPIError(t, res.err, "")
|
|
tc.expected(t, res.data)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func setupTestTargetRetriever(t *testing.T) *testTargetRetriever {
|
|
t.Helper()
|
|
|
|
targets := []*testTargetParams{
|
|
{
|
|
Identifier: "test",
|
|
Labels: labels.FromMap(map[string]string{
|
|
model.SchemeLabel: "http",
|
|
model.AddressLabel: "example.com:8080",
|
|
model.MetricsPathLabel: "/metrics",
|
|
model.JobLabel: "test",
|
|
model.ScrapeIntervalLabel: "15s",
|
|
model.ScrapeTimeoutLabel: "5s",
|
|
}),
|
|
DiscoveredLabels: labels.EmptyLabels(),
|
|
Params: url.Values{},
|
|
Reports: []*testReport{{scrapeStart, 70 * time.Millisecond, nil}},
|
|
Active: true,
|
|
},
|
|
{
|
|
Identifier: "blackbox",
|
|
Labels: labels.FromMap(map[string]string{
|
|
model.SchemeLabel: "http",
|
|
model.AddressLabel: "localhost:9115",
|
|
model.MetricsPathLabel: "/probe",
|
|
model.JobLabel: "blackbox",
|
|
model.ScrapeIntervalLabel: "20s",
|
|
model.ScrapeTimeoutLabel: "10s",
|
|
}),
|
|
DiscoveredLabels: labels.EmptyLabels(),
|
|
Params: url.Values{"target": []string{"example.com"}},
|
|
Reports: []*testReport{{scrapeStart, 100 * time.Millisecond, errors.New("failed")}},
|
|
Active: true,
|
|
},
|
|
{
|
|
Identifier: "blackbox",
|
|
Labels: labels.EmptyLabels(),
|
|
DiscoveredLabels: labels.FromMap(map[string]string{
|
|
model.SchemeLabel: "http",
|
|
model.AddressLabel: "http://dropped.example.com:9115",
|
|
model.MetricsPathLabel: "/probe",
|
|
model.JobLabel: "blackbox",
|
|
model.ScrapeIntervalLabel: "30s",
|
|
model.ScrapeTimeoutLabel: "15s",
|
|
}),
|
|
Params: url.Values{},
|
|
Active: false,
|
|
},
|
|
}
|
|
|
|
return newTestTargetRetriever(targets)
|
|
}
|
|
|
|
func setupRemote(s storage.Storage) *httptest.Server {
|
|
handler := http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
|
req, err := remote.DecodeReadRequest(r)
|
|
if err != nil {
|
|
http.Error(w, err.Error(), http.StatusBadRequest)
|
|
return
|
|
}
|
|
resp := prompb.ReadResponse{
|
|
Results: make([]*prompb.QueryResult, len(req.Queries)),
|
|
}
|
|
for i, query := range req.Queries {
|
|
matchers, err := remote.FromLabelMatchers(query.Matchers)
|
|
if err != nil {
|
|
http.Error(w, err.Error(), http.StatusBadRequest)
|
|
return
|
|
}
|
|
|
|
var hints *storage.SelectHints
|
|
if query.Hints != nil {
|
|
hints = &storage.SelectHints{
|
|
Start: query.Hints.StartMs,
|
|
End: query.Hints.EndMs,
|
|
Step: query.Hints.StepMs,
|
|
Func: query.Hints.Func,
|
|
}
|
|
}
|
|
|
|
querier, err := s.Querier(query.StartTimestampMs, query.EndTimestampMs)
|
|
if err != nil {
|
|
http.Error(w, err.Error(), http.StatusInternalServerError)
|
|
return
|
|
}
|
|
defer querier.Close()
|
|
|
|
set := querier.Select(r.Context(), false, hints, matchers...)
|
|
resp.Results[i], _, err = remote.ToQueryResult(set, 1e6)
|
|
if err != nil {
|
|
http.Error(w, err.Error(), http.StatusInternalServerError)
|
|
return
|
|
}
|
|
}
|
|
|
|
w.Header().Set("Content-Type", "application/x-protobuf")
|
|
w.Header().Set("Content-Encoding", "snappy")
|
|
|
|
if err := remote.EncodeReadResponse(&resp, w); err != nil {
|
|
http.Error(w, err.Error(), http.StatusInternalServerError)
|
|
return
|
|
}
|
|
})
|
|
|
|
return httptest.NewServer(handler)
|
|
}
|
|
|
|
func testEndpoints(t *testing.T, api *API, tr *testTargetRetriever, es storage.ExemplarStorage, testLabelAPI bool) {
|
|
start := time.Unix(0, 0)
|
|
|
|
type targetMetadata struct {
|
|
identifier string
|
|
metadata []scrape.MetricMetadata
|
|
}
|
|
|
|
type test struct {
|
|
endpoint apiFunc
|
|
params map[string]string
|
|
query url.Values
|
|
response interface{}
|
|
responseLen int // If nonzero, check only the length; `response` is ignored.
|
|
responseMetadataTotal int
|
|
responseAsJSON string
|
|
warningsCount int
|
|
errType errorType
|
|
sorter func(interface{})
|
|
metadata []targetMetadata
|
|
exemplars []exemplar.QueryResult
|
|
zeroFunc func(interface{})
|
|
}
|
|
|
|
rulesZeroFunc := func(i interface{}) {
|
|
if i != nil {
|
|
v := i.(*RuleDiscovery)
|
|
for _, ruleGroup := range v.RuleGroups {
|
|
ruleGroup.EvaluationTime = float64(0)
|
|
ruleGroup.LastEvaluation = time.Time{}
|
|
for k, rule := range ruleGroup.Rules {
|
|
switch r := rule.(type) {
|
|
case AlertingRule:
|
|
r.LastEvaluation = time.Time{}
|
|
r.EvaluationTime = float64(0)
|
|
r.LastError = ""
|
|
r.Health = "ok"
|
|
for _, alert := range r.Alerts {
|
|
alert.ActiveAt = nil
|
|
}
|
|
ruleGroup.Rules[k] = r
|
|
case RecordingRule:
|
|
r.LastEvaluation = time.Time{}
|
|
r.EvaluationTime = float64(0)
|
|
r.LastError = ""
|
|
r.Health = "ok"
|
|
ruleGroup.Rules[k] = r
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
tests := []test{
|
|
{
|
|
endpoint: api.query,
|
|
query: url.Values{
|
|
"query": []string{"2"},
|
|
"time": []string{"123.4"},
|
|
},
|
|
response: &QueryData{
|
|
ResultType: parser.ValueTypeScalar,
|
|
Result: promql.Scalar{
|
|
V: 2,
|
|
T: timestamp.FromTime(start.Add(123*time.Second + 400*time.Millisecond)),
|
|
},
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.query,
|
|
query: url.Values{
|
|
"query": []string{"0.333"},
|
|
"time": []string{"1970-01-01T00:02:03Z"},
|
|
},
|
|
response: &QueryData{
|
|
ResultType: parser.ValueTypeScalar,
|
|
Result: promql.Scalar{
|
|
V: 0.333,
|
|
T: timestamp.FromTime(start.Add(123 * time.Second)),
|
|
},
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.query,
|
|
query: url.Values{
|
|
"query": []string{"0.333"},
|
|
"time": []string{"1970-01-01T01:02:03+01:00"},
|
|
},
|
|
response: &QueryData{
|
|
ResultType: parser.ValueTypeScalar,
|
|
Result: promql.Scalar{
|
|
V: 0.333,
|
|
T: timestamp.FromTime(start.Add(123 * time.Second)),
|
|
},
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.query,
|
|
query: url.Values{
|
|
"query": []string{"0.333"},
|
|
},
|
|
response: &QueryData{
|
|
ResultType: parser.ValueTypeScalar,
|
|
Result: promql.Scalar{
|
|
V: 0.333,
|
|
T: timestamp.FromTime(api.now()),
|
|
},
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.queryRange,
|
|
query: url.Values{
|
|
"query": []string{"time()"},
|
|
"start": []string{"0"},
|
|
"end": []string{"2"},
|
|
"step": []string{"1"},
|
|
},
|
|
response: &QueryData{
|
|
ResultType: parser.ValueTypeMatrix,
|
|
Result: promql.Matrix{
|
|
promql.Series{
|
|
Floats: []promql.FPoint{
|
|
{F: 0, T: timestamp.FromTime(start)},
|
|
{F: 1, T: timestamp.FromTime(start.Add(1 * time.Second))},
|
|
{F: 2, T: timestamp.FromTime(start.Add(2 * time.Second))},
|
|
},
|
|
// No Metric returned - use zero value for comparison.
|
|
},
|
|
},
|
|
},
|
|
},
|
|
// Test empty vector result
|
|
{
|
|
endpoint: api.query,
|
|
query: url.Values{
|
|
"query": []string{"bottomk(2, notExists)"},
|
|
},
|
|
responseAsJSON: `{"resultType":"vector","result":[]}`,
|
|
},
|
|
// Test empty matrix result
|
|
{
|
|
endpoint: api.queryRange,
|
|
query: url.Values{
|
|
"query": []string{"bottomk(2, notExists)"},
|
|
"start": []string{"0"},
|
|
"end": []string{"2"},
|
|
"step": []string{"1"},
|
|
},
|
|
responseAsJSON: `{"resultType":"matrix","result":[]}`,
|
|
},
|
|
// Missing query params in range queries.
|
|
{
|
|
endpoint: api.queryRange,
|
|
query: url.Values{
|
|
"query": []string{"time()"},
|
|
"end": []string{"2"},
|
|
"step": []string{"1"},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
{
|
|
endpoint: api.queryRange,
|
|
query: url.Values{
|
|
"query": []string{"time()"},
|
|
"start": []string{"0"},
|
|
"step": []string{"1"},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
{
|
|
endpoint: api.queryRange,
|
|
query: url.Values{
|
|
"query": []string{"time()"},
|
|
"start": []string{"0"},
|
|
"end": []string{"2"},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
// Bad query expression.
|
|
{
|
|
endpoint: api.query,
|
|
query: url.Values{
|
|
"query": []string{"invalid][query"},
|
|
"time": []string{"1970-01-01T01:02:03+01:00"},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
{
|
|
endpoint: api.queryRange,
|
|
query: url.Values{
|
|
"query": []string{"invalid][query"},
|
|
"start": []string{"0"},
|
|
"end": []string{"100"},
|
|
"step": []string{"1"},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
// Invalid step.
|
|
{
|
|
endpoint: api.queryRange,
|
|
query: url.Values{
|
|
"query": []string{"time()"},
|
|
"start": []string{"1"},
|
|
"end": []string{"2"},
|
|
"step": []string{"0"},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
// Start after end.
|
|
{
|
|
endpoint: api.queryRange,
|
|
query: url.Values{
|
|
"query": []string{"time()"},
|
|
"start": []string{"2"},
|
|
"end": []string{"1"},
|
|
"step": []string{"1"},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
// Start overflows int64 internally.
|
|
{
|
|
endpoint: api.queryRange,
|
|
query: url.Values{
|
|
"query": []string{"time()"},
|
|
"start": []string{"148966367200.372"},
|
|
"end": []string{"1489667272.372"},
|
|
"step": []string{"1"},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
{
|
|
endpoint: api.formatQuery,
|
|
query: url.Values{
|
|
"query": []string{"foo+bar"},
|
|
},
|
|
response: "foo + bar",
|
|
},
|
|
{
|
|
endpoint: api.formatQuery,
|
|
query: url.Values{
|
|
"query": []string{"invalid_expression/"},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
{
|
|
endpoint: api.series,
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric2`},
|
|
},
|
|
response: []labels.Labels{
|
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"),
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.series,
|
|
query: url.Values{
|
|
"match[]": []string{`{foo=""}`},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
{
|
|
endpoint: api.series,
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric1{foo=~".+o"}`},
|
|
},
|
|
response: []labels.Labels{
|
|
labels.FromStrings("__name__", "test_metric1", "foo", "boo"),
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.series,
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric1{foo=~".+o$"}`, `test_metric1{foo=~".+o"}`},
|
|
},
|
|
response: []labels.Labels{
|
|
labels.FromStrings("__name__", "test_metric1", "foo", "boo"),
|
|
},
|
|
},
|
|
// Try to overlap the selected series set as much as possible to test the result de-duplication works well.
|
|
{
|
|
endpoint: api.series,
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric4{foo=~".+o$"}`, `test_metric4{dup=~"^1"}`},
|
|
},
|
|
response: []labels.Labels{
|
|
labels.FromStrings("__name__", "test_metric4", "dup", "1", "foo", "bar"),
|
|
labels.FromStrings("__name__", "test_metric4", "dup", "1", "foo", "boo"),
|
|
labels.FromStrings("__name__", "test_metric4", "foo", "boo"),
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.series,
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric1{foo=~".+o"}`, `none`},
|
|
},
|
|
response: []labels.Labels{
|
|
labels.FromStrings("__name__", "test_metric1", "foo", "boo"),
|
|
},
|
|
},
|
|
// Start and end before series starts.
|
|
{
|
|
endpoint: api.series,
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric2`},
|
|
"start": []string{"-2"},
|
|
"end": []string{"-1"},
|
|
},
|
|
response: []labels.Labels{},
|
|
},
|
|
// Start and end after series ends.
|
|
{
|
|
endpoint: api.series,
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric2`},
|
|
"start": []string{"100000"},
|
|
"end": []string{"100001"},
|
|
},
|
|
response: []labels.Labels{},
|
|
},
|
|
// Start before series starts, end after series ends.
|
|
{
|
|
endpoint: api.series,
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric2`},
|
|
"start": []string{"-1"},
|
|
"end": []string{"100000"},
|
|
},
|
|
response: []labels.Labels{
|
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"),
|
|
},
|
|
},
|
|
// Start and end within series.
|
|
{
|
|
endpoint: api.series,
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric2`},
|
|
"start": []string{"1"},
|
|
"end": []string{"100"},
|
|
},
|
|
response: []labels.Labels{
|
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"),
|
|
},
|
|
},
|
|
// Start within series, end after.
|
|
{
|
|
endpoint: api.series,
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric2`},
|
|
"start": []string{"1"},
|
|
"end": []string{"100000"},
|
|
},
|
|
response: []labels.Labels{
|
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"),
|
|
},
|
|
},
|
|
// Start before series, end within series.
|
|
{
|
|
endpoint: api.series,
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric2`},
|
|
"start": []string{"-1"},
|
|
"end": []string{"1"},
|
|
},
|
|
response: []labels.Labels{
|
|
labels.FromStrings("__name__", "test_metric2", "foo", "boo"),
|
|
},
|
|
},
|
|
// Series request with limit.
|
|
{
|
|
endpoint: api.series,
|
|
query: url.Values{
|
|
"match[]": []string{"test_metric1"},
|
|
"limit": []string{"1"},
|
|
},
|
|
responseLen: 1, // API does not specify which particular value will come back.
|
|
warningsCount: 1,
|
|
},
|
|
{
|
|
endpoint: api.series,
|
|
query: url.Values{
|
|
"match[]": []string{"test_metric1"},
|
|
"limit": []string{"2"},
|
|
},
|
|
responseLen: 2, // API does not specify which particular value will come back.
|
|
warningsCount: 0, // No warnings if limit isn't exceeded.
|
|
},
|
|
{
|
|
endpoint: api.series,
|
|
query: url.Values{
|
|
"match[]": []string{"test_metric1"},
|
|
"limit": []string{"0"},
|
|
},
|
|
responseLen: 2, // API does not specify which particular value will come back.
|
|
warningsCount: 0, // No warnings if limit isn't exceeded.
|
|
},
|
|
// Missing match[] query params in series requests.
|
|
{
|
|
endpoint: api.series,
|
|
errType: errorBadData,
|
|
},
|
|
{
|
|
endpoint: api.dropSeries,
|
|
errType: errorInternal,
|
|
},
|
|
{
|
|
endpoint: api.targets,
|
|
response: &TargetDiscovery{
|
|
ActiveTargets: []*Target{
|
|
{
|
|
DiscoveredLabels: labels.FromStrings(),
|
|
Labels: labels.FromStrings("job", "blackbox"),
|
|
ScrapePool: "blackbox",
|
|
ScrapeURL: "http://localhost:9115/probe?target=example.com",
|
|
GlobalURL: "http://localhost:9115/probe?target=example.com",
|
|
Health: "down",
|
|
LastError: "failed: missing port in address",
|
|
LastScrape: scrapeStart,
|
|
LastScrapeDuration: 0.1,
|
|
ScrapeInterval: "20s",
|
|
ScrapeTimeout: "10s",
|
|
},
|
|
{
|
|
DiscoveredLabels: labels.FromStrings(),
|
|
Labels: labels.FromStrings("job", "test"),
|
|
ScrapePool: "test",
|
|
ScrapeURL: "http://example.com:8080/metrics",
|
|
GlobalURL: "http://example.com:8080/metrics",
|
|
Health: "up",
|
|
LastError: "",
|
|
LastScrape: scrapeStart,
|
|
LastScrapeDuration: 0.07,
|
|
ScrapeInterval: "15s",
|
|
ScrapeTimeout: "5s",
|
|
},
|
|
},
|
|
DroppedTargets: []*DroppedTarget{
|
|
{
|
|
DiscoveredLabels: labels.FromStrings(
|
|
"__address__", "http://dropped.example.com:9115",
|
|
"__metrics_path__", "/probe",
|
|
"__scheme__", "http",
|
|
"job", "blackbox",
|
|
"__scrape_interval__", "30s",
|
|
"__scrape_timeout__", "15s",
|
|
),
|
|
},
|
|
},
|
|
DroppedTargetCounts: map[string]int{"blackbox": 1},
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.targets,
|
|
query: url.Values{
|
|
"state": []string{"any"},
|
|
},
|
|
response: &TargetDiscovery{
|
|
ActiveTargets: []*Target{
|
|
{
|
|
DiscoveredLabels: labels.FromStrings(),
|
|
Labels: labels.FromStrings("job", "blackbox"),
|
|
ScrapePool: "blackbox",
|
|
ScrapeURL: "http://localhost:9115/probe?target=example.com",
|
|
GlobalURL: "http://localhost:9115/probe?target=example.com",
|
|
Health: "down",
|
|
LastError: "failed: missing port in address",
|
|
LastScrape: scrapeStart,
|
|
LastScrapeDuration: 0.1,
|
|
ScrapeInterval: "20s",
|
|
ScrapeTimeout: "10s",
|
|
},
|
|
{
|
|
DiscoveredLabels: labels.FromStrings(),
|
|
Labels: labels.FromStrings("job", "test"),
|
|
ScrapePool: "test",
|
|
ScrapeURL: "http://example.com:8080/metrics",
|
|
GlobalURL: "http://example.com:8080/metrics",
|
|
Health: "up",
|
|
LastError: "",
|
|
LastScrape: scrapeStart,
|
|
LastScrapeDuration: 0.07,
|
|
ScrapeInterval: "15s",
|
|
ScrapeTimeout: "5s",
|
|
},
|
|
},
|
|
DroppedTargets: []*DroppedTarget{
|
|
{
|
|
DiscoveredLabels: labels.FromStrings(
|
|
"__address__", "http://dropped.example.com:9115",
|
|
"__metrics_path__", "/probe",
|
|
"__scheme__", "http",
|
|
"job", "blackbox",
|
|
"__scrape_interval__", "30s",
|
|
"__scrape_timeout__", "15s",
|
|
),
|
|
},
|
|
},
|
|
DroppedTargetCounts: map[string]int{"blackbox": 1},
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.targets,
|
|
query: url.Values{
|
|
"state": []string{"active"},
|
|
},
|
|
response: &TargetDiscovery{
|
|
ActiveTargets: []*Target{
|
|
{
|
|
DiscoveredLabels: labels.FromStrings(),
|
|
Labels: labels.FromStrings("job", "blackbox"),
|
|
ScrapePool: "blackbox",
|
|
ScrapeURL: "http://localhost:9115/probe?target=example.com",
|
|
GlobalURL: "http://localhost:9115/probe?target=example.com",
|
|
Health: "down",
|
|
LastError: "failed: missing port in address",
|
|
LastScrape: scrapeStart,
|
|
LastScrapeDuration: 0.1,
|
|
ScrapeInterval: "20s",
|
|
ScrapeTimeout: "10s",
|
|
},
|
|
{
|
|
DiscoveredLabels: labels.FromStrings(),
|
|
Labels: labels.FromStrings("job", "test"),
|
|
ScrapePool: "test",
|
|
ScrapeURL: "http://example.com:8080/metrics",
|
|
GlobalURL: "http://example.com:8080/metrics",
|
|
Health: "up",
|
|
LastError: "",
|
|
LastScrape: scrapeStart,
|
|
LastScrapeDuration: 0.07,
|
|
ScrapeInterval: "15s",
|
|
ScrapeTimeout: "5s",
|
|
},
|
|
},
|
|
DroppedTargets: []*DroppedTarget{},
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.targets,
|
|
query: url.Values{
|
|
"state": []string{"Dropped"},
|
|
},
|
|
response: &TargetDiscovery{
|
|
ActiveTargets: []*Target{},
|
|
DroppedTargets: []*DroppedTarget{
|
|
{
|
|
DiscoveredLabels: labels.FromStrings(
|
|
"__address__", "http://dropped.example.com:9115",
|
|
"__metrics_path__", "/probe",
|
|
"__scheme__", "http",
|
|
"job", "blackbox",
|
|
"__scrape_interval__", "30s",
|
|
"__scrape_timeout__", "15s",
|
|
),
|
|
},
|
|
},
|
|
DroppedTargetCounts: map[string]int{"blackbox": 1},
|
|
},
|
|
},
|
|
// With a matching metric.
|
|
{
|
|
endpoint: api.targetMetadata,
|
|
query: url.Values{
|
|
"metric": []string{"go_threads"},
|
|
},
|
|
metadata: []targetMetadata{
|
|
{
|
|
identifier: "test",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Number of OS threads created.",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
response: []metricMetadata{
|
|
{
|
|
Target: labels.FromMap(map[string]string{
|
|
"job": "test",
|
|
}),
|
|
Help: "Number of OS threads created.",
|
|
Type: model.MetricTypeGauge,
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
// With a matching target.
|
|
{
|
|
endpoint: api.targetMetadata,
|
|
query: url.Values{
|
|
"match_target": []string{"{job=\"blackbox\"}"},
|
|
},
|
|
metadata: []targetMetadata{
|
|
{
|
|
identifier: "blackbox",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "prometheus_tsdb_storage_blocks_bytes",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "The number of bytes that are currently used for local storage by all blocks.",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
response: []metricMetadata{
|
|
{
|
|
Target: labels.FromMap(map[string]string{
|
|
"job": "blackbox",
|
|
}),
|
|
Metric: "prometheus_tsdb_storage_blocks_bytes",
|
|
Help: "The number of bytes that are currently used for local storage by all blocks.",
|
|
Type: model.MetricTypeGauge,
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
// Without a target or metric.
|
|
{
|
|
endpoint: api.targetMetadata,
|
|
metadata: []targetMetadata{
|
|
{
|
|
identifier: "test",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Number of OS threads created.",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
{
|
|
identifier: "blackbox",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "prometheus_tsdb_storage_blocks_bytes",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "The number of bytes that are currently used for local storage by all blocks.",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
response: []metricMetadata{
|
|
{
|
|
Target: labels.FromMap(map[string]string{
|
|
"job": "test",
|
|
}),
|
|
Metric: "go_threads",
|
|
Help: "Number of OS threads created.",
|
|
Type: model.MetricTypeGauge,
|
|
Unit: "",
|
|
},
|
|
{
|
|
Target: labels.FromMap(map[string]string{
|
|
"job": "blackbox",
|
|
}),
|
|
Metric: "prometheus_tsdb_storage_blocks_bytes",
|
|
Help: "The number of bytes that are currently used for local storage by all blocks.",
|
|
Type: model.MetricTypeGauge,
|
|
Unit: "",
|
|
},
|
|
},
|
|
sorter: func(m interface{}) {
|
|
sort.Slice(m.([]metricMetadata), func(i, j int) bool {
|
|
s := m.([]metricMetadata)
|
|
return s[i].Metric < s[j].Metric
|
|
})
|
|
},
|
|
},
|
|
// Without a matching metric.
|
|
{
|
|
endpoint: api.targetMetadata,
|
|
query: url.Values{
|
|
"match_target": []string{"{job=\"non-existentblackbox\"}"},
|
|
},
|
|
response: []metricMetadata{},
|
|
},
|
|
{
|
|
endpoint: api.alertmanagers,
|
|
response: &AlertmanagerDiscovery{
|
|
ActiveAlertmanagers: []*AlertmanagerTarget{
|
|
{
|
|
URL: "http://alertmanager.example.com:8080/api/v1/alerts",
|
|
},
|
|
},
|
|
DroppedAlertmanagers: []*AlertmanagerTarget{
|
|
{
|
|
URL: "http://dropped.alertmanager.example.com:8080/api/v1/alerts",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
// With metadata available.
|
|
{
|
|
endpoint: api.metricMetadata,
|
|
metadata: []targetMetadata{
|
|
{
|
|
identifier: "test",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "prometheus_engine_query_duration_seconds",
|
|
Type: model.MetricTypeSummary,
|
|
Help: "Query timings",
|
|
Unit: "",
|
|
},
|
|
{
|
|
Metric: "go_info",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Information about the Go environment.",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
response: map[string][]metadata.Metadata{
|
|
"prometheus_engine_query_duration_seconds": {{Type: model.MetricTypeSummary, Help: "Query timings", Unit: ""}},
|
|
"go_info": {{Type: model.MetricTypeGauge, Help: "Information about the Go environment.", Unit: ""}},
|
|
},
|
|
responseAsJSON: `{"prometheus_engine_query_duration_seconds":[{"type":"summary","unit":"",
|
|
"help":"Query timings"}], "go_info":[{"type":"gauge","unit":"",
|
|
"help":"Information about the Go environment."}]}`,
|
|
},
|
|
// With duplicate metadata for a metric that comes from different targets.
|
|
{
|
|
endpoint: api.metricMetadata,
|
|
metadata: []targetMetadata{
|
|
{
|
|
identifier: "test",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Number of OS threads created",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
{
|
|
identifier: "blackbox",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Number of OS threads created",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
response: map[string][]metadata.Metadata{
|
|
"go_threads": {{Type: model.MetricTypeGauge, Help: "Number of OS threads created"}},
|
|
},
|
|
responseAsJSON: `{"go_threads": [{"type":"gauge","unit":"",
|
|
"help":"Number of OS threads created"}]}`,
|
|
},
|
|
// With non-duplicate metadata for the same metric from different targets.
|
|
{
|
|
endpoint: api.metricMetadata,
|
|
metadata: []targetMetadata{
|
|
{
|
|
identifier: "test",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Number of OS threads created",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
{
|
|
identifier: "blackbox",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Number of OS threads that were created.",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
response: map[string][]metadata.Metadata{
|
|
"go_threads": {
|
|
{Type: model.MetricTypeGauge, Help: "Number of OS threads created"},
|
|
{Type: model.MetricTypeGauge, Help: "Number of OS threads that were created."},
|
|
},
|
|
},
|
|
responseAsJSON: `{"go_threads": [{"type":"gauge","unit":"",
|
|
"help":"Number of OS threads created"},{"type":"gauge","unit":"",
|
|
"help":"Number of OS threads that were created."}]}`,
|
|
sorter: func(m interface{}) {
|
|
v := m.(map[string][]metadata.Metadata)["go_threads"]
|
|
|
|
sort.Slice(v, func(i, j int) bool {
|
|
return v[i].Help < v[j].Help
|
|
})
|
|
},
|
|
},
|
|
// With a limit for the number of metrics returned.
|
|
{
|
|
endpoint: api.metricMetadata,
|
|
query: url.Values{
|
|
"limit": []string{"2"},
|
|
},
|
|
metadata: []targetMetadata{
|
|
{
|
|
identifier: "test",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Number of OS threads created",
|
|
Unit: "",
|
|
},
|
|
{
|
|
Metric: "prometheus_engine_query_duration_seconds",
|
|
Type: model.MetricTypeSummary,
|
|
Help: "Query Timings.",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
{
|
|
identifier: "blackbox",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "go_gc_duration_seconds",
|
|
Type: model.MetricTypeSummary,
|
|
Help: "A summary of the GC invocation durations.",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
responseLen: 2,
|
|
},
|
|
// With a limit for the number of metadata per metric.
|
|
{
|
|
endpoint: api.metricMetadata,
|
|
query: url.Values{"limit_per_metric": []string{"1"}},
|
|
metadata: []targetMetadata{
|
|
{
|
|
identifier: "test",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Number of OS threads created",
|
|
Unit: "",
|
|
},
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Repeated metadata",
|
|
Unit: "",
|
|
},
|
|
{
|
|
Metric: "go_gc_duration_seconds",
|
|
Type: model.MetricTypeSummary,
|
|
Help: "A summary of the GC invocation durations.",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
response: map[string][]metadata.Metadata{
|
|
"go_threads": {
|
|
{Type: model.MetricTypeGauge, Help: "Number of OS threads created"},
|
|
},
|
|
"go_gc_duration_seconds": {
|
|
{Type: model.MetricTypeSummary, Help: "A summary of the GC invocation durations."},
|
|
},
|
|
},
|
|
responseAsJSON: `{"go_gc_duration_seconds":[{"help":"A summary of the GC invocation durations.","type":"summary","unit":""}],"go_threads": [{"type":"gauge","unit":"","help":"Number of OS threads created"}]}`,
|
|
},
|
|
// With a limit for the number of metadata per metric and per metric.
|
|
{
|
|
endpoint: api.metricMetadata,
|
|
query: url.Values{"limit_per_metric": []string{"1"}, "limit": []string{"1"}},
|
|
metadata: []targetMetadata{
|
|
{
|
|
identifier: "test",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Number of OS threads created",
|
|
Unit: "",
|
|
},
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Repeated metadata",
|
|
Unit: "",
|
|
},
|
|
{
|
|
Metric: "go_gc_duration_seconds",
|
|
Type: model.MetricTypeSummary,
|
|
Help: "A summary of the GC invocation durations.",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
responseLen: 1,
|
|
responseMetadataTotal: 1,
|
|
},
|
|
|
|
// With a limit for the number of metadata per metric and per metric, while having multiple targets.
|
|
{
|
|
endpoint: api.metricMetadata,
|
|
query: url.Values{"limit_per_metric": []string{"1"}, "limit": []string{"1"}},
|
|
metadata: []targetMetadata{
|
|
{
|
|
identifier: "test",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Number of OS threads created",
|
|
Unit: "",
|
|
},
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Repeated metadata",
|
|
Unit: "",
|
|
},
|
|
{
|
|
Metric: "go_gc_duration_seconds",
|
|
Type: model.MetricTypeSummary,
|
|
Help: "A summary of the GC invocation durations.",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
{
|
|
identifier: "secondTarget",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Number of OS threads created, but from a different target",
|
|
Unit: "",
|
|
},
|
|
{
|
|
Metric: "go_gc_duration_seconds",
|
|
Type: model.MetricTypeSummary,
|
|
Help: "A summary of the GC invocation durations, but from a different target.",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
responseLen: 1,
|
|
responseMetadataTotal: 1,
|
|
},
|
|
// When requesting a specific metric that is present.
|
|
{
|
|
endpoint: api.metricMetadata,
|
|
query: url.Values{"metric": []string{"go_threads"}},
|
|
metadata: []targetMetadata{
|
|
{
|
|
identifier: "test",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Number of OS threads created",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
{
|
|
identifier: "blackbox",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "go_gc_duration_seconds",
|
|
Type: model.MetricTypeSummary,
|
|
Help: "A summary of the GC invocation durations.",
|
|
Unit: "",
|
|
},
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Number of OS threads that were created.",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
response: map[string][]metadata.Metadata{
|
|
"go_threads": {
|
|
{Type: model.MetricTypeGauge, Help: "Number of OS threads created"},
|
|
{Type: model.MetricTypeGauge, Help: "Number of OS threads that were created."},
|
|
},
|
|
},
|
|
responseAsJSON: `{"go_threads": [{"type":"gauge","unit":"","help":"Number of OS threads created"},{"type":"gauge","unit":"","help":"Number of OS threads that were created."}]}`,
|
|
sorter: func(m interface{}) {
|
|
v := m.(map[string][]metadata.Metadata)["go_threads"]
|
|
|
|
sort.Slice(v, func(i, j int) bool {
|
|
return v[i].Help < v[j].Help
|
|
})
|
|
},
|
|
},
|
|
// With a specific metric that is not present.
|
|
{
|
|
endpoint: api.metricMetadata,
|
|
query: url.Values{"metric": []string{"go_gc_duration_seconds"}},
|
|
metadata: []targetMetadata{
|
|
{
|
|
identifier: "test",
|
|
metadata: []scrape.MetricMetadata{
|
|
{
|
|
Metric: "go_threads",
|
|
Type: model.MetricTypeGauge,
|
|
Help: "Number of OS threads created",
|
|
Unit: "",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
response: map[string][]metadata.Metadata{},
|
|
},
|
|
// With no available metadata.
|
|
{
|
|
endpoint: api.metricMetadata,
|
|
response: map[string][]metadata.Metadata{},
|
|
},
|
|
{
|
|
endpoint: api.serveConfig,
|
|
response: &prometheusConfig{
|
|
YAML: samplePrometheusCfg.String(),
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.serveFlags,
|
|
response: sampleFlagMap,
|
|
},
|
|
{
|
|
endpoint: api.alerts,
|
|
response: &AlertDiscovery{
|
|
Alerts: []*Alert{
|
|
{
|
|
Labels: labels.FromStrings("alertname", "test_metric5", "name", "tm5"),
|
|
Annotations: labels.Labels{},
|
|
State: "pending",
|
|
Value: "1e+00",
|
|
},
|
|
},
|
|
},
|
|
zeroFunc: func(i interface{}) {
|
|
if i != nil {
|
|
v := i.(*AlertDiscovery)
|
|
for _, alert := range v.Alerts {
|
|
alert.ActiveAt = nil
|
|
}
|
|
}
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.rules,
|
|
response: &RuleDiscovery{
|
|
RuleGroups: []*RuleGroup{
|
|
{
|
|
Name: "grp",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric3",
|
|
Query: "absent(test_metric3) != 1",
|
|
Duration: 1,
|
|
Labels: labels.Labels{},
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric4",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.Labels{},
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "pending",
|
|
Name: "test_metric5",
|
|
Query: "vector(1)",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("name", "tm5"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{
|
|
{
|
|
Labels: labels.FromStrings("alertname", "test_metric5", "name", "tm5"),
|
|
Annotations: labels.Labels{},
|
|
State: "pending",
|
|
Value: "1e+00",
|
|
},
|
|
},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric6",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("testlabel", "rule"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric7",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("templatedlabel", "{{ $externalURL }}"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
RecordingRule{
|
|
Name: "recording-rule-1",
|
|
Query: "vector(1)",
|
|
Labels: labels.Labels{},
|
|
Health: "ok",
|
|
Type: "recording",
|
|
},
|
|
RecordingRule{
|
|
Name: "recording-rule-2",
|
|
Query: "vector(1)",
|
|
Labels: labels.FromStrings("testlabel", "rule"),
|
|
Health: "ok",
|
|
Type: "recording",
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Name: "grp2",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric3",
|
|
Query: "absent(test_metric3) != 1",
|
|
Duration: 1,
|
|
Labels: labels.Labels{},
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{
|
|
endpoint: api.rules,
|
|
query: url.Values{
|
|
"exclude_alerts": []string{"true"},
|
|
},
|
|
response: &RuleDiscovery{
|
|
RuleGroups: []*RuleGroup{
|
|
{
|
|
Name: "grp",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric3",
|
|
Query: "absent(test_metric3) != 1",
|
|
Duration: 1,
|
|
Labels: labels.Labels{},
|
|
Annotations: labels.Labels{},
|
|
Alerts: nil,
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric4",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.Labels{},
|
|
Annotations: labels.Labels{},
|
|
Alerts: nil,
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "pending",
|
|
Name: "test_metric5",
|
|
Query: "vector(1)",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("name", "tm5"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: nil,
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric6",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("testlabel", "rule"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: nil,
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric7",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("templatedlabel", "{{ $externalURL }}"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: nil,
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
RecordingRule{
|
|
Name: "recording-rule-1",
|
|
Query: "vector(1)",
|
|
Labels: labels.Labels{},
|
|
Health: "ok",
|
|
Type: "recording",
|
|
},
|
|
RecordingRule{
|
|
Name: "recording-rule-2",
|
|
Query: "vector(1)",
|
|
Labels: labels.FromStrings("testlabel", "rule"),
|
|
Health: "ok",
|
|
Type: "recording",
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Name: "grp2",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric3",
|
|
Query: "absent(test_metric3) != 1",
|
|
Duration: 1,
|
|
Labels: labels.Labels{},
|
|
Annotations: labels.Labels{},
|
|
Alerts: nil,
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{
|
|
endpoint: api.rules,
|
|
query: url.Values{
|
|
"type": []string{"alert"},
|
|
},
|
|
response: &RuleDiscovery{
|
|
RuleGroups: []*RuleGroup{
|
|
{
|
|
Name: "grp",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric3",
|
|
Query: "absent(test_metric3) != 1",
|
|
Duration: 1,
|
|
Labels: labels.Labels{},
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric4",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.Labels{},
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "pending",
|
|
Name: "test_metric5",
|
|
Query: "vector(1)",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("name", "tm5"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{
|
|
{
|
|
Labels: labels.FromStrings("alertname", "test_metric5", "name", "tm5"),
|
|
Annotations: labels.Labels{},
|
|
State: "pending",
|
|
Value: "1e+00",
|
|
},
|
|
},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric6",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("testlabel", "rule"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric7",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("templatedlabel", "{{ $externalURL }}"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
},
|
|
},
|
|
{
|
|
Name: "grp2",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric3",
|
|
Query: "absent(test_metric3) != 1",
|
|
Duration: 1,
|
|
Labels: labels.Labels{},
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{
|
|
endpoint: api.rules,
|
|
query: url.Values{
|
|
"type": []string{"record"},
|
|
},
|
|
response: &RuleDiscovery{
|
|
RuleGroups: []*RuleGroup{
|
|
{
|
|
Name: "grp",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
RecordingRule{
|
|
Name: "recording-rule-1",
|
|
Query: "vector(1)",
|
|
Labels: labels.Labels{},
|
|
Health: "ok",
|
|
Type: "recording",
|
|
},
|
|
RecordingRule{
|
|
Name: "recording-rule-2",
|
|
Query: "vector(1)",
|
|
Labels: labels.FromStrings("testlabel", "rule"),
|
|
Health: "ok",
|
|
Type: "recording",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{
|
|
endpoint: api.rules,
|
|
query: url.Values{"rule_name[]": []string{"test_metric4"}},
|
|
response: &RuleDiscovery{
|
|
RuleGroups: []*RuleGroup{
|
|
{
|
|
Name: "grp",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric4",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.Labels{},
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{
|
|
endpoint: api.rules,
|
|
query: url.Values{"rule_group[]": []string{"respond-with-nothing"}},
|
|
response: &RuleDiscovery{RuleGroups: []*RuleGroup{}},
|
|
},
|
|
{
|
|
endpoint: api.rules,
|
|
query: url.Values{"file[]": []string{"/path/to/file"}, "rule_name[]": []string{"test_metric4"}},
|
|
response: &RuleDiscovery{
|
|
RuleGroups: []*RuleGroup{
|
|
{
|
|
Name: "grp",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric4",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.Labels{},
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{
|
|
endpoint: api.rules,
|
|
query: url.Values{
|
|
"match[]": []string{`{testlabel="rule"}`},
|
|
},
|
|
response: &RuleDiscovery{
|
|
RuleGroups: []*RuleGroup{
|
|
{
|
|
Name: "grp",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric6",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("testlabel", "rule"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
RecordingRule{
|
|
Name: "recording-rule-2",
|
|
Query: "vector(1)",
|
|
Labels: labels.FromStrings("testlabel", "rule"),
|
|
Health: "ok",
|
|
Type: "recording",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{
|
|
endpoint: api.rules,
|
|
query: url.Values{
|
|
"type": []string{"alert"},
|
|
"match[]": []string{`{templatedlabel="{{ $externalURL }}"}`},
|
|
},
|
|
response: &RuleDiscovery{
|
|
RuleGroups: []*RuleGroup{
|
|
{
|
|
Name: "grp",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric7",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("templatedlabel", "{{ $externalURL }}"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{
|
|
endpoint: api.rules,
|
|
query: url.Values{
|
|
"match[]": []string{`{testlabel="abc"}`},
|
|
},
|
|
response: &RuleDiscovery{
|
|
RuleGroups: []*RuleGroup{},
|
|
},
|
|
},
|
|
// This is testing OR condition, the api response should return rule if it matches one of the label selector
|
|
{
|
|
endpoint: api.rules,
|
|
query: url.Values{
|
|
"match[]": []string{`{testlabel="abc"}`, `{testlabel="rule"}`},
|
|
},
|
|
response: &RuleDiscovery{
|
|
RuleGroups: []*RuleGroup{
|
|
{
|
|
Name: "grp",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric6",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("testlabel", "rule"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
RecordingRule{
|
|
Name: "recording-rule-2",
|
|
Query: "vector(1)",
|
|
Labels: labels.FromStrings("testlabel", "rule"),
|
|
Health: "ok",
|
|
Type: "recording",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{
|
|
endpoint: api.rules,
|
|
query: url.Values{
|
|
"type": []string{"record"},
|
|
"match[]": []string{`{testlabel="rule"}`},
|
|
},
|
|
response: &RuleDiscovery{
|
|
RuleGroups: []*RuleGroup{
|
|
{
|
|
Name: "grp",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
RecordingRule{
|
|
Name: "recording-rule-2",
|
|
Query: "vector(1)",
|
|
Labels: labels.FromStrings("testlabel", "rule"),
|
|
Health: "ok",
|
|
Type: "recording",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{
|
|
endpoint: api.rules,
|
|
query: url.Values{
|
|
"type": []string{"alert"},
|
|
"match[]": []string{`{testlabel="rule"}`},
|
|
},
|
|
response: &RuleDiscovery{
|
|
RuleGroups: []*RuleGroup{
|
|
{
|
|
Name: "grp",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric6",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("testlabel", "rule"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{
|
|
endpoint: api.rules,
|
|
query: url.Values{
|
|
"group_limit": []string{"1"},
|
|
},
|
|
response: &RuleDiscovery{
|
|
GroupNextToken: getRuleGroupNextToken("/path/to/file", "grp2"),
|
|
RuleGroups: []*RuleGroup{
|
|
{
|
|
Name: "grp",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric3",
|
|
Query: "absent(test_metric3) != 1",
|
|
Duration: 1,
|
|
Labels: labels.Labels{},
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric4",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.Labels{},
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "pending",
|
|
Name: "test_metric5",
|
|
Query: "vector(1)",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("name", "tm5"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{
|
|
{
|
|
Labels: labels.FromStrings("alertname", "test_metric5", "name", "tm5"),
|
|
Annotations: labels.Labels{},
|
|
State: "pending",
|
|
Value: "1e+00",
|
|
},
|
|
},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric6",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("testlabel", "rule"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric7",
|
|
Query: "up == 1",
|
|
Duration: 1,
|
|
Labels: labels.FromStrings("templatedlabel", "{{ $externalURL }}"),
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
RecordingRule{
|
|
Name: "recording-rule-1",
|
|
Query: "vector(1)",
|
|
Labels: labels.Labels{},
|
|
Health: "ok",
|
|
Type: "recording",
|
|
},
|
|
RecordingRule{
|
|
Name: "recording-rule-2",
|
|
Query: "vector(1)",
|
|
Labels: labels.FromStrings("testlabel", "rule"),
|
|
Health: "ok",
|
|
Type: "recording",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{
|
|
endpoint: api.rules,
|
|
query: url.Values{
|
|
"group_limit": []string{"1"},
|
|
"group_next_token": []string{getRuleGroupNextToken("/path/to/file", "grp2")},
|
|
},
|
|
response: &RuleDiscovery{
|
|
RuleGroups: []*RuleGroup{
|
|
{
|
|
Name: "grp2",
|
|
File: "/path/to/file",
|
|
Interval: 1,
|
|
Limit: 0,
|
|
Rules: []Rule{
|
|
AlertingRule{
|
|
State: "inactive",
|
|
Name: "test_metric3",
|
|
Query: "absent(test_metric3) != 1",
|
|
Duration: 1,
|
|
Labels: labels.Labels{},
|
|
Annotations: labels.Labels{},
|
|
Alerts: []*Alert{},
|
|
Health: "ok",
|
|
Type: "alerting",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{ // invalid pagination request
|
|
endpoint: api.rules,
|
|
query: url.Values{
|
|
"group_next_token": []string{getRuleGroupNextToken("/path/to/file", "grp2")},
|
|
},
|
|
errType: errorBadData,
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{ // invalid group_limit
|
|
endpoint: api.rules,
|
|
query: url.Values{
|
|
"group_limit": []string{"0"},
|
|
"group_next_token": []string{getRuleGroupNextToken("/path/to/file", "grp2")},
|
|
},
|
|
errType: errorBadData,
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{ // Pagination token is invalid due to changes in the rule groups
|
|
endpoint: api.rules,
|
|
query: url.Values{
|
|
"group_limit": []string{"1"},
|
|
"group_next_token": []string{getRuleGroupNextToken("/removed/file", "notfound")},
|
|
},
|
|
errType: errorBadData,
|
|
zeroFunc: rulesZeroFunc,
|
|
},
|
|
{
|
|
endpoint: api.queryExemplars,
|
|
query: url.Values{
|
|
"query": []string{`test_metric3{foo="boo"} - test_metric4{foo="bar"}`},
|
|
"start": []string{"0"},
|
|
"end": []string{"4"},
|
|
},
|
|
// Note extra integer length of timestamps for exemplars because of millisecond preservation
|
|
// of timestamps within Prometheus (see timestamp package).
|
|
|
|
response: []exemplar.QueryResult{
|
|
{
|
|
SeriesLabels: labels.FromStrings("__name__", "test_metric3", "foo", "boo", "dup", "1"),
|
|
Exemplars: []exemplar.Exemplar{
|
|
{
|
|
Labels: labels.FromStrings("id", "abc"),
|
|
Value: 10,
|
|
Ts: timestamp.FromTime(start.Add(2 * time.Second)),
|
|
},
|
|
},
|
|
},
|
|
{
|
|
SeriesLabels: labels.FromStrings("__name__", "test_metric4", "foo", "bar", "dup", "1"),
|
|
Exemplars: []exemplar.Exemplar{
|
|
{
|
|
Labels: labels.FromStrings("id", "lul"),
|
|
Value: 10,
|
|
Ts: timestamp.FromTime(start.Add(4 * time.Second)),
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.queryExemplars,
|
|
query: url.Values{
|
|
"query": []string{`{foo="boo"}`},
|
|
"start": []string{"4"},
|
|
"end": []string{"4.1"},
|
|
},
|
|
response: []exemplar.QueryResult{
|
|
{
|
|
SeriesLabels: labels.FromStrings("__name__", "test_metric3", "foo", "boo", "dup", "1"),
|
|
Exemplars: []exemplar.Exemplar{
|
|
{
|
|
Labels: labels.FromStrings("id", "abc2"),
|
|
Value: 10,
|
|
Ts: 4053,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.queryExemplars,
|
|
query: url.Values{
|
|
"query": []string{`{foo="boo"}`},
|
|
},
|
|
response: []exemplar.QueryResult{
|
|
{
|
|
SeriesLabels: labels.FromStrings("__name__", "test_metric3", "foo", "boo", "dup", "1"),
|
|
Exemplars: []exemplar.Exemplar{
|
|
{
|
|
Labels: labels.FromStrings("id", "abc"),
|
|
Value: 10,
|
|
Ts: 2000,
|
|
},
|
|
{
|
|
Labels: labels.FromStrings("id", "abc2"),
|
|
Value: 10,
|
|
Ts: 4053,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.queryExemplars,
|
|
query: url.Values{
|
|
"query": []string{`{__name__="test_metric5"}`},
|
|
},
|
|
response: []exemplar.QueryResult{},
|
|
},
|
|
}
|
|
|
|
if testLabelAPI {
|
|
tests = append(tests, []test{
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "__name__",
|
|
},
|
|
response: []string{
|
|
"test_metric1",
|
|
"test_metric2",
|
|
"test_metric3",
|
|
"test_metric4",
|
|
},
|
|
},
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
response: []string{
|
|
"bar",
|
|
"boo",
|
|
},
|
|
},
|
|
// Bad name parameter.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "not!!!allowed",
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
// Start and end before LabelValues starts.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"start": []string{"-2"},
|
|
"end": []string{"-1"},
|
|
},
|
|
response: []string{},
|
|
},
|
|
// Start and end within LabelValues.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"start": []string{"1"},
|
|
"end": []string{"100"},
|
|
},
|
|
response: []string{
|
|
"bar",
|
|
"boo",
|
|
},
|
|
},
|
|
// Start before LabelValues, end within LabelValues.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"start": []string{"-1"},
|
|
"end": []string{"3"},
|
|
},
|
|
response: []string{
|
|
"bar",
|
|
"boo",
|
|
},
|
|
},
|
|
// Start before LabelValues starts, end after LabelValues ends.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"start": []string{"1969-12-31T00:00:00Z"},
|
|
"end": []string{"1970-02-01T00:02:03Z"},
|
|
},
|
|
response: []string{
|
|
"bar",
|
|
"boo",
|
|
},
|
|
},
|
|
// Start with bad data, end within LabelValues.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"start": []string{"boop"},
|
|
"end": []string{"1"},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
// Start within LabelValues, end after.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"start": []string{"1"},
|
|
"end": []string{"100000000"},
|
|
},
|
|
response: []string{
|
|
"bar",
|
|
"boo",
|
|
},
|
|
},
|
|
// Start and end after LabelValues ends.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"start": []string{"148966367200.372"},
|
|
"end": []string{"148966367200.972"},
|
|
},
|
|
response: []string{},
|
|
},
|
|
// Only provide Start within LabelValues, don't provide an end time.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"start": []string{"2"},
|
|
},
|
|
response: []string{
|
|
"bar",
|
|
"boo",
|
|
},
|
|
},
|
|
// Only provide end within LabelValues, don't provide a start time.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"end": []string{"100"},
|
|
},
|
|
response: []string{
|
|
"bar",
|
|
"boo",
|
|
},
|
|
},
|
|
// Label values with bad matchers.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"match[]": []string{`{foo=""`, `test_metric2`},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
// Label values with empty matchers.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"match[]": []string{`{foo=""}`},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
// Label values with matcher.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric2`},
|
|
},
|
|
response: []string{
|
|
"boo",
|
|
},
|
|
},
|
|
// Label values with matcher.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric1`},
|
|
},
|
|
response: []string{
|
|
"bar",
|
|
"boo",
|
|
},
|
|
},
|
|
// Label values with matcher using label filter.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric1{foo="bar"}`},
|
|
},
|
|
response: []string{
|
|
"bar",
|
|
},
|
|
},
|
|
// Label values with matcher and time range.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric1`},
|
|
"start": []string{"1"},
|
|
"end": []string{"100000000"},
|
|
},
|
|
response: []string{
|
|
"bar",
|
|
"boo",
|
|
},
|
|
},
|
|
// Try to overlap the selected series set as much as possible to test that the value de-duplication works.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric4{dup=~"^1"}`, `test_metric4{foo=~".+o$"}`},
|
|
},
|
|
response: []string{
|
|
"bar",
|
|
"boo",
|
|
},
|
|
},
|
|
// Label values with limit.
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "__name__",
|
|
},
|
|
query: url.Values{
|
|
"limit": []string{"2"},
|
|
},
|
|
responseLen: 2, // API does not specify which particular values will come back.
|
|
warningsCount: 1,
|
|
},
|
|
{
|
|
endpoint: api.labelValues,
|
|
params: map[string]string{
|
|
"name": "__name__",
|
|
},
|
|
query: url.Values{
|
|
"limit": []string{"4"},
|
|
},
|
|
responseLen: 4, // API does not specify which particular values will come back.
|
|
warningsCount: 0, // No warnings if limit isn't exceeded.
|
|
},
|
|
// Label names.
|
|
{
|
|
endpoint: api.labelNames,
|
|
response: []string{"__name__", "dup", "foo"},
|
|
},
|
|
// Start and end before Label names starts.
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"start": []string{"-2"},
|
|
"end": []string{"-1"},
|
|
},
|
|
response: []string{},
|
|
},
|
|
// Start and end within Label names.
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"start": []string{"1"},
|
|
"end": []string{"100"},
|
|
},
|
|
response: []string{"__name__", "dup", "foo"},
|
|
},
|
|
// Start before Label names, end within Label names.
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"start": []string{"-1"},
|
|
"end": []string{"10"},
|
|
},
|
|
response: []string{"__name__", "dup", "foo"},
|
|
},
|
|
|
|
// Start before Label names starts, end after Label names ends.
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"start": []string{"-1"},
|
|
"end": []string{"100000"},
|
|
},
|
|
response: []string{"__name__", "dup", "foo"},
|
|
},
|
|
// Start with bad data for Label names, end within Label names.
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"start": []string{"boop"},
|
|
"end": []string{"1"},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
// Start within Label names, end after.
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"start": []string{"1"},
|
|
"end": []string{"1000000006"},
|
|
},
|
|
response: []string{"__name__", "dup", "foo"},
|
|
},
|
|
// Start and end after Label names ends.
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"start": []string{"148966367200.372"},
|
|
"end": []string{"148966367200.972"},
|
|
},
|
|
response: []string{},
|
|
},
|
|
// Only provide Start within Label names, don't provide an end time.
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"start": []string{"4"},
|
|
},
|
|
response: []string{"__name__", "dup", "foo"},
|
|
},
|
|
// Only provide End within Label names, don't provide a start time.
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"end": []string{"20"},
|
|
},
|
|
response: []string{"__name__", "dup", "foo"},
|
|
},
|
|
// Label names with bad matchers.
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"match[]": []string{`{foo=""`, `test_metric2`},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
// Label values with empty matchers.
|
|
{
|
|
endpoint: api.labelNames,
|
|
params: map[string]string{
|
|
"name": "foo",
|
|
},
|
|
query: url.Values{
|
|
"match[]": []string{`{foo=""}`},
|
|
},
|
|
errType: errorBadData,
|
|
},
|
|
// Label names with matcher.
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric2`},
|
|
},
|
|
response: []string{"__name__", "foo"},
|
|
},
|
|
// Label names with matcher.
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric3`},
|
|
},
|
|
response: []string{"__name__", "dup", "foo"},
|
|
},
|
|
// Label names with matcher using label filter.
|
|
// There is no matching series.
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric1{foo="test"}`},
|
|
},
|
|
response: []string{},
|
|
},
|
|
// Label names with matcher and time range.
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"match[]": []string{`test_metric2`},
|
|
"start": []string{"1"},
|
|
"end": []string{"100000000"},
|
|
},
|
|
response: []string{"__name__", "foo"},
|
|
},
|
|
// Label names with limit.
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"limit": []string{"2"},
|
|
},
|
|
responseLen: 2, // API does not specify which particular values will come back.
|
|
warningsCount: 1,
|
|
},
|
|
{
|
|
endpoint: api.labelNames,
|
|
query: url.Values{
|
|
"limit": []string{"3"},
|
|
},
|
|
responseLen: 3, // API does not specify which particular values will come back.
|
|
warningsCount: 0, // No warnings if limit isn't exceeded.
|
|
},
|
|
}...)
|
|
}
|
|
|
|
methods := func(f apiFunc) []string {
|
|
fp := reflect.ValueOf(f).Pointer()
|
|
if fp == reflect.ValueOf(api.query).Pointer() || fp == reflect.ValueOf(api.queryRange).Pointer() || fp == reflect.ValueOf(api.series).Pointer() {
|
|
return []string{http.MethodGet, http.MethodPost}
|
|
}
|
|
return []string{http.MethodGet}
|
|
}
|
|
|
|
request := func(m string, q url.Values) (*http.Request, error) {
|
|
if m == http.MethodPost {
|
|
r, err := http.NewRequest(m, "http://example.com", strings.NewReader(q.Encode()))
|
|
r.Header.Set("Content-Type", "application/x-www-form-urlencoded")
|
|
r.RemoteAddr = "127.0.0.1:20201"
|
|
return r, err
|
|
}
|
|
r, err := http.NewRequest(m, fmt.Sprintf("http://example.com?%s", q.Encode()), nil)
|
|
r.RemoteAddr = "127.0.0.1:20201"
|
|
return r, err
|
|
}
|
|
|
|
for i, test := range tests {
|
|
t.Run(fmt.Sprintf("run %d %s %q", i, describeAPIFunc(test.endpoint), test.query.Encode()), func(t *testing.T) {
|
|
for _, method := range methods(test.endpoint) {
|
|
t.Run(method, func(t *testing.T) {
|
|
// Build a context with the correct request params.
|
|
ctx := context.Background()
|
|
for p, v := range test.params {
|
|
ctx = route.WithParam(ctx, p, v)
|
|
}
|
|
|
|
req, err := request(method, test.query)
|
|
require.NoError(t, err)
|
|
|
|
tr.ResetMetadataStore()
|
|
for _, tm := range test.metadata {
|
|
tr.SetMetadataStoreForTargets(tm.identifier, &testMetaStore{Metadata: tm.metadata})
|
|
}
|
|
|
|
for _, te := range test.exemplars {
|
|
for _, e := range te.Exemplars {
|
|
_, err := es.AppendExemplar(0, te.SeriesLabels, e)
|
|
require.NoError(t, err)
|
|
}
|
|
}
|
|
|
|
res := test.endpoint(req.WithContext(ctx))
|
|
assertAPIError(t, res.err, test.errType)
|
|
|
|
if test.sorter != nil {
|
|
test.sorter(res.data)
|
|
}
|
|
|
|
if test.responseLen != 0 {
|
|
assertAPIResponseLength(t, res.data, test.responseLen)
|
|
if test.responseMetadataTotal != 0 {
|
|
assertAPIResponseMetadataLen(t, res.data, test.responseMetadataTotal)
|
|
}
|
|
} else {
|
|
if test.zeroFunc != nil {
|
|
test.zeroFunc(res.data)
|
|
}
|
|
if test.response != nil {
|
|
assertAPIResponse(t, res.data, test.response)
|
|
}
|
|
}
|
|
|
|
if test.responseAsJSON != "" {
|
|
json := jsoniter.ConfigCompatibleWithStandardLibrary
|
|
s, err := json.Marshal(res.data)
|
|
require.NoError(t, err)
|
|
require.JSONEq(t, test.responseAsJSON, string(s))
|
|
}
|
|
|
|
require.Len(t, res.warnings, test.warningsCount)
|
|
})
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func describeAPIFunc(f apiFunc) string {
|
|
name := runtime.FuncForPC(reflect.ValueOf(f).Pointer()).Name()
|
|
return strings.Split(name[strings.LastIndex(name, ".")+1:], "-")[0]
|
|
}
|
|
|
|
func assertAPIError(t *testing.T, got *apiError, exp errorType) {
|
|
t.Helper()
|
|
|
|
if exp == errorNone {
|
|
require.Nil(t, got)
|
|
} else {
|
|
require.NotNil(t, got)
|
|
require.Equal(t, exp, got.typ, "(%q)", got)
|
|
}
|
|
}
|
|
|
|
func assertAPIResponse(t *testing.T, got, exp interface{}) {
|
|
t.Helper()
|
|
|
|
testutil.RequireEqual(t, exp, got)
|
|
}
|
|
|
|
func assertAPIResponseLength(t *testing.T, got interface{}, expLen int) {
|
|
t.Helper()
|
|
|
|
gotLen := reflect.ValueOf(got).Len()
|
|
require.Equal(t, expLen, gotLen, "Response length does not match")
|
|
}
|
|
|
|
func assertAPIResponseMetadataLen(t *testing.T, got interface{}, expLen int) {
|
|
t.Helper()
|
|
|
|
var gotLen int
|
|
response := got.(map[string][]metadata.Metadata)
|
|
for _, m := range response {
|
|
gotLen += len(m)
|
|
}
|
|
|
|
require.Equal(t, expLen, gotLen, "Amount of metadata in the response does not match")
|
|
}
|
|
|
|
type fakeDB struct {
|
|
err error
|
|
}
|
|
|
|
func (f *fakeDB) CleanTombstones() error { return f.err }
|
|
func (f *fakeDB) Delete(context.Context, int64, int64, ...*labels.Matcher) error { return f.err }
|
|
func (f *fakeDB) Snapshot(string, bool) error { return f.err }
|
|
func (f *fakeDB) Stats(statsByLabelName string, limit int) (_ *tsdb.Stats, retErr error) {
|
|
dbDir, err := os.MkdirTemp("", "tsdb-api-ready")
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
defer func() {
|
|
err := os.RemoveAll(dbDir)
|
|
if retErr != nil {
|
|
retErr = err
|
|
}
|
|
}()
|
|
opts := tsdb.DefaultHeadOptions()
|
|
opts.ChunkRange = 1000
|
|
h, _ := tsdb.NewHead(nil, nil, nil, nil, opts, nil)
|
|
return h.Stats(statsByLabelName, limit), nil
|
|
}
|
|
|
|
func (f *fakeDB) WALReplayStatus() (tsdb.WALReplayStatus, error) {
|
|
return tsdb.WALReplayStatus{}, nil
|
|
}
|
|
|
|
func TestAdminEndpoints(t *testing.T) {
|
|
tsdb, tsdbWithError, tsdbNotReady := &fakeDB{}, &fakeDB{err: errors.New("some error")}, &fakeDB{err: fmt.Errorf("wrap: %w", tsdb.ErrNotReady)}
|
|
snapshotAPI := func(api *API) apiFunc { return api.snapshot }
|
|
cleanAPI := func(api *API) apiFunc { return api.cleanTombstones }
|
|
deleteAPI := func(api *API) apiFunc { return api.deleteSeries }
|
|
|
|
for _, tc := range []struct {
|
|
db *fakeDB
|
|
enableAdmin bool
|
|
endpoint func(api *API) apiFunc
|
|
method string
|
|
values url.Values
|
|
|
|
errType errorType
|
|
}{
|
|
// Tests for the snapshot endpoint.
|
|
{
|
|
db: tsdb,
|
|
enableAdmin: false,
|
|
endpoint: snapshotAPI,
|
|
|
|
errType: errorUnavailable,
|
|
},
|
|
{
|
|
db: tsdb,
|
|
enableAdmin: true,
|
|
endpoint: snapshotAPI,
|
|
|
|
errType: errorNone,
|
|
},
|
|
{
|
|
db: tsdb,
|
|
enableAdmin: true,
|
|
endpoint: snapshotAPI,
|
|
values: map[string][]string{"skip_head": {"true"}},
|
|
|
|
errType: errorNone,
|
|
},
|
|
{
|
|
db: tsdb,
|
|
enableAdmin: true,
|
|
endpoint: snapshotAPI,
|
|
values: map[string][]string{"skip_head": {"xxx"}},
|
|
|
|
errType: errorBadData,
|
|
},
|
|
{
|
|
db: tsdbWithError,
|
|
enableAdmin: true,
|
|
endpoint: snapshotAPI,
|
|
|
|
errType: errorInternal,
|
|
},
|
|
{
|
|
db: tsdbNotReady,
|
|
enableAdmin: true,
|
|
endpoint: snapshotAPI,
|
|
|
|
errType: errorUnavailable,
|
|
},
|
|
// Tests for the cleanTombstones endpoint.
|
|
{
|
|
db: tsdb,
|
|
enableAdmin: false,
|
|
endpoint: cleanAPI,
|
|
|
|
errType: errorUnavailable,
|
|
},
|
|
{
|
|
db: tsdb,
|
|
enableAdmin: true,
|
|
endpoint: cleanAPI,
|
|
|
|
errType: errorNone,
|
|
},
|
|
{
|
|
db: tsdbWithError,
|
|
enableAdmin: true,
|
|
endpoint: cleanAPI,
|
|
|
|
errType: errorInternal,
|
|
},
|
|
{
|
|
db: tsdbNotReady,
|
|
enableAdmin: true,
|
|
endpoint: cleanAPI,
|
|
|
|
errType: errorUnavailable,
|
|
},
|
|
// Tests for the deleteSeries endpoint.
|
|
{
|
|
db: tsdb,
|
|
enableAdmin: false,
|
|
endpoint: deleteAPI,
|
|
|
|
errType: errorUnavailable,
|
|
},
|
|
{
|
|
db: tsdb,
|
|
enableAdmin: true,
|
|
endpoint: deleteAPI,
|
|
|
|
errType: errorBadData,
|
|
},
|
|
{
|
|
db: tsdb,
|
|
enableAdmin: true,
|
|
endpoint: deleteAPI,
|
|
values: map[string][]string{"match[]": {"123"}},
|
|
|
|
errType: errorBadData,
|
|
},
|
|
{
|
|
db: tsdb,
|
|
enableAdmin: true,
|
|
endpoint: deleteAPI,
|
|
values: map[string][]string{"match[]": {"up"}, "start": {"xxx"}},
|
|
|
|
errType: errorBadData,
|
|
},
|
|
{
|
|
db: tsdb,
|
|
enableAdmin: true,
|
|
endpoint: deleteAPI,
|
|
values: map[string][]string{"match[]": {"up"}, "end": {"xxx"}},
|
|
|
|
errType: errorBadData,
|
|
},
|
|
{
|
|
db: tsdb,
|
|
enableAdmin: true,
|
|
endpoint: deleteAPI,
|
|
values: map[string][]string{"match[]": {"up"}},
|
|
|
|
errType: errorNone,
|
|
},
|
|
{
|
|
db: tsdb,
|
|
enableAdmin: true,
|
|
endpoint: deleteAPI,
|
|
values: map[string][]string{"match[]": {"up{job!=\"foo\"}", "{job=~\"bar.+\"}", "up{instance!~\"fred.+\"}"}},
|
|
|
|
errType: errorNone,
|
|
},
|
|
{
|
|
db: tsdbWithError,
|
|
enableAdmin: true,
|
|
endpoint: deleteAPI,
|
|
values: map[string][]string{"match[]": {"up"}},
|
|
|
|
errType: errorInternal,
|
|
},
|
|
{
|
|
db: tsdbNotReady,
|
|
enableAdmin: true,
|
|
endpoint: deleteAPI,
|
|
values: map[string][]string{"match[]": {"up"}},
|
|
|
|
errType: errorUnavailable,
|
|
},
|
|
} {
|
|
tc := tc
|
|
t.Run("", func(t *testing.T) {
|
|
dir := t.TempDir()
|
|
|
|
api := &API{
|
|
db: tc.db,
|
|
dbDir: dir,
|
|
ready: func(f http.HandlerFunc) http.HandlerFunc { return f },
|
|
enableAdmin: tc.enableAdmin,
|
|
}
|
|
|
|
endpoint := tc.endpoint(api)
|
|
req, err := http.NewRequest(tc.method, fmt.Sprintf("?%s", tc.values.Encode()), nil)
|
|
require.NoError(t, err)
|
|
|
|
res := setUnavailStatusOnTSDBNotReady(endpoint(req))
|
|
assertAPIError(t, res.err, tc.errType)
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestRespondSuccess(t *testing.T) {
|
|
api := API{
|
|
logger: promslog.NewNopLogger(),
|
|
}
|
|
|
|
api.ClearCodecs()
|
|
api.InstallCodec(JSONCodec{})
|
|
api.InstallCodec(&testCodec{contentType: MIMEType{"test", "cannot-encode"}, canEncode: false})
|
|
api.InstallCodec(&testCodec{contentType: MIMEType{"test", "can-encode"}, canEncode: true})
|
|
api.InstallCodec(&testCodec{contentType: MIMEType{"test", "can-encode-2"}, canEncode: true})
|
|
|
|
s := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
|
api.respond(w, r, "test", nil, "")
|
|
}))
|
|
defer s.Close()
|
|
|
|
for _, tc := range []struct {
|
|
name string
|
|
acceptHeader string
|
|
expectedContentType string
|
|
expectedBody string
|
|
}{
|
|
{
|
|
name: "no Accept header",
|
|
expectedContentType: "application/json",
|
|
expectedBody: `{"status":"success","data":"test"}`,
|
|
},
|
|
{
|
|
name: "Accept header with single content type which is suitable",
|
|
acceptHeader: "test/can-encode",
|
|
expectedContentType: "test/can-encode",
|
|
expectedBody: `response from test/can-encode codec`,
|
|
},
|
|
{
|
|
name: "Accept header with single content type which is not available",
|
|
acceptHeader: "test/not-registered",
|
|
expectedContentType: "application/json",
|
|
expectedBody: `{"status":"success","data":"test"}`,
|
|
},
|
|
{
|
|
name: "Accept header with single content type which cannot encode the response payload",
|
|
acceptHeader: "test/cannot-encode",
|
|
expectedContentType: "application/json",
|
|
expectedBody: `{"status":"success","data":"test"}`,
|
|
},
|
|
{
|
|
name: "Accept header with multiple content types, all of which are suitable",
|
|
acceptHeader: "test/can-encode, test/can-encode-2",
|
|
expectedContentType: "test/can-encode",
|
|
expectedBody: `response from test/can-encode codec`,
|
|
},
|
|
{
|
|
name: "Accept header with multiple content types, only one of which is available",
|
|
acceptHeader: "test/not-registered, test/can-encode",
|
|
expectedContentType: "test/can-encode",
|
|
expectedBody: `response from test/can-encode codec`,
|
|
},
|
|
{
|
|
name: "Accept header with multiple content types, only one of which can encode the response payload",
|
|
acceptHeader: "test/cannot-encode, test/can-encode",
|
|
expectedContentType: "test/can-encode",
|
|
expectedBody: `response from test/can-encode codec`,
|
|
},
|
|
{
|
|
name: "Accept header with multiple content types, none of which are available",
|
|
acceptHeader: "test/not-registered, test/also-not-registered",
|
|
expectedContentType: "application/json",
|
|
expectedBody: `{"status":"success","data":"test"}`,
|
|
},
|
|
} {
|
|
t.Run(tc.name, func(t *testing.T) {
|
|
req, err := http.NewRequest(http.MethodGet, s.URL, nil)
|
|
require.NoError(t, err)
|
|
|
|
if tc.acceptHeader != "" {
|
|
req.Header.Set("Accept", tc.acceptHeader)
|
|
}
|
|
|
|
resp, err := http.DefaultClient.Do(req)
|
|
require.NoError(t, err)
|
|
|
|
body, err := io.ReadAll(resp.Body)
|
|
defer resp.Body.Close()
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, http.StatusOK, resp.StatusCode)
|
|
require.Equal(t, tc.expectedContentType, resp.Header.Get("Content-Type"))
|
|
require.Equal(t, tc.expectedBody, string(body))
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestRespondSuccess_DefaultCodecCannotEncodeResponse(t *testing.T) {
|
|
api := API{
|
|
logger: promslog.NewNopLogger(),
|
|
}
|
|
|
|
api.ClearCodecs()
|
|
api.InstallCodec(&testCodec{contentType: MIMEType{"application", "default-format"}, canEncode: false})
|
|
|
|
s := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
|
api.respond(w, r, "test", nil, "")
|
|
}))
|
|
defer s.Close()
|
|
|
|
req, err := http.NewRequest(http.MethodGet, s.URL, nil)
|
|
require.NoError(t, err)
|
|
|
|
resp, err := http.DefaultClient.Do(req)
|
|
require.NoError(t, err)
|
|
|
|
body, err := io.ReadAll(resp.Body)
|
|
defer resp.Body.Close()
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, http.StatusNotAcceptable, resp.StatusCode)
|
|
require.Equal(t, "application/json", resp.Header.Get("Content-Type"))
|
|
require.Equal(t, `{"status":"error","errorType":"not_acceptable","error":"cannot encode response as application/default-format"}`, string(body))
|
|
}
|
|
|
|
func TestRespondError(t *testing.T) {
|
|
s := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
|
api := API{}
|
|
api.respondError(w, &apiError{errorTimeout, errors.New("message")}, "test")
|
|
}))
|
|
defer s.Close()
|
|
|
|
resp, err := http.Get(s.URL)
|
|
require.NoError(t, err, "Error on test request")
|
|
body, err := io.ReadAll(resp.Body)
|
|
defer resp.Body.Close()
|
|
require.NoError(t, err, "Error reading response body")
|
|
want, have := http.StatusServiceUnavailable, resp.StatusCode
|
|
require.Equal(t, want, have, "Return code %d expected in error response but got %d", want, have)
|
|
h := resp.Header.Get("Content-Type")
|
|
require.Equal(t, "application/json", h, "Expected Content-Type %q but got %q", "application/json", h)
|
|
require.JSONEq(t, `{"status": "error", "data": "test", "errorType": "timeout", "error": "message"}`, string(body))
|
|
}
|
|
|
|
func TestParseTimeParam(t *testing.T) {
|
|
type resultType struct {
|
|
asTime time.Time
|
|
asError func() error
|
|
}
|
|
|
|
ts, err := parseTime("1582468023986")
|
|
require.NoError(t, err)
|
|
|
|
tests := []struct {
|
|
paramName string
|
|
paramValue string
|
|
defaultValue time.Time
|
|
result resultType
|
|
}{
|
|
{ // When data is valid.
|
|
paramName: "start",
|
|
paramValue: "1582468023986",
|
|
defaultValue: MinTime,
|
|
result: resultType{
|
|
asTime: ts,
|
|
asError: nil,
|
|
},
|
|
},
|
|
{ // When data is empty string.
|
|
paramName: "end",
|
|
paramValue: "",
|
|
defaultValue: MaxTime,
|
|
result: resultType{
|
|
asTime: MaxTime,
|
|
asError: nil,
|
|
},
|
|
},
|
|
{ // When data is not valid.
|
|
paramName: "foo",
|
|
paramValue: "baz",
|
|
defaultValue: MaxTime,
|
|
result: resultType{
|
|
asTime: time.Time{},
|
|
asError: func() error {
|
|
_, err := parseTime("baz")
|
|
return fmt.Errorf("Invalid time value for '%s': %w", "foo", err)
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
for _, test := range tests {
|
|
req, err := http.NewRequest(http.MethodGet, "localhost:42/foo?"+test.paramName+"="+test.paramValue, nil)
|
|
require.NoError(t, err)
|
|
|
|
result := test.result
|
|
asTime, err := parseTimeParam(req, test.paramName, test.defaultValue)
|
|
|
|
if err != nil {
|
|
require.EqualError(t, err, result.asError().Error())
|
|
} else {
|
|
require.True(t, asTime.Equal(result.asTime), "time as return value: %s not parsed correctly. Expected %s. Actual %s", test.paramValue, result.asTime, asTime)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestParseTime(t *testing.T) {
|
|
ts, err := time.Parse(time.RFC3339Nano, "2015-06-03T13:21:58.555Z")
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
tests := []struct {
|
|
input string
|
|
fail bool
|
|
result time.Time
|
|
}{
|
|
{
|
|
input: "",
|
|
fail: true,
|
|
},
|
|
{
|
|
input: "abc",
|
|
fail: true,
|
|
},
|
|
{
|
|
input: "30s",
|
|
fail: true,
|
|
},
|
|
{
|
|
input: "123",
|
|
result: time.Unix(123, 0),
|
|
},
|
|
{
|
|
input: "123.123",
|
|
result: time.Unix(123, 123000000),
|
|
},
|
|
{
|
|
input: "2015-06-03T13:21:58.555Z",
|
|
result: ts,
|
|
},
|
|
{
|
|
input: "2015-06-03T14:21:58.555+01:00",
|
|
result: ts,
|
|
},
|
|
{
|
|
// Test float rounding.
|
|
input: "1543578564.705",
|
|
result: time.Unix(1543578564, 705*1e6),
|
|
},
|
|
{
|
|
input: MinTime.Format(time.RFC3339Nano),
|
|
result: MinTime,
|
|
},
|
|
{
|
|
input: MaxTime.Format(time.RFC3339Nano),
|
|
result: MaxTime,
|
|
},
|
|
}
|
|
|
|
for _, test := range tests {
|
|
ts, err := parseTime(test.input)
|
|
if !test.fail {
|
|
require.NoError(t, err, "Unexpected error for %q", test.input)
|
|
require.NotNil(t, ts)
|
|
require.True(t, ts.Equal(test.result), "Expected time %v for input %q but got %v", test.result, test.input, ts)
|
|
continue
|
|
}
|
|
require.Error(t, err, "Expected error for %q but got none", test.input)
|
|
}
|
|
}
|
|
|
|
func TestParseDuration(t *testing.T) {
|
|
tests := []struct {
|
|
input string
|
|
fail bool
|
|
result time.Duration
|
|
}{
|
|
{
|
|
input: "",
|
|
fail: true,
|
|
}, {
|
|
input: "abc",
|
|
fail: true,
|
|
}, {
|
|
input: "2015-06-03T13:21:58.555Z",
|
|
fail: true,
|
|
}, {
|
|
// Internal int64 overflow.
|
|
input: "-148966367200.372",
|
|
fail: true,
|
|
}, {
|
|
// Internal int64 overflow.
|
|
input: "148966367200.372",
|
|
fail: true,
|
|
}, {
|
|
input: "123",
|
|
result: 123 * time.Second,
|
|
}, {
|
|
input: "123.333",
|
|
result: 123*time.Second + 333*time.Millisecond,
|
|
}, {
|
|
input: "15s",
|
|
result: 15 * time.Second,
|
|
}, {
|
|
input: "5m",
|
|
result: 5 * time.Minute,
|
|
},
|
|
}
|
|
|
|
for _, test := range tests {
|
|
d, err := parseDuration(test.input)
|
|
if !test.fail {
|
|
require.NoError(t, err, "Unexpected error for %q", test.input)
|
|
require.Equal(t, test.result, d, "Expected duration %v for input %q but got %v", test.result, test.input, d)
|
|
continue
|
|
}
|
|
require.Error(t, err, "Expected error for %q but got none", test.input)
|
|
}
|
|
}
|
|
|
|
func TestOptionsMethod(t *testing.T) {
|
|
r := route.New()
|
|
api := &API{ready: func(f http.HandlerFunc) http.HandlerFunc { return f }}
|
|
api.Register(r)
|
|
|
|
s := httptest.NewServer(r)
|
|
defer s.Close()
|
|
|
|
req, err := http.NewRequest(http.MethodOptions, s.URL+"/any_path", nil)
|
|
require.NoError(t, err, "Error creating OPTIONS request")
|
|
client := &http.Client{}
|
|
resp, err := client.Do(req)
|
|
require.NoError(t, err, "Error executing OPTIONS request")
|
|
require.Equal(t, http.StatusNoContent, resp.StatusCode)
|
|
}
|
|
|
|
func TestTSDBStatus(t *testing.T) {
|
|
tsdb := &fakeDB{}
|
|
tsdbStatusAPI := func(api *API) apiFunc { return api.serveTSDBStatus }
|
|
|
|
for i, tc := range []struct {
|
|
db *fakeDB
|
|
endpoint func(api *API) apiFunc
|
|
method string
|
|
values url.Values
|
|
|
|
errType errorType
|
|
}{
|
|
// Tests for the TSDB Status endpoint.
|
|
{
|
|
db: tsdb,
|
|
endpoint: tsdbStatusAPI,
|
|
errType: errorNone,
|
|
},
|
|
{
|
|
db: tsdb,
|
|
endpoint: tsdbStatusAPI,
|
|
values: map[string][]string{"limit": {"20"}},
|
|
errType: errorNone,
|
|
},
|
|
{
|
|
db: tsdb,
|
|
endpoint: tsdbStatusAPI,
|
|
values: map[string][]string{"limit": {"0"}},
|
|
errType: errorBadData,
|
|
},
|
|
} {
|
|
tc := tc
|
|
t.Run(strconv.Itoa(i), func(t *testing.T) {
|
|
api := &API{db: tc.db, gatherer: prometheus.DefaultGatherer}
|
|
endpoint := tc.endpoint(api)
|
|
req, err := http.NewRequest(tc.method, fmt.Sprintf("?%s", tc.values.Encode()), nil)
|
|
require.NoError(t, err, "Error when creating test request")
|
|
res := endpoint(req)
|
|
assertAPIError(t, res.err, tc.errType)
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestReturnAPIError(t *testing.T) {
|
|
cases := []struct {
|
|
err error
|
|
expected errorType
|
|
}{
|
|
{
|
|
err: promql.ErrStorage{Err: errors.New("storage error")},
|
|
expected: errorInternal,
|
|
}, {
|
|
err: fmt.Errorf("wrapped: %w", promql.ErrStorage{Err: errors.New("storage error")}),
|
|
expected: errorInternal,
|
|
}, {
|
|
err: promql.ErrQueryTimeout("timeout error"),
|
|
expected: errorTimeout,
|
|
}, {
|
|
err: fmt.Errorf("wrapped: %w", promql.ErrQueryTimeout("timeout error")),
|
|
expected: errorTimeout,
|
|
}, {
|
|
err: promql.ErrQueryCanceled("canceled error"),
|
|
expected: errorCanceled,
|
|
}, {
|
|
err: fmt.Errorf("wrapped: %w", promql.ErrQueryCanceled("canceled error")),
|
|
expected: errorCanceled,
|
|
}, {
|
|
err: errors.New("exec error"),
|
|
expected: errorExec,
|
|
}, {
|
|
err: context.Canceled,
|
|
expected: errorCanceled,
|
|
},
|
|
}
|
|
|
|
for ix, c := range cases {
|
|
actual := returnAPIError(c.err)
|
|
require.Error(t, actual, ix)
|
|
require.Equal(t, c.expected, actual.typ, ix)
|
|
}
|
|
}
|
|
|
|
// This is a global to avoid the benchmark being optimized away.
|
|
var testResponseWriter = httptest.ResponseRecorder{}
|
|
|
|
func BenchmarkRespond(b *testing.B) {
|
|
points := []promql.FPoint{}
|
|
for i := 0; i < 10000; i++ {
|
|
points = append(points, promql.FPoint{F: float64(i * 1000000), T: int64(i)})
|
|
}
|
|
matrix := promql.Matrix{}
|
|
for i := 0; i < 1000; i++ {
|
|
matrix = append(matrix, promql.Series{
|
|
Metric: labels.FromStrings("__name__", fmt.Sprintf("series%v", i),
|
|
"label", fmt.Sprintf("series%v", i),
|
|
"label2", fmt.Sprintf("series%v", i)),
|
|
Floats: points[:10],
|
|
})
|
|
}
|
|
series := []labels.Labels{}
|
|
for i := 0; i < 1000; i++ {
|
|
series = append(series, labels.FromStrings("__name__", fmt.Sprintf("series%v", i),
|
|
"label", fmt.Sprintf("series%v", i),
|
|
"label2", fmt.Sprintf("series%v", i)))
|
|
}
|
|
|
|
cases := []struct {
|
|
name string
|
|
response interface{}
|
|
}{
|
|
{name: "10000 points no labels", response: &QueryData{
|
|
ResultType: parser.ValueTypeMatrix,
|
|
Result: promql.Matrix{
|
|
promql.Series{
|
|
Floats: points,
|
|
Metric: labels.EmptyLabels(),
|
|
},
|
|
},
|
|
}},
|
|
{name: "1000 labels", response: series},
|
|
{name: "1000 series 10 points", response: &QueryData{
|
|
ResultType: parser.ValueTypeMatrix,
|
|
Result: matrix,
|
|
}},
|
|
}
|
|
for _, c := range cases {
|
|
b.Run(c.name, func(b *testing.B) {
|
|
b.ReportAllocs()
|
|
request, err := http.NewRequest(http.MethodGet, "/does-not-matter", nil)
|
|
require.NoError(b, err)
|
|
b.ResetTimer()
|
|
api := API{}
|
|
api.InstallCodec(JSONCodec{})
|
|
for n := 0; n < b.N; n++ {
|
|
api.respond(&testResponseWriter, request, c.response, nil, "")
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestGetGlobalURL(t *testing.T) {
|
|
mustParseURL := func(t *testing.T, u string) *url.URL {
|
|
parsed, err := url.Parse(u)
|
|
require.NoError(t, err)
|
|
return parsed
|
|
}
|
|
|
|
testcases := []struct {
|
|
input *url.URL
|
|
opts GlobalURLOptions
|
|
expected *url.URL
|
|
errorful bool
|
|
}{
|
|
{
|
|
mustParseURL(t, "http://127.0.0.1:9090"),
|
|
GlobalURLOptions{
|
|
ListenAddress: "127.0.0.1:9090",
|
|
Host: "127.0.0.1:9090",
|
|
Scheme: "http",
|
|
},
|
|
mustParseURL(t, "http://127.0.0.1:9090"),
|
|
false,
|
|
},
|
|
{
|
|
mustParseURL(t, "http://127.0.0.1:9090"),
|
|
GlobalURLOptions{
|
|
ListenAddress: "127.0.0.1:9090",
|
|
Host: "prometheus.io",
|
|
Scheme: "https",
|
|
},
|
|
mustParseURL(t, "https://prometheus.io"),
|
|
false,
|
|
},
|
|
{
|
|
mustParseURL(t, "http://example.com"),
|
|
GlobalURLOptions{
|
|
ListenAddress: "127.0.0.1:9090",
|
|
Host: "prometheus.io",
|
|
Scheme: "https",
|
|
},
|
|
mustParseURL(t, "http://example.com"),
|
|
false,
|
|
},
|
|
{
|
|
mustParseURL(t, "http://localhost:8080"),
|
|
GlobalURLOptions{
|
|
ListenAddress: "127.0.0.1:9090",
|
|
Host: "prometheus.io",
|
|
Scheme: "https",
|
|
},
|
|
mustParseURL(t, "http://prometheus.io:8080"),
|
|
false,
|
|
},
|
|
{
|
|
mustParseURL(t, "http://[::1]:8080"),
|
|
GlobalURLOptions{
|
|
ListenAddress: "127.0.0.1:9090",
|
|
Host: "prometheus.io",
|
|
Scheme: "https",
|
|
},
|
|
mustParseURL(t, "http://prometheus.io:8080"),
|
|
false,
|
|
},
|
|
{
|
|
mustParseURL(t, "http://localhost"),
|
|
GlobalURLOptions{
|
|
ListenAddress: "127.0.0.1:9090",
|
|
Host: "prometheus.io",
|
|
Scheme: "https",
|
|
},
|
|
mustParseURL(t, "http://prometheus.io"),
|
|
false,
|
|
},
|
|
{
|
|
mustParseURL(t, "http://localhost:9091"),
|
|
GlobalURLOptions{
|
|
ListenAddress: "[::1]:9090",
|
|
Host: "[::1]",
|
|
Scheme: "https",
|
|
},
|
|
mustParseURL(t, "http://[::1]:9091"),
|
|
false,
|
|
},
|
|
{
|
|
mustParseURL(t, "http://localhost:9091"),
|
|
GlobalURLOptions{
|
|
ListenAddress: "[::1]:9090",
|
|
Host: "[::1]:9090",
|
|
Scheme: "https",
|
|
},
|
|
mustParseURL(t, "http://[::1]:9091"),
|
|
false,
|
|
},
|
|
}
|
|
|
|
for i, tc := range testcases {
|
|
t.Run(fmt.Sprintf("Test %d", i), func(t *testing.T) {
|
|
output, err := getGlobalURL(tc.input, tc.opts)
|
|
if tc.errorful {
|
|
require.Error(t, err)
|
|
return
|
|
}
|
|
require.NoError(t, err)
|
|
require.Equal(t, tc.expected, output)
|
|
})
|
|
}
|
|
}
|
|
|
|
type testCodec struct {
|
|
contentType MIMEType
|
|
canEncode bool
|
|
}
|
|
|
|
func (t *testCodec) ContentType() MIMEType {
|
|
return t.contentType
|
|
}
|
|
|
|
func (t *testCodec) CanEncode(_ *Response) bool {
|
|
return t.canEncode
|
|
}
|
|
|
|
func (t *testCodec) Encode(_ *Response) ([]byte, error) {
|
|
return []byte(fmt.Sprintf("response from %v codec", t.contentType)), nil
|
|
}
|
|
|
|
func TestExtractQueryOpts(t *testing.T) {
|
|
tests := []struct {
|
|
name string
|
|
form url.Values
|
|
expect promql.QueryOpts
|
|
err error
|
|
}{
|
|
{
|
|
name: "with stats all",
|
|
form: url.Values{
|
|
"stats": []string{"all"},
|
|
},
|
|
expect: promql.NewPrometheusQueryOpts(true, 0),
|
|
|
|
err: nil,
|
|
},
|
|
{
|
|
name: "with stats none",
|
|
form: url.Values{
|
|
"stats": []string{"none"},
|
|
},
|
|
expect: promql.NewPrometheusQueryOpts(false, 0),
|
|
err: nil,
|
|
},
|
|
{
|
|
name: "with lookback delta",
|
|
form: url.Values{
|
|
"stats": []string{"all"},
|
|
"lookback_delta": []string{"30s"},
|
|
},
|
|
expect: promql.NewPrometheusQueryOpts(true, 30*time.Second),
|
|
err: nil,
|
|
},
|
|
{
|
|
name: "with invalid lookback delta",
|
|
form: url.Values{
|
|
"lookback_delta": []string{"invalid"},
|
|
},
|
|
expect: nil,
|
|
err: errors.New(`error parsing lookback delta duration: cannot parse "invalid" to a valid duration`),
|
|
},
|
|
}
|
|
|
|
for _, test := range tests {
|
|
t.Run(test.name, func(t *testing.T) {
|
|
req := &http.Request{Form: test.form}
|
|
opts, err := extractQueryOpts(req)
|
|
require.Equal(t, test.expect, opts)
|
|
if test.err == nil {
|
|
require.NoError(t, err)
|
|
} else {
|
|
require.EqualError(t, err, test.err.Error())
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
// Test query timeout parameter.
|
|
func TestQueryTimeout(t *testing.T) {
|
|
storage := promqltest.LoadedStorage(t, `
|
|
load 1m
|
|
test_metric1{foo="bar"} 0+100x100
|
|
`)
|
|
t.Cleanup(func() {
|
|
_ = storage.Close()
|
|
})
|
|
|
|
now := time.Now()
|
|
|
|
for _, tc := range []struct {
|
|
name string
|
|
method string
|
|
}{
|
|
{
|
|
name: "GET method",
|
|
method: http.MethodGet,
|
|
},
|
|
{
|
|
name: "POST method",
|
|
method: http.MethodPost,
|
|
},
|
|
} {
|
|
t.Run(tc.name, func(t *testing.T) {
|
|
engine := &fakeEngine{}
|
|
api := &API{
|
|
Queryable: storage,
|
|
QueryEngine: engine,
|
|
ExemplarQueryable: storage.ExemplarQueryable(),
|
|
alertmanagerRetriever: testAlertmanagerRetriever{}.toFactory(),
|
|
flagsMap: sampleFlagMap,
|
|
now: func() time.Time { return now },
|
|
config: func() config.Config { return samplePrometheusCfg },
|
|
ready: func(f http.HandlerFunc) http.HandlerFunc { return f },
|
|
}
|
|
|
|
query := url.Values{
|
|
"query": []string{"2"},
|
|
"timeout": []string{"1s"},
|
|
}
|
|
ctx := context.Background()
|
|
req, err := http.NewRequest(tc.method, fmt.Sprintf("http://example.com?%s", query.Encode()), nil)
|
|
require.NoError(t, err)
|
|
req.RemoteAddr = "127.0.0.1:20201"
|
|
|
|
res := api.query(req.WithContext(ctx))
|
|
assertAPIError(t, res.err, errorNone)
|
|
|
|
require.Len(t, engine.query.execCalls, 1)
|
|
deadline, ok := engine.query.execCalls[0].Deadline()
|
|
require.True(t, ok)
|
|
require.Equal(t, now.Add(time.Second), deadline)
|
|
})
|
|
}
|
|
}
|
|
|
|
// fakeEngine is a fake QueryEngine implementation.
|
|
type fakeEngine struct {
|
|
query fakeQuery
|
|
}
|
|
|
|
func (e *fakeEngine) NewInstantQuery(ctx context.Context, q storage.Queryable, opts promql.QueryOpts, qs string, ts time.Time) (promql.Query, error) {
|
|
return &e.query, nil
|
|
}
|
|
|
|
func (e *fakeEngine) NewRangeQuery(ctx context.Context, q storage.Queryable, opts promql.QueryOpts, qs string, start, end time.Time, interval time.Duration) (promql.Query, error) {
|
|
return &e.query, nil
|
|
}
|
|
|
|
// fakeQuery is a fake Query implementation.
|
|
type fakeQuery struct {
|
|
query string
|
|
execCalls []context.Context
|
|
}
|
|
|
|
func (q *fakeQuery) Exec(ctx context.Context) *promql.Result {
|
|
q.execCalls = append(q.execCalls, ctx)
|
|
return &promql.Result{
|
|
Value: &parser.StringLiteral{
|
|
Val: "test",
|
|
},
|
|
}
|
|
}
|
|
|
|
func (q *fakeQuery) Close() {}
|
|
|
|
func (q *fakeQuery) Statement() parser.Statement {
|
|
return nil
|
|
}
|
|
|
|
func (q *fakeQuery) Stats() *stats.Statistics {
|
|
return nil
|
|
}
|
|
|
|
func (q *fakeQuery) Cancel() {}
|
|
|
|
func (q *fakeQuery) String() string {
|
|
return q.query
|
|
}
|