mirror of https://github.com/prometheus/prometheus
2303 lines
75 KiB
Go
2303 lines
75 KiB
Go
// Copyright 2015 The Prometheus Authors
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package config
|
|
|
|
import (
|
|
"crypto/tls"
|
|
"encoding/json"
|
|
"net/url"
|
|
"os"
|
|
"path/filepath"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/alecthomas/units"
|
|
"github.com/go-kit/log"
|
|
"github.com/grafana/regexp"
|
|
"github.com/prometheus/common/config"
|
|
"github.com/prometheus/common/model"
|
|
"github.com/stretchr/testify/require"
|
|
"gopkg.in/yaml.v2"
|
|
|
|
"github.com/prometheus/prometheus/discovery"
|
|
"github.com/prometheus/prometheus/discovery/aws"
|
|
"github.com/prometheus/prometheus/discovery/azure"
|
|
"github.com/prometheus/prometheus/discovery/consul"
|
|
"github.com/prometheus/prometheus/discovery/digitalocean"
|
|
"github.com/prometheus/prometheus/discovery/dns"
|
|
"github.com/prometheus/prometheus/discovery/eureka"
|
|
"github.com/prometheus/prometheus/discovery/file"
|
|
"github.com/prometheus/prometheus/discovery/hetzner"
|
|
"github.com/prometheus/prometheus/discovery/http"
|
|
"github.com/prometheus/prometheus/discovery/ionos"
|
|
"github.com/prometheus/prometheus/discovery/kubernetes"
|
|
"github.com/prometheus/prometheus/discovery/linode"
|
|
"github.com/prometheus/prometheus/discovery/marathon"
|
|
"github.com/prometheus/prometheus/discovery/moby"
|
|
"github.com/prometheus/prometheus/discovery/nomad"
|
|
"github.com/prometheus/prometheus/discovery/openstack"
|
|
"github.com/prometheus/prometheus/discovery/ovhcloud"
|
|
"github.com/prometheus/prometheus/discovery/puppetdb"
|
|
"github.com/prometheus/prometheus/discovery/scaleway"
|
|
"github.com/prometheus/prometheus/discovery/targetgroup"
|
|
"github.com/prometheus/prometheus/discovery/triton"
|
|
"github.com/prometheus/prometheus/discovery/uyuni"
|
|
"github.com/prometheus/prometheus/discovery/vultr"
|
|
"github.com/prometheus/prometheus/discovery/xds"
|
|
"github.com/prometheus/prometheus/discovery/zookeeper"
|
|
"github.com/prometheus/prometheus/model/labels"
|
|
"github.com/prometheus/prometheus/model/relabel"
|
|
"github.com/prometheus/prometheus/util/testutil"
|
|
)
|
|
|
|
func mustParseURL(u string) *config.URL {
|
|
parsed, err := url.Parse(u)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
return &config.URL{URL: parsed}
|
|
}
|
|
|
|
const (
|
|
globBodySizeLimit = 15 * units.MiB
|
|
globSampleLimit = 1500
|
|
globTargetLimit = 30
|
|
globLabelLimit = 30
|
|
globLabelNameLengthLimit = 200
|
|
globLabelValueLengthLimit = 200
|
|
globalGoGC = 42
|
|
)
|
|
|
|
var expectedConf = &Config{
|
|
GlobalConfig: GlobalConfig{
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EvaluationInterval: model.Duration(30 * time.Second),
|
|
QueryLogFile: "",
|
|
|
|
ExternalLabels: labels.FromStrings("foo", "bar", "monitor", "codelab"),
|
|
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
},
|
|
|
|
Runtime: RuntimeConfig{
|
|
GoGC: globalGoGC,
|
|
},
|
|
|
|
RuleFiles: []string{
|
|
filepath.FromSlash("testdata/first.rules"),
|
|
filepath.FromSlash("testdata/my/*.rules"),
|
|
},
|
|
|
|
RemoteWriteConfigs: []*RemoteWriteConfig{
|
|
{
|
|
URL: mustParseURL("http://remote1/push"),
|
|
ProtobufMessage: RemoteWriteProtoMsgV1,
|
|
RemoteTimeout: model.Duration(30 * time.Second),
|
|
Name: "drop_expensive",
|
|
WriteRelabelConfigs: []*relabel.Config{
|
|
{
|
|
SourceLabels: model.LabelNames{"__name__"},
|
|
Separator: ";",
|
|
Regex: relabel.MustNewRegexp("expensive.*"),
|
|
Replacement: "$1",
|
|
Action: relabel.Drop,
|
|
},
|
|
},
|
|
QueueConfig: DefaultQueueConfig,
|
|
MetadataConfig: DefaultMetadataConfig,
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
OAuth2: &config.OAuth2{
|
|
ClientID: "123",
|
|
ClientSecret: "456",
|
|
TokenURL: "http://remote1/auth",
|
|
TLSConfig: config.TLSConfig{
|
|
CertFile: filepath.FromSlash("testdata/valid_cert_file"),
|
|
KeyFile: filepath.FromSlash("testdata/valid_key_file"),
|
|
},
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
},
|
|
{
|
|
URL: mustParseURL("http://remote2/push"),
|
|
ProtobufMessage: RemoteWriteProtoMsgV2,
|
|
RemoteTimeout: model.Duration(30 * time.Second),
|
|
QueueConfig: DefaultQueueConfig,
|
|
MetadataConfig: DefaultMetadataConfig,
|
|
Name: "rw_tls",
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
TLSConfig: config.TLSConfig{
|
|
CertFile: filepath.FromSlash("testdata/valid_cert_file"),
|
|
KeyFile: filepath.FromSlash("testdata/valid_key_file"),
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
Headers: map[string]string{"name": "value"},
|
|
},
|
|
},
|
|
|
|
OTLPConfig: OTLPConfig{
|
|
PromoteResourceAttributes: []string{
|
|
"k8s.cluster.name", "k8s.job.name", "k8s.namespace.name",
|
|
},
|
|
},
|
|
|
|
RemoteReadConfigs: []*RemoteReadConfig{
|
|
{
|
|
URL: mustParseURL("http://remote1/read"),
|
|
RemoteTimeout: model.Duration(1 * time.Minute),
|
|
ReadRecent: true,
|
|
Name: "default",
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
FollowRedirects: true,
|
|
EnableHTTP2: false,
|
|
},
|
|
FilterExternalLabels: true,
|
|
},
|
|
{
|
|
URL: mustParseURL("http://remote3/read"),
|
|
RemoteTimeout: model.Duration(1 * time.Minute),
|
|
ReadRecent: false,
|
|
Name: "read_special",
|
|
RequiredMatchers: model.LabelSet{"job": "special"},
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
TLSConfig: config.TLSConfig{
|
|
CertFile: filepath.FromSlash("testdata/valid_cert_file"),
|
|
KeyFile: filepath.FromSlash("testdata/valid_key_file"),
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
FilterExternalLabels: true,
|
|
},
|
|
},
|
|
|
|
ScrapeConfigs: []*ScrapeConfig{
|
|
{
|
|
JobName: "prometheus",
|
|
|
|
HonorLabels: true,
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
Authorization: &config.Authorization{
|
|
Type: "Bearer",
|
|
CredentialsFile: filepath.FromSlash("testdata/valid_token_file"),
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
TLSConfig: config.TLSConfig{
|
|
MinVersion: config.TLSVersion(tls.VersionTLS10),
|
|
},
|
|
},
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&file.SDConfig{
|
|
Files: []string{"testdata/foo/*.slow.json", "testdata/foo/*.slow.yml", "testdata/single/file.yml"},
|
|
RefreshInterval: model.Duration(10 * time.Minute),
|
|
},
|
|
&file.SDConfig{
|
|
Files: []string{"testdata/bar/*.yaml"},
|
|
RefreshInterval: model.Duration(5 * time.Minute),
|
|
},
|
|
discovery.StaticConfig{
|
|
{
|
|
Targets: []model.LabelSet{
|
|
{model.AddressLabel: "localhost:9090"},
|
|
{model.AddressLabel: "localhost:9191"},
|
|
},
|
|
Labels: model.LabelSet{
|
|
"my": "label",
|
|
"your": "label",
|
|
},
|
|
Source: "0",
|
|
},
|
|
},
|
|
},
|
|
|
|
RelabelConfigs: []*relabel.Config{
|
|
{
|
|
SourceLabels: model.LabelNames{"job", "__meta_dns_name"},
|
|
TargetLabel: "job",
|
|
Separator: ";",
|
|
Regex: relabel.MustNewRegexp("(.*)some-[regex]"),
|
|
Replacement: "foo-${1}",
|
|
Action: relabel.Replace,
|
|
},
|
|
{
|
|
SourceLabels: model.LabelNames{"abc"},
|
|
TargetLabel: "cde",
|
|
Separator: ";",
|
|
Regex: relabel.DefaultRelabelConfig.Regex,
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
Action: relabel.Replace,
|
|
},
|
|
{
|
|
TargetLabel: "abc",
|
|
Separator: ";",
|
|
Regex: relabel.DefaultRelabelConfig.Regex,
|
|
Replacement: "static",
|
|
Action: relabel.Replace,
|
|
},
|
|
{
|
|
TargetLabel: "abc",
|
|
Separator: ";",
|
|
Regex: relabel.MustNewRegexp(""),
|
|
Replacement: "static",
|
|
Action: relabel.Replace,
|
|
},
|
|
{
|
|
SourceLabels: model.LabelNames{"foo"},
|
|
TargetLabel: "abc",
|
|
Action: relabel.KeepEqual,
|
|
Regex: relabel.DefaultRelabelConfig.Regex,
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
Separator: relabel.DefaultRelabelConfig.Separator,
|
|
},
|
|
{
|
|
SourceLabels: model.LabelNames{"foo"},
|
|
TargetLabel: "abc",
|
|
Action: relabel.DropEqual,
|
|
Regex: relabel.DefaultRelabelConfig.Regex,
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
Separator: relabel.DefaultRelabelConfig.Separator,
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-x",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(50 * time.Second),
|
|
ScrapeTimeout: model.Duration(5 * time.Second),
|
|
EnableCompression: true,
|
|
BodySizeLimit: 10 * units.MiB,
|
|
SampleLimit: 1000,
|
|
TargetLimit: 35,
|
|
LabelLimit: 35,
|
|
LabelNameLengthLimit: 210,
|
|
LabelValueLengthLimit: 210,
|
|
ScrapeProtocols: []ScrapeProtocol{PrometheusText0_0_4},
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
BasicAuth: &config.BasicAuth{
|
|
Username: "admin_name",
|
|
Password: "multiline\nmysecret\ntest",
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
MetricsPath: "/my_path",
|
|
Scheme: "https",
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&dns.SDConfig{
|
|
Names: []string{
|
|
"first.dns.address.domain.com",
|
|
"second.dns.address.domain.com",
|
|
},
|
|
RefreshInterval: model.Duration(15 * time.Second),
|
|
Type: "SRV",
|
|
},
|
|
&dns.SDConfig{
|
|
Names: []string{
|
|
"first.dns.address.domain.com",
|
|
},
|
|
RefreshInterval: model.Duration(30 * time.Second),
|
|
Type: "SRV",
|
|
},
|
|
},
|
|
|
|
RelabelConfigs: []*relabel.Config{
|
|
{
|
|
SourceLabels: model.LabelNames{"job"},
|
|
Regex: relabel.MustNewRegexp("(.*)some-[regex]"),
|
|
Separator: ";",
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
Action: relabel.Drop,
|
|
},
|
|
{
|
|
SourceLabels: model.LabelNames{"__address__"},
|
|
TargetLabel: "__tmp_hash",
|
|
Regex: relabel.DefaultRelabelConfig.Regex,
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
Modulus: 8,
|
|
Separator: ";",
|
|
Action: relabel.HashMod,
|
|
},
|
|
{
|
|
SourceLabels: model.LabelNames{"__tmp_hash"},
|
|
Regex: relabel.MustNewRegexp("1"),
|
|
Separator: ";",
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
Action: relabel.Keep,
|
|
},
|
|
{
|
|
Regex: relabel.MustNewRegexp("1"),
|
|
Separator: ";",
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
Action: relabel.LabelMap,
|
|
},
|
|
{
|
|
Regex: relabel.MustNewRegexp("d"),
|
|
Separator: ";",
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
Action: relabel.LabelDrop,
|
|
},
|
|
{
|
|
Regex: relabel.MustNewRegexp("k"),
|
|
Separator: ";",
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
Action: relabel.LabelKeep,
|
|
},
|
|
},
|
|
MetricRelabelConfigs: []*relabel.Config{
|
|
{
|
|
SourceLabels: model.LabelNames{"__name__"},
|
|
Regex: relabel.MustNewRegexp("expensive_metric.*"),
|
|
Separator: ";",
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
Action: relabel.Drop,
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-y",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&consul.SDConfig{
|
|
Server: "localhost:1234",
|
|
PathPrefix: "/consul",
|
|
Token: "mysecret",
|
|
Services: []string{"nginx", "cache", "mysql"},
|
|
ServiceTags: []string{"canary", "v1"},
|
|
NodeMeta: map[string]string{"rack": "123"},
|
|
TagSeparator: consul.DefaultSDConfig.TagSeparator,
|
|
Scheme: "https",
|
|
RefreshInterval: consul.DefaultSDConfig.RefreshInterval,
|
|
AllowStale: true,
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
TLSConfig: config.TLSConfig{
|
|
CertFile: filepath.FromSlash("testdata/valid_cert_file"),
|
|
KeyFile: filepath.FromSlash("testdata/valid_key_file"),
|
|
CAFile: filepath.FromSlash("testdata/valid_ca_file"),
|
|
InsecureSkipVerify: false,
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
},
|
|
},
|
|
|
|
RelabelConfigs: []*relabel.Config{
|
|
{
|
|
SourceLabels: model.LabelNames{"__meta_sd_consul_tags"},
|
|
Regex: relabel.MustNewRegexp("label:([^=]+)=([^,]+)"),
|
|
Separator: ",",
|
|
TargetLabel: "${1}",
|
|
Replacement: "${2}",
|
|
Action: relabel.Replace,
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-z",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: model.Duration(10 * time.Second),
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: "/metrics",
|
|
Scheme: "http",
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
TLSConfig: config.TLSConfig{
|
|
CertFile: filepath.FromSlash("testdata/valid_cert_file"),
|
|
KeyFile: filepath.FromSlash("testdata/valid_key_file"),
|
|
},
|
|
|
|
Authorization: &config.Authorization{
|
|
Type: "Bearer",
|
|
Credentials: "mysecret",
|
|
},
|
|
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-kubernetes",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&kubernetes.SDConfig{
|
|
APIServer: kubernetesSDHostURL(),
|
|
Role: kubernetes.RoleEndpoint,
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
BasicAuth: &config.BasicAuth{
|
|
Username: "myusername",
|
|
Password: "mysecret",
|
|
},
|
|
TLSConfig: config.TLSConfig{
|
|
CertFile: filepath.FromSlash("testdata/valid_cert_file"),
|
|
KeyFile: filepath.FromSlash("testdata/valid_key_file"),
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
NamespaceDiscovery: kubernetes.NamespaceDiscovery{},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-kubernetes-namespaces",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
BasicAuth: &config.BasicAuth{
|
|
Username: "myusername",
|
|
PasswordFile: filepath.FromSlash("testdata/valid_password_file"),
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&kubernetes.SDConfig{
|
|
APIServer: kubernetesSDHostURL(),
|
|
Role: kubernetes.RoleEndpoint,
|
|
NamespaceDiscovery: kubernetes.NamespaceDiscovery{
|
|
Names: []string{
|
|
"default",
|
|
},
|
|
},
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-kuma",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&xds.KumaSDConfig{
|
|
Server: "http://kuma-control-plane.kuma-system.svc:5676",
|
|
ClientID: "main-prometheus",
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
RefreshInterval: model.Duration(15 * time.Second),
|
|
FetchTimeout: model.Duration(2 * time.Minute),
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-marathon",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&marathon.SDConfig{
|
|
Servers: []string{
|
|
"https://marathon.example.com:443",
|
|
},
|
|
RefreshInterval: model.Duration(30 * time.Second),
|
|
AuthToken: "mysecret",
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
TLSConfig: config.TLSConfig{
|
|
CertFile: filepath.FromSlash("testdata/valid_cert_file"),
|
|
KeyFile: filepath.FromSlash("testdata/valid_key_file"),
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-nomad",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&nomad.SDConfig{
|
|
AllowStale: true,
|
|
Namespace: "default",
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
Region: "global",
|
|
Server: "http://localhost:4646",
|
|
TagSeparator: ",",
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-ec2",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&aws.EC2SDConfig{
|
|
Region: "us-east-1",
|
|
AccessKey: "access",
|
|
SecretKey: "mysecret",
|
|
Profile: "profile",
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
Port: 80,
|
|
Filters: []*aws.EC2Filter{
|
|
{
|
|
Name: "tag:environment",
|
|
Values: []string{"prod"},
|
|
},
|
|
{
|
|
Name: "tag:service",
|
|
Values: []string{"web", "db"},
|
|
},
|
|
},
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-lightsail",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&aws.LightsailSDConfig{
|
|
Region: "us-east-1",
|
|
AccessKey: "access",
|
|
SecretKey: "mysecret",
|
|
Profile: "profile",
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
Port: 80,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-azure",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&azure.SDConfig{
|
|
Environment: "AzurePublicCloud",
|
|
SubscriptionID: "11AAAA11-A11A-111A-A111-1111A1111A11",
|
|
ResourceGroup: "my-resource-group",
|
|
TenantID: "BBBB222B-B2B2-2B22-B222-2BB2222BB2B2",
|
|
ClientID: "333333CC-3C33-3333-CCC3-33C3CCCCC33C",
|
|
ClientSecret: "mysecret",
|
|
AuthenticationMethod: "OAuth",
|
|
RefreshInterval: model.Duration(5 * time.Minute),
|
|
Port: 9100,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-nerve",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&zookeeper.NerveSDConfig{
|
|
Servers: []string{"localhost"},
|
|
Paths: []string{"/monitoring"},
|
|
Timeout: model.Duration(10 * time.Second),
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "0123service-xxx",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
discovery.StaticConfig{
|
|
{
|
|
Targets: []model.LabelSet{
|
|
{model.AddressLabel: "localhost:9090"},
|
|
},
|
|
Source: "0",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "badfederation",
|
|
|
|
HonorTimestamps: false,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: "/federate",
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
discovery.StaticConfig{
|
|
{
|
|
Targets: []model.LabelSet{
|
|
{model.AddressLabel: "localhost:9090"},
|
|
},
|
|
Source: "0",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "測試",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
discovery.StaticConfig{
|
|
{
|
|
Targets: []model.LabelSet{
|
|
{model.AddressLabel: "localhost:9090"},
|
|
},
|
|
Source: "0",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "httpsd",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&http.SDConfig{
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
URL: "http://example.com/prometheus",
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-triton",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&triton.SDConfig{
|
|
Account: "testAccount",
|
|
Role: "container",
|
|
DNSSuffix: "triton.example.com",
|
|
Endpoint: "triton.example.com",
|
|
Port: 9163,
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
Version: 1,
|
|
TLSConfig: config.TLSConfig{
|
|
CertFile: "testdata/valid_cert_file",
|
|
KeyFile: "testdata/valid_key_file",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "digitalocean-droplets",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&digitalocean.SDConfig{
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
Authorization: &config.Authorization{
|
|
Type: "Bearer",
|
|
Credentials: "abcdef",
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
Port: 80,
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "docker",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&moby.DockerSDConfig{
|
|
Filters: []moby.Filter{},
|
|
Host: "unix:///var/run/docker.sock",
|
|
Port: 80,
|
|
HostNetworkingHost: "localhost",
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
MatchFirstNetwork: true,
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "dockerswarm",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&moby.DockerSwarmSDConfig{
|
|
Filters: []moby.Filter{},
|
|
Host: "http://127.0.0.1:2375",
|
|
Role: "nodes",
|
|
Port: 80,
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-openstack",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&openstack.SDConfig{
|
|
Role: "instance",
|
|
Region: "RegionOne",
|
|
Port: 80,
|
|
Availability: "public",
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
TLSConfig: config.TLSConfig{
|
|
CAFile: "testdata/valid_ca_file",
|
|
CertFile: "testdata/valid_cert_file",
|
|
KeyFile: "testdata/valid_key_file",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-puppetdb",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&puppetdb.SDConfig{
|
|
URL: "https://puppetserver/",
|
|
Query: "resources { type = \"Package\" and title = \"httpd\" }",
|
|
IncludeParameters: true,
|
|
Port: 80,
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
TLSConfig: config.TLSConfig{
|
|
CAFile: "testdata/valid_ca_file",
|
|
CertFile: "testdata/valid_cert_file",
|
|
KeyFile: "testdata/valid_key_file",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "hetzner",
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
RelabelConfigs: []*relabel.Config{
|
|
{
|
|
Action: relabel.Uppercase,
|
|
Regex: relabel.DefaultRelabelConfig.Regex,
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
Separator: relabel.DefaultRelabelConfig.Separator,
|
|
SourceLabels: model.LabelNames{"instance"},
|
|
TargetLabel: "instance",
|
|
},
|
|
},
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&hetzner.SDConfig{
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
Authorization: &config.Authorization{
|
|
Type: "Bearer",
|
|
Credentials: "abcdef",
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
Port: 80,
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
Role: "hcloud",
|
|
},
|
|
&hetzner.SDConfig{
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
BasicAuth: &config.BasicAuth{Username: "abcdef", Password: "abcdef"},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
Port: 80,
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
Role: "robot",
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "service-eureka",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&eureka.SDConfig{
|
|
Server: "http://eureka.example.com:8761/eureka",
|
|
RefreshInterval: model.Duration(30 * time.Second),
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "ovhcloud",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&ovhcloud.SDConfig{
|
|
Endpoint: "ovh-eu",
|
|
ApplicationKey: "testAppKey",
|
|
ApplicationSecret: "testAppSecret",
|
|
ConsumerKey: "testConsumerKey",
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
Service: "vps",
|
|
},
|
|
&ovhcloud.SDConfig{
|
|
Endpoint: "ovh-eu",
|
|
ApplicationKey: "testAppKey",
|
|
ApplicationSecret: "testAppSecret",
|
|
ConsumerKey: "testConsumerKey",
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
Service: "dedicated_server",
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "scaleway",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&scaleway.SDConfig{
|
|
APIURL: "https://api.scaleway.com",
|
|
AccessKey: "SCWXXXXXXXXXXXXXXXXX",
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
Port: 80,
|
|
Project: "11111111-1111-1111-1111-111111111112",
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
Role: "instance",
|
|
SecretKey: "11111111-1111-1111-1111-111111111111",
|
|
Zone: "fr-par-1",
|
|
},
|
|
&scaleway.SDConfig{
|
|
APIURL: "https://api.scaleway.com",
|
|
AccessKey: "SCWXXXXXXXXXXXXXXXXX",
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
Port: 80,
|
|
Project: "11111111-1111-1111-1111-111111111112",
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
Role: "baremetal",
|
|
SecretKey: "11111111-1111-1111-1111-111111111111",
|
|
Zone: "fr-par-1",
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "linode-instances",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&linode.SDConfig{
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
Authorization: &config.Authorization{
|
|
Type: "Bearer",
|
|
Credentials: "abcdef",
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
Port: 80,
|
|
TagSeparator: linode.DefaultSDConfig.TagSeparator,
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "uyuni",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&uyuni.SDConfig{
|
|
Server: "https://localhost:1234",
|
|
Username: "gopher",
|
|
Password: "hole",
|
|
Entitlement: "monitoring_entitled",
|
|
Separator: ",",
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "ionos",
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&ionos.SDConfig{
|
|
DatacenterID: "8feda53f-15f0-447f-badf-ebe32dad2fc0",
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
Authorization: &config.Authorization{Type: "Bearer", Credentials: "abcdef"},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
Port: 80,
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "vultr",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
EnableCompression: true,
|
|
BodySizeLimit: globBodySizeLimit,
|
|
SampleLimit: globSampleLimit,
|
|
TargetLimit: globTargetLimit,
|
|
LabelLimit: globLabelLimit,
|
|
LabelNameLengthLimit: globLabelNameLengthLimit,
|
|
LabelValueLengthLimit: globLabelValueLengthLimit,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&vultr.SDConfig{
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
Authorization: &config.Authorization{
|
|
Type: "Bearer",
|
|
Credentials: "abcdef",
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
Port: 80,
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
},
|
|
},
|
|
},
|
|
},
|
|
AlertingConfig: AlertingConfig{
|
|
AlertmanagerConfigs: []*AlertmanagerConfig{
|
|
{
|
|
Scheme: "https",
|
|
Timeout: model.Duration(10 * time.Second),
|
|
APIVersion: AlertmanagerAPIVersionV2,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
discovery.StaticConfig{
|
|
{
|
|
Targets: []model.LabelSet{
|
|
{model.AddressLabel: "1.2.3.4:9093"},
|
|
{model.AddressLabel: "1.2.3.5:9093"},
|
|
{model.AddressLabel: "1.2.3.6:9093"},
|
|
},
|
|
Source: "0",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
StorageConfig: StorageConfig{
|
|
TSDBConfig: &TSDBConfig{
|
|
OutOfOrderTimeWindow: 30 * time.Minute.Milliseconds(),
|
|
OutOfOrderTimeWindowFlag: model.Duration(30 * time.Minute),
|
|
},
|
|
},
|
|
TracingConfig: TracingConfig{
|
|
Endpoint: "localhost:4317",
|
|
ClientType: TracingClientGRPC,
|
|
Insecure: false,
|
|
Compression: "gzip",
|
|
Timeout: model.Duration(5 * time.Second),
|
|
Headers: map[string]string{"foo": "bar"},
|
|
TLSConfig: config.TLSConfig{
|
|
CertFile: "testdata/valid_cert_file",
|
|
KeyFile: "testdata/valid_key_file",
|
|
InsecureSkipVerify: true,
|
|
},
|
|
},
|
|
}
|
|
|
|
func TestYAMLRoundtrip(t *testing.T) {
|
|
want, err := LoadFile("testdata/roundtrip.good.yml", false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
|
|
out, err := yaml.Marshal(want)
|
|
|
|
require.NoError(t, err)
|
|
got := &Config{}
|
|
require.NoError(t, yaml.UnmarshalStrict(out, got))
|
|
|
|
require.Equal(t, want, got)
|
|
}
|
|
|
|
func TestRemoteWriteRetryOnRateLimit(t *testing.T) {
|
|
want, err := LoadFile("testdata/remote_write_retry_on_rate_limit.good.yml", false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
|
|
out, err := yaml.Marshal(want)
|
|
|
|
require.NoError(t, err)
|
|
got := &Config{}
|
|
require.NoError(t, yaml.UnmarshalStrict(out, got))
|
|
|
|
require.True(t, got.RemoteWriteConfigs[0].QueueConfig.RetryOnRateLimit)
|
|
require.False(t, got.RemoteWriteConfigs[1].QueueConfig.RetryOnRateLimit)
|
|
}
|
|
|
|
func TestOTLPSanitizeResourceAttributes(t *testing.T) {
|
|
t.Run("good config", func(t *testing.T) {
|
|
want, err := LoadFile(filepath.Join("testdata", "otlp_sanitize_resource_attributes.good.yml"), false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
|
|
out, err := yaml.Marshal(want)
|
|
require.NoError(t, err)
|
|
var got Config
|
|
require.NoError(t, yaml.UnmarshalStrict(out, &got))
|
|
|
|
require.Equal(t, []string{"k8s.cluster.name", "k8s.job.name", "k8s.namespace.name"}, got.OTLPConfig.PromoteResourceAttributes)
|
|
})
|
|
|
|
t.Run("bad config", func(t *testing.T) {
|
|
_, err := LoadFile(filepath.Join("testdata", "otlp_sanitize_resource_attributes.bad.yml"), false, false, log.NewNopLogger())
|
|
require.ErrorContains(t, err, `duplicated promoted OTel resource attribute "k8s.job.name"`)
|
|
require.ErrorContains(t, err, `empty promoted OTel resource attribute`)
|
|
})
|
|
}
|
|
|
|
func TestLoadConfig(t *testing.T) {
|
|
// Parse a valid file that sets a global scrape timeout. This tests whether parsing
|
|
// an overwritten default field in the global config permanently changes the default.
|
|
_, err := LoadFile("testdata/global_timeout.good.yml", false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
|
|
c, err := LoadFile("testdata/conf.good.yml", false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
require.Equal(t, expectedConf, c)
|
|
}
|
|
|
|
func TestScrapeIntervalLarger(t *testing.T) {
|
|
c, err := LoadFile("testdata/scrape_interval_larger.good.yml", false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
require.Len(t, c.ScrapeConfigs, 1)
|
|
for _, sc := range c.ScrapeConfigs {
|
|
require.GreaterOrEqual(t, sc.ScrapeInterval, sc.ScrapeTimeout)
|
|
}
|
|
}
|
|
|
|
// YAML marshaling must not reveal authentication credentials.
|
|
func TestElideSecrets(t *testing.T) {
|
|
c, err := LoadFile("testdata/conf.good.yml", false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
|
|
secretRe := regexp.MustCompile(`\\u003csecret\\u003e|<secret>`)
|
|
|
|
config, err := yaml.Marshal(c)
|
|
require.NoError(t, err)
|
|
yamlConfig := string(config)
|
|
|
|
matches := secretRe.FindAllStringIndex(yamlConfig, -1)
|
|
require.Len(t, matches, 22, "wrong number of secret matches found")
|
|
require.NotContains(t, yamlConfig, "mysecret",
|
|
"yaml marshal reveals authentication credentials.")
|
|
}
|
|
|
|
func TestLoadConfigRuleFilesAbsolutePath(t *testing.T) {
|
|
// Parse a valid file that sets a rule files with an absolute path
|
|
c, err := LoadFile(ruleFilesConfigFile, false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
require.Equal(t, ruleFilesExpectedConf, c)
|
|
}
|
|
|
|
func TestKubernetesEmptyAPIServer(t *testing.T) {
|
|
_, err := LoadFile("testdata/kubernetes_empty_apiserver.good.yml", false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
}
|
|
|
|
func TestKubernetesWithKubeConfig(t *testing.T) {
|
|
_, err := LoadFile("testdata/kubernetes_kubeconfig_without_apiserver.good.yml", false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
}
|
|
|
|
func TestKubernetesSelectors(t *testing.T) {
|
|
_, err := LoadFile("testdata/kubernetes_selectors_endpoints.good.yml", false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
_, err = LoadFile("testdata/kubernetes_selectors_node.good.yml", false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
_, err = LoadFile("testdata/kubernetes_selectors_ingress.good.yml", false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
_, err = LoadFile("testdata/kubernetes_selectors_pod.good.yml", false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
_, err = LoadFile("testdata/kubernetes_selectors_service.good.yml", false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
}
|
|
|
|
var expectedErrors = []struct {
|
|
filename string
|
|
errMsg string
|
|
}{
|
|
{
|
|
filename: "jobname.bad.yml",
|
|
errMsg: `job_name is empty`,
|
|
},
|
|
{
|
|
filename: "jobname_dup.bad.yml",
|
|
errMsg: `found multiple scrape configs with job name "prometheus"`,
|
|
},
|
|
{
|
|
filename: "scrape_interval.bad.yml",
|
|
errMsg: `scrape timeout greater than scrape interval`,
|
|
},
|
|
{
|
|
filename: "labelname.bad.yml",
|
|
errMsg: `"not$allowed" is not a valid label name`,
|
|
},
|
|
{
|
|
filename: "labelname2.bad.yml",
|
|
errMsg: `"not:allowed" is not a valid label name`,
|
|
},
|
|
{
|
|
filename: "labelvalue.bad.yml",
|
|
errMsg: `"\xff" is not a valid label value`,
|
|
},
|
|
{
|
|
filename: "regex.bad.yml",
|
|
errMsg: "error parsing regexp",
|
|
},
|
|
{
|
|
filename: "modulus_missing.bad.yml",
|
|
errMsg: "relabel configuration for hashmod requires non-zero modulus",
|
|
},
|
|
{
|
|
filename: "labelkeep.bad.yml",
|
|
errMsg: "labelkeep action requires only 'regex', and no other fields",
|
|
},
|
|
{
|
|
filename: "labelkeep2.bad.yml",
|
|
errMsg: "labelkeep action requires only 'regex', and no other fields",
|
|
},
|
|
{
|
|
filename: "labelkeep3.bad.yml",
|
|
errMsg: "labelkeep action requires only 'regex', and no other fields",
|
|
},
|
|
{
|
|
filename: "labelkeep4.bad.yml",
|
|
errMsg: "labelkeep action requires only 'regex', and no other fields",
|
|
},
|
|
{
|
|
filename: "labelkeep5.bad.yml",
|
|
errMsg: "labelkeep action requires only 'regex', and no other fields",
|
|
},
|
|
{
|
|
filename: "labeldrop.bad.yml",
|
|
errMsg: "labeldrop action requires only 'regex', and no other fields",
|
|
},
|
|
{
|
|
filename: "labeldrop2.bad.yml",
|
|
errMsg: "labeldrop action requires only 'regex', and no other fields",
|
|
},
|
|
{
|
|
filename: "labeldrop3.bad.yml",
|
|
errMsg: "labeldrop action requires only 'regex', and no other fields",
|
|
},
|
|
{
|
|
filename: "labeldrop4.bad.yml",
|
|
errMsg: "labeldrop action requires only 'regex', and no other fields",
|
|
},
|
|
{
|
|
filename: "labeldrop5.bad.yml",
|
|
errMsg: "labeldrop action requires only 'regex', and no other fields",
|
|
},
|
|
{
|
|
filename: "dropequal.bad.yml",
|
|
errMsg: "relabel configuration for dropequal action requires 'target_label' value",
|
|
},
|
|
{
|
|
filename: "dropequal1.bad.yml",
|
|
errMsg: "dropequal action requires only 'source_labels' and `target_label`, and no other fields",
|
|
},
|
|
{
|
|
filename: "keepequal.bad.yml",
|
|
errMsg: "relabel configuration for keepequal action requires 'target_label' value",
|
|
},
|
|
{
|
|
filename: "keepequal1.bad.yml",
|
|
errMsg: "keepequal action requires only 'source_labels' and `target_label`, and no other fields",
|
|
},
|
|
{
|
|
filename: "labelmap.bad.yml",
|
|
errMsg: "\"l-$1\" is invalid 'replacement' for labelmap action",
|
|
},
|
|
{
|
|
filename: "lowercase.bad.yml",
|
|
errMsg: "relabel configuration for lowercase action requires 'target_label' value",
|
|
},
|
|
{
|
|
filename: "lowercase2.bad.yml",
|
|
errMsg: "\"42lab\" is invalid 'target_label' for lowercase action",
|
|
},
|
|
{
|
|
filename: "lowercase3.bad.yml",
|
|
errMsg: "'replacement' can not be set for lowercase action",
|
|
},
|
|
{
|
|
filename: "uppercase.bad.yml",
|
|
errMsg: "relabel configuration for uppercase action requires 'target_label' value",
|
|
},
|
|
{
|
|
filename: "uppercase2.bad.yml",
|
|
errMsg: "\"42lab\" is invalid 'target_label' for uppercase action",
|
|
},
|
|
{
|
|
filename: "uppercase3.bad.yml",
|
|
errMsg: "'replacement' can not be set for uppercase action",
|
|
},
|
|
{
|
|
filename: "rules.bad.yml",
|
|
errMsg: "invalid rule file path",
|
|
},
|
|
{
|
|
filename: "unknown_attr.bad.yml",
|
|
errMsg: "field consult_sd_configs not found in type",
|
|
},
|
|
{
|
|
filename: "bearertoken.bad.yml",
|
|
errMsg: "at most one of bearer_token & bearer_token_file must be configured",
|
|
},
|
|
{
|
|
filename: "bearertoken_basicauth.bad.yml",
|
|
errMsg: "at most one of basic_auth, oauth2, bearer_token & bearer_token_file must be configured",
|
|
},
|
|
{
|
|
filename: "kubernetes_http_config_without_api_server.bad.yml",
|
|
errMsg: "to use custom HTTP client configuration please provide the 'api_server' URL explicitly",
|
|
},
|
|
{
|
|
filename: "kubernetes_kubeconfig_with_own_namespace.bad.yml",
|
|
errMsg: "cannot use 'kubeconfig_file' and 'namespaces.own_namespace' simultaneously",
|
|
},
|
|
{
|
|
filename: "kubernetes_api_server_with_own_namespace.bad.yml",
|
|
errMsg: "cannot use 'api_server' and 'namespaces.own_namespace' simultaneously",
|
|
},
|
|
{
|
|
filename: "kubernetes_kubeconfig_with_apiserver.bad.yml",
|
|
errMsg: "cannot use 'kubeconfig_file' and 'api_server' simultaneously",
|
|
},
|
|
{
|
|
filename: "kubernetes_kubeconfig_with_http_config.bad.yml",
|
|
errMsg: "cannot use a custom HTTP client configuration together with 'kubeconfig_file'",
|
|
},
|
|
{
|
|
filename: "kubernetes_bearertoken.bad.yml",
|
|
errMsg: "at most one of bearer_token & bearer_token_file must be configured",
|
|
},
|
|
{
|
|
filename: "kubernetes_role.bad.yml",
|
|
errMsg: "role",
|
|
},
|
|
{
|
|
filename: "kubernetes_selectors_endpoints.bad.yml",
|
|
errMsg: "endpoints role supports only pod, service, endpoints selectors",
|
|
},
|
|
{
|
|
filename: "kubernetes_selectors_ingress.bad.yml",
|
|
errMsg: "ingress role supports only ingress selectors",
|
|
},
|
|
{
|
|
filename: "kubernetes_selectors_node.bad.yml",
|
|
errMsg: "node role supports only node selectors",
|
|
},
|
|
{
|
|
filename: "kubernetes_selectors_pod.bad.yml",
|
|
errMsg: "pod role supports only pod selectors",
|
|
},
|
|
{
|
|
filename: "kubernetes_selectors_service.bad.yml",
|
|
errMsg: "service role supports only service selectors",
|
|
},
|
|
{
|
|
filename: "kubernetes_namespace_discovery.bad.yml",
|
|
errMsg: "field foo not found in type kubernetes.plain",
|
|
},
|
|
{
|
|
filename: "kubernetes_selectors_duplicated_role.bad.yml",
|
|
errMsg: "duplicated selector role: pod",
|
|
},
|
|
{
|
|
filename: "kubernetes_selectors_incorrect_selector.bad.yml",
|
|
errMsg: "invalid selector: 'metadata.status-Running'; can't understand 'metadata.status-Running'",
|
|
},
|
|
{
|
|
filename: "kubernetes_bearertoken_basicauth.bad.yml",
|
|
errMsg: "at most one of basic_auth, oauth2, bearer_token & bearer_token_file must be configured",
|
|
},
|
|
{
|
|
filename: "kubernetes_authorization_basicauth.bad.yml",
|
|
errMsg: "at most one of basic_auth, oauth2 & authorization must be configured",
|
|
},
|
|
{
|
|
filename: "marathon_no_servers.bad.yml",
|
|
errMsg: "marathon_sd: must contain at least one Marathon server",
|
|
},
|
|
{
|
|
filename: "marathon_authtoken_authtokenfile.bad.yml",
|
|
errMsg: "marathon_sd: at most one of auth_token & auth_token_file must be configured",
|
|
},
|
|
{
|
|
filename: "marathon_authtoken_basicauth.bad.yml",
|
|
errMsg: "marathon_sd: at most one of basic_auth, auth_token & auth_token_file must be configured",
|
|
},
|
|
{
|
|
filename: "marathon_authtoken_bearertoken.bad.yml",
|
|
errMsg: "marathon_sd: at most one of bearer_token, bearer_token_file, auth_token & auth_token_file must be configured",
|
|
},
|
|
{
|
|
filename: "marathon_authtoken_authorization.bad.yml",
|
|
errMsg: "marathon_sd: at most one of auth_token, auth_token_file & authorization must be configured",
|
|
},
|
|
{
|
|
filename: "openstack_role.bad.yml",
|
|
errMsg: "unknown OpenStack SD role",
|
|
},
|
|
{
|
|
filename: "openstack_availability.bad.yml",
|
|
errMsg: "unknown availability invalid, must be one of admin, internal or public",
|
|
},
|
|
{
|
|
filename: "url_in_targetgroup.bad.yml",
|
|
errMsg: "\"http://bad\" is not a valid hostname",
|
|
},
|
|
{
|
|
filename: "target_label_missing.bad.yml",
|
|
errMsg: "relabel configuration for replace action requires 'target_label' value",
|
|
},
|
|
{
|
|
filename: "target_label_hashmod_missing.bad.yml",
|
|
errMsg: "relabel configuration for hashmod action requires 'target_label' value",
|
|
},
|
|
{
|
|
filename: "unknown_global_attr.bad.yml",
|
|
errMsg: "field nonexistent_field not found in type config.plain",
|
|
},
|
|
{
|
|
filename: "remote_read_url_missing.bad.yml",
|
|
errMsg: `url for remote_read is empty`,
|
|
},
|
|
{
|
|
filename: "remote_write_header.bad.yml",
|
|
errMsg: `x-prometheus-remote-write-version is a reserved header. It must not be changed`,
|
|
},
|
|
{
|
|
filename: "remote_read_header.bad.yml",
|
|
errMsg: `x-prometheus-remote-write-version is a reserved header. It must not be changed`,
|
|
},
|
|
{
|
|
filename: "remote_write_authorization_header.bad.yml",
|
|
errMsg: `authorization header must be changed via the basic_auth, authorization, oauth2, sigv4, azuread or google_iam parameter`,
|
|
},
|
|
{
|
|
filename: "remote_write_wrong_msg.bad.yml",
|
|
errMsg: `invalid protobuf_message value: unknown remote write protobuf message io.prometheus.writet.v2.Request, supported: prometheus.WriteRequest, io.prometheus.write.v2.Request`,
|
|
},
|
|
{
|
|
filename: "remote_write_url_missing.bad.yml",
|
|
errMsg: `url for remote_write is empty`,
|
|
},
|
|
{
|
|
filename: "remote_write_dup.bad.yml",
|
|
errMsg: `found multiple remote write configs with job name "queue1"`,
|
|
},
|
|
{
|
|
filename: "remote_read_dup.bad.yml",
|
|
errMsg: `found multiple remote read configs with job name "queue1"`,
|
|
},
|
|
{
|
|
filename: "ec2_filters_empty_values.bad.yml",
|
|
errMsg: `EC2 SD configuration filter values cannot be empty`,
|
|
},
|
|
{
|
|
filename: "ec2_token_file.bad.yml",
|
|
errMsg: `at most one of bearer_token & bearer_token_file must be configured`,
|
|
},
|
|
{
|
|
filename: "lightsail_token_file.bad.yml",
|
|
errMsg: `at most one of bearer_token & bearer_token_file must be configured`,
|
|
},
|
|
{
|
|
filename: "section_key_dup.bad.yml",
|
|
errMsg: "field scrape_configs already set in type config.plain",
|
|
},
|
|
{
|
|
filename: "azure_client_id_missing.bad.yml",
|
|
errMsg: "azure SD configuration requires a client_id",
|
|
},
|
|
{
|
|
filename: "azure_client_secret_missing.bad.yml",
|
|
errMsg: "azure SD configuration requires a client_secret",
|
|
},
|
|
{
|
|
filename: "azure_subscription_id_missing.bad.yml",
|
|
errMsg: "azure SD configuration requires a subscription_id",
|
|
},
|
|
{
|
|
filename: "azure_tenant_id_missing.bad.yml",
|
|
errMsg: "azure SD configuration requires a tenant_id",
|
|
},
|
|
{
|
|
filename: "azure_authentication_method.bad.yml",
|
|
errMsg: "unknown authentication_type \"invalid\". Supported types are \"OAuth\", \"ManagedIdentity\" or \"SDK\"",
|
|
},
|
|
{
|
|
filename: "azure_bearertoken_basicauth.bad.yml",
|
|
errMsg: "at most one of basic_auth, oauth2, bearer_token & bearer_token_file must be configured",
|
|
},
|
|
{
|
|
filename: "empty_scrape_config.bad.yml",
|
|
errMsg: "empty or null scrape config section",
|
|
},
|
|
{
|
|
filename: "empty_rw_config.bad.yml",
|
|
errMsg: "empty or null remote write config section",
|
|
},
|
|
{
|
|
filename: "empty_rr_config.bad.yml",
|
|
errMsg: "empty or null remote read config section",
|
|
},
|
|
{
|
|
filename: "empty_target_relabel_config.bad.yml",
|
|
errMsg: "empty or null target relabeling rule",
|
|
},
|
|
{
|
|
filename: "empty_metric_relabel_config.bad.yml",
|
|
errMsg: "empty or null metric relabeling rule",
|
|
},
|
|
{
|
|
filename: "empty_alert_relabel_config.bad.yml",
|
|
errMsg: "empty or null alert relabeling rule",
|
|
},
|
|
{
|
|
filename: "empty_alertmanager_relabel_config.bad.yml",
|
|
errMsg: "empty or null Alertmanager target relabeling rule",
|
|
},
|
|
{
|
|
filename: "empty_rw_relabel_config.bad.yml",
|
|
errMsg: "empty or null relabeling rule in remote write config",
|
|
},
|
|
{
|
|
filename: "empty_static_config.bad.yml",
|
|
errMsg: "empty or null section in static_configs",
|
|
},
|
|
{
|
|
filename: "puppetdb_no_query.bad.yml",
|
|
errMsg: "query missing",
|
|
},
|
|
{
|
|
filename: "puppetdb_no_url.bad.yml",
|
|
errMsg: "URL is missing",
|
|
},
|
|
{
|
|
filename: "puppetdb_bad_url.bad.yml",
|
|
errMsg: "host is missing in URL",
|
|
},
|
|
{
|
|
filename: "puppetdb_no_scheme.bad.yml",
|
|
errMsg: "URL scheme must be 'http' or 'https'",
|
|
},
|
|
{
|
|
filename: "puppetdb_token_file.bad.yml",
|
|
errMsg: "at most one of bearer_token & bearer_token_file must be configured",
|
|
},
|
|
{
|
|
filename: "hetzner_role.bad.yml",
|
|
errMsg: "unknown role",
|
|
},
|
|
{
|
|
filename: "eureka_no_server.bad.yml",
|
|
errMsg: "empty or null eureka server",
|
|
},
|
|
{
|
|
filename: "eureka_invalid_server.bad.yml",
|
|
errMsg: "invalid eureka server URL",
|
|
},
|
|
{
|
|
filename: "scaleway_role.bad.yml",
|
|
errMsg: `unknown role "invalid"`,
|
|
},
|
|
{
|
|
filename: "scaleway_no_secret.bad.yml",
|
|
errMsg: "one of secret_key & secret_key_file must be configured",
|
|
},
|
|
{
|
|
filename: "scaleway_two_secrets.bad.yml",
|
|
errMsg: "at most one of secret_key & secret_key_file must be configured",
|
|
},
|
|
{
|
|
filename: "scrape_body_size_limit.bad.yml",
|
|
errMsg: "units: unknown unit in 100",
|
|
},
|
|
{
|
|
filename: "http_url_no_scheme.bad.yml",
|
|
errMsg: "URL scheme must be 'http' or 'https'",
|
|
},
|
|
{
|
|
filename: "http_url_no_host.bad.yml",
|
|
errMsg: "host is missing in URL",
|
|
},
|
|
{
|
|
filename: "http_token_file.bad.yml",
|
|
errMsg: "at most one of bearer_token & bearer_token_file must be configured",
|
|
},
|
|
{
|
|
filename: "http_url_bad_scheme.bad.yml",
|
|
errMsg: "URL scheme must be 'http' or 'https'",
|
|
},
|
|
{
|
|
filename: "empty_scrape_config_action.bad.yml",
|
|
errMsg: "relabel action cannot be empty",
|
|
},
|
|
{
|
|
filename: "tracing_missing_endpoint.bad.yml",
|
|
errMsg: "tracing endpoint must be set",
|
|
},
|
|
{
|
|
filename: "tracing_invalid_header.bad.yml",
|
|
errMsg: "x-prometheus-remote-write-version is a reserved header. It must not be changed",
|
|
},
|
|
{
|
|
filename: "tracing_invalid_authorization_header.bad.yml",
|
|
errMsg: "authorization header configuration is not yet supported",
|
|
},
|
|
{
|
|
filename: "tracing_invalid_compression.bad.yml",
|
|
errMsg: "invalid compression type foo provided, valid options: gzip",
|
|
},
|
|
{
|
|
filename: "uyuni_no_server.bad.yml",
|
|
errMsg: "Uyuni SD configuration requires server host",
|
|
},
|
|
{
|
|
filename: "uyuni_token_file.bad.yml",
|
|
errMsg: "at most one of bearer_token & bearer_token_file must be configured",
|
|
},
|
|
{
|
|
filename: "ionos_datacenter.bad.yml",
|
|
errMsg: "datacenter id can't be empty",
|
|
},
|
|
{
|
|
filename: "ovhcloud_no_secret.bad.yml",
|
|
errMsg: "application secret can not be empty",
|
|
},
|
|
{
|
|
filename: "ovhcloud_bad_service.bad.yml",
|
|
errMsg: "unknown service: fakeservice",
|
|
},
|
|
{
|
|
filename: "scrape_config_files_glob.bad.yml",
|
|
errMsg: `parsing YAML file testdata/scrape_config_files_glob.bad.yml: invalid scrape config file path "scrape_configs/*/*"`,
|
|
},
|
|
{
|
|
filename: "scrape_config_files_scrape_protocols.bad.yml",
|
|
errMsg: `parsing YAML file testdata/scrape_config_files_scrape_protocols.bad.yml: scrape_protocols: unknown scrape protocol prometheusproto, supported: [OpenMetricsText0.0.1 OpenMetricsText1.0.0 PrometheusProto PrometheusText0.0.4] for scrape config with job name "node"`,
|
|
},
|
|
{
|
|
filename: "scrape_config_files_scrape_protocols2.bad.yml",
|
|
errMsg: `parsing YAML file testdata/scrape_config_files_scrape_protocols2.bad.yml: duplicated protocol in scrape_protocols, got [OpenMetricsText1.0.0 PrometheusProto OpenMetricsText1.0.0] for scrape config with job name "node"`,
|
|
},
|
|
}
|
|
|
|
func TestBadConfigs(t *testing.T) {
|
|
for _, ee := range expectedErrors {
|
|
_, err := LoadFile("testdata/"+ee.filename, false, false, log.NewNopLogger())
|
|
require.Error(t, err, "%s", ee.filename)
|
|
require.Contains(t, err.Error(), ee.errMsg,
|
|
"Expected error for %s to contain %q but got: %s", ee.filename, ee.errMsg, err)
|
|
}
|
|
}
|
|
|
|
func TestBadStaticConfigsJSON(t *testing.T) {
|
|
content, err := os.ReadFile("testdata/static_config.bad.json")
|
|
require.NoError(t, err)
|
|
var tg targetgroup.Group
|
|
err = json.Unmarshal(content, &tg)
|
|
require.Error(t, err)
|
|
}
|
|
|
|
func TestBadStaticConfigsYML(t *testing.T) {
|
|
content, err := os.ReadFile("testdata/static_config.bad.yml")
|
|
require.NoError(t, err)
|
|
var tg targetgroup.Group
|
|
err = yaml.UnmarshalStrict(content, &tg)
|
|
require.Error(t, err)
|
|
}
|
|
|
|
func TestEmptyConfig(t *testing.T) {
|
|
c, err := Load("", false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
exp := DefaultConfig
|
|
require.Equal(t, exp, *c)
|
|
}
|
|
|
|
func TestExpandExternalLabels(t *testing.T) {
|
|
// Cleanup ant TEST env variable that could exist on the system.
|
|
os.Setenv("TEST", "")
|
|
|
|
c, err := LoadFile("testdata/external_labels.good.yml", false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
testutil.RequireEqual(t, labels.FromStrings("bar", "foo", "baz", "foo${TEST}bar", "foo", "${TEST}", "qux", "foo$${TEST}", "xyz", "foo$$bar"), c.GlobalConfig.ExternalLabels)
|
|
|
|
c, err = LoadFile("testdata/external_labels.good.yml", false, true, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
testutil.RequireEqual(t, labels.FromStrings("bar", "foo", "baz", "foobar", "foo", "", "qux", "foo${TEST}", "xyz", "foo$bar"), c.GlobalConfig.ExternalLabels)
|
|
|
|
os.Setenv("TEST", "TestValue")
|
|
c, err = LoadFile("testdata/external_labels.good.yml", false, true, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
testutil.RequireEqual(t, labels.FromStrings("bar", "foo", "baz", "fooTestValuebar", "foo", "TestValue", "qux", "foo${TEST}", "xyz", "foo$bar"), c.GlobalConfig.ExternalLabels)
|
|
}
|
|
|
|
func TestAgentMode(t *testing.T) {
|
|
_, err := LoadFile("testdata/agent_mode.with_alert_manager.yml", true, false, log.NewNopLogger())
|
|
require.ErrorContains(t, err, "field alerting is not allowed in agent mode")
|
|
|
|
_, err = LoadFile("testdata/agent_mode.with_alert_relabels.yml", true, false, log.NewNopLogger())
|
|
require.ErrorContains(t, err, "field alerting is not allowed in agent mode")
|
|
|
|
_, err = LoadFile("testdata/agent_mode.with_rule_files.yml", true, false, log.NewNopLogger())
|
|
require.ErrorContains(t, err, "field rule_files is not allowed in agent mode")
|
|
|
|
_, err = LoadFile("testdata/agent_mode.with_remote_reads.yml", true, false, log.NewNopLogger())
|
|
require.ErrorContains(t, err, "field remote_read is not allowed in agent mode")
|
|
|
|
c, err := LoadFile("testdata/agent_mode.without_remote_writes.yml", true, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
require.Empty(t, c.RemoteWriteConfigs)
|
|
|
|
c, err = LoadFile("testdata/agent_mode.good.yml", true, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
require.Len(t, c.RemoteWriteConfigs, 1)
|
|
require.Equal(
|
|
t,
|
|
"http://remote1/push",
|
|
c.RemoteWriteConfigs[0].URL.String(),
|
|
)
|
|
}
|
|
|
|
func TestEmptyGlobalBlock(t *testing.T) {
|
|
c, err := Load("global:\n", false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
exp := DefaultConfig
|
|
exp.Runtime = DefaultRuntimeConfig
|
|
require.Equal(t, exp, *c)
|
|
}
|
|
|
|
func TestGetScrapeConfigs(t *testing.T) {
|
|
sc := func(jobName string, scrapeInterval, scrapeTimeout model.Duration) *ScrapeConfig {
|
|
return &ScrapeConfig{
|
|
JobName: jobName,
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: scrapeInterval,
|
|
ScrapeTimeout: scrapeTimeout,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: "/metrics",
|
|
Scheme: "http",
|
|
EnableCompression: true,
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
discovery.StaticConfig{
|
|
{
|
|
Targets: []model.LabelSet{
|
|
{
|
|
model.AddressLabel: "localhost:8080",
|
|
},
|
|
},
|
|
Source: "0",
|
|
},
|
|
},
|
|
},
|
|
}
|
|
}
|
|
|
|
testCases := []struct {
|
|
name string
|
|
configFile string
|
|
expectedResult []*ScrapeConfig
|
|
expectedError string
|
|
}{
|
|
{
|
|
name: "An included config file should be a valid global config.",
|
|
configFile: "testdata/scrape_config_files.good.yml",
|
|
expectedResult: []*ScrapeConfig{sc("prometheus", model.Duration(60*time.Second), model.Duration(10*time.Second))},
|
|
},
|
|
{
|
|
name: "An global config that only include a scrape config file.",
|
|
configFile: "testdata/scrape_config_files_only.good.yml",
|
|
expectedResult: []*ScrapeConfig{sc("prometheus", model.Duration(60*time.Second), model.Duration(10*time.Second))},
|
|
},
|
|
{
|
|
name: "An global config that combine scrape config files and scrape configs.",
|
|
configFile: "testdata/scrape_config_files_combined.good.yml",
|
|
expectedResult: []*ScrapeConfig{
|
|
sc("node", model.Duration(60*time.Second), model.Duration(10*time.Second)),
|
|
sc("prometheus", model.Duration(60*time.Second), model.Duration(10*time.Second)),
|
|
sc("alertmanager", model.Duration(60*time.Second), model.Duration(10*time.Second)),
|
|
},
|
|
},
|
|
{
|
|
name: "An global config that includes a scrape config file with globs",
|
|
configFile: "testdata/scrape_config_files_glob.good.yml",
|
|
expectedResult: []*ScrapeConfig{
|
|
{
|
|
JobName: "prometheus",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(60 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
EnableCompression: true,
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
TLSConfig: config.TLSConfig{
|
|
CertFile: filepath.FromSlash("testdata/scrape_configs/valid_cert_file"),
|
|
KeyFile: filepath.FromSlash("testdata/scrape_configs/valid_key_file"),
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
discovery.StaticConfig{
|
|
{
|
|
Targets: []model.LabelSet{
|
|
{model.AddressLabel: "localhost:8080"},
|
|
},
|
|
Source: "0",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
JobName: "node",
|
|
|
|
HonorTimestamps: true,
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
ScrapeProtocols: DefaultGlobalConfig.ScrapeProtocols,
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
TLSConfig: config.TLSConfig{
|
|
CertFile: filepath.FromSlash("testdata/valid_cert_file"),
|
|
KeyFile: filepath.FromSlash("testdata/valid_key_file"),
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
EnableCompression: true,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
&vultr.SDConfig{
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
Authorization: &config.Authorization{
|
|
Type: "Bearer",
|
|
Credentials: "abcdef",
|
|
},
|
|
FollowRedirects: true,
|
|
EnableHTTP2: true,
|
|
},
|
|
Port: 80,
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
{
|
|
name: "An global config that includes twice the same scrape configs.",
|
|
configFile: "testdata/scrape_config_files_double_import.bad.yml",
|
|
expectedError: `found multiple scrape configs with job name "prometheus"`,
|
|
},
|
|
{
|
|
name: "An global config that includes a scrape config identical to a scrape config in the main file.",
|
|
configFile: "testdata/scrape_config_files_duplicate.bad.yml",
|
|
expectedError: `found multiple scrape configs with job name "prometheus"`,
|
|
},
|
|
{
|
|
name: "An global config that includes a scrape config file with errors.",
|
|
configFile: "testdata/scrape_config_files_global.bad.yml",
|
|
expectedError: `scrape timeout greater than scrape interval for scrape config with job name "prometheus"`,
|
|
},
|
|
}
|
|
|
|
for _, tc := range testCases {
|
|
t.Run(tc.name, func(t *testing.T) {
|
|
c, err := LoadFile(tc.configFile, false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
|
|
scfgs, err := c.GetScrapeConfigs()
|
|
if len(tc.expectedError) > 0 {
|
|
require.ErrorContains(t, err, tc.expectedError)
|
|
}
|
|
require.Equal(t, tc.expectedResult, scfgs)
|
|
})
|
|
}
|
|
}
|
|
|
|
func kubernetesSDHostURL() config.URL {
|
|
tURL, _ := url.Parse("https://localhost:1234")
|
|
return config.URL{URL: tURL}
|
|
}
|
|
|
|
func TestScrapeConfigDisableCompression(t *testing.T) {
|
|
want, err := LoadFile("testdata/scrape_config_disable_compression.good.yml", false, false, log.NewNopLogger())
|
|
require.NoError(t, err)
|
|
|
|
out, err := yaml.Marshal(want)
|
|
|
|
require.NoError(t, err)
|
|
got := &Config{}
|
|
require.NoError(t, yaml.UnmarshalStrict(out, got))
|
|
|
|
require.False(t, got.ScrapeConfigs[0].EnableCompression)
|
|
}
|