|
|
|
// Copyright 2015 The Prometheus Authors
|
|
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
// you may not use this file except in compliance with the License.
|
|
|
|
// You may obtain a copy of the License at
|
|
|
|
//
|
|
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
//
|
|
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
// See the License for the specific language governing permissions and
|
|
|
|
// limitations under the License.
|
|
|
|
|
|
|
|
package config
|
|
|
|
|
|
|
|
import (
|
|
|
|
"encoding/json"
|
|
|
|
"io/ioutil"
|
|
|
|
"net/url"
|
|
|
|
"os"
|
|
|
|
"path/filepath"
|
|
|
|
"regexp"
|
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/alecthomas/units"
|
|
|
|
"github.com/go-kit/log"
|
|
|
|
"github.com/prometheus/common/config"
|
|
|
|
"github.com/prometheus/common/model"
|
|
|
|
"github.com/stretchr/testify/require"
|
|
|
|
"gopkg.in/yaml.v2"
|
|
|
|
|
|
|
|
"github.com/prometheus/prometheus/discovery"
|
|
|
|
"github.com/prometheus/prometheus/discovery/aws"
|
Refactor SD configuration to remove `config` dependency (#3629)
* refactor: move targetGroup struct and CheckOverflow() to their own package
* refactor: move auth and security related structs to a utility package, fix import error in utility package
* refactor: Azure SD, remove SD struct from config
* refactor: DNS SD, remove SD struct from config into dns package
* refactor: ec2 SD, move SD struct from config into the ec2 package
* refactor: file SD, move SD struct from config to file discovery package
* refactor: gce, move SD struct from config to gce discovery package
* refactor: move HTTPClientConfig and URL into util/config, fix import error in httputil
* refactor: consul, move SD struct from config into consul discovery package
* refactor: marathon, move SD struct from config into marathon discovery package
* refactor: triton, move SD struct from config to triton discovery package, fix test
* refactor: zookeeper, move SD structs from config to zookeeper discovery package
* refactor: openstack, remove SD struct from config, move into openstack discovery package
* refactor: kubernetes, move SD struct from config into kubernetes discovery package
* refactor: notifier, use targetgroup package instead of config
* refactor: tests for file, marathon, triton SD - use targetgroup package instead of config.TargetGroup
* refactor: retrieval, use targetgroup package instead of config.TargetGroup
* refactor: storage, use config util package
* refactor: discovery manager, use targetgroup package instead of config.TargetGroup
* refactor: use HTTPClient and TLS config from configUtil instead of config
* refactor: tests, use targetgroup package instead of config.TargetGroup
* refactor: fix tagetgroup.Group pointers that were removed by mistake
* refactor: openstack, kubernetes: drop prefixes
* refactor: remove import aliases forced due to vscode bug
* refactor: move main SD struct out of config into discovery/config
* refactor: rename configUtil to config_util
* refactor: rename yamlUtil to yaml_config
* refactor: kubernetes, remove prefixes
* refactor: move the TargetGroup package to discovery/
* refactor: fix order of imports
7 years ago
|
|
|
"github.com/prometheus/prometheus/discovery/azure"
|
|
|
|
"github.com/prometheus/prometheus/discovery/consul"
|
|
|
|
"github.com/prometheus/prometheus/discovery/digitalocean"
|
Refactor SD configuration to remove `config` dependency (#3629)
* refactor: move targetGroup struct and CheckOverflow() to their own package
* refactor: move auth and security related structs to a utility package, fix import error in utility package
* refactor: Azure SD, remove SD struct from config
* refactor: DNS SD, remove SD struct from config into dns package
* refactor: ec2 SD, move SD struct from config into the ec2 package
* refactor: file SD, move SD struct from config to file discovery package
* refactor: gce, move SD struct from config to gce discovery package
* refactor: move HTTPClientConfig and URL into util/config, fix import error in httputil
* refactor: consul, move SD struct from config into consul discovery package
* refactor: marathon, move SD struct from config into marathon discovery package
* refactor: triton, move SD struct from config to triton discovery package, fix test
* refactor: zookeeper, move SD structs from config to zookeeper discovery package
* refactor: openstack, remove SD struct from config, move into openstack discovery package
* refactor: kubernetes, move SD struct from config into kubernetes discovery package
* refactor: notifier, use targetgroup package instead of config
* refactor: tests for file, marathon, triton SD - use targetgroup package instead of config.TargetGroup
* refactor: retrieval, use targetgroup package instead of config.TargetGroup
* refactor: storage, use config util package
* refactor: discovery manager, use targetgroup package instead of config.TargetGroup
* refactor: use HTTPClient and TLS config from configUtil instead of config
* refactor: tests, use targetgroup package instead of config.TargetGroup
* refactor: fix tagetgroup.Group pointers that were removed by mistake
* refactor: openstack, kubernetes: drop prefixes
* refactor: remove import aliases forced due to vscode bug
* refactor: move main SD struct out of config into discovery/config
* refactor: rename configUtil to config_util
* refactor: rename yamlUtil to yaml_config
* refactor: kubernetes, remove prefixes
* refactor: move the TargetGroup package to discovery/
* refactor: fix order of imports
7 years ago
|
|
|
"github.com/prometheus/prometheus/discovery/dns"
|
|
|
|
"github.com/prometheus/prometheus/discovery/eureka"
|
Refactor SD configuration to remove `config` dependency (#3629)
* refactor: move targetGroup struct and CheckOverflow() to their own package
* refactor: move auth and security related structs to a utility package, fix import error in utility package
* refactor: Azure SD, remove SD struct from config
* refactor: DNS SD, remove SD struct from config into dns package
* refactor: ec2 SD, move SD struct from config into the ec2 package
* refactor: file SD, move SD struct from config to file discovery package
* refactor: gce, move SD struct from config to gce discovery package
* refactor: move HTTPClientConfig and URL into util/config, fix import error in httputil
* refactor: consul, move SD struct from config into consul discovery package
* refactor: marathon, move SD struct from config into marathon discovery package
* refactor: triton, move SD struct from config to triton discovery package, fix test
* refactor: zookeeper, move SD structs from config to zookeeper discovery package
* refactor: openstack, remove SD struct from config, move into openstack discovery package
* refactor: kubernetes, move SD struct from config into kubernetes discovery package
* refactor: notifier, use targetgroup package instead of config
* refactor: tests for file, marathon, triton SD - use targetgroup package instead of config.TargetGroup
* refactor: retrieval, use targetgroup package instead of config.TargetGroup
* refactor: storage, use config util package
* refactor: discovery manager, use targetgroup package instead of config.TargetGroup
* refactor: use HTTPClient and TLS config from configUtil instead of config
* refactor: tests, use targetgroup package instead of config.TargetGroup
* refactor: fix tagetgroup.Group pointers that were removed by mistake
* refactor: openstack, kubernetes: drop prefixes
* refactor: remove import aliases forced due to vscode bug
* refactor: move main SD struct out of config into discovery/config
* refactor: rename configUtil to config_util
* refactor: rename yamlUtil to yaml_config
* refactor: kubernetes, remove prefixes
* refactor: move the TargetGroup package to discovery/
* refactor: fix order of imports
7 years ago
|
|
|
"github.com/prometheus/prometheus/discovery/file"
|
|
|
|
"github.com/prometheus/prometheus/discovery/hetzner"
|
|
|
|
"github.com/prometheus/prometheus/discovery/http"
|
Refactor SD configuration to remove `config` dependency (#3629)
* refactor: move targetGroup struct and CheckOverflow() to their own package
* refactor: move auth and security related structs to a utility package, fix import error in utility package
* refactor: Azure SD, remove SD struct from config
* refactor: DNS SD, remove SD struct from config into dns package
* refactor: ec2 SD, move SD struct from config into the ec2 package
* refactor: file SD, move SD struct from config to file discovery package
* refactor: gce, move SD struct from config to gce discovery package
* refactor: move HTTPClientConfig and URL into util/config, fix import error in httputil
* refactor: consul, move SD struct from config into consul discovery package
* refactor: marathon, move SD struct from config into marathon discovery package
* refactor: triton, move SD struct from config to triton discovery package, fix test
* refactor: zookeeper, move SD structs from config to zookeeper discovery package
* refactor: openstack, remove SD struct from config, move into openstack discovery package
* refactor: kubernetes, move SD struct from config into kubernetes discovery package
* refactor: notifier, use targetgroup package instead of config
* refactor: tests for file, marathon, triton SD - use targetgroup package instead of config.TargetGroup
* refactor: retrieval, use targetgroup package instead of config.TargetGroup
* refactor: storage, use config util package
* refactor: discovery manager, use targetgroup package instead of config.TargetGroup
* refactor: use HTTPClient and TLS config from configUtil instead of config
* refactor: tests, use targetgroup package instead of config.TargetGroup
* refactor: fix tagetgroup.Group pointers that were removed by mistake
* refactor: openstack, kubernetes: drop prefixes
* refactor: remove import aliases forced due to vscode bug
* refactor: move main SD struct out of config into discovery/config
* refactor: rename configUtil to config_util
* refactor: rename yamlUtil to yaml_config
* refactor: kubernetes, remove prefixes
* refactor: move the TargetGroup package to discovery/
* refactor: fix order of imports
7 years ago
|
|
|
"github.com/prometheus/prometheus/discovery/kubernetes"
|
|
|
|
"github.com/prometheus/prometheus/discovery/linode"
|
Refactor SD configuration to remove `config` dependency (#3629)
* refactor: move targetGroup struct and CheckOverflow() to their own package
* refactor: move auth and security related structs to a utility package, fix import error in utility package
* refactor: Azure SD, remove SD struct from config
* refactor: DNS SD, remove SD struct from config into dns package
* refactor: ec2 SD, move SD struct from config into the ec2 package
* refactor: file SD, move SD struct from config to file discovery package
* refactor: gce, move SD struct from config to gce discovery package
* refactor: move HTTPClientConfig and URL into util/config, fix import error in httputil
* refactor: consul, move SD struct from config into consul discovery package
* refactor: marathon, move SD struct from config into marathon discovery package
* refactor: triton, move SD struct from config to triton discovery package, fix test
* refactor: zookeeper, move SD structs from config to zookeeper discovery package
* refactor: openstack, remove SD struct from config, move into openstack discovery package
* refactor: kubernetes, move SD struct from config into kubernetes discovery package
* refactor: notifier, use targetgroup package instead of config
* refactor: tests for file, marathon, triton SD - use targetgroup package instead of config.TargetGroup
* refactor: retrieval, use targetgroup package instead of config.TargetGroup
* refactor: storage, use config util package
* refactor: discovery manager, use targetgroup package instead of config.TargetGroup
* refactor: use HTTPClient and TLS config from configUtil instead of config
* refactor: tests, use targetgroup package instead of config.TargetGroup
* refactor: fix tagetgroup.Group pointers that were removed by mistake
* refactor: openstack, kubernetes: drop prefixes
* refactor: remove import aliases forced due to vscode bug
* refactor: move main SD struct out of config into discovery/config
* refactor: rename configUtil to config_util
* refactor: rename yamlUtil to yaml_config
* refactor: kubernetes, remove prefixes
* refactor: move the TargetGroup package to discovery/
* refactor: fix order of imports
7 years ago
|
|
|
"github.com/prometheus/prometheus/discovery/marathon"
|
|
|
|
"github.com/prometheus/prometheus/discovery/moby"
|
|
|
|
"github.com/prometheus/prometheus/discovery/openstack"
|
|
|
|
"github.com/prometheus/prometheus/discovery/scaleway"
|
Refactor SD configuration to remove `config` dependency (#3629)
* refactor: move targetGroup struct and CheckOverflow() to their own package
* refactor: move auth and security related structs to a utility package, fix import error in utility package
* refactor: Azure SD, remove SD struct from config
* refactor: DNS SD, remove SD struct from config into dns package
* refactor: ec2 SD, move SD struct from config into the ec2 package
* refactor: file SD, move SD struct from config to file discovery package
* refactor: gce, move SD struct from config to gce discovery package
* refactor: move HTTPClientConfig and URL into util/config, fix import error in httputil
* refactor: consul, move SD struct from config into consul discovery package
* refactor: marathon, move SD struct from config into marathon discovery package
* refactor: triton, move SD struct from config to triton discovery package, fix test
* refactor: zookeeper, move SD structs from config to zookeeper discovery package
* refactor: openstack, remove SD struct from config, move into openstack discovery package
* refactor: kubernetes, move SD struct from config into kubernetes discovery package
* refactor: notifier, use targetgroup package instead of config
* refactor: tests for file, marathon, triton SD - use targetgroup package instead of config.TargetGroup
* refactor: retrieval, use targetgroup package instead of config.TargetGroup
* refactor: storage, use config util package
* refactor: discovery manager, use targetgroup package instead of config.TargetGroup
* refactor: use HTTPClient and TLS config from configUtil instead of config
* refactor: tests, use targetgroup package instead of config.TargetGroup
* refactor: fix tagetgroup.Group pointers that were removed by mistake
* refactor: openstack, kubernetes: drop prefixes
* refactor: remove import aliases forced due to vscode bug
* refactor: move main SD struct out of config into discovery/config
* refactor: rename configUtil to config_util
* refactor: rename yamlUtil to yaml_config
* refactor: kubernetes, remove prefixes
* refactor: move the TargetGroup package to discovery/
* refactor: fix order of imports
7 years ago
|
|
|
"github.com/prometheus/prometheus/discovery/targetgroup"
|
|
|
|
"github.com/prometheus/prometheus/discovery/triton"
|
|
|
|
"github.com/prometheus/prometheus/discovery/zookeeper"
|
|
|
|
"github.com/prometheus/prometheus/pkg/labels"
|
|
|
|
"github.com/prometheus/prometheus/pkg/relabel"
|
|
|
|
)
|
|
|
|
|
|
|
|
func mustParseURL(u string) *config.URL {
|
|
|
|
parsed, err := url.Parse(u)
|
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
|
|
|
return &config.URL{URL: parsed}
|
|
|
|
}
|
|
|
|
|
|
|
|
var expectedConf = &Config{
|
|
|
|
GlobalConfig: GlobalConfig{
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
EvaluationInterval: model.Duration(30 * time.Second),
|
|
|
|
QueryLogFile: "",
|
|
|
|
|
|
|
|
ExternalLabels: labels.Labels{
|
|
|
|
{Name: "foo", Value: "bar"},
|
|
|
|
{Name: "monitor", Value: "codelab"},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
|
|
|
|
RuleFiles: []string{
|
|
|
|
filepath.FromSlash("testdata/first.rules"),
|
|
|
|
filepath.FromSlash("testdata/my/*.rules"),
|
|
|
|
},
|
|
|
|
|
|
|
|
RemoteWriteConfigs: []*RemoteWriteConfig{
|
|
|
|
{
|
|
|
|
URL: mustParseURL("http://remote1/push"),
|
|
|
|
RemoteTimeout: model.Duration(30 * time.Second),
|
|
|
|
Name: "drop_expensive",
|
|
|
|
WriteRelabelConfigs: []*relabel.Config{
|
|
|
|
{
|
|
|
|
SourceLabels: model.LabelNames{"__name__"},
|
|
|
|
Separator: ";",
|
|
|
|
Regex: relabel.MustNewRegexp("expensive.*"),
|
|
|
|
Replacement: "$1",
|
|
|
|
Action: relabel.Drop,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
QueueConfig: DefaultQueueConfig,
|
|
|
|
MetadataConfig: DefaultMetadataConfig,
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
|
|
OAuth2: &config.OAuth2{
|
|
|
|
ClientID: "123",
|
|
|
|
ClientSecret: "456",
|
|
|
|
TokenURL: "http://remote1/auth",
|
|
|
|
},
|
|
|
|
FollowRedirects: true,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
URL: mustParseURL("http://remote2/push"),
|
|
|
|
RemoteTimeout: model.Duration(30 * time.Second),
|
|
|
|
QueueConfig: DefaultQueueConfig,
|
|
|
|
MetadataConfig: DefaultMetadataConfig,
|
|
|
|
Name: "rw_tls",
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
|
|
TLSConfig: config.TLSConfig{
|
|
|
|
CertFile: filepath.FromSlash("testdata/valid_cert_file"),
|
|
|
|
KeyFile: filepath.FromSlash("testdata/valid_key_file"),
|
|
|
|
},
|
|
|
|
FollowRedirects: true,
|
|
|
|
},
|
|
|
|
Headers: map[string]string{"name": "value"},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
|
|
|
|
RemoteReadConfigs: []*RemoteReadConfig{
|
|
|
|
{
|
|
|
|
URL: mustParseURL("http://remote1/read"),
|
|
|
|
RemoteTimeout: model.Duration(1 * time.Minute),
|
|
|
|
ReadRecent: true,
|
|
|
|
Name: "default",
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
URL: mustParseURL("http://remote3/read"),
|
|
|
|
RemoteTimeout: model.Duration(1 * time.Minute),
|
|
|
|
ReadRecent: false,
|
|
|
|
Name: "read_special",
|
|
|
|
RequiredMatchers: model.LabelSet{"job": "special"},
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
|
|
TLSConfig: config.TLSConfig{
|
|
|
|
CertFile: filepath.FromSlash("testdata/valid_cert_file"),
|
|
|
|
KeyFile: filepath.FromSlash("testdata/valid_key_file"),
|
|
|
|
},
|
|
|
|
FollowRedirects: true,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
|
|
|
|
ScrapeConfigs: []*ScrapeConfig{
|
|
|
|
{
|
|
|
|
JobName: "prometheus",
|
|
|
|
|
|
|
|
HonorLabels: true,
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
|
|
Authorization: &config.Authorization{
|
|
|
|
Type: "Bearer",
|
|
|
|
CredentialsFile: filepath.FromSlash("testdata/valid_token_file"),
|
|
|
|
},
|
|
|
|
FollowRedirects: true,
|
|
|
|
},
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&file.SDConfig{
|
|
|
|
Files: []string{"testdata/foo/*.slow.json", "testdata/foo/*.slow.yml", "testdata/single/file.yml"},
|
|
|
|
RefreshInterval: model.Duration(10 * time.Minute),
|
|
|
|
},
|
|
|
|
&file.SDConfig{
|
|
|
|
Files: []string{"testdata/bar/*.yaml"},
|
|
|
|
RefreshInterval: model.Duration(5 * time.Minute),
|
|
|
|
},
|
|
|
|
discovery.StaticConfig{
|
|
|
|
{
|
|
|
|
Targets: []model.LabelSet{
|
|
|
|
{model.AddressLabel: "localhost:9090"},
|
|
|
|
{model.AddressLabel: "localhost:9191"},
|
|
|
|
},
|
|
|
|
Labels: model.LabelSet{
|
|
|
|
"my": "label",
|
|
|
|
"your": "label",
|
|
|
|
},
|
|
|
|
Source: "0",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
|
|
|
|
RelabelConfigs: []*relabel.Config{
|
|
|
|
{
|
|
|
|
SourceLabels: model.LabelNames{"job", "__meta_dns_name"},
|
|
|
|
TargetLabel: "job",
|
|
|
|
Separator: ";",
|
|
|
|
Regex: relabel.MustNewRegexp("(.*)some-[regex]"),
|
|
|
|
Replacement: "foo-${1}",
|
|
|
|
Action: relabel.Replace,
|
|
|
|
}, {
|
|
|
|
SourceLabels: model.LabelNames{"abc"},
|
|
|
|
TargetLabel: "cde",
|
|
|
|
Separator: ";",
|
|
|
|
Regex: relabel.DefaultRelabelConfig.Regex,
|
|
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
|
|
Action: relabel.Replace,
|
|
|
|
}, {
|
|
|
|
TargetLabel: "abc",
|
|
|
|
Separator: ";",
|
|
|
|
Regex: relabel.DefaultRelabelConfig.Regex,
|
|
|
|
Replacement: "static",
|
|
|
|
Action: relabel.Replace,
|
|
|
|
}, {
|
|
|
|
TargetLabel: "abc",
|
|
|
|
Separator: ";",
|
|
|
|
Regex: relabel.MustNewRegexp(""),
|
|
|
|
Replacement: "static",
|
|
|
|
Action: relabel.Replace,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
|
|
|
|
JobName: "service-x",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(50 * time.Second),
|
|
|
|
ScrapeTimeout: model.Duration(5 * time.Second),
|
|
|
|
BodySizeLimit: 10 * units.MiB,
|
|
|
|
SampleLimit: 1000,
|
|
|
|
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
|
|
BasicAuth: &config.BasicAuth{
|
|
|
|
Username: "admin_name",
|
|
|
|
Password: "multiline\nmysecret\ntest",
|
|
|
|
},
|
|
|
|
FollowRedirects: true,
|
|
|
|
},
|
|
|
|
MetricsPath: "/my_path",
|
|
|
|
Scheme: "https",
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&dns.SDConfig{
|
|
|
|
Names: []string{
|
|
|
|
"first.dns.address.domain.com",
|
|
|
|
"second.dns.address.domain.com",
|
|
|
|
},
|
|
|
|
RefreshInterval: model.Duration(15 * time.Second),
|
|
|
|
Type: "SRV",
|
|
|
|
},
|
|
|
|
&dns.SDConfig{
|
|
|
|
Names: []string{
|
|
|
|
"first.dns.address.domain.com",
|
|
|
|
},
|
|
|
|
RefreshInterval: model.Duration(30 * time.Second),
|
|
|
|
Type: "SRV",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
|
|
|
|
RelabelConfigs: []*relabel.Config{
|
|
|
|
{
|
|
|
|
SourceLabels: model.LabelNames{"job"},
|
|
|
|
Regex: relabel.MustNewRegexp("(.*)some-[regex]"),
|
|
|
|
Separator: ";",
|
|
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
|
|
Action: relabel.Drop,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
SourceLabels: model.LabelNames{"__address__"},
|
|
|
|
TargetLabel: "__tmp_hash",
|
|
|
|
Regex: relabel.DefaultRelabelConfig.Regex,
|
|
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
|
|
Modulus: 8,
|
|
|
|
Separator: ";",
|
|
|
|
Action: relabel.HashMod,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
SourceLabels: model.LabelNames{"__tmp_hash"},
|
|
|
|
Regex: relabel.MustNewRegexp("1"),
|
|
|
|
Separator: ";",
|
|
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
|
|
Action: relabel.Keep,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Regex: relabel.MustNewRegexp("1"),
|
|
|
|
Separator: ";",
|
|
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
|
|
Action: relabel.LabelMap,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Regex: relabel.MustNewRegexp("d"),
|
|
|
|
Separator: ";",
|
|
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
|
|
Action: relabel.LabelDrop,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Regex: relabel.MustNewRegexp("k"),
|
|
|
|
Separator: ";",
|
|
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
|
|
Action: relabel.LabelKeep,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
MetricRelabelConfigs: []*relabel.Config{
|
|
|
|
{
|
|
|
|
SourceLabels: model.LabelNames{"__name__"},
|
|
|
|
Regex: relabel.MustNewRegexp("expensive_metric.*"),
|
|
|
|
Separator: ";",
|
|
|
|
Replacement: relabel.DefaultRelabelConfig.Replacement,
|
|
|
|
Action: relabel.Drop,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "service-y",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&consul.SDConfig{
|
|
|
|
Server: "localhost:1234",
|
|
|
|
Token: "mysecret",
|
|
|
|
Services: []string{"nginx", "cache", "mysql"},
|
|
|
|
ServiceTags: []string{"canary", "v1"},
|
|
|
|
NodeMeta: map[string]string{"rack": "123"},
|
|
|
|
TagSeparator: consul.DefaultSDConfig.TagSeparator,
|
|
|
|
Scheme: "https",
|
|
|
|
RefreshInterval: consul.DefaultSDConfig.RefreshInterval,
|
|
|
|
AllowStale: true,
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
|
|
TLSConfig: config.TLSConfig{
|
|
|
|
CertFile: filepath.FromSlash("testdata/valid_cert_file"),
|
|
|
|
KeyFile: filepath.FromSlash("testdata/valid_key_file"),
|
|
|
|
CAFile: filepath.FromSlash("testdata/valid_ca_file"),
|
|
|
|
InsecureSkipVerify: false,
|
|
|
|
},
|
|
|
|
FollowRedirects: true,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
|
|
|
|
RelabelConfigs: []*relabel.Config{
|
|
|
|
{
|
|
|
|
SourceLabels: model.LabelNames{"__meta_sd_consul_tags"},
|
|
|
|
Regex: relabel.MustNewRegexp("label:([^=]+)=([^,]+)"),
|
|
|
|
Separator: ",",
|
|
|
|
TargetLabel: "${1}",
|
|
|
|
Replacement: "${2}",
|
|
|
|
Action: relabel.Replace,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "service-z",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: model.Duration(10 * time.Second),
|
|
|
|
|
|
|
|
MetricsPath: "/metrics",
|
|
|
|
Scheme: "http",
|
|
|
|
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
|
|
TLSConfig: config.TLSConfig{
|
|
|
|
CertFile: filepath.FromSlash("testdata/valid_cert_file"),
|
|
|
|
KeyFile: filepath.FromSlash("testdata/valid_key_file"),
|
|
|
|
},
|
|
|
|
|
|
|
|
Authorization: &config.Authorization{
|
|
|
|
Type: "Bearer",
|
|
|
|
Credentials: "mysecret",
|
|
|
|
},
|
|
|
|
|
|
|
|
FollowRedirects: true,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "service-kubernetes",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&kubernetes.SDConfig{
|
|
|
|
APIServer: kubernetesSDHostURL(),
|
|
|
|
Role: kubernetes.RoleEndpoint,
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
|
|
BasicAuth: &config.BasicAuth{
|
|
|
|
Username: "myusername",
|
|
|
|
Password: "mysecret",
|
|
|
|
},
|
|
|
|
TLSConfig: config.TLSConfig{
|
|
|
|
CertFile: filepath.FromSlash("testdata/valid_cert_file"),
|
|
|
|
KeyFile: filepath.FromSlash("testdata/valid_key_file"),
|
|
|
|
},
|
|
|
|
FollowRedirects: true,
|
|
|
|
},
|
|
|
|
NamespaceDiscovery: kubernetes.NamespaceDiscovery{},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "service-kubernetes-namespaces",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
|
|
BasicAuth: &config.BasicAuth{
|
|
|
|
Username: "myusername",
|
|
|
|
PasswordFile: filepath.FromSlash("testdata/valid_password_file"),
|
|
|
|
},
|
|
|
|
FollowRedirects: true,
|
|
|
|
},
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&kubernetes.SDConfig{
|
|
|
|
APIServer: kubernetesSDHostURL(),
|
|
|
|
Role: kubernetes.RoleEndpoint,
|
|
|
|
NamespaceDiscovery: kubernetes.NamespaceDiscovery{
|
|
|
|
Names: []string{
|
|
|
|
"default",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "service-marathon",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&marathon.SDConfig{
|
|
|
|
Servers: []string{
|
|
|
|
"https://marathon.example.com:443",
|
|
|
|
},
|
|
|
|
RefreshInterval: model.Duration(30 * time.Second),
|
|
|
|
AuthToken: "mysecret",
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
|
|
TLSConfig: config.TLSConfig{
|
|
|
|
CertFile: filepath.FromSlash("testdata/valid_cert_file"),
|
|
|
|
KeyFile: filepath.FromSlash("testdata/valid_key_file"),
|
|
|
|
},
|
|
|
|
FollowRedirects: true,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "service-ec2",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&aws.EC2SDConfig{
|
|
|
|
Region: "us-east-1",
|
|
|
|
AccessKey: "access",
|
|
|
|
SecretKey: "mysecret",
|
|
|
|
Profile: "profile",
|
|
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
|
|
Port: 80,
|
|
|
|
Filters: []*aws.EC2Filter{
|
|
|
|
{
|
|
|
|
Name: "tag:environment",
|
|
|
|
Values: []string{"prod"},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "tag:service",
|
|
|
|
Values: []string{"web", "db"},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "service-lightsail",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&aws.LightsailSDConfig{
|
|
|
|
Region: "us-east-1",
|
|
|
|
AccessKey: "access",
|
|
|
|
SecretKey: "mysecret",
|
|
|
|
Profile: "profile",
|
|
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
|
|
Port: 80,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "service-azure",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&azure.SDConfig{
|
|
|
|
Environment: "AzurePublicCloud",
|
|
|
|
SubscriptionID: "11AAAA11-A11A-111A-A111-1111A1111A11",
|
|
|
|
TenantID: "BBBB222B-B2B2-2B22-B222-2BB2222BB2B2",
|
|
|
|
ClientID: "333333CC-3C33-3333-CCC3-33C3CCCCC33C",
|
|
|
|
ClientSecret: "mysecret",
|
|
|
|
AuthenticationMethod: "OAuth",
|
|
|
|
RefreshInterval: model.Duration(5 * time.Minute),
|
|
|
|
Port: 9100,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "service-nerve",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&zookeeper.NerveSDConfig{
|
|
|
|
Servers: []string{"localhost"},
|
|
|
|
Paths: []string{"/monitoring"},
|
|
|
|
Timeout: model.Duration(10 * time.Second),
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "0123service-xxx",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
discovery.StaticConfig{
|
|
|
|
{
|
|
|
|
Targets: []model.LabelSet{
|
|
|
|
{model.AddressLabel: "localhost:9090"},
|
|
|
|
},
|
|
|
|
Source: "0",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "badfederation",
|
|
|
|
|
|
|
|
HonorTimestamps: false,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: "/federate",
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
discovery.StaticConfig{
|
|
|
|
{
|
|
|
|
Targets: []model.LabelSet{
|
|
|
|
{model.AddressLabel: "localhost:9090"},
|
|
|
|
},
|
|
|
|
Source: "0",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "測試",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
discovery.StaticConfig{
|
|
|
|
{
|
|
|
|
Targets: []model.LabelSet{
|
|
|
|
{model.AddressLabel: "localhost:9090"},
|
|
|
|
},
|
|
|
|
Source: "0",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "httpsd",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&http.SDConfig{
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
URL: "http://example.com/prometheus",
|
|
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "service-triton",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&triton.SDConfig{
|
|
|
|
Account: "testAccount",
|
|
|
|
Role: "container",
|
|
|
|
DNSSuffix: "triton.example.com",
|
|
|
|
Endpoint: "triton.example.com",
|
|
|
|
Port: 9163,
|
|
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
|
|
Version: 1,
|
|
|
|
TLSConfig: config.TLSConfig{
|
|
|
|
CertFile: "testdata/valid_cert_file",
|
|
|
|
KeyFile: "testdata/valid_key_file",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "digitalocean-droplets",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&digitalocean.SDConfig{
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
|
|
Authorization: &config.Authorization{
|
|
|
|
Type: "Bearer",
|
|
|
|
Credentials: "abcdef",
|
|
|
|
},
|
|
|
|
FollowRedirects: true,
|
|
|
|
},
|
|
|
|
Port: 80,
|
|
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "docker",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&moby.DockerSDConfig{
|
|
|
|
Filters: []moby.Filter{},
|
|
|
|
Host: "unix:///var/run/docker.sock",
|
|
|
|
Port: 80,
|
|
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "dockerswarm",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&moby.DockerSwarmSDConfig{
|
|
|
|
Filters: []moby.Filter{},
|
|
|
|
Host: "http://127.0.0.1:2375",
|
|
|
|
Role: "nodes",
|
|
|
|
Port: 80,
|
|
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "service-openstack",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{&openstack.SDConfig{
|
|
|
|
Role: "instance",
|
|
|
|
Region: "RegionOne",
|
|
|
|
Port: 80,
|
|
|
|
Availability: "public",
|
|
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
|
|
TLSConfig: config.TLSConfig{
|
|
|
|
CAFile: "testdata/valid_ca_file",
|
|
|
|
CertFile: "testdata/valid_cert_file",
|
|
|
|
KeyFile: "testdata/valid_key_file",
|
|
|
|
}},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "hetzner",
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&hetzner.SDConfig{
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
|
|
Authorization: &config.Authorization{
|
|
|
|
Type: "Bearer",
|
|
|
|
Credentials: "abcdef",
|
|
|
|
},
|
|
|
|
FollowRedirects: true,
|
|
|
|
},
|
|
|
|
Port: 80,
|
|
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
|
|
Role: "hcloud",
|
|
|
|
},
|
|
|
|
&hetzner.SDConfig{
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
|
|
BasicAuth: &config.BasicAuth{Username: "abcdef", Password: "abcdef"},
|
|
|
|
FollowRedirects: true,
|
|
|
|
},
|
|
|
|
Port: 80,
|
|
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
|
|
Role: "robot",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "service-eureka",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&eureka.SDConfig{
|
|
|
|
Server: "http://eureka.example.com:8761/eureka",
|
|
|
|
RefreshInterval: model.Duration(30 * time.Second),
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "scaleway",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{FollowRedirects: true},
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&scaleway.SDConfig{
|
|
|
|
APIURL: "https://api.scaleway.com",
|
|
|
|
AccessKey: "SCWXXXXXXXXXXXXXXXXX",
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{FollowRedirects: true},
|
|
|
|
Port: 80,
|
|
|
|
Project: "11111111-1111-1111-1111-111111111112",
|
|
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
|
|
Role: "instance",
|
|
|
|
SecretKey: "11111111-1111-1111-1111-111111111111",
|
|
|
|
Zone: "fr-par-1",
|
|
|
|
},
|
|
|
|
&scaleway.SDConfig{
|
|
|
|
APIURL: "https://api.scaleway.com",
|
|
|
|
AccessKey: "SCWXXXXXXXXXXXXXXXXX",
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{FollowRedirects: true},
|
|
|
|
Port: 80,
|
|
|
|
Project: "11111111-1111-1111-1111-111111111112",
|
|
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
|
|
Role: "baremetal",
|
|
|
|
SecretKey: "11111111-1111-1111-1111-111111111111",
|
|
|
|
Zone: "fr-par-1",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
JobName: "linode-instances",
|
|
|
|
|
|
|
|
HonorTimestamps: true,
|
|
|
|
ScrapeInterval: model.Duration(15 * time.Second),
|
|
|
|
ScrapeTimeout: DefaultGlobalConfig.ScrapeTimeout,
|
|
|
|
|
|
|
|
MetricsPath: DefaultScrapeConfig.MetricsPath,
|
|
|
|
Scheme: DefaultScrapeConfig.Scheme,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
&linode.SDConfig{
|
|
|
|
HTTPClientConfig: config.HTTPClientConfig{
|
|
|
|
Authorization: &config.Authorization{
|
|
|
|
Type: "Bearer",
|
|
|
|
Credentials: "abcdef",
|
|
|
|
},
|
|
|
|
FollowRedirects: true,
|
|
|
|
},
|
|
|
|
Port: 80,
|
|
|
|
TagSeparator: linode.DefaultSDConfig.TagSeparator,
|
|
|
|
RefreshInterval: model.Duration(60 * time.Second),
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
AlertingConfig: AlertingConfig{
|
|
|
|
AlertmanagerConfigs: []*AlertmanagerConfig{
|
|
|
|
{
|
|
|
|
Scheme: "https",
|
|
|
|
Timeout: model.Duration(10 * time.Second),
|
|
|
|
APIVersion: AlertmanagerAPIVersionV2,
|
|
|
|
HTTPClientConfig: config.DefaultHTTPClientConfig,
|
|
|
|
ServiceDiscoveryConfigs: discovery.Configs{
|
|
|
|
discovery.StaticConfig{
|
|
|
|
{
|
|
|
|
Targets: []model.LabelSet{
|
|
|
|
{model.AddressLabel: "1.2.3.4:9093"},
|
|
|
|
{model.AddressLabel: "1.2.3.5:9093"},
|
|
|
|
{model.AddressLabel: "1.2.3.6:9093"},
|
|
|
|
},
|
|
|
|
Source: "0",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestYAMLRoundtrip(t *testing.T) {
|
|
|
|
want, err := LoadFile("testdata/roundtrip.good.yml", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
out, err := yaml.Marshal(want)
|
|
|
|
|
|
|
|
require.NoError(t, err)
|
|
|
|
got := &Config{}
|
|
|
|
require.NoError(t, yaml.UnmarshalStrict(out, got))
|
|
|
|
|
|
|
|
require.Equal(t, want, got)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestRemoteWriteRetryOnRateLimit(t *testing.T) {
|
|
|
|
want, err := LoadFile("testdata/remote_write_retry_on_rate_limit.good.yml", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
out, err := yaml.Marshal(want)
|
|
|
|
|
|
|
|
require.NoError(t, err)
|
|
|
|
got := &Config{}
|
|
|
|
require.NoError(t, yaml.UnmarshalStrict(out, got))
|
|
|
|
|
|
|
|
require.Equal(t, true, got.RemoteWriteConfigs[0].QueueConfig.RetryOnRateLimit)
|
|
|
|
require.Equal(t, false, got.RemoteWriteConfigs[1].QueueConfig.RetryOnRateLimit)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestLoadConfig(t *testing.T) {
|
|
|
|
// Parse a valid file that sets a global scrape timeout. This tests whether parsing
|
|
|
|
// an overwritten default field in the global config permanently changes the default.
|
|
|
|
_, err := LoadFile("testdata/global_timeout.good.yml", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
c, err := LoadFile("testdata/conf.good.yml", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, expectedConf, c)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestScrapeIntervalLarger(t *testing.T) {
|
|
|
|
c, err := LoadFile("testdata/scrape_interval_larger.good.yml", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, 1, len(c.ScrapeConfigs))
|
|
|
|
for _, sc := range c.ScrapeConfigs {
|
|
|
|
require.Equal(t, true, sc.ScrapeInterval >= sc.ScrapeTimeout)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// YAML marshaling must not reveal authentication credentials.
|
|
|
|
func TestElideSecrets(t *testing.T) {
|
|
|
|
c, err := LoadFile("testdata/conf.good.yml", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
secretRe := regexp.MustCompile(`\\u003csecret\\u003e|<secret>`)
|
|
|
|
|
|
|
|
config, err := yaml.Marshal(c)
|
|
|
|
require.NoError(t, err)
|
|
|
|
yamlConfig := string(config)
|
|
|
|
|
|
|
|
matches := secretRe.FindAllStringIndex(yamlConfig, -1)
|
|
|
|
require.Equal(t, 15, len(matches), "wrong number of secret matches found")
|
|
|
|
require.NotContains(t, yamlConfig, "mysecret",
|
|
|
|
"yaml marshal reveals authentication credentials.")
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestLoadConfigRuleFilesAbsolutePath(t *testing.T) {
|
|
|
|
// Parse a valid file that sets a rule files with an absolute path
|
|
|
|
c, err := LoadFile(ruleFilesConfigFile, false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, ruleFilesExpectedConf, c)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestKubernetesEmptyAPIServer(t *testing.T) {
|
|
|
|
_, err := LoadFile("testdata/kubernetes_empty_apiserver.good.yml", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestKubernetesWithKubeConfig(t *testing.T) {
|
|
|
|
_, err := LoadFile("testdata/kubernetes_kubeconfig_without_apiserver.good.yml", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestKubernetesSelectors(t *testing.T) {
|
|
|
|
_, err := LoadFile("testdata/kubernetes_selectors_endpoints.good.yml", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = LoadFile("testdata/kubernetes_selectors_node.good.yml", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = LoadFile("testdata/kubernetes_selectors_ingress.good.yml", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = LoadFile("testdata/kubernetes_selectors_pod.good.yml", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = LoadFile("testdata/kubernetes_selectors_service.good.yml", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
var expectedErrors = []struct {
|
|
|
|
filename string
|
|
|
|
errMsg string
|
|
|
|
}{
|
|
|
|
{
|
|
|
|
filename: "jobname.bad.yml",
|
|
|
|
errMsg: `job_name is empty`,
|
|
|
|
}, {
|
|
|
|
filename: "jobname_dup.bad.yml",
|
|
|
|
errMsg: `found multiple scrape configs with job name "prometheus"`,
|
|
|
|
}, {
|
|
|
|
filename: "scrape_interval.bad.yml",
|
|
|
|
errMsg: `scrape timeout greater than scrape interval`,
|
|
|
|
}, {
|
|
|
|
filename: "labelname.bad.yml",
|
|
|
|
errMsg: `"not$allowed" is not a valid label name`,
|
|
|
|
}, {
|
|
|
|
filename: "labelname2.bad.yml",
|
|
|
|
errMsg: `"not:allowed" is not a valid label name`,
|
|
|
|
}, {
|
|
|
|
filename: "labelvalue.bad.yml",
|
|
|
|
errMsg: `"\xff" is not a valid label value`,
|
|
|
|
}, {
|
|
|
|
filename: "regex.bad.yml",
|
|
|
|
errMsg: "error parsing regexp",
|
|
|
|
}, {
|
|
|
|
filename: "modulus_missing.bad.yml",
|
|
|
|
errMsg: "relabel configuration for hashmod requires non-zero modulus",
|
|
|
|
}, {
|
|
|
|
filename: "labelkeep.bad.yml",
|
|
|
|
errMsg: "labelkeep action requires only 'regex', and no other fields",
|
|
|
|
}, {
|
|
|
|
filename: "labelkeep2.bad.yml",
|
|
|
|
errMsg: "labelkeep action requires only 'regex', and no other fields",
|
|
|
|
}, {
|
|
|
|
filename: "labelkeep3.bad.yml",
|
|
|
|
errMsg: "labelkeep action requires only 'regex', and no other fields",
|
|
|
|
}, {
|
|
|
|
filename: "labelkeep4.bad.yml",
|
|
|
|
errMsg: "labelkeep action requires only 'regex', and no other fields",
|
|
|
|
}, {
|
|
|
|
filename: "labelkeep5.bad.yml",
|
|
|
|
errMsg: "labelkeep action requires only 'regex', and no other fields",
|
|
|
|
}, {
|
|
|
|
filename: "labeldrop.bad.yml",
|
|
|
|
errMsg: "labeldrop action requires only 'regex', and no other fields",
|
|
|
|
}, {
|
|
|
|
filename: "labeldrop2.bad.yml",
|
|
|
|
errMsg: "labeldrop action requires only 'regex', and no other fields",
|
|
|
|
}, {
|
|
|
|
filename: "labeldrop3.bad.yml",
|
|
|
|
errMsg: "labeldrop action requires only 'regex', and no other fields",
|
|
|
|
}, {
|
|
|
|
filename: "labeldrop4.bad.yml",
|
|
|
|
errMsg: "labeldrop action requires only 'regex', and no other fields",
|
|
|
|
}, {
|
|
|
|
filename: "labeldrop5.bad.yml",
|
|
|
|
errMsg: "labeldrop action requires only 'regex', and no other fields",
|
|
|
|
}, {
|
|
|
|
filename: "labelmap.bad.yml",
|
|
|
|
errMsg: "\"l-$1\" is invalid 'replacement' for labelmap action",
|
|
|
|
}, {
|
|
|
|
filename: "rules.bad.yml",
|
|
|
|
errMsg: "invalid rule file path",
|
|
|
|
}, {
|
|
|
|
filename: "unknown_attr.bad.yml",
|
|
|
|
errMsg: "field consult_sd_configs not found in type",
|
|
|
|
}, {
|
|
|
|
filename: "bearertoken.bad.yml",
|
|
|
|
errMsg: "at most one of bearer_token & bearer_token_file must be configured",
|
|
|
|
}, {
|
|
|
|
filename: "bearertoken_basicauth.bad.yml",
|
|
|
|
errMsg: "at most one of basic_auth, oauth2, bearer_token & bearer_token_file must be configured",
|
|
|
|
}, {
|
|
|
|
filename: "kubernetes_http_config_without_api_server.bad.yml",
|
|
|
|
errMsg: "to use custom HTTP client configuration please provide the 'api_server' URL explicitly",
|
|
|
|
}, {
|
|
|
|
filename: "kubernetes_kubeconfig_with_apiserver.bad.yml",
|
|
|
|
errMsg: "cannot use 'kubeconfig_file' and 'api_server' simultaneously",
|
|
|
|
}, {
|
|
|
|
filename: "kubernetes_kubeconfig_with_http_config.bad.yml",
|
|
|
|
errMsg: "cannot use a custom HTTP client configuration together with 'kubeconfig_file'",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "kubernetes_bearertoken.bad.yml",
|
|
|
|
errMsg: "at most one of bearer_token & bearer_token_file must be configured",
|
|
|
|
}, {
|
|
|
|
filename: "kubernetes_role.bad.yml",
|
|
|
|
errMsg: "role",
|
|
|
|
}, {
|
|
|
|
filename: "kubernetes_selectors_endpoints.bad.yml",
|
|
|
|
errMsg: "endpoints role supports only pod, service, endpoints selectors",
|
|
|
|
}, {
|
|
|
|
filename: "kubernetes_selectors_ingress.bad.yml",
|
|
|
|
errMsg: "ingress role supports only ingress selectors",
|
|
|
|
}, {
|
|
|
|
filename: "kubernetes_selectors_node.bad.yml",
|
|
|
|
errMsg: "node role supports only node selectors",
|
|
|
|
}, {
|
|
|
|
filename: "kubernetes_selectors_pod.bad.yml",
|
|
|
|
errMsg: "pod role supports only pod selectors",
|
|
|
|
}, {
|
|
|
|
filename: "kubernetes_selectors_service.bad.yml",
|
|
|
|
errMsg: "service role supports only service selectors",
|
|
|
|
}, {
|
|
|
|
filename: "kubernetes_namespace_discovery.bad.yml",
|
|
|
|
errMsg: "field foo not found in type kubernetes.plain",
|
|
|
|
}, {
|
|
|
|
filename: "kubernetes_selectors_duplicated_role.bad.yml",
|
|
|
|
errMsg: "duplicated selector role: pod",
|
|
|
|
}, {
|
|
|
|
filename: "kubernetes_selectors_incorrect_selector.bad.yml",
|
|
|
|
errMsg: "invalid selector: 'metadata.status-Running'; can't understand 'metadata.status-Running'",
|
|
|
|
}, {
|
|
|
|
filename: "kubernetes_bearertoken_basicauth.bad.yml",
|
|
|
|
errMsg: "at most one of basic_auth, oauth2, bearer_token & bearer_token_file must be configured",
|
|
|
|
}, {
|
|
|
|
filename: "kubernetes_authorization_basicauth.bad.yml",
|
|
|
|
errMsg: "at most one of basic_auth, oauth2 & authorization must be configured",
|
|
|
|
}, {
|
|
|
|
filename: "marathon_no_servers.bad.yml",
|
|
|
|
errMsg: "marathon_sd: must contain at least one Marathon server",
|
|
|
|
}, {
|
|
|
|
filename: "marathon_authtoken_authtokenfile.bad.yml",
|
|
|
|
errMsg: "marathon_sd: at most one of auth_token & auth_token_file must be configured",
|
|
|
|
}, {
|
|
|
|
filename: "marathon_authtoken_basicauth.bad.yml",
|
|
|
|
errMsg: "marathon_sd: at most one of basic_auth, auth_token & auth_token_file must be configured",
|
|
|
|
}, {
|
|
|
|
filename: "marathon_authtoken_bearertoken.bad.yml",
|
|
|
|
errMsg: "marathon_sd: at most one of bearer_token, bearer_token_file, auth_token & auth_token_file must be configured",
|
|
|
|
}, {
|
|
|
|
filename: "marathon_authtoken_authorization.bad.yml",
|
|
|
|
errMsg: "marathon_sd: at most one of auth_token, auth_token_file & authorization must be configured",
|
|
|
|
}, {
|
|
|
|
filename: "openstack_role.bad.yml",
|
|
|
|
errMsg: "unknown OpenStack SD role",
|
|
|
|
}, {
|
|
|
|
filename: "openstack_availability.bad.yml",
|
|
|
|
errMsg: "unknown availability invalid, must be one of admin, internal or public",
|
|
|
|
}, {
|
|
|
|
filename: "url_in_targetgroup.bad.yml",
|
|
|
|
errMsg: "\"http://bad\" is not a valid hostname",
|
|
|
|
}, {
|
|
|
|
filename: "target_label_missing.bad.yml",
|
|
|
|
errMsg: "relabel configuration for replace action requires 'target_label' value",
|
|
|
|
}, {
|
|
|
|
filename: "target_label_hashmod_missing.bad.yml",
|
|
|
|
errMsg: "relabel configuration for hashmod action requires 'target_label' value",
|
|
|
|
}, {
|
|
|
|
filename: "unknown_global_attr.bad.yml",
|
|
|
|
errMsg: "field nonexistent_field not found in type config.plain",
|
|
|
|
}, {
|
|
|
|
filename: "remote_read_url_missing.bad.yml",
|
|
|
|
errMsg: `url for remote_read is empty`,
|
|
|
|
}, {
|
|
|
|
filename: "remote_write_header.bad.yml",
|
|
|
|
errMsg: `x-prometheus-remote-write-version is a reserved header. It must not be changed`,
|
|
|
|
}, {
|
|
|
|
filename: "remote_read_header.bad.yml",
|
|
|
|
errMsg: `x-prometheus-remote-write-version is a reserved header. It must not be changed`,
|
|
|
|
}, {
|
|
|
|
filename: "remote_write_authorization_header.bad.yml",
|
|
|
|
errMsg: `authorization header must be changed via the basic_auth, authorization, oauth2, or sigv4 parameter`,
|
|
|
|
}, {
|
|
|
|
filename: "remote_write_url_missing.bad.yml",
|
|
|
|
errMsg: `url for remote_write is empty`,
|
|
|
|
}, {
|
|
|
|
filename: "remote_write_dup.bad.yml",
|
|
|
|
errMsg: `found multiple remote write configs with job name "queue1"`,
|
|
|
|
}, {
|
|
|
|
filename: "remote_read_dup.bad.yml",
|
|
|
|
errMsg: `found multiple remote read configs with job name "queue1"`,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "ec2_filters_empty_values.bad.yml",
|
|
|
|
errMsg: `EC2 SD configuration filter values cannot be empty`,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "section_key_dup.bad.yml",
|
|
|
|
errMsg: "field scrape_configs already set in type config.plain",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "azure_client_id_missing.bad.yml",
|
|
|
|
errMsg: "azure SD configuration requires a client_id",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "azure_client_secret_missing.bad.yml",
|
|
|
|
errMsg: "azure SD configuration requires a client_secret",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "azure_subscription_id_missing.bad.yml",
|
|
|
|
errMsg: "azure SD configuration requires a subscription_id",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "azure_tenant_id_missing.bad.yml",
|
|
|
|
errMsg: "azure SD configuration requires a tenant_id",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "azure_authentication_method.bad.yml",
|
|
|
|
errMsg: "unknown authentication_type \"invalid\". Supported types are \"OAuth\" or \"ManagedIdentity\"",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "empty_scrape_config.bad.yml",
|
|
|
|
errMsg: "empty or null scrape config section",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "empty_rw_config.bad.yml",
|
|
|
|
errMsg: "empty or null remote write config section",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "empty_rr_config.bad.yml",
|
|
|
|
errMsg: "empty or null remote read config section",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "empty_target_relabel_config.bad.yml",
|
|
|
|
errMsg: "empty or null target relabeling rule",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "empty_metric_relabel_config.bad.yml",
|
|
|
|
errMsg: "empty or null metric relabeling rule",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "empty_alert_relabel_config.bad.yml",
|
|
|
|
errMsg: "empty or null alert relabeling rule",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "empty_alertmanager_relabel_config.bad.yml",
|
|
|
|
errMsg: "empty or null Alertmanager target relabeling rule",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "empty_rw_relabel_config.bad.yml",
|
|
|
|
errMsg: "empty or null relabeling rule in remote write config",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "empty_static_config.bad.yml",
|
|
|
|
errMsg: "empty or null section in static_configs",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "hetzner_role.bad.yml",
|
|
|
|
errMsg: "unknown role",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "eureka_no_server.bad.yml",
|
|
|
|
errMsg: "empty or null eureka server",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "eureka_invalid_server.bad.yml",
|
|
|
|
errMsg: "invalid eureka server URL",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "scaleway_role.bad.yml",
|
|
|
|
errMsg: `unknown role "invalid"`,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "scaleway_no_secret.bad.yml",
|
|
|
|
errMsg: "one of secret_key & secret_key_file must be configured",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "scaleway_two_secrets.bad.yml",
|
|
|
|
errMsg: "at most one of secret_key & secret_key_file must be configured",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "scrape_body_size_limit.bad.yml",
|
|
|
|
errMsg: "units: unknown unit in 100",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "http_url_no_scheme.bad.yml",
|
|
|
|
errMsg: "URL scheme must be 'http' or 'https'",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "http_url_no_host.bad.yml",
|
|
|
|
errMsg: "host is missing in URL",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
filename: "http_url_bad_scheme.bad.yml",
|
|
|
|
errMsg: "URL scheme must be 'http' or 'https'",
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestBadConfigs(t *testing.T) {
|
|
|
|
for _, ee := range expectedErrors {
|
|
|
|
_, err := LoadFile("testdata/"+ee.filename, false, log.NewNopLogger())
|
|
|
|
require.Error(t, err, "%s", ee.filename)
|
|
|
|
require.Contains(t, err.Error(), ee.errMsg,
|
|
|
|
"Expected error for %s to contain %q but got: %s", ee.filename, ee.errMsg, err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestBadStaticConfigsJSON(t *testing.T) {
|
|
|
|
content, err := ioutil.ReadFile("testdata/static_config.bad.json")
|
|
|
|
require.NoError(t, err)
|
Refactor SD configuration to remove `config` dependency (#3629)
* refactor: move targetGroup struct and CheckOverflow() to their own package
* refactor: move auth and security related structs to a utility package, fix import error in utility package
* refactor: Azure SD, remove SD struct from config
* refactor: DNS SD, remove SD struct from config into dns package
* refactor: ec2 SD, move SD struct from config into the ec2 package
* refactor: file SD, move SD struct from config to file discovery package
* refactor: gce, move SD struct from config to gce discovery package
* refactor: move HTTPClientConfig and URL into util/config, fix import error in httputil
* refactor: consul, move SD struct from config into consul discovery package
* refactor: marathon, move SD struct from config into marathon discovery package
* refactor: triton, move SD struct from config to triton discovery package, fix test
* refactor: zookeeper, move SD structs from config to zookeeper discovery package
* refactor: openstack, remove SD struct from config, move into openstack discovery package
* refactor: kubernetes, move SD struct from config into kubernetes discovery package
* refactor: notifier, use targetgroup package instead of config
* refactor: tests for file, marathon, triton SD - use targetgroup package instead of config.TargetGroup
* refactor: retrieval, use targetgroup package instead of config.TargetGroup
* refactor: storage, use config util package
* refactor: discovery manager, use targetgroup package instead of config.TargetGroup
* refactor: use HTTPClient and TLS config from configUtil instead of config
* refactor: tests, use targetgroup package instead of config.TargetGroup
* refactor: fix tagetgroup.Group pointers that were removed by mistake
* refactor: openstack, kubernetes: drop prefixes
* refactor: remove import aliases forced due to vscode bug
* refactor: move main SD struct out of config into discovery/config
* refactor: rename configUtil to config_util
* refactor: rename yamlUtil to yaml_config
* refactor: kubernetes, remove prefixes
* refactor: move the TargetGroup package to discovery/
* refactor: fix order of imports
7 years ago
|
|
|
var tg targetgroup.Group
|
|
|
|
err = json.Unmarshal(content, &tg)
|
|
|
|
require.Error(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestBadStaticConfigsYML(t *testing.T) {
|
|
|
|
content, err := ioutil.ReadFile("testdata/static_config.bad.yml")
|
|
|
|
require.NoError(t, err)
|
|
|
|
var tg targetgroup.Group
|
|
|
|
err = yaml.UnmarshalStrict(content, &tg)
|
|
|
|
require.Error(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestEmptyConfig(t *testing.T) {
|
|
|
|
c, err := Load("", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
exp := DefaultConfig
|
|
|
|
require.Equal(t, exp, *c)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestExpandExternalLabels(t *testing.T) {
|
|
|
|
// Cleanup ant TEST env variable that could exist on the system.
|
|
|
|
os.Setenv("TEST", "")
|
|
|
|
|
|
|
|
c, err := LoadFile("testdata/external_labels.good.yml", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, labels.Label{Name: "bar", Value: "foo"}, c.GlobalConfig.ExternalLabels[0])
|
|
|
|
require.Equal(t, labels.Label{Name: "baz", Value: "foo${TEST}bar"}, c.GlobalConfig.ExternalLabels[1])
|
|
|
|
require.Equal(t, labels.Label{Name: "foo", Value: "${TEST}"}, c.GlobalConfig.ExternalLabels[2])
|
|
|
|
|
|
|
|
c, err = LoadFile("testdata/external_labels.good.yml", true, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, labels.Label{Name: "bar", Value: "foo"}, c.GlobalConfig.ExternalLabels[0])
|
|
|
|
require.Equal(t, labels.Label{Name: "baz", Value: "foobar"}, c.GlobalConfig.ExternalLabels[1])
|
|
|
|
require.Equal(t, labels.Label{Name: "foo", Value: ""}, c.GlobalConfig.ExternalLabels[2])
|
|
|
|
|
|
|
|
os.Setenv("TEST", "TestValue")
|
|
|
|
c, err = LoadFile("testdata/external_labels.good.yml", true, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, labels.Label{Name: "bar", Value: "foo"}, c.GlobalConfig.ExternalLabels[0])
|
|
|
|
require.Equal(t, labels.Label{Name: "baz", Value: "fooTestValuebar"}, c.GlobalConfig.ExternalLabels[1])
|
|
|
|
require.Equal(t, labels.Label{Name: "foo", Value: "TestValue"}, c.GlobalConfig.ExternalLabels[2])
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestEmptyGlobalBlock(t *testing.T) {
|
|
|
|
c, err := Load("global:\n", false, log.NewNopLogger())
|
|
|
|
require.NoError(t, err)
|
|
|
|
exp := DefaultConfig
|
|
|
|
require.Equal(t, exp, *c)
|
|
|
|
}
|
|
|
|
|
|
|
|
func kubernetesSDHostURL() config.URL {
|
|
|
|
tURL, _ := url.Parse("https://localhost:1234")
|
|
|
|
return config.URL{URL: tURL}
|
|
|
|
}
|