mirror of https://github.com/k3s-io/k3s
245 lines
7.7 KiB
Go
245 lines
7.7 KiB
Go
/*
|
|
Copyright 2016 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package e2e
|
|
|
|
import (
|
|
"encoding/json"
|
|
"fmt"
|
|
"strconv"
|
|
"time"
|
|
|
|
meta_v1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
"k8s.io/apimachinery/pkg/fields"
|
|
"k8s.io/kubernetes/pkg/api"
|
|
"k8s.io/kubernetes/test/e2e/framework"
|
|
|
|
. "github.com/onsi/ginkgo"
|
|
. "github.com/onsi/gomega"
|
|
)
|
|
|
|
const (
|
|
// esRetryTimeout is how long to keep retrying requesting elasticsearch for status information.
|
|
esRetryTimeout = 5 * time.Minute
|
|
// esRetryDelay is how much time to wait between two attempts to send a request to elasticsearch
|
|
esRetryDelay = 5 * time.Second
|
|
)
|
|
|
|
type esLogsProvider struct {
|
|
Framework *framework.Framework
|
|
}
|
|
|
|
func newEsLogsProvider(f *framework.Framework) (*esLogsProvider, error) {
|
|
return &esLogsProvider{Framework: f}, nil
|
|
}
|
|
|
|
func (logsProvider *esLogsProvider) FluentdApplicationName() string {
|
|
return "fluentd-es"
|
|
}
|
|
|
|
// Ensures that elasticsearch is running and ready to serve requests
|
|
func (logsProvider *esLogsProvider) EnsureWorking() error {
|
|
f := logsProvider.Framework
|
|
// Check for the existence of the Elasticsearch service.
|
|
By("Checking the Elasticsearch service exists.")
|
|
s := f.ClientSet.Core().Services(api.NamespaceSystem)
|
|
// Make a few attempts to connect. This makes the test robust against
|
|
// being run as the first e2e test just after the e2e cluster has been created.
|
|
var err error
|
|
for start := time.Now(); time.Since(start) < esRetryTimeout; time.Sleep(esRetryDelay) {
|
|
if _, err = s.Get("elasticsearch-logging", meta_v1.GetOptions{}); err == nil {
|
|
break
|
|
}
|
|
framework.Logf("Attempt to check for the existence of the Elasticsearch service failed after %v", time.Since(start))
|
|
}
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Wait for the Elasticsearch pods to enter the running state.
|
|
By("Checking to make sure the Elasticsearch pods are running")
|
|
labelSelector := fields.SelectorFromSet(fields.Set(map[string]string{"k8s-app": "elasticsearch-logging"})).String()
|
|
options := meta_v1.ListOptions{LabelSelector: labelSelector}
|
|
pods, err := f.ClientSet.Core().Pods(api.NamespaceSystem).List(options)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
for _, pod := range pods.Items {
|
|
err = framework.WaitForPodRunningInNamespace(f.ClientSet, &pod)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
}
|
|
|
|
By("Checking to make sure we are talking to an Elasticsearch service.")
|
|
// Perform a few checks to make sure this looks like an Elasticsearch cluster.
|
|
var statusCode int
|
|
err = nil
|
|
var body []byte
|
|
for start := time.Now(); time.Since(start) < esRetryTimeout; time.Sleep(esRetryDelay) {
|
|
proxyRequest, errProxy := framework.GetServicesProxyRequest(f.ClientSet, f.ClientSet.Core().RESTClient().Get())
|
|
if errProxy != nil {
|
|
framework.Logf("After %v failed to get services proxy request: %v", time.Since(start), errProxy)
|
|
continue
|
|
}
|
|
// Query against the root URL for Elasticsearch.
|
|
response := proxyRequest.Namespace(api.NamespaceSystem).
|
|
Name("elasticsearch-logging").
|
|
Do()
|
|
err = response.Error()
|
|
response.StatusCode(&statusCode)
|
|
|
|
if err != nil {
|
|
framework.Logf("After %v proxy call to elasticsearch-loigging failed: %v", time.Since(start), err)
|
|
continue
|
|
}
|
|
if int(statusCode) != 200 {
|
|
framework.Logf("After %v Elasticsearch cluster has a bad status: %v", time.Since(start), statusCode)
|
|
continue
|
|
}
|
|
break
|
|
}
|
|
Expect(err).NotTo(HaveOccurred())
|
|
if int(statusCode) != 200 {
|
|
framework.Failf("Elasticsearch cluster has a bad status: %v", statusCode)
|
|
}
|
|
|
|
// Now assume we really are talking to an Elasticsearch instance.
|
|
// Check the cluster health.
|
|
By("Checking health of Elasticsearch service.")
|
|
healthy := false
|
|
for start := time.Now(); time.Since(start) < esRetryTimeout; time.Sleep(esRetryDelay) {
|
|
proxyRequest, errProxy := framework.GetServicesProxyRequest(f.ClientSet, f.ClientSet.Core().RESTClient().Get())
|
|
if errProxy != nil {
|
|
framework.Logf("After %v failed to get services proxy request: %v", time.Since(start), errProxy)
|
|
continue
|
|
}
|
|
body, err = proxyRequest.Namespace(api.NamespaceSystem).
|
|
Name("elasticsearch-logging").
|
|
Suffix("_cluster/health").
|
|
Param("level", "indices").
|
|
DoRaw()
|
|
if err != nil {
|
|
continue
|
|
}
|
|
health := make(map[string]interface{})
|
|
err := json.Unmarshal(body, &health)
|
|
if err != nil {
|
|
framework.Logf("Bad json response from elasticsearch: %v", err)
|
|
continue
|
|
}
|
|
statusIntf, ok := health["status"]
|
|
if !ok {
|
|
framework.Logf("No status field found in cluster health response: %v", health)
|
|
continue
|
|
}
|
|
status := statusIntf.(string)
|
|
if status != "green" && status != "yellow" {
|
|
framework.Logf("Cluster health has bad status: %v", health)
|
|
continue
|
|
}
|
|
if err == nil && ok {
|
|
healthy = true
|
|
break
|
|
}
|
|
}
|
|
if !healthy {
|
|
return fmt.Errorf("after %v elasticsearch cluster is not healthy", esRetryTimeout)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (logsProvider *esLogsProvider) ReadEntries(pod *loggingPod) []*logEntry {
|
|
f := logsProvider.Framework
|
|
|
|
proxyRequest, errProxy := framework.GetServicesProxyRequest(f.ClientSet, f.ClientSet.Core().RESTClient().Get())
|
|
if errProxy != nil {
|
|
framework.Logf("Failed to get services proxy request: %v", errProxy)
|
|
return nil
|
|
}
|
|
|
|
// Ask Elasticsearch to return all the log lines that were tagged with the
|
|
// pod name. Ask for ten times as many log lines because duplication is possible.
|
|
body, err := proxyRequest.Namespace(api.NamespaceSystem).
|
|
Name("elasticsearch-logging").
|
|
Suffix("_search").
|
|
// TODO: Change filter to only match records from current test run
|
|
// after fluent-plugin-kubernetes_metadata_filter is enabled
|
|
// and optimize current query
|
|
Param("q", fmt.Sprintf("tag:*%s*", pod.Name)).
|
|
// Ask for more in case we included some unrelated records in our query
|
|
Param("size", strconv.Itoa(pod.ExpectedLinesNumber*10)).
|
|
DoRaw()
|
|
if err != nil {
|
|
framework.Logf("Failed to make proxy call to elasticsearch-logging: %v", err)
|
|
return nil
|
|
}
|
|
|
|
var response map[string]interface{}
|
|
err = json.Unmarshal(body, &response)
|
|
if err != nil {
|
|
framework.Logf("Failed to unmarshal response: %v", err)
|
|
return nil
|
|
}
|
|
|
|
hits, ok := response["hits"].(map[string]interface{})
|
|
if !ok {
|
|
framework.Logf("response[hits] not of the expected type: %T", response["hits"])
|
|
return nil
|
|
}
|
|
|
|
h, ok := hits["hits"].([]interface{})
|
|
if !ok {
|
|
framework.Logf("Hits not of the expected type: %T", hits["hits"])
|
|
return nil
|
|
}
|
|
|
|
entries := []*logEntry{}
|
|
// Iterate over the hits and populate the observed array.
|
|
for _, e := range h {
|
|
l, ok := e.(map[string]interface{})
|
|
if !ok {
|
|
framework.Logf("Element of hit not of expected type: %T", e)
|
|
continue
|
|
}
|
|
|
|
source, ok := l["_source"].(map[string]interface{})
|
|
if !ok {
|
|
framework.Logf("_source not of the expected type: %T", l["_source"])
|
|
continue
|
|
}
|
|
|
|
msg, ok := source["log"].(string)
|
|
if !ok {
|
|
framework.Logf("Log not of the expected type: %T", source["log"])
|
|
continue
|
|
}
|
|
|
|
timestampString, ok := source["@timestamp"].(string)
|
|
if !ok {
|
|
framework.Logf("Timestamp not of the expected type: %T", source["@timestamp"])
|
|
continue
|
|
}
|
|
timestamp, err := time.Parse(time.RFC3339, timestampString)
|
|
if err != nil {
|
|
framework.Logf("Timestamp was not in correct format: %s", timestampString)
|
|
continue
|
|
}
|
|
|
|
entries = append(entries, &logEntry{
|
|
Payload: msg,
|
|
Timestamp: timestamp,
|
|
})
|
|
}
|
|
|
|
return entries
|
|
}
|