mirror of https://github.com/k3s-io/k3s
Move kubemark, skew, upgrade, leak logic into hack/e2e.go
parent
e84a8ec45a
commit
7011d8826e
|
@ -1,32 +0,0 @@
|
|||
#!/bin/bash
|
||||
|
||||
# Copyright 2015 The Kubernetes Authors.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
set -o errexit
|
||||
set -o nounset
|
||||
set -o pipefail
|
||||
|
||||
KUBE_ROOT=$(dirname "${BASH_SOURCE}")/../..
|
||||
|
||||
: ${KUBECTL:=${KUBE_ROOT}/cluster/kubectl.sh}
|
||||
: ${KUBE_CONFIG_FILE:="config-test.sh"}
|
||||
|
||||
export KUBECTL KUBE_CONFIG_FILE
|
||||
|
||||
source "${KUBE_ROOT}/cluster/kube-util.sh"
|
||||
|
||||
prepare-e2e
|
||||
|
||||
test-build-release
|
398
hack/e2e.go
398
hack/e2e.go
|
@ -19,10 +19,11 @@ package main
|
|||
|
||||
import (
|
||||
"flag"
|
||||
"fmt"
|
||||
"io/ioutil"
|
||||
"log"
|
||||
"os"
|
||||
"os/exec"
|
||||
"path"
|
||||
"path/filepath"
|
||||
"strconv"
|
||||
"strings"
|
||||
|
@ -30,52 +31,51 @@ import (
|
|||
)
|
||||
|
||||
var (
|
||||
isup = flag.Bool("isup", false, "Check to see if the e2e cluster is up, then exit.")
|
||||
build = flag.Bool("build", false, "If true, build a new release. Otherwise, use whatever is there.")
|
||||
up = flag.Bool("up", false, "If true, start the the e2e cluster. If cluster is already up, recreate it.")
|
||||
push = flag.Bool("push", false, "If true, push to e2e cluster. Has no effect if -up is true.")
|
||||
pushup = flag.Bool("pushup", false, "If true, push to e2e cluster if it's up, otherwise start the e2e cluster.")
|
||||
down = flag.Bool("down", false, "If true, tear down the cluster before exiting.")
|
||||
test = flag.Bool("test", false, "Run Ginkgo tests.")
|
||||
testArgs = flag.String("test_args", "", "Space-separated list of arguments to pass to Ginkgo test runner.")
|
||||
root = flag.String("root", absOrDie(filepath.Clean(filepath.Join(path.Base(os.Args[0]), ".."))), "Root directory of kubernetes repository.")
|
||||
verbose = flag.Bool("v", false, "If true, print all command output.")
|
||||
checkVersionSkew = flag.Bool("check_version_skew", true, ""+
|
||||
"By default, verify that client and server have exact version match. "+
|
||||
"You can explicitly set to false if you're, e.g., testing client changes "+
|
||||
"for which the server version doesn't make a difference.")
|
||||
build = flag.Bool("build", false, "If true, build a new release. Otherwise, use whatever is there.")
|
||||
checkNodeCount = flag.Bool("check_node_count", true, ""+
|
||||
"By default, verify that the cluster has at least two nodes."+
|
||||
"You can explicitly set to false if you're, e.g., testing single-node clusters "+
|
||||
"for which the node count is supposed to be one.")
|
||||
|
||||
ctlCmd = flag.String("ctl", "", "If nonempty, pass this as an argument, and call kubectl. Implies -v. (-test, -cfg, -ctl are mutually exclusive)")
|
||||
checkVersionSkew = flag.Bool("check_version_skew", true, ""+
|
||||
"By default, verify that client and server have exact version match. "+
|
||||
"You can explicitly set to false if you're, e.g., testing client changes "+
|
||||
"for which the server version doesn't make a difference.")
|
||||
checkLeakedResources = flag.Bool("check_leaked_resources", false, "Ensure project ends with the same resources")
|
||||
ctlCmd = flag.String("ctl", "", "If nonempty, pass this as an argument, and call kubectl. Implies -v.")
|
||||
down = flag.Bool("down", false, "If true, tear down the cluster before exiting.")
|
||||
dump = flag.String("dump", "", "If set, dump cluster logs to this location")
|
||||
kubemark = flag.Bool("kubemark", false, "If true, run kubemark tests.")
|
||||
isup = flag.Bool("isup", false, "Check to see if the e2e cluster is up, then exit.")
|
||||
push = flag.Bool("push", false, "If true, push to e2e cluster. Has no effect if -up is true.")
|
||||
pushup = flag.Bool("pushup", false, "If true, push to e2e cluster if it's up, otherwise start the e2e cluster.")
|
||||
skewTests = flag.Bool("skew", false, "If true, run tests in another version at ../kubernetes/hack/e2e.go")
|
||||
testArgs = flag.String("test_args", "", "Space-separated list of arguments to pass to Ginkgo test runner.")
|
||||
test = flag.Bool("test", false, "Run Ginkgo tests.")
|
||||
up = flag.Bool("up", false, "If true, start the the e2e cluster. If cluster is already up, recreate it.")
|
||||
upgradeArgs = flag.String("upgrade_args", "", "If set, run upgrade tests before other tests")
|
||||
verbose = flag.Bool("v", false, "If true, print all command output.")
|
||||
)
|
||||
|
||||
const (
|
||||
minNodeCount = 2
|
||||
)
|
||||
|
||||
func absOrDie(path string) string {
|
||||
out, err := filepath.Abs(path)
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
return out
|
||||
}
|
||||
|
||||
type TestResult struct {
|
||||
Pass int
|
||||
Fail int
|
||||
}
|
||||
|
||||
type ResultsByTest map[string]TestResult
|
||||
|
||||
func main() {
|
||||
log.SetFlags(log.LstdFlags | log.Lshortfile)
|
||||
flag.Parse()
|
||||
|
||||
os.Setenv("KUBECTL", *root+`/cluster/kubectl.sh`+kubectlArgs())
|
||||
cwd, err := os.Getwd()
|
||||
if err != nil {
|
||||
log.Fatalf("Could not get pwd: %v", err)
|
||||
}
|
||||
acwd, err := filepath.Abs(cwd)
|
||||
if err != nil {
|
||||
log.Fatalf("Failed to convert to an absolute path: %v", err)
|
||||
}
|
||||
if !strings.Contains(filepath.Base(acwd), "kubernetes") {
|
||||
// TODO(fejta): cd up into the kubernetes directory
|
||||
log.Fatalf("Must run from kubernetes directory: %v", cwd)
|
||||
}
|
||||
|
||||
if *isup {
|
||||
status := 1
|
||||
|
@ -91,46 +91,97 @@ func main() {
|
|||
if *build {
|
||||
// The build-release script needs stdin to ask the user whether
|
||||
// it's OK to download the docker image.
|
||||
cmd := exec.Command(path.Join(*root, "hack/e2e-internal/build-release.sh"))
|
||||
cmd := exec.Command("make", "quick-release")
|
||||
cmd.Stdin = os.Stdin
|
||||
if !finishRunning("build-release", cmd) {
|
||||
log.Fatal("Error building. Aborting.")
|
||||
}
|
||||
}
|
||||
|
||||
if *pushup {
|
||||
if IsUp() {
|
||||
log.Printf("e2e cluster is up, pushing.")
|
||||
*up = false
|
||||
*push = true
|
||||
} else {
|
||||
log.Printf("e2e cluster is down, creating.")
|
||||
*up = true
|
||||
*push = false
|
||||
if *up && !TearDown() {
|
||||
log.Fatal("Could not tear down previous cluster")
|
||||
}
|
||||
|
||||
beforeResources := ""
|
||||
if *checkLeakedResources {
|
||||
beforeResources = ListResources()
|
||||
}
|
||||
|
||||
os.Setenv("KUBECTL", strings.Join(append([]string{"./cluster/kubectl.sh"}, kubectlArgs()...), " "))
|
||||
|
||||
if *upgradeArgs != "" { // Start the cluster using a previous version.
|
||||
if !UpgradeUp() {
|
||||
log.Fatal("Failed to start cluster to upgrade. Aborting.")
|
||||
}
|
||||
} else { // Start the cluster using this version.
|
||||
if *pushup {
|
||||
if IsUp() {
|
||||
log.Printf("e2e cluster is up, pushing.")
|
||||
*up = false
|
||||
*push = true
|
||||
} else {
|
||||
log.Printf("e2e cluster is down, creating.")
|
||||
*up = true
|
||||
*push = false
|
||||
}
|
||||
}
|
||||
if *up {
|
||||
if !Up() {
|
||||
log.Fatal("Error starting e2e cluster. Aborting.")
|
||||
}
|
||||
} else if *push {
|
||||
if !finishRunning("push", exec.Command("./hack/e2e-internal/e2e-push.sh")) {
|
||||
log.Fatal("Error pushing e2e cluster. Aborting.")
|
||||
}
|
||||
}
|
||||
}
|
||||
if *up {
|
||||
if !Up() {
|
||||
log.Fatal("Error starting e2e cluster. Aborting.")
|
||||
}
|
||||
} else if *push {
|
||||
if !finishRunning("push", exec.Command(path.Join(*root, "hack/e2e-internal/e2e-push.sh"))) {
|
||||
log.Fatal("Error pushing e2e cluster. Aborting.")
|
||||
}
|
||||
|
||||
upResources := ""
|
||||
if *checkLeakedResources {
|
||||
upResources = ListResources()
|
||||
}
|
||||
|
||||
success := true
|
||||
switch {
|
||||
case *ctlCmd != "":
|
||||
|
||||
if *ctlCmd != "" {
|
||||
ctlArgs := strings.Fields(*ctlCmd)
|
||||
os.Setenv("KUBE_CONFIG_FILE", "config-test.sh")
|
||||
success = finishRunning("'kubectl "+*ctlCmd+"'", exec.Command(path.Join(*root, "cluster/kubectl.sh"), ctlArgs...))
|
||||
case *test:
|
||||
success = Test()
|
||||
ctlSuccess := finishRunning("'kubectl "+*ctlCmd+"'", exec.Command("./cluster/kubectl.sh", ctlArgs...))
|
||||
success = success && ctlSuccess
|
||||
}
|
||||
|
||||
if *upgradeArgs != "" {
|
||||
upgradeSuccess := UpgradeTest(*upgradeArgs)
|
||||
success = success && upgradeSuccess
|
||||
}
|
||||
|
||||
if *test {
|
||||
if *skewTests {
|
||||
skewSuccess := SkewTest()
|
||||
success = success && skewSuccess
|
||||
} else {
|
||||
testSuccess := Test()
|
||||
success = success && testSuccess
|
||||
}
|
||||
}
|
||||
|
||||
if *kubemark {
|
||||
kubeSuccess := KubemarkTest()
|
||||
success = success && kubeSuccess
|
||||
}
|
||||
|
||||
if *down {
|
||||
TearDown()
|
||||
if !success && *dump != "" {
|
||||
DumpClusterLogs(*dump)
|
||||
}
|
||||
tearSuccess := TearDown()
|
||||
success = success && tearSuccess
|
||||
}
|
||||
|
||||
if *checkLeakedResources {
|
||||
log.Print("Sleeping for 30 seconds...") // Wait for eventually consistent listing
|
||||
time.Sleep(30 * time.Second)
|
||||
DiffResources(beforeResources, upResources, ListResources(), *dump)
|
||||
}
|
||||
|
||||
if !success {
|
||||
|
@ -138,29 +189,86 @@ func main() {
|
|||
}
|
||||
}
|
||||
|
||||
func writeOrDie(dir, name, data string) string {
|
||||
f, err := os.Create(filepath.Join(dir, name))
|
||||
if err != nil {
|
||||
log.Fatal(err)
|
||||
}
|
||||
if _, err := f.WriteString(data); err != nil {
|
||||
log.Fatal(err)
|
||||
}
|
||||
if err := f.Close(); err != nil {
|
||||
log.Fatal(err)
|
||||
}
|
||||
log.Printf("Created file: %s", f.Name())
|
||||
return f.Name()
|
||||
}
|
||||
|
||||
func DiffResources(before, clusterUp, after, location string) {
|
||||
if location == "" {
|
||||
var err error
|
||||
location, err = ioutil.TempDir("", "e2e-check-resources")
|
||||
if err != nil {
|
||||
log.Fatal(err)
|
||||
}
|
||||
}
|
||||
bp := writeOrDie(location, "gcp-resources-before.txt", before)
|
||||
writeOrDie(location, "gcp-resources-cluster-up.txt", clusterUp)
|
||||
ap := writeOrDie(location, "gcp-resources-after.txt", after)
|
||||
|
||||
cmd := exec.Command("diff", "-sw", "-U0", "-F^\\[.*\\]$", bp, ap)
|
||||
if *verbose {
|
||||
cmd.Stderr = os.Stderr
|
||||
}
|
||||
o, err := cmd.Output()
|
||||
stdout := string(o)
|
||||
writeOrDie(location, "gcp-resources-diff.txt", stdout)
|
||||
if err == nil {
|
||||
return
|
||||
}
|
||||
lines := strings.Split(stdout, "\n")
|
||||
if len(lines) < 3 { // Ignore the +++ and --- header lines
|
||||
return
|
||||
}
|
||||
|
||||
var added []string
|
||||
for _, l := range lines {
|
||||
if strings.HasPrefix(l, "+") {
|
||||
added = append(added, l)
|
||||
}
|
||||
}
|
||||
if len(added) > 0 {
|
||||
log.Printf("Error: %d leaked resources", len(added))
|
||||
log.Fatal(strings.Join(added, "\n"))
|
||||
}
|
||||
}
|
||||
|
||||
func ListResources() string {
|
||||
log.Printf("Listing resources...")
|
||||
cmd := exec.Command("./cluster/gce/list-resources.sh")
|
||||
if *verbose {
|
||||
cmd.Stderr = os.Stderr
|
||||
}
|
||||
stdout, err := cmd.Output()
|
||||
if err != nil {
|
||||
log.Fatalf("Failed to list resources (%s):\n%s", err, stdout)
|
||||
}
|
||||
return string(stdout)
|
||||
}
|
||||
|
||||
func TearDown() bool {
|
||||
return finishRunning("teardown", exec.Command(path.Join(*root, "hack/e2e-internal/e2e-down.sh")))
|
||||
return finishRunning("teardown", exec.Command("./hack/e2e-internal/e2e-down.sh"))
|
||||
}
|
||||
|
||||
// Up brings an e2e cluster up, recreating it if one is already running.
|
||||
func Up() bool {
|
||||
if IsUp() {
|
||||
log.Printf("e2e cluster already running; will teardown")
|
||||
if res := TearDown(); !res {
|
||||
return false
|
||||
}
|
||||
}
|
||||
return finishRunning("up", exec.Command(path.Join(*root, "hack/e2e-internal/e2e-up.sh")))
|
||||
return finishRunning("up", exec.Command("./hack/e2e-internal/e2e-up.sh"))
|
||||
}
|
||||
|
||||
// Ensure that the cluster is large engough to run the e2e tests.
|
||||
func ValidateClusterSize() {
|
||||
if os.Getenv("FEDERATION") == "true" {
|
||||
//TODO(colhom): federated equivalent of ValidateClusterSize
|
||||
return
|
||||
}
|
||||
// Check that there are at least minNodeCount nodes running
|
||||
cmd := exec.Command(path.Join(*root, "hack/e2e-internal/e2e-cluster-size.sh"))
|
||||
cmd := exec.Command("./hack/e2e-internal/e2e-cluster-size.sh")
|
||||
if *verbose {
|
||||
cmd.Stderr = os.Stderr
|
||||
}
|
||||
|
@ -181,7 +289,128 @@ func ValidateClusterSize() {
|
|||
|
||||
// Is the e2e cluster up?
|
||||
func IsUp() bool {
|
||||
return finishRunning("get status", exec.Command(path.Join(*root, "hack/e2e-internal/e2e-status.sh")))
|
||||
return finishRunning("get status", exec.Command("./hack/e2e-internal/e2e-status.sh"))
|
||||
}
|
||||
|
||||
func DumpClusterLogs(location string) {
|
||||
log.Printf("Dumping cluster logs to: %v", location)
|
||||
finishRunning("dump cluster logs", exec.Command("./cluster/log-dump.sh", location))
|
||||
}
|
||||
|
||||
func KubemarkTest() bool {
|
||||
// Stop previous run
|
||||
if !finishRunning("Stop kubemark", exec.Command("./test/kubemark/stop-kubemark.sh")) {
|
||||
log.Print("stop kubemark failed")
|
||||
return false
|
||||
}
|
||||
|
||||
// Start new run
|
||||
backups := []string{"NUM_NODES", "MASTER_SIZE"}
|
||||
for _, item := range backups {
|
||||
old, present := os.LookupEnv(item)
|
||||
if present {
|
||||
defer os.Setenv(item, old)
|
||||
} else {
|
||||
defer os.Unsetenv(item)
|
||||
}
|
||||
}
|
||||
os.Setenv("NUM_NODES", os.Getenv("KUBEMARK_NUM_NODES"))
|
||||
os.Setenv("MASTER_SIZE", os.Getenv("KUBEMARK_MASTER_SIZE"))
|
||||
if !finishRunning("Start Kubemark", exec.Command("./test/kubemark/start-kubemark.sh")) {
|
||||
log.Print("Error: start kubemark failed")
|
||||
return false
|
||||
}
|
||||
|
||||
// Run kubemark tests
|
||||
focus, present := os.LookupEnv("KUBEMARK_TESTS")
|
||||
if !present {
|
||||
focus = "starting\\s30\\pods"
|
||||
}
|
||||
test_args := os.Getenv("KUBEMARK_TEST_ARGS")
|
||||
|
||||
if !finishRunning("Run kubemark tests", exec.Command("./test/kubemark/run-e2e-tests.sh", "--ginkgo.focus="+focus, test_args)) {
|
||||
log.Print("Error: run kubemark tests failed")
|
||||
return false
|
||||
}
|
||||
|
||||
// Stop kubemark
|
||||
if !finishRunning("Stop kubemark", exec.Command("./test/kubemark/stop-kubemark.sh")) {
|
||||
log.Print("Error: stop kubemark failed")
|
||||
return false
|
||||
}
|
||||
return true
|
||||
}
|
||||
|
||||
func UpgradeUp() bool {
|
||||
old, err := os.Getwd()
|
||||
if err != nil {
|
||||
log.Printf("Failed to os.Getwd(): %v", err)
|
||||
return false
|
||||
}
|
||||
defer os.Chdir(old)
|
||||
err = os.Chdir("../kubernetes_skew")
|
||||
if err != nil {
|
||||
log.Printf("Failed to cd ../kubernetes_skew: %v", err)
|
||||
return false
|
||||
}
|
||||
return finishRunning("UpgradeUp",
|
||||
exec.Command(
|
||||
"go", "run", "./hack/e2e.go",
|
||||
fmt.Sprintf("--check_version_skew=%t", *checkVersionSkew),
|
||||
fmt.Sprintf("--push=%t", *push),
|
||||
fmt.Sprintf("--pushup=%t", *pushup),
|
||||
fmt.Sprintf("--up=%t", *up),
|
||||
fmt.Sprintf("--v=%t", *verbose),
|
||||
))
|
||||
}
|
||||
|
||||
func UpgradeTest(args string) bool {
|
||||
old, err := os.Getwd()
|
||||
if err != nil {
|
||||
log.Printf("Failed to os.Getwd(): %v", err)
|
||||
return false
|
||||
}
|
||||
defer os.Chdir(old)
|
||||
err = os.Chdir("../kubernetes_skew")
|
||||
if err != nil {
|
||||
log.Printf("Failed to cd ../kubernetes_skew: %v", err)
|
||||
return false
|
||||
}
|
||||
previous, present := os.LookupEnv("E2E_REPORT_PREFIX")
|
||||
if present {
|
||||
defer os.Setenv("E2E_REPORT_PREFIX", previous)
|
||||
} else {
|
||||
defer os.Unsetenv("E2E_REPORT_PREFIX")
|
||||
}
|
||||
os.Setenv("E2E_REPORT_PREFIX", "upgrade")
|
||||
return finishRunning("Upgrade Ginkgo tests",
|
||||
exec.Command(
|
||||
"go", "run", "./hack/e2e.go",
|
||||
"--test",
|
||||
"--test_args="+args,
|
||||
fmt.Sprintf("--v=%t", *verbose),
|
||||
fmt.Sprintf("--check_version_skew=%t", *checkVersionSkew)))
|
||||
}
|
||||
|
||||
func SkewTest() bool {
|
||||
old, err := os.Getwd()
|
||||
if err != nil {
|
||||
log.Printf("Failed to Getwd: %v", err)
|
||||
return false
|
||||
}
|
||||
defer os.Chdir(old)
|
||||
err = os.Chdir("../kubernetes_skew")
|
||||
if err != nil {
|
||||
log.Printf("Failed to cd ../kubernetes_skew: %v", err)
|
||||
return false
|
||||
}
|
||||
return finishRunning("Skewed Ginkgo tests",
|
||||
exec.Command(
|
||||
"go", "run", "./hack/e2e.go",
|
||||
"--test",
|
||||
"--test_args="+*testArgs,
|
||||
fmt.Sprintf("--v=%t", *verbose),
|
||||
fmt.Sprintf("--check_version_skew=%t", *checkVersionSkew)))
|
||||
}
|
||||
|
||||
func Test() bool {
|
||||
|
@ -189,18 +418,18 @@ func Test() bool {
|
|||
log.Fatal("Testing requested, but e2e cluster not up!")
|
||||
}
|
||||
|
||||
if *checkNodeCount {
|
||||
ValidateClusterSize()
|
||||
// TODO(fejta): add a --federated or something similar
|
||||
if os.Getenv("FEDERATION") == "" {
|
||||
if *checkNodeCount {
|
||||
ValidateClusterSize()
|
||||
}
|
||||
return finishRunning("Ginkgo tests", exec.Command("./hack/ginkgo-e2e.sh", strings.Fields(*testArgs)...))
|
||||
}
|
||||
|
||||
if os.Getenv("FEDERATION") == "true" {
|
||||
if *testArgs == "" {
|
||||
*testArgs = "--ginkgo.focus=\\[Feature:Federation\\]"
|
||||
}
|
||||
return finishRunning("Federated Ginkgo tests", exec.Command(filepath.Join(*root, "hack/federated-ginkgo-e2e.sh"), strings.Fields(*testArgs)...))
|
||||
} else {
|
||||
return finishRunning("Ginkgo tests", exec.Command(filepath.Join(*root, "hack/ginkgo-e2e.sh"), strings.Fields(*testArgs)...))
|
||||
if *testArgs == "" {
|
||||
*testArgs = "--ginkgo.focus=\\[Feature:Federation\\]"
|
||||
}
|
||||
return finishRunning("Federated Ginkgo tests", exec.Command("./hack/federated-ginkgo-e2e.sh", strings.Fields(*testArgs)...))
|
||||
}
|
||||
|
||||
func finishRunning(stepName string, cmd *exec.Cmd) bool {
|
||||
|
@ -222,10 +451,9 @@ func finishRunning(stepName string, cmd *exec.Cmd) bool {
|
|||
|
||||
// returns either "", or a list of args intended for appending with the
|
||||
// kubectl command (beginning with a space).
|
||||
func kubectlArgs() string {
|
||||
args := []string{""}
|
||||
if *checkVersionSkew {
|
||||
args = append(args, "--match-server-version")
|
||||
func kubectlArgs() []string {
|
||||
if !*checkVersionSkew {
|
||||
return []string{}
|
||||
}
|
||||
return strings.Join(args, " ")
|
||||
return []string{"--match-server-version"}
|
||||
}
|
||||
|
|
|
@ -161,35 +161,6 @@ function install_google_cloud_sdk_tarball() {
|
|||
export PATH=${install_dir}/google-cloud-sdk/bin:${PATH}
|
||||
}
|
||||
|
||||
# Only call after attempting to bring the cluster up. Don't call after
|
||||
# bringing the cluster down.
|
||||
function dump_cluster_logs_and_exit() {
|
||||
local -r exit_status=$?
|
||||
dump_cluster_logs
|
||||
if [[ "${USE_KUBEMARK:-}" == "true" ]]; then
|
||||
# If we tried to bring the Kubemark cluster up, make a courtesy
|
||||
# attempt to bring it down so we're not leaving resources around.
|
||||
./test/kubemark/stop-kubemark.sh || true
|
||||
fi
|
||||
if [[ "${E2E_DOWN,,}" == "true" ]]; then
|
||||
# If we tried to bring the cluster up, make a courtesy attempt
|
||||
# to bring the cluster down so we're not leaving resources
|
||||
# around. Unlike later, don't sleep beforehand, though. (We're
|
||||
# just trying to tear down as many resources as we can as fast
|
||||
# as possible and don't even know if we brought the master up.)
|
||||
go run ./hack/e2e.go ${E2E_OPT:-} -v --down || true
|
||||
fi
|
||||
exit ${exit_status}
|
||||
}
|
||||
|
||||
# Only call after attempting to bring the cluster up. Don't call after
|
||||
# bringing the cluster down.
|
||||
function dump_cluster_logs() {
|
||||
if [[ -x "cluster/log-dump.sh" ]]; then
|
||||
./cluster/log-dump.sh "${ARTIFACTS}"
|
||||
fi
|
||||
}
|
||||
|
||||
# Figures out the builtin k8s version of a GCI image.
|
||||
function get_gci_k8s_version() {
|
||||
local -r image_description=$(gcloud compute images describe ${KUBE_GCE_MASTER_IMAGE} --project=${KUBE_GCE_MASTER_PROJECT})
|
||||
|
@ -265,24 +236,6 @@ elif [[ -n "${KUBEKINS_SERVICE_ACCOUNT_FILE:-}" ]]; then
|
|||
echo "ERROR: cannot access service account file at: ${KUBEKINS_SERVICE_ACCOUNT_FILE}"
|
||||
fi
|
||||
|
||||
|
||||
function e2e_test() {
|
||||
local -r ginkgo_test_args="${1}"
|
||||
# Check to make sure the cluster is up before running tests, and fail if it's not.
|
||||
go run ./hack/e2e.go ${E2E_OPT:-} -v --isup
|
||||
# Jenkins will look at the junit*.xml files for test failures, so don't exit with a nonzero
|
||||
# error code if it was only tests that failed.
|
||||
go run ./hack/e2e.go ${E2E_OPT:-} -v --test \
|
||||
${ginkgo_test_args:+--test_args="${ginkgo_test_args}"} \
|
||||
&& exitcode=0 || exitcode=$?
|
||||
if [[ "${E2E_PUBLISH_GREEN_VERSION:-}" == "true" && ${exitcode} == 0 ]]; then
|
||||
# Use plaintext version file packaged with kubernetes.tar.gz
|
||||
echo "Publish version to ci/latest-green.txt: $(cat version)"
|
||||
gsutil cp ./version "gs://${KUBE_GCS_DEV_RELEASE_BUCKET}/ci/latest-green.txt"
|
||||
fi
|
||||
return ${exitcode}
|
||||
}
|
||||
|
||||
echo "--------------------------------------------------------------------------------"
|
||||
echo "Test Environment:"
|
||||
printenv | sort
|
||||
|
@ -360,47 +313,6 @@ case "${KUBERNETES_PROVIDER}" in
|
|||
;;
|
||||
esac
|
||||
|
||||
cd kubernetes
|
||||
|
||||
# Upload build start time and k8s version to GCS, but not on PR Jenkins.
|
||||
# On PR Jenkins this is done before the build.
|
||||
if [[ ! "${JOB_NAME}" =~ -pull- ]]; then
|
||||
JENKINS_BUILD_STARTED=true bash <(curl -fsS --retry 3 --keepalive-time 2 "https://raw.githubusercontent.com/kubernetes/kubernetes/master/hack/jenkins/upload-to-gcs.sh")
|
||||
fi
|
||||
|
||||
# When run inside Docker, we need to make sure all files are world-readable
|
||||
# (since they will be owned by root on the host).
|
||||
trap "chmod -R o+r '${ARTIFACTS}'" EXIT SIGINT SIGTERM
|
||||
export E2E_REPORT_DIR=${ARTIFACTS}
|
||||
declare -r gcp_list_resources_script="./cluster/gce/list-resources.sh"
|
||||
declare -r gcp_resources_before="${ARTIFACTS}/gcp-resources-before.txt"
|
||||
declare -r gcp_resources_cluster_up="${ARTIFACTS}/gcp-resources-cluster-up.txt"
|
||||
declare -r gcp_resources_after="${ARTIFACTS}/gcp-resources-after.txt"
|
||||
if [[ ( ${KUBERNETES_PROVIDER} == "gce" || ${KUBERNETES_PROVIDER} == "gke" ) && -x "${gcp_list_resources_script}" ]]; then
|
||||
gcp_list_resources="true"
|
||||
# Always pull the script from HEAD, overwriting the local one if it exists.
|
||||
# We do this to pick up fixes if we are running tests from a branch or tag.
|
||||
curl -fsS --retry 3 --keepalive-time 2 "https://raw.githubusercontent.com/kubernetes/kubernetes/master/cluster/gce/list-resources.sh" > "${gcp_list_resources_script}"
|
||||
else
|
||||
gcp_list_resources="false"
|
||||
fi
|
||||
|
||||
### Set up ###
|
||||
if [[ "${E2E_UP,,}" == "true" ]]; then
|
||||
go run ./hack/e2e.go ${E2E_OPT:-} -v --down
|
||||
fi
|
||||
if [[ "${gcp_list_resources}" == "true" ]]; then
|
||||
${gcp_list_resources_script} > "${gcp_resources_before}"
|
||||
fi
|
||||
if [[ "${E2E_UP,,}" == "true" ]]; then
|
||||
# We want to try to gather logs even if kube-up fails, so collect the
|
||||
# result here and fail after dumping logs if it's nonzero.
|
||||
go run ./hack/e2e.go ${E2E_OPT:-} -v --up || dump_cluster_logs_and_exit
|
||||
go run ./hack/e2e.go -v --ctl="version --match-server-version=false"
|
||||
if [[ "${gcp_list_resources}" == "true" ]]; then
|
||||
${gcp_list_resources_script} > "${gcp_resources_cluster_up}"
|
||||
fi
|
||||
fi
|
||||
|
||||
# Allow download & unpack of alternate version of tests, for cross-version & upgrade testing.
|
||||
#
|
||||
|
@ -414,87 +326,75 @@ fi
|
|||
# JENKINS_USE_SKEW_TESTS=true will run tests from the skewed version rather
|
||||
# than the original version.
|
||||
if [[ -n "${JENKINS_PUBLISHED_SKEW_VERSION:-}" ]]; then
|
||||
cd ..
|
||||
mv kubernetes kubernetes_old
|
||||
fetch_published_version_tars "${JENKINS_PUBLISHED_SKEW_VERSION}"
|
||||
cd kubernetes
|
||||
# Upgrade the cluster before running other tests
|
||||
if [[ "${E2E_UPGRADE_TEST:-}" == "true" ]]; then
|
||||
# Add a report prefix for the e2e tests so that the tests don't get overwritten when we run
|
||||
# the rest of the e2es.
|
||||
E2E_REPORT_PREFIX='upgrade' e2e_test "${GINKGO_UPGRADE_TEST_ARGS:-}" || EXIT_CODE=1
|
||||
fi
|
||||
if [[ "${JENKINS_USE_SKEW_TESTS:-}" != "true" ]]; then
|
||||
# Back out into the old tests now that we've downloaded & maybe upgraded.
|
||||
cd ../kubernetes_old
|
||||
# Append kubectl-path of skewed kubectl to test args, since we always
|
||||
# want that to use the skewed kubectl version:
|
||||
#
|
||||
# - for upgrade jobs, we want kubectl to be at the same version as master.
|
||||
# - for client skew tests, we want to use the skewed kubectl (that's what we're testing).
|
||||
GINKGO_TEST_ARGS="${GINKGO_TEST_ARGS:-} --kubectl-path=$(pwd)/../kubernetes/cluster/kubectl.sh"
|
||||
fi
|
||||
mv kubernetes kubernetes_orig
|
||||
fetch_published_version_tars "${JENKINS_PUBLISHED_SKEW_VERSION}"
|
||||
mv kubernetes kubernetes_skew
|
||||
mv kubernetes_orig kubernetes
|
||||
if [[ "${JENKINS_USE_SKEW_TESTS:-}" != "true" ]]; then
|
||||
# Append kubectl-path of skewed kubectl to test args, since we always
|
||||
# # want that to use the skewed kubectl version:
|
||||
# #
|
||||
# # - for upgrade jobs, we want kubectl to be at the same version as
|
||||
# master.
|
||||
# # - for client skew tests, we want to use the skewed kubectl
|
||||
# (that's what we're testing).
|
||||
GINKGO_TEST_ARGS="${GINKGO_TEST_ARGS:-} --kubectl-path=$(pwd)/../kubernetes_skew/cluster/kubectl.sh"
|
||||
fi
|
||||
fi
|
||||
|
||||
cd kubernetes
|
||||
|
||||
# Upload build start time and k8s version to GCS, but not on PR Jenkins.
|
||||
# On PR Jenkins this is done before the build.
|
||||
if [[ ! "${JOB_NAME}" =~ -pull- ]]; then
|
||||
JENKINS_BUILD_STARTED=true bash <(curl -fsS --retry 3 --keepalive-time 2 "https://raw.githubusercontent.com/kubernetes/kubernetes/master/hack/jenkins/upload-to-gcs.sh")
|
||||
fi
|
||||
|
||||
# When run inside Docker, we need to make sure all files are world-readable
|
||||
# (since they will be owned by root on the host).
|
||||
trap "chmod -R o+r '${ARTIFACTS}'" EXIT SIGINT SIGTERM
|
||||
export E2E_REPORT_DIR=${ARTIFACTS}
|
||||
|
||||
e2e_go_args=( \
|
||||
-v \
|
||||
--dump="${ARTIFACTS}" \
|
||||
)
|
||||
|
||||
|
||||
case "${KUBERNETES_PROVIDER}" in
|
||||
gce|gke)
|
||||
e2e_go_args+=(--check_leaked_resources)
|
||||
;;
|
||||
esac
|
||||
|
||||
if [[ "${E2E_UP,,}" == "true" ]]; then
|
||||
e2e_go_args+=(--up --ctl="version --match-server-version=false")
|
||||
fi
|
||||
|
||||
if [[ "${E2E_DOWN,,}" == "true" ]]; then
|
||||
e2e_go_args+=(--down)
|
||||
fi
|
||||
|
||||
if [[ "${E2E_TEST,,}" == "true" ]]; then
|
||||
e2e_test "${GINKGO_TEST_ARGS:-}" || EXIT_CODE=1
|
||||
e2e_go_args+=(--test --test_args="${GINKGO_TEST_ARGS}")
|
||||
fi
|
||||
|
||||
### Start Kubemark ###
|
||||
if [[ "${USE_KUBEMARK:-}" == "true" ]]; then
|
||||
export RUN_FROM_DISTRO=true
|
||||
NUM_NODES_BKP=${NUM_NODES}
|
||||
MASTER_SIZE_BKP=${MASTER_SIZE}
|
||||
./test/kubemark/stop-kubemark.sh
|
||||
NUM_NODES=${KUBEMARK_NUM_NODES:-$NUM_NODES}
|
||||
MASTER_SIZE=${KUBEMARK_MASTER_SIZE:-$MASTER_SIZE}
|
||||
./test/kubemark/start-kubemark.sh || dump_cluster_logs_and_exit
|
||||
# Similarly, if tests fail, we trigger empty set of tests that would trigger storing logs from the base cluster.
|
||||
# We intentionally overwrite the exit-code from `run-e2e-tests.sh` because we want jenkins to look at the
|
||||
# junit.xml results for test failures and not process the exit code. This is needed by jenkins to more gracefully
|
||||
# handle blocking the merge queue as a result of test failure flakes. Infrastructure failures should continue to
|
||||
# exit non-0.
|
||||
# TODO: The above comment is no longer accurate. Need to fix this before
|
||||
# turning xunit off for the postsubmit tests. See: #28200
|
||||
./test/kubemark/run-e2e-tests.sh --ginkgo.focus="${KUBEMARK_TESTS:-starting\s30\spods}" "${KUBEMARK_TEST_ARGS:-}" || dump_cluster_logs_and_exit
|
||||
./test/kubemark/stop-kubemark.sh
|
||||
NUM_NODES=${NUM_NODES_BKP}
|
||||
MASTER_SIZE=${MASTER_SIZE_BKP}
|
||||
unset RUN_FROM_DISTRO
|
||||
unset NUM_NODES_BKP
|
||||
unset MASTER_SIZE_BKP
|
||||
# Optionally run tests from the version in kubernetes_skew
|
||||
if [[ "${JENKINS_USE_SKEW_TESTS:-}" == "true" ]]; then
|
||||
e2e_go_args+=(--skew)
|
||||
fi
|
||||
|
||||
### Clean up ###
|
||||
if [[ "${E2E_DOWN,,}" == "true" ]]; then
|
||||
# Sleep before deleting the cluster to give the controller manager time to
|
||||
# delete any cloudprovider resources still around from the last test.
|
||||
# This is calibrated to allow enough time for 3 attempts to delete the
|
||||
# resources. Each attempt is allocated 5 seconds for requests to the
|
||||
# cloudprovider plus the processingRetryInterval from servicecontroller.go
|
||||
# for the wait between attempts.
|
||||
sleep 30
|
||||
go run ./hack/e2e.go ${E2E_OPT:-} -v --down
|
||||
fi
|
||||
if [[ "${gcp_list_resources}" == "true" ]]; then
|
||||
${gcp_list_resources_script} > "${gcp_resources_after}"
|
||||
# Optionally run upgrade tests before other tests.
|
||||
if [[ "${E2E_UPGRADE_TEST:-}" == "true" ]]; then
|
||||
e2e_go_args+=(--upgrade_args="${GINKGO_UPGRADE_TEST_ARGS}")
|
||||
fi
|
||||
|
||||
# Compare resources if either the cluster was
|
||||
# * started and destroyed (normal e2e)
|
||||
# * neither started nor destroyed (soak test)
|
||||
if [[ "${E2E_UP:-}" == "${E2E_DOWN:-}" && -f "${gcp_resources_before}" && -f "${gcp_resources_after}" ]]; then
|
||||
difference=$(diff -sw -U0 -F'^\[.*\]$' "${gcp_resources_before}" "${gcp_resources_after}") || true
|
||||
noleak=true
|
||||
if [[ -n $(echo "${difference}" | tail -n +3 | grep -E "^\+") ]] && [[ "${FAIL_ON_GCP_RESOURCE_LEAK:-}" == "true" ]]; then
|
||||
noleak=false
|
||||
fi
|
||||
if ! ${noleak} ; then
|
||||
echo "${difference}"
|
||||
echo "!!! FAIL: Google Cloud Platform resources leaked while running tests!"
|
||||
EXIT_CODE=1
|
||||
fi
|
||||
record_command "${STAGE_CLEANUP}" "gcp_resource_leak_check" ${noleak}
|
||||
fi
|
||||
go run ./hack/e2e.go \
|
||||
${E2E_OPT:-} \
|
||||
"${e2e_go_args[@]}"
|
||||
|
||||
exit ${EXIT_CODE}
|
||||
if [[ "${E2E_PUBLISH_GREEN_VERSION:-}" == "true" ]]; then
|
||||
# Use plaintext version file packaged with kubernetes.tar.gz
|
||||
echo "Publish version to ci/latest-green.txt: $(cat version)"
|
||||
gsutil cp ./version "gs://${KUBE_GCS_DEV_RELEASE_BUCKET}/ci/latest-green.txt"
|
||||
fi
|
||||
|
|
|
@ -1,5 +1,6 @@
|
|||
check_version_skew
|
||||
check_leaked_resources
|
||||
check_node_count
|
||||
check_version_skew
|
||||
concurrent_rc_syncs
|
||||
file_content
|
||||
file_mode
|
||||
|
@ -18,6 +19,7 @@ pods_per_node
|
|||
test_args
|
||||
up_to
|
||||
up_to
|
||||
upgrade_args
|
||||
valid_flag
|
||||
retry_time
|
||||
file_content_in_loop
|
||||
|
|
|
@ -42,17 +42,23 @@ EOF
|
|||
sed -i'' -e "s/\"//g" "${RESOURCE_DIRECTORY}/controllers_flags"
|
||||
}
|
||||
|
||||
RUN_FROM_DISTRO=${RUN_FROM_DISTRO:-false}
|
||||
MAKE_DIR="${KUBE_ROOT}/cluster/images/kubemark"
|
||||
|
||||
if [ "${RUN_FROM_DISTRO}" == "false" ]; then
|
||||
# Running from repository
|
||||
echo "Copying kubemark to ${MAKE_DIR}"
|
||||
if [[ -f "${KUBE_ROOT}/_output/release-tars/kubernetes-server-linux-amd64.tar.gz" ]]; then
|
||||
# Running from distro
|
||||
SERVER_TARBALL="${KUBE_ROOT}/_output/release-tars/kubernetes-server-linux-amd64.tar.gz"
|
||||
cp "${KUBE_ROOT}/_output/release-stage/server/linux-amd64/kubernetes/server/bin/kubemark" "${MAKE_DIR}"
|
||||
elif [[ -f "${KUBE_ROOT}/server/kubernetes-server-linux-amd64.tar.gz" ]]; then
|
||||
# Running from an extracted release tarball (kubernetes.tar.gz)
|
||||
SERVER_TARBALL="${KUBE_ROOT}/server/kubernetes-server-linux-amd64.tar.gz"
|
||||
tar \
|
||||
--strip-components=3 \
|
||||
-xzf kubernetes-server-linux-amd64.tar.gz \
|
||||
-C "${MAKE_DIR}" 'kubernetes/server/bin/kubemark' || exit 1
|
||||
else
|
||||
cp "${KUBE_ROOT}/server/kubernetes-server-linux-amd64.tar.gz" "."
|
||||
tar -xzf kubernetes-server-linux-amd64.tar.gz
|
||||
cp "kubernetes/server/bin/kubemark" "${MAKE_DIR}"
|
||||
rm -rf "kubernetes-server-linux-amd64.tar.gz" "kubernetes"
|
||||
echo 'Cannot find kubernetes/server/bin/kubemark binary'
|
||||
exit 1
|
||||
fi
|
||||
|
||||
CURR_DIR=`pwd`
|
||||
|
@ -167,25 +173,14 @@ gcloud compute ssh --zone="${ZONE}" --project="${PROJECT}" "${MASTER_NAME}" \
|
|||
|
||||
writeEnvironmentFiles
|
||||
|
||||
if [ "${RUN_FROM_DISTRO}" == "false" ]; then
|
||||
gcloud compute copy-files --zone="${ZONE}" --project="${PROJECT}" \
|
||||
"${KUBE_ROOT}/_output/release-tars/kubernetes-server-linux-amd64.tar.gz" \
|
||||
"${KUBEMARK_DIRECTORY}/start-kubemark-master.sh" \
|
||||
"${KUBEMARK_DIRECTORY}/configure-kubectl.sh" \
|
||||
"${RESOURCE_DIRECTORY}/apiserver_flags" \
|
||||
"${RESOURCE_DIRECTORY}/scheduler_flags" \
|
||||
"${RESOURCE_DIRECTORY}/controllers_flags" \
|
||||
"${MASTER_NAME}":~
|
||||
else
|
||||
gcloud compute copy-files --zone="${ZONE}" --project="${PROJECT}" \
|
||||
"${KUBE_ROOT}/server/kubernetes-server-linux-amd64.tar.gz" \
|
||||
"${KUBEMARK_DIRECTORY}/start-kubemark-master.sh" \
|
||||
"${KUBEMARK_DIRECTORY}/configure-kubectl.sh" \
|
||||
"${RESOURCE_DIRECTORY}/apiserver_flags" \
|
||||
"${RESOURCE_DIRECTORY}/scheduler_flags" \
|
||||
"${RESOURCE_DIRECTORY}/controllers_flags" \
|
||||
"${MASTER_NAME}":~
|
||||
fi
|
||||
gcloud compute copy-files --zone="${ZONE}" --project="${PROJECT}" \
|
||||
"${SERVER_TARBALL}" \
|
||||
"${KUBEMARK_DIRECTORY}/start-kubemark-master.sh" \
|
||||
"${KUBEMARK_DIRECTORY}/configure-kubectl.sh" \
|
||||
"${RESOURCE_DIRECTORY}/apiserver_flags" \
|
||||
"${RESOURCE_DIRECTORY}/scheduler_flags" \
|
||||
"${RESOURCE_DIRECTORY}/controllers_flags" \
|
||||
"${MASTER_NAME}":~
|
||||
|
||||
gcloud compute ssh "${MASTER_NAME}" --zone="${ZONE}" --project="${PROJECT}" \
|
||||
--command="chmod a+x configure-kubectl.sh && chmod a+x start-kubemark-master.sh && sudo ./start-kubemark-master.sh ${EVENT_STORE_IP:-127.0.0.1}"
|
||||
|
|
Loading…
Reference in New Issue