mirror of https://github.com/k3s-io/k3s
285 lines
9.8 KiB
Go
285 lines
9.8 KiB
Go
/*
|
|
Copyright 2016 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package storage
|
|
|
|
import (
|
|
"fmt"
|
|
"time"
|
|
|
|
. "github.com/onsi/ginkgo"
|
|
. "github.com/onsi/gomega"
|
|
"k8s.io/api/core/v1"
|
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
"k8s.io/apimachinery/pkg/labels"
|
|
utilerrors "k8s.io/apimachinery/pkg/util/errors"
|
|
clientset "k8s.io/client-go/kubernetes"
|
|
"k8s.io/kubernetes/test/e2e/framework"
|
|
"k8s.io/kubernetes/test/e2e/storage/utils"
|
|
)
|
|
|
|
type testBody func(c clientset.Interface, f *framework.Framework, clientPod *v1.Pod, pvc *v1.PersistentVolumeClaim, pv *v1.PersistentVolume)
|
|
type disruptiveTest struct {
|
|
testItStmt string
|
|
runTest testBody
|
|
}
|
|
|
|
const (
|
|
MinNodes = 2
|
|
)
|
|
|
|
var _ = utils.SIGDescribe("PersistentVolumes[Disruptive][Flaky]", func() {
|
|
|
|
f := framework.NewDefaultFramework("disruptive-pv")
|
|
var (
|
|
c clientset.Interface
|
|
ns string
|
|
nfsServerPod *v1.Pod
|
|
nfsPVconfig framework.PersistentVolumeConfig
|
|
pvcConfig framework.PersistentVolumeClaimConfig
|
|
nfsServerIP, clientNodeIP string
|
|
clientNode *v1.Node
|
|
volLabel labels.Set
|
|
selector *metav1.LabelSelector
|
|
)
|
|
|
|
BeforeEach(func() {
|
|
// To protect the NFS volume pod from the kubelet restart, we isolate it on its own node.
|
|
framework.SkipUnlessNodeCountIsAtLeast(MinNodes)
|
|
framework.SkipIfProviderIs("local")
|
|
|
|
c = f.ClientSet
|
|
ns = f.Namespace.Name
|
|
volLabel = labels.Set{framework.VolumeSelectorKey: ns}
|
|
selector = metav1.SetAsLabelSelector(volLabel)
|
|
// Start the NFS server pod.
|
|
_, nfsServerPod, nfsServerIP = framework.NewNFSServer(c, ns, []string{"-G", "777", "/exports"})
|
|
nfsPVconfig = framework.PersistentVolumeConfig{
|
|
NamePrefix: "nfs-",
|
|
Labels: volLabel,
|
|
PVSource: v1.PersistentVolumeSource{
|
|
NFS: &v1.NFSVolumeSource{
|
|
Server: nfsServerIP,
|
|
Path: "/exports",
|
|
ReadOnly: false,
|
|
},
|
|
},
|
|
}
|
|
pvcConfig = framework.PersistentVolumeClaimConfig{
|
|
Annotations: map[string]string{
|
|
v1.BetaStorageClassAnnotation: "",
|
|
},
|
|
Selector: selector,
|
|
}
|
|
// Get the first ready node IP that is not hosting the NFS pod.
|
|
if clientNodeIP == "" {
|
|
framework.Logf("Designating test node")
|
|
nodes := framework.GetReadySchedulableNodesOrDie(c)
|
|
for _, node := range nodes.Items {
|
|
if node.Name != nfsServerPod.Spec.NodeName {
|
|
clientNode = &node
|
|
clientNodeIP = framework.GetNodeExternalIP(clientNode)
|
|
break
|
|
}
|
|
}
|
|
Expect(clientNodeIP).NotTo(BeEmpty())
|
|
}
|
|
})
|
|
|
|
AfterEach(func() {
|
|
framework.DeletePodWithWait(f, c, nfsServerPod)
|
|
})
|
|
|
|
Context("when kube-controller-manager restarts", func() {
|
|
var (
|
|
diskName1, diskName2 string
|
|
err error
|
|
pvConfig1, pvConfig2 framework.PersistentVolumeConfig
|
|
pv1, pv2 *v1.PersistentVolume
|
|
pvSource1, pvSource2 *v1.PersistentVolumeSource
|
|
pvc1, pvc2 *v1.PersistentVolumeClaim
|
|
clientPod *v1.Pod
|
|
)
|
|
|
|
BeforeEach(func() {
|
|
framework.SkipUnlessProviderIs("gce")
|
|
framework.SkipUnlessSSHKeyPresent()
|
|
|
|
By("Initializing first PD with PVPVC binding")
|
|
pvSource1, diskName1 = framework.CreateGCEVolume()
|
|
Expect(err).NotTo(HaveOccurred())
|
|
pvConfig1 = framework.PersistentVolumeConfig{
|
|
NamePrefix: "gce-",
|
|
Labels: volLabel,
|
|
PVSource: *pvSource1,
|
|
Prebind: nil,
|
|
}
|
|
pv1, pvc1, err = framework.CreatePVPVC(c, pvConfig1, pvcConfig, ns, false)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
framework.ExpectNoError(framework.WaitOnPVandPVC(c, ns, pv1, pvc1))
|
|
|
|
By("Initializing second PD with PVPVC binding")
|
|
pvSource2, diskName2 = framework.CreateGCEVolume()
|
|
Expect(err).NotTo(HaveOccurred())
|
|
pvConfig2 = framework.PersistentVolumeConfig{
|
|
NamePrefix: "gce-",
|
|
Labels: volLabel,
|
|
PVSource: *pvSource2,
|
|
Prebind: nil,
|
|
}
|
|
pv2, pvc2, err = framework.CreatePVPVC(c, pvConfig2, pvcConfig, ns, false)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
framework.ExpectNoError(framework.WaitOnPVandPVC(c, ns, pv2, pvc2))
|
|
|
|
By("Attaching both PVC's to a single pod")
|
|
clientPod, err = framework.CreatePod(c, ns, nil, []*v1.PersistentVolumeClaim{pvc1, pvc2}, true, "")
|
|
Expect(err).NotTo(HaveOccurred())
|
|
})
|
|
|
|
AfterEach(func() {
|
|
// Delete client/user pod first
|
|
framework.ExpectNoError(framework.DeletePodWithWait(f, c, clientPod))
|
|
|
|
// Delete PV and PVCs
|
|
if errs := framework.PVPVCCleanup(c, ns, pv1, pvc1); len(errs) > 0 {
|
|
framework.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs))
|
|
}
|
|
pv1, pvc1 = nil, nil
|
|
if errs := framework.PVPVCCleanup(c, ns, pv2, pvc2); len(errs) > 0 {
|
|
framework.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs))
|
|
}
|
|
pv2, pvc2 = nil, nil
|
|
|
|
// Delete the actual disks
|
|
if diskName1 != "" {
|
|
framework.ExpectNoError(framework.DeletePDWithRetry(diskName1))
|
|
}
|
|
if diskName2 != "" {
|
|
framework.ExpectNoError(framework.DeletePDWithRetry(diskName2))
|
|
}
|
|
})
|
|
|
|
It("should delete a bound PVC from a clientPod, restart the kube-control-manager, and ensure the kube-controller-manager does not crash", func() {
|
|
By("Deleting PVC for volume 2")
|
|
err = framework.DeletePersistentVolumeClaim(c, pvc2.Name, ns)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
pvc2 = nil
|
|
|
|
By("Restarting the kube-controller-manager")
|
|
err = framework.RestartControllerManager()
|
|
Expect(err).NotTo(HaveOccurred())
|
|
err = framework.WaitForControllerManagerUp()
|
|
Expect(err).NotTo(HaveOccurred())
|
|
framework.Logf("kube-controller-manager restarted")
|
|
|
|
By("Observing the kube-controller-manager healthy for at least 2 minutes")
|
|
// Continue checking for 2 minutes to make sure kube-controller-manager is healthy
|
|
err = framework.CheckForControllerManagerHealthy(2 * time.Minute)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
})
|
|
|
|
})
|
|
|
|
Context("when kubelet restarts", func() {
|
|
var (
|
|
clientPod *v1.Pod
|
|
pv *v1.PersistentVolume
|
|
pvc *v1.PersistentVolumeClaim
|
|
)
|
|
|
|
BeforeEach(func() {
|
|
framework.Logf("Initializing test spec")
|
|
clientPod, pv, pvc = initTestCase(f, c, nfsPVconfig, pvcConfig, ns, clientNode.Name)
|
|
})
|
|
|
|
AfterEach(func() {
|
|
framework.Logf("Tearing down test spec")
|
|
tearDownTestCase(c, f, ns, clientPod, pvc, pv)
|
|
pv, pvc, clientPod = nil, nil, nil
|
|
})
|
|
|
|
// Test table housing the It() title string and test spec. runTest is type testBody, defined at
|
|
// the start of this file. To add tests, define a function mirroring the testBody signature and assign
|
|
// to runTest.
|
|
disruptiveTestTable := []disruptiveTest{
|
|
{
|
|
testItStmt: "Should test that a file written to the mount before kubelet restart is readable after restart.",
|
|
runTest: utils.TestKubeletRestartsAndRestoresMount,
|
|
},
|
|
{
|
|
testItStmt: "Should test that a volume mounted to a pod that is deleted while the kubelet is down unmounts when the kubelet returns.",
|
|
runTest: utils.TestVolumeUnmountsFromDeletedPod,
|
|
},
|
|
{
|
|
testItStmt: "Should test that a volume mounted to a pod that is force deleted while the kubelet is down unmounts when the kubelet returns.",
|
|
runTest: utils.TestVolumeUnmountsFromForceDeletedPod,
|
|
},
|
|
}
|
|
|
|
// Test loop executes each disruptiveTest iteratively.
|
|
for _, test := range disruptiveTestTable {
|
|
func(t disruptiveTest) {
|
|
It(t.testItStmt, func() {
|
|
By("Executing Spec")
|
|
t.runTest(c, f, clientPod, pvc, pv)
|
|
})
|
|
}(test)
|
|
}
|
|
})
|
|
})
|
|
|
|
// initTestCase initializes spec resources (pv, pvc, and pod) and returns pointers to be consumed
|
|
// by the test.
|
|
func initTestCase(f *framework.Framework, c clientset.Interface, pvConfig framework.PersistentVolumeConfig, pvcConfig framework.PersistentVolumeClaimConfig, ns, nodeName string) (*v1.Pod, *v1.PersistentVolume, *v1.PersistentVolumeClaim) {
|
|
pv, pvc, err := framework.CreatePVPVC(c, pvConfig, pvcConfig, ns, false)
|
|
defer func() {
|
|
if err != nil {
|
|
framework.DeletePersistentVolumeClaim(c, pvc.Name, ns)
|
|
framework.DeletePersistentVolume(c, pv.Name)
|
|
}
|
|
}()
|
|
Expect(err).NotTo(HaveOccurred())
|
|
pod := framework.MakePod(ns, nil, []*v1.PersistentVolumeClaim{pvc}, true, "")
|
|
pod.Spec.NodeName = nodeName
|
|
framework.Logf("Creating NFS client pod.")
|
|
pod, err = c.CoreV1().Pods(ns).Create(pod)
|
|
framework.Logf("NFS client Pod %q created on Node %q", pod.Name, nodeName)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
defer func() {
|
|
if err != nil {
|
|
framework.DeletePodWithWait(f, c, pod)
|
|
}
|
|
}()
|
|
err = framework.WaitForPodRunningInNamespace(c, pod)
|
|
Expect(err).NotTo(HaveOccurred(), fmt.Sprintf("Pod %q timed out waiting for phase: Running", pod.Name))
|
|
// Return created api objects
|
|
pod, err = c.CoreV1().Pods(ns).Get(pod.Name, metav1.GetOptions{})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
pvc, err = c.CoreV1().PersistentVolumeClaims(ns).Get(pvc.Name, metav1.GetOptions{})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
pv, err = c.CoreV1().PersistentVolumes().Get(pv.Name, metav1.GetOptions{})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
return pod, pv, pvc
|
|
}
|
|
|
|
// tearDownTestCase destroy resources created by initTestCase.
|
|
func tearDownTestCase(c clientset.Interface, f *framework.Framework, ns string, client *v1.Pod, pvc *v1.PersistentVolumeClaim, pv *v1.PersistentVolume) {
|
|
// Ignore deletion errors. Failing on them will interrupt test cleanup.
|
|
framework.DeletePodWithWait(f, c, client)
|
|
framework.DeletePersistentVolumeClaim(c, pvc.Name, ns)
|
|
framework.DeletePersistentVolume(c, pv.Name)
|
|
}
|