mirror of https://github.com/k3s-io/k3s
Merge pull request #67783 from dashpole/fix_eviction_tests
Automatic merge from submit-queue. If you want to cherry-pick this change to another branch, please follow the instructions here: https://github.com/kubernetes/community/blob/master/contributors/devel/cherry-picks.md. Fix eviction test panics **What this PR does / why we need it**: https://k8s-testgrid.appspot.com/sig-node-kubelet#kubelet-serial-gce-e2e&include-filter-by-regex=EvictionOrdering e2e_node tests for eviction are failing because the BeforeEach and AfterEach added in are not correctly placed inside of a context. This PR moves the BeforeEach and AfterEach inside the context. **Which issue(s) this PR fixes** *(optional, in `fixes #<issue number>(, fixes #<issue_number>, ...)` format, will close the issue(s) when PR gets merged)*: Issue: https://github.com/kubernetes/kubernetes/issues/66357 **Release note**: ```release-note NONE ``` /sig node /kind bug /assign @yujuhong @Random-Liupull/8/head
commit
b90ca3e2f9
|
@ -289,16 +289,6 @@ var _ = framework.KubeDescribe("PriorityMemoryEvictionOrdering [Slow] [Serial] [
|
||||||
highPriorityClassName := f.BaseName + "-high-priority"
|
highPriorityClassName := f.BaseName + "-high-priority"
|
||||||
highPriority := int32(999999999)
|
highPriority := int32(999999999)
|
||||||
|
|
||||||
BeforeEach(func() {
|
|
||||||
_, err := f.ClientSet.SchedulingV1beta1().PriorityClasses().Create(&schedulerapi.PriorityClass{ObjectMeta: metav1.ObjectMeta{Name: highPriorityClassName}, Value: highPriority})
|
|
||||||
Expect(err == nil || errors.IsAlreadyExists(err)).To(BeTrue())
|
|
||||||
})
|
|
||||||
|
|
||||||
AfterEach(func() {
|
|
||||||
err := f.ClientSet.SchedulingV1beta1().PriorityClasses().Delete(highPriorityClassName, &metav1.DeleteOptions{})
|
|
||||||
Expect(err).NotTo(HaveOccurred())
|
|
||||||
})
|
|
||||||
|
|
||||||
Context(fmt.Sprintf(testContextFmt, expectedNodeCondition), func() {
|
Context(fmt.Sprintf(testContextFmt, expectedNodeCondition), func() {
|
||||||
tempSetCurrentKubeletConfig(f, func(initialConfig *kubeletconfig.KubeletConfiguration) {
|
tempSetCurrentKubeletConfig(f, func(initialConfig *kubeletconfig.KubeletConfiguration) {
|
||||||
memoryConsumed := resource.MustParse("600Mi")
|
memoryConsumed := resource.MustParse("600Mi")
|
||||||
|
@ -310,6 +300,14 @@ var _ = framework.KubeDescribe("PriorityMemoryEvictionOrdering [Slow] [Serial] [
|
||||||
initialConfig.EvictionHard = map[string]string{"memory.available": fmt.Sprintf("%d", availableBytes-uint64(memoryConsumed.Value()))}
|
initialConfig.EvictionHard = map[string]string{"memory.available": fmt.Sprintf("%d", availableBytes-uint64(memoryConsumed.Value()))}
|
||||||
initialConfig.EvictionMinimumReclaim = map[string]string{}
|
initialConfig.EvictionMinimumReclaim = map[string]string{}
|
||||||
})
|
})
|
||||||
|
BeforeEach(func() {
|
||||||
|
_, err := f.ClientSet.SchedulingV1beta1().PriorityClasses().Create(&schedulerapi.PriorityClass{ObjectMeta: metav1.ObjectMeta{Name: highPriorityClassName}, Value: highPriority})
|
||||||
|
Expect(err == nil || errors.IsAlreadyExists(err)).To(BeTrue())
|
||||||
|
})
|
||||||
|
AfterEach(func() {
|
||||||
|
err := f.ClientSet.SchedulingV1beta1().PriorityClasses().Delete(highPriorityClassName, &metav1.DeleteOptions{})
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
})
|
||||||
specs := []podEvictSpec{
|
specs := []podEvictSpec{
|
||||||
{
|
{
|
||||||
evictionPriority: 2,
|
evictionPriority: 2,
|
||||||
|
@ -348,16 +346,6 @@ var _ = framework.KubeDescribe("PriorityLocalStorageEvictionOrdering [Slow] [Ser
|
||||||
highPriorityClassName := f.BaseName + "-high-priority"
|
highPriorityClassName := f.BaseName + "-high-priority"
|
||||||
highPriority := int32(999999999)
|
highPriority := int32(999999999)
|
||||||
|
|
||||||
BeforeEach(func() {
|
|
||||||
_, err := f.ClientSet.SchedulingV1beta1().PriorityClasses().Create(&schedulerapi.PriorityClass{ObjectMeta: metav1.ObjectMeta{Name: highPriorityClassName}, Value: highPriority})
|
|
||||||
Expect(err == nil || errors.IsAlreadyExists(err)).To(BeTrue())
|
|
||||||
})
|
|
||||||
|
|
||||||
AfterEach(func() {
|
|
||||||
err := f.ClientSet.SchedulingV1beta1().PriorityClasses().Delete(highPriorityClassName, &metav1.DeleteOptions{})
|
|
||||||
Expect(err).NotTo(HaveOccurred())
|
|
||||||
})
|
|
||||||
|
|
||||||
Context(fmt.Sprintf(testContextFmt, expectedNodeCondition), func() {
|
Context(fmt.Sprintf(testContextFmt, expectedNodeCondition), func() {
|
||||||
tempSetCurrentKubeletConfig(f, func(initialConfig *kubeletconfig.KubeletConfiguration) {
|
tempSetCurrentKubeletConfig(f, func(initialConfig *kubeletconfig.KubeletConfiguration) {
|
||||||
diskConsumed := resource.MustParse("350Mi")
|
diskConsumed := resource.MustParse("350Mi")
|
||||||
|
@ -369,6 +357,14 @@ var _ = framework.KubeDescribe("PriorityLocalStorageEvictionOrdering [Slow] [Ser
|
||||||
initialConfig.EvictionHard = map[string]string{"nodefs.available": fmt.Sprintf("%d", availableBytes-uint64(diskConsumed.Value()))}
|
initialConfig.EvictionHard = map[string]string{"nodefs.available": fmt.Sprintf("%d", availableBytes-uint64(diskConsumed.Value()))}
|
||||||
initialConfig.EvictionMinimumReclaim = map[string]string{}
|
initialConfig.EvictionMinimumReclaim = map[string]string{}
|
||||||
})
|
})
|
||||||
|
BeforeEach(func() {
|
||||||
|
_, err := f.ClientSet.SchedulingV1beta1().PriorityClasses().Create(&schedulerapi.PriorityClass{ObjectMeta: metav1.ObjectMeta{Name: highPriorityClassName}, Value: highPriority})
|
||||||
|
Expect(err == nil || errors.IsAlreadyExists(err)).To(BeTrue())
|
||||||
|
})
|
||||||
|
AfterEach(func() {
|
||||||
|
err := f.ClientSet.SchedulingV1beta1().PriorityClasses().Delete(highPriorityClassName, &metav1.DeleteOptions{})
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
})
|
||||||
specs := []podEvictSpec{
|
specs := []podEvictSpec{
|
||||||
{
|
{
|
||||||
evictionPriority: 2,
|
evictionPriority: 2,
|
||||||
|
|
Loading…
Reference in New Issue