mirror of https://github.com/k3s-io/k3s
1462 lines
56 KiB
Go
1462 lines
56 KiB
Go
/*
|
|
Copyright 2017 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package deployment
|
|
|
|
import (
|
|
"fmt"
|
|
"reflect"
|
|
"strings"
|
|
"testing"
|
|
|
|
apps "k8s.io/api/apps/v1"
|
|
"k8s.io/api/core/v1"
|
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
"k8s.io/apimachinery/pkg/util/intstr"
|
|
"k8s.io/apimachinery/pkg/util/uuid"
|
|
"k8s.io/apimachinery/pkg/util/wait"
|
|
"k8s.io/client-go/util/retry"
|
|
deploymentutil "k8s.io/kubernetes/pkg/controller/deployment/util"
|
|
"k8s.io/kubernetes/test/integration/framework"
|
|
"k8s.io/utils/pointer"
|
|
)
|
|
|
|
func TestNewDeployment(t *testing.T) {
|
|
s, closeFn, rm, dc, informers, c := dcSetup(t)
|
|
defer closeFn()
|
|
name := "test-new-deployment"
|
|
ns := framework.CreateTestingNamespace(name, s, t)
|
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
|
|
|
replicas := int32(20)
|
|
tester := &deploymentTester{t: t, c: c, deployment: newDeployment(name, ns.Name, replicas)}
|
|
tester.deployment.Spec.MinReadySeconds = 4
|
|
|
|
tester.deployment.Annotations = map[string]string{"test": "should-copy-to-replica-set", v1.LastAppliedConfigAnnotation: "should-not-copy-to-replica-set"}
|
|
var err error
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Create(tester.deployment)
|
|
if err != nil {
|
|
t.Fatalf("failed to create deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
|
|
// Start informer and controllers
|
|
stopCh := make(chan struct{})
|
|
defer close(stopCh)
|
|
informers.Start(stopCh)
|
|
go rm.Run(5, stopCh)
|
|
go dc.Run(5, stopCh)
|
|
|
|
// Wait for the Deployment to be updated to revision 1
|
|
if err := tester.waitForDeploymentRevisionAndImage("1", fakeImage); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Make sure the Deployment completes while manually marking Deployment pods as ready at the same time.
|
|
// Use soft check because this deployment was just created and rolling update strategy might be violated.
|
|
if err := tester.waitForDeploymentCompleteAndMarkPodsReady(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Check new RS annotations
|
|
newRS, err := tester.expectNewReplicaSet()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if newRS.Annotations["test"] != "should-copy-to-replica-set" {
|
|
t.Errorf("expected new ReplicaSet annotations copied from Deployment %s, got: %v", tester.deployment.Name, newRS.Annotations)
|
|
}
|
|
if newRS.Annotations[v1.LastAppliedConfigAnnotation] != "" {
|
|
t.Errorf("expected new ReplicaSet last-applied annotation not copied from Deployment %s", tester.deployment.Name)
|
|
}
|
|
|
|
// New RS should contain pod-template-hash in its selector, label, and template label
|
|
rsHash, err := checkRSHashLabels(newRS)
|
|
if err != nil {
|
|
t.Error(err)
|
|
}
|
|
|
|
// All pods targeted by the deployment should contain pod-template-hash in their labels
|
|
selector, err := metav1.LabelSelectorAsSelector(tester.deployment.Spec.Selector)
|
|
if err != nil {
|
|
t.Fatalf("failed to parse deployment %s selector: %v", name, err)
|
|
}
|
|
pods, err := c.CoreV1().Pods(ns.Name).List(metav1.ListOptions{LabelSelector: selector.String()})
|
|
if err != nil {
|
|
t.Fatalf("failed to list pods of deployment %s: %v", name, err)
|
|
}
|
|
if len(pods.Items) != int(replicas) {
|
|
t.Errorf("expected %d pods, got %d pods", replicas, len(pods.Items))
|
|
}
|
|
podHash, err := checkPodsHashLabel(pods)
|
|
if err != nil {
|
|
t.Error(err)
|
|
}
|
|
if rsHash != podHash {
|
|
t.Errorf("found mismatching pod-template-hash value: rs hash = %s whereas pod hash = %s", rsHash, podHash)
|
|
}
|
|
}
|
|
|
|
// Deployments should support roll out, roll back, and roll over.
|
|
// TODO: drop the rollback portions of this test when extensions/v1beta1 is no longer served
|
|
// and rollback endpoint is no longer supported.
|
|
func TestDeploymentRollingUpdate(t *testing.T) {
|
|
s, closeFn, rm, dc, informers, c := dcSetup(t)
|
|
defer closeFn()
|
|
name := "test-rolling-update-deployment"
|
|
ns := framework.CreateTestingNamespace(name, s, t)
|
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
|
|
|
// Start informer and controllers
|
|
stopCh := make(chan struct{})
|
|
defer close(stopCh)
|
|
informers.Start(stopCh)
|
|
go rm.Run(5, stopCh)
|
|
go dc.Run(5, stopCh)
|
|
|
|
replicas := int32(20)
|
|
tester := &deploymentTester{t: t, c: c, deployment: newDeployment(name, ns.Name, replicas)}
|
|
tester.deployment.Spec.MinReadySeconds = 4
|
|
quarter := intstr.FromString("25%")
|
|
tester.deployment.Spec.Strategy.RollingUpdate = &apps.RollingUpdateDeployment{
|
|
MaxUnavailable: &quarter,
|
|
MaxSurge: &quarter,
|
|
}
|
|
|
|
// Create a deployment.
|
|
var err error
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Create(tester.deployment)
|
|
if err != nil {
|
|
t.Fatalf("failed to create deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
oriImage := tester.deployment.Spec.Template.Spec.Containers[0].Image
|
|
if err := tester.waitForDeploymentRevisionAndImage("1", oriImage); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := tester.waitForDeploymentCompleteAndMarkPodsReady(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// 1. Roll out a new image.
|
|
image := "new-image"
|
|
if oriImage == image {
|
|
t.Fatalf("bad test setup, deployment %s roll out with the same image", tester.deployment.Name)
|
|
}
|
|
imageFn := func(update *apps.Deployment) {
|
|
update.Spec.Template.Spec.Containers[0].Image = image
|
|
}
|
|
tester.deployment, err = tester.updateDeployment(imageFn)
|
|
if err != nil {
|
|
t.Fatalf("failed to update deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
if err := tester.waitForDeploymentRevisionAndImage("2", image); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := tester.waitForDeploymentCompleteAndCheckRollingAndMarkPodsReady(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// 2. Roll back to the last revision.
|
|
revision := int64(0)
|
|
rollback := newDeploymentRollback(tester.deployment.Name, nil, revision)
|
|
if err = c.ExtensionsV1beta1().Deployments(ns.Name).Rollback(rollback); err != nil {
|
|
t.Fatalf("failed to roll back deployment %s to last revision: %v", tester.deployment.Name, err)
|
|
}
|
|
// Wait for the deployment to start rolling back
|
|
if err = tester.waitForDeploymentRollbackCleared(); err != nil {
|
|
t.Fatalf("failed to roll back deployment %s to last revision: %v", tester.deployment.Name, err)
|
|
}
|
|
// Wait for the deployment to be rolled back to the template stored in revision 1 and rolled forward to revision 3.
|
|
if err := tester.waitForDeploymentRevisionAndImage("3", oriImage); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := tester.waitForDeploymentCompleteAndCheckRollingAndMarkPodsReady(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// 3. Roll over a deployment before the previous rolling update finishes.
|
|
image = "dont-finish"
|
|
imageFn = func(update *apps.Deployment) {
|
|
update.Spec.Template.Spec.Containers[0].Image = image
|
|
}
|
|
tester.deployment, err = tester.updateDeployment(imageFn)
|
|
if err != nil {
|
|
t.Fatalf("failed to update deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
if err := tester.waitForDeploymentRevisionAndImage("4", image); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
// We don't mark pods as ready so that rollout won't finish.
|
|
// Before the rollout finishes, trigger another rollout.
|
|
image = "rollover"
|
|
imageFn = func(update *apps.Deployment) {
|
|
update.Spec.Template.Spec.Containers[0].Image = image
|
|
}
|
|
tester.deployment, err = tester.updateDeployment(imageFn)
|
|
if err != nil {
|
|
t.Fatalf("failed to update deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
if err := tester.waitForDeploymentRevisionAndImage("5", image); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := tester.waitForDeploymentCompleteAndCheckRollingAndMarkPodsReady(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
_, allOldRSs, err := deploymentutil.GetOldReplicaSets(tester.deployment, c.AppsV1())
|
|
if err != nil {
|
|
t.Fatalf("failed retrieving old replicasets of deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
for _, oldRS := range allOldRSs {
|
|
if *oldRS.Spec.Replicas != 0 {
|
|
t.Errorf("expected old replicaset %s of deployment %s to have 0 replica, got %d", oldRS.Name, tester.deployment.Name, *oldRS.Spec.Replicas)
|
|
}
|
|
}
|
|
}
|
|
|
|
// selectors are IMMUTABLE for all API versions except apps/v1beta1 and extensions/v1beta1
|
|
func TestDeploymentSelectorImmutability(t *testing.T) {
|
|
s, closeFn, c := dcSimpleSetup(t)
|
|
defer closeFn()
|
|
name := "test-deployment-selector-immutability"
|
|
ns := framework.CreateTestingNamespace(name, s, t)
|
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
|
|
|
tester := &deploymentTester{t: t, c: c, deployment: newDeployment(name, ns.Name, int32(20))}
|
|
var err error
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Create(tester.deployment)
|
|
if err != nil {
|
|
t.Fatalf("failed to create apps/v1 deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
|
|
// test to ensure extensions/v1beta1 selector is mutable
|
|
// TODO: drop the extensions/v1beta1 portion of this test when extensions/v1beta1 is no longer served
|
|
newSelectorLabels := map[string]string{"name_extensions_v1beta1": "test_extensions_v1beta1"}
|
|
deploymentExtensionsV1beta1, err := c.ExtensionsV1beta1().Deployments(ns.Name).Get(name, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed to get extensions/v1beta deployment %s: %v", name, err)
|
|
}
|
|
deploymentExtensionsV1beta1.Spec.Selector.MatchLabels = newSelectorLabels
|
|
deploymentExtensionsV1beta1.Spec.Template.Labels = newSelectorLabels
|
|
updatedDeploymentExtensionsV1beta1, err := c.ExtensionsV1beta1().Deployments(ns.Name).Update(deploymentExtensionsV1beta1)
|
|
if err != nil {
|
|
t.Fatalf("failed to update extensions/v1beta1 deployment %s: %v", deploymentExtensionsV1beta1.Name, err)
|
|
}
|
|
if !reflect.DeepEqual(updatedDeploymentExtensionsV1beta1.Spec.Selector.MatchLabels, newSelectorLabels) {
|
|
t.Errorf("selector should be changed for extensions/v1beta1, expected: %v, got: %v", newSelectorLabels, updatedDeploymentExtensionsV1beta1.Spec.Selector.MatchLabels)
|
|
}
|
|
|
|
// test to ensure apps/v1beta1 selector is mutable
|
|
deploymentAppsV1beta1, err := c.AppsV1beta1().Deployments(ns.Name).Get(updatedDeploymentExtensionsV1beta1.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed to get apps/v1beta1 deployment %s: %v", updatedDeploymentExtensionsV1beta1.Name, err)
|
|
}
|
|
|
|
newSelectorLabels = map[string]string{"name_apps_v1beta1": "test_apps_v1beta1"}
|
|
deploymentAppsV1beta1.Spec.Selector.MatchLabels = newSelectorLabels
|
|
deploymentAppsV1beta1.Spec.Template.Labels = newSelectorLabels
|
|
updatedDeploymentAppsV1beta1, err := c.AppsV1beta1().Deployments(ns.Name).Update(deploymentAppsV1beta1)
|
|
if err != nil {
|
|
t.Fatalf("failed to update apps/v1beta1 deployment %s: %v", deploymentAppsV1beta1.Name, err)
|
|
}
|
|
if !reflect.DeepEqual(updatedDeploymentAppsV1beta1.Spec.Selector.MatchLabels, newSelectorLabels) {
|
|
t.Errorf("selector should be changed for apps/v1beta1, expected: %v, got: %v", newSelectorLabels, updatedDeploymentAppsV1beta1.Spec.Selector.MatchLabels)
|
|
}
|
|
|
|
// test to ensure apps/v1beta2 selector is immutable
|
|
deploymentAppsV1beta2, err := c.AppsV1beta2().Deployments(ns.Name).Get(updatedDeploymentAppsV1beta1.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed to get apps/v1beta2 deployment %s: %v", updatedDeploymentAppsV1beta1.Name, err)
|
|
}
|
|
newSelectorLabels = map[string]string{"name_apps_v1beta2": "test_apps_v1beta2"}
|
|
deploymentAppsV1beta2.Spec.Selector.MatchLabels = newSelectorLabels
|
|
deploymentAppsV1beta2.Spec.Template.Labels = newSelectorLabels
|
|
_, err = c.AppsV1beta2().Deployments(ns.Name).Update(deploymentAppsV1beta2)
|
|
if err == nil {
|
|
t.Fatalf("failed to provide validation error when changing immutable selector when updating apps/v1beta2 deployment %s", deploymentAppsV1beta2.Name)
|
|
}
|
|
expectedErrType := "Invalid value"
|
|
expectedErrDetail := "field is immutable"
|
|
if !strings.Contains(err.Error(), expectedErrType) || !strings.Contains(err.Error(), expectedErrDetail) {
|
|
t.Errorf("error message does not match, expected type: %s, expected detail: %s, got: %s", expectedErrType, expectedErrDetail, err.Error())
|
|
}
|
|
|
|
// test to ensure apps/v1 selector is immutable
|
|
deploymentAppsV1, err := c.AppsV1().Deployments(ns.Name).Get(updatedDeploymentAppsV1beta1.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed to get apps/v1 deployment %s: %v", updatedDeploymentAppsV1beta1.Name, err)
|
|
}
|
|
newSelectorLabels = map[string]string{"name_apps_v1": "test_apps_v1"}
|
|
deploymentAppsV1.Spec.Selector.MatchLabels = newSelectorLabels
|
|
deploymentAppsV1.Spec.Template.Labels = newSelectorLabels
|
|
_, err = c.AppsV1().Deployments(ns.Name).Update(deploymentAppsV1)
|
|
if err == nil {
|
|
t.Fatalf("failed to provide validation error when changing immutable selector when updating apps/v1 deployment %s", deploymentAppsV1.Name)
|
|
}
|
|
if !strings.Contains(err.Error(), expectedErrType) || !strings.Contains(err.Error(), expectedErrDetail) {
|
|
t.Errorf("error message does not match, expected type: %s, expected detail: %s, got: %s", expectedErrType, expectedErrDetail, err.Error())
|
|
}
|
|
}
|
|
|
|
// Paused deployment should not start new rollout
|
|
func TestPausedDeployment(t *testing.T) {
|
|
s, closeFn, rm, dc, informers, c := dcSetup(t)
|
|
defer closeFn()
|
|
name := "test-paused-deployment"
|
|
ns := framework.CreateTestingNamespace(name, s, t)
|
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
|
|
|
replicas := int32(1)
|
|
tester := &deploymentTester{t: t, c: c, deployment: newDeployment(name, ns.Name, replicas)}
|
|
tester.deployment.Spec.Paused = true
|
|
tgps := int64(1)
|
|
tester.deployment.Spec.Template.Spec.TerminationGracePeriodSeconds = &tgps
|
|
|
|
var err error
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Create(tester.deployment)
|
|
if err != nil {
|
|
t.Fatalf("failed to create deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
|
|
// Start informer and controllers
|
|
stopCh := make(chan struct{})
|
|
defer close(stopCh)
|
|
informers.Start(stopCh)
|
|
go rm.Run(5, stopCh)
|
|
go dc.Run(5, stopCh)
|
|
|
|
// Verify that the paused deployment won't create new replica set.
|
|
if err := tester.expectNoNewReplicaSet(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Resume the deployment
|
|
tester.deployment, err = tester.updateDeployment(resumeFn)
|
|
if err != nil {
|
|
t.Fatalf("failed to resume deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
|
|
// Wait for the controller to notice the resume.
|
|
if err := tester.waitForObservedDeployment(tester.deployment.Generation); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Wait for the Deployment to be updated to revision 1
|
|
if err := tester.waitForDeploymentRevisionAndImage("1", fakeImage); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Make sure the Deployment completes while manually marking Deployment pods as ready at the same time.
|
|
// Use soft check because this deployment was just created and rolling update strategy might be violated.
|
|
if err := tester.waitForDeploymentCompleteAndMarkPodsReady(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// A new replicaset should be created.
|
|
if _, err := tester.expectNewReplicaSet(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Pause the deployment.
|
|
// The paused deployment shouldn't trigger a new rollout.
|
|
tester.deployment, err = tester.updateDeployment(pauseFn)
|
|
if err != nil {
|
|
t.Fatalf("failed to pause deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
|
|
// Wait for the controller to notice the pause.
|
|
if err := tester.waitForObservedDeployment(tester.deployment.Generation); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Update the deployment template
|
|
newTGPS := int64(0)
|
|
tester.deployment, err = tester.updateDeployment(func(update *apps.Deployment) {
|
|
update.Spec.Template.Spec.TerminationGracePeriodSeconds = &newTGPS
|
|
})
|
|
if err != nil {
|
|
t.Fatalf("failed updating template of deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
|
|
// Wait for the controller to notice the rollout.
|
|
if err := tester.waitForObservedDeployment(tester.deployment.Generation); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Verify that the paused deployment won't create new replica set.
|
|
if err := tester.expectNoNewReplicaSet(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
_, allOldRs, err := deploymentutil.GetOldReplicaSets(tester.deployment, c.AppsV1())
|
|
if err != nil {
|
|
t.Fatalf("failed retrieving old replicasets of deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
if len(allOldRs) != 1 {
|
|
t.Errorf("expected an old replica set, got %v", allOldRs)
|
|
}
|
|
if *allOldRs[0].Spec.Template.Spec.TerminationGracePeriodSeconds == newTGPS {
|
|
t.Errorf("TerminationGracePeriodSeconds on the replica set should be %d, got %d", tgps, newTGPS)
|
|
}
|
|
}
|
|
|
|
// Paused deployment can be scaled
|
|
func TestScalePausedDeployment(t *testing.T) {
|
|
s, closeFn, rm, dc, informers, c := dcSetup(t)
|
|
defer closeFn()
|
|
name := "test-scale-paused-deployment"
|
|
ns := framework.CreateTestingNamespace(name, s, t)
|
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
|
|
|
replicas := int32(1)
|
|
tester := &deploymentTester{t: t, c: c, deployment: newDeployment(name, ns.Name, replicas)}
|
|
tgps := int64(1)
|
|
tester.deployment.Spec.Template.Spec.TerminationGracePeriodSeconds = &tgps
|
|
|
|
var err error
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Create(tester.deployment)
|
|
if err != nil {
|
|
t.Fatalf("failed to create deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
|
|
// Start informer and controllers
|
|
stopCh := make(chan struct{})
|
|
defer close(stopCh)
|
|
informers.Start(stopCh)
|
|
go rm.Run(5, stopCh)
|
|
go dc.Run(5, stopCh)
|
|
|
|
// Wait for the Deployment to be updated to revision 1
|
|
if err := tester.waitForDeploymentRevisionAndImage("1", fakeImage); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Make sure the Deployment completes while manually marking Deployment pods as ready at the same time.
|
|
// Use soft check because this deployment was just created and rolling update strategy might be violated.
|
|
if err := tester.waitForDeploymentCompleteAndMarkPodsReady(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// A new replicaset should be created.
|
|
if _, err := tester.expectNewReplicaSet(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Pause the deployment.
|
|
tester.deployment, err = tester.updateDeployment(pauseFn)
|
|
if err != nil {
|
|
t.Fatalf("failed to pause deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
|
|
// Wait for the controller to notice the scale.
|
|
if err := tester.waitForObservedDeployment(tester.deployment.Generation); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Scale the paused deployment.
|
|
newReplicas := int32(10)
|
|
tester.deployment, err = tester.updateDeployment(func(update *apps.Deployment) {
|
|
update.Spec.Replicas = &newReplicas
|
|
})
|
|
if err != nil {
|
|
t.Fatalf("failed updating deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
|
|
// Wait for the controller to notice the scale.
|
|
if err := tester.waitForObservedDeployment(tester.deployment.Generation); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Verify that the new replicaset is scaled.
|
|
rs, err := tester.expectNewReplicaSet()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if *rs.Spec.Replicas != newReplicas {
|
|
t.Errorf("expected new replicaset replicas = %d, got %d", newReplicas, *rs.Spec.Replicas)
|
|
}
|
|
|
|
// Make sure the Deployment completes while manually marking Deployment pods as ready at the same time.
|
|
// Use soft check because this deployment was just scaled and rolling update strategy might be violated.
|
|
if err := tester.waitForDeploymentCompleteAndMarkPodsReady(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
// Deployment rollout shouldn't be blocked on hash collisions
|
|
func TestDeploymentHashCollision(t *testing.T) {
|
|
s, closeFn, rm, dc, informers, c := dcSetup(t)
|
|
defer closeFn()
|
|
name := "test-hash-collision-deployment"
|
|
ns := framework.CreateTestingNamespace(name, s, t)
|
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
|
|
|
replicas := int32(1)
|
|
tester := &deploymentTester{t: t, c: c, deployment: newDeployment(name, ns.Name, replicas)}
|
|
|
|
var err error
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Create(tester.deployment)
|
|
if err != nil {
|
|
t.Fatalf("failed to create deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
|
|
// Start informer and controllers
|
|
stopCh := make(chan struct{})
|
|
defer close(stopCh)
|
|
informers.Start(stopCh)
|
|
go rm.Run(5, stopCh)
|
|
go dc.Run(5, stopCh)
|
|
|
|
// Wait for the Deployment to be updated to revision 1
|
|
if err := tester.waitForDeploymentRevisionAndImage("1", fakeImage); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Mock a hash collision
|
|
newRS, err := deploymentutil.GetNewReplicaSet(tester.deployment, c.AppsV1())
|
|
if err != nil {
|
|
t.Fatalf("failed getting new replicaset of deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
if newRS == nil {
|
|
t.Fatalf("unable to find new replicaset of deployment %s", tester.deployment.Name)
|
|
}
|
|
_, err = tester.updateReplicaSet(newRS.Name, func(update *apps.ReplicaSet) {
|
|
*update.Spec.Template.Spec.TerminationGracePeriodSeconds = int64(5)
|
|
})
|
|
if err != nil {
|
|
t.Fatalf("failed updating replicaset %s template: %v", newRS.Name, err)
|
|
}
|
|
|
|
// Expect deployment collision counter to increment
|
|
if err := wait.PollImmediate(pollInterval, pollTimeout, func() (bool, error) {
|
|
d, err := c.AppsV1().Deployments(ns.Name).Get(tester.deployment.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
return false, nil
|
|
}
|
|
return d.Status.CollisionCount != nil && *d.Status.CollisionCount == int32(1), nil
|
|
}); err != nil {
|
|
t.Fatalf("Failed to increment collision counter for deployment %q: %v", tester.deployment.Name, err)
|
|
}
|
|
|
|
// Expect a new ReplicaSet to be created
|
|
if err := tester.waitForDeploymentRevisionAndImage("2", fakeImage); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
// Deployment supports rollback even when there's old replica set without revision.
|
|
// TODO: drop this test when extensions/v1beta1 is no longer served
|
|
func TestRollbackDeploymentRSNoRevision(t *testing.T) {
|
|
s, closeFn, rm, dc, informers, c := dcSetup(t)
|
|
defer closeFn()
|
|
name := "test-rollback-no-revision-deployment"
|
|
ns := framework.CreateTestingNamespace(name, s, t)
|
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
|
|
|
// Create an old RS without revision
|
|
rsName := "test-rollback-no-revision-controller"
|
|
rsReplicas := int32(1)
|
|
rs := newReplicaSet(rsName, ns.Name, rsReplicas)
|
|
rs.Annotations = make(map[string]string)
|
|
rs.Annotations["make"] = "difference"
|
|
rs.Spec.Template.Spec.Containers[0].Image = "different-image"
|
|
_, err := c.AppsV1().ReplicaSets(ns.Name).Create(rs)
|
|
if err != nil {
|
|
t.Fatalf("failed to create replicaset %s: %v", rsName, err)
|
|
}
|
|
|
|
replicas := int32(1)
|
|
tester := &deploymentTester{t: t, c: c, deployment: newDeployment(name, ns.Name, replicas)}
|
|
oriImage := tester.deployment.Spec.Template.Spec.Containers[0].Image
|
|
// Set absolute rollout limits (defaults changed to percentages)
|
|
max := intstr.FromInt(1)
|
|
tester.deployment.Spec.Strategy.RollingUpdate.MaxUnavailable = &max
|
|
tester.deployment.Spec.Strategy.RollingUpdate.MaxSurge = &max
|
|
|
|
// Create a deployment which have different template than the replica set created above.
|
|
if tester.deployment, err = c.AppsV1().Deployments(ns.Name).Create(tester.deployment); err != nil {
|
|
t.Fatalf("failed to create deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
|
|
// Start informer and controllers
|
|
stopCh := make(chan struct{})
|
|
defer close(stopCh)
|
|
informers.Start(stopCh)
|
|
go rm.Run(5, stopCh)
|
|
go dc.Run(5, stopCh)
|
|
|
|
// Wait for the Deployment to be updated to revision 1
|
|
if err = tester.waitForDeploymentRevisionAndImage("1", fakeImage); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// 1. Rollback to the last revision
|
|
// Since there's only 1 revision in history, it should still be revision 1
|
|
revision := int64(0)
|
|
rollback := newDeploymentRollback(tester.deployment.Name, nil, revision)
|
|
if err = c.ExtensionsV1beta1().Deployments(ns.Name).Rollback(rollback); err != nil {
|
|
t.Fatalf("failed to roll back deployment %s to last revision: %v", tester.deployment.Name, err)
|
|
}
|
|
|
|
// Wait for the deployment to start rolling back
|
|
if err = tester.waitForDeploymentRollbackCleared(); err != nil {
|
|
t.Fatalf("failed to roll back deployment %s to last revision: %v", tester.deployment.Name, err)
|
|
}
|
|
// TODO: report RollbackRevisionNotFound in deployment status and check it here
|
|
|
|
// The pod template shouldn't change since there's no last revision
|
|
// Check if the deployment is still revision 1 and still has the old pod template
|
|
err = tester.checkDeploymentRevisionAndImage("1", oriImage)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// 2. Update the deployment to revision 2.
|
|
updatedImage := "update"
|
|
tester.deployment, err = tester.updateDeployment(func(update *apps.Deployment) {
|
|
update.Spec.Template.Spec.Containers[0].Name = updatedImage
|
|
update.Spec.Template.Spec.Containers[0].Image = updatedImage
|
|
})
|
|
if err != nil {
|
|
t.Fatalf("failed updating deployment %s: %v", tester.deployment.Name, err)
|
|
}
|
|
|
|
// Use observedGeneration to determine if the controller noticed the pod template update.
|
|
// Wait for the controller to notice the resume.
|
|
if err = tester.waitForObservedDeployment(tester.deployment.Generation); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Wait for it to be updated to revision 2
|
|
if err = tester.waitForDeploymentRevisionAndImage("2", updatedImage); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Wait for the Deployment to complete while manually marking Deployment pods as ready at the same time
|
|
if err = tester.waitForDeploymentCompleteAndCheckRollingAndMarkPodsReady(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// 3. Update the deploymentRollback to rollback to revision 1
|
|
revision = int64(1)
|
|
rollback = newDeploymentRollback(tester.deployment.Name, nil, revision)
|
|
if err = c.ExtensionsV1beta1().Deployments(ns.Name).Rollback(rollback); err != nil {
|
|
t.Fatalf("failed to roll back deployment %s to revision %d: %v", tester.deployment.Name, revision, err)
|
|
}
|
|
|
|
// Wait for the deployment to start rolling back
|
|
if err = tester.waitForDeploymentRollbackCleared(); err != nil {
|
|
t.Fatalf("failed to roll back deployment %s to revision %d: %v", tester.deployment.Name, revision, err)
|
|
}
|
|
// TODO: report RollbackDone in deployment status and check it here
|
|
|
|
// The pod template should be updated to the one in revision 1
|
|
// Wait for it to be updated to revision 3
|
|
if err = tester.waitForDeploymentRevisionAndImage("3", oriImage); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Wait for the Deployment to complete while manually marking Deployment pods as ready at the same time
|
|
if err = tester.waitForDeploymentCompleteAndCheckRollingAndMarkPodsReady(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
func checkRSHashLabels(rs *apps.ReplicaSet) (string, error) {
|
|
hash := rs.Labels[apps.DefaultDeploymentUniqueLabelKey]
|
|
selectorHash := rs.Spec.Selector.MatchLabels[apps.DefaultDeploymentUniqueLabelKey]
|
|
templateLabelHash := rs.Spec.Template.Labels[apps.DefaultDeploymentUniqueLabelKey]
|
|
|
|
if hash != selectorHash || selectorHash != templateLabelHash {
|
|
return "", fmt.Errorf("mismatching hash value found in replicaset %s: %#v", rs.Name, rs)
|
|
}
|
|
if len(hash) == 0 {
|
|
return "", fmt.Errorf("unexpected replicaset %s missing required pod-template-hash labels", rs.Name)
|
|
}
|
|
|
|
if !strings.HasSuffix(rs.Name, hash) {
|
|
return "", fmt.Errorf("unexpected replicaset %s name suffix doesn't match hash %s", rs.Name, hash)
|
|
}
|
|
|
|
return hash, nil
|
|
}
|
|
|
|
func checkPodsHashLabel(pods *v1.PodList) (string, error) {
|
|
if len(pods.Items) == 0 {
|
|
return "", fmt.Errorf("no pods given")
|
|
}
|
|
var hash string
|
|
for _, pod := range pods.Items {
|
|
podHash := pod.Labels[apps.DefaultDeploymentUniqueLabelKey]
|
|
if len(podHash) == 0 {
|
|
return "", fmt.Errorf("found pod %s missing pod-template-hash label: %#v", pod.Name, pods)
|
|
}
|
|
// Save the first valid hash
|
|
if len(hash) == 0 {
|
|
hash = podHash
|
|
}
|
|
if podHash != hash {
|
|
return "", fmt.Errorf("found pod %s with mismatching pod-template-hash value %s: %#v", pod.Name, podHash, pods)
|
|
}
|
|
}
|
|
return hash, nil
|
|
}
|
|
|
|
// Deployment should have a timeout condition when it fails to progress after given deadline.
|
|
func TestFailedDeployment(t *testing.T) {
|
|
s, closeFn, rm, dc, informers, c := dcSetup(t)
|
|
defer closeFn()
|
|
name := "test-failed-deployment"
|
|
ns := framework.CreateTestingNamespace(name, s, t)
|
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
|
|
|
deploymentName := "progress-check"
|
|
replicas := int32(1)
|
|
three := int32(3)
|
|
tester := &deploymentTester{t: t, c: c, deployment: newDeployment(deploymentName, ns.Name, replicas)}
|
|
tester.deployment.Spec.ProgressDeadlineSeconds = &three
|
|
var err error
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Create(tester.deployment)
|
|
if err != nil {
|
|
t.Fatalf("failed to create deployment %q: %v", deploymentName, err)
|
|
}
|
|
|
|
// Start informer and controllers
|
|
stopCh := make(chan struct{})
|
|
defer close(stopCh)
|
|
informers.Start(stopCh)
|
|
go rm.Run(5, stopCh)
|
|
go dc.Run(5, stopCh)
|
|
|
|
if err = tester.waitForDeploymentUpdatedReplicasGTE(replicas); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Pods are not marked as Ready, therefore the deployment progress will eventually timeout after progressDeadlineSeconds has passed.
|
|
// Wait for the deployment to have a progress timeout condition.
|
|
if err = tester.waitForDeploymentWithCondition(deploymentutil.TimedOutReason, apps.DeploymentProgressing); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Manually mark pods as Ready and wait for deployment to complete.
|
|
if err := tester.waitForDeploymentCompleteAndMarkPodsReady(); err != nil {
|
|
t.Fatalf("deployment %q fails to have its status becoming valid: %v", deploymentName, err)
|
|
}
|
|
|
|
// Wait for the deployment to have a progress complete condition.
|
|
if err = tester.waitForDeploymentWithCondition(deploymentutil.NewRSAvailableReason, apps.DeploymentProgressing); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
func TestOverlappingDeployments(t *testing.T) {
|
|
s, closeFn, rm, dc, informers, c := dcSetup(t)
|
|
defer closeFn()
|
|
name := "test-overlapping-deployments"
|
|
ns := framework.CreateTestingNamespace(name, s, t)
|
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
|
|
|
replicas := int32(1)
|
|
firstDeploymentName := "first-deployment"
|
|
secondDeploymentName := "second-deployment"
|
|
testers := []*deploymentTester{
|
|
{t: t, c: c, deployment: newDeployment(firstDeploymentName, ns.Name, replicas)},
|
|
{t: t, c: c, deployment: newDeployment(secondDeploymentName, ns.Name, replicas)},
|
|
}
|
|
// Start informer and controllers
|
|
stopCh := make(chan struct{})
|
|
defer close(stopCh)
|
|
informers.Start(stopCh)
|
|
go rm.Run(5, stopCh)
|
|
go dc.Run(5, stopCh)
|
|
|
|
// Create 2 deployments with overlapping selectors
|
|
var err error
|
|
var rss []*apps.ReplicaSet
|
|
for _, tester := range testers {
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Create(tester.deployment)
|
|
dname := tester.deployment.Name
|
|
if err != nil {
|
|
t.Fatalf("failed to create deployment %q: %v", dname, err)
|
|
}
|
|
// Wait for the deployment to be updated to revision 1
|
|
if err = tester.waitForDeploymentRevisionAndImage("1", fakeImage); err != nil {
|
|
t.Fatalf("failed to update deployment %q to revision 1: %v", dname, err)
|
|
}
|
|
// Make sure the deployment completes while manually marking its pods as ready at the same time
|
|
if err = tester.waitForDeploymentCompleteAndMarkPodsReady(); err != nil {
|
|
t.Fatalf("deployment %q failed to complete: %v", dname, err)
|
|
}
|
|
// Get replicaset of the deployment
|
|
newRS, err := tester.getNewReplicaSet()
|
|
if err != nil {
|
|
t.Fatalf("failed to get new replicaset of deployment %q: %v", dname, err)
|
|
}
|
|
if newRS == nil {
|
|
t.Fatalf("unable to find new replicaset of deployment %q", dname)
|
|
}
|
|
// Store the replicaset for future usage
|
|
rss = append(rss, newRS)
|
|
}
|
|
|
|
// Both deployments should proceed independently, so their respective replicaset should not be the same replicaset
|
|
if rss[0].UID == rss[1].UID {
|
|
t.Fatalf("overlapping deployments should not share the same replicaset")
|
|
}
|
|
|
|
// Scale only the first deployment by 1
|
|
newReplicas := replicas + 1
|
|
testers[0].deployment, err = testers[0].updateDeployment(func(update *apps.Deployment) {
|
|
update.Spec.Replicas = &newReplicas
|
|
})
|
|
if err != nil {
|
|
t.Fatalf("failed updating deployment %q: %v", firstDeploymentName, err)
|
|
}
|
|
|
|
// Make sure the deployment completes after scaling
|
|
if err := testers[0].waitForDeploymentCompleteAndMarkPodsReady(); err != nil {
|
|
t.Fatalf("deployment %q failed to complete after scaling: %v", firstDeploymentName, err)
|
|
}
|
|
|
|
// Verify replicaset of both deployments has updated number of replicas
|
|
for i, tester := range testers {
|
|
rs, err := c.AppsV1().ReplicaSets(ns.Name).Get(rss[i].Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed to get replicaset %q: %v", rss[i].Name, err)
|
|
}
|
|
if *rs.Spec.Replicas != *tester.deployment.Spec.Replicas {
|
|
t.Errorf("expected replicaset %q of deployment %q has %d replicas, but found %d replicas", rs.Name, firstDeploymentName, *tester.deployment.Spec.Replicas, *rs.Spec.Replicas)
|
|
}
|
|
}
|
|
}
|
|
|
|
// Deployment should not block rollout when updating spec replica number and template at the same time.
|
|
func TestScaledRolloutDeployment(t *testing.T) {
|
|
s, closeFn, rm, dc, informers, c := dcSetup(t)
|
|
defer closeFn()
|
|
name := "test-scaled-rollout-deployment"
|
|
ns := framework.CreateTestingNamespace(name, s, t)
|
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
|
|
|
stopCh := make(chan struct{})
|
|
defer close(stopCh)
|
|
informers.Start(stopCh)
|
|
go rm.Run(5, stopCh)
|
|
go dc.Run(5, stopCh)
|
|
|
|
// Create a deployment with rolling update strategy, max surge = 3, and max unavailable = 2
|
|
var err error
|
|
replicas := int32(10)
|
|
tester := &deploymentTester{t: t, c: c, deployment: newDeployment(name, ns.Name, replicas)}
|
|
tester.deployment.Spec.Strategy.RollingUpdate.MaxSurge = intOrStrP(3)
|
|
tester.deployment.Spec.Strategy.RollingUpdate.MaxUnavailable = intOrStrP(2)
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Create(tester.deployment)
|
|
if err != nil {
|
|
t.Fatalf("failed to create deployment %q: %v", name, err)
|
|
}
|
|
if err = tester.waitForDeploymentRevisionAndImage("1", fakeImage); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err = tester.waitForDeploymentCompleteAndMarkPodsReady(); err != nil {
|
|
t.Fatalf("deployment %q failed to complete: %v", name, err)
|
|
}
|
|
|
|
// Record current replicaset before starting new rollout
|
|
firstRS, err := tester.expectNewReplicaSet()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Update the deployment with another new image but do not mark the pods as ready to block new replicaset
|
|
fakeImage2 := "fakeimage2"
|
|
tester.deployment, err = tester.updateDeployment(func(update *apps.Deployment) {
|
|
update.Spec.Template.Spec.Containers[0].Image = fakeImage2
|
|
})
|
|
if err != nil {
|
|
t.Fatalf("failed updating deployment %q: %v", name, err)
|
|
}
|
|
if err = tester.waitForDeploymentRevisionAndImage("2", fakeImage2); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Verify the deployment has minimum available replicas after 2nd rollout
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Get(name, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed to get deployment %q: %v", name, err)
|
|
}
|
|
minAvailableReplicas := deploymentutil.MinAvailable(tester.deployment)
|
|
if tester.deployment.Status.AvailableReplicas < minAvailableReplicas {
|
|
t.Fatalf("deployment %q does not have minimum number of available replicas after 2nd rollout", name)
|
|
}
|
|
|
|
// Wait for old replicaset of 1st rollout to have desired replicas
|
|
firstRS, err = c.AppsV1().ReplicaSets(ns.Name).Get(firstRS.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed to get replicaset %q: %v", firstRS.Name, err)
|
|
}
|
|
if err = tester.waitRSStable(firstRS); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Wait for new replicaset of 2nd rollout to have desired replicas
|
|
secondRS, err := tester.expectNewReplicaSet()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err = tester.waitRSStable(secondRS); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Scale up the deployment and update its image to another new image simultaneously (this time marks all pods as ready)
|
|
newReplicas := int32(20)
|
|
fakeImage3 := "fakeimage3"
|
|
tester.deployment, err = tester.updateDeployment(func(update *apps.Deployment) {
|
|
update.Spec.Replicas = &newReplicas
|
|
update.Spec.Template.Spec.Containers[0].Image = fakeImage3
|
|
})
|
|
if err != nil {
|
|
t.Fatalf("failed updating deployment %q: %v", name, err)
|
|
}
|
|
if err = tester.waitForDeploymentRevisionAndImage("3", fakeImage3); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err = tester.waitForDeploymentCompleteAndMarkPodsReady(); err != nil {
|
|
t.Fatalf("deployment %q failed to complete: %v", name, err)
|
|
}
|
|
|
|
// Verify every replicaset has correct desiredReplicas annotation after 3rd rollout
|
|
thirdRS, err := deploymentutil.GetNewReplicaSet(tester.deployment, c.AppsV1())
|
|
if err != nil {
|
|
t.Fatalf("failed getting new revision 3 replicaset for deployment %q: %v", name, err)
|
|
}
|
|
rss := []*apps.ReplicaSet{firstRS, secondRS, thirdRS}
|
|
for _, curRS := range rss {
|
|
curRS, err = c.AppsV1().ReplicaSets(ns.Name).Get(curRS.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed to get replicaset when checking desired replicas annotation: %v", err)
|
|
}
|
|
desired, ok := deploymentutil.GetDesiredReplicasAnnotation(curRS)
|
|
if !ok {
|
|
t.Fatalf("failed to retrieve desiredReplicas annotation for replicaset %q", curRS.Name)
|
|
}
|
|
if desired != *(tester.deployment.Spec.Replicas) {
|
|
t.Fatalf("unexpected desiredReplicas annotation for replicaset %q: expected %d, got %d", curRS.Name, *(tester.deployment.Spec.Replicas), desired)
|
|
}
|
|
}
|
|
|
|
// Update the deployment with another new image but do not mark the pods as ready to block new replicaset
|
|
fakeImage4 := "fakeimage4"
|
|
tester.deployment, err = tester.updateDeployment(func(update *apps.Deployment) {
|
|
update.Spec.Template.Spec.Containers[0].Image = fakeImage4
|
|
})
|
|
if err != nil {
|
|
t.Fatalf("failed updating deployment %q: %v", name, err)
|
|
}
|
|
if err = tester.waitForDeploymentRevisionAndImage("4", fakeImage4); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Verify the deployment has minimum available replicas after 4th rollout
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Get(name, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed to get deployment %q: %v", name, err)
|
|
}
|
|
minAvailableReplicas = deploymentutil.MinAvailable(tester.deployment)
|
|
if tester.deployment.Status.AvailableReplicas < minAvailableReplicas {
|
|
t.Fatalf("deployment %q does not have minimum number of available replicas after 4th rollout", name)
|
|
}
|
|
|
|
// Wait for old replicaset of 3rd rollout to have desired replicas
|
|
thirdRS, err = c.AppsV1().ReplicaSets(ns.Name).Get(thirdRS.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed to get replicaset %q: %v", thirdRS.Name, err)
|
|
}
|
|
if err = tester.waitRSStable(thirdRS); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Wait for new replicaset of 4th rollout to have desired replicas
|
|
fourthRS, err := tester.expectNewReplicaSet()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err = tester.waitRSStable(fourthRS); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Scale down the deployment and update its image to another new image simultaneously (this time marks all pods as ready)
|
|
newReplicas = int32(5)
|
|
fakeImage5 := "fakeimage5"
|
|
tester.deployment, err = tester.updateDeployment(func(update *apps.Deployment) {
|
|
update.Spec.Replicas = &newReplicas
|
|
update.Spec.Template.Spec.Containers[0].Image = fakeImage5
|
|
})
|
|
if err != nil {
|
|
t.Fatalf("failed updating deployment %q: %v", name, err)
|
|
}
|
|
if err = tester.waitForDeploymentRevisionAndImage("5", fakeImage5); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err = tester.waitForDeploymentCompleteAndMarkPodsReady(); err != nil {
|
|
t.Fatalf("deployment %q failed to complete: %v", name, err)
|
|
}
|
|
|
|
// Verify every replicaset has correct desiredReplicas annotation after 5th rollout
|
|
fifthRS, err := deploymentutil.GetNewReplicaSet(tester.deployment, c.AppsV1())
|
|
if err != nil {
|
|
t.Fatalf("failed getting new revision 5 replicaset for deployment %q: %v", name, err)
|
|
}
|
|
rss = []*apps.ReplicaSet{thirdRS, fourthRS, fifthRS}
|
|
for _, curRS := range rss {
|
|
curRS, err = c.AppsV1().ReplicaSets(ns.Name).Get(curRS.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed to get replicaset when checking desired replicas annotation: %v", err)
|
|
}
|
|
desired, ok := deploymentutil.GetDesiredReplicasAnnotation(curRS)
|
|
if !ok {
|
|
t.Fatalf("failed to retrieve desiredReplicas annotation for replicaset %q", curRS.Name)
|
|
}
|
|
if desired != *(tester.deployment.Spec.Replicas) {
|
|
t.Fatalf("unexpected desiredReplicas annotation for replicaset %q: expected %d, got %d", curRS.Name, *(tester.deployment.Spec.Replicas), desired)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestSpecReplicasChange(t *testing.T) {
|
|
s, closeFn, rm, dc, informers, c := dcSetup(t)
|
|
defer closeFn()
|
|
name := "test-spec-replicas-change"
|
|
ns := framework.CreateTestingNamespace(name, s, t)
|
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
|
|
|
deploymentName := "deployment"
|
|
replicas := int32(1)
|
|
tester := &deploymentTester{t: t, c: c, deployment: newDeployment(deploymentName, ns.Name, replicas)}
|
|
tester.deployment.Spec.Strategy.Type = apps.RecreateDeploymentStrategyType
|
|
tester.deployment.Spec.Strategy.RollingUpdate = nil
|
|
var err error
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Create(tester.deployment)
|
|
if err != nil {
|
|
t.Fatalf("failed to create deployment %q: %v", deploymentName, err)
|
|
}
|
|
|
|
// Start informer and controllers
|
|
stopCh := make(chan struct{})
|
|
defer close(stopCh)
|
|
informers.Start(stopCh)
|
|
go rm.Run(5, stopCh)
|
|
go dc.Run(5, stopCh)
|
|
|
|
// Scale up/down deployment and verify its replicaset has matching .spec.replicas
|
|
if err = tester.scaleDeployment(2); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err = tester.scaleDeployment(0); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err = tester.scaleDeployment(1); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Add a template annotation change to test deployment's status does update
|
|
// without .spec.replicas change
|
|
var oldGeneration int64
|
|
tester.deployment, err = tester.updateDeployment(func(update *apps.Deployment) {
|
|
oldGeneration = update.Generation
|
|
update.Spec.RevisionHistoryLimit = pointer.Int32Ptr(4)
|
|
})
|
|
if err != nil {
|
|
t.Fatalf("failed updating deployment %q: %v", tester.deployment.Name, err)
|
|
}
|
|
|
|
savedGeneration := tester.deployment.Generation
|
|
if savedGeneration == oldGeneration {
|
|
t.Fatalf("Failed to verify .Generation has incremented for deployment %q", deploymentName)
|
|
}
|
|
if err = tester.waitForObservedDeployment(savedGeneration); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
func TestDeploymentAvailableCondition(t *testing.T) {
|
|
s, closeFn, rm, dc, informers, c := dcSetup(t)
|
|
defer closeFn()
|
|
name := "test-deployment-available-condition"
|
|
ns := framework.CreateTestingNamespace(name, s, t)
|
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
|
|
|
deploymentName := "deployment"
|
|
replicas := int32(10)
|
|
tester := &deploymentTester{t: t, c: c, deployment: newDeployment(deploymentName, ns.Name, replicas)}
|
|
// Assign a high value to the deployment's minReadySeconds
|
|
tester.deployment.Spec.MinReadySeconds = 3600
|
|
// progressDeadlineSeconds must be greater than minReadySeconds
|
|
tester.deployment.Spec.ProgressDeadlineSeconds = pointer.Int32Ptr(7200)
|
|
var err error
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Create(tester.deployment)
|
|
if err != nil {
|
|
t.Fatalf("failed to create deployment %q: %v", deploymentName, err)
|
|
}
|
|
|
|
// Start informer and controllers
|
|
stopCh := make(chan struct{})
|
|
defer close(stopCh)
|
|
informers.Start(stopCh)
|
|
go rm.Run(5, stopCh)
|
|
go dc.Run(5, stopCh)
|
|
|
|
// Wait for the deployment to be observed by the controller and has at least specified number of updated replicas
|
|
if err = tester.waitForDeploymentUpdatedReplicasGTE(replicas); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Wait for the deployment to have MinimumReplicasUnavailable reason because the pods are not marked as ready
|
|
if err = tester.waitForDeploymentWithCondition(deploymentutil.MinimumReplicasUnavailable, apps.DeploymentAvailable); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Verify all replicas fields of DeploymentStatus have desired counts
|
|
if err = tester.checkDeploymentStatusReplicasFields(10, 10, 0, 0, 10); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Mark the pods as ready without waiting for the deployment to complete
|
|
if err = tester.markUpdatedPodsReadyWithoutComplete(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Wait for number of ready replicas to equal number of replicas.
|
|
if err = tester.waitForReadyReplicas(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Wait for the deployment to still have MinimumReplicasUnavailable reason within minReadySeconds period
|
|
if err = tester.waitForDeploymentWithCondition(deploymentutil.MinimumReplicasUnavailable, apps.DeploymentAvailable); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Verify all replicas fields of DeploymentStatus have desired counts
|
|
if err = tester.checkDeploymentStatusReplicasFields(10, 10, 10, 0, 10); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Update the deployment's minReadySeconds to a small value
|
|
tester.deployment, err = tester.updateDeployment(func(update *apps.Deployment) {
|
|
update.Spec.MinReadySeconds = 1
|
|
})
|
|
if err != nil {
|
|
t.Fatalf("failed updating deployment %q: %v", deploymentName, err)
|
|
}
|
|
|
|
// Wait for the deployment to notice minReadySeconds has changed
|
|
if err := tester.waitForObservedDeployment(tester.deployment.Generation); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Wait for the deployment to have MinimumReplicasAvailable reason after minReadySeconds period
|
|
if err = tester.waitForDeploymentWithCondition(deploymentutil.MinimumReplicasAvailable, apps.DeploymentAvailable); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Verify all replicas fields of DeploymentStatus have desired counts
|
|
if err = tester.checkDeploymentStatusReplicasFields(10, 10, 10, 10, 0); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
// Wait for deployment to automatically patch incorrect ControllerRef of RS
|
|
func testRSControllerRefPatch(t *testing.T, tester *deploymentTester, rs *apps.ReplicaSet, ownerReference *metav1.OwnerReference, expectedOwnerReferenceNum int) {
|
|
ns := rs.Namespace
|
|
rsClient := tester.c.AppsV1().ReplicaSets(ns)
|
|
rs, err := tester.updateReplicaSet(rs.Name, func(update *apps.ReplicaSet) {
|
|
update.OwnerReferences = []metav1.OwnerReference{*ownerReference}
|
|
})
|
|
if err != nil {
|
|
t.Fatalf("failed to update replicaset %q: %v", rs.Name, err)
|
|
}
|
|
|
|
if err := wait.PollImmediate(pollInterval, pollTimeout, func() (bool, error) {
|
|
newRS, err := rsClient.Get(rs.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
return metav1.GetControllerOf(newRS) != nil, nil
|
|
}); err != nil {
|
|
t.Fatalf("failed to wait for controllerRef of the replicaset %q to become nil: %v", rs.Name, err)
|
|
}
|
|
|
|
newRS, err := rsClient.Get(rs.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("failed to obtain replicaset %q: %v", rs.Name, err)
|
|
}
|
|
controllerRef := metav1.GetControllerOf(newRS)
|
|
if controllerRef.UID != tester.deployment.UID {
|
|
t.Fatalf("controllerRef of replicaset %q has a different UID: Expected %v, got %v", newRS.Name, tester.deployment.UID, controllerRef.UID)
|
|
}
|
|
ownerReferenceNum := len(newRS.GetOwnerReferences())
|
|
if ownerReferenceNum != expectedOwnerReferenceNum {
|
|
t.Fatalf("unexpected number of owner references for replicaset %q: Expected %d, got %d", newRS.Name, expectedOwnerReferenceNum, ownerReferenceNum)
|
|
}
|
|
}
|
|
|
|
func TestGeneralReplicaSetAdoption(t *testing.T) {
|
|
s, closeFn, rm, dc, informers, c := dcSetup(t)
|
|
defer closeFn()
|
|
name := "test-general-replicaset-adoption"
|
|
ns := framework.CreateTestingNamespace(name, s, t)
|
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
|
|
|
deploymentName := "deployment"
|
|
replicas := int32(1)
|
|
tester := &deploymentTester{t: t, c: c, deployment: newDeployment(deploymentName, ns.Name, replicas)}
|
|
var err error
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Create(tester.deployment)
|
|
if err != nil {
|
|
t.Fatalf("failed to create deployment %q: %v", deploymentName, err)
|
|
}
|
|
|
|
// Start informer and controllers
|
|
stopCh := make(chan struct{})
|
|
defer close(stopCh)
|
|
informers.Start(stopCh)
|
|
go rm.Run(5, stopCh)
|
|
go dc.Run(5, stopCh)
|
|
|
|
// Wait for the Deployment to be updated to revision 1
|
|
if err := tester.waitForDeploymentRevisionAndImage("1", fakeImage); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Ensure the deployment completes while marking its pods as ready simultaneously
|
|
if err := tester.waitForDeploymentCompleteAndMarkPodsReady(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Get replicaset of the deployment
|
|
rs, err := deploymentutil.GetNewReplicaSet(tester.deployment, c.AppsV1())
|
|
if err != nil {
|
|
t.Fatalf("failed to get replicaset of deployment %q: %v", deploymentName, err)
|
|
}
|
|
if rs == nil {
|
|
t.Fatalf("unable to find replicaset of deployment %q", deploymentName)
|
|
}
|
|
|
|
// When the only OwnerReference of the RS points to another type of API object such as statefulset
|
|
// with Controller=false, the deployment should add a second OwnerReference (ControllerRef) pointing to itself
|
|
// with Controller=true
|
|
var falseVar = false
|
|
ownerReference := metav1.OwnerReference{UID: uuid.NewUUID(), APIVersion: "apps/v1", Kind: "StatefulSet", Name: deploymentName, Controller: &falseVar}
|
|
testRSControllerRefPatch(t, tester, rs, &ownerReference, 2)
|
|
|
|
// When the only OwnerReference of the RS points to the deployment with Controller=false,
|
|
// the deployment should set Controller=true for the only OwnerReference
|
|
ownerReference = metav1.OwnerReference{UID: tester.deployment.UID, APIVersion: "apps/v1", Kind: "Deployment", Name: deploymentName, Controller: &falseVar}
|
|
testRSControllerRefPatch(t, tester, rs, &ownerReference, 1)
|
|
}
|
|
|
|
func testScalingUsingScaleSubresource(t *testing.T, tester *deploymentTester, replicas int32) {
|
|
ns := tester.deployment.Namespace
|
|
deploymentName := tester.deployment.Name
|
|
deploymentClient := tester.c.AppsV1().Deployments(ns)
|
|
deployment, err := deploymentClient.Get(deploymentName, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("Failed to obtain deployment %q: %v", deploymentName, err)
|
|
}
|
|
scale, err := tester.c.AppsV1().Deployments(ns).GetScale(deploymentName, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("Failed to obtain scale subresource for deployment %q: %v", deploymentName, err)
|
|
}
|
|
if scale.Spec.Replicas != *deployment.Spec.Replicas {
|
|
t.Fatalf("Scale subresource for deployment %q does not match .Spec.Replicas: expected %d, got %d", deploymentName, *deployment.Spec.Replicas, scale.Spec.Replicas)
|
|
}
|
|
|
|
if err := retry.RetryOnConflict(retry.DefaultBackoff, func() error {
|
|
scale, err := tester.c.AppsV1().Deployments(ns).GetScale(deploymentName, metav1.GetOptions{})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
scale.Spec.Replicas = replicas
|
|
_, err = tester.c.AppsV1().Deployments(ns).UpdateScale(deploymentName, scale)
|
|
return err
|
|
}); err != nil {
|
|
t.Fatalf("Failed to set .Spec.Replicas of scale subresource for deployment %q: %v", deploymentName, err)
|
|
}
|
|
|
|
deployment, err = deploymentClient.Get(deploymentName, metav1.GetOptions{})
|
|
if err != nil {
|
|
t.Fatalf("Failed to obtain deployment %q: %v", deploymentName, err)
|
|
}
|
|
if *deployment.Spec.Replicas != replicas {
|
|
t.Fatalf(".Spec.Replicas of deployment %q does not match its scale subresource: expected %d, got %d", deploymentName, replicas, *deployment.Spec.Replicas)
|
|
}
|
|
}
|
|
|
|
func TestDeploymentScaleSubresource(t *testing.T) {
|
|
s, closeFn, rm, dc, informers, c := dcSetup(t)
|
|
defer closeFn()
|
|
name := "test-deployment-scale-subresource"
|
|
ns := framework.CreateTestingNamespace(name, s, t)
|
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
|
|
|
deploymentName := "deployment"
|
|
replicas := int32(2)
|
|
tester := &deploymentTester{t: t, c: c, deployment: newDeployment(deploymentName, ns.Name, replicas)}
|
|
var err error
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Create(tester.deployment)
|
|
if err != nil {
|
|
t.Fatalf("failed to create deployment %q: %v", deploymentName, err)
|
|
}
|
|
|
|
// Start informer and controllers
|
|
stopCh := make(chan struct{})
|
|
defer close(stopCh)
|
|
informers.Start(stopCh)
|
|
go rm.Run(5, stopCh)
|
|
go dc.Run(5, stopCh)
|
|
|
|
// Wait for the Deployment to be updated to revision 1
|
|
if err := tester.waitForDeploymentRevisionAndImage("1", fakeImage); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Ensure the deployment completes while marking its pods as ready simultaneously
|
|
if err := tester.waitForDeploymentCompleteAndMarkPodsReady(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Use scale subresource to scale the deployment up to 3
|
|
testScalingUsingScaleSubresource(t, tester, 3)
|
|
// Use the scale subresource to scale the deployment down to 0
|
|
testScalingUsingScaleSubresource(t, tester, 0)
|
|
}
|
|
|
|
// This test verifies that the Deployment does orphan a ReplicaSet when the ReplicaSet's
|
|
// .Labels field is changed to no longer match the Deployment's selector. It also partially
|
|
// verifies that collision avoidance mechanism is triggered when a Deployment's new ReplicaSet
|
|
// is orphaned, even without PodTemplateSpec change. Refer comment below for more info:
|
|
// https://github.com/kubernetes/kubernetes/pull/59212#discussion_r166465113
|
|
func TestReplicaSetOrphaningAndAdoptionWhenLabelsChange(t *testing.T) {
|
|
s, closeFn, rm, dc, informers, c := dcSetup(t)
|
|
defer closeFn()
|
|
name := "test-replicaset-orphaning-and-adoption-when-labels-change"
|
|
ns := framework.CreateTestingNamespace(name, s, t)
|
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
|
|
|
deploymentName := "deployment"
|
|
replicas := int32(1)
|
|
tester := &deploymentTester{t: t, c: c, deployment: newDeployment(deploymentName, ns.Name, replicas)}
|
|
var err error
|
|
tester.deployment, err = c.AppsV1().Deployments(ns.Name).Create(tester.deployment)
|
|
if err != nil {
|
|
t.Fatalf("failed to create deployment %q: %v", deploymentName, err)
|
|
}
|
|
|
|
// Start informer and controllers
|
|
stopCh := make(chan struct{})
|
|
defer close(stopCh)
|
|
informers.Start(stopCh)
|
|
go rm.Run(5, stopCh)
|
|
go dc.Run(5, stopCh)
|
|
|
|
// Wait for the Deployment to be updated to revision 1
|
|
if err := tester.waitForDeploymentRevisionAndImage("1", fakeImage); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Ensure the deployment completes while marking its pods as ready simultaneously
|
|
if err := tester.waitForDeploymentCompleteAndMarkPodsReady(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Orphaning: deployment should remove OwnerReference from a RS when the RS's labels change to not match its labels
|
|
|
|
// Get replicaset of the deployment
|
|
rs, err := deploymentutil.GetNewReplicaSet(tester.deployment, c.AppsV1())
|
|
if err != nil {
|
|
t.Fatalf("failed to get replicaset of deployment %q: %v", deploymentName, err)
|
|
}
|
|
if rs == nil {
|
|
t.Fatalf("unable to find replicaset of deployment %q", deploymentName)
|
|
}
|
|
|
|
// Verify controllerRef of the replicaset is not nil and pointing to the deployment
|
|
controllerRef := metav1.GetControllerOf(rs)
|
|
if controllerRef == nil {
|
|
t.Fatalf("controllerRef of replicaset %q is nil", rs.Name)
|
|
}
|
|
if controllerRef.UID != tester.deployment.UID {
|
|
t.Fatalf("controllerRef of replicaset %q has a different UID: Expected %v, got %v", rs.Name, tester.deployment.UID, controllerRef.UID)
|
|
}
|
|
|
|
// Change the replicaset's labels to not match the deployment's labels
|
|
labelMap := map[string]string{"new-name": "new-test"}
|
|
rs, err = tester.updateReplicaSet(rs.Name, func(update *apps.ReplicaSet) {
|
|
update.Labels = labelMap
|
|
})
|
|
if err != nil {
|
|
t.Fatalf("failed to update replicaset %q: %v", rs.Name, err)
|
|
}
|
|
|
|
// Wait for the controllerRef of the replicaset to become nil
|
|
rsClient := tester.c.AppsV1().ReplicaSets(ns.Name)
|
|
if err = wait.PollImmediate(pollInterval, pollTimeout, func() (bool, error) {
|
|
rs, err = rsClient.Get(rs.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
return metav1.GetControllerOf(rs) == nil, nil
|
|
}); err != nil {
|
|
t.Fatalf("failed to wait for controllerRef of replicaset %q to become nil: %v", rs.Name, err)
|
|
}
|
|
|
|
// Wait for the deployment to create a new replicaset
|
|
// This will trigger collision avoidance due to deterministic nature of replicaset name
|
|
// i.e., the new replicaset will have a name with different hash to preserve name uniqueness
|
|
var newRS *apps.ReplicaSet
|
|
if err = wait.PollImmediate(pollInterval, pollTimeout, func() (bool, error) {
|
|
newRS, err = deploymentutil.GetNewReplicaSet(tester.deployment, c.AppsV1())
|
|
if err != nil {
|
|
return false, fmt.Errorf("failed to get new replicaset of deployment %q after orphaning: %v", deploymentName, err)
|
|
}
|
|
return newRS != nil, nil
|
|
}); err != nil {
|
|
t.Fatalf("failed to wait for deployment %q to create a new replicaset after orphaning: %v", deploymentName, err)
|
|
}
|
|
if newRS.UID == rs.UID {
|
|
t.Fatalf("expect deployment %q to create a new replicaset different from the orphaned one, but it isn't", deploymentName)
|
|
}
|
|
|
|
// Adoption: deployment should add controllerRef to a RS when the RS's labels change to match its labels
|
|
|
|
// Change the old replicaset's labels to match the deployment's labels
|
|
rs, err = tester.updateReplicaSet(rs.Name, func(update *apps.ReplicaSet) {
|
|
update.Labels = testLabels()
|
|
})
|
|
if err != nil {
|
|
t.Fatalf("failed to update replicaset %q: %v", rs.Name, err)
|
|
}
|
|
|
|
// Wait for the deployment to adopt the old replicaset
|
|
if err = wait.PollImmediate(pollInterval, pollTimeout, func() (bool, error) {
|
|
rs, err := rsClient.Get(rs.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
controllerRef = metav1.GetControllerOf(rs)
|
|
return controllerRef != nil && controllerRef.UID == tester.deployment.UID, nil
|
|
}); err != nil {
|
|
t.Fatalf("failed waiting for replicaset adoption by deployment %q to complete: %v", deploymentName, err)
|
|
}
|
|
}
|