mirror of https://github.com/k3s-io/k3s
Address reviewer comments.
parent
29dd7e3bb8
commit
723cb73057
|
@ -172,6 +172,7 @@ func NewDaemonSetsController(kubeClient client.Interface) *DaemonSetsController
|
|||
|
||||
// Run begins watching and syncing daemon sets.
|
||||
func (dsc *DaemonSetsController) Run(workers int, stopCh <-chan struct{}) {
|
||||
defer util.HandleCrash()
|
||||
go dsc.dsController.Run(stopCh)
|
||||
go dsc.podController.Run(stopCh)
|
||||
go dsc.nodeController.Run(stopCh)
|
||||
|
@ -217,6 +218,8 @@ func (dsc *DaemonSetsController) enqueueDaemonSet(obj interface{}) {
|
|||
glog.Errorf("Couldn't get key for object %+v: %v", obj, err)
|
||||
return
|
||||
}
|
||||
|
||||
// TODO: Handle overlapping controllers better. See comment in ReplicationManager.
|
||||
dsc.queue.Add(key)
|
||||
}
|
||||
|
||||
|
@ -483,7 +486,7 @@ func (dsc *DaemonSetsController) syncDaemonSet(key string) error {
|
|||
// Sleep so we give the pod reflector goroutine a chance to run.
|
||||
time.Sleep(PodStoreSyncedPollPeriod)
|
||||
glog.Infof("Waiting for pods controller to sync, requeuing ds %v", ds.Name)
|
||||
dsc.enqueueDaemonSet(&ds)
|
||||
dsc.enqueueDaemonSet(ds)
|
||||
return nil
|
||||
}
|
||||
|
||||
|
|
|
@ -35,12 +35,22 @@ var (
|
|||
simpleDaemonSetLabel2 = map[string]string{"name": "simple-daemon", "type": "test"}
|
||||
simpleNodeLabel = map[string]string{"color": "blue", "speed": "fast"}
|
||||
simpleNodeLabel2 = map[string]string{"color": "red", "speed": "fast"}
|
||||
alwaysReady = func() bool { return true }
|
||||
)
|
||||
|
||||
func init() {
|
||||
api.ForTesting_ReferencesAllowBlankSelfLinks = true
|
||||
}
|
||||
|
||||
func getKey(ds *experimental.DaemonSet, t *testing.T) string {
|
||||
if key, err := controller.KeyFunc(ds); err != nil {
|
||||
t.Errorf("Unexpected error getting key for ds %v: %v", ds.Name, err)
|
||||
return ""
|
||||
} else {
|
||||
return key
|
||||
}
|
||||
}
|
||||
|
||||
func newDaemonSet(name string) *experimental.DaemonSet {
|
||||
return &experimental.DaemonSet{
|
||||
TypeMeta: unversioned.TypeMeta{APIVersion: testapi.Experimental.Version()},
|
||||
|
@ -121,6 +131,7 @@ func addPods(podStore cache.Store, nodeName string, label map[string]string, num
|
|||
func newTestController() (*DaemonSetsController, *controller.FakePodControl) {
|
||||
client := client.NewOrDie(&client.Config{Host: "", Version: testapi.Default.GroupAndVersion()})
|
||||
manager := NewDaemonSetsController(client)
|
||||
manager.podStoreSynced = alwaysReady
|
||||
podControl := &controller.FakePodControl{}
|
||||
manager.podControl = podControl
|
||||
return manager, podControl
|
||||
|
@ -282,3 +293,25 @@ func TestInconsistentNameSelectorDaemonSetDoesNothing(t *testing.T) {
|
|||
manager.dsStore.Add(ds)
|
||||
syncAndValidateDaemonSets(t, manager, ds, podControl, 0, 0)
|
||||
}
|
||||
|
||||
func TestDSManagerNotReady(t *testing.T) {
|
||||
manager, podControl := newTestController()
|
||||
manager.podStoreSynced = func() bool { return false }
|
||||
addNodes(manager.nodeStore.Store, 0, 1, nil)
|
||||
|
||||
// Simulates the ds reflector running before the pod reflector. We don't
|
||||
// want to end up creating daemon pods in this case until the pod reflector
|
||||
// has synced, so the ds manager should just requeue the ds.
|
||||
ds := newDaemonSet("foo")
|
||||
manager.dsStore.Add(ds)
|
||||
|
||||
dsKey := getKey(ds, t)
|
||||
syncAndValidateDaemonSets(t, manager, ds, podControl, 0, 0)
|
||||
queueDS, _ := manager.queue.Get()
|
||||
if queueDS != dsKey {
|
||||
t.Fatalf("Expected to find key %v in queue, found %v", dsKey, queueDS)
|
||||
}
|
||||
|
||||
manager.podStoreSynced = alwaysReady
|
||||
syncAndValidateDaemonSets(t, manager, ds, podControl, 1, 0)
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue