2015-08-17 12:18:26 +00:00
|
|
|
/*
|
2016-06-03 00:25:58 +00:00
|
|
|
Copyright 2015 The Kubernetes Authors.
|
2015-08-17 12:18:26 +00:00
|
|
|
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
you may not use this file except in compliance with the License.
|
|
|
|
You may obtain a copy of the License at
|
|
|
|
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
See the License for the specific language governing permissions and
|
|
|
|
limitations under the License.
|
|
|
|
*/
|
|
|
|
|
2015-09-10 13:10:07 +00:00
|
|
|
package podautoscaler
|
2015-08-17 12:18:26 +00:00
|
|
|
|
|
|
|
import (
|
|
|
|
"fmt"
|
2015-09-07 10:25:04 +00:00
|
|
|
"math"
|
2015-08-17 12:18:26 +00:00
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/golang/glog"
|
2017-03-08 07:02:34 +00:00
|
|
|
"k8s.io/apimachinery/pkg/api/errors"
|
2017-01-25 13:13:07 +00:00
|
|
|
"k8s.io/apimachinery/pkg/api/resource"
|
2017-01-11 14:09:48 +00:00
|
|
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
2016-12-02 20:18:26 +00:00
|
|
|
"k8s.io/apimachinery/pkg/labels"
|
|
|
|
"k8s.io/apimachinery/pkg/runtime"
|
|
|
|
"k8s.io/apimachinery/pkg/runtime/schema"
|
2017-01-11 14:09:48 +00:00
|
|
|
utilruntime "k8s.io/apimachinery/pkg/util/runtime"
|
2017-03-08 07:02:34 +00:00
|
|
|
"k8s.io/apimachinery/pkg/util/wait"
|
2017-01-30 18:39:54 +00:00
|
|
|
v1core "k8s.io/client-go/kubernetes/typed/core/v1"
|
|
|
|
clientv1 "k8s.io/client-go/pkg/api/v1"
|
2017-01-24 14:11:51 +00:00
|
|
|
"k8s.io/client-go/tools/cache"
|
2017-01-30 18:39:54 +00:00
|
|
|
"k8s.io/client-go/tools/record"
|
2017-03-08 07:02:34 +00:00
|
|
|
"k8s.io/client-go/util/workqueue"
|
2017-01-30 18:39:54 +00:00
|
|
|
"k8s.io/kubernetes/pkg/api"
|
2016-11-18 20:50:17 +00:00
|
|
|
"k8s.io/kubernetes/pkg/api/v1"
|
2016-12-02 20:18:26 +00:00
|
|
|
autoscalingv1 "k8s.io/kubernetes/pkg/apis/autoscaling/v1"
|
|
|
|
autoscalingv2 "k8s.io/kubernetes/pkg/apis/autoscaling/v2alpha1"
|
|
|
|
extensions "k8s.io/kubernetes/pkg/apis/extensions/v1beta1"
|
|
|
|
autoscalingclient "k8s.io/kubernetes/pkg/client/clientset_generated/clientset/typed/autoscaling/v1"
|
|
|
|
extensionsclient "k8s.io/kubernetes/pkg/client/clientset_generated/clientset/typed/extensions/v1beta1"
|
2017-02-08 21:18:21 +00:00
|
|
|
autoscalinginformers "k8s.io/kubernetes/pkg/client/informers/informers_generated/externalversions/autoscaling/v1"
|
2017-02-09 19:59:19 +00:00
|
|
|
autoscalinglisters "k8s.io/kubernetes/pkg/client/listers/autoscaling/v1"
|
2017-03-08 07:02:34 +00:00
|
|
|
"k8s.io/kubernetes/pkg/controller"
|
2015-08-20 12:55:28 +00:00
|
|
|
)
|
|
|
|
|
2015-12-13 08:54:43 +00:00
|
|
|
const (
|
|
|
|
// Usage shoud exceed the tolerance before we start downscale or upscale the pods.
|
|
|
|
// TODO: make it a flag or HPA spec element.
|
|
|
|
tolerance = 0.1
|
2016-01-29 11:20:19 +00:00
|
|
|
|
2016-10-17 15:14:15 +00:00
|
|
|
scaleUpLimitFactor = 2
|
|
|
|
scaleUpLimitMinimum = 4
|
2015-12-13 08:54:43 +00:00
|
|
|
)
|
|
|
|
|
2016-10-17 15:14:15 +00:00
|
|
|
func calculateScaleUpLimit(currentReplicas int32) int32 {
|
|
|
|
return int32(math.Max(scaleUpLimitFactor*float64(currentReplicas), scaleUpLimitMinimum))
|
|
|
|
}
|
|
|
|
|
2017-02-21 15:46:29 +00:00
|
|
|
// UnsafeConvertToVersionVia is like api.Scheme.UnsafeConvertToVersion, but it does so via an internal version first.
|
|
|
|
// We use it since working with v2alpha1 is convenient here, but we want to use the v1 client (and
|
|
|
|
// can't just use the internal version). Note that conversion mutates the object, so you need to deepcopy
|
|
|
|
// *before* you call this if the input object came out of a shared cache.
|
2016-12-02 20:18:26 +00:00
|
|
|
func UnsafeConvertToVersionVia(obj runtime.Object, externalVersion schema.GroupVersion) (runtime.Object, error) {
|
|
|
|
objInt, err := api.Scheme.UnsafeConvertToVersion(obj, schema.GroupVersion{Group: externalVersion.Group, Version: runtime.APIVersionInternal})
|
|
|
|
if err != nil {
|
|
|
|
return nil, fmt.Errorf("failed to convert the given object to the internal version: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
objExt, err := api.Scheme.UnsafeConvertToVersion(objInt, externalVersion)
|
|
|
|
if err != nil {
|
|
|
|
return nil, fmt.Errorf("failed to convert the given object back to the external version: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
return objExt, err
|
|
|
|
}
|
|
|
|
|
2015-09-10 13:10:07 +00:00
|
|
|
type HorizontalController struct {
|
2016-12-02 20:18:26 +00:00
|
|
|
scaleNamespacer extensionsclient.ScalesGetter
|
|
|
|
hpaNamespacer autoscalingclient.HorizontalPodAutoscalersGetter
|
2015-11-02 15:18:53 +00:00
|
|
|
|
HPA: Consider unready pods and missing metrics
Currently, the HPA considers unready pods the same as ready pods when
looking at their CPU and custom metric usage. However, pods frequently
use extra CPU during initialization, so we want to consider them
separately.
This commit causes the HPA to consider unready pods as having 0 CPU
usage when scaling up, and ignores them when scaling down. If, when
scaling up, factoring the unready pods as having 0 CPU would cause a
downscale instead, we simply choose not to scale. Otherwise, we simply
scale up at the reduced amount caculated by factoring the pods in at
zero CPU usage.
The effect is that unready pods cause the autoscaler to be a bit more
conservative -- large increases in CPU usage can still cause scales,
even with unready pods in the mix, but will not cause the scale factors
to be as large, in anticipation of the new pods later becoming ready and
handling load.
Similarly, if there are pods for which no metrics have been retrieved,
these pods are treated as having 100% of the requested metric when
scaling down, and 0% when scaling up. As above, this cannot change the
direction of the scale.
This commit also changes the HPA to ignore superfluous metrics -- as
long as metrics for all ready pods are present, the HPA we make scaling
decisions. Currently, this only works for CPU. For custom metrics, we
cannot identify which metrics go to which pods if we get superfluous
metrics, so we abort the scale.
2016-09-27 18:47:52 +00:00
|
|
|
replicaCalc *ReplicaCalculator
|
2015-12-13 08:54:43 +00:00
|
|
|
eventRecorder record.EventRecorder
|
2016-03-02 08:29:17 +00:00
|
|
|
|
2017-02-09 19:59:19 +00:00
|
|
|
// hpaLister is able to list/get HPAs from the shared cache from the informer passed in to
|
|
|
|
// NewHorizontalController.
|
|
|
|
hpaLister autoscalinglisters.HorizontalPodAutoscalerLister
|
|
|
|
hpaListerSynced cache.InformerSynced
|
2017-03-08 07:02:34 +00:00
|
|
|
|
|
|
|
// Controllers that need to be synced
|
|
|
|
queue workqueue.RateLimitingInterface
|
2015-08-25 17:16:47 +00:00
|
|
|
}
|
|
|
|
|
2015-12-13 08:54:43 +00:00
|
|
|
var downscaleForbiddenWindow = 5 * time.Minute
|
|
|
|
var upscaleForbiddenWindow = 3 * time.Minute
|
|
|
|
|
2017-02-09 19:59:19 +00:00
|
|
|
func NewHorizontalController(
|
|
|
|
evtNamespacer v1core.EventsGetter,
|
2016-12-02 20:18:26 +00:00
|
|
|
scaleNamespacer extensionsclient.ScalesGetter,
|
|
|
|
hpaNamespacer autoscalingclient.HorizontalPodAutoscalersGetter,
|
2017-02-09 19:59:19 +00:00
|
|
|
replicaCalc *ReplicaCalculator,
|
|
|
|
hpaInformer autoscalinginformers.HorizontalPodAutoscalerInformer,
|
|
|
|
resyncPeriod time.Duration,
|
|
|
|
) *HorizontalController {
|
|
|
|
broadcaster := record.NewBroadcaster()
|
|
|
|
// TODO: remove the wrapper when every clients have moved to use the clientset.
|
|
|
|
broadcaster.StartRecordingToSink(&v1core.EventSinkImpl{Interface: evtNamespacer.Events("")})
|
|
|
|
recorder := broadcaster.NewRecorder(api.Scheme, clientv1.EventSource{Component: "horizontal-pod-autoscaler"})
|
|
|
|
|
2017-03-08 07:02:34 +00:00
|
|
|
hpaController := &HorizontalController{
|
2017-02-09 19:59:19 +00:00
|
|
|
replicaCalc: replicaCalc,
|
|
|
|
eventRecorder: recorder,
|
|
|
|
scaleNamespacer: scaleNamespacer,
|
|
|
|
hpaNamespacer: hpaNamespacer,
|
2017-03-08 07:02:34 +00:00
|
|
|
queue: workqueue.NewNamedRateLimitingQueue(NewDefaultHPARateLimiter(resyncPeriod), "horizontalpodautoscaler"),
|
2017-02-09 19:59:19 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
hpaInformer.Informer().AddEventHandlerWithResyncPeriod(
|
2016-09-14 18:35:38 +00:00
|
|
|
cache.ResourceEventHandlerFuncs{
|
2017-03-08 07:02:34 +00:00
|
|
|
AddFunc: hpaController.enqueueHPA,
|
|
|
|
UpdateFunc: hpaController.updateHPA,
|
|
|
|
DeleteFunc: hpaController.deleteHPA,
|
2016-03-02 08:29:17 +00:00
|
|
|
},
|
2017-02-09 19:59:19 +00:00
|
|
|
resyncPeriod,
|
2016-03-02 08:29:17 +00:00
|
|
|
)
|
2017-03-08 07:02:34 +00:00
|
|
|
hpaController.hpaLister = hpaInformer.Lister()
|
|
|
|
hpaController.hpaListerSynced = hpaInformer.Informer().HasSynced
|
2016-03-02 08:29:17 +00:00
|
|
|
|
2017-03-08 07:02:34 +00:00
|
|
|
return hpaController
|
2015-08-17 12:18:26 +00:00
|
|
|
}
|
|
|
|
|
2016-03-02 08:29:17 +00:00
|
|
|
func (a *HorizontalController) Run(stopCh <-chan struct{}) {
|
|
|
|
defer utilruntime.HandleCrash()
|
2017-03-08 07:02:34 +00:00
|
|
|
defer a.queue.ShutDown()
|
2017-02-09 19:59:19 +00:00
|
|
|
|
2016-03-02 08:29:17 +00:00
|
|
|
glog.Infof("Starting HPA Controller")
|
2017-02-09 19:59:19 +00:00
|
|
|
|
|
|
|
if !cache.WaitForCacheSync(stopCh, a.hpaListerSynced) {
|
|
|
|
utilruntime.HandleError(fmt.Errorf("timed out waiting for caches to sync"))
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2017-03-08 07:02:34 +00:00
|
|
|
// start a single worker (we may wish to start more in the future)
|
|
|
|
go wait.Until(a.worker, time.Second, stopCh)
|
|
|
|
|
2016-03-02 08:29:17 +00:00
|
|
|
<-stopCh
|
|
|
|
glog.Infof("Shutting down HPA Controller")
|
2015-08-17 12:18:26 +00:00
|
|
|
}
|
|
|
|
|
2017-03-08 07:02:34 +00:00
|
|
|
// obj could be an *v1.HorizontalPodAutoscaler, or a DeletionFinalStateUnknown marker item.
|
|
|
|
func (a *HorizontalController) updateHPA(old, cur interface{}) {
|
|
|
|
a.enqueueHPA(cur)
|
|
|
|
}
|
|
|
|
|
|
|
|
// obj could be an *v1.HorizontalPodAutoscaler, or a DeletionFinalStateUnknown marker item.
|
|
|
|
func (a *HorizontalController) enqueueHPA(obj interface{}) {
|
|
|
|
key, err := controller.KeyFunc(obj)
|
|
|
|
if err != nil {
|
|
|
|
utilruntime.HandleError(fmt.Errorf("couldn't get key for object %+v: %v", obj, err))
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// always add rate-limitted so we don't fetch metrics more that once per resync interval
|
|
|
|
a.queue.AddRateLimited(key)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (a *HorizontalController) deleteHPA(obj interface{}) {
|
|
|
|
key, err := controller.KeyFunc(obj)
|
|
|
|
if err != nil {
|
|
|
|
utilruntime.HandleError(fmt.Errorf("couldn't get key for object %+v: %v", obj, err))
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// TODO: could we leak if we fail to get the key?
|
|
|
|
a.queue.Forget(key)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (a *HorizontalController) worker() {
|
|
|
|
for a.processNextWorkItem() {
|
|
|
|
}
|
|
|
|
glog.Infof("horizontal pod autoscaler controller worker shutting down")
|
|
|
|
}
|
|
|
|
|
|
|
|
func (a *HorizontalController) processNextWorkItem() bool {
|
|
|
|
key, quit := a.queue.Get()
|
|
|
|
if quit {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
defer a.queue.Done(key)
|
|
|
|
|
|
|
|
err := a.reconcileKey(key.(string))
|
|
|
|
if err == nil {
|
|
|
|
// don't "forget" here because we want to only process a given HPA once per resync interval
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
a.queue.AddRateLimited(key)
|
|
|
|
utilruntime.HandleError(err)
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
2016-12-02 20:18:26 +00:00
|
|
|
// Computes the desired number of replicas for the metric specifications listed in the HPA, returning the maximum
|
|
|
|
// of the computed replica counts, a description of the associated metric, and the statuses of all metrics
|
|
|
|
// computed.
|
|
|
|
func (a *HorizontalController) computeReplicasForMetrics(hpa *autoscalingv2.HorizontalPodAutoscaler, scale *extensions.Scale,
|
|
|
|
metricSpecs []autoscalingv2.MetricSpec) (replicas int32, metric string, statuses []autoscalingv2.MetricStatus, timestamp time.Time, err error) {
|
2016-01-29 11:20:19 +00:00
|
|
|
|
2016-07-11 06:42:51 +00:00
|
|
|
currentReplicas := scale.Status.Replicas
|
|
|
|
|
2016-12-02 20:18:26 +00:00
|
|
|
statuses = make([]autoscalingv2.MetricStatus, len(metricSpecs))
|
2016-01-29 11:20:19 +00:00
|
|
|
|
2016-12-02 20:18:26 +00:00
|
|
|
for i, metricSpec := range metricSpecs {
|
|
|
|
if len(scale.Status.Selector) == 0 && len(scale.Status.TargetSelector) == 0 {
|
2016-03-09 00:27:13 +00:00
|
|
|
errMsg := "selector is required"
|
2016-11-18 20:50:17 +00:00
|
|
|
a.eventRecorder.Event(hpa, v1.EventTypeWarning, "SelectorRequired", errMsg)
|
2016-12-02 20:18:26 +00:00
|
|
|
return 0, "", nil, time.Time{}, fmt.Errorf(errMsg)
|
2016-03-09 00:27:13 +00:00
|
|
|
}
|
|
|
|
|
2016-12-02 20:18:26 +00:00
|
|
|
var selector labels.Selector
|
|
|
|
var err error
|
|
|
|
if len(scale.Status.Selector) > 0 {
|
|
|
|
selector = labels.SelectorFromSet(labels.Set(scale.Status.Selector))
|
|
|
|
err = nil
|
|
|
|
} else {
|
|
|
|
selector, err = labels.Parse(scale.Status.TargetSelector)
|
|
|
|
}
|
2016-03-09 00:27:13 +00:00
|
|
|
if err != nil {
|
2016-12-02 20:18:26 +00:00
|
|
|
errMsg := fmt.Sprintf("couldn't convert selector into a corresponding internal selector object: %v", err)
|
2016-11-18 20:50:17 +00:00
|
|
|
a.eventRecorder.Event(hpa, v1.EventTypeWarning, "InvalidSelector", errMsg)
|
2016-12-02 20:18:26 +00:00
|
|
|
return 0, "", nil, time.Time{}, fmt.Errorf(errMsg)
|
2016-03-09 00:27:13 +00:00
|
|
|
}
|
2016-12-02 20:18:26 +00:00
|
|
|
|
|
|
|
var replicaCountProposal int32
|
|
|
|
var utilizationProposal int64
|
|
|
|
var timestampProposal time.Time
|
|
|
|
var metricNameProposal string
|
|
|
|
|
|
|
|
switch metricSpec.Type {
|
|
|
|
case autoscalingv2.ObjectMetricSourceType:
|
|
|
|
replicaCountProposal, utilizationProposal, timestampProposal, err = a.replicaCalc.GetObjectMetricReplicas(currentReplicas, metricSpec.Object.TargetValue.MilliValue(), metricSpec.Object.MetricName, hpa.Namespace, &metricSpec.Object.Target)
|
|
|
|
if err != nil {
|
|
|
|
a.eventRecorder.Event(hpa, v1.EventTypeWarning, "FailedGetObjectMetric", err.Error())
|
|
|
|
return 0, "", nil, time.Time{}, fmt.Errorf("failed to get object metric value: %v", err)
|
|
|
|
}
|
|
|
|
metricNameProposal = fmt.Sprintf("%s metric %s", metricSpec.Object.Target.Kind, metricSpec.Object.MetricName)
|
|
|
|
statuses[i] = autoscalingv2.MetricStatus{
|
|
|
|
Type: autoscalingv2.ObjectMetricSourceType,
|
|
|
|
Object: &autoscalingv2.ObjectMetricStatus{
|
|
|
|
Target: metricSpec.Object.Target,
|
|
|
|
MetricName: metricSpec.Object.MetricName,
|
|
|
|
CurrentValue: *resource.NewMilliQuantity(utilizationProposal, resource.DecimalSI),
|
|
|
|
},
|
2016-11-04 14:38:06 +00:00
|
|
|
}
|
2016-12-02 20:18:26 +00:00
|
|
|
case autoscalingv2.PodsMetricSourceType:
|
|
|
|
replicaCountProposal, utilizationProposal, timestampProposal, err = a.replicaCalc.GetMetricReplicas(currentReplicas, metricSpec.Pods.TargetAverageValue.MilliValue(), metricSpec.Pods.MetricName, hpa.Namespace, selector)
|
|
|
|
if err != nil {
|
|
|
|
a.eventRecorder.Event(hpa, v1.EventTypeWarning, "FailedGetPodsMetric", err.Error())
|
|
|
|
return 0, "", nil, time.Time{}, fmt.Errorf("failed to get pods metric value: %v", err)
|
|
|
|
}
|
|
|
|
metricNameProposal = fmt.Sprintf("pods metric %s", metricSpec.Pods.MetricName)
|
|
|
|
statuses[i] = autoscalingv2.MetricStatus{
|
|
|
|
Type: autoscalingv2.PodsMetricSourceType,
|
|
|
|
Pods: &autoscalingv2.PodsMetricStatus{
|
|
|
|
MetricName: metricSpec.Pods.MetricName,
|
|
|
|
CurrentAverageValue: *resource.NewMilliQuantity(utilizationProposal, resource.DecimalSI),
|
|
|
|
},
|
|
|
|
}
|
|
|
|
case autoscalingv2.ResourceMetricSourceType:
|
|
|
|
if metricSpec.Resource.TargetAverageValue != nil {
|
|
|
|
var rawProposal int64
|
|
|
|
replicaCountProposal, rawProposal, timestampProposal, err = a.replicaCalc.GetRawResourceReplicas(currentReplicas, metricSpec.Resource.TargetAverageValue.MilliValue(), metricSpec.Resource.Name, hpa.Namespace, selector)
|
|
|
|
if err != nil {
|
|
|
|
a.eventRecorder.Event(hpa, v1.EventTypeWarning, "FailedGetResourceMetric", err.Error())
|
|
|
|
return 0, "", nil, time.Time{}, fmt.Errorf("failed to get %s utilization: %v", metricSpec.Resource.Name, err)
|
|
|
|
}
|
|
|
|
metricNameProposal = fmt.Sprintf("%s resource", metricSpec.Resource.Name)
|
|
|
|
statuses[i] = autoscalingv2.MetricStatus{
|
|
|
|
Type: autoscalingv2.ResourceMetricSourceType,
|
|
|
|
Resource: &autoscalingv2.ResourceMetricStatus{
|
|
|
|
Name: metricSpec.Resource.Name,
|
|
|
|
CurrentAverageValue: *resource.NewMilliQuantity(rawProposal, resource.DecimalSI),
|
|
|
|
},
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// set a default utilization percentage if none is set
|
|
|
|
if metricSpec.Resource.TargetAverageUtilization == nil {
|
|
|
|
errMsg := "invalid resource metric source: neither a utilization target nor a value target was set"
|
|
|
|
a.eventRecorder.Event(hpa, v1.EventTypeWarning, "FailedGetResourceMetric", errMsg)
|
|
|
|
return 0, "", nil, time.Time{}, fmt.Errorf(errMsg)
|
|
|
|
}
|
|
|
|
|
|
|
|
targetUtilization := *metricSpec.Resource.TargetAverageUtilization
|
2016-11-04 14:38:06 +00:00
|
|
|
|
2016-12-02 20:18:26 +00:00
|
|
|
var percentageProposal int32
|
|
|
|
var rawProposal int64
|
|
|
|
replicaCountProposal, percentageProposal, rawProposal, timestampProposal, err = a.replicaCalc.GetResourceReplicas(currentReplicas, targetUtilization, metricSpec.Resource.Name, hpa.Namespace, selector)
|
|
|
|
if err != nil {
|
|
|
|
a.eventRecorder.Event(hpa, v1.EventTypeWarning, "FailedGetResourceMetric", err.Error())
|
|
|
|
return 0, "", nil, time.Time{}, fmt.Errorf("failed to get %s utilization: %v", metricSpec.Resource.Name, err)
|
|
|
|
}
|
|
|
|
metricNameProposal = fmt.Sprintf("%s resource utilization (percentage of request)", metricSpec.Resource.Name)
|
|
|
|
statuses[i] = autoscalingv2.MetricStatus{
|
|
|
|
Type: autoscalingv2.ResourceMetricSourceType,
|
|
|
|
Resource: &autoscalingv2.ResourceMetricStatus{
|
|
|
|
Name: metricSpec.Resource.Name,
|
|
|
|
CurrentAverageUtilization: &percentageProposal,
|
|
|
|
CurrentAverageValue: *resource.NewMilliQuantity(rawProposal, resource.DecimalSI),
|
|
|
|
},
|
|
|
|
}
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
errMsg := fmt.Sprintf("unknown metric source type %q", string(metricSpec.Type))
|
|
|
|
a.eventRecorder.Event(hpa, v1.EventTypeWarning, "InvalidMetricSourceType", errMsg)
|
|
|
|
return 0, "", nil, time.Time{}, fmt.Errorf(errMsg)
|
2016-01-29 11:20:19 +00:00
|
|
|
}
|
|
|
|
|
2016-12-02 20:18:26 +00:00
|
|
|
if replicas == 0 || replicaCountProposal > replicas {
|
HPA: Consider unready pods and missing metrics
Currently, the HPA considers unready pods the same as ready pods when
looking at their CPU and custom metric usage. However, pods frequently
use extra CPU during initialization, so we want to consider them
separately.
This commit causes the HPA to consider unready pods as having 0 CPU
usage when scaling up, and ignores them when scaling down. If, when
scaling up, factoring the unready pods as having 0 CPU would cause a
downscale instead, we simply choose not to scale. Otherwise, we simply
scale up at the reduced amount caculated by factoring the pods in at
zero CPU usage.
The effect is that unready pods cause the autoscaler to be a bit more
conservative -- large increases in CPU usage can still cause scales,
even with unready pods in the mix, but will not cause the scale factors
to be as large, in anticipation of the new pods later becoming ready and
handling load.
Similarly, if there are pods for which no metrics have been retrieved,
these pods are treated as having 100% of the requested metric when
scaling down, and 0% when scaling up. As above, this cannot change the
direction of the scale.
This commit also changes the HPA to ignore superfluous metrics -- as
long as metrics for all ready pods are present, the HPA we make scaling
decisions. Currently, this only works for CPU. For custom metrics, we
cannot identify which metrics go to which pods if we get superfluous
metrics, so we abort the scale.
2016-09-27 18:47:52 +00:00
|
|
|
timestamp = timestampProposal
|
2016-01-29 11:20:19 +00:00
|
|
|
replicas = replicaCountProposal
|
2016-12-02 20:18:26 +00:00
|
|
|
metric = metricNameProposal
|
2016-01-29 11:20:19 +00:00
|
|
|
}
|
2016-11-08 12:11:08 +00:00
|
|
|
}
|
|
|
|
|
2016-12-02 20:18:26 +00:00
|
|
|
return replicas, metric, statuses, timestamp, nil
|
2016-01-29 11:20:19 +00:00
|
|
|
}
|
|
|
|
|
2017-03-08 07:02:34 +00:00
|
|
|
func (a *HorizontalController) reconcileKey(key string) error {
|
|
|
|
namespace, name, err := cache.SplitMetaNamespaceKey(key)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
hpa, err := a.hpaLister.HorizontalPodAutoscalers(namespace).Get(name)
|
|
|
|
if errors.IsNotFound(err) {
|
|
|
|
glog.Infof("Horizontal Pod Autoscaler has been deleted %v", key)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
return a.reconcileAutoscaler(hpa)
|
|
|
|
}
|
|
|
|
|
2017-02-21 15:46:29 +00:00
|
|
|
func (a *HorizontalController) reconcileAutoscaler(hpav1Shared *autoscalingv1.HorizontalPodAutoscaler) error {
|
|
|
|
// make a copy so that we never mutate the shared informer cache (conversion can mutate the object)
|
|
|
|
hpav1Raw, err := api.Scheme.DeepCopy(hpav1Shared)
|
|
|
|
if err != nil {
|
|
|
|
a.eventRecorder.Event(hpav1Shared, v1.EventTypeWarning, "FailedConvertHPA", err.Error())
|
|
|
|
return fmt.Errorf("failed to deep-copy the HPA: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// then, convert to autoscaling/v2, which makes our lives easier when calculating metrics
|
|
|
|
hpav1 := hpav1Raw.(*autoscalingv1.HorizontalPodAutoscaler)
|
2016-12-02 20:18:26 +00:00
|
|
|
hpaRaw, err := UnsafeConvertToVersionVia(hpav1, autoscalingv2.SchemeGroupVersion)
|
|
|
|
if err != nil {
|
|
|
|
a.eventRecorder.Event(hpav1, v1.EventTypeWarning, "FailedConvertHPA", err.Error())
|
|
|
|
return fmt.Errorf("failed to convert the given HPA to %s: %v", autoscalingv2.SchemeGroupVersion.String(), err)
|
|
|
|
}
|
|
|
|
hpa := hpaRaw.(*autoscalingv2.HorizontalPodAutoscaler)
|
|
|
|
|
2016-05-05 10:27:24 +00:00
|
|
|
reference := fmt.Sprintf("%s/%s/%s", hpa.Spec.ScaleTargetRef.Kind, hpa.Namespace, hpa.Spec.ScaleTargetRef.Name)
|
2015-09-14 13:08:43 +00:00
|
|
|
|
2016-05-05 10:27:24 +00:00
|
|
|
scale, err := a.scaleNamespacer.Scales(hpa.Namespace).Get(hpa.Spec.ScaleTargetRef.Kind, hpa.Spec.ScaleTargetRef.Name)
|
2015-08-17 12:18:26 +00:00
|
|
|
if err != nil {
|
2016-11-18 20:50:17 +00:00
|
|
|
a.eventRecorder.Event(hpa, v1.EventTypeWarning, "FailedGetScale", err.Error())
|
2015-09-14 13:08:43 +00:00
|
|
|
return fmt.Errorf("failed to query scale subresource for %s: %v", reference, err)
|
2015-08-17 12:18:26 +00:00
|
|
|
}
|
2015-09-14 13:08:43 +00:00
|
|
|
currentReplicas := scale.Status.Replicas
|
2015-08-20 12:55:28 +00:00
|
|
|
|
2016-12-02 20:18:26 +00:00
|
|
|
var metricStatuses []autoscalingv2.MetricStatus
|
|
|
|
metricDesiredReplicas := int32(0)
|
|
|
|
metricName := ""
|
|
|
|
metricTimestamp := time.Time{}
|
2016-01-29 11:20:19 +00:00
|
|
|
|
2016-04-27 04:35:14 +00:00
|
|
|
desiredReplicas := int32(0)
|
2016-03-02 09:08:17 +00:00
|
|
|
rescaleReason := ""
|
2016-02-12 15:26:59 +00:00
|
|
|
timestamp := time.Now()
|
2016-01-29 11:20:19 +00:00
|
|
|
|
2016-11-21 09:32:00 +00:00
|
|
|
rescale := true
|
|
|
|
|
2016-07-19 15:54:38 +00:00
|
|
|
if scale.Spec.Replicas == 0 {
|
|
|
|
// Autoscaling is disabled for this resource
|
|
|
|
desiredReplicas = 0
|
2016-11-21 09:32:00 +00:00
|
|
|
rescale = false
|
2016-07-19 15:54:38 +00:00
|
|
|
} else if currentReplicas > hpa.Spec.MaxReplicas {
|
2016-03-02 09:08:17 +00:00
|
|
|
rescaleReason = "Current number of replicas above Spec.MaxReplicas"
|
2016-02-12 15:26:59 +00:00
|
|
|
desiredReplicas = hpa.Spec.MaxReplicas
|
|
|
|
} else if hpa.Spec.MinReplicas != nil && currentReplicas < *hpa.Spec.MinReplicas {
|
2016-03-02 09:08:17 +00:00
|
|
|
rescaleReason = "Current number of replicas below Spec.MinReplicas"
|
2016-02-12 15:26:59 +00:00
|
|
|
desiredReplicas = *hpa.Spec.MinReplicas
|
|
|
|
} else if currentReplicas == 0 {
|
2016-03-02 09:08:17 +00:00
|
|
|
rescaleReason = "Current number of replicas must be greater than 0"
|
2016-02-12 15:26:59 +00:00
|
|
|
desiredReplicas = 1
|
|
|
|
} else {
|
2016-12-02 20:18:26 +00:00
|
|
|
metricDesiredReplicas, metricName, metricStatuses, metricTimestamp, err = a.computeReplicasForMetrics(hpa, scale, hpa.Spec.Metrics)
|
|
|
|
if err != nil {
|
|
|
|
a.updateCurrentReplicasInStatus(hpa, currentReplicas)
|
|
|
|
a.eventRecorder.Event(hpa, v1.EventTypeWarning, "FailedComputeMetricsReplicas", err.Error())
|
|
|
|
return fmt.Errorf("failed to compute desired number of replicas based on listed metrics for %s: %v", reference, err)
|
2016-01-29 11:20:19 +00:00
|
|
|
}
|
|
|
|
|
2016-12-02 20:18:26 +00:00
|
|
|
glog.V(4).Infof("proposing %v desired replicas (based on %s from %s) for %s", metricDesiredReplicas, metricName, timestamp, reference)
|
2016-01-29 11:20:19 +00:00
|
|
|
|
2016-03-02 09:08:17 +00:00
|
|
|
rescaleMetric := ""
|
2016-12-02 20:18:26 +00:00
|
|
|
if metricDesiredReplicas > desiredReplicas {
|
|
|
|
desiredReplicas = metricDesiredReplicas
|
|
|
|
timestamp = metricTimestamp
|
|
|
|
rescaleMetric = metricName
|
2016-03-02 09:08:17 +00:00
|
|
|
}
|
|
|
|
if desiredReplicas > currentReplicas {
|
|
|
|
rescaleReason = fmt.Sprintf("%s above target", rescaleMetric)
|
2016-07-11 06:42:51 +00:00
|
|
|
}
|
|
|
|
if desiredReplicas < currentReplicas {
|
2016-03-02 09:08:17 +00:00
|
|
|
rescaleReason = "All metrics below target"
|
2016-02-12 15:26:59 +00:00
|
|
|
}
|
2015-08-20 12:55:28 +00:00
|
|
|
|
2016-02-12 15:26:59 +00:00
|
|
|
if hpa.Spec.MinReplicas != nil && desiredReplicas < *hpa.Spec.MinReplicas {
|
|
|
|
desiredReplicas = *hpa.Spec.MinReplicas
|
|
|
|
}
|
2015-08-25 17:16:47 +00:00
|
|
|
|
2016-07-19 15:54:38 +00:00
|
|
|
// never scale down to 0, reserved for disabling autoscaling
|
2016-02-12 15:26:59 +00:00
|
|
|
if desiredReplicas == 0 {
|
|
|
|
desiredReplicas = 1
|
|
|
|
}
|
2015-09-07 10:25:04 +00:00
|
|
|
|
2016-02-12 15:26:59 +00:00
|
|
|
if desiredReplicas > hpa.Spec.MaxReplicas {
|
|
|
|
desiredReplicas = hpa.Spec.MaxReplicas
|
|
|
|
}
|
2016-10-17 15:14:15 +00:00
|
|
|
|
|
|
|
// Do not upscale too much to prevent incorrect rapid increase of the number of master replicas caused by
|
|
|
|
// bogus CPU usage report from heapster/kubelet (like in issue #32304).
|
|
|
|
if desiredReplicas > calculateScaleUpLimit(currentReplicas) {
|
|
|
|
desiredReplicas = calculateScaleUpLimit(currentReplicas)
|
|
|
|
}
|
2016-11-21 09:32:00 +00:00
|
|
|
|
|
|
|
rescale = shouldScale(hpa, currentReplicas, desiredReplicas, timestamp)
|
2015-09-14 13:08:43 +00:00
|
|
|
}
|
|
|
|
|
2016-02-23 10:29:40 +00:00
|
|
|
if rescale {
|
|
|
|
scale.Spec.Replicas = desiredReplicas
|
2016-05-05 10:27:24 +00:00
|
|
|
_, err = a.scaleNamespacer.Scales(hpa.Namespace).Update(hpa.Spec.ScaleTargetRef.Kind, scale)
|
2016-02-23 10:29:40 +00:00
|
|
|
if err != nil {
|
2016-11-18 20:50:17 +00:00
|
|
|
a.eventRecorder.Eventf(hpa, v1.EventTypeWarning, "FailedRescale", "New size: %d; reason: %s; error: %v", desiredReplicas, rescaleReason, err.Error())
|
2016-02-23 10:29:40 +00:00
|
|
|
return fmt.Errorf("failed to rescale %s: %v", reference, err)
|
|
|
|
}
|
2016-11-18 20:50:17 +00:00
|
|
|
a.eventRecorder.Eventf(hpa, v1.EventTypeNormal, "SuccessfulRescale", "New size: %d; reason: %s", desiredReplicas, rescaleReason)
|
2016-03-02 09:08:17 +00:00
|
|
|
glog.Infof("Successfull rescale of %s, old size: %d, new size: %d, reason: %s",
|
|
|
|
hpa.Name, currentReplicas, desiredReplicas, rescaleReason)
|
2016-02-23 10:29:40 +00:00
|
|
|
} else {
|
2016-12-02 20:18:26 +00:00
|
|
|
glog.V(4).Infof("decided not to scale %s to %v (last scale time was %s)", reference, desiredReplicas, hpa.Status.LastScaleTime)
|
2016-02-23 10:29:40 +00:00
|
|
|
desiredReplicas = currentReplicas
|
|
|
|
}
|
|
|
|
|
2016-12-02 20:18:26 +00:00
|
|
|
return a.updateStatus(hpa, currentReplicas, desiredReplicas, metricStatuses, rescale)
|
2016-02-23 10:29:40 +00:00
|
|
|
}
|
|
|
|
|
2016-12-02 20:18:26 +00:00
|
|
|
func shouldScale(hpa *autoscalingv2.HorizontalPodAutoscaler, currentReplicas, desiredReplicas int32, timestamp time.Time) bool {
|
2016-07-11 06:42:51 +00:00
|
|
|
if desiredReplicas == currentReplicas {
|
|
|
|
return false
|
|
|
|
}
|
2015-09-07 10:25:04 +00:00
|
|
|
|
2016-08-16 11:02:08 +00:00
|
|
|
if hpa.Status.LastScaleTime == nil {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
2016-07-11 06:42:51 +00:00
|
|
|
// Going down only if the usageRatio dropped significantly below the target
|
|
|
|
// and there was no rescaling in the last downscaleForbiddenWindow.
|
2016-08-16 11:02:08 +00:00
|
|
|
if desiredReplicas < currentReplicas && hpa.Status.LastScaleTime.Add(downscaleForbiddenWindow).Before(timestamp) {
|
2016-07-11 06:42:51 +00:00
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// Going up only if the usage ratio increased significantly above the target
|
|
|
|
// and there was no rescaling in the last upscaleForbiddenWindow.
|
2016-08-16 11:02:08 +00:00
|
|
|
if desiredReplicas > currentReplicas && hpa.Status.LastScaleTime.Add(upscaleForbiddenWindow).Before(timestamp) {
|
2016-07-11 06:42:51 +00:00
|
|
|
return true
|
2015-09-14 13:08:43 +00:00
|
|
|
}
|
2016-12-02 20:18:26 +00:00
|
|
|
|
2016-02-23 10:29:40 +00:00
|
|
|
return false
|
|
|
|
}
|
2015-08-25 17:16:47 +00:00
|
|
|
|
2016-12-02 20:18:26 +00:00
|
|
|
func (a *HorizontalController) updateCurrentReplicasInStatus(hpa *autoscalingv2.HorizontalPodAutoscaler, currentReplicas int32) {
|
|
|
|
err := a.updateStatus(hpa, currentReplicas, hpa.Status.DesiredReplicas, hpa.Status.CurrentMetrics, false)
|
2016-02-23 12:05:07 +00:00
|
|
|
if err != nil {
|
2017-02-09 19:59:19 +00:00
|
|
|
utilruntime.HandleError(err)
|
2016-02-23 12:05:07 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-12-02 20:18:26 +00:00
|
|
|
func (a *HorizontalController) updateStatus(hpa *autoscalingv2.HorizontalPodAutoscaler, currentReplicas, desiredReplicas int32, metricStatuses []autoscalingv2.MetricStatus, rescale bool) error {
|
|
|
|
hpa.Status = autoscalingv2.HorizontalPodAutoscalerStatus{
|
|
|
|
CurrentReplicas: currentReplicas,
|
|
|
|
DesiredReplicas: desiredReplicas,
|
|
|
|
LastScaleTime: hpa.Status.LastScaleTime,
|
|
|
|
CurrentMetrics: metricStatuses,
|
2016-01-29 11:20:19 +00:00
|
|
|
}
|
|
|
|
|
2015-09-14 13:08:43 +00:00
|
|
|
if rescale {
|
2016-12-03 18:57:26 +00:00
|
|
|
now := metav1.NewTime(time.Now())
|
2015-10-13 15:24:23 +00:00
|
|
|
hpa.Status.LastScaleTime = &now
|
2015-09-14 13:08:43 +00:00
|
|
|
}
|
|
|
|
|
2016-12-02 20:18:26 +00:00
|
|
|
// convert back to autoscalingv1
|
|
|
|
hpaRaw, err := UnsafeConvertToVersionVia(hpa, autoscalingv1.SchemeGroupVersion)
|
|
|
|
if err != nil {
|
|
|
|
a.eventRecorder.Event(hpa, v1.EventTypeWarning, "FailedConvertHPA", err.Error())
|
|
|
|
return fmt.Errorf("failed to convert the given HPA to %s: %v", autoscalingv2.SchemeGroupVersion.String(), err)
|
|
|
|
}
|
|
|
|
hpav1 := hpaRaw.(*autoscalingv1.HorizontalPodAutoscaler)
|
|
|
|
|
|
|
|
_, err = a.hpaNamespacer.HorizontalPodAutoscalers(hpav1.Namespace).UpdateStatus(hpav1)
|
2015-09-14 13:08:43 +00:00
|
|
|
if err != nil {
|
2016-11-18 20:50:17 +00:00
|
|
|
a.eventRecorder.Event(hpa, v1.EventTypeWarning, "FailedUpdateStatus", err.Error())
|
2015-09-14 13:08:43 +00:00
|
|
|
return fmt.Errorf("failed to update status for %s: %v", hpa.Name, err)
|
|
|
|
}
|
2016-03-02 08:29:17 +00:00
|
|
|
glog.V(2).Infof("Successfully updated status for %s", hpa.Name)
|
2015-08-17 12:18:26 +00:00
|
|
|
return nil
|
|
|
|
}
|