mirror of https://github.com/k3s-io/k3s
Merge pull request #59728 from wgliang/master.append
Automatic merge from submit-queue (batch tested with PRs 59740, 59728, 60080, 60086, 58714). If you want to cherry-pick this change to another branch, please follow the instructions <a href="https://github.com/kubernetes/community/blob/master/contributors/devel/cherry-picks.md">here</a>. more concise to merge the slice **What this PR does / why we need it**: more concise to merge the slice **Special notes for your reviewer**:pull/8/head
commit
c64f19dd1b
|
@ -247,9 +247,8 @@ func TestUpdateNodesInExternalLoadBalancer(t *testing.T) {
|
||||||
controller, cloud, _ := newController()
|
controller, cloud, _ := newController()
|
||||||
|
|
||||||
var services []*v1.Service
|
var services []*v1.Service
|
||||||
for _, service := range item.services {
|
services = append(services, item.services...)
|
||||||
services = append(services, service)
|
|
||||||
}
|
|
||||||
if err := controller.updateLoadBalancerHosts(services, nodes); err != nil {
|
if err := controller.updateLoadBalancerHosts(services, nodes); err != nil {
|
||||||
t.Errorf("unexpected error: %v", err)
|
t.Errorf("unexpected error: %v", err)
|
||||||
}
|
}
|
||||||
|
|
|
@ -404,9 +404,8 @@ func TestGenericScheduler(t *testing.T) {
|
||||||
cache.AddNode(&v1.Node{ObjectMeta: metav1.ObjectMeta{Name: name}})
|
cache.AddNode(&v1.Node{ObjectMeta: metav1.ObjectMeta{Name: name}})
|
||||||
}
|
}
|
||||||
pvcs := []*v1.PersistentVolumeClaim{}
|
pvcs := []*v1.PersistentVolumeClaim{}
|
||||||
for _, pvc := range test.pvcs {
|
pvcs = append(pvcs, test.pvcs...)
|
||||||
pvcs = append(pvcs, pvc)
|
|
||||||
}
|
|
||||||
pvcLister := schedulertesting.FakePersistentVolumeClaimLister(pvcs)
|
pvcLister := schedulertesting.FakePersistentVolumeClaimLister(pvcs)
|
||||||
|
|
||||||
scheduler := NewGenericScheduler(
|
scheduler := NewGenericScheduler(
|
||||||
|
|
|
@ -590,9 +590,7 @@ func (r *rbdVolumeProvisioner) Provision() (*v1.PersistentVolume, error) {
|
||||||
switch dstrings.ToLower(k) {
|
switch dstrings.ToLower(k) {
|
||||||
case "monitors":
|
case "monitors":
|
||||||
arr := dstrings.Split(v, ",")
|
arr := dstrings.Split(v, ",")
|
||||||
for _, m := range arr {
|
r.Mon = append(r.Mon, arr...)
|
||||||
r.Mon = append(r.Mon, m)
|
|
||||||
}
|
|
||||||
case "adminid":
|
case "adminid":
|
||||||
r.adminId = v
|
r.adminId = v
|
||||||
case "adminsecretname":
|
case "adminsecretname":
|
||||||
|
|
|
@ -276,9 +276,8 @@ func TestGCAdmission(t *testing.T) {
|
||||||
func TestBlockOwnerDeletionAdmission(t *testing.T) {
|
func TestBlockOwnerDeletionAdmission(t *testing.T) {
|
||||||
podWithOwnerRefs := func(refs ...metav1.OwnerReference) *api.Pod {
|
podWithOwnerRefs := func(refs ...metav1.OwnerReference) *api.Pod {
|
||||||
var refSlice []metav1.OwnerReference
|
var refSlice []metav1.OwnerReference
|
||||||
for _, ref := range refs {
|
refSlice = append(refSlice, refs...)
|
||||||
refSlice = append(refSlice, ref)
|
|
||||||
}
|
|
||||||
return &api.Pod{
|
return &api.Pod{
|
||||||
ObjectMeta: metav1.ObjectMeta{
|
ObjectMeta: metav1.ObjectMeta{
|
||||||
OwnerReferences: refSlice,
|
OwnerReferences: refSlice,
|
||||||
|
|
|
@ -112,9 +112,8 @@ func TestNamedCertKeyArrayFlag(t *testing.T) {
|
||||||
for i, test := range tests {
|
for i, test := range tests {
|
||||||
fs := pflag.NewFlagSet("testNamedCertKeyArray", pflag.ContinueOnError)
|
fs := pflag.NewFlagSet("testNamedCertKeyArray", pflag.ContinueOnError)
|
||||||
var nkcs []NamedCertKey
|
var nkcs []NamedCertKey
|
||||||
for _, d := range test.def {
|
nkcs = append(nkcs, test.def...)
|
||||||
nkcs = append(nkcs, d)
|
|
||||||
}
|
|
||||||
fs.Var(NewNamedCertKeyArray(&nkcs), "tls-sni-cert-key", "usage")
|
fs.Var(NewNamedCertKeyArray(&nkcs), "tls-sni-cert-key", "usage")
|
||||||
|
|
||||||
args := []string{}
|
args := []string{}
|
||||||
|
|
|
@ -220,9 +220,7 @@ func (r *ResourceCollector) GetBasicCPUStats(containerName string) map[float64]f
|
||||||
|
|
||||||
// We must make a copy of array, otherwise the timeseries order is changed.
|
// We must make a copy of array, otherwise the timeseries order is changed.
|
||||||
usages := make([]*framework.ContainerResourceUsage, 0)
|
usages := make([]*framework.ContainerResourceUsage, 0)
|
||||||
for _, usage := range r.buffers[containerName] {
|
usages = append(usages, r.buffers[containerName]...)
|
||||||
usages = append(usages, usage)
|
|
||||||
}
|
|
||||||
|
|
||||||
sort.Sort(resourceUsageByCPU(usages))
|
sort.Sort(resourceUsageByCPU(usages))
|
||||||
for _, q := range percentiles {
|
for _, q := range percentiles {
|
||||||
|
|
Loading…
Reference in New Issue