mirror of https://github.com/k3s-io/k3s
Remove check for SubResourcePodProxyVersion and SubResourceServiceAndNodeProxyVersion
parent
026197fb04
commit
32675e6f62
|
@ -280,17 +280,10 @@ func HighLatencyKubeletOperations(c clientset.Interface, threshold time.Duration
|
|||
|
||||
// getStatsSummary contacts kubelet for the container information.
|
||||
func getStatsSummary(c clientset.Interface, nodeName string) (*stats.Summary, error) {
|
||||
subResourceProxyAvailable, err := ServerVersionGTE(SubResourceServiceAndNodeProxyVersion, c.Discovery())
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
ctx, cancel := context.WithTimeout(context.Background(), SingleCallTimeout)
|
||||
defer cancel()
|
||||
|
||||
var data []byte
|
||||
if subResourceProxyAvailable {
|
||||
data, err = c.CoreV1().RESTClient().Get().
|
||||
data, err := c.CoreV1().RESTClient().Get().
|
||||
Context(ctx).
|
||||
Resource("nodes").
|
||||
SubResource("proxy").
|
||||
|
@ -298,15 +291,6 @@ func getStatsSummary(c clientset.Interface, nodeName string) (*stats.Summary, er
|
|||
Suffix("stats/summary").
|
||||
Do().Raw()
|
||||
|
||||
} else {
|
||||
data, err = c.CoreV1().RESTClient().Get().
|
||||
Context(ctx).
|
||||
Prefix("proxy").
|
||||
Resource("nodes").
|
||||
Name(fmt.Sprintf("%v:%v", nodeName, ports.KubeletPort)).
|
||||
Suffix("stats/summary").
|
||||
Do().Raw()
|
||||
}
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
@ -406,14 +390,7 @@ func getOneTimeResourceUsageOnNode(
|
|||
}
|
||||
|
||||
func getNodeStatsSummary(c clientset.Interface, nodeName string) (*stats.Summary, error) {
|
||||
subResourceProxyAvailable, err := ServerVersionGTE(SubResourceServiceAndNodeProxyVersion, c.Discovery())
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
var data []byte
|
||||
if subResourceProxyAvailable {
|
||||
data, err = c.CoreV1().RESTClient().Get().
|
||||
data, err := c.CoreV1().RESTClient().Get().
|
||||
Resource("nodes").
|
||||
SubResource("proxy").
|
||||
Name(fmt.Sprintf("%v:%v", nodeName, ports.KubeletPort)).
|
||||
|
@ -421,15 +398,6 @@ func getNodeStatsSummary(c clientset.Interface, nodeName string) (*stats.Summary
|
|||
SetHeader("Content-Type", "application/json").
|
||||
Do().Raw()
|
||||
|
||||
} else {
|
||||
data, err = c.CoreV1().RESTClient().Get().
|
||||
Prefix("proxy").
|
||||
Resource("nodes").
|
||||
Name(fmt.Sprintf("%v:%v", nodeName, ports.KubeletPort)).
|
||||
Suffix("stats/summary").
|
||||
SetHeader("Content-Type", "application/json").
|
||||
Do().Raw()
|
||||
}
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
|
|
@ -443,11 +443,6 @@ func getSchedulingLatency(c clientset.Interface) (*SchedulingLatency, error) {
|
|||
nodes, err := c.CoreV1().Nodes().List(metav1.ListOptions{})
|
||||
ExpectNoError(err)
|
||||
|
||||
subResourceProxyAvailable, err := ServerVersionGTE(SubResourcePodProxyVersion, c.Discovery())
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
var data string
|
||||
var masterRegistered = false
|
||||
for _, node := range nodes.Items {
|
||||
|
@ -460,7 +455,6 @@ func getSchedulingLatency(c clientset.Interface) (*SchedulingLatency, error) {
|
|||
defer cancel()
|
||||
|
||||
var rawData []byte
|
||||
if subResourceProxyAvailable {
|
||||
rawData, err = c.CoreV1().RESTClient().Get().
|
||||
Context(ctx).
|
||||
Namespace(metav1.NamespaceSystem).
|
||||
|
@ -469,16 +463,6 @@ func getSchedulingLatency(c clientset.Interface) (*SchedulingLatency, error) {
|
|||
SubResource("proxy").
|
||||
Suffix("metrics").
|
||||
Do().Raw()
|
||||
} else {
|
||||
rawData, err = c.CoreV1().RESTClient().Get().
|
||||
Context(ctx).
|
||||
Prefix("proxy").
|
||||
Namespace(metav1.NamespaceSystem).
|
||||
SubResource("pods").
|
||||
Name(fmt.Sprintf("kube-scheduler-%v:%v", TestContext.CloudConfig.MasterName, ports.SchedulerPort)).
|
||||
Suffix("metrics").
|
||||
Do().Raw()
|
||||
}
|
||||
|
||||
ExpectNoError(err)
|
||||
data = string(rawData)
|
||||
|
|
|
@ -244,25 +244,9 @@ func GetPauseImageNameForHostArch() string {
|
|||
return currentPodInfraContainerImageName + "-" + goruntime.GOARCH + ":" + currentPodInfraContainerImageVersion
|
||||
}
|
||||
|
||||
// SubResource proxy should have been functional in v1.0.0, but SubResource
|
||||
// proxy via tunneling is known to be broken in v1.0. See
|
||||
// https://github.com/kubernetes/kubernetes/pull/15224#issuecomment-146769463
|
||||
//
|
||||
// TODO(ihmccreery): remove once we don't care about v1.0 anymore, (tentatively
|
||||
// in v1.3).
|
||||
var SubResourcePodProxyVersion = utilversion.MustParseSemantic("v1.1.0")
|
||||
var SubResourceServiceAndNodeProxyVersion = utilversion.MustParseSemantic("v1.2.0")
|
||||
|
||||
func GetServicesProxyRequest(c clientset.Interface, request *restclient.Request) (*restclient.Request, error) {
|
||||
subResourceProxyAvailable, err := ServerVersionGTE(SubResourceServiceAndNodeProxyVersion, c.Discovery())
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
if subResourceProxyAvailable {
|
||||
return request.Resource("services").SubResource("proxy"), nil
|
||||
}
|
||||
return request.Prefix("proxy").Resource("services"), nil
|
||||
}
|
||||
|
||||
// unique identifier of the e2e run
|
||||
var RunId = uuid.NewUUID()
|
||||
|
@ -1672,17 +1656,11 @@ func (r podProxyResponseChecker) CheckAllResponses() (done bool, err error) {
|
|||
if !isElementOf(pod.UID, currentPods) {
|
||||
return false, fmt.Errorf("pod with UID %s is no longer a member of the replica set. Must have been restarted for some reason. Current replica set: %v", pod.UID, currentPods)
|
||||
}
|
||||
subResourceProxyAvailable, err := ServerVersionGTE(SubResourcePodProxyVersion, r.c.Discovery())
|
||||
if err != nil {
|
||||
return false, err
|
||||
}
|
||||
|
||||
ctx, cancel := context.WithTimeout(context.Background(), SingleCallTimeout)
|
||||
defer cancel()
|
||||
|
||||
var body []byte
|
||||
if subResourceProxyAvailable {
|
||||
body, err = r.c.CoreV1().RESTClient().Get().
|
||||
body, err := r.c.CoreV1().RESTClient().Get().
|
||||
Context(ctx).
|
||||
Namespace(r.ns).
|
||||
Resource("pods").
|
||||
|
@ -1690,16 +1668,7 @@ func (r podProxyResponseChecker) CheckAllResponses() (done bool, err error) {
|
|||
Name(string(pod.Name)).
|
||||
Do().
|
||||
Raw()
|
||||
} else {
|
||||
body, err = r.c.CoreV1().RESTClient().Get().
|
||||
Context(ctx).
|
||||
Prefix("proxy").
|
||||
Namespace(r.ns).
|
||||
Resource("pods").
|
||||
Name(string(pod.Name)).
|
||||
Do().
|
||||
Raw()
|
||||
}
|
||||
|
||||
if err != nil {
|
||||
if ctx.Err() != nil {
|
||||
// We may encounter errors here because of a race between the pod readiness and apiserver
|
||||
|
@ -4404,14 +4373,9 @@ const proxyTimeout = 2 * time.Minute
|
|||
func NodeProxyRequest(c clientset.Interface, node, endpoint string) (restclient.Result, error) {
|
||||
// proxy tends to hang in some cases when Node is not ready. Add an artificial timeout for this call.
|
||||
// This will leak a goroutine if proxy hangs. #22165
|
||||
subResourceProxyAvailable, err := ServerVersionGTE(SubResourceServiceAndNodeProxyVersion, c.Discovery())
|
||||
if err != nil {
|
||||
return restclient.Result{}, err
|
||||
}
|
||||
var result restclient.Result
|
||||
finished := make(chan struct{})
|
||||
go func() {
|
||||
if subResourceProxyAvailable {
|
||||
result = c.CoreV1().RESTClient().Get().
|
||||
Resource("nodes").
|
||||
SubResource("proxy").
|
||||
|
@ -4419,14 +4383,6 @@ func NodeProxyRequest(c clientset.Interface, node, endpoint string) (restclient.
|
|||
Suffix(endpoint).
|
||||
Do()
|
||||
|
||||
} else {
|
||||
result = c.CoreV1().RESTClient().Get().
|
||||
Prefix("proxy").
|
||||
Resource("nodes").
|
||||
Name(fmt.Sprintf("%v:%v", node, ports.KubeletPort)).
|
||||
Suffix(endpoint).
|
||||
Do()
|
||||
}
|
||||
finished <- struct{}{}
|
||||
}()
|
||||
select {
|
||||
|
|
|
@ -66,17 +66,10 @@ var (
|
|||
|
||||
// Query sends a command to the server and returns the Response
|
||||
func Query(c clientset.Interface, query string) (*influxdb.Response, error) {
|
||||
subResourceProxyAvailable, err := framework.ServerVersionGTE(framework.SubResourceServiceAndNodeProxyVersion, c.Discovery())
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
ctx, cancel := context.WithTimeout(context.Background(), framework.SingleCallTimeout)
|
||||
defer cancel()
|
||||
|
||||
var result []byte
|
||||
if subResourceProxyAvailable {
|
||||
result, err = c.CoreV1().RESTClient().Get().
|
||||
result, err := c.CoreV1().RESTClient().Get().
|
||||
Context(ctx).
|
||||
Namespace("kube-system").
|
||||
Resource("services").
|
||||
|
@ -88,20 +81,6 @@ func Query(c clientset.Interface, query string) (*influxdb.Response, error) {
|
|||
Param("epoch", "s").
|
||||
Do().
|
||||
Raw()
|
||||
} else {
|
||||
result, err = c.CoreV1().RESTClient().Get().
|
||||
Context(ctx).
|
||||
Prefix("proxy").
|
||||
Namespace("kube-system").
|
||||
Resource("services").
|
||||
Name(influxdbService+":api").
|
||||
Suffix("query").
|
||||
Param("q", query).
|
||||
Param("db", influxdbDatabaseName).
|
||||
Param("epoch", "s").
|
||||
Do().
|
||||
Raw()
|
||||
}
|
||||
|
||||
if err != nil {
|
||||
if ctx.Err() != nil {
|
||||
|
|
|
@ -1990,17 +1990,11 @@ func getUDData(jpgExpected string, ns string) func(clientset.Interface, string)
|
|||
// getUDData validates data.json in the update-demo (returns nil if data is ok).
|
||||
return func(c clientset.Interface, podID string) error {
|
||||
framework.Logf("validating pod %s", podID)
|
||||
subResourceProxyAvailable, err := framework.ServerVersionGTE(framework.SubResourcePodProxyVersion, c.Discovery())
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
ctx, cancel := context.WithTimeout(context.Background(), framework.SingleCallTimeout)
|
||||
defer cancel()
|
||||
|
||||
var body []byte
|
||||
if subResourceProxyAvailable {
|
||||
body, err = c.CoreV1().RESTClient().Get().
|
||||
body, err := c.CoreV1().RESTClient().Get().
|
||||
Namespace(ns).
|
||||
Resource("pods").
|
||||
SubResource("proxy").
|
||||
|
@ -2008,16 +2002,7 @@ func getUDData(jpgExpected string, ns string) func(clientset.Interface, string)
|
|||
Suffix("data.json").
|
||||
Do().
|
||||
Raw()
|
||||
} else {
|
||||
body, err = c.CoreV1().RESTClient().Get().
|
||||
Prefix("proxy").
|
||||
Namespace(ns).
|
||||
Resource("pods").
|
||||
Name(podID).
|
||||
Suffix("data.json").
|
||||
Do().
|
||||
Raw()
|
||||
}
|
||||
|
||||
if err != nil {
|
||||
if ctx.Err() != nil {
|
||||
framework.Failf("Failed to retrieve data from container: %v", err)
|
||||
|
|
|
@ -169,18 +169,12 @@ func assertFilesContain(fileNames []string, fileDir string, pod *v1.Pod, client
|
|||
|
||||
framework.ExpectNoError(wait.Poll(time.Second*10, time.Second*600, func() (bool, error) {
|
||||
failed = []string{}
|
||||
subResourceProxyAvailable, err := framework.ServerVersionGTE(framework.SubResourcePodProxyVersion, client.Discovery())
|
||||
if err != nil {
|
||||
return false, err
|
||||
}
|
||||
|
||||
ctx, cancel := context.WithTimeout(context.Background(), framework.SingleCallTimeout)
|
||||
defer cancel()
|
||||
|
||||
var contents []byte
|
||||
for _, fileName := range fileNames {
|
||||
if subResourceProxyAvailable {
|
||||
contents, err = client.CoreV1().RESTClient().Get().
|
||||
contents, err := client.CoreV1().RESTClient().Get().
|
||||
Context(ctx).
|
||||
Namespace(pod.Namespace).
|
||||
Resource("pods").
|
||||
|
@ -188,16 +182,7 @@ func assertFilesContain(fileNames []string, fileDir string, pod *v1.Pod, client
|
|||
Name(pod.Name).
|
||||
Suffix(fileDir, fileName).
|
||||
Do().Raw()
|
||||
} else {
|
||||
contents, err = client.CoreV1().RESTClient().Get().
|
||||
Context(ctx).
|
||||
Prefix("proxy").
|
||||
Resource("pods").
|
||||
Namespace(pod.Namespace).
|
||||
Name(pod.Name).
|
||||
Suffix(fileDir, fileName).
|
||||
Do().Raw()
|
||||
}
|
||||
|
||||
if err != nil {
|
||||
if ctx.Err() != nil {
|
||||
framework.Failf("Unable to read %s from pod %s: %v", fileName, pod.Name, err)
|
||||
|
|
|
@ -121,16 +121,11 @@ func testPreStop(c clientset.Interface, ns string) {
|
|||
|
||||
// Validate that the server received the web poke.
|
||||
err = wait.Poll(time.Second*5, time.Second*60, func() (bool, error) {
|
||||
subResourceProxyAvailable, err := framework.ServerVersionGTE(framework.SubResourcePodProxyVersion, c.Discovery())
|
||||
if err != nil {
|
||||
return false, err
|
||||
}
|
||||
|
||||
ctx, cancel := context.WithTimeout(context.Background(), framework.SingleCallTimeout)
|
||||
defer cancel()
|
||||
|
||||
var body []byte
|
||||
if subResourceProxyAvailable {
|
||||
body, err = c.CoreV1().RESTClient().Get().
|
||||
Context(ctx).
|
||||
Namespace(ns).
|
||||
|
@ -139,16 +134,7 @@ func testPreStop(c clientset.Interface, ns string) {
|
|||
Name(podDescr.Name).
|
||||
Suffix("read").
|
||||
DoRaw()
|
||||
} else {
|
||||
body, err = c.CoreV1().RESTClient().Get().
|
||||
Context(ctx).
|
||||
Prefix("proxy").
|
||||
Namespace(ns).
|
||||
Resource("pods").
|
||||
Name(podDescr.Name).
|
||||
Suffix("read").
|
||||
DoRaw()
|
||||
}
|
||||
|
||||
if err != nil {
|
||||
if ctx.Err() != nil {
|
||||
framework.Failf("Error validating prestop: %v", err)
|
||||
|
|
Loading…
Reference in New Issue