mirror of https://github.com/k3s-io/k3s
Merge pull request #38763 from aaron12134/spell-corrective-doctrine
Automatic merge from submit-queue make spellcheck for test/* **What this PR does / why we need it**: Increase code readability **Which issue this PR fixes** *(optional, in `fixes #<issue number>(, fixes #<issue_number>, ...)` format, will close that issue when PR gets merged)*: fixes # **Special notes for your reviewer**: only a slight effort **Release note**: ```release-note ```pull/6/head
commit
6c50e73d4d
|
@ -43,7 +43,7 @@ kube::util::gen-docs "${KUBE_TEMP}"
|
||||||
# remove all of the old docs
|
# remove all of the old docs
|
||||||
kube::util::remove-gen-docs
|
kube::util::remove-gen-docs
|
||||||
|
|
||||||
# Copy fresh docs into the repo.
|
# copy fresh docs into the repo.
|
||||||
# the shopt is so that we get .generated_docs from the glob.
|
# the shopt is so that we get .generated_docs from the glob.
|
||||||
shopt -s dotglob
|
shopt -s dotglob
|
||||||
cp -af "${KUBE_TEMP}"/* "${KUBE_ROOT}"
|
cp -af "${KUBE_TEMP}"/* "${KUBE_ROOT}"
|
||||||
|
|
|
@ -91,7 +91,7 @@ function start_discovery {
|
||||||
${kubectl} config set-cluster local-up-cluster --kubeconfig="${CERT_DIR}/admin-discovery.kubeconfig" --certificate-authority="${CERT_DIR}/discovery-ca.crt" --embed-certs --server="https://${API_HOST_IP}:${DISCOVERY_SECURE_PORT}"
|
${kubectl} config set-cluster local-up-cluster --kubeconfig="${CERT_DIR}/admin-discovery.kubeconfig" --certificate-authority="${CERT_DIR}/discovery-ca.crt" --embed-certs --server="https://${API_HOST_IP}:${DISCOVERY_SECURE_PORT}"
|
||||||
|
|
||||||
# Wait for kubernetes-discovery to come up before launching the rest of the components.
|
# Wait for kubernetes-discovery to come up before launching the rest of the components.
|
||||||
# this should work since we're creating a node port service
|
# This should work since we're creating a node port service.
|
||||||
echo "Waiting for kubernetes-discovery to come up: https://${API_HOST_IP}:${DISCOVERY_SECURE_PORT}/version"
|
echo "Waiting for kubernetes-discovery to come up: https://${API_HOST_IP}:${DISCOVERY_SECURE_PORT}/version"
|
||||||
kube::util::wait_for_url "https://${API_HOST_IP}:${DISCOVERY_SECURE_PORT}/version" "kubernetes-discovery: " 1 60 || exit 1
|
kube::util::wait_for_url "https://${API_HOST_IP}:${DISCOVERY_SECURE_PORT}/version" "kubernetes-discovery: " 1 60 || exit 1
|
||||||
|
|
||||||
|
|
|
@ -49,4 +49,4 @@ PATH="${KUBE_ROOT}/_output/bin:${PATH}" \
|
||||||
"${gotoprotobuf}" \
|
"${gotoprotobuf}" \
|
||||||
--proto-import="${KUBE_ROOT}/vendor" \
|
--proto-import="${KUBE_ROOT}/vendor" \
|
||||||
--proto-import="${KUBE_ROOT}/third_party/protobuf" \
|
--proto-import="${KUBE_ROOT}/third_party/protobuf" \
|
||||||
$@
|
"$@"
|
||||||
|
|
|
@ -25,7 +25,7 @@ kube::golang::setup_env
|
||||||
|
|
||||||
make -C "${KUBE_ROOT}" WHAT=cmd/hyperkube
|
make -C "${KUBE_ROOT}" WHAT=cmd/hyperkube
|
||||||
|
|
||||||
# add other BADSYMBOLS here.
|
# Add other BADSYMBOLS here.
|
||||||
BADSYMBOLS=(
|
BADSYMBOLS=(
|
||||||
"httptest"
|
"httptest"
|
||||||
"testify"
|
"testify"
|
||||||
|
|
|
@ -428,7 +428,7 @@ func TestOrphaning(t *testing.T) {
|
||||||
t.Fatalf("Failed to create replication controller: %v", err)
|
t.Fatalf("Failed to create replication controller: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
// these pods should be ophaned.
|
// these pods should be orphaned.
|
||||||
var podUIDs []types.UID
|
var podUIDs []types.UID
|
||||||
podsNum := 3
|
podsNum := 3
|
||||||
for i := 0; i < podsNum; i++ {
|
for i := 0; i < podsNum; i++ {
|
||||||
|
@ -478,7 +478,7 @@ func TestOrphaning(t *testing.T) {
|
||||||
}
|
}
|
||||||
for _, pod := range pods.Items {
|
for _, pod := range pods.Items {
|
||||||
if len(pod.ObjectMeta.OwnerReferences) != 0 {
|
if len(pod.ObjectMeta.OwnerReferences) != 0 {
|
||||||
t.Errorf("pod %s still has non-empty OwnerRefereces: %v", pod.ObjectMeta.Name, pod.ObjectMeta.OwnerReferences)
|
t.Errorf("pod %s still has non-empty OwnerReferences: %v", pod.ObjectMeta.Name, pod.ObjectMeta.OwnerReferences)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -174,7 +174,7 @@ func DoTestUnschedulableNodes(t *testing.T, cs clientset.Interface, ns *v1.Names
|
||||||
err = waitForReflection(t, s, nodeKey, func(node interface{}) bool {
|
err = waitForReflection(t, s, nodeKey, func(node interface{}) bool {
|
||||||
// An unschedulable node should still be present in the store
|
// An unschedulable node should still be present in the store
|
||||||
// Nodes that are unschedulable or that are not ready or
|
// Nodes that are unschedulable or that are not ready or
|
||||||
// have their disk full (Node.Spec.Conditions) are exluded
|
// have their disk full (Node.Spec.Conditions) are excluded
|
||||||
// based on NodeConditionPredicate, a separate check
|
// based on NodeConditionPredicate, a separate check
|
||||||
return node != nil && node.(*v1.Node).Spec.Unschedulable == true
|
return node != nil && node.(*v1.Node).Spec.Unschedulable == true
|
||||||
})
|
})
|
||||||
|
@ -319,7 +319,7 @@ func TestMultiScheduler(t *testing.T) {
|
||||||
8. create 2 pods: testPodNoAnnotation2 and testPodWithAnnotationFitsDefault2
|
8. create 2 pods: testPodNoAnnotation2 and testPodWithAnnotationFitsDefault2
|
||||||
- note: these two pods belong to default scheduler which no longer exists
|
- note: these two pods belong to default scheduler which no longer exists
|
||||||
9. **check point-3**:
|
9. **check point-3**:
|
||||||
- testPodNoAnnotation2 and testPodWithAnnotationFitsDefault2 shoule NOT be scheduled
|
- testPodNoAnnotation2 and testPodWithAnnotationFitsDefault2 should NOT be scheduled
|
||||||
*/
|
*/
|
||||||
// 1. create and start default-scheduler
|
// 1. create and start default-scheduler
|
||||||
clientSet := clientset.NewForConfigOrDie(&restclient.Config{Host: s.URL, ContentConfig: restclient.ContentConfig{GroupVersion: ®istered.GroupOrDie(v1.GroupName).GroupVersion}})
|
clientSet := clientset.NewForConfigOrDie(&restclient.Config{Host: s.URL, ContentConfig: restclient.ContentConfig{GroupVersion: ®istered.GroupOrDie(v1.GroupName).GroupVersion}})
|
||||||
|
@ -453,7 +453,7 @@ func TestMultiScheduler(t *testing.T) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// 9. **check point-3**:
|
// 9. **check point-3**:
|
||||||
// - testPodNoAnnotation2 and testPodWithAnnotationFitsDefault2 shoule NOT be scheduled
|
// - testPodNoAnnotation2 and testPodWithAnnotationFitsDefault2 should NOT be scheduled
|
||||||
err = wait.Poll(time.Second, time.Second*5, podScheduled(clientSet, testPodNoAnnotation2.Namespace, testPodNoAnnotation2.Name))
|
err = wait.Poll(time.Second, time.Second*5, podScheduled(clientSet, testPodNoAnnotation2.Namespace, testPodNoAnnotation2.Name))
|
||||||
if err == nil {
|
if err == nil {
|
||||||
t.Errorf("Test MultiScheduler: %s Pod got scheduled, %v", testPodNoAnnotation2.Name, err)
|
t.Errorf("Test MultiScheduler: %s Pod got scheduled, %v", testPodNoAnnotation2.Name, err)
|
||||||
|
@ -587,7 +587,7 @@ func TestAllocatable(t *testing.T) {
|
||||||
// 7. Test: this test pod should not be scheduled since it request more than Allocatable
|
// 7. Test: this test pod should not be scheduled since it request more than Allocatable
|
||||||
err = wait.Poll(time.Second, time.Second*5, podScheduled(clientSet, testAllocPod2.Namespace, testAllocPod2.Name))
|
err = wait.Poll(time.Second, time.Second*5, podScheduled(clientSet, testAllocPod2.Namespace, testAllocPod2.Name))
|
||||||
if err == nil {
|
if err == nil {
|
||||||
t.Errorf("Test allocatable awareness: %s Pod got scheduled unexpectly, %v", testAllocPod2.Name, err)
|
t.Errorf("Test allocatable awareness: %s Pod got scheduled unexpectedly, %v", testAllocPod2.Name, err)
|
||||||
} else {
|
} else {
|
||||||
t.Logf("Test allocatable awareness: %s Pod not scheduled as expected", testAllocPod2.Name)
|
t.Logf("Test allocatable awareness: %s Pod not scheduled as expected", testAllocPod2.Name)
|
||||||
}
|
}
|
||||||
|
|
|
@ -184,7 +184,7 @@ func defaultSchedulerBenchmarkConfig(numNodes, numPods int) *testConfig {
|
||||||
// This is used to learn the scheduling throughput on various
|
// This is used to learn the scheduling throughput on various
|
||||||
// sizes of cluster and changes as more and more pods are scheduled.
|
// sizes of cluster and changes as more and more pods are scheduled.
|
||||||
// It won't stop until all pods are scheduled.
|
// It won't stop until all pods are scheduled.
|
||||||
// It retruns the minimum of throughput over whole run.
|
// It returns the minimum of throughput over whole run.
|
||||||
func schedulePods(config *testConfig) int32 {
|
func schedulePods(config *testConfig) int32 {
|
||||||
defer config.destroyFunc()
|
defer config.destroyFunc()
|
||||||
if err := config.nodePreparer.PrepareNodes(); err != nil {
|
if err := config.nodePreparer.PrepareNodes(); err != nil {
|
||||||
|
|
|
@ -78,7 +78,7 @@ type FooList struct {
|
||||||
Items []Foo `json:"items"`
|
Items []Foo `json:"items"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// installThirdParty installs a third party resoure and returns a defer func
|
// installThirdParty installs a third party resource and returns a defer func
|
||||||
func installThirdParty(t *testing.T, client clientset.Interface, clientConfig *restclient.Config, tpr *extensions.ThirdPartyResource, group, version, resource string) func() {
|
func installThirdParty(t *testing.T, client clientset.Interface, clientConfig *restclient.Config, tpr *extensions.ThirdPartyResource, group, version, resource string) func() {
|
||||||
var err error
|
var err error
|
||||||
_, err = client.Extensions().ThirdPartyResources().Create(tpr)
|
_, err = client.Extensions().ThirdPartyResources().Create(tpr)
|
||||||
|
|
|
@ -372,7 +372,7 @@ func TestPersistentVolumeClaimLabelSelectorMatchExpressions(t *testing.T) {
|
||||||
_, s := framework.RunAMaster(nil)
|
_, s := framework.RunAMaster(nil)
|
||||||
defer s.Close()
|
defer s.Close()
|
||||||
|
|
||||||
ns := framework.CreateTestingNamespace("pvc-match-expresssions", s, t)
|
ns := framework.CreateTestingNamespace("pvc-match-expressions", s, t)
|
||||||
defer framework.DeleteTestingNamespace(ns, s, t)
|
defer framework.DeleteTestingNamespace(ns, s, t)
|
||||||
|
|
||||||
testClient, controller, watchPV, watchPVC := createClients(ns, t, s, defaultSyncPeriod)
|
testClient, controller, watchPV, watchPVC := createClients(ns, t, s, defaultSyncPeriod)
|
||||||
|
@ -778,7 +778,7 @@ func TestPersistentVolumeControllerStartup(t *testing.T) {
|
||||||
}
|
}
|
||||||
pvs[i] = newPV
|
pvs[i] = newPV
|
||||||
// Drain watchPV with all events generated by the PV until it's bound
|
// Drain watchPV with all events generated by the PV until it's bound
|
||||||
// We don't want to catch "PV craated with Status.Phase == Pending"
|
// We don't want to catch "PV created with Status.Phase == Pending"
|
||||||
// later in this test.
|
// later in this test.
|
||||||
waitForAnyPersistentVolumePhase(watchPV, v1.VolumeBound)
|
waitForAnyPersistentVolumePhase(watchPV, v1.VolumeBound)
|
||||||
}
|
}
|
||||||
|
@ -1135,7 +1135,7 @@ func createClients(ns *v1.Namespace, t *testing.T, s *httptest.Server, syncPerio
|
||||||
}
|
}
|
||||||
watchPVC, err := testClient.PersistentVolumeClaims(ns.Name).Watch(v1.ListOptions{})
|
watchPVC, err := testClient.PersistentVolumeClaims(ns.Name).Watch(v1.ListOptions{})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("Failed to watch PersistentVolumeClaimss: %v", err)
|
t.Fatalf("Failed to watch PersistentVolumeClaims: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
return testClient, ctrl, watchPV, watchPVC
|
return testClient, ctrl, watchPV, watchPVC
|
||||||
|
|
|
@ -40,5 +40,5 @@ else
|
||||||
fi
|
fi
|
||||||
|
|
||||||
go run ./hack/e2e.go -v --check_version_skew=false --test --test_args="--e2e-verify-service-account=false --dump-logs-on-failure=false ${ARGS}"
|
go run ./hack/e2e.go -v --check_version_skew=false --test --test_args="--e2e-verify-service-account=false --dump-logs-on-failure=false ${ARGS}"
|
||||||
# Just make local testing easier...
|
# Just make local test easier...
|
||||||
# ${KUBE_ROOT}/hack/ginkgo-e2e.sh "--e2e-verify-service-account=false" "--dump-logs-on-failure=false" $ARGS
|
# ${KUBE_ROOT}/hack/ginkgo-e2e.sh "--e2e-verify-service-account=false" "--dump-logs-on-failure=false" $ARGS
|
||||||
|
|
Loading…
Reference in New Issue