Commit Graph

50814 Commits (58dce96947abe36b1ec2142313eb06d69b1eb03a)

Author SHA1 Message Date
Kubernetes Submit Queue ede78d9ee7 Merge pull request #47513 from gmarek/subnet
Automatic merge from submit-queue

Make big clusters work again after introduction of subnets

This PR does two things: 
  - make IP aliases automatically pick Node IP Range based on number of Nodes,
  - fix logic for starting clusters >4095 Nodes that was broken by introduction of subnets,

cc @wojtek-t @shyamjvs 

```release-note
Setting env var ENABLE_BIG_CLUSTER_SUBNETS=true will allow kube-up.sh to start clusters bigger that 4095 Nodes on GCE.
```

Ref https://github.com/kubernetes/kubernetes/issues/47344
2017-06-27 08:52:50 -07:00
Kubernetes Submit Queue d65b87a00d Merge pull request #47847 from chuckbutler/cluster-juju-approvers
Automatic merge from submit-queue

Insert Cynerva and Kjackal to approvers list

**What this PR does / why we need it**:
Per the membership reviews, we're looking to promote Konstantinos and
George to approvers to help distribute the review/bug load for the `cluster/juju` code
tree.

**Which issue this PR fixes** *(optional, in `fixes #<issue number>(, fixes #<issue_number>, ...)` format, will close that issue when PR gets merged)*: 

**Special notes for your reviewer**:
cc @marcoceppi and @tvansteenburgh 

**Release note**:

```release-note
NONE
```
2017-06-27 08:36:30 -07:00
Kubernetes Submit Queue 0d84a4fedb Merge pull request #47105 from feiskyer/cleanup
Automatic merge from submit-queue (batch tested with PRs 47038, 47105)

kuberuntime: cleanup TestGenerateContainerConfig

Followup of #47009, cleanup TestGenerateContainerConfig and remove unused expectedConfig.

/assign @yujuhong
2017-06-27 08:29:29 -07:00
Kubernetes Submit Queue 6acfb0f868 Merge pull request #47038 from yastij/DefaultExternalHost-for-registredCP
Automatic merge from submit-queue (batch tested with PRs 47038, 47105)

extending DefaultExternalHost for any registred cloud provider

**What this PR does / why we need it**: this PR enables DefaultExternalHost to work with any registered cloud provider.

**Which issue this PR fixes** : fixes #46567 

**Special notes for your reviewer**:

**Release note**:

```release-note
When determining the default external host of the kube apiserver, any configured cloud provider is now consulted
```
2017-06-27 08:29:25 -07:00
Mik Vyatskov b6a0e442ce Fix fluentd-gcp configuration to facilitate JSON parsing 2017-06-27 16:16:00 +02:00
Dan Winship 127eb53d4d Fix the names of some iptables tests
The delete tests were copy+pasted from the create tests, and the names
not fully updated to match.
2017-06-27 09:26:10 -04:00
Dan Winship 3d48091ce8 Fix ebtables_test.go to actually get run, and to pass 2017-06-27 09:26:10 -04:00
Kubernetes Submit Queue 6d1da16456 Merge pull request #47731 from jsravn/use-endpoints-cache-for-endpoint-controller
Automatic merge from submit-queue

Use endpoints informer for the endpoint controller

This substantially reduces the number of API calls made by the endpoint
controller. Currently the controller makes an API call per endpoint for
each service that is synced. When the 30s resync is triggered, this
results in an API call for every single endpoint in the cluster. This
quickly exceeds the default qps/burst limit of 20/30 even in small
clusters, leading to delays in endpoint updates.

This change modifies the controller to use the endpoint informer cache
for all endpoint GETs. This means we only make API calls for changes in
endpoints. As a result, qps only depends on the pod activity in the
cluster, rather than the number of services.



**What this PR does / why we need it**:

Address endpoint update delays as described in https://github.com/kubernetes/kubernetes/issues/47597.

**Which issue this PR fixes** *(optional, in `fixes #<issue number>(, fixes #<issue_number>, ...)` format, will close that issue when PR gets merged)*: fixes #

https://github.com/kubernetes/kubernetes/issues/47597

**Special notes for your reviewer**:

**Release note**:

```release-note
```
2017-06-27 05:20:12 -07:00
Marcin Wielgus 53ac028473 Merge pull request #48132 from MaciekPytel/cluster_autoscaler_node_balance_2
Fix typo in cluster-autoscaler config
2017-06-27 13:52:22 +02:00
Maciej Pytel 04f7a96340 Fix typo in cluster-autoscaler config 2017-06-27 13:49:19 +02:00
Kubernetes Submit Queue 49c151e433 Merge pull request #48049 from luxas/kubeadm_rolling_proxy
Automatic merge from submit-queue

kubeadm: Make kube-proxy RollingUpgradeable

**What this PR does / why we need it**:
Sets the right updateStrategy for kube-proxy.

**Which issue this PR fixes** *(optional, in `fixes #<issue number>(, fixes #<issue_number>, ...)` format, will close that issue when PR gets merged)*: fixes #

Fixes: https://github.com/kubernetes/kubeadm/issues/319

**Special notes for your reviewer**:

**Release note**:

```release-note
NONE
```
@pipejakob @timothysc @kubernetes/sig-cluster-lifecycle-pr-reviews
2017-06-27 04:29:39 -07:00
Kubernetes Submit Queue 0e509854c7 Merge pull request #48072 from MaciekPytel/ca_scale_to_0_e2e
Automatic merge from submit-queue

Add e2e for cluster-autoscaler scale-up from 0

Ref: https://github.com/kubernetes/autoscaler/issues/43
2017-06-27 03:35:34 -07:00
Marcin Wielgus 7174455168 Merge pull request #48129 from MaciekPytel/cluster_autoscaler_node_balance
Set cluster-autoscaler node balancing flag
2017-06-27 12:24:31 +02:00
Maciej Pytel b11175f73f Set cluster-autoscaler node balancing flag 2017-06-27 12:10:33 +02:00
Konstantinos Tsakalozos 0b01cd743b Improve security of Juju deployed clusters 2017-06-27 12:19:21 +03:00
James Ravn 9fc5a547ae Use endpoints informer for the endpoint controller
This substantially reduces the number of API calls made by the endpoint
controller. Currently the controller makes an API call per endpoint for
each service that is synced. When the 30s resync is triggered, this
results in an API call for every single endpoint in the cluster. This
quickly exceeds the default qps/burst limit of 20/30 even in small
clusters, leading to delays in endpoint updates.

This change modifies the controller to use the endpoint informer cache
for all endpoint GETs. This means we only make API calls for changes in
endpoints. As a result, qps only depends on the pod activity in the
cluster, rather than the number of services.
2017-06-27 10:17:09 +01:00
xiangpengzhao 0f65b218a0
Add Pod UID (metadata.uid) to downward API env var 2017-06-27 16:54:35 +08:00
Kubernetes Submit Queue f39c1712c2 Merge pull request #47052 from gyliu513/godoc
Automatic merge from submit-queue

godoc update for scheduler predicates.

**What this PR does / why we need it**:
This is a follow up PR for https://github.com/kubernetes/kubernetes/pull/46621 

/cc @timothysc 

**Which issue this PR fixes** *(optional, in `fixes #<issue number>(, fixes #<issue_number>, ...)` format, will close that issue when PR gets merged)*: fixes #

**Special notes for your reviewer**:

**Release note**:

```release-note
none
```
2017-06-26 23:59:20 -07:00
Kubernetes Submit Queue eead1729d2 Merge pull request #48034 from smarterclayton/more_printers_2
Automatic merge from submit-queue (batch tested with PRs 47484, 47904, 48034)

Add a DaemonSet table printer

Only the last commit is new
2017-06-26 22:22:00 -07:00
Kubernetes Submit Queue 353633b877 Merge pull request #47904 from deads2k/log-01-supress
Automatic merge from submit-queue (batch tested with PRs 47484, 47904, 48034)

prioritize messages for long steps

This pull prioritizes the trace messages, so steps that are unusually large come out at the info level and all details come out a v(4) level.
2017-06-26 22:21:58 -07:00
Kubernetes Submit Queue cf15b17319 Merge pull request #47484 from allencloud/fix-nit-in-kubelet-server
Automatic merge from submit-queue (batch tested with PRs 47484, 47904, 48034)

fix nits in kubelet server

Signed-off-by: allencloud <allen.sun@daocloud.io>

**What this PR does / why we need it**:
fix nits in kubelet server

**Which issue this PR fixes** *(optional, in `fixes #<issue number>(, fixes #<issue_number>, ...)` format, will close that issue when PR gets merged)*: fixes #
NONE

**Special notes for your reviewer**:
NONE

**Release note**:

```release-note
NONE
```
2017-06-26 22:21:57 -07:00
xiangpengzhao 415f13ea4e Modify e2e tests for service type update. 2017-06-27 12:30:40 +08:00
xiangpengzhao 4e9abca6d0 Allocate clusterIP when change service type from ExternalName to ClusterIP 2017-06-27 12:30:22 +08:00
Kubernetes Submit Queue 0dad2d0803 Merge pull request #47983 from yguo0905/memcg
Automatic merge from submit-queue (batch tested with PRs 48092, 47894, 47983)

Enables memcg notification in cluster/node e2e tests

Ref: https://github.com/kubernetes/kubernetes/issues/42676

This PR sets Kubelet flag `--experimental-kernel-memcg-notification=true` when running cluster/node e2e tests on COS and Ubuntu images.

Tested:
```
e2e-node-cos:
I0623 00:09:06.641776    1080 server.go:147] Starting server "kubelet" with command "/usr/bin/systemd-run --unit=kubelet-777178888.service --slice=runtime.slice --remain-after-exit /tmp/node-e2e-20170622T170739/kubelet --kubelet-cgroups=/kubelet.slice --cgroup-root=/ --api-servers http://localhost:8080 --address 0.0.0.0 --port 10250 --read-only-port 10255 --volume-stats-agg-period 10s --allow-privileged true --serialize-image-pulls false --pod-manifest-path /tmp/node-e2e-20170622T170739/pod-manifest571288056 --file-check-frequency 10s --pod-cidr 10.100.0.0/24 --eviction-pressure-transition-period 30s --feature-gates  --eviction-hard memory.available<250Mi,nodefs.available<10%%,nodefs.inodesFree<5%% --eviction-minimum-reclaim nodefs.available=5%%,nodefs.inodesFree=5%% --v 4 --logtostderr --network-plugin=kubenet --cni-bin-dir /tmp/node-e2e-20170622T170739/cni/bin --cni-conf-dir /tmp/node-e2e-20170622T170739/cni/net.d --hostname-override tmp-node-e2e-bfe5799d-cos-stable-59-9460-64-0 --experimental-mounter-path=/tmp/node-e2e-20170622T170739/cluster/gce/gci/mounter/mounter --experimental-kernel-memcg-notification=true"

e2e-node-ubuntu:
I0623 00:03:28.526984    2279 server.go:147] Starting server "kubelet" with command "/usr/bin/systemd-run --unit=kubelet-1407651753.service --slice=runtime.slice --remain-after-exit /tmp/node-e2e-20170622T170203/kubelet --kubelet-cgroups=/kubelet.slice --cgroup-root=/ --api-servers http://localhost:8080 --address 0.0.0.0 --port 10250 --read-only-port 10255 --volume-stats-agg-period 10s --allow-privileged true --serialize-image-pulls false --pod-manifest-path /tmp/node-e2e-20170622T170203/pod-manifest083943734 --file-check-frequency 10s --pod-cidr 10.100.0.0/24 --eviction-pressure-transition-period 30s --feature-gates  --eviction-hard memory.available<250Mi,nodefs.available<10%%,nodefs.inodesFree<5%% --eviction-minimum-reclaim nodefs.available=5%%,nodefs.inodesFree=5%% --v 4 --logtostderr --network-plugin=kubenet --cni-bin-dir /tmp/node-e2e-20170622T170203/cni/bin --cni-conf-dir /tmp/node-e2e-20170622T170203/cni/net.d --hostname-override tmp-node-e2e-e48cdd73-ubuntu-gke-1604-xenial-v20170420-1 --experimental-kernel-memcg-notification=true"

e2e-node-containervm:
I0623 00:14:35.392383    2774 server.go:147] Starting server "kubelet" with command "/tmp/node-e2e-20170622T171318/kubelet --runtime-cgroups=/docker-daemon --kubelet-cgroups=/kubelet --cgroup-root=/ --system-cgroups=/system --api-servers http://localhost:8080 --address 0.0.0.0 --port 10250 --read-only-port 10255 --volume-stats-agg-period 10s --allow-privileged true --serialize-image-pulls false --pod-manifest-path /tmp/node-e2e-20170622T171318/pod-manifest507536807 --file-check-frequency 10s --pod-cidr 10.100.0.0/24 --eviction-pressure-transition-period 30s --feature-gates  --eviction-hard memory.available<250Mi,nodefs.available<10%,nodefs.inodesFree<5% --eviction-minimum-reclaim nodefs.available=5%,nodefs.inodesFree=5% --v 4 --logtostderr --network-plugin=kubenet --cni-bin-dir /tmp/node-e2e-20170622T171318/cni/bin --cni-conf-dir /tmp/node-e2e-20170622T171318/cni/net.d --hostname-override tmp-node-e2e-9e3fdd7c-e2e-node-containervm-v20161208-image"

e2e-cos:
Jun 23 17:54:38 e2e-test-ygg-minion-group-t5r0 kubelet[2005]: I0623 17:54:38.646374    2005 flags.go:52] FLAG: --experimental-kernel-memcg-notification="true"

e2e-ubuntu:
Jun 23 18:25:27 e2e-test-ygg-minion-group-19qp kubelet[1547]: I0623 18:25:27.722253    1547 flags.go:52] FLAG: --experimental-kernel-memcg-notification="true"

e2e-containervm:
I0623 18:55:51.886632    3385 flags.go:52] FLAG: --experimental-kernel-memcg-notification="false"
```

**Release note**:
```
None
```

/sig node
/area node-e2e
/assign @dchen1107 @dashpole
2017-06-26 21:08:10 -07:00
Kubernetes Submit Queue e8701e6746 Merge pull request #47894 from xuxinkun/fixclustercentos20170622
Automatic merge from submit-queue (batch tested with PRs 48092, 47894, 47983)

fix systemd service file for custom args.

`KUBE_SCHEDULER_ARGS` and `KUBELET_ARGS` are used to custom args for scheduler or kubelet by users. 
But if there are more than one params in `KUBELET_ARGS`, for example, if I set  KUBELET_ARGS="--cgroups-per-qos=false --enforce-node-allocatable=", the kubelet will judge the `false --enforce-node-allocatable=` as the value of `cgroups-per-qos`.  Because `${KUBELET_ARGS}` in kubelet.service will expands the variable into one word. And if I take `$KUBELET_ARGS` instead, kubelet will worker perfectly.
For more info, please click [EnvironmentFiles and support for /etc/sysconfig files](http://fedoraproject.org/wiki/Packaging:Systemd#EnvironmentFiles_and_support_for_.2Fetc.2Fsysconfig_files). This bug is reported by @huanxingyouyoutoo. And I make this PR for her to fix it.

**Release note**:

```
NONE
```
2017-06-26 21:08:08 -07:00
Kubernetes Submit Queue eb2c702aaa Merge pull request #48092 from enisoc/upgrade-skip-dep
Automatic merge from submit-queue (batch tested with PRs 48092, 47894, 47983)

Skip Deployment upgrade test on 1.5 and earlier.

The test relies on implementation details and would need a rewrite to work for older clusters.

xref #47685
2017-06-26 21:08:06 -07:00
Kubernetes Submit Queue 95be540bd4 Merge pull request #47242 from k82cn/make_golint_happy
Automatic merge from submit-queue

Updated comments for functions.
2017-06-26 20:02:40 -07:00
Cheng Xing 78d111b75f Ensures node becomes schedulable at the end of tests that delete nodes 2017-06-26 18:41:14 -07:00
Clayton Coleman 2f211b8870
Move DaemonSet to table printer 2017-06-26 20:58:02 -04:00
Kubernetes Submit Queue 9a4a0cb31d Merge pull request #47178 from ktsakalozos/bug/worker-labels
Automatic merge from submit-queue (batch tested with PRs 48012, 47443, 47702, 47178)

Fix setting juju worker labels during deployment

**What this PR does / why we need it**: Allows for setting the labels of juju workers during deployment (eg inside a bundle)

**Which issue this PR fixes** *(optional, in `fixes #<issue number>(, fixes #<issue_number>, ...)` format, will close that issue when PR gets merged)*: fixes #47176

**Special notes for your reviewer**:

**Release note**:

```
Fix bug in setting Juju kubernetes-worker labels in bundle.yaml files. 
```
2017-06-26 17:48:06 -07:00
Kubernetes Submit Queue d31bcad496 Merge pull request #47702 from smarterclayton/avoid_mutable
Automatic merge from submit-queue (batch tested with PRs 48012, 47443, 47702, 47178)

Don't bother with a mutable transformer for identity

The default value transformer can safely be the identity transformer - mutability is not required if the caller doesn't need transformation.
2017-06-26 17:48:03 -07:00
Kubernetes Submit Queue aac42add77 Merge pull request #47443 from p0lyn0mial/use_incluster_cfg_when_creating_ext_informers
Automatic merge from submit-queue (batch tested with PRs 48012, 47443, 47702, 47178)

incluster config will be used when creating external shared informers.

**What this PR does / why we need it**:
Previously the loopback configuration was used to talk to the server.
As a consequence a custom API server was unable to talk to the root API server.
This PR changes the above by using incluster configuration to create shared informers.

**Release note**:

```release-note
NONE
```
2017-06-26 17:48:01 -07:00
Kubernetes Submit Queue 1258a7ae43 Merge pull request #48012 from verult/FlakyDetachE2E
Automatic merge from submit-queue (batch tested with PRs 48012, 47443, 47702, 47178)

Extending timeout waiting for delete node to become ready before the test ends.

**What this PR does / why we need it**: It seems to take longer than 5 minutes for the node to recover. Changing the timeout to 10 minutes.
This is an extension of PR #46746

**Which issue this PR fixes** *(optional, in `fixes #<issue number>(, fixes #<issue_number>, ...)` format, will close that issue when PR gets merged)*: fixes #48008 

/release-note-none
2017-06-26 17:47:59 -07:00
shikhi7 7a279e5972 Formatted Dockerfile to be cleaner and precise 2017-06-27 06:12:56 +05:30
Kubernetes Submit Queue eadd20ef48 Merge pull request #47748 from nikhita/client-go-examples-readme
Automatic merge from submit-queue

Update custom-resources example in client-go

- Update client-go examples `README.md` to point to the CustomResources example instead of the deprecated TPR one.
- Delete `staging/src/k8s.io/client-go/examples/custom-resources`. 

Fixing #47743.

**Release note**:

```release-note
NONE
```

/cc @ahmetb @sttts
2017-06-26 16:30:55 -07:00
Kubernetes Submit Queue df41787b1a Merge pull request #47823 from sjenning/fix-waiting-reason
Automatic merge from submit-queue (batch tested with PRs 44058, 48085, 48077, 48076, 47823)

don't pass CRI error through to waiting state reason

Raw gRPC errors are getting into the `Reason` field of the container status `State`, causing it to be output inline on a `kubectl get pod`

xref https://bugzilla.redhat.com/show_bug.cgi?id=1449820

Basically the issue is that the err and msg are reversed in `startContainer()`.  The msg is short and the err is long.  It should be the other way around.

This PR changes `startContainer()` to return a short error that becomes the Reason and the extracted gPRC error description that becomes the Message.

@derekwaynecarr @smarterclayton @eparis
2017-06-26 15:29:33 -07:00
Kubernetes Submit Queue 36e17f6d07 Merge pull request #48076 from nikhita/fix-local-cluster-up
Automatic merge from submit-queue (batch tested with PRs 44058, 48085, 48077, 48076, 47823)

Fix error in local-cluster-up

When $GO_OUT is not set, line 152 outputs an error.

```
./hack/local-up-cluster.sh: line 152: [: ==: unary operator expected
```
This occurs because the if condition expands as `if [  == "" ]`. This results in an error because == is a binary operator and expects something on the LHS.

**Release note**:

```release-note
NONE
```

cc @sttts 

FYI @Gouthamve
2017-06-26 15:29:31 -07:00
Kubernetes Submit Queue babd565908 Merge pull request #48077 from liggitt/kubefed-discovery-cache
Automatic merge from submit-queue (batch tested with PRs 44058, 48085, 48077, 48076, 47823)

Retry finding RBAC version if not found in discovery cache

Alternate to https://github.com/kubernetes/kubernetes/pull/47995

xref #47977

The caching discovery client can indicate whether it used fresh discovery data. `kubefed init` should invalidate and recheck if it doesn't find an RBAC API group

```release-note
`kubefed init` correctly checks for RBAC API enablement.
```
2017-06-26 15:29:29 -07:00
Kubernetes Submit Queue a3df4bf764 Merge pull request #48085 from shyamjvs/reduce-kubeproxy-logs
Automatic merge from submit-queue (batch tested with PRs 44058, 48085, 48077, 48076, 47823)

Move iptables logging in kubeproxy from Errorf to V(2).Infof

Fixes https://github.com/kubernetes/kubernetes/issues/48052

This will stop fluentd from OOM'ing in reasonably large clusters with services due to kube-proxy. You'll still get iptables printed on setups which run at >= v2, but we can at least optout.
@bowei Does this look reasonable?

cc @kubernetes/sig-network-misc
2017-06-26 15:29:27 -07:00
Kubernetes Submit Queue 6a28658ca1 Merge pull request #44058 from caesarxuchao/background-cascading
Automatic merge from submit-queue (batch tested with PRs 44058, 48085, 48077, 48076, 47823)

Make background garbage collection cascading

Fix #44046, fix #47843 where user reported that the garbage collector didn't delete pods when a deployment was deleted with PropagationPolicy=Background.

The cause is that when propagating background garbage collection request, the garbage collector deletes dependents with DeleteOptions.PropagationPolicy=nil, which means the default GC policy of a resource (defined by its REST strategy) and the existing GC-related finalizers will decide how the delete request is propagated further. Unfortunately, the default GC policy for RS is orphaning, so the pods are behind when a deployment is deleted.

This PR changes the garbage collector to delete dependents with DeleteOptions.PropagationPolicy=Background when the owner is deleted in background. This means the dependent's existing GC finalizers will be overridden, making orphaning less flexible (see this made-up [case](https://github.com/kubernetes/kubeadm/issues/149#issuecomment-278942012)). I think sacrificing the flexibility of orphaning is worthwhile, because making the behavior of background garbage collection matching users' expectation is more important.

cc @lavalamp @kargakis @krmayankk @enisoc 

```release-note
The garbage collector now cascades deletion properly when deleting an object with propagationPolicy="background". This resolves issue [#44046](https://github.com/kubernetes/kubernetes/issues/44046), so that when a deployment is deleted with propagationPolicy="background", the garbage collector ensures dependent pods are deleted as well.
```
2017-06-26 15:29:25 -07:00
Nikhita Raghunath bfa9ce4308
Update CR example in client-go
Remove custom-resources directory from client-go

Add TPR example back

Mention CRD is successor to TPR
2017-06-27 03:39:29 +05:30
Bobby (Babak) Salamat 403b30f6e9 Autogenerated files 2017-06-26 15:02:49 -07:00
Tim St. Clair b34d6ab890
s/count/total/ in audit prometheus metrics 2017-06-26 14:41:20 -07:00
Jeff Grafton 211554a37f bazel: update rules_docker and use official busybox base image 2017-06-26 14:29:58 -07:00
Kubernetes Submit Queue ec1fcbed74 Merge pull request #47170 from ktsakalozos/bug/restart
Automatic merge from submit-queue (batch tested with PRs 47860, 47170)

Fix restart action on juju kubernetes-master

**What this PR does / why we need it**: Restart action of kubernetes-master of Juju is not functioning. 

**Which issue this PR fixes** *(optional, in `fixes #<issue number>(, fixes #<issue_number>, ...)` format, will close that issue when PR gets merged)*: fixes https://github.com/juju-solutions/bundle-canonical-kubernetes/issues/299

**Special notes for your reviewer**:

**Release note**:

```
Fix: Restart action of juju's kubernetes-master restarts the respective snap based services
```
2017-06-26 14:03:06 -07:00
Kubernetes Submit Queue 93ef2f2f59 Merge pull request #47860 from crassirostris/fluentd-log-to-stdio
Automatic merge from submit-queue (batch tested with PRs 47860, 47170)

Make fluentd log to stdio instead of a dedicated file

Lower verbosity also, to reduce volume of system logs exported to the backend.

Fix https://github.com/kubernetes/kubernetes/issues/43772

/cc @piosz
2017-06-26 14:03:04 -07:00
Bobby Salamat 91f893eebe Add priority to Kubernetes API 2017-06-26 13:43:19 -07:00
Kubernetes Submit Queue 53a5ee7929 Merge pull request #48022 from smarterclayton/more_printers
Automatic merge from submit-queue (batch tested with PRs 48036, 48022)

Move more printers to TablePrinter

Continues work from 1.7
2017-06-26 12:52:45 -07:00
Kubernetes Submit Queue 6b973991a8 Merge pull request #48036 from hongchaodeng/master
Automatic merge from submit-queue (batch tested with PRs 48036, 48022)

apiextensions-apiserver: fix build

Can't build CRD due to this bug. This PR will fix it.
2017-06-26 12:52:39 -07:00
Anthony Yeh 4ab06a6922
Skip Deployment upgrade test on 1.5 and earlier.
The test relies on implementation details and would need a rewrite
to work for older clusters.
2017-06-26 12:32:39 -07:00