Change minion to node

Contination of #1111

I tried to keep this PR down to just a simple search-n-replace to keep
things simple.  I may have gone too far in some spots but its easy to
roll those back if needed.

I avoided renaming `contrib/mesos/pkg/minion` because there's already
a `contrib/mesos/pkg/node` dir and fixing that will require a bit of work
due to a circular import chain that pops up. So I'm saving that for a
follow-on PR.

I rolled back some of this from a previous commit because it just got
to big/messy. Will follow up with additional PRs

Signed-off-by: Doug Davis <dug@us.ibm.com>
pull/6/head
Doug Davis 2016-05-05 13:41:49 -07:00
parent 5af1b25235
commit 9d5bac6330
40 changed files with 195 additions and 186 deletions

View File

@ -16760,7 +16760,7 @@
}, },
"v1.Node": { "v1.Node": {
"id": "v1.Node", "id": "v1.Node",
"description": "Node is a worker node in Kubernetes, formerly known as minion. Each node will have a unique identifier in the cache (i.e. in etcd).", "description": "Node is a worker node in Kubernetes. Each node will have a unique identifier in the cache (i.e. in etcd).",
"properties": { "properties": {
"kind": { "kind": {
"type": "string", "type": "string",

View File

@ -481,7 +481,7 @@ span.icon > [class^="icon-"], span.icon > [class*=" icon-"] { cursor: default; }
<div class="sect2"> <div class="sect2">
<h3 id="_v1_node">v1.Node</h3> <h3 id="_v1_node">v1.Node</h3>
<div class="paragraph"> <div class="paragraph">
<p>Node is a worker node in Kubernetes, formerly known as minion. Each node will have a unique identifier in the cache (i.e. in etcd).</p> <p>Node is a worker node in Kubernetes. Each node will have a unique identifier in the cache (i.e. in etcd).</p>
</div> </div>
<table class="tableblock frame-all grid-all" style="width:100%; "> <table class="tableblock frame-all grid-all" style="width:100%; ">
<colgroup> <colgroup>

View File

@ -176,7 +176,7 @@ Some important differences between v1beta1/2 and v1beta3:
* The resource `id` is now called `name`. * The resource `id` is now called `name`.
* `name`, `labels`, `annotations`, and other metadata are now nested in a map called `metadata` * `name`, `labels`, `annotations`, and other metadata are now nested in a map called `metadata`
* `desiredState` is now called `spec`, and `currentState` is now called `status` * `desiredState` is now called `spec`, and `currentState` is now called `status`
* `/minions` has been moved to `/nodes`, and the resource has kind `Node` * `/nodes` has been moved to `/nodes`, and the resource has kind `Node`
* The namespace is required (for all namespaced resources) and has moved from a URL parameter to the path: `/api/v1beta3/namespaces/{namespace}/{resource_collection}/{resource_name}`. If you were not using a namespace before, use `default` here. * The namespace is required (for all namespaced resources) and has moved from a URL parameter to the path: `/api/v1beta3/namespaces/{namespace}/{resource_collection}/{resource_name}`. If you were not using a namespace before, use `default` here.
* The names of all resource collections are now lower cased - instead of `replicationControllers`, use `replicationcontrollers`. * The names of all resource collections are now lower cased - instead of `replicationControllers`, use `replicationcontrollers`.
* To watch for changes to a resource, open an HTTP or Websocket connection to the collection query and provide the `?watch=true` query parameter along with the desired `resourceVersion` parameter to watch from. * To watch for changes to a resource, open an HTTP or Websocket connection to the collection query and provide the `?watch=true` query parameter along with the desired `resourceVersion` parameter to watch from.

View File

@ -64,7 +64,7 @@ you manually created or configured your cluster.
### Architecture overview ### Architecture overview
Kubernetes is a cluster of several machines that consists of a Kubernetes Kubernetes is a cluster of several machines that consists of a Kubernetes
master and a set number of nodes (previously known as 'minions') for which the master and a set number of nodes (previously known as 'nodes') for which the
master which is responsible. See the [Architecture](architecture.md) topic for master which is responsible. See the [Architecture](architecture.md) topic for
more details. more details.
@ -161,7 +161,7 @@ Note that we do not automatically open NodePort services in the AWS firewall
NodePort services are more of a building block for things like inter-cluster NodePort services are more of a building block for things like inter-cluster
services or for LoadBalancer. To consume a NodePort service externally, you services or for LoadBalancer. To consume a NodePort service externally, you
will likely have to open the port in the node security group will likely have to open the port in the node security group
(`kubernetes-minion-<clusterid>`). (`kubernetes-node-<clusterid>`).
For SSL support, starting with 1.3 two annotations can be added to a service: For SSL support, starting with 1.3 two annotations can be added to a service:
@ -194,7 +194,7 @@ modifying the headers.
kube-proxy sets up two IAM roles, one for the master called kube-proxy sets up two IAM roles, one for the master called
[kubernetes-master](../../cluster/aws/templates/iam/kubernetes-master-policy.json) [kubernetes-master](../../cluster/aws/templates/iam/kubernetes-master-policy.json)
and one for the nodes called and one for the nodes called
[kubernetes-minion](../../cluster/aws/templates/iam/kubernetes-minion-policy.json). [kubernetes-node](../../cluster/aws/templates/iam/kubernetes-minion-policy.json).
The master is responsible for creating ELBs and configuring them, as well as The master is responsible for creating ELBs and configuring them, as well as
setting up advanced VPC routing. Currently it has blanket permissions on EC2, setting up advanced VPC routing. Currently it has blanket permissions on EC2,
@ -242,7 +242,7 @@ HTTP URLs are passed to instances; this is how Kubernetes code gets onto the
machines. machines.
* Creates two IAM profiles based on templates in [cluster/aws/templates/iam](../../cluster/aws/templates/iam/): * Creates two IAM profiles based on templates in [cluster/aws/templates/iam](../../cluster/aws/templates/iam/):
* `kubernetes-master` is used by the master. * `kubernetes-master` is used by the master.
* `kubernetes-minion` is used by nodes. * `kubernetes-node` is used by nodes.
* Creates an AWS SSH key named `kubernetes-<fingerprint>`. Fingerprint here is * Creates an AWS SSH key named `kubernetes-<fingerprint>`. Fingerprint here is
the OpenSSH key fingerprint, so that multiple users can run the script with the OpenSSH key fingerprint, so that multiple users can run the script with
different keys and their keys will not collide (with near-certainty). It will different keys and their keys will not collide (with near-certainty). It will
@ -265,7 +265,7 @@ The debate is open here, where cluster-per-AZ is discussed as more robust but
cross-AZ-clusters are more convenient. cross-AZ-clusters are more convenient.
* Associates the subnet to the route table * Associates the subnet to the route table
* Creates security groups for the master (`kubernetes-master-<clusterid>`) * Creates security groups for the master (`kubernetes-master-<clusterid>`)
and the nodes (`kubernetes-minion-<clusterid>`). and the nodes (`kubernetes-node-<clusterid>`).
* Configures security groups so that masters and nodes can communicate. This * Configures security groups so that masters and nodes can communicate. This
includes intercommunication between masters and nodes, opening SSH publicly includes intercommunication between masters and nodes, opening SSH publicly
for both masters and nodes, and opening port 443 on the master for the HTTPS for both masters and nodes, and opening port 443 on the master for the HTTPS
@ -281,8 +281,8 @@ information that must be passed in this way.
routing rule for the internal network range (`MASTER_IP_RANGE`, defaults to routing rule for the internal network range (`MASTER_IP_RANGE`, defaults to
10.246.0.0/24). 10.246.0.0/24).
* For auto-scaling, on each nodes it creates a launch configuration and group. * For auto-scaling, on each nodes it creates a launch configuration and group.
The name for both is <*KUBE_AWS_INSTANCE_PREFIX*>-minion-group. The default The name for both is <*KUBE_AWS_INSTANCE_PREFIX*>-node-group. The default
name is kubernetes-minion-group. The auto-scaling group has a min and max size name is kubernetes-node-group. The auto-scaling group has a min and max size
that are both set to NUM_NODES. You can change the size of the auto-scaling that are both set to NUM_NODES. You can change the size of the auto-scaling
group to add or remove the total number of nodes from within the AWS API or group to add or remove the total number of nodes from within the AWS API or
Console. Each nodes self-configures, meaning that they come up; run Salt with Console. Each nodes self-configures, meaning that they come up; run Salt with

View File

@ -170,10 +170,10 @@ Sample kubectl output:
```console ```console
FIRSTSEEN LASTSEEN COUNT NAME KIND SUBOBJECT REASON SOURCE MESSAGE FIRSTSEEN LASTSEEN COUNT NAME KIND SUBOBJECT REASON SOURCE MESSAGE
Thu, 12 Feb 2015 01:13:02 +0000 Thu, 12 Feb 2015 01:13:02 +0000 1 kubernetes-node-4.c.saad-dev-vms.internal Minion starting {kubelet kubernetes-node-4.c.saad-dev-vms.internal} Starting kubelet. Thu, 12 Feb 2015 01:13:02 +0000 Thu, 12 Feb 2015 01:13:02 +0000 1 kubernetes-node-4.c.saad-dev-vms.internal Node starting {kubelet kubernetes-node-4.c.saad-dev-vms.internal} Starting kubelet.
Thu, 12 Feb 2015 01:13:09 +0000 Thu, 12 Feb 2015 01:13:09 +0000 1 kubernetes-node-1.c.saad-dev-vms.internal Minion starting {kubelet kubernetes-node-1.c.saad-dev-vms.internal} Starting kubelet. Thu, 12 Feb 2015 01:13:09 +0000 Thu, 12 Feb 2015 01:13:09 +0000 1 kubernetes-node-1.c.saad-dev-vms.internal Node starting {kubelet kubernetes-node-1.c.saad-dev-vms.internal} Starting kubelet.
Thu, 12 Feb 2015 01:13:09 +0000 Thu, 12 Feb 2015 01:13:09 +0000 1 kubernetes-node-3.c.saad-dev-vms.internal Minion starting {kubelet kubernetes-node-3.c.saad-dev-vms.internal} Starting kubelet. Thu, 12 Feb 2015 01:13:09 +0000 Thu, 12 Feb 2015 01:13:09 +0000 1 kubernetes-node-3.c.saad-dev-vms.internal Node starting {kubelet kubernetes-node-3.c.saad-dev-vms.internal} Starting kubelet.
Thu, 12 Feb 2015 01:13:09 +0000 Thu, 12 Feb 2015 01:13:09 +0000 1 kubernetes-node-2.c.saad-dev-vms.internal Minion starting {kubelet kubernetes-node-2.c.saad-dev-vms.internal} Starting kubelet. Thu, 12 Feb 2015 01:13:09 +0000 Thu, 12 Feb 2015 01:13:09 +0000 1 kubernetes-node-2.c.saad-dev-vms.internal Node starting {kubelet kubernetes-node-2.c.saad-dev-vms.internal} Starting kubelet.
Thu, 12 Feb 2015 01:13:05 +0000 Thu, 12 Feb 2015 01:13:12 +0000 4 monitoring-influx-grafana-controller-0133o Pod failedScheduling {scheduler } Error scheduling: no nodes available to schedule pods Thu, 12 Feb 2015 01:13:05 +0000 Thu, 12 Feb 2015 01:13:12 +0000 4 monitoring-influx-grafana-controller-0133o Pod failedScheduling {scheduler } Error scheduling: no nodes available to schedule pods
Thu, 12 Feb 2015 01:13:05 +0000 Thu, 12 Feb 2015 01:13:12 +0000 4 elasticsearch-logging-controller-fplln Pod failedScheduling {scheduler } Error scheduling: no nodes available to schedule pods Thu, 12 Feb 2015 01:13:05 +0000 Thu, 12 Feb 2015 01:13:12 +0000 4 elasticsearch-logging-controller-fplln Pod failedScheduling {scheduler } Error scheduling: no nodes available to schedule pods
Thu, 12 Feb 2015 01:13:05 +0000 Thu, 12 Feb 2015 01:13:12 +0000 4 kibana-logging-controller-gziey Pod failedScheduling {scheduler } Error scheduling: no nodes available to schedule pods Thu, 12 Feb 2015 01:13:05 +0000 Thu, 12 Feb 2015 01:13:12 +0000 4 kibana-logging-controller-gziey Pod failedScheduling {scheduler } Error scheduling: no nodes available to schedule pods

View File

@ -1182,7 +1182,7 @@ than capitalization of the initial letter, the two should almost always match.
No underscores nor dashes in either. No underscores nor dashes in either.
* Field and resource names should be declarative, not imperative (DoSomething, * Field and resource names should be declarative, not imperative (DoSomething,
SomethingDoer, DoneBy, DoneAt). SomethingDoer, DoneBy, DoneAt).
* `Minion` has been deprecated in favor of `Node`. Use `Node` where referring to * Use `Node` where referring to
the node resource in the context of the cluster. Use `Host` where referring to the node resource in the context of the cluster. Use `Host` where referring to
properties of the individual physical/virtual system, such as `hostname`, properties of the individual physical/virtual system, such as `hostname`,
`hostPath`, `hostNetwork`, etc. `hostPath`, `hostNetwork`, etc.

View File

@ -371,8 +371,8 @@ provisioned.
#### I have Vagrant up but the nodes won't validate! #### I have Vagrant up but the nodes won't validate!
Log on to one of the nodes (`vagrant ssh node-1`) and inspect the salt minion Log on to one of the nodes (`vagrant ssh node-1`) and inspect the salt node
log (`sudo cat /var/log/salt/minion`). log (`sudo cat /var/log/salt/node`).
#### I want to change the number of nodes! #### I want to change the number of nodes!

View File

@ -92,9 +92,9 @@ Use the `examples/guestbook-go/redis-master-controller.json` file to create a [r
4. To verify what containers are running in the redis-master pod, you can SSH to that machine with `gcloud compute ssh --zone` *`zone_name`* *`host_name`* and then run `docker ps`: 4. To verify what containers are running in the redis-master pod, you can SSH to that machine with `gcloud compute ssh --zone` *`zone_name`* *`host_name`* and then run `docker ps`:
```console ```console
me@workstation$ gcloud compute ssh --zone us-central1-b kubernetes-minion-bz1p me@workstation$ gcloud compute ssh --zone us-central1-b kubernetes-node-bz1p
me@kubernetes-minion-3:~$ sudo docker ps me@kubernetes-node-3:~$ sudo docker ps
CONTAINER ID IMAGE COMMAND CREATED STATUS CONTAINER ID IMAGE COMMAND CREATED STATUS
d5c458dabe50 redis "/entrypoint.sh redis" 5 minutes ago Up 5 minutes d5c458dabe50 redis "/entrypoint.sh redis" 5 minutes ago Up 5 minutes
``` ```

View File

@ -322,7 +322,7 @@ You can get information about a pod, including the machine that it is running on
```console ```console
$ kubectl describe pods redis-master-2353460263-1ecey $ kubectl describe pods redis-master-2353460263-1ecey
Name: redis-master-2353460263-1ecey Name: redis-master-2353460263-1ecey
Node: kubernetes-minion-m0k7/10.240.0.5 Node: kubernetes-node-m0k7/10.240.0.5
... ...
Labels: app=redis,pod-template-hash=2353460263,role=master,tier=backend Labels: app=redis,pod-template-hash=2353460263,role=master,tier=backend
Status: Running Status: Running
@ -337,7 +337,7 @@ Containers:
... ...
``` ```
The `Node` is the name and IP of the machine, e.g. `kubernetes-minion-m0k7` in the example above. You can find more details about this node with `kubectl describe nodes kubernetes-minion-m0k7`. The `Node` is the name and IP of the machine, e.g. `kubernetes-node-m0k7` in the example above. You can find more details about this node with `kubectl describe nodes kubernetes-node-m0k7`.
If you want to view the container logs for a given pod, you can run: If you want to view the container logs for a given pod, you can run:
@ -356,7 +356,7 @@ me@workstation$ gcloud compute ssh <NODE-NAME>
Then, you can look at the Docker containers on the remote machine. You should see something like this (the specifics of the IDs will be different): Then, you can look at the Docker containers on the remote machine. You should see something like this (the specifics of the IDs will be different):
```console ```console
me@kubernetes-minion-krxw:~$ sudo docker ps me@kubernetes-node-krxw:~$ sudo docker ps
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
... ...
0ffef9649265 redis:latest "/entrypoint.sh redi" About a minute ago Up About a minute k8s_master.869d22f3_redis-master-dz33o_default_1449a58a-5ead-11e5-a104-688f84ef8ef6_d74cb2b5 0ffef9649265 redis:latest "/entrypoint.sh redi" About a minute ago Up About a minute k8s_master.869d22f3_redis-master-dz33o_default_1449a58a-5ead-11e5-a104-688f84ef8ef6_d74cb2b5
@ -718,10 +718,10 @@ NAME REGION IP_ADDRESS IP_PROTOCOL TARGET
frontend us-central1 130.211.188.51 TCP us-central1/targetPools/frontend frontend us-central1 130.211.188.51 TCP us-central1/targetPools/frontend
``` ```
In Google Compute Engine, you also may need to open the firewall for port 80 using the [console][cloud-console] or the `gcloud` tool. The following command will allow traffic from any source to instances tagged `kubernetes-minion` (replace with your tags as appropriate): In Google Compute Engine, you also may need to open the firewall for port 80 using the [console][cloud-console] or the `gcloud` tool. The following command will allow traffic from any source to instances tagged `kubernetes-node` (replace with your tags as appropriate):
```console ```console
$ gcloud compute firewall-rules create --allow=tcp:80 --target-tags=kubernetes-minion kubernetes-minion-80 $ gcloud compute firewall-rules create --allow=tcp:80 --target-tags=kubernetes-node kubernetes-node-80
``` ```
For GCE Kubernetes startup details, see the [Getting started on Google Compute Engine](../../docs/getting-started-guides/gce.md) For GCE Kubernetes startup details, see the [Getting started on Google Compute Engine](../../docs/getting-started-guides/gce.md)

View File

@ -143,12 +143,12 @@ kubectl get -o template po wildfly-rc-w2kk5 --template={{.status.podIP}}
10.246.1.23 10.246.1.23
``` ```
Log in to minion and access the application: Log in to node and access the application:
```sh ```sh
vagrant ssh minion-1 vagrant ssh node-1
Last login: Thu Jul 16 00:24:36 2015 from 10.0.2.2 Last login: Thu Jul 16 00:24:36 2015 from 10.0.2.2
[vagrant@kubernetes-minion-1 ~]$ curl http://10.246.1.23:8080/employees/resources/employees/ [vagrant@kubernetes-node-1 ~]$ curl http://10.246.1.23:8080/employees/resources/employees/
<?xml version="1.0" encoding="UTF-8" standalone="yes"?><collection><employee><id>1</id><name>Penny</name></employee><employee><id>2</id><name>Sheldon</name></employee><employee><id>3</id><name>Amy</name></employee><employee><id>4</id><name>Leonard</name></employee><employee><id>5</id><name>Bernadette</name></employee><employee><id>6</id><name>Raj</name></employee><employee><id>7</id><name>Howard</name></employee><employee><id>8</id><name>Priya</name></employee></collection> <?xml version="1.0" encoding="UTF-8" standalone="yes"?><collection><employee><id>1</id><name>Penny</name></employee><employee><id>2</id><name>Sheldon</name></employee><employee><id>3</id><name>Amy</name></employee><employee><id>4</id><name>Leonard</name></employee><employee><id>5</id><name>Bernadette</name></employee><employee><id>6</id><name>Raj</name></employee><employee><id>7</id><name>Howard</name></employee><employee><id>8</id><name>Priya</name></employee></collection>
``` ```

View File

@ -180,7 +180,7 @@ You will have to open up port 80 if it's not open yet in your
environment. On Google Compute Engine, you may run the below command. environment. On Google Compute Engine, you may run the below command.
``` ```
gcloud compute firewall-rules create meteor-80 --allow=tcp:80 --target-tags kubernetes-minion gcloud compute firewall-rules create meteor-80 --allow=tcp:80 --target-tags kubernetes-node
``` ```
What is going on? What is going on?

View File

@ -59,7 +59,7 @@ $ vi cluster/saltbase/pillar/privilege.sls
allow_privileged: true allow_privileged: true
``` ```
Now spin up a cluster using your preferred KUBERNETES_PROVIDER. Remember that `kube-up.sh` may start other pods on your minion nodes, so ensure that you have enough resources to run the five pods for this example. Now spin up a cluster using your preferred KUBERNETES_PROVIDER. Remember that `kube-up.sh` may start other pods on your nodes, so ensure that you have enough resources to run the five pods for this example.
```sh ```sh

View File

@ -160,7 +160,7 @@ phabricator-controller-9vy68 1/1 Running 0 1m
If you ssh to that machine, you can run `docker ps` to see the actual pod: If you ssh to that machine, you can run `docker ps` to see the actual pod:
```sh ```sh
me@workstation$ gcloud compute ssh --zone us-central1-b kubernetes-minion-2 me@workstation$ gcloud compute ssh --zone us-central1-b kubernetes-node-2
$ sudo docker ps $ sudo docker ps
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
@ -230,10 +230,10 @@ and then visit port 80 of that IP address.
**Note**: Provisioning of the external IP address may take few minutes. **Note**: Provisioning of the external IP address may take few minutes.
**Note**: You may need to open the firewall for port 80 using the [console][cloud-console] or the `gcloud` tool. The following command will allow traffic from any source to instances tagged `kubernetes-minion`: **Note**: You may need to open the firewall for port 80 using the [console][cloud-console] or the `gcloud` tool. The following command will allow traffic from any source to instances tagged `kubernetes-node`:
```sh ```sh
$ gcloud compute firewall-rules create phabricator-node-80 --allow=tcp:80 --target-tags kubernetes-minion $ gcloud compute firewall-rules create phabricator-node-80 --allow=tcp:80 --target-tags kubernetes-node
``` ```
### Step Six: Cleanup ### Step Six: Cleanup

View File

@ -16,5 +16,5 @@
echo "Create Phabricator replication controller" && kubectl create -f phabricator-controller.json echo "Create Phabricator replication controller" && kubectl create -f phabricator-controller.json
echo "Create Phabricator service" && kubectl create -f phabricator-service.json echo "Create Phabricator service" && kubectl create -f phabricator-service.json
echo "Create firewall rule" && gcloud compute firewall-rules create phabricator-node-80 --allow=tcp:80 --target-tags kubernetes-minion echo "Create firewall rule" && gcloud compute firewall-rules create phabricator-node-80 --allow=tcp:80 --target-tags kubernetes-node

View File

@ -79,13 +79,13 @@ $ cluster/kubectl.sh run cpuhog \
-- md5sum /dev/urandom -- md5sum /dev/urandom
``` ```
This will create a single pod on your minion that requests 1/10 of a CPU, but it has no limit on how much CPU it may actually consume This will create a single pod on your node that requests 1/10 of a CPU, but it has no limit on how much CPU it may actually consume
on the node. on the node.
To demonstrate this, if you SSH into your machine, you will see it is consuming as much CPU as possible on the node. To demonstrate this, if you SSH into your machine, you will see it is consuming as much CPU as possible on the node.
``` ```
$ vagrant ssh minion-1 $ vagrant ssh node-1
$ sudo docker stats $(sudo docker ps -q) $ sudo docker stats $(sudo docker ps -q)
CONTAINER CPU % MEM USAGE/LIMIT MEM % NET I/O CONTAINER CPU % MEM USAGE/LIMIT MEM % NET I/O
6b593b1a9658 0.00% 1.425 MB/1.042 GB 0.14% 1.038 kB/738 B 6b593b1a9658 0.00% 1.425 MB/1.042 GB 0.14% 1.038 kB/738 B
@ -150,7 +150,7 @@ $ cluster/kubectl.sh run cpuhog \
Let's SSH into the node, and look at usage stats. Let's SSH into the node, and look at usage stats.
``` ```
$ vagrant ssh minion-1 $ vagrant ssh node-1
$ sudo su $ sudo su
$ docker stats $(docker ps -q) $ docker stats $(docker ps -q)
CONTAINER CPU % MEM USAGE/LIMIT MEM % NET I/O CONTAINER CPU % MEM USAGE/LIMIT MEM % NET I/O

View File

@ -88,18 +88,18 @@ And kubectl get nodes should agree:
``` ```
$ kubectl get nodes $ kubectl get nodes
NAME LABELS STATUS NAME LABELS STATUS
eu-minion-0n61 kubernetes.io/hostname=eu-minion-0n61 Ready eu-node-0n61 kubernetes.io/hostname=eu-node-0n61 Ready
eu-minion-79ua kubernetes.io/hostname=eu-minion-79ua Ready eu-node-79ua kubernetes.io/hostname=eu-node-79ua Ready
eu-minion-7wz7 kubernetes.io/hostname=eu-minion-7wz7 Ready eu-node-7wz7 kubernetes.io/hostname=eu-node-7wz7 Ready
eu-minion-loh2 kubernetes.io/hostname=eu-minion-loh2 Ready eu-node-loh2 kubernetes.io/hostname=eu-node-loh2 Ready
$ kubectl config use-context <clustername_us> $ kubectl config use-context <clustername_us>
$ kubectl get nodes $ kubectl get nodes
NAME LABELS STATUS NAME LABELS STATUS
kubernetes-minion-5jtd kubernetes.io/hostname=kubernetes-minion-5jtd Ready kubernetes-node-5jtd kubernetes.io/hostname=kubernetes-node-5jtd Ready
kubernetes-minion-lqfc kubernetes.io/hostname=kubernetes-minion-lqfc Ready kubernetes-node-lqfc kubernetes.io/hostname=kubernetes-node-lqfc Ready
kubernetes-minion-sjra kubernetes.io/hostname=kubernetes-minion-sjra Ready kubernetes-node-sjra kubernetes.io/hostname=kubernetes-node-sjra Ready
kubernetes-minion-wul8 kubernetes.io/hostname=kubernetes-minion-wul8 Ready kubernetes-node-wul8 kubernetes.io/hostname=kubernetes-node-wul8 Ready
``` ```
## Testing reachability ## Testing reachability
@ -207,10 +207,10 @@ $ kubectl exec -it kubectl-tester bash
kubectl-tester $ kubectl get nodes kubectl-tester $ kubectl get nodes
NAME LABELS STATUS NAME LABELS STATUS
eu-minion-0n61 kubernetes.io/hostname=eu-minion-0n61 Ready eu-node-0n61 kubernetes.io/hostname=eu-node-0n61 Ready
eu-minion-79ua kubernetes.io/hostname=eu-minion-79ua Ready eu-node-79ua kubernetes.io/hostname=eu-node-79ua Ready
eu-minion-7wz7 kubernetes.io/hostname=eu-minion-7wz7 Ready eu-node-7wz7 kubernetes.io/hostname=eu-node-7wz7 Ready
eu-minion-loh2 kubernetes.io/hostname=eu-minion-loh2 Ready eu-node-loh2 kubernetes.io/hostname=eu-node-loh2 Ready
``` ```
For a more advanced example of sharing clusters, see the [service-loadbalancer](https://github.com/kubernetes/contrib/tree/master/service-loadbalancer/README.md) For a more advanced example of sharing clusters, see the [service-loadbalancer](https://github.com/kubernetes/contrib/tree/master/service-loadbalancer/README.md)

View File

@ -70,7 +70,7 @@ gs://kubernetes-jenkins/logs/kubernetes-e2e-gce/
gcp-resources-{before, after}.txt gcp-resources-{before, after}.txt
junit_{00, 01, ...}.xml junit_{00, 01, ...}.xml
jenkins-e2e-master/{kube-apiserver.log, ...} jenkins-e2e-master/{kube-apiserver.log, ...}
jenkins-e2e-minion-abcd/{kubelet.log, ...} jenkins-e2e-node-abcd/{kubelet.log, ...}
12344/ 12344/
... ...
``` ```

View File

@ -44,7 +44,6 @@ cluster/saltbase/salt/etcd/etcd.manifest: "value": "{{ storage_backend }}
cluster/saltbase/salt/etcd/etcd.manifest:{% set storage_backend = pillar.get('storage_backend', 'etcd2') -%} cluster/saltbase/salt/etcd/etcd.manifest:{% set storage_backend = pillar.get('storage_backend', 'etcd2') -%}
cluster/saltbase/salt/kube-admission-controls/init.sls:{% if 'LimitRanger' in pillar.get('admission_control', '') %} cluster/saltbase/salt/kube-admission-controls/init.sls:{% if 'LimitRanger' in pillar.get('admission_control', '') %}
cluster/saltbase/salt/kube-apiserver/kube-apiserver.manifest:{% set params = address + " " + storage_backend + " " + etcd_servers + " " + etcd_servers_overrides + " " + cloud_provider + " " + cloud_config + " " + runtime_config + " " + feature_gates + " " + admission_control + " " + target_ram_mb + " " + service_cluster_ip_range + " " + client_ca_file + basic_auth_file + " " + min_request_timeout + " " + enable_garbage_collector -%} cluster/saltbase/salt/kube-apiserver/kube-apiserver.manifest:{% set params = address + " " + storage_backend + " " + etcd_servers + " " + etcd_servers_overrides + " " + cloud_provider + " " + cloud_config + " " + runtime_config + " " + feature_gates + " " + admission_control + " " + target_ram_mb + " " + service_cluster_ip_range + " " + client_ca_file + basic_auth_file + " " + min_request_timeout + " " + enable_garbage_collector -%}
cluster/saltbase/salt/kube-apiserver/kube-apiserver.manifest:{% set params = address + " " + storage_backend + " " + etcd_servers + " " + etcd_servers_overrides + " " + cloud_provider + " " + cloud_config + " " + runtime_config + " " + admission_control + " " + target_ram_mb + " " + service_cluster_ip_range + " " + client_ca_file + basic_auth_file + " " + min_request_timeout + " " + enable_garbage_collector -%}
cluster/saltbase/salt/kube-controller-manager/kube-controller-manager.manifest:{% set params = "--master=127.0.0.1:8080" + " " + cluster_name + " " + cluster_cidr + " " + allocate_node_cidrs + " " + service_cluster_ip_range + " " + terminated_pod_gc + " " + enable_garbage_collector + " " + cloud_provider + " " + cloud_config + " " + service_account_key + " " + log_level + " " + root_ca_file -%} cluster/saltbase/salt/kube-controller-manager/kube-controller-manager.manifest:{% set params = "--master=127.0.0.1:8080" + " " + cluster_name + " " + cluster_cidr + " " + allocate_node_cidrs + " " + service_cluster_ip_range + " " + terminated_pod_gc + " " + enable_garbage_collector + " " + cloud_provider + " " + cloud_config + " " + service_account_key + " " + log_level + " " + root_ca_file -%}
cluster/saltbase/salt/kube-controller-manager/kube-controller-manager.manifest:{% set params = params + " " + feature_gates -%} cluster/saltbase/salt/kube-controller-manager/kube-controller-manager.manifest:{% set params = params + " " + feature_gates -%}
cluster/saltbase/salt/kube-controller-manager/kube-controller-manager.manifest:{% if pillar.get('enable_hostpath_provisioner', '').lower() == 'true' -%} cluster/saltbase/salt/kube-controller-manager/kube-controller-manager.manifest:{% if pillar.get('enable_hostpath_provisioner', '').lower() == 'true' -%}

View File

@ -14,6 +14,7 @@ api-external-dns-names
api-burst api-burst
api-prefix api-prefix
api-rate api-rate
apiserver-count
api-server-port api-server-port
api-servers api-servers
api-token api-token
@ -31,9 +32,12 @@ authentication-token-webhook-config-file
authorization-mode authorization-mode
authorization-policy-file authorization-policy-file
authorization-rbac-super-user authorization-rbac-super-user
authorization-webhook-config-file
authorization-webhook-cache-authorized-ttl authorization-webhook-cache-authorized-ttl
authorization-webhook-cache-unauthorized-ttl authorization-webhook-cache-unauthorized-ttl
authorization-webhook-config-file
auth-path
auth-provider
auth-provider-arg
babysit-daemons babysit-daemons
basic-auth-file basic-auth-file
bench-pods bench-pods
@ -50,8 +54,8 @@ build-tag
cadvisor-port cadvisor-port
cert-dir cert-dir
certificate-authority certificate-authority
cgroups-per-qos
cgroup-root cgroup-root
cgroups-per-qos
chaos-chance chaos-chance
clean-start clean-start
cleanup cleanup
@ -68,21 +72,21 @@ cluster-cidr
cluster-dns cluster-dns
cluster-domain cluster-domain
cluster-ip cluster-ip
cluster-name
cluster-tag
cluster-monitor-period cluster-monitor-period
cluster-name
cluster-signing-cert-file cluster-signing-cert-file
cluster-signing-key-file cluster-signing-key-file
cni-bin-dir cni-bin-dir
cni-conf-dir cni-conf-dir
cluster-tag
concurrent-deployment-syncs concurrent-deployment-syncs
concurrent-endpoint-syncs concurrent-endpoint-syncs
concurrent-gc-syncs
concurrent-namespace-syncs concurrent-namespace-syncs
concurrent-replicaset-syncs concurrent-replicaset-syncs
concurrent-service-syncs
concurrent-resource-quota-syncs concurrent-resource-quota-syncs
concurrent-serviceaccount-token-syncs concurrent-serviceaccount-token-syncs
concurrent-gc-syncs concurrent-service-syncs
config-sync-period config-sync-period
configure-cbr0 configure-cbr0
configure-cloud-routes configure-cloud-routes
@ -93,10 +97,10 @@ conntrack-tcp-timeout-established
consumer-port consumer-port
consumer-service-name consumer-service-name
consumer-service-namespace consumer-service-namespace
contain-pod-resources
container-port container-port
container-runtime container-runtime
container-runtime-endpoint container-runtime-endpoint
contain-pod-resources
controller-start-interval controller-start-interval
cors-allowed-origins cors-allowed-origins
cpu-cfs-quota cpu-cfs-quota
@ -124,13 +128,13 @@ disable-kubenet
dns-port dns-port
dns-provider dns-provider
dns-provider-config dns-provider-config
dockercfg-path
docker-email docker-email
docker-endpoint docker-endpoint
docker-exec-handler docker-exec-handler
docker-password docker-password
docker-server docker-server
docker-username docker-username
dockercfg-path
driver-port driver-port
drop-embedded-fields drop-embedded-fields
dry-run dry-run
@ -141,8 +145,9 @@ e2e-verify-service-account
enable-controller-attach-detach enable-controller-attach-detach
enable-custom-metrics enable-custom-metrics
enable-debugging-handlers enable-debugging-handlers
enable-garbage-collector
enable-dynamic-provisioning enable-dynamic-provisioning
enable-garbage-collector
enable-garbage-collector
enable-hostpath-provisioner enable-hostpath-provisioner
enable-server enable-server
enable-swagger-ui enable-swagger-ui
@ -162,11 +167,11 @@ event-burst
event-qps event-qps
event-ttl event-ttl
eviction-hard eviction-hard
eviction-soft
eviction-soft-grace-period
eviction-pressure-transition-period
eviction-max-pod-grace-period eviction-max-pod-grace-period
eviction-minimum-reclaim eviction-minimum-reclaim
eviction-pressure-transition-period
eviction-soft
eviction-soft-grace-period
executor-bindall executor-bindall
executor-logv executor-logv
executor-path executor-path
@ -195,8 +200,8 @@ federated-api-qps
federated-kube-context federated-kube-context
federation-name federation-name
file-check-frequency file-check-frequency
file-suffix
file_content_in_loop file_content_in_loop
file-suffix
flex-volume-plugin-dir flex-volume-plugin-dir
forward-services forward-services
framework-name framework-name
@ -219,16 +224,16 @@ google-json-key
grace-period grace-period
ha-domain ha-domain
hairpin-mode hairpin-mode
hard-pod-affinity-symmetric-weight
hard hard
hard-pod-affinity-symmetric-weight
healthz-bind-address healthz-bind-address
healthz-port healthz-port
horizontal-pod-autoscaler-sync-period horizontal-pod-autoscaler-sync-period
host-ipc-sources host-ipc-sources
hostname-override
host-network-sources host-network-sources
host-pid-sources host-pid-sources
host-port-endpoints host-port-endpoints
hostname-override
http-check-frequency http-check-frequency
http-port http-port
ignore-daemonsets ignore-daemonsets
@ -241,6 +246,7 @@ image-pull-policy
image-service-endpoint image-service-endpoint
include-extended-apis include-extended-apis
included-types-overrides included-types-overrides
include-extended-apis
input-base input-base
input-dirs input-dirs
insecure-experimental-approve-all-kubelet-csrs-for-group insecure-experimental-approve-all-kubelet-csrs-for-group
@ -273,10 +279,6 @@ kops-zones
kube-api-burst kube-api-burst
kube-api-content-type kube-api-content-type
kube-api-qps kube-api-qps
kube-master
kube-master
kube-master-url
kube-reserved
kubecfg-file kubecfg-file
kubectl-path kubectl-path
kubelet-address kubelet-address
@ -298,6 +300,10 @@ kubelet-read-only-port
kubelet-root-dir kubelet-root-dir
kubelet-sync-frequency kubelet-sync-frequency
kubelet-timeout kubelet-timeout
kube-master
kube-master
kube-master-url
kube-reserved
kubernetes-service-node-port kubernetes-service-node-port
label-columns label-columns
large-cluster-size-threshold large-cluster-size-threshold
@ -324,6 +330,8 @@ master-os-distro
master-service-namespace master-service-namespace
max-concurrency max-concurrency
max-connection-bytes-per-sec max-connection-bytes-per-sec
maximum-dead-containers
maximum-dead-containers-per-container
max-log-age max-log-age
max-log-backups max-log-backups
max-log-size max-log-size
@ -332,8 +340,6 @@ max-outgoing-burst
max-outgoing-qps max-outgoing-qps
max-pods max-pods
max-requests-inflight max-requests-inflight
maximum-dead-containers
maximum-dead-containers-per-container
mesos-authentication-principal mesos-authentication-principal
mesos-authentication-provider mesos-authentication-provider
mesos-authentication-secret-file mesos-authentication-secret-file
@ -347,15 +353,15 @@ mesos-launch-grace-period
mesos-master mesos-master
mesos-sandbox-overlay mesos-sandbox-overlay
mesos-user mesos-user
min-pr-number
min-request-timeout
min-resync-period
minimum-container-ttl-duration minimum-container-ttl-duration
minimum-image-ttl-duration minimum-image-ttl-duration
minion-max-log-age minion-max-log-age
minion-max-log-backups minion-max-log-backups
minion-max-log-size minion-max-log-size
minion-path-override minion-path-override
min-pr-number
min-request-timeout
min-resync-period
namespace-sync-period namespace-sync-period
network-plugin network-plugin
network-plugin-dir network-plugin-dir
@ -367,14 +373,20 @@ node-eviction-rate
node-instance-group node-instance-group
node-ip node-ip
node-labels node-labels
node-max-log-age
node-max-log-backups
node-max-log-size
node-monitor-grace-period node-monitor-grace-period
node-monitor-period node-monitor-period
node-name node-name
node-os-distro node-os-distro
node-path-override
node-startup-grace-period node-startup-grace-period
node-status-update-frequency node-status-update-frequency
node-sync-period node-sync-period
no-headers
non-masquerade-cidr non-masquerade-cidr
no-suggestions
num-nodes num-nodes
oidc-ca-file oidc-ca-file
oidc-client-id oidc-client-id
@ -383,7 +395,6 @@ oidc-issuer-url
oidc-username-claim oidc-username-claim
only-idl only-idl
oom-score-adj oom-score-adj
out-version
outofdisk-transition-frequency outofdisk-transition-frequency
output-base output-base
output-directory output-directory
@ -391,6 +402,7 @@ output-file-base
output-package output-package
output-print-type output-print-type
output-version output-version
out-version
path-override path-override
pod-cidr pod-cidr
pod-eviction-timeout pod-eviction-timeout
@ -413,6 +425,7 @@ proxy-logv
proxy-mode proxy-mode
proxy-port-range proxy-port-range
public-address-override public-address-override
pvclaimbinder-sync-period
pv-recycler-increment-timeout-nfs pv-recycler-increment-timeout-nfs
pv-recycler-maximum-retry pv-recycler-maximum-retry
pv-recycler-minimum-timeout-hostpath pv-recycler-minimum-timeout-hostpath
@ -420,7 +433,6 @@ pv-recycler-minimum-timeout-nfs
pv-recycler-pod-template-filepath-hostpath pv-recycler-pod-template-filepath-hostpath
pv-recycler-pod-template-filepath-nfs pv-recycler-pod-template-filepath-nfs
pv-recycler-timeout-increment-hostpath pv-recycler-timeout-increment-hostpath
pvclaimbinder-sync-period
read-only-port read-only-port
really-crash-for-testing really-crash-for-testing
reconcile-cidr reconcile-cidr
@ -524,9 +536,9 @@ test-timeout
tls-ca-file tls-ca-file
tls-cert-file tls-cert-file
tls-private-key-file tls-private-key-file
to-version
token-auth-file token-auth-file
ttl-keys-prefix ttl-keys-prefix
to-version
ttl-secs ttl-secs
type-src type-src
udp-port udp-port

View File

@ -31,11 +31,9 @@ import (
const ( const (
StatusUnprocessableEntity = 422 StatusUnprocessableEntity = 422
StatusTooManyRequests = 429 StatusTooManyRequests = 429
// HTTP recommendations are for servers to define 5xx error codes // StatusServerTimeout is an indication that a transient server error has
// for scenarios not covered by behavior. In this case, ServerTimeout // occurred and the client *should* retry, with an optional Retry-After
// is an indication that a transient server error has occurred and the // header to specify the back off window.
// client *should* retry, with an optional Retry-After header to specify
// the back off window.
StatusServerTimeout = 504 StatusServerTimeout = 504
) )

View File

@ -2,8 +2,8 @@
"kind": "Node", "kind": "Node",
"apiVersion": "v1", "apiVersion": "v1",
"metadata": { "metadata": {
"name": "e2e-test-wojtekt-minion-etd6", "name": "e2e-test-wojtekt-node-etd6",
"selfLink": "/api/v1/nodes/e2e-test-wojtekt-minion-etd6", "selfLink": "/api/v1/nodes/e2e-test-wojtekt-node-etd6",
"uid": "a7e89222-e8e5-11e4-8fde-42010af09327", "uid": "a7e89222-e8e5-11e4-8fde-42010af09327",
"resourceVersion": "379", "resourceVersion": "379",
"creationTimestamp": "2015-04-22T11:49:39Z" "creationTimestamp": "2015-04-22T11:49:39Z"

View File

@ -331,7 +331,7 @@ func TestBadJSONRejection(t *testing.T) {
t.Errorf("Did not reject despite use of unknown type: %s", badJSONUnknownType) t.Errorf("Did not reject despite use of unknown type: %s", badJSONUnknownType)
} }
/*badJSONKindMismatch := []byte(`{"kind": "Pod"}`) /*badJSONKindMismatch := []byte(`{"kind": "Pod"}`)
if err2 := DecodeInto(badJSONKindMismatch, &Minion{}); err2 == nil { if err2 := DecodeInto(badJSONKindMismatch, &Node{}); err2 == nil {
t.Errorf("Kind is set but doesn't match the object type: %s", badJSONKindMismatch) t.Errorf("Kind is set but doesn't match the object type: %s", badJSONKindMismatch)
}*/ }*/
} }

View File

@ -1205,7 +1205,7 @@ message NamespaceStatus {
optional string phase = 1; optional string phase = 1;
} }
// Node is a worker node in Kubernetes, formerly known as minion. // Node is a worker node in Kubernetes.
// Each node will have a unique identifier in the cache (i.e. in etcd). // Each node will have a unique identifier in the cache (i.e. in etcd).
message Node { message Node {
// Standard object's metadata. // Standard object's metadata.

View File

@ -2678,7 +2678,7 @@ type ResourceList map[ResourceName]resource.Quantity
// +genclient=true // +genclient=true
// +nonNamespaced=true // +nonNamespaced=true
// Node is a worker node in Kubernetes, formerly known as minion. // Node is a worker node in Kubernetes.
// Each node will have a unique identifier in the cache (i.e. in etcd). // Each node will have a unique identifier in the cache (i.e. in etcd).
type Node struct { type Node struct {
unversioned.TypeMeta `json:",inline"` unversioned.TypeMeta `json:",inline"`

View File

@ -795,7 +795,7 @@ func (NamespaceStatus) SwaggerDoc() map[string]string {
} }
var map_Node = map[string]string{ var map_Node = map[string]string{
"": "Node is a worker node in Kubernetes, formerly known as minion. Each node will have a unique identifier in the cache (i.e. in etcd).", "": "Node is a worker node in Kubernetes. Each node will have a unique identifier in the cache (i.e. in etcd).",
"metadata": "Standard object's metadata. More info: http://releases.k8s.io/HEAD/docs/devel/api-conventions.md#metadata", "metadata": "Standard object's metadata. More info: http://releases.k8s.io/HEAD/docs/devel/api-conventions.md#metadata",
"spec": "Spec defines the behavior of a node. http://releases.k8s.io/HEAD/docs/devel/api-conventions.md#spec-and-status", "spec": "Spec defines the behavior of a node. http://releases.k8s.io/HEAD/docs/devel/api-conventions.md#spec-and-status",
"status": "Most recently observed status of the node. Populated by the system. Read-only. More info: http://releases.k8s.io/HEAD/docs/devel/api-conventions.md#spec-and-status", "status": "Most recently observed status of the node. Populated by the system. Read-only. More info: http://releases.k8s.io/HEAD/docs/devel/api-conventions.md#spec-and-status",

View File

@ -2885,7 +2885,7 @@ func (c *Cloud) updateInstanceSecurityGroupsForLoadBalancer(lb *elb.LoadBalancer
// Open the firewall from the load balancer to the instance // Open the firewall from the load balancer to the instance
// We don't actually have a trivial way to know in advance which security group the instance is in // We don't actually have a trivial way to know in advance which security group the instance is in
// (it is probably the minion security group, but we don't easily have that). // (it is probably the node security group, but we don't easily have that).
// However, we _do_ have the list of security groups on the instance records. // However, we _do_ have the list of security groups on the instance records.
// Map containing the changes we want to make; true to add, false to remove // Map containing the changes we want to make; true to add, false to remove

View File

@ -2085,7 +2085,7 @@ func (gce *GCECloud) GetInstanceGroup(name string, zone string) (*compute.Instan
// Take a GCE instance 'hostname' and break it down to something that can be fed // Take a GCE instance 'hostname' and break it down to something that can be fed
// to the GCE API client library. Basically this means reducing 'kubernetes- // to the GCE API client library. Basically this means reducing 'kubernetes-
// minion-2.c.my-proj.internal' to 'kubernetes-minion-2' if necessary. // node-2.c.my-proj.internal' to 'kubernetes-node-2' if necessary.
func canonicalizeInstanceName(name string) string { func canonicalizeInstanceName(name string) string {
ix := strings.Index(name, ".") ix := strings.Index(name, ".")
if ix != -1 { if ix != -1 {

View File

@ -283,7 +283,7 @@ func (m *OVirtInstanceMap) ListSortedNames() []string {
return names return names
} }
// List enumerates the set of minions instances known by the cloud provider // List enumerates the set of nodes instances known by the cloud provider
func (v *OVirtCloud) List(filter string) ([]types.NodeName, error) { func (v *OVirtCloud) List(filter string) ([]types.NodeName, error) {
instances, err := v.fetchAllInstances() instances, err := v.fetchAllInstances()
if err != nil { if err != nil {

View File

@ -573,7 +573,7 @@ func Example_printPodWithWideFormat() {
NegotiatedSerializer: ns, NegotiatedSerializer: ns,
Client: nil, Client: nil,
} }
nodeName := "kubernetes-minion-abcd" nodeName := "kubernetes-node-abcd"
cmd := NewCmdRun(f, os.Stdin, os.Stdout, os.Stderr) cmd := NewCmdRun(f, os.Stdin, os.Stdout, os.Stderr)
pod := &api.Pod{ pod := &api.Pod{
ObjectMeta: api.ObjectMeta{ ObjectMeta: api.ObjectMeta{
@ -600,7 +600,7 @@ func Example_printPodWithWideFormat() {
} }
// Output: // Output:
// NAME READY STATUS RESTARTS AGE IP NODE // NAME READY STATUS RESTARTS AGE IP NODE
// test1 1/2 podPhase 6 10y 10.1.1.3 kubernetes-minion-abcd // test1 1/2 podPhase 6 10y 10.1.1.3 kubernetes-node-abcd
} }
func Example_printPodWithShowLabels() { func Example_printPodWithShowLabels() {
@ -613,7 +613,7 @@ func Example_printPodWithShowLabels() {
NegotiatedSerializer: ns, NegotiatedSerializer: ns,
Client: nil, Client: nil,
} }
nodeName := "kubernetes-minion-abcd" nodeName := "kubernetes-node-abcd"
cmd := NewCmdRun(f, os.Stdin, os.Stdout, os.Stderr) cmd := NewCmdRun(f, os.Stdin, os.Stdout, os.Stderr)
pod := &api.Pod{ pod := &api.Pod{
ObjectMeta: api.ObjectMeta{ ObjectMeta: api.ObjectMeta{
@ -647,7 +647,7 @@ func Example_printPodWithShowLabels() {
} }
func newAllPhasePodList() *api.PodList { func newAllPhasePodList() *api.PodList {
nodeName := "kubernetes-minion-abcd" nodeName := "kubernetes-node-abcd"
return &api.PodList{ return &api.PodList{
Items: []api.Pod{ Items: []api.Pod{
{ {

View File

@ -48,7 +48,7 @@ var (
describe_example = dedent.Dedent(` describe_example = dedent.Dedent(`
# Describe a node # Describe a node
kubectl describe nodes kubernetes-minion-emt8.c.myproject.internal kubectl describe nodes kubernetes-node-emt8.c.myproject.internal
# Describe a pod # Describe a pod
kubectl describe pods/nginx kubectl describe pods/nginx

View File

@ -15,5 +15,5 @@ limitations under the License.
*/ */
// Package registrytest provides tests for Registry implementations // Package registrytest provides tests for Registry implementations
// for storing Minions, Pods, Schedulers and Services. // for storing Nodes, Pods, Schedulers and Services.
package registrytest // import "k8s.io/kubernetes/pkg/registry/registrytest" package registrytest // import "k8s.io/kubernetes/pkg/registry/registrytest"

View File

@ -161,7 +161,7 @@ func TestBadJSONRejection(t *testing.T) {
t.Errorf("Did not reject despite use of unknown type: %s", badJSONUnknownType) t.Errorf("Did not reject despite use of unknown type: %s", badJSONUnknownType)
} }
/*badJSONKindMismatch := []byte(`{"kind": "Pod"}`) /*badJSONKindMismatch := []byte(`{"kind": "Pod"}`)
if err2 := DecodeInto(badJSONKindMismatch, &Minion{}); err2 == nil { if err2 := DecodeInto(badJSONKindMismatch, &Node{}); err2 == nil {
t.Errorf("Kind is set but doesn't match the object type: %s", badJSONKindMismatch) t.Errorf("Kind is set but doesn't match the object type: %s", badJSONKindMismatch)
}*/ }*/
} }

View File

@ -38,20 +38,20 @@ func TestProxyTransport(t *testing.T) {
testTransport := &Transport{ testTransport := &Transport{
Scheme: "http", Scheme: "http",
Host: "foo.com", Host: "foo.com",
PathPrepend: "/proxy/minion/minion1:10250", PathPrepend: "/proxy/node/node1:10250",
} }
testTransport2 := &Transport{ testTransport2 := &Transport{
Scheme: "https", Scheme: "https",
Host: "foo.com", Host: "foo.com",
PathPrepend: "/proxy/minion/minion1:8080", PathPrepend: "/proxy/node/node1:8080",
} }
emptyHostTransport := &Transport{ emptyHostTransport := &Transport{
Scheme: "https", Scheme: "https",
PathPrepend: "/proxy/minion/minion1:10250", PathPrepend: "/proxy/node/node1:10250",
} }
emptySchemeTransport := &Transport{ emptySchemeTransport := &Transport{
Host: "foo.com", Host: "foo.com",
PathPrepend: "/proxy/minion/minion1:10250", PathPrepend: "/proxy/node/node1:10250",
} }
type Item struct { type Item struct {
input string input string
@ -67,120 +67,120 @@ func TestProxyTransport(t *testing.T) {
table := map[string]Item{ table := map[string]Item{
"normal": { "normal": {
input: `<pre><a href="kubelet.log">kubelet.log</a><a href="/google.log">google.log</a></pre>`, input: `<pre><a href="kubelet.log">kubelet.log</a><a href="/google.log">google.log</a></pre>`,
sourceURL: "http://myminion.com/logs/log.log", sourceURL: "http://mynode.com/logs/log.log",
transport: testTransport, transport: testTransport,
output: `<pre><a href="kubelet.log">kubelet.log</a><a href="http://foo.com/proxy/minion/minion1:10250/google.log">google.log</a></pre>`, output: `<pre><a href="kubelet.log">kubelet.log</a><a href="http://foo.com/proxy/node/node1:10250/google.log">google.log</a></pre>`,
contentType: "text/html", contentType: "text/html",
forwardedURI: "/proxy/minion/minion1:10250/logs/log.log", forwardedURI: "/proxy/node/node1:10250/logs/log.log",
}, },
"full document": { "full document": {
input: `<html><header></header><body><pre><a href="kubelet.log">kubelet.log</a><a href="/google.log">google.log</a></pre></body></html>`, input: `<html><header></header><body><pre><a href="kubelet.log">kubelet.log</a><a href="/google.log">google.log</a></pre></body></html>`,
sourceURL: "http://myminion.com/logs/log.log", sourceURL: "http://mynode.com/logs/log.log",
transport: testTransport, transport: testTransport,
output: `<html><header></header><body><pre><a href="kubelet.log">kubelet.log</a><a href="http://foo.com/proxy/minion/minion1:10250/google.log">google.log</a></pre></body></html>`, output: `<html><header></header><body><pre><a href="kubelet.log">kubelet.log</a><a href="http://foo.com/proxy/node/node1:10250/google.log">google.log</a></pre></body></html>`,
contentType: "text/html", contentType: "text/html",
forwardedURI: "/proxy/minion/minion1:10250/logs/log.log", forwardedURI: "/proxy/node/node1:10250/logs/log.log",
}, },
"trailing slash": { "trailing slash": {
input: `<pre><a href="kubelet.log">kubelet.log</a><a href="/google.log/">google.log</a></pre>`, input: `<pre><a href="kubelet.log">kubelet.log</a><a href="/google.log/">google.log</a></pre>`,
sourceURL: "http://myminion.com/logs/log.log", sourceURL: "http://mynode.com/logs/log.log",
transport: testTransport, transport: testTransport,
output: `<pre><a href="kubelet.log">kubelet.log</a><a href="http://foo.com/proxy/minion/minion1:10250/google.log/">google.log</a></pre>`, output: `<pre><a href="kubelet.log">kubelet.log</a><a href="http://foo.com/proxy/node/node1:10250/google.log/">google.log</a></pre>`,
contentType: "text/html", contentType: "text/html",
forwardedURI: "/proxy/minion/minion1:10250/logs/log.log", forwardedURI: "/proxy/node/node1:10250/logs/log.log",
}, },
"content-type charset": { "content-type charset": {
input: `<pre><a href="kubelet.log">kubelet.log</a><a href="/google.log">google.log</a></pre>`, input: `<pre><a href="kubelet.log">kubelet.log</a><a href="/google.log">google.log</a></pre>`,
sourceURL: "http://myminion.com/logs/log.log", sourceURL: "http://mynode.com/logs/log.log",
transport: testTransport, transport: testTransport,
output: `<pre><a href="kubelet.log">kubelet.log</a><a href="http://foo.com/proxy/minion/minion1:10250/google.log">google.log</a></pre>`, output: `<pre><a href="kubelet.log">kubelet.log</a><a href="http://foo.com/proxy/node/node1:10250/google.log">google.log</a></pre>`,
contentType: "text/html; charset=utf-8", contentType: "text/html; charset=utf-8",
forwardedURI: "/proxy/minion/minion1:10250/logs/log.log", forwardedURI: "/proxy/node/node1:10250/logs/log.log",
}, },
"content-type passthrough": { "content-type passthrough": {
input: `<pre><a href="kubelet.log">kubelet.log</a><a href="/google.log">google.log</a></pre>`, input: `<pre><a href="kubelet.log">kubelet.log</a><a href="/google.log">google.log</a></pre>`,
sourceURL: "http://myminion.com/logs/log.log", sourceURL: "http://mynode.com/logs/log.log",
transport: testTransport, transport: testTransport,
output: `<pre><a href="kubelet.log">kubelet.log</a><a href="/google.log">google.log</a></pre>`, output: `<pre><a href="kubelet.log">kubelet.log</a><a href="/google.log">google.log</a></pre>`,
contentType: "text/plain", contentType: "text/plain",
forwardedURI: "/proxy/minion/minion1:10250/logs/log.log", forwardedURI: "/proxy/node/node1:10250/logs/log.log",
}, },
"subdir": { "subdir": {
input: `<a href="kubelet.log">kubelet.log</a><a href="/google.log">google.log</a>`, input: `<a href="kubelet.log">kubelet.log</a><a href="/google.log">google.log</a>`,
sourceURL: "http://myminion.com/whatever/apt/somelog.log", sourceURL: "http://mynode.com/whatever/apt/somelog.log",
transport: testTransport2, transport: testTransport2,
output: `<a href="kubelet.log">kubelet.log</a><a href="https://foo.com/proxy/minion/minion1:8080/google.log">google.log</a>`, output: `<a href="kubelet.log">kubelet.log</a><a href="https://foo.com/proxy/node/node1:8080/google.log">google.log</a>`,
contentType: "text/html", contentType: "text/html",
forwardedURI: "/proxy/minion/minion1:8080/whatever/apt/somelog.log", forwardedURI: "/proxy/node/node1:8080/whatever/apt/somelog.log",
}, },
"image": { "image": {
input: `<pre><img src="kubernetes.jpg"/><img src="/kubernetes_abs.jpg"/></pre>`, input: `<pre><img src="kubernetes.jpg"/><img src="/kubernetes_abs.jpg"/></pre>`,
sourceURL: "http://myminion.com/", sourceURL: "http://mynode.com/",
transport: testTransport, transport: testTransport,
output: `<pre><img src="kubernetes.jpg"/><img src="http://foo.com/proxy/minion/minion1:10250/kubernetes_abs.jpg"/></pre>`, output: `<pre><img src="kubernetes.jpg"/><img src="http://foo.com/proxy/node/node1:10250/kubernetes_abs.jpg"/></pre>`,
contentType: "text/html", contentType: "text/html",
forwardedURI: "/proxy/minion/minion1:10250/", forwardedURI: "/proxy/node/node1:10250/",
}, },
"abs": { "abs": {
input: `<script src="http://google.com/kubernetes.js"/>`, input: `<script src="http://google.com/kubernetes.js"/>`,
sourceURL: "http://myminion.com/any/path/", sourceURL: "http://mynode.com/any/path/",
transport: testTransport, transport: testTransport,
output: `<script src="http://google.com/kubernetes.js"/>`, output: `<script src="http://google.com/kubernetes.js"/>`,
contentType: "text/html", contentType: "text/html",
forwardedURI: "/proxy/minion/minion1:10250/any/path/", forwardedURI: "/proxy/node/node1:10250/any/path/",
}, },
"abs but same host": { "abs but same host": {
input: `<script src="http://myminion.com/kubernetes.js"/>`, input: `<script src="http://mynode.com/kubernetes.js"/>`,
sourceURL: "http://myminion.com/any/path/", sourceURL: "http://mynode.com/any/path/",
transport: testTransport, transport: testTransport,
output: `<script src="http://foo.com/proxy/minion/minion1:10250/kubernetes.js"/>`, output: `<script src="http://foo.com/proxy/node/node1:10250/kubernetes.js"/>`,
contentType: "text/html", contentType: "text/html",
forwardedURI: "/proxy/minion/minion1:10250/any/path/", forwardedURI: "/proxy/node/node1:10250/any/path/",
}, },
"redirect rel": { "redirect rel": {
sourceURL: "http://myminion.com/redirect", sourceURL: "http://mynode.com/redirect",
transport: testTransport, transport: testTransport,
redirect: "/redirected/target/", redirect: "/redirected/target/",
redirectWant: "http://foo.com/proxy/minion/minion1:10250/redirected/target/", redirectWant: "http://foo.com/proxy/node/node1:10250/redirected/target/",
forwardedURI: "/proxy/minion/minion1:10250/redirect", forwardedURI: "/proxy/node/node1:10250/redirect",
}, },
"redirect abs same host": { "redirect abs same host": {
sourceURL: "http://myminion.com/redirect", sourceURL: "http://mynode.com/redirect",
transport: testTransport, transport: testTransport,
redirect: "http://myminion.com/redirected/target/", redirect: "http://mynode.com/redirected/target/",
redirectWant: "http://foo.com/proxy/minion/minion1:10250/redirected/target/", redirectWant: "http://foo.com/proxy/node/node1:10250/redirected/target/",
forwardedURI: "/proxy/minion/minion1:10250/redirect", forwardedURI: "/proxy/node/node1:10250/redirect",
}, },
"redirect abs other host": { "redirect abs other host": {
sourceURL: "http://myminion.com/redirect", sourceURL: "http://mynode.com/redirect",
transport: testTransport, transport: testTransport,
redirect: "http://example.com/redirected/target/", redirect: "http://example.com/redirected/target/",
redirectWant: "http://example.com/redirected/target/", redirectWant: "http://example.com/redirected/target/",
forwardedURI: "/proxy/minion/minion1:10250/redirect", forwardedURI: "/proxy/node/node1:10250/redirect",
}, },
"source contains the redirect already": { "source contains the redirect already": {
input: `<pre><a href="kubelet.log">kubelet.log</a><a href="http://foo.com/proxy/minion/minion1:10250/google.log">google.log</a></pre>`, input: `<pre><a href="kubelet.log">kubelet.log</a><a href="http://foo.com/proxy/node/node1:10250/google.log">google.log</a></pre>`,
sourceURL: "http://foo.com/logs/log.log", sourceURL: "http://foo.com/logs/log.log",
transport: testTransport, transport: testTransport,
output: `<pre><a href="kubelet.log">kubelet.log</a><a href="http://foo.com/proxy/minion/minion1:10250/google.log">google.log</a></pre>`, output: `<pre><a href="kubelet.log">kubelet.log</a><a href="http://foo.com/proxy/node/node1:10250/google.log">google.log</a></pre>`,
contentType: "text/html", contentType: "text/html",
forwardedURI: "/proxy/minion/minion1:10250/logs/log.log", forwardedURI: "/proxy/node/node1:10250/logs/log.log",
}, },
"no host": { "no host": {
input: "<html></html>", input: "<html></html>",
sourceURL: "http://myminion.com/logs/log.log", sourceURL: "http://mynode.com/logs/log.log",
transport: emptyHostTransport, transport: emptyHostTransport,
output: "<html></html>", output: "<html></html>",
contentType: "text/html", contentType: "text/html",
forwardedURI: "/proxy/minion/minion1:10250/logs/log.log", forwardedURI: "/proxy/node/node1:10250/logs/log.log",
}, },
"no scheme": { "no scheme": {
input: "<html></html>", input: "<html></html>",
sourceURL: "http://myminion.com/logs/log.log", sourceURL: "http://mynode.com/logs/log.log",
transport: emptySchemeTransport, transport: emptySchemeTransport,
output: "<html></html>", output: "<html></html>",
contentType: "text/html", contentType: "text/html",
forwardedURI: "/proxy/minion/minion1:10250/logs/log.log", forwardedURI: "/proxy/node/node1:10250/logs/log.log",
}, },
} }

View File

@ -1200,7 +1200,7 @@ message NamespaceStatus {
optional string phase = 1; optional string phase = 1;
} }
// Node is a worker node in Kubernetes, formerly known as minion. // Node is a worker node in Kubernetes.
// Each node will have a unique identifier in the cache (i.e. in etcd). // Each node will have a unique identifier in the cache (i.e. in etcd).
message Node { message Node {
// Standard object's metadata. // Standard object's metadata.

View File

@ -2666,7 +2666,7 @@ type ResourceList map[ResourceName]resource.Quantity
// +genclient=true // +genclient=true
// +nonNamespaced=true // +nonNamespaced=true
// Node is a worker node in Kubernetes, formerly known as minion. // Node is a worker node in Kubernetes.
// Each node will have a unique identifier in the cache (i.e. in etcd). // Each node will have a unique identifier in the cache (i.e. in etcd).
type Node struct { type Node struct {
unversioned.TypeMeta `json:",inline"` unversioned.TypeMeta `json:",inline"`

View File

@ -794,7 +794,7 @@ func (NamespaceStatus) SwaggerDoc() map[string]string {
} }
var map_Node = map[string]string{ var map_Node = map[string]string{
"": "Node is a worker node in Kubernetes, formerly known as minion. Each node will have a unique identifier in the cache (i.e. in etcd).", "": "Node is a worker node in Kubernetes.",
"metadata": "Standard object's metadata. More info: http://releases.k8s.io/HEAD/docs/devel/api-conventions.md#metadata", "metadata": "Standard object's metadata. More info: http://releases.k8s.io/HEAD/docs/devel/api-conventions.md#metadata",
"spec": "Spec defines the behavior of a node. http://releases.k8s.io/HEAD/docs/devel/api-conventions.md#spec-and-status", "spec": "Spec defines the behavior of a node. http://releases.k8s.io/HEAD/docs/devel/api-conventions.md#spec-and-status",
"status": "Most recently observed status of the node. Populated by the system. Read-only. More info: http://releases.k8s.io/HEAD/docs/devel/api-conventions.md#spec-and-status", "status": "Most recently observed status of the node. Populated by the system. Read-only. More info: http://releases.k8s.io/HEAD/docs/devel/api-conventions.md#spec-and-status",

View File

@ -706,7 +706,7 @@ func testRollbackDeployment(f *framework.Framework) {
deploymentStrategyType := extensions.RollingUpdateDeploymentStrategyType deploymentStrategyType := extensions.RollingUpdateDeploymentStrategyType
framework.Logf("Creating deployment %s", deploymentName) framework.Logf("Creating deployment %s", deploymentName)
d := newDeployment(deploymentName, deploymentReplicas, deploymentPodLabels, deploymentImageName, deploymentImage, deploymentStrategyType, nil) d := newDeployment(deploymentName, deploymentReplicas, deploymentPodLabels, deploymentImageName, deploymentImage, deploymentStrategyType, nil)
createAnnotation := map[string]string{"action": "create", "author": "minion"} createAnnotation := map[string]string{"action": "create", "author": "node"}
d.Annotations = createAnnotation d.Annotations = createAnnotation
deploy, err := c.Extensions().Deployments(ns).Create(d) deploy, err := c.Extensions().Deployments(ns).Create(d)
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())

View File

@ -488,7 +488,7 @@ type ResourceUsagePerNode map[string]ResourceUsagePerContainer
func formatResourceUsageStats(nodeName string, containerStats ResourceUsagePerContainer) string { func formatResourceUsageStats(nodeName string, containerStats ResourceUsagePerContainer) string {
// Example output: // Example output:
// //
// Resource usage for node "e2e-test-foo-minion-abcde": // Resource usage for node "e2e-test-foo-node-abcde":
// container cpu(cores) memory(MB) // container cpu(cores) memory(MB)
// "/" 0.363 2942.09 // "/" 0.363 2942.09
// "/docker-daemon" 0.088 521.80 // "/docker-daemon" 0.088 521.80
@ -794,7 +794,7 @@ type NodesCPUSummary map[string]ContainersCPUSummary
func (r *ResourceMonitor) FormatCPUSummary(summary NodesCPUSummary) string { func (r *ResourceMonitor) FormatCPUSummary(summary NodesCPUSummary) string {
// Example output for a node (the percentiles may differ): // Example output for a node (the percentiles may differ):
// CPU usage of containers on node "e2e-test-foo-minion-0vj7": // CPU usage of containers on node "e2e-test-foo-node-0vj7":
// container 5th% 50th% 90th% 95th% // container 5th% 50th% 90th% 95th%
// "/" 0.051 0.159 0.387 0.455 // "/" 0.051 0.159 0.387 0.455
// "/runtime 0.000 0.000 0.146 0.166 // "/runtime 0.000 0.000 0.146 0.166

View File

@ -237,7 +237,7 @@ func (r *ResourceCollector) GetBasicCPUStats(containerName string) map[float64]f
func formatResourceUsageStats(containerStats framework.ResourceUsagePerContainer) string { func formatResourceUsageStats(containerStats framework.ResourceUsagePerContainer) string {
// Example output: // Example output:
// //
// Resource usage for node "e2e-test-foo-minion-abcde": // Resource usage for node "e2e-test-foo-node-abcde":
// container cpu(cores) memory(MB) // container cpu(cores) memory(MB)
// "/" 0.363 2942.09 // "/" 0.363 2942.09
// "/docker-daemon" 0.088 521.80 // "/docker-daemon" 0.088 521.80
@ -255,7 +255,7 @@ func formatResourceUsageStats(containerStats framework.ResourceUsagePerContainer
func formatCPUSummary(summary framework.ContainersCPUSummary) string { func formatCPUSummary(summary framework.ContainersCPUSummary) string {
// Example output for a node (the percentiles may differ): // Example output for a node (the percentiles may differ):
// CPU usage of containers on node "e2e-test-foo-minion-0vj7": // CPU usage of containers on node "e2e-test-foo-node-0vj7":
// container 5th% 50th% 90th% 95th% // container 5th% 50th% 90th% 95th%
// "/" 0.051 0.159 0.387 0.455 // "/" 0.051 0.159 0.387 0.455
// "/runtime 0.000 0.000 0.146 0.166 // "/runtime 0.000 0.000 0.146 0.166

View File

@ -15,22 +15,22 @@ Here is some representative output.
$ ./serve_hostnames $ ./serve_hostnames
I0326 14:21:04.179893 11434 serve_hostnames.go:60] Starting serve_hostnames soak test with queries=10 and podsPerNode=1 upTo=1 I0326 14:21:04.179893 11434 serve_hostnames.go:60] Starting serve_hostnames soak test with queries=10 and podsPerNode=1 upTo=1
I0326 14:21:04.507252 11434 serve_hostnames.go:85] Nodes found on this cluster: I0326 14:21:04.507252 11434 serve_hostnames.go:85] Nodes found on this cluster:
I0326 14:21:04.507282 11434 serve_hostnames.go:87] 0: kubernetes-minion-5h4m.c.kubernetes-satnam.internal I0326 14:21:04.507282 11434 serve_hostnames.go:87] 0: kubernetes-node-5h4m.c.kubernetes-satnam.internal
I0326 14:21:04.507297 11434 serve_hostnames.go:87] 1: kubernetes-minion-9i4n.c.kubernetes-satnam.internal I0326 14:21:04.507297 11434 serve_hostnames.go:87] 1: kubernetes-node-9i4n.c.kubernetes-satnam.internal
I0326 14:21:04.507309 11434 serve_hostnames.go:87] 2: kubernetes-minion-d0yo.c.kubernetes-satnam.internal I0326 14:21:04.507309 11434 serve_hostnames.go:87] 2: kubernetes-node-d0yo.c.kubernetes-satnam.internal
I0326 14:21:04.507320 11434 serve_hostnames.go:87] 3: kubernetes-minion-jay1.c.kubernetes-satnam.internal I0326 14:21:04.507320 11434 serve_hostnames.go:87] 3: kubernetes-node-jay1.c.kubernetes-satnam.internal
I0326 14:21:04.507347 11434 serve_hostnames.go:95] Using namespace serve-hostnames-8145 for this test. I0326 14:21:04.507347 11434 serve_hostnames.go:95] Using namespace serve-hostnames-8145 for this test.
I0326 14:21:04.507363 11434 serve_hostnames.go:98] Creating service serve-hostnames-8145/serve-hostnames I0326 14:21:04.507363 11434 serve_hostnames.go:98] Creating service serve-hostnames-8145/serve-hostnames
I0326 14:21:04.559849 11434 serve_hostnames.go:148] Creating pod serve-hostnames-8145/serve-hostname-0-0 on node kubernetes-minion-5h4m.c.kubernetes-satnam.internal I0326 14:21:04.559849 11434 serve_hostnames.go:148] Creating pod serve-hostnames-8145/serve-hostname-0-0 on node kubernetes-node-5h4m.c.kubernetes-satnam.internal
I0326 14:21:04.605603 11434 serve_hostnames.go:148] Creating pod serve-hostnames-8145/serve-hostname-1-0 on node kubernetes-minion-9i4n.c.kubernetes-satnam.internal I0326 14:21:04.605603 11434 serve_hostnames.go:148] Creating pod serve-hostnames-8145/serve-hostname-1-0 on node kubernetes-node-9i4n.c.kubernetes-satnam.internal
I0326 14:21:04.662099 11434 serve_hostnames.go:148] Creating pod serve-hostnames-8145/serve-hostname-2-0 on node kubernetes-minion-d0yo.c.kubernetes-satnam.internal I0326 14:21:04.662099 11434 serve_hostnames.go:148] Creating pod serve-hostnames-8145/serve-hostname-2-0 on node kubernetes-node-d0yo.c.kubernetes-satnam.internal
I0326 14:21:04.707179 11434 serve_hostnames.go:148] Creating pod serve-hostnames-8145/serve-hostname-3-0 on node kubernetes-minion-jay1.c.kubernetes-satnam.internal I0326 14:21:04.707179 11434 serve_hostnames.go:148] Creating pod serve-hostnames-8145/serve-hostname-3-0 on node kubernetes-node-jay1.c.kubernetes-satnam.internal
I0326 14:21:04.757646 11434 serve_hostnames.go:194] Waiting for the serve-hostname pods to be ready I0326 14:21:04.757646 11434 serve_hostnames.go:194] Waiting for the serve-hostname pods to be ready
I0326 14:23:31.125188 11434 serve_hostnames.go:211] serve-hostnames-8145/serve-hostname-0-0 is running I0326 14:23:31.125188 11434 serve_hostnames.go:211] serve-hostnames-8145/serve-hostname-0-0 is running
I0326 14:23:31.165984 11434 serve_hostnames.go:211] serve-hostnames-8145/serve-hostname-1-0 is running I0326 14:23:31.165984 11434 serve_hostnames.go:211] serve-hostnames-8145/serve-hostname-1-0 is running
I0326 14:25:22.213751 11434 serve_hostnames.go:211] serve-hostnames-8145/serve-hostname-2-0 is running I0326 14:25:22.213751 11434 serve_hostnames.go:211] serve-hostnames-8145/serve-hostname-2-0 is running
I0326 14:25:37.387257 11434 serve_hostnames.go:211] serve-hostnames-8145/serve-hostname-3-0 is running I0326 14:25:37.387257 11434 serve_hostnames.go:211] serve-hostnames-8145/serve-hostname-3-0 is running
W0326 14:25:39.243813 11434 serve_hostnames.go:265] No response from pod serve-hostname-3-0 on node kubernetes-minion-jay1.c.kubernetes-satnam.internal at iteration 0 W0326 14:25:39.243813 11434 serve_hostnames.go:265] No response from pod serve-hostname-3-0 on node kubernetes-node-jay1.c.kubernetes-satnam.internal at iteration 0
I0326 14:25:39.243844 11434 serve_hostnames.go:269] Iteration 0 took 1.814483599s for 40 queries (22.04 QPS) I0326 14:25:39.243844 11434 serve_hostnames.go:269] Iteration 0 took 1.814483599s for 40 queries (22.04 QPS)
I0326 14:25:39.243871 11434 serve_hostnames.go:182] Cleaning up pods I0326 14:25:39.243871 11434 serve_hostnames.go:182] Cleaning up pods
I0326 14:25:39.434619 11434 serve_hostnames.go:130] Cleaning up service serve-hostnames-8145/server-hostnames I0326 14:25:39.434619 11434 serve_hostnames.go:130] Cleaning up service serve-hostnames-8145/server-hostnames
@ -45,20 +45,20 @@ The number of iterations to perform for issuing queries can be changed from the
$ ./serve_hostnames --up_to=3 --pods_per_node=2 $ ./serve_hostnames --up_to=3 --pods_per_node=2
I0326 14:27:27.584378 11808 serve_hostnames.go:60] Starting serve_hostnames soak test with queries=10 and podsPerNode=2 upTo=3 I0326 14:27:27.584378 11808 serve_hostnames.go:60] Starting serve_hostnames soak test with queries=10 and podsPerNode=2 upTo=3
I0326 14:27:27.913713 11808 serve_hostnames.go:85] Nodes found on this cluster: I0326 14:27:27.913713 11808 serve_hostnames.go:85] Nodes found on this cluster:
I0326 14:27:27.913774 11808 serve_hostnames.go:87] 0: kubernetes-minion-5h4m.c.kubernetes-satnam.internal I0326 14:27:27.913774 11808 serve_hostnames.go:87] 0: kubernetes-node-5h4m.c.kubernetes-satnam.internal
I0326 14:27:27.913800 11808 serve_hostnames.go:87] 1: kubernetes-minion-9i4n.c.kubernetes-satnam.internal I0326 14:27:27.913800 11808 serve_hostnames.go:87] 1: kubernetes-node-9i4n.c.kubernetes-satnam.internal
I0326 14:27:27.913825 11808 serve_hostnames.go:87] 2: kubernetes-minion-d0yo.c.kubernetes-satnam.internal I0326 14:27:27.913825 11808 serve_hostnames.go:87] 2: kubernetes-node-d0yo.c.kubernetes-satnam.internal
I0326 14:27:27.913846 11808 serve_hostnames.go:87] 3: kubernetes-minion-jay1.c.kubernetes-satnam.internal I0326 14:27:27.913846 11808 serve_hostnames.go:87] 3: kubernetes-node-jay1.c.kubernetes-satnam.internal
I0326 14:27:27.913904 11808 serve_hostnames.go:95] Using namespace serve-hostnames-4997 for this test. I0326 14:27:27.913904 11808 serve_hostnames.go:95] Using namespace serve-hostnames-4997 for this test.
I0326 14:27:27.913931 11808 serve_hostnames.go:98] Creating service serve-hostnames-4997/serve-hostnames I0326 14:27:27.913931 11808 serve_hostnames.go:98] Creating service serve-hostnames-4997/serve-hostnames
I0326 14:27:27.969083 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-0-0 on node kubernetes-minion-5h4m.c.kubernetes-satnam.internal I0326 14:27:27.969083 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-0-0 on node kubernetes-node-5h4m.c.kubernetes-satnam.internal
I0326 14:27:28.020133 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-0-1 on node kubernetes-minion-5h4m.c.kubernetes-satnam.internal I0326 14:27:28.020133 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-0-1 on node kubernetes-node-5h4m.c.kubernetes-satnam.internal
I0326 14:27:28.070054 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-1-0 on node kubernetes-minion-9i4n.c.kubernetes-satnam.internal I0326 14:27:28.070054 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-1-0 on node kubernetes-node-9i4n.c.kubernetes-satnam.internal
I0326 14:27:28.118641 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-1-1 on node kubernetes-minion-9i4n.c.kubernetes-satnam.internal I0326 14:27:28.118641 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-1-1 on node kubernetes-node-9i4n.c.kubernetes-satnam.internal
I0326 14:27:28.168786 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-2-0 on node kubernetes-minion-d0yo.c.kubernetes-satnam.internal I0326 14:27:28.168786 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-2-0 on node kubernetes-node-d0yo.c.kubernetes-satnam.internal
I0326 14:27:28.214730 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-2-1 on node kubernetes-minion-d0yo.c.kubernetes-satnam.internal I0326 14:27:28.214730 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-2-1 on node kubernetes-node-d0yo.c.kubernetes-satnam.internal
I0326 14:27:28.261685 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-3-0 on node kubernetes-minion-jay1.c.kubernetes-satnam.internal I0326 14:27:28.261685 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-3-0 on node kubernetes-node-jay1.c.kubernetes-satnam.internal
I0326 14:27:28.320224 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-3-1 on node kubernetes-minion-jay1.c.kubernetes-satnam.internal I0326 14:27:28.320224 11808 serve_hostnames.go:148] Creating pod serve-hostnames-4997/serve-hostname-3-1 on node kubernetes-node-jay1.c.kubernetes-satnam.internal
I0326 14:27:28.387007 11808 serve_hostnames.go:194] Waiting for the serve-hostname pods to be ready I0326 14:27:28.387007 11808 serve_hostnames.go:194] Waiting for the serve-hostname pods to be ready
I0326 14:28:28.969149 11808 serve_hostnames.go:211] serve-hostnames-4997/serve-hostname-0-0 is running I0326 14:28:28.969149 11808 serve_hostnames.go:211] serve-hostnames-4997/serve-hostname-0-0 is running
I0326 14:28:29.010376 11808 serve_hostnames.go:211] serve-hostnames-4997/serve-hostname-0-1 is running I0326 14:28:29.010376 11808 serve_hostnames.go:211] serve-hostnames-4997/serve-hostname-0-1 is running
@ -68,11 +68,11 @@ I0326 14:30:00.850461 11808 serve_hostnames.go:211] serve-hostnames-4997/serve
I0326 14:30:00.891559 11808 serve_hostnames.go:211] serve-hostnames-4997/serve-hostname-2-1 is running I0326 14:30:00.891559 11808 serve_hostnames.go:211] serve-hostnames-4997/serve-hostname-2-1 is running
I0326 14:30:00.932829 11808 serve_hostnames.go:211] serve-hostnames-4997/serve-hostname-3-0 is running I0326 14:30:00.932829 11808 serve_hostnames.go:211] serve-hostnames-4997/serve-hostname-3-0 is running
I0326 14:30:00.973941 11808 serve_hostnames.go:211] serve-hostnames-4997/serve-hostname-3-1 is running I0326 14:30:00.973941 11808 serve_hostnames.go:211] serve-hostnames-4997/serve-hostname-3-1 is running
W0326 14:30:04.726582 11808 serve_hostnames.go:265] No response from pod serve-hostname-2-0 on node kubernetes-minion-d0yo.c.kubernetes-satnam.internal at iteration 0 W0326 14:30:04.726582 11808 serve_hostnames.go:265] No response from pod serve-hostname-2-0 on node kubernetes-node-d0yo.c.kubernetes-satnam.internal at iteration 0
W0326 14:30:04.726658 11808 serve_hostnames.go:265] No response from pod serve-hostname-2-1 on node kubernetes-minion-d0yo.c.kubernetes-satnam.internal at iteration 0 W0326 14:30:04.726658 11808 serve_hostnames.go:265] No response from pod serve-hostname-2-1 on node kubernetes-node-d0yo.c.kubernetes-satnam.internal at iteration 0
I0326 14:30:04.726696 11808 serve_hostnames.go:269] Iteration 0 took 3.711080213s for 80 queries (21.56 QPS) I0326 14:30:04.726696 11808 serve_hostnames.go:269] Iteration 0 took 3.711080213s for 80 queries (21.56 QPS)
W0326 14:30:08.267297 11808 serve_hostnames.go:265] No response from pod serve-hostname-2-0 on node kubernetes-minion-d0yo.c.kubernetes-satnam.internal at iteration 1 W0326 14:30:08.267297 11808 serve_hostnames.go:265] No response from pod serve-hostname-2-0 on node kubernetes-node-d0yo.c.kubernetes-satnam.internal at iteration 1
W0326 14:30:08.267365 11808 serve_hostnames.go:265] No response from pod serve-hostname-2-1 on node kubernetes-minion-d0yo.c.kubernetes-satnam.internal at iteration 1 W0326 14:30:08.267365 11808 serve_hostnames.go:265] No response from pod serve-hostname-2-1 on node kubernetes-node-d0yo.c.kubernetes-satnam.internal at iteration 1
I0326 14:30:08.267404 11808 serve_hostnames.go:269] Iteration 1 took 3.540635303s for 80 queries (22.59 QPS) I0326 14:30:08.267404 11808 serve_hostnames.go:269] Iteration 1 took 3.540635303s for 80 queries (22.59 QPS)
I0326 14:30:11.971349 11808 serve_hostnames.go:269] Iteration 2 took 3.703884372s for 80 queries (21.60 QPS) I0326 14:30:11.971349 11808 serve_hostnames.go:269] Iteration 2 took 3.703884372s for 80 queries (21.60 QPS)
I0326 14:30:11.971425 11808 serve_hostnames.go:182] Cleaning up pods I0326 14:30:11.971425 11808 serve_hostnames.go:182] Cleaning up pods
@ -98,20 +98,20 @@ pod on node 3 returned 12 responses and the pod on node 2 did not respond at all
$ ./serve_hostnames --v=4 $ ./serve_hostnames --v=4
I0326 14:33:26.020917 12099 serve_hostnames.go:60] Starting serve_hostnames soak test with queries=10 and podsPerNode=1 upTo=1 I0326 14:33:26.020917 12099 serve_hostnames.go:60] Starting serve_hostnames soak test with queries=10 and podsPerNode=1 upTo=1
I0326 14:33:26.365201 12099 serve_hostnames.go:85] Nodes found on this cluster: I0326 14:33:26.365201 12099 serve_hostnames.go:85] Nodes found on this cluster:
I0326 14:33:26.365260 12099 serve_hostnames.go:87] 0: kubernetes-minion-5h4m.c.kubernetes-satnam.internal I0326 14:33:26.365260 12099 serve_hostnames.go:87] 0: kubernetes-node-5h4m.c.kubernetes-satnam.internal
I0326 14:33:26.365288 12099 serve_hostnames.go:87] 1: kubernetes-minion-9i4n.c.kubernetes-satnam.internal I0326 14:33:26.365288 12099 serve_hostnames.go:87] 1: kubernetes-node-9i4n.c.kubernetes-satnam.internal
I0326 14:33:26.365313 12099 serve_hostnames.go:87] 2: kubernetes-minion-d0yo.c.kubernetes-satnam.internal I0326 14:33:26.365313 12099 serve_hostnames.go:87] 2: kubernetes-node-d0yo.c.kubernetes-satnam.internal
I0326 14:33:26.365334 12099 serve_hostnames.go:87] 3: kubernetes-minion-jay1.c.kubernetes-satnam.internal I0326 14:33:26.365334 12099 serve_hostnames.go:87] 3: kubernetes-node-jay1.c.kubernetes-satnam.internal
I0326 14:33:26.365392 12099 serve_hostnames.go:95] Using namespace serve-hostnames-1631 for this test. I0326 14:33:26.365392 12099 serve_hostnames.go:95] Using namespace serve-hostnames-1631 for this test.
I0326 14:33:26.365419 12099 serve_hostnames.go:98] Creating service serve-hostnames-1631/serve-hostnames I0326 14:33:26.365419 12099 serve_hostnames.go:98] Creating service serve-hostnames-1631/serve-hostnames
I0326 14:33:26.423927 12099 serve_hostnames.go:118] Service create serve-hostnames-1631/server-hostnames took 58.473361ms I0326 14:33:26.423927 12099 serve_hostnames.go:118] Service create serve-hostnames-1631/server-hostnames took 58.473361ms
I0326 14:33:26.423981 12099 serve_hostnames.go:148] Creating pod serve-hostnames-1631/serve-hostname-0-0 on node kubernetes-minion-5h4m.c.kubernetes-satnam.internal I0326 14:33:26.423981 12099 serve_hostnames.go:148] Creating pod serve-hostnames-1631/serve-hostname-0-0 on node kubernetes-node-5h4m.c.kubernetes-satnam.internal
I0326 14:33:26.480185 12099 serve_hostnames.go:168] Pod create serve-hostnames-1631/serve-hostname-0-0 request took 56.178906ms I0326 14:33:26.480185 12099 serve_hostnames.go:168] Pod create serve-hostnames-1631/serve-hostname-0-0 request took 56.178906ms
I0326 14:33:26.480271 12099 serve_hostnames.go:148] Creating pod serve-hostnames-1631/serve-hostname-1-0 on node kubernetes-minion-9i4n.c.kubernetes-satnam.internal I0326 14:33:26.480271 12099 serve_hostnames.go:148] Creating pod serve-hostnames-1631/serve-hostname-1-0 on node kubernetes-node-9i4n.c.kubernetes-satnam.internal
I0326 14:33:26.534300 12099 serve_hostnames.go:168] Pod create serve-hostnames-1631/serve-hostname-1-0 request took 53.981761ms I0326 14:33:26.534300 12099 serve_hostnames.go:168] Pod create serve-hostnames-1631/serve-hostname-1-0 request took 53.981761ms
I0326 14:33:26.534396 12099 serve_hostnames.go:148] Creating pod serve-hostnames-1631/serve-hostname-2-0 on node kubernetes-minion-d0yo.c.kubernetes-satnam.internal I0326 14:33:26.534396 12099 serve_hostnames.go:148] Creating pod serve-hostnames-1631/serve-hostname-2-0 on node kubernetes-node-d0yo.c.kubernetes-satnam.internal
I0326 14:33:26.590188 12099 serve_hostnames.go:168] Pod create serve-hostnames-1631/serve-hostname-2-0 request took 55.752115ms I0326 14:33:26.590188 12099 serve_hostnames.go:168] Pod create serve-hostnames-1631/serve-hostname-2-0 request took 55.752115ms
I0326 14:33:26.590222 12099 serve_hostnames.go:148] Creating pod serve-hostnames-1631/serve-hostname-3-0 on node kubernetes-minion-jay1.c.kubernetes-satnam.internal I0326 14:33:26.590222 12099 serve_hostnames.go:148] Creating pod serve-hostnames-1631/serve-hostname-3-0 on node kubernetes-node-jay1.c.kubernetes-satnam.internal
I0326 14:33:26.650024 12099 serve_hostnames.go:168] Pod create serve-hostnames-1631/serve-hostname-3-0 request took 59.781614ms I0326 14:33:26.650024 12099 serve_hostnames.go:168] Pod create serve-hostnames-1631/serve-hostname-3-0 request took 59.781614ms
I0326 14:33:26.650083 12099 serve_hostnames.go:194] Waiting for the serve-hostname pods to be ready I0326 14:33:26.650083 12099 serve_hostnames.go:194] Waiting for the serve-hostname pods to be ready
I0326 14:33:32.776651 12099 serve_hostnames.go:211] serve-hostnames-1631/serve-hostname-0-0 is running I0326 14:33:32.776651 12099 serve_hostnames.go:211] serve-hostnames-1631/serve-hostname-0-0 is running
@ -161,7 +161,7 @@ I0326 14:35:05.607126 12099 serve_hostnames.go:249] Proxy call in namespace se
I0326 14:35:05.607164 12099 serve_hostnames.go:258] serve-hostname-3-0: 12 I0326 14:35:05.607164 12099 serve_hostnames.go:258] serve-hostname-3-0: 12
I0326 14:35:05.607176 12099 serve_hostnames.go:258] serve-hostname-1-0: 10 I0326 14:35:05.607176 12099 serve_hostnames.go:258] serve-hostname-1-0: 10
I0326 14:35:05.607186 12099 serve_hostnames.go:258] serve-hostname-0-0: 18 I0326 14:35:05.607186 12099 serve_hostnames.go:258] serve-hostname-0-0: 18
W0326 14:35:05.607199 12099 serve_hostnames.go:265] No response from pod serve-hostname-2-0 on node kubernetes-minion-d0yo.c.kubernetes-satnam.internal at iteration 0 W0326 14:35:05.607199 12099 serve_hostnames.go:265] No response from pod serve-hostname-2-0 on node kubernetes-node-d0yo.c.kubernetes-satnam.internal at iteration 0
I0326 14:35:05.607211 12099 serve_hostnames.go:269] Iteration 0 took 1.774856469s for 40 queries (22.54 QPS) I0326 14:35:05.607211 12099 serve_hostnames.go:269] Iteration 0 took 1.774856469s for 40 queries (22.54 QPS)
I0326 14:35:05.607236 12099 serve_hostnames.go:182] Cleaning up pods I0326 14:35:05.607236 12099 serve_hostnames.go:182] Cleaning up pods
I0326 14:35:05.797893 12099 serve_hostnames.go:130] Cleaning up service serve-hostnames-1631/server-hostnames I0326 14:35:05.797893 12099 serve_hostnames.go:130] Cleaning up service serve-hostnames-1631/server-hostnames