![]() This commit brings two main changes, notably: Two new options that can be set as environment variables - DOCKER_OPTS: any arbitrary set of docker options. Example: --tlsverify - DOCKER_NATIVE: This forces the use of the native docker available. This is most useful if you're on OS X and do not want to use boot2docker. Now uses 'docker cp' instead of tar piping to transfer files. This currently must be done by copying the binaries off of the docker volume and into a local filesystem (/tmp) before a docker cp is done. This workaround will no longer be necessary after bug fix https://github.com/docker/docker/pull/8509 makes it into stable. This was necessary because the tar | tar method was creating corrupted archives on OS X even with the < /dev/null workaround. |
||
---|---|---|
.. | ||
build-image | ||
hello-kubernetes | ||
pause | ||
run-images | ||
README.md | ||
common.sh | ||
copy-output.sh | ||
json-extractor.py | ||
make-build-image.sh | ||
make-clean.sh | ||
make-run-images.sh | ||
mark-new-version.sh | ||
push-official-release.sh | ||
release.sh | ||
run.sh | ||
shell.sh |
README.md
Building Kubernetes
To build Kubernetes you need to have access to a Docker installation through either of the following methods:
Requirements
- Be running Docker. 2 options supported/tested:
- Mac OS X The best way to go is to use
boot2docker
. See instructions here. - Linux with local Docker Install Docker according to the instructions for your OS. The scripts here assume that they are using a local Docker server and that they can "reach around" docker and grab results directly from the file system.
- Have python installed. Pretty much it is installed everywhere at this point so you can probably ignore this.
- Optional For uploading your release to Google Cloud Storage, have the Google Cloud SDK installed and configured.
Overview
While it is possible to build Kubernetes using a local golang installation, we have a build process that runs in a Docker container. This simplifies initial set up and provides for a very consistent build and test environment.
There is also early support for building Docker "run" containers
Key scripts
run.sh
: Run a command in a build docker container. Common invocations:run.sh hack/build-go.sh
: Build just linux binaries in the container. Pass options and packages as necessary.run.sh hack/build-cross.sh
: Build all binaries for all platformsrun.sh hack/test-go.sh
: Run all unit testsrun.sh hack/test-integration.sh
: Run integration test
copy-output.sh
: This will copy the contents of_output/dockerized/bin
from any remote Docker container to the local_output/dockerized/bin
. Right now this is only necessary on Mac OS X withboot2docker
when your git repo isn't under/Users
.make-clean.sh
: Clean out the contents of_output/dockerized
and remove any local built container images.shell.sh
: Drop into abash
shell in a build container with a snapshot of the current repo code.release.sh
: Build everything, test it, and (optionally) upload the results to a GCS bucket.
Releasing
The release.sh
script will build a release. It will build binaries, run tests, (optionally) build runtime Docker images and then (optionally) upload all build artifacts to a GCS bucket.
The main output is a tar file: kubernetes.tar.gz
. This includes:
- Cross compiled client utilities.
- Script (
cluster/kubecfg.sh
) for picking and running the right client binary based on platform. - Examples
- Cluster deployment scripts for various clouds
- Tar file containing all server binaries
- Tar file containing salt deployment tree shared across multiple cloud deployments.
In addition, there are some other tar files that are created:
kubernetes-client-*.tar.gz
Client binaries for a specific platform.kubernetes-server-*.tar.gz
Server binaries for a specific platform.kubernetes-salt.tar.gz
The salt script/tree shared across multiple deployment scripts.
The release utilities grab a set of environment variables to modify behavior. Arguably, these should be command line flags:
Env Variable | Default | Description |
---|---|---|
KUBE_SKIP_CONFIRMATIONS |
n |
If y then no questions are asked and the scripts just continue. |
KUBE_GCS_UPLOAD_RELEASE |
n |
Upload release artifacts to GCS |
KUBE_GCS_RELEASE_BUCKET |
kubernetes-releases-${project_hash} |
The bucket to upload releases to |
KUBE_GCS_RELEASE_PREFIX |
devel |
The path under the release bucket to put releases |
KUBE_GCS_MAKE_PUBLIC |
y |
Make GCS links readable from anywhere |
KUBE_GCS_NO_CACHING |
y |
Disable HTTP caching of GCS release artifacts. By default GCS will cache public objects for up to an hour. When doing "devel" releases this can cause problems. |
KUBE_BUILD_RUN_IMAGES |
n |
Experimental Build Docker images for running most server components. |
KUBE_GCS_DOCKER_REG_PREFIX |
docker-reg |
Experimental When uploading docker images, the bucket that backs the registry. |
Basic Flow
The scripts directly under build/
are used to build and test. They will ensure that the kube-build
Docker image is built (based on build/build-image/Dockerfile
) and then execute the appropriate command in that container. If necessary (for Mac OS X), the scripts will also copy results out.
The kube-build
container image is built by first creating a "context" directory in _output/images/build-image
. It is done there instead of at the root of the Kubernetes repo to minimize the amount of data we need to package up when building the image.
Everything in build/build-image/
is meant to be run inside of the container. If it doesn't think it is running in the container it'll throw a warning. While you can run some of that stuff outside of the container, it wasn't built to do so.
When building final release tars, they are first staged into _output/release-stage
before being tar'd up and put into _output/release-tars
.
Runtime Docker Images
This support is experimental and hasn't been used yet to deploy a cluster.
The files necessarily for the release Docker images are in build/run-images/*
. All of this is staged into _output/images
similar to build-image. The base
image is used as a base for each of the specialized containers and is generally never pushed to a shared repository.
If the release script is set to upload to GCS, it'll do the following:
- Start up a local
google/docker-registry
registry that is backed by GCS. - Rename/push the runtime images to that registry.
TODOs
These are in no particular order
- Harmonize with scripts in
hack/
. How much do we support building outside of Docker and these scripts? - Get a cluster up and running with the Docker images. Perhaps start with a local cluster and move up to a GCE cluster.
- Implement (#186)[https://github.com/GoogleCloudPlatform/kubernetes/issues/186]. This will make it easier to develop Kubernetes.
- Deprecate/replace most of the stuff in the hack/
- Create an install script that'll let us do a
curl https://[URL] | bash
to get that tarball down and ensure that other dependencies (cloud SDK?) are installed and configured correctly. - Support/test Windows as a client.
- Finish support for the Dockerized runtime. Issue (#19)[https://github.com/GoogleCloudPlatform/kubernetes/issues/19]. A key issue here is to make this fast/light enough that we can use it for development workflows.
- Support uploading to the Docker index instead of the GCS bucket. This'll allow easier installs for those not running on GCE