Merge branch 'master' of github.com:GoogleCloudPlatform/kubernetes

pull/6/head
Max Forbes 2014-09-10 13:36:55 -07:00
commit 6a583ebcf5
36 changed files with 467 additions and 251 deletions

View File

@ -7,6 +7,7 @@
# clean: Clean up. # clean: Clean up.
OUT_DIR = _output OUT_DIR = _output
GODEPS_PKG_DIR = Godeps/_workspace/pkg
export GOFLAGS export GOFLAGS
@ -48,4 +49,5 @@ check test:
# make clean # make clean
clean: clean:
rm -rf $(OUT_DIR) rm -rf $(OUT_DIR)
rm -rf $(GODEPS_PKG_DIR)
.PHONY: clean .PHONY: clean

View File

@ -24,6 +24,7 @@ While the concepts and architecture in Kubernetes represent years of experience
* [Microsoft Azure](docs/getting-started-guides/azure.md) * [Microsoft Azure](docs/getting-started-guides/azure.md)
* [Rackspace](docs/getting-started-guides/rackspace.md) * [Rackspace](docs/getting-started-guides/rackspace.md)
* [Circle CI](https://circleci.com/docs/docker#google-compute-engine-and-kubernetes) * [Circle CI](https://circleci.com/docs/docker#google-compute-engine-and-kubernetes)
* [Digital Ocean](https://github.com/bketelsen/coreos-kubernetes-digitalocean)
* [kubecfg command line tool](https://github.com/GoogleCloudPlatform/kubernetes/blob/master/docs/cli.md) * [kubecfg command line tool](https://github.com/GoogleCloudPlatform/kubernetes/blob/master/docs/cli.md)
* [Kubernetes API Documentation](http://cdn.rawgit.com/GoogleCloudPlatform/kubernetes/31a0daae3627c91bc96e1f02a6344cd76e294791/api/kubernetes.html) * [Kubernetes API Documentation](http://cdn.rawgit.com/GoogleCloudPlatform/kubernetes/31a0daae3627c91bc96e1f02a6344cd76e294791/api/kubernetes.html)
* [Kubernetes Client Libraries](https://github.com/GoogleCloudPlatform/kubernetes/blob/master/docs/client-libraries.md) * [Kubernetes Client Libraries](https://github.com/GoogleCloudPlatform/kubernetes/blob/master/docs/client-libraries.md)

View File

@ -49,13 +49,21 @@ ENV GOPATH /go
ENV GOOS linux ENV GOOS linux
ENV GOARCH amd64 ENV GOARCH amd64
# Get the code coverage tool and etcd for integration tests # Get the code coverage tool and godep
RUN go get code.google.com/p/go.tools/cmd/cover github.com/coreos/etcd github.com/tools/godep RUN go get code.google.com/p/go.tools/cmd/cover github.com/tools/godep
RUN mkdir -p /go/src/github.com/coreos/etcd && \
cd /go/src/github.com/coreos/etcd && \
git clone https://github.com/coreos/etcd.git . -b v0.4.6 --depth=1 && \
go install github.com/coreos/etcd
# Mark this as a kube-build container # Mark this as a kube-build container
RUN touch /kube-build-image RUN touch /kube-build-image
WORKDIR /go/src/github.com/GoogleCloudPlatform/kubernetes WORKDIR /go/src/github.com/GoogleCloudPlatform/kubernetes
# Propagate the git tree version into the build image
ADD kube-version-defs /kube-version-defs
# Upload Kubernetes source # Upload Kubernetes source
ADD kube-source.tar.gz /go/src/github.com/GoogleCloudPlatform/kubernetes ADD kube-source.tar.gz /go/src/github.com/GoogleCloudPlatform/kubernetes

View File

@ -25,34 +25,53 @@ readonly KUBE_GO_PACKAGE=github.com/GoogleCloudPlatform/kubernetes
mkdir -p "${KUBE_TARGET}" mkdir -p "${KUBE_TARGET}"
if [[ ! -f "/kube-build-image" ]]; then if [[ ! -f "/kube-build-image" ]]; then
echo "WARNING: This script should be run in the kube-build conrtainer image!" >&2 echo "WARNING: This script should be run in the kube-build container image!" >&2
fi fi
function make-binaries() { if [[ -f "/kube-version-defs" ]]; then
readonly BINARIES=" source "/kube-version-defs"
proxy else
integration echo "WARNING: No version information provided in build image"
apiserver fi
controller-manager
kubelet function kube::build::make_binary() {
kubecfg" local -r gopkg=$1
local -r bin=${gopkg##*/}
echo "+++ Building ${bin} for ${GOOS}/${GOARCH}"
pushd "${KUBE_REPO_ROOT}" >/dev/null
godep go build -ldflags "${KUBE_LD_FLAGS-}" -o "${ARCH_TARGET}/${bin}" "${gopkg}"
popd >/dev/null
}
function kube::build::make_binaries() {
if [[ ${#targets[@]} -eq 0 ]]; then
targets=(
cmd/proxy
cmd/apiserver
cmd/controller-manager
cmd/kubelet
cmd/kubecfg
plugin/cmd/scheduler
)
fi
binaries=()
local target
for target in "${targets[@]}"; do
binaries+=("${KUBE_GO_PACKAGE}/${target}")
done
ARCH_TARGET="${KUBE_TARGET}/${GOOS}/${GOARCH}" ARCH_TARGET="${KUBE_TARGET}/${GOOS}/${GOARCH}"
mkdir -p "${ARCH_TARGET}" mkdir -p "${ARCH_TARGET}"
function make-binary() { if [[ -n "$1" ]]; then
echo "+++ Building $1 for ${GOOS}/${GOARCH}" kube::build::make_binary "$1"
godep go build \
-o "${ARCH_TARGET}/$1" \
github.com/GoogleCloudPlatform/kubernetes/cmd/$1
}
if [[ -n $1 ]]; then
make-binary $1
exit 0 exit 0
fi fi
for b in ${BINARIES}; do local b
make-binary $b for b in "${binaries[@]}"; do
kube::build::make_binary "$b"
done done
} }

View File

@ -14,10 +14,8 @@
# See the License for the specific language governing permissions and # See the License for the specific language governing permissions and
# limitations under the License. # limitations under the License.
# This and builds all go components.
set -e set -e
source $(dirname $0)/common.sh source $(dirname $0)/common.sh
make-binaries "$@" kube::build::make_binaries "$@"

View File

@ -14,22 +14,20 @@
# See the License for the specific language governing permissions and # See the License for the specific language governing permissions and
# limitations under the License. # limitations under the License.
# This and builds all go components.
set -e set -e
source $(dirname $0)/common.sh source $(dirname $0)/common.sh
readonly CROSS_BINARIES=" readonly CROSS_BINARIES=(
kubecfg ./cmd/kubecfg
" )
for platform in ${KUBE_CROSSPLATFORMS}; do for platform in ${KUBE_CROSSPLATFORMS}; do
( (
export GOOS=${platform%/*} export GOOS=${platform%/*}
export GOARCH=${platform##*/} export GOARCH=${platform##*/}
for binary in ${CROSS_BINARIES}; do for binary in "${CROSS_BINARIES[@]}"; do
make-binaries "${binary}" kube::build::make_binaries "${binary}"
done done
) )
done done

View File

@ -18,13 +18,15 @@ set -e
source $(dirname $0)/common.sh source $(dirname $0)/common.sh
ETCD_DIR="${KUBE_REPO_ROOT}/_output/etcd" kube::build::make_binaries "./cmd/integration"
readonly ETCD_DIR="${KUBE_REPO_ROOT}/_output/etcd"
mkdir -p "${ETCD_DIR}" mkdir -p "${ETCD_DIR}"
echo "+++ Running integration test" echo "+++ Running integration test"
etcd -name test -data-dir ${ETCD_DIR} > "${KUBE_REPO_ROOT}/_output/etcd.log" & etcd -name test -data-dir ${ETCD_DIR} > "${KUBE_REPO_ROOT}/_output/etcd.log" &
ETCD_PID=$! readonly ETCD_PID=$!
sleep 5 sleep 5

View File

@ -14,7 +14,6 @@
# See the License for the specific language governing permissions and # See the License for the specific language governing permissions and
# limitations under the License. # limitations under the License.
set -e set -e
source $(dirname $0)/common.sh source $(dirname $0)/common.sh

View File

@ -17,6 +17,9 @@
# Common utilties, variables and checks for all build scripts. # Common utilties, variables and checks for all build scripts.
cd $(dirname "${BASH_SOURCE}")/.. cd $(dirname "${BASH_SOURCE}")/..
source hack/config-go.sh
readonly KUBE_REPO_ROOT="${PWD}" readonly KUBE_REPO_ROOT="${PWD}"
readonly KUBE_GIT_BRANCH=$(git rev-parse --abbrev-ref HEAD 2>/dev/null) readonly KUBE_GIT_BRANCH=$(git rev-parse --abbrev-ref HEAD 2>/dev/null)
@ -38,11 +41,13 @@ readonly DOCKER_CONTAINER_NAME=kube-build
readonly DOCKER_MOUNT="-v ${LOCAL_OUTPUT_DIR}:${REMOTE_OUTPUT_DIR}" readonly DOCKER_MOUNT="-v ${LOCAL_OUTPUT_DIR}:${REMOTE_OUTPUT_DIR}"
readonly KUBE_RUN_IMAGE_BASE="kubernetes" readonly KUBE_RUN_IMAGE_BASE="kubernetes"
readonly KUBE_RUN_BINARIES=" readonly KUBE_RUN_BINARIES=(
apiserver apiserver
controller-manager controller-manager
proxy proxy
" scheduler
)
# This is where the final release artifacts are created locally # This is where the final release artifacts are created locally
readonly RELEASE_DIR="${KUBE_REPO_ROOT}/_output/release" readonly RELEASE_DIR="${KUBE_REPO_ROOT}/_output/release"
@ -51,7 +56,7 @@ readonly RELEASE_DIR="${KUBE_REPO_ROOT}/_output/release"
# Basic setup functions # Basic setup functions
# Verify that the right utilities and such are installed for building Kube. # Verify that the right utilities and such are installed for building Kube.
function verify-prereqs() { function kube::build::verify_prereqs() {
if [[ -z "$(which docker)" ]]; then if [[ -z "$(which docker)" ]]; then
echo "Can't find 'docker' in PATH, please fix and retry." >&2 echo "Can't find 'docker' in PATH, please fix and retry." >&2
echo "See https://docs.docker.com/installation/#installation for installation instructions." >&2 echo "See https://docs.docker.com/installation/#installation for installation instructions." >&2
@ -82,42 +87,13 @@ function verify-prereqs() {
fi fi
} }
# Verify things are set up for uploading to GCS
function verify-gcs-prereqs() {
if [[ -z "$(which gsutil)" || -z "$(which gcloud)" ]]; then
echo "Releasing Kubernetes requires gsutil and gcloud. Please download,"
echo "install and authorize through the Google Cloud SDK: "
echo
echo " https://developers.google.com/cloud/sdk/"
return 1
fi
FIND_ACCOUNT="gcloud auth list 2>/dev/null | grep '(active)' | awk '{ print \$2 }'"
GCLOUD_ACCOUNT=${GCLOUD_ACCOUNT-$(eval ${FIND_ACCOUNT})}
if [[ -z "${GCLOUD_ACCOUNT}" ]]; then
echo "No account authorized through gcloud. Please fix with:"
echo
echo " gcloud auth login"
return 1
fi
FIND_PROJECT="gcloud config list project | tail -n 1 | awk '{ print \$3 }'"
GCLOUD_PROJECT=${GCLOUD_PROJECT-$(eval ${FIND_PROJECT})}
if [[ -z "${GCLOUD_PROJECT}" ]]; then
echo "No account authorized through gcloud. Please fix with:"
echo
echo " gcloud config set project <project id>"
return 1
fi
}
# --------------------------------------------------------------------------- # ---------------------------------------------------------------------------
# Building # Building
# Set up the context directory for the kube-build image and build it. # Set up the context directory for the kube-build image and build it.
function build-image() { function kube::build::build_image() {
local -r BUILD_CONTEXT_DIR="${KUBE_REPO_ROOT}/_output/images/${KUBE_BUILD_IMAGE}" local -r build_context_dir="${KUBE_REPO_ROOT}/_output/images/${KUBE_BUILD_IMAGE}"
local -r SOURCE=" local -r source=(
api api
build build
cmd cmd
@ -125,78 +101,103 @@ function build-image() {
Godeps Godeps
hack hack
LICENSE LICENSE
README.md
pkg pkg
plugin plugin
README.md
third_party third_party
" )
mkdir -p ${BUILD_CONTEXT_DIR} mkdir -p "${build_context_dir}"
tar czf ${BUILD_CONTEXT_DIR}/kube-source.tar.gz ${SOURCE} tar czf "${build_context_dir}/kube-source.tar.gz" "${source[@]}"
cp build/build-image/Dockerfile ${BUILD_CONTEXT_DIR}/Dockerfile cat >"${build_context_dir}/kube-version-defs" <<EOF
docker-build "${KUBE_BUILD_IMAGE}" "${BUILD_CONTEXT_DIR}" KUBE_LD_FLAGS="$(kube::version_ldflags)"
EOF
cp build/build-image/Dockerfile ${build_context_dir}/Dockerfile
kube::build::docker_build "${KUBE_BUILD_IMAGE}" "${build_context_dir}"
} }
# Builds the runtime image. Assumes that the appropriate binaries are already # Builds the runtime image. Assumes that the appropriate binaries are already
# built and in _output/build/. # built and in _output/build/.
function run-image() { function kube::build::run_image() {
local -r BUILD_CONTEXT_BASE="${KUBE_REPO_ROOT}/_output/images/${KUBE_RUN_IMAGE_BASE}" local -r build_context_base="${KUBE_REPO_ROOT}/_output/images/${KUBE_RUN_IMAGE_BASE}"
# First build the base image. This one brings in all of the binaries. # First build the base image. This one brings in all of the binaries.
mkdir -p "${BUILD_CONTEXT_BASE}" mkdir -p "${build_context_base}"
tar czf ${BUILD_CONTEXT_BASE}/kube-bins.tar.gz \ tar czf "${build_context_base}/kube-bins.tar.gz" \
-C "_output/build/linux/amd64" \ -C "_output/build/linux/amd64" \
${KUBE_RUN_BINARIES} "${KUBE_RUN_BINARIES[@]}"
cp -R build/run-images/base/* "${BUILD_CONTEXT_BASE}/" cp -R build/run-images/base/* "${build_context_base}/"
docker-build "${KUBE_RUN_IMAGE_BASE}" "${BUILD_CONTEXT_BASE}" kube::build::docker_build "${KUBE_RUN_IMAGE_BASE}" "${build_context_base}"
for b in $KUBE_RUN_BINARIES ; do local b
local SUB_CONTEXT_DIR="${BUILD_CONTEXT_BASE}-$b" for b in "${KUBE_RUN_BINARIES[@]}" ; do
mkdir -p "${SUB_CONTEXT_DIR}" local sub_context_dir="${build_context_base}-$b"
cp -R build/run-images/$b/* "${SUB_CONTEXT_DIR}/" mkdir -p "${sub_context_dir}"
docker-build "${KUBE_RUN_IMAGE_BASE}-$b" "${SUB_CONTEXT_DIR}" cp -R build/run-images/$b/* "${sub_context_dir}/"
kube::build::docker_build "${KUBE_RUN_IMAGE_BASE}-$b" "${sub_context_dir}"
done done
} }
function kube::build::clean_images() {
kube::build::clean_image "${KUBE_BUILD_IMAGE}"
kube::build::clean_image "${KUBE_RUN_IMAGE_BASE}"
local b
for b in "${KUBE_RUN_BINARIES[@]}" ; do
kube::build::clean_image "${KUBE_RUN_IMAGE_BASE}-${b}"
done
echo "+++ Cleaning all other untagged docker images"
docker rmi $(docker images | awk '/^<none>/ {print $3}') 2> /dev/null || true
}
# Build a docker image from a Dockerfile. # Build a docker image from a Dockerfile.
# $1 is the name of the image to build # $1 is the name of the image to build
# $2 is the location of the "context" directory, with the Dockerfile at the root. # $2 is the location of the "context" directory, with the Dockerfile at the root.
function docker-build() { function kube::build::docker_build() {
local -r IMAGE=$1 local -r image=$1
local -r CONTEXT_DIR=$2 local -r context_dir=$2
local -r BUILD_CMD="docker build -t ${IMAGE} ${CONTEXT_DIR}" local -r build_cmd="docker build -t ${image} ${context_dir}"
echo "+++ Building Docker image ${IMAGE}. This can take a while." echo "+++ Building Docker image ${image}. This can take a while."
set +e # We are handling the error here manually set +e # We are handling the error here manually
local -r DOCKER_OUTPUT="$(${BUILD_CMD} 2>&1)" local -r docker_output="$(${build_cmd} 2>&1)"
if [ $? -ne 0 ]; then if [ $? -ne 0 ]; then
set -e set -e
echo "+++ Docker build command failed for ${IMAGE}" >&2 echo "+++ Docker build command failed for ${image}" >&2
echo >&2 echo >&2
echo "${DOCKER_OUTPUT}" >&2 echo "${docker_output}" >&2
echo >&2 echo >&2
echo "To retry manually, run:" >&2 echo "To retry manually, run:" >&2
echo >&2 echo >&2
echo " ${DOCKER_BUILD_CMD}" >&2 echo " ${build_cmd}" >&2
echo >&2 echo >&2
return 1 return 1
fi fi
set -e set -e
} }
function kube::build::clean_image() {
local -r image=$1
echo "+++ Deleting docker image ${image}"
docker rmi ${image} 2> /dev/null || true
}
# Run a command in the kube-build image. This assumes that the image has # Run a command in the kube-build image. This assumes that the image has
# already been built. This will sync out all output data from the build. # already been built. This will sync out all output data from the build.
function run-build-command() { function kube::build::run_build_command() {
[[ -n "$@" ]] || { echo "Invalid input." >&2; return 4; } [[ -n "$@" ]] || { echo "Invalid input." >&2; return 4; }
local -r DOCKER="docker run --rm --name=${DOCKER_CONTAINER_NAME} -it ${DOCKER_MOUNT} ${KUBE_BUILD_IMAGE}" local -r docker="docker run --rm --name=${DOCKER_CONTAINER_NAME} -it ${DOCKER_MOUNT} ${KUBE_BUILD_IMAGE}"
docker rm ${DOCKER_CONTAINER_NAME} >/dev/null 2>&1 || true docker rm ${DOCKER_CONTAINER_NAME} >/dev/null 2>&1 || true
${DOCKER} "$@" ${docker} "$@"
} }
# If the Docker server is remote, copy the results back out. # If the Docker server is remote, copy the results back out.
function copy-output() { function kube::build::copy_output() {
if [[ "$OSTYPE" == "darwin"* ]]; then if [[ "$OSTYPE" == "darwin"* ]]; then
# When we are on the Mac with boot2docker we need to copy the results back # When we are on the Mac with boot2docker we need to copy the results back
# out. Ideally we would leave the container around and use 'docker cp' to # out. Ideally we would leave the container around and use 'docker cp' to
@ -207,7 +208,7 @@ function copy-output() {
# The easiest thing I (jbeda) could figure out was to launch another # The easiest thing I (jbeda) could figure out was to launch another
# container pointed at the same volume, tar the output directory and ship # container pointed at the same volume, tar the output directory and ship
# that tar over stdou. # that tar over stdou.
local DOCKER="docker run -a stdout --rm --name=${DOCKER_CONTAINER_NAME} ${DOCKER_MOUNT} ${KUBE_BUILD_IMAGE}" local -r docker="docker run -a stdout --rm --name=${DOCKER_CONTAINER_NAME} ${DOCKER_MOUNT} ${KUBE_BUILD_IMAGE}"
# Kill any leftover container # Kill any leftover container
docker rm ${DOCKER_CONTAINER_NAME} >/dev/null 2>&1 || true docker rm ${DOCKER_CONTAINER_NAME} >/dev/null 2>&1 || true
@ -215,7 +216,7 @@ function copy-output() {
echo "+++ Syncing back _output directory from boot2docker VM" echo "+++ Syncing back _output directory from boot2docker VM"
mkdir -p "${LOCAL_OUTPUT_DIR}" mkdir -p "${LOCAL_OUTPUT_DIR}"
rm -rf "${LOCAL_OUTPUT_DIR}/*" rm -rf "${LOCAL_OUTPUT_DIR}/*"
${DOCKER} sh -c "tar c -C ${REMOTE_OUTPUT_DIR} ." \ ${docker} sh -c "tar c -C ${REMOTE_OUTPUT_DIR} ." \
| tar xv -C "${LOCAL_OUTPUT_DIR}" | tar xv -C "${LOCAL_OUTPUT_DIR}"
# I (jbeda) also tried getting rsync working using 'docker run' as the # I (jbeda) also tried getting rsync working using 'docker run' as the
@ -229,17 +230,82 @@ function copy-output() {
} }
# --------------------------------------------------------------------------- # ---------------------------------------------------------------------------
# Release # Build final release artifacts
# Package up all of the cross compiled clients
function kube::build::package_tarballs() {
mkdir -p "${RELEASE_DIR}"
# Find all of the built kubecfg binaries
local platform
for platform in _output/build/*/* ; do
local platform_tag=${platform}
platform_tag=${platform_tag#*/*/} # remove the first two path components
platform_tag=${platform_tag/\//-} # Replace a "/" for a "-"
echo "+++ Building client package for $platform_tag"
local client_release_stage="${KUBE_REPO_ROOT}/_output/release-stage/${platform_tag}/kubernetes"
mkdir -p "${client_release_stage}"
mkdir -p "${client_release_stage}/bin"
cp "${platform}"/* "${client_release_stage}/bin"
local client_package_name="${RELEASE_DIR}/kubernetes-${platform_tag}.tar.gz"
tar czf "${client_package_name}" -C "${client_release_stage}/.." .
done
}
# ---------------------------------------------------------------------------
# GCS Release
function kube::release::gcs::release() {
kube::release::gcs::verify_prereqs
kube::release::gcs::ensure_release_bucket
kube::release::gcs::push_images
kube::release::gcs::copy_release_tarballs
}
# Verify things are set up for uploading to GCS
function kube::release::gcs::verify_prereqs() {
if [[ -z "$(which gsutil)" || -z "$(which gcloud)" ]]; then
echo "Releasing Kubernetes requires gsutil and gcloud. Please download,"
echo "install and authorize through the Google Cloud SDK: "
echo
echo " https://developers.google.com/cloud/sdk/"
return 1
fi
if [[ -z "${GCLOUD_ACCOUNT-}" ]]; then
GCLOUD_ACCOUNT=$(gcloud auth list 2>/dev/null | awk '/(active)/ { print $2 }')
fi
if [[ -z "${GCLOUD_ACCOUNT}" ]]; then
echo "No account authorized through gcloud. Please fix with:"
echo
echo " gcloud auth login"
return 1
fi
if [[ -z "${GCLOUD_PROJECT-}" ]]; then
GCLOUD_PROJECT=$(gcloud config list project | awk '{project = $3} END {print project}')
fi
if [[ -z "${GCLOUD_PROJECT}" ]]; then
echo "No account authorized through gcloud. Please fix with:"
echo
echo " gcloud config set project <project id>"
return 1
fi
}
# Create a unique bucket name for releasing Kube and make sure it exists. # Create a unique bucket name for releasing Kube and make sure it exists.
function ensure-gcs-release-bucket() { function kube::release::gcs::ensure_release_bucket() {
local project_hash
if which md5 > /dev/null 2>&1; then if which md5 > /dev/null 2>&1; then
HASH=$(md5 -q -s "$GCLOUD_PROJECT") project_hash=$(md5 -q -s "$GCLOUD_PROJECT")
else else
HASH=$(echo -n "$GCLOUD_PROJECT" | md5sum) project_hash=$(echo -n "$GCLOUD_PROJECT" | md5sum)
fi fi
HASH=${HASH:0:5} project_hash=${project_hash:0:5}
KUBE_RELEASE_BUCKET=${KUBE_RELEASE_BUCKET-kubernetes-releases-$HASH} KUBE_RELEASE_BUCKET=${KUBE_RELEASE_BUCKET-kubernetes-releases-${project_hash}}
KUBE_RELEASE_PREFIX=${KUBE_RELEASE_PREFIX-devel/} KUBE_RELEASE_PREFIX=${KUBE_RELEASE_PREFIX-devel/}
KUBE_DOCKER_REG_PREFIX=${KUBE_DOCKER_REG_PREFIX-docker-reg/} KUBE_DOCKER_REG_PREFIX=${KUBE_DOCKER_REG_PREFIX-docker-reg/}
@ -249,87 +315,66 @@ function ensure-gcs-release-bucket() {
fi fi
} }
function ensure-gcs-docker-registry() { function kube::release::gcs::ensure_docker_registry() {
local -r REG_CONTAINER_NAME="gcs-registry" local -r reg_container_name="gcs-registry"
local -r RUNNING=$(docker inspect ${REG_CONTAINER_NAME} 2>/dev/null \ local -r running=$(docker inspect ${reg_container_name} 2>/dev/null \
| build/json-extractor.py 0.State.Running 2>/dev/null) | build/json-extractor.py 0.State.Running 2>/dev/null)
[[ "$RUNNING" != "true" ]] || return 0 [[ "$running" != "true" ]] || return 0
# Grovel around and find the OAuth token in the gcloud config # Grovel around and find the OAuth token in the gcloud config
local -r BOTO=~/.config/gcloud/legacy_credentials/${GCLOUD_ACCOUNT}/.boto local -r boto=~/.config/gcloud/legacy_credentials/${GCLOUD_ACCOUNT}/.boto
local -r REFRESH_TOKEN=$(grep 'gs_oauth2_refresh_token =' $BOTO | awk '{ print $3 }') local -r refresh_token=$(grep 'gs_oauth2_refresh_token =' $boto | awk '{ print $3 }')
if [[ -z $REFRESH_TOKEN ]]; then if [[ -z "$refresh_token" ]]; then
echo "Couldn't find OAuth 2 refresh token in ${BOTO}" >&2 echo "Couldn't find OAuth 2 refresh token in ${boto}" >&2
return 1 return 1
fi fi
# If we have an old one sitting around, remove it # If we have an old one sitting around, remove it
docker rm ${REG_CONTAINER_NAME} >/dev/null 2>&1 || true docker rm ${reg_container_name} >/dev/null 2>&1 || true
echo "+++ Starting GCS backed Docker registry" echo "+++ Starting GCS backed Docker registry"
local DOCKER="docker run -d --name=${REG_CONTAINER_NAME} " local docker="docker run -d --name=${reg_container_name} "
DOCKER+="-e GCS_BUCKET=${KUBE_RELEASE_BUCKET} " docker+="-e GCS_BUCKET=${KUBE_RELEASE_BUCKET} "
DOCKER+="-e STORAGE_PATH=${KUBE_DOCKER_REG_PREFIX} " docker+="-e STORAGE_PATH=${KUBE_DOCKER_REG_PREFIX} "
DOCKER+="-e GCP_OAUTH2_REFRESH_TOKEN=${REFRESH_TOKEN} " docker+="-e GCP_OAUTH2_REFRESH_TOKEN=${refresh_token} "
DOCKER+="-p 127.0.0.1:5000:5000 " docker+="-p 127.0.0.1:5000:5000 "
DOCKER+="google/docker-registry" docker+="google/docker-registry"
${DOCKER} ${docker}
# Give it time to spin up before we start throwing stuff at it # Give it time to spin up before we start throwing stuff at it
sleep 5 sleep 5
} }
function push-images-to-gcs() { function kube::release::gcs::push_images() {
ensure-gcs-docker-registry kube::release::gcs::ensure_docker_registry
# Tag each of our run binaries with the right registry and push # Tag each of our run binaries with the right registry and push
for b in ${KUBE_RUN_BINARIES} ; do local b image_name
echo "+++ Tagging and pushing ${KUBE_RUN_IMAGE_BASE}-$b to GCS bucket ${KUBE_RELEASE_BUCKET}" for b in "${KUBE_RUN_BINARIES[@]}" ; do
docker tag "${KUBE_RUN_IMAGE_BASE}-$b" "localhost:5000/${KUBE_RUN_IMAGE_BASE}-$b" image_name="${KUBE_RUN_IMAGE_BASE}-${b}"
docker push "localhost:5000/${KUBE_RUN_IMAGE_BASE}-$b" echo "+++ Tagging and pushing ${image_name} to GCS bucket ${KUBE_RELEASE_BUCKET}"
docker rmi "localhost:5000/${KUBE_RUN_IMAGE_BASE}-$b" docker tag "${KUBE_RUN_IMAGE_BASE}-$b" "localhost:5000/${image_name}"
docker push "localhost:5000/${image_name}"
docker rmi "localhost:5000/${image_name}"
done done
} }
# Package up all of the cross compiled clients function kube::release::gcs::copy_release_tarballs() {
function package-tarballs() {
mkdir -p "${RELEASE_DIR}"
# Find all of the built kubecfg binaries
for platform in _output/build/*/* ; do
echo $platform
local PLATFORM_TAG=$(echo $platform | awk -F / '{ printf "%s-%s", $3, $4 }')
echo "+++ Building client package for $PLATFORM_TAG"
local CLIENT_RELEASE_STAGE="${KUBE_REPO_ROOT}/_output/release-stage/${PLATFORM_TAG}/kubernetes"
mkdir -p "${CLIENT_RELEASE_STAGE}"
mkdir -p "${CLIENT_RELEASE_STAGE}/bin"
cp $platform/* "${CLIENT_RELEASE_STAGE}/bin"
local CLIENT_PACKAGE_NAME="${RELEASE_DIR}/kubernetes-${PLATFORM_TAG}.tar.gz"
tar czf ${CLIENT_PACKAGE_NAME} \
-C "${CLIENT_RELEASE_STAGE}/.." \
.
done
}
function copy-release-to-gcs() {
# TODO: This isn't atomic. There will be points in time where there will be # TODO: This isn't atomic. There will be points in time where there will be
# no active release. Also, if something fails, the release could be half- # no active release. Also, if something fails, the release could be half-
# copied. The real way to do this would perhaps to have some sort of release # copied. The real way to do this would perhaps to have some sort of release
# version so that we are never overwriting a destination. # version so that we are never overwriting a destination.
local -r GCS_DESTINATION="gs://${KUBE_RELEASE_BUCKET}/${KUBE_RELEASE_PREFIX}" local -r gcs_destination="gs://${KUBE_RELEASE_BUCKET}/${KUBE_RELEASE_PREFIX}"
echo "+++ Copying client tarballs to ${GCS_DESTINATION}" echo "+++ Copying client tarballs to ${gcs_destination}"
# First delete all objects at the destination # First delete all objects at the destination
gsutil -q rm -f -R "${GCS_DESTINATION}" >/dev/null 2>&1 || true gsutil -q rm -f -R "${gcs_destination}" >/dev/null 2>&1 || true
# Now upload everything in release directory # Now upload everything in release directory
gsutil -m cp -r "${RELEASE_DIR}" "${GCS_DESTINATION}" >/dev/null 2>&1 gsutil -m cp -r "${RELEASE_DIR}" "${gcs_destination}" >/dev/null 2>&1
} }

View File

@ -23,5 +23,5 @@ set -e
source $(dirname $0)/common.sh source $(dirname $0)/common.sh
verify-prereqs kube::build::verify_prereqs
copy-output kube::build::copy_output

View File

@ -23,6 +23,6 @@ set -e
source $(dirname $0)/common.sh source $(dirname $0)/common.sh
verify-prereqs kube::build::verify_prereqs
build-image kube::build::build_image
run-build-command build/build-image/make-binaries.sh "$@" kube::build::run_build_command build/build-image/make-binaries.sh "$@"

View File

@ -25,5 +25,5 @@ set -e
source $(dirname $0)/common.sh source $(dirname $0)/common.sh
verify-prereqs kube::build::verify_prereqs
build-image kube::build::build_image

View File

@ -20,6 +20,10 @@ set -e
source $(dirname $0)/common.sh source $(dirname $0)/common.sh
verify-prereqs kube::build::verify_prereqs
build-image kube::build::build_image
run-build-command rm -rf _output/build/*
echo "+++ Cleaning out _output/build/*"
kube::build::run_build_command rm -rf _output/build/*
kube::build::clean_images

View File

@ -23,6 +23,6 @@ set -e
source $(dirname $0)/common.sh source $(dirname $0)/common.sh
verify-prereqs kube::build::verify_prereqs
build-image kube::build::build_image
run-build-command build/build-image/make-cross.sh kube::build::run_build_command build/build-image/make-cross.sh

View File

@ -23,8 +23,8 @@ set -e
source $(dirname $0)/common.sh source $(dirname $0)/common.sh
verify-prereqs kube::build::verify_prereqs
build-image kube::build::build_image
run-build-command build/build-image/make-binaries.sh "$@" kube::build::run_build_command build/build-image/make-binaries.sh "$@"
copy-output kube::build::copy_output
run-image kube::build::run_image

View File

@ -22,16 +22,13 @@ set -e
source $(dirname $0)/common.sh source $(dirname $0)/common.sh
verify-prereqs kube::build::verify_prereqs
verify-gcs-prereqs kube::build::build_image
ensure-gcs-release-bucket kube::build::run_build_command build/build-image/make-binaries.sh
build-image kube::build::run_build_command build/build-image/make-cross.sh
run-build-command build/build-image/make-binaries.sh kube::build::run_build_command build/build-image/run-tests.sh
run-build-command build/build-image/make-cross.sh kube::build::run_build_command build/build-image/run-integration.sh
run-build-command build/build-image/run-tests.sh kube::build::copy_output
run-build-command build/build-image/run-integration.sh kube::build::run_image
copy-output kube::build::package_tarballs
run-image kube::release::gcs::release
package-tarballs
push-images-to-gcs
copy-release-to-gcs

View File

@ -19,5 +19,5 @@ MAINTAINER Joe Beda <jbeda@google.com>
WORKDIR /kubernetes WORKDIR /kubernetes
# Upload Kubernetes # Upload Kubernetes server binaries
ADD kube-bins.tar.gz /kubernetes ADD kube-bins.tar.gz /kubernetes

View File

@ -0,0 +1,24 @@
# Copyright 2014 Google Inc. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# This file creates a minimal container for running Kubernetes binaries
FROM kubernetes
MAINTAINER Joe Beda <jbeda@google.com>
ENV API_SERVER 127.0.0.1:8080
ADD . /kubernetes
CMD ["/kubernetes/run.sh"]

View File

@ -0,0 +1,17 @@
#! /bin/bash
# Copyright 2014 Google Inc. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
./scheduler -master="${API_SERVER}"

View File

@ -20,7 +20,7 @@ set -e
source $(dirname $0)/common.sh source $(dirname $0)/common.sh
verify-prereqs kube::build::verify_prereqs
build-image kube::build::build_image
run-build-command build/build-image/make-binaries.sh "integration" kube::build::run_build_command build/build-image/make-binaries.sh "./cmd/integration"
run-build-command build/build-image/run-integration.sh kube::build::run_build_command build/build-image/run-integration.sh

View File

@ -20,6 +20,6 @@ set -e
source $(dirname $0)/common.sh source $(dirname $0)/common.sh
verify-prereqs kube::build::verify_prereqs
build-image kube::build::build_image
run-build-command build/build-image/run-tests.sh "$@" kube::build::run_build_command build/build-image/run-tests.sh "$@"

View File

@ -22,6 +22,6 @@ set -e
source $(dirname $0)/common.sh source $(dirname $0)/common.sh
verify-prereqs kube::build::verify_prereqs
build-image kube::build::build_image
run-build-command bash kube::build::run_build_command bash

View File

@ -24,11 +24,11 @@ function public-key {
fi fi
done done
echo "Can't find public key file..." echo "Can't find public key file..." 1>&2
exit 1 exit 1
} }
DISK=kube.vmdk DISK=./kube/kube.vmdk
GUEST_ID=debian7_64Guest GUEST_ID=debian7_64Guest
PUBLIC_KEY_FILE=${PUBLIC_KEY_FILE-$(public-key)} PUBLIC_KEY_FILE=${PUBLIC_KEY_FILE-$(public-key)}
SSH_OPTS="-oStrictHostKeyChecking=no -oUserKnownHostsFile=/dev/null" SSH_OPTS="-oStrictHostKeyChecking=no -oUserKnownHostsFile=/dev/null"
@ -40,3 +40,7 @@ SSH_OPTS="-oStrictHostKeyChecking=no -oUserKnownHostsFile=/dev/null"
#export GOVC_RESOURCE_POOL= #export GOVC_RESOURCE_POOL=
#export GOVC_NETWORK= #export GOVC_NETWORK=
#export GOVC_GUEST_LOGIN='kube:kube' #export GOVC_GUEST_LOGIN='kube:kube'
# Set GOVC_INSECURE if the host in GOVC_URL is using a certificate that cannot
# be verified (i.e. a self-signed certificate), but IS trusted.
#export GOVC_INSECURE=1

View File

@ -20,6 +20,7 @@ import (
"flag" "flag"
"fmt" "fmt"
"io/ioutil" "io/ioutil"
"net/http"
"os" "os"
"reflect" "reflect"
"sort" "sort"
@ -39,10 +40,10 @@ import (
) )
var ( var (
serverVersion = flag.Bool("server_version", false, "Print the server's version number.") serverVersion = verflag.Version("server_version", verflag.VersionFalse, "Print the server's version information and quit")
preventSkew = flag.Bool("expect_version_match", false, "Fail if server's version doesn't match own version.") preventSkew = flag.Bool("expect_version_match", false, "Fail if server's version doesn't match own version.")
httpServer = flag.String("h", "", "The host to connect to.") httpServer = flag.String("h", "", "The host to connect to.")
config = flag.String("c", "", "Path to the config file.") config = flag.String("c", "", "Path or URL to the config file.")
selector = flag.String("l", "", "Selector (label query) to use for listing") selector = flag.String("l", "", "Selector (label query) to use for listing")
updatePeriod = flag.Duration("u", 60*time.Second, "Update interval period") updatePeriod = flag.Duration("u", 60*time.Second, "Update interval period")
portSpec = flag.String("p", "", "The port spec, comma-separated list of <external>:<internal>,...") portSpec = flag.String("p", "", "The port spec, comma-separated list of <external>:<internal>,...")
@ -95,17 +96,41 @@ func prettyWireStorage() string {
return strings.Join(types, "|") return strings.Join(types, "|")
} }
// readConfigData reads the bytes from the specified filesytem or network location associated with the *config flag
func readConfigData() []byte {
// we look for http:// or https:// to determine if valid URL, otherwise do normal file IO
if strings.Index(*config, "http://") == 0 || strings.Index(*config, "https://") == 0 {
resp, err := http.Get(*config)
if err != nil {
glog.Fatalf("Unable to access URL %v: %v\n", *config, err)
}
defer resp.Body.Close()
if resp.StatusCode != 200 {
glog.Fatalf("Unable to read URL, server reported %d %s", resp.StatusCode, resp.Status)
}
data, err := ioutil.ReadAll(resp.Body)
if err != nil {
glog.Fatalf("Unable to read URL %v: %v\n", *config, err)
}
return data
} else {
data, err := ioutil.ReadFile(*config)
if err != nil {
glog.Fatalf("Unable to read %v: %v\n", *config, err)
}
return data
}
}
// readConfig reads and parses pod, replicationController, and service // readConfig reads and parses pod, replicationController, and service
// configuration files. If any errors log and exit non-zero. // configuration files. If any errors log and exit non-zero.
func readConfig(storage string) []byte { func readConfig(storage string) []byte {
if len(*config) == 0 { if len(*config) == 0 {
glog.Fatal("Need config file (-c)") glog.Fatal("Need config file (-c)")
} }
data, err := ioutil.ReadFile(*config)
if err != nil { data, err := parser.ToWireFormat(readConfigData(), storage, runtime.DefaultCodec)
glog.Fatalf("Unable to read %v: %v\n", *config, err)
}
data, err = parser.ToWireFormat(data, storage, runtime.DefaultCodec)
if err != nil { if err != nil {
glog.Fatalf("Error parsing %v as an object for %v: %v\n", *config, storage, err) glog.Fatalf("Error parsing %v as an object for %v: %v\n", *config, storage, err)
} }
@ -152,14 +177,19 @@ func main() {
} }
} }
if *serverVersion { if *serverVersion != verflag.VersionFalse {
got, err := kubeClient.ServerVersion() got, err := kubeClient.ServerVersion()
if err != nil { if err != nil {
fmt.Printf("Couldn't read version from server: %v\n", err) fmt.Printf("Couldn't read version from server: %v\n", err)
os.Exit(1) os.Exit(1)
} }
fmt.Printf("Server Version: %#v\n", got) if *serverVersion == verflag.VersionRaw {
fmt.Printf("%#v\n", *got)
os.Exit(0) os.Exit(0)
} else {
fmt.Printf("Server: Kubernetes %s\n", got)
os.Exit(0)
}
} }
if *preventSkew { if *preventSkew {

View File

@ -41,16 +41,17 @@ Upload this VMDK to your vSphere instance:
```sh ```sh
export GOVC_URL='https://user:pass@hostname/sdk' export GOVC_URL='https://user:pass@hostname/sdk'
export GOVC_INSECURE=1 # If the host above uses a self-signed cert
export GOVC_DATASTORE='target datastore' export GOVC_DATASTORE='target datastore'
export GOVC_RESOURCE_POOL='resource pool with access to datastore' export GOVC_RESOURCE_POOL='resource pool or cluster with access to datastore'
govc datastore.import kube.vmdk govc import.vmdk kube.vmdk ./kube/
``` ```
Verify that the VMDK was correctly uploaded and expanded to 10GiB: Verify that the VMDK was correctly uploaded and expanded to 10GiB:
```sh ```sh
govc datastore.ls govc datastore.ls ./kube/
``` ```
Take a look at the file `cluster/vsphere/config-common.sh` fill in the required Take a look at the file `cluster/vsphere/config-common.sh` fill in the required

View File

@ -48,20 +48,10 @@ for arg; do
done done
if [[ ${#targets[@]} -eq 0 ]]; then if [[ ${#targets[@]} -eq 0 ]]; then
targets=( targets=($(kube::default_build_targets))
cmd/proxy
cmd/apiserver
cmd/controller-manager
cmd/kubelet
cmd/kubecfg
plugin/cmd/scheduler
)
fi fi
binaries=() binaries=($(kube::binaries_from_targets "${targets[@]}"))
for target in ${targets[@]}; do
binaries+=("${KUBE_GO_PACKAGE}/${target}")
done
echo "Building local go components" echo "Building local go components"
# Note that the flags to 'go build' are duplicated in the salt build setup # Note that the flags to 'go build' are duplicated in the salt build setup

View File

@ -114,6 +114,24 @@ kube::setup_go_environment() {
} }
# kube::default_build_targets return list of all build targets
kube::default_build_targets() {
echo "cmd/proxy"
echo "cmd/apiserver"
echo "cmd/controller-manager"
echo "cmd/kubelet"
echo "cmd/kubecfg"
echo "plugin/cmd/scheduler"
}
# kube::binaries_from_targets take a list of build targets and return the
# full go package to be built
kube::binaries_from_targets() {
local target
for target; do
echo "${KUBE_GO_PACKAGE}/${target}"
done
}
# --- Environment Variables --- # --- Environment Variables ---
# KUBE_REPO_ROOT - Path to the top of the build tree. # KUBE_REPO_ROOT - Path to the top of the build tree.

View File

@ -326,6 +326,7 @@ func (*ReplicationControllerList) IsAnAPIObject() {}
type ReplicationController struct { type ReplicationController struct {
JSONBase `json:",inline" yaml:",inline"` JSONBase `json:",inline" yaml:",inline"`
DesiredState ReplicationControllerState `json:"desiredState,omitempty" yaml:"desiredState,omitempty"` DesiredState ReplicationControllerState `json:"desiredState,omitempty" yaml:"desiredState,omitempty"`
CurrentState ReplicationControllerState `json:"currentState,omitempty" yaml:"currentState,omitempty"`
Labels map[string]string `json:"labels,omitempty" yaml:"labels,omitempty"` Labels map[string]string `json:"labels,omitempty" yaml:"labels,omitempty"`
} }

View File

@ -339,6 +339,7 @@ func (*ReplicationControllerList) IsAnAPIObject() {}
type ReplicationController struct { type ReplicationController struct {
JSONBase `json:",inline" yaml:",inline"` JSONBase `json:",inline" yaml:",inline"`
DesiredState ReplicationControllerState `json:"desiredState,omitempty" yaml:"desiredState,omitempty"` DesiredState ReplicationControllerState `json:"desiredState,omitempty" yaml:"desiredState,omitempty"`
CurrentState ReplicationControllerState `json:"currentState,omitempty" yaml:"currentState,omitempty"`
Labels map[string]string `json:"labels,omitempty" yaml:"labels,omitempty"` Labels map[string]string `json:"labels,omitempty" yaml:"labels,omitempty"`
} }

View File

@ -228,8 +228,8 @@ func ValidateManifest(manifest *api.ContainerManifest) errs.ErrorList {
} else if !supportedManifestVersions.Has(strings.ToLower(manifest.Version)) { } else if !supportedManifestVersions.Has(strings.ToLower(manifest.Version)) {
allErrs = append(allErrs, errs.NewFieldNotSupported("version", manifest.Version)) allErrs = append(allErrs, errs.NewFieldNotSupported("version", manifest.Version))
} }
allVolumes, errs := validateVolumes(manifest.Volumes) allVolumes, vErrs := validateVolumes(manifest.Volumes)
allErrs = append(allErrs, errs.Prefix("volumes")...) allErrs = append(allErrs, vErrs.Prefix("volumes")...)
allErrs = append(allErrs, validateContainers(manifest.Containers, allVolumes).Prefix("containers")...) allErrs = append(allErrs, validateContainers(manifest.Containers, allVolumes).Prefix("containers")...)
allErrs = append(allErrs, validateRestartPolicy(&manifest.RestartPolicy).Prefix("restartPolicy")...) allErrs = append(allErrs, validateRestartPolicy(&manifest.RestartPolicy).Prefix("restartPolicy")...)
return allErrs return allErrs

View File

@ -29,6 +29,7 @@ import (
"github.com/GoogleCloudPlatform/kubernetes/pkg/api" "github.com/GoogleCloudPlatform/kubernetes/pkg/api"
"github.com/GoogleCloudPlatform/kubernetes/pkg/client" "github.com/GoogleCloudPlatform/kubernetes/pkg/client"
"github.com/GoogleCloudPlatform/kubernetes/pkg/labels" "github.com/GoogleCloudPlatform/kubernetes/pkg/labels"
"github.com/GoogleCloudPlatform/kubernetes/pkg/util"
"github.com/GoogleCloudPlatform/kubernetes/pkg/util/wait" "github.com/GoogleCloudPlatform/kubernetes/pkg/util/wait"
"github.com/GoogleCloudPlatform/kubernetes/pkg/version" "github.com/GoogleCloudPlatform/kubernetes/pkg/version"
"github.com/golang/glog" "github.com/golang/glog"
@ -174,6 +175,9 @@ func portsFromString(spec string) []api.Port {
// RunController creates a new replication controller named 'name' which creates 'replicas' pods running 'image'. // RunController creates a new replication controller named 'name' which creates 'replicas' pods running 'image'.
func RunController(image, name string, replicas int, client client.Interface, portSpec string, servicePort int) error { func RunController(image, name string, replicas int, client client.Interface, portSpec string, servicePort int) error {
if servicePort > 0 && !util.IsDNSLabel(name) {
return fmt.Errorf("Service creation requested, but an invalid name for a service was provided (%s). Service names must be valid DNS labels.", name)
}
controller := &api.ReplicationController{ controller := &api.ReplicationController{
JSONBase: api.JSONBase{ JSONBase: api.JSONBase{
ID: name, ID: name,

View File

@ -92,6 +92,7 @@ func (rs *REST) Get(id string) (runtime.Object, error) {
if err != nil { if err != nil {
return nil, err return nil, err
} }
rs.fillCurrentState(controller)
return controller, err return controller, err
} }
@ -104,6 +105,7 @@ func (rs *REST) List(selector labels.Selector) (runtime.Object, error) {
filtered := []api.ReplicationController{} filtered := []api.ReplicationController{}
for _, controller := range controllers.Items { for _, controller := range controllers.Items {
if selector.Matches(labels.Set(controller.Labels)) { if selector.Matches(labels.Set(controller.Labels)) {
rs.fillCurrentState(&controller)
filtered = append(filtered, controller) filtered = append(filtered, controller)
} }
} }
@ -147,7 +149,11 @@ func (rs *REST) Watch(label, field labels.Selector, resourceVersion uint64) (wat
} }
return watch.Filter(incoming, func(e watch.Event) (watch.Event, bool) { return watch.Filter(incoming, func(e watch.Event) (watch.Event, bool) {
repController := e.Object.(*api.ReplicationController) repController := e.Object.(*api.ReplicationController)
return e, label.Matches(labels.Set(repController.Labels)) match := label.Matches(labels.Set(repController.Labels))
if match {
rs.fillCurrentState(repController)
}
return e, match
}), nil }), nil
} }
@ -164,3 +170,15 @@ func (rs *REST) waitForController(ctrl *api.ReplicationController) (runtime.Obje
} }
return ctrl, nil return ctrl, nil
} }
func (rs *REST) fillCurrentState(ctrl *api.ReplicationController) error {
if rs.podLister == nil {
return nil
}
list, err := rs.podLister.ListPods(labels.Set(ctrl.DesiredState.ReplicaSelector).AsSelector())
if err != nil {
return err
}
ctrl.CurrentState.Replicas = len(list.Items)
return nil
}

View File

@ -316,3 +316,44 @@ func TestControllerStorageValidatesUpdate(t *testing.T) {
} }
} }
} }
type fakePodLister struct {
e error
l api.PodList
s labels.Selector
}
func (f *fakePodLister) ListPods(s labels.Selector) (*api.PodList, error) {
f.s = s
return &f.l, f.e
}
func TestFillCurrentState(t *testing.T) {
fakeLister := fakePodLister{
l: api.PodList{
Items: []api.Pod{
{JSONBase: api.JSONBase{ID: "foo"}},
{JSONBase: api.JSONBase{ID: "bar"}},
},
},
}
mockRegistry := registrytest.ControllerRegistry{}
storage := REST{
registry: &mockRegistry,
podLister: &fakeLister,
}
controller := api.ReplicationController{
DesiredState: api.ReplicationControllerState{
ReplicaSelector: map[string]string{
"foo": "bar",
},
},
}
storage.fillCurrentState(&controller)
if controller.CurrentState.Replicas != 2 {
t.Errorf("expected 2, got: %d", controller.CurrentState.Replicas)
}
if !reflect.DeepEqual(fakeLister.s, labels.Set(controller.DesiredState.ReplicaSelector).AsSelector()) {
t.Errorf("unexpected output: %#v %#v", labels.Set(controller.DesiredState.ReplicaSelector).AsSelector(), fakeLister.s)
}
}

View File

@ -23,9 +23,10 @@ package version
// version for ad-hoc builds (e.g. `go build`) that cannot get the version // version for ad-hoc builds (e.g. `go build`) that cannot get the version
// information from git. // information from git.
// //
// The "+" in the version info indicates that fact, and it means the current // The "-dev" suffix in the version info indicates that fact, and it means the
// build is from a version greater or equal to that version. // current build is from a version greater that version. For example, v0.7-dev
// (e.g. v0.7+ means version >= 0.7 and < 0.8) // means version > 0.7 and < 0.8. (There's exceptions to this rule, see
// docs/releasing.md for more details.)
// //
// When releasing a new Kubernetes version, this file should be updated to // When releasing a new Kubernetes version, this file should be updated to
// reflect the new version, and then a git annotated tag (using format vX.Y // reflect the new version, and then a git annotated tag (using format vX.Y
@ -33,9 +34,10 @@ package version
// to the commit that updates pkg/version/base.go // to the commit that updates pkg/version/base.go
var ( var (
// TODO: Deprecate gitMajor and gitMinor, use only gitVersion instead.
gitMajor string = "0" // major version, always numeric gitMajor string = "0" // major version, always numeric
gitMinor string = "2+" // minor version, numeric possibly followed by "+" gitMinor string = "2+" // minor version, numeric possibly followed by "+"
gitVersion string = "v0.2+" // version from git, output of $(git describe) gitVersion string = "v0.2-dev" // version from git, output of $(git describe)
gitCommit string = "" // sha1 from git, output of $(git rev-parse HEAD) gitCommit string = "" // sha1 from git, output of $(git rev-parse HEAD)
gitTreeState string = "not a git tree" // state of git tree, either "clean" or "dirty" gitTreeState string = "not a git tree" // state of git tree, either "clean" or "dirty"
) )

View File

@ -16,10 +16,6 @@ limitations under the License.
package version package version
import (
"fmt"
)
// Info contains versioning information. // Info contains versioning information.
// TODO: Add []string of api versions supported? It's still unclear // TODO: Add []string of api versions supported? It's still unclear
// how we'll want to distribute that information. // how we'll want to distribute that information.
@ -47,9 +43,5 @@ func Get() Info {
// String returns info as a human-friendly version string. // String returns info as a human-friendly version string.
func (info Info) String() string { func (info Info) String() string {
commit := info.GitCommit return info.GitVersion
if commit == "" {
commit = "(unknown)"
}
return fmt.Sprintf("version %s.%s, build %s", info.Major, info.Minor, commit)
} }