Modularize local-up into clean functions so its self documenting

pull/6/head
jayunit100 2015-06-09 20:26:18 -04:00
parent f62a2a1bb6
commit a2ea0ba467
1 changed files with 167 additions and 131 deletions

View File

@ -22,21 +22,27 @@ DOCKER_NATIVE=${DOCKER_NATIVE:-""}
DOCKER=(docker ${DOCKER_OPTS}) DOCKER=(docker ${DOCKER_OPTS})
DOCKERIZE_KUBELET=${DOCKERIZE_KUBELET:-""} DOCKERIZE_KUBELET=${DOCKERIZE_KUBELET:-""}
ALLOW_PRIVILEGED=${ALLOW_PRIVILEGED:-""} ALLOW_PRIVILEGED=${ALLOW_PRIVILEGED:-""}
KUBE_ROOT=$(dirname "${BASH_SOURCE}")/.. KUBE_ROOT=$(dirname "${BASH_SOURCE}")/..
cd "${KUBE_ROOT}" cd "${KUBE_ROOT}"
if [ "$(id -u)" != "0" ]; then
echo "WARNING : This script MAY be run as root for docker socket / iptables functionality... if failures occur... Retry as root." 2>&1
fi
# Stop right away if the build fails # Stop right away if the build fails
set -e set -e
source "${KUBE_ROOT}/hack/lib/init.sh" source "${KUBE_ROOT}/hack/lib/init.sh"
"${KUBE_ROOT}/hack/build-go.sh" "${KUBE_ROOT}/hack/build-go.sh"
${DOCKER[@]} ps 2> /dev/null 1> /dev/null function test_docker {
if [ "$?" != "0" ]; then ${DOCKER[@]} ps 2> /dev/null 1> /dev/null
echo "Failed to successfully run 'docker ps', please verify that docker is installed and \$DOCKER_HOST is set correctly." if [ "$?" != "0" ]; then
exit 1 echo "Failed to successfully run 'docker ps', please verify that docker is installed and \$DOCKER_HOST is set correctly."
fi exit 1
fi
}
# Shut down anyway if there's an error. # Shut down anyway if there's an error.
set +e set +e
@ -50,54 +56,55 @@ LOG_LEVEL=${LOG_LEVEL:-3}
CONTAINER_RUNTIME=${CONTAINER_RUNTIME:-"docker"} CONTAINER_RUNTIME=${CONTAINER_RUNTIME:-"docker"}
CHAOS_CHANCE=${CHAOS_CHANCE:-0.0} CHAOS_CHANCE=${CHAOS_CHANCE:-0.0}
# For the common local scenario, fail fast if server is already running. function test_apiserver_off {
# this can happen if you run local-up-cluster.sh twice and kill etcd in between. # For the common local scenario, fail fast if server is already running.
curl $API_HOST:$API_PORT # this can happen if you run local-up-cluster.sh twice and kill etcd in between.
if [ ! $? -eq 0 ]; then curl $API_HOST:$API_PORT
echo "API SERVER port is free, proceeding..." if [ ! $? -eq 0 ]; then
else echo "API SERVER port is free, proceeding..."
echo "ERROR starting API SERVER, exiting. Some host on $API_HOST is serving already on $API_PORT" else
exit 1 echo "ERROR starting API SERVER, exiting. Some host on $API_HOST is serving already on $API_PORT"
fi exit 1
fi
}
# Detect the OS name/arch so that we can find our binary function detect_binary {
case "$(uname -s)" in # Detect the OS name/arch so that we can find our binary
Darwin) case "$(uname -s)" in
host_os=darwin Darwin)
;; host_os=darwin
Linux) ;;
host_os=linux Linux)
;; host_os=linux
*) ;;
echo "Unsupported host OS. Must be Linux or Mac OS X." >&2 *)
exit 1 echo "Unsupported host OS. Must be Linux or Mac OS X." >&2
;; exit 1
esac ;;
esac
case "$(uname -m)" in case "$(uname -m)" in
x86_64*) x86_64*)
host_arch=amd64 host_arch=amd64
;; ;;
i?86_64*) i?86_64*)
host_arch=amd64 host_arch=amd64
;; ;;
amd64*) amd64*)
host_arch=amd64 host_arch=amd64
;; ;;
arm*) arm*)
host_arch=arm host_arch=arm
;; ;;
i?86*) i?86*)
host_arch=x86 host_arch=x86
;; ;;
*) *)
echo "Unsupported host arch. Must be x86_64, 386 or arm." >&2 echo "Unsupported host arch. Must be x86_64, 386 or arm." >&2
exit 1 exit 1
;; ;;
esac esac
}
GO_OUT="${KUBE_ROOT}/_output/local/bin/${host_os}/${host_arch}"
KUBELET_CIDFILE=/tmp/kubelet.cid
cleanup_dockerized_kubelet() cleanup_dockerized_kubelet()
{ {
@ -141,96 +148,107 @@ cleanup()
exit 0 exit 0
} }
trap cleanup EXIT function startETCD {
echo "Starting etcd"
kube::etcd::start
}
echo "Starting etcd" function set_service_accounts {
kube::etcd::start SERVICE_ACCOUNT_LOOKUP=${SERVICE_ACCOUNT_LOOKUP:-false}
SERVICE_ACCOUNT_KEY=${SERVICE_ACCOUNT_KEY:-"/tmp/kube-serviceaccount.key"}
# Generate ServiceAccount key if needed
if [[ ! -f "${SERVICE_ACCOUNT_KEY}" ]]; then
mkdir -p "$(dirname ${SERVICE_ACCOUNT_KEY})"
openssl genrsa -out "${SERVICE_ACCOUNT_KEY}" 2048 2>/dev/null
fi
}
SERVICE_ACCOUNT_LOOKUP=${SERVICE_ACCOUNT_LOOKUP:-false} function start_apiserver {
SERVICE_ACCOUNT_KEY=${SERVICE_ACCOUNT_KEY:-"/tmp/kube-serviceaccount.key"} # Admission Controllers to invoke prior to persisting objects in cluster
# Generate ServiceAccount key if needed ADMISSION_CONTROL=NamespaceLifecycle,NamespaceAutoProvision,LimitRanger,SecurityContextDeny,ServiceAccount,ResourceQuota
if [[ ! -f "${SERVICE_ACCOUNT_KEY}" ]]; then
mkdir -p "$(dirname ${SERVICE_ACCOUNT_KEY})"
openssl genrsa -out "${SERVICE_ACCOUNT_KEY}" 2048 2>/dev/null
fi
# Admission Controllers to invoke prior to persisting objects in cluster priv_arg=""
ADMISSION_CONTROL=NamespaceLifecycle,NamespaceAutoProvision,LimitRanger,SecurityContextDeny,ServiceAccount,ResourceQuota if [[ -n "${ALLOW_PRIVILEGED}" ]]; then
priv_arg="--allow-privileged "
fi
priv_arg="" APISERVER_LOG=/tmp/kube-apiserver.log
if [[ -n "${ALLOW_PRIVILEGED}" ]]; then sudo -E "${GO_OUT}/kube-apiserver" ${priv_arg}\
priv_arg="--allow-privileged " --v=${LOG_LEVEL} \
fi --service_account_key_file="${SERVICE_ACCOUNT_KEY}" \
--service_account_lookup="${SERVICE_ACCOUNT_LOOKUP}" \
--admission_control="${ADMISSION_CONTROL}" \
--address="${API_HOST}" \
--port="${API_PORT}" \
--runtime_config=api/v1beta3 \
--etcd_servers="http://127.0.0.1:4001" \
--service-cluster-ip-range="10.0.0.0/24" \
--cors_allowed_origins="${API_CORS_ALLOWED_ORIGINS}" >"${APISERVER_LOG}" 2>&1 &
APISERVER_PID=$!
APISERVER_LOG=/tmp/kube-apiserver.log # Wait for kube-apiserver to come up before launching the rest of the components.
sudo -E "${GO_OUT}/kube-apiserver" ${priv_arg}\ echo "Waiting for apiserver to come up"
--v=${LOG_LEVEL} \ kube::util::wait_for_url "http://${API_HOST}:${API_PORT}/api/v1beta3/pods" "apiserver: " 1 10 || exit 1
--service_account_key_file="${SERVICE_ACCOUNT_KEY}" \ }
--service_account_lookup="${SERVICE_ACCOUNT_LOOKUP}" \
--admission_control="${ADMISSION_CONTROL}" \
--address="${API_HOST}" \
--port="${API_PORT}" \
--runtime_config=api/v1beta3 \
--etcd_servers="http://127.0.0.1:4001" \
--service-cluster-ip-range="10.0.0.0/24" \
--cors_allowed_origins="${API_CORS_ALLOWED_ORIGINS}" >"${APISERVER_LOG}" 2>&1 &
APISERVER_PID=$!
# Wait for kube-apiserver to come up before launching the rest of the components. function start_controller_manager {
echo "Waiting for apiserver to come up" CTLRMGR_LOG=/tmp/kube-controller-manager.log
kube::util::wait_for_url "http://${API_HOST}:${API_PORT}/api/v1beta3/pods" "apiserver: " 1 10 || exit 1 sudo -E "${GO_OUT}/kube-controller-manager" \
--v=${LOG_LEVEL} \
--machines="127.0.0.1" \
--service_account_private_key_file="${SERVICE_ACCOUNT_KEY}" \
--master="${API_HOST}:${API_PORT}" >"${CTLRMGR_LOG}" 2>&1 &
CTLRMGR_PID=$!
}
CTLRMGR_LOG=/tmp/kube-controller-manager.log function start_kubelet {
sudo -E "${GO_OUT}/kube-controller-manager" \ KUBELET_LOG=/tmp/kubelet.log
--v=${LOG_LEVEL} \ if [[ -z "${DOCKERIZE_KUBELET}" ]]; then
--machines="127.0.0.1" \ sudo -E "${GO_OUT}/kubelet" ${priv_arg}\
--service_account_private_key_file="${SERVICE_ACCOUNT_KEY}" \ --v=${LOG_LEVEL} \
--master="${API_HOST}:${API_PORT}" >"${CTLRMGR_LOG}" 2>&1 & --chaos_chance="${CHAOS_CHANCE}" \
CTLRMGR_PID=$! --container_runtime="${CONTAINER_RUNTIME}" \
--hostname_override="127.0.0.1" \
--address="127.0.0.1" \
--api_servers="${API_HOST}:${API_PORT}" \
--port="$KUBELET_PORT" >"${KUBELET_LOG}" 2>&1 &
KUBELET_PID=$!
else
# Docker won't run a container with a cidfile (container id file)
# unless that file does not already exist; clean up an existing
# dockerized kubelet that might be running.
cleanup_dockerized_kubelet
KUBELET_LOG=/tmp/kubelet.log docker run \
if [[ -z "${DOCKERIZE_KUBELET}" ]]; then --volume=/:/rootfs:ro \
sudo -E "${GO_OUT}/kubelet" ${priv_arg}\ --volume=/var/run:/var/run:rw \
--v=${LOG_LEVEL} \ --volume=/sys:/sys:ro \
--chaos_chance="${CHAOS_CHANCE}" \ --volume=/var/lib/docker/:/var/lib/docker:ro \
--container_runtime="${CONTAINER_RUNTIME}" \ --volume=/var/lib/kubelet/:/var/lib/kubelet:rw \
--hostname_override="127.0.0.1" \ --net=host \
--address="127.0.0.1" \ --privileged=true \
--api_servers="${API_HOST}:${API_PORT}" \ -i \
--port="$KUBELET_PORT" >"${KUBELET_LOG}" 2>&1 & --cidfile=$KUBELET_CIDFILE \
KUBELET_PID=$! gcr.io/google_containers/kubelet \
else /kubelet --v=3 --containerized ${priv_arg}--chaos-chance="${CHAOS_CHANCE}" --hostname-override="127.0.0.1" --address="127.0.0.1" --api-servers="${API_HOST}:${API_PORT}" --port="$KUBELET_PORT" --resource-container="" &> $KUBELET_LOG &
# Docker won't run a container with a cidfile (container id file) fi
# unless that file does not already exist; clean up an existing }
# dockerized kubelet that might be running.
cleanup_dockerized_kubelet
docker run \ function start_kubeproxy {
--volume=/:/rootfs:ro \ PROXY_LOG=/tmp/kube-proxy.log
--volume=/var/run:/var/run:rw \ sudo -E "${GO_OUT}/kube-proxy" \
--volume=/sys:/sys:ro \ --v=${LOG_LEVEL} \
--volume=/var/lib/docker/:/var/lib/docker:ro \ --master="http://${API_HOST}:${API_PORT}" >"${PROXY_LOG}" 2>&1 &
--volume=/var/lib/kubelet/:/var/lib/kubelet:rw \ PROXY_PID=$!
--net=host \
--privileged=true \
-i \
--cidfile=$KUBELET_CIDFILE \
gcr.io/google_containers/kubelet \
/kubelet --v=3 --containerized ${priv_arg}--chaos-chance="${CHAOS_CHANCE}" --hostname-override="127.0.0.1" --address="127.0.0.1" --api-servers="${API_HOST}:${API_PORT}" --port="$KUBELET_PORT" --resource-container="" &> $KUBELET_LOG &
fi
PROXY_LOG=/tmp/kube-proxy.log SCHEDULER_LOG=/tmp/kube-scheduler.log
sudo -E "${GO_OUT}/kube-proxy" \ sudo -E "${GO_OUT}/kube-scheduler" \
--v=${LOG_LEVEL} \ --v=${LOG_LEVEL} \
--master="http://${API_HOST}:${API_PORT}" >"${PROXY_LOG}" 2>&1 & --master="http://${API_HOST}:${API_PORT}" >"${SCHEDULER_LOG}" 2>&1 &
PROXY_PID=$! SCHEDULER_PID=$!
}
SCHEDULER_LOG=/tmp/kube-scheduler.log
sudo -E "${GO_OUT}/kube-scheduler" \
--v=${LOG_LEVEL} \
--master="http://${API_HOST}:${API_PORT}" >"${SCHEDULER_LOG}" 2>&1 &
SCHEDULER_PID=$!
function print_success {
cat <<EOF cat <<EOF
Local Kubernetes cluster is running. Press Ctrl-C to shut it down. Local Kubernetes cluster is running. Press Ctrl-C to shut it down.
@ -248,5 +266,23 @@ To start using your cluster, open up another terminal/tab and run:
cluster/kubectl.sh config use-context local cluster/kubectl.sh config use-context local
cluster/kubectl.sh cluster/kubectl.sh
EOF EOF
}
test_docker
test_apiserver_off
detect_binary
echo "Detected host and ready to start services. Doing some housekeeping first..."
GO_OUT="${KUBE_ROOT}/_output/local/bin/${host_os}/${host_arch}"
KUBELET_CIDFILE=/tmp/kubelet.cid
trap cleanup EXIT
echo "Starting services now!"
startETCD
set_service_accounts
start_apiserver
start_controller_manager
start_kubelet
start_kubeproxy
print_success
while true; do sleep 1; done while true; do sleep 1; done