mirror of https://github.com/k3s-io/k3s
Merge branch 'master' of https://github.com/dutchcoders/kubernetes into dutchcoders-master
Conflicts: cluster/aws/config-default.sh cluster/saltbase/salt/kube-controller-manager/defaultpull/6/head
commit
bd712b61e6
|
@ -42,3 +42,5 @@ ENABLE_NODE_MONITORING=true
|
||||||
# Optional: Install node logging
|
# Optional: Install node logging
|
||||||
ENABLE_NODE_LOGGING=true
|
ENABLE_NODE_LOGGING=true
|
||||||
LOGGING_DESTINATION=elasticsearch # options: elasticsearch, gcp
|
LOGGING_DESTINATION=elasticsearch # options: elasticsearch, gcp
|
||||||
|
IAM_PROFILE="kubernetes"
|
||||||
|
LOG="/dev/null"
|
||||||
|
|
|
@ -26,11 +26,8 @@ grains:
|
||||||
EOF
|
EOF
|
||||||
|
|
||||||
cat <<EOF > /etc/aws.conf
|
cat <<EOF > /etc/aws.conf
|
||||||
{
|
[Global]
|
||||||
"Global": {
|
Region:${AWS_ZONE}
|
||||||
"Region": "${AWS_ZONE}"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
EOF
|
EOF
|
||||||
|
|
||||||
# Auto accept all keys from minions that try to join
|
# Auto accept all keys from minions that try to join
|
||||||
|
|
|
@ -29,7 +29,7 @@ function json_val {
|
||||||
# TODO (ayurchuk) Refactor the get_* functions to use filters
|
# TODO (ayurchuk) Refactor the get_* functions to use filters
|
||||||
# TODO (bburns) Parameterize this for multiple cluster per project
|
# TODO (bburns) Parameterize this for multiple cluster per project
|
||||||
function get_instance_ids {
|
function get_instance_ids {
|
||||||
python -c 'import json,sys; lst = [str(instance["InstanceId"]) for reservation in json.load(sys.stdin)["Reservations"] for instance in reservation["Instances"] for tag in instance["Tags"] if tag["Value"].startswith("kubernetes-minion") or tag["Value"].startswith("kubernetes-master")]; print " ".join(lst)'
|
python -c 'import json,sys; lst = [str(instance["InstanceId"]) for reservation in json.load(sys.stdin)["Reservations"] for instance in reservation["Instances"] for tag in instance.get("Tags", []) if tag["Value"].startswith("kubernetes-minion") or tag["Value"].startswith("kubernetes-master")]; print " ".join(lst)'
|
||||||
}
|
}
|
||||||
|
|
||||||
function get_vpc_id {
|
function get_vpc_id {
|
||||||
|
@ -57,7 +57,7 @@ function expect_instance_states {
|
||||||
}
|
}
|
||||||
|
|
||||||
function get_instance_public_ip {
|
function get_instance_public_ip {
|
||||||
python -c "import json,sys; lst = [str(instance['NetworkInterfaces'][0]['Association']['PublicIp']) for reservation in json.load(sys.stdin)['Reservations'] for instance in reservation['Instances'] for tag in instance['Tags'] if tag['Value'] == '$1' and instance['State']['Name'] == 'running']; print ' '.join(lst)"
|
python -c "import json,sys; lst = [str(instance['NetworkInterfaces'][0]['Association']['PublicIp']) for reservation in json.load(sys.stdin)['Reservations'] for instance in reservation['Instances'] for tag in instance.get('Tags', []) if tag['Value'] == '$1' and instance['State']['Name'] == 'running']; print ' '.join(lst)"
|
||||||
}
|
}
|
||||||
|
|
||||||
function detect-master () {
|
function detect-master () {
|
||||||
|
@ -228,20 +228,55 @@ function kube-up {
|
||||||
ssh-keygen -f $AWS_SSH_KEY -N ''
|
ssh-keygen -f $AWS_SSH_KEY -N ''
|
||||||
fi
|
fi
|
||||||
|
|
||||||
$AWS_CMD import-key-pair --key-name kubernetes --public-key-material file://$AWS_SSH_KEY.pub > /dev/null 2>&1 || true
|
$AWS_CMD import-key-pair --key-name kubernetes --public-key-material file://$AWS_SSH_KEY.pub > $LOG 2>&1 || true
|
||||||
VPC_ID=$($AWS_CMD create-vpc --cidr-block 172.20.0.0/16 | json_val '["Vpc"]["VpcId"]')
|
|
||||||
$AWS_CMD modify-vpc-attribute --vpc-id $VPC_ID --enable-dns-support '{"Value": true}' > /dev/null
|
VPC_ID=$($AWS_CMD describe-vpcs | get_vpc_id)
|
||||||
$AWS_CMD modify-vpc-attribute --vpc-id $VPC_ID --enable-dns-hostnames '{"Value": true}' > /dev/null
|
|
||||||
$AWS_CMD create-tags --resources $VPC_ID --tags Key=Name,Value=kubernetes-vpc > /dev/null
|
if [ -z "$VPC_ID" ]; then
|
||||||
SUBNET_ID=$($AWS_CMD create-subnet --cidr-block 172.20.0.0/24 --vpc-id $VPC_ID | json_val '["Subnet"]["SubnetId"]')
|
echo "Creating vpc."
|
||||||
IGW_ID=$($AWS_CMD create-internet-gateway | json_val '["InternetGateway"]["InternetGatewayId"]')
|
VPC_ID=$($AWS_CMD create-vpc --cidr-block 172.20.0.0/16 | json_val '["Vpc"]["VpcId"]')
|
||||||
$AWS_CMD attach-internet-gateway --internet-gateway-id $IGW_ID --vpc-id $VPC_ID > /dev/null
|
$AWS_CMD modify-vpc-attribute --vpc-id $VPC_ID --enable-dns-support '{"Value": true}' > $LOG
|
||||||
|
$AWS_CMD modify-vpc-attribute --vpc-id $VPC_ID --enable-dns-hostnames '{"Value": true}' > $LOG
|
||||||
|
$AWS_CMD create-tags --resources $VPC_ID --tags Key=Name,Value=kubernetes-vpc > $LOG
|
||||||
|
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo "Using VPC $VPC_ID"
|
||||||
|
|
||||||
|
SUBNET_ID=$($AWS_CMD describe-subnets | get_subnet_id $VPC_ID)
|
||||||
|
if [ -z "$SUBNET_ID" ]; then
|
||||||
|
echo "Creating subnet."
|
||||||
|
SUBNET_ID=$($AWS_CMD create-subnet --cidr-block 172.20.0.0/24 --vpc-id $VPC_ID | json_val '["Subnet"]["SubnetId"]')
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo "Using subnet $SUBNET_ID"
|
||||||
|
|
||||||
|
IGW_ID=$($AWS_CMD describe-internet-gateways | get_igw_id $VPC_ID)
|
||||||
|
if [ -z "$IGW_ID" ]; then
|
||||||
|
echo "Creating Internet Gateway."
|
||||||
|
IGW_ID=$($AWS_CMD create-internet-gateway | json_val '["InternetGateway"]["InternetGatewayId"]')
|
||||||
|
$AWS_CMD attach-internet-gateway --internet-gateway-id $IGW_ID --vpc-id $VPC_ID > $LOG
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo "Using Internet Gateway $IGW_ID"
|
||||||
|
|
||||||
|
echo "Associating route table."
|
||||||
ROUTE_TABLE_ID=$($AWS_CMD describe-route-tables --filters Name=vpc-id,Values=$VPC_ID | json_val '["RouteTables"][0]["RouteTableId"]')
|
ROUTE_TABLE_ID=$($AWS_CMD describe-route-tables --filters Name=vpc-id,Values=$VPC_ID | json_val '["RouteTables"][0]["RouteTableId"]')
|
||||||
$AWS_CMD associate-route-table --route-table-id $ROUTE_TABLE_ID --subnet-id $SUBNET_ID > /dev/null
|
$AWS_CMD associate-route-table --route-table-id $ROUTE_TABLE_ID --subnet-id $SUBNET_ID > $LOG || true
|
||||||
$AWS_CMD describe-route-tables --filters Name=vpc-id,Values=$VPC_ID > /dev/null
|
echo "Configuring route table."
|
||||||
$AWS_CMD create-route --route-table-id $ROUTE_TABLE_ID --destination-cidr-block 0.0.0.0/0 --gateway-id $IGW_ID > /dev/null
|
$AWS_CMD describe-route-tables --filters Name=vpc-id,Values=$VPC_ID > $LOG || true
|
||||||
SEC_GROUP_ID=$($AWS_CMD create-security-group --group-name kubernetes-sec-group --description kubernetes-sec-group --vpc-id $VPC_ID | json_val '["GroupId"]')
|
echo "Adding route to route table."
|
||||||
$AWS_CMD authorize-security-group-ingress --group-id $SEC_GROUP_ID --protocol -1 --port all --cidr 0.0.0.0/0 > /dev/null
|
$AWS_CMD create-route --route-table-id $ROUTE_TABLE_ID --destination-cidr-block 0.0.0.0/0 --gateway-id $IGW_ID > $LOG || true
|
||||||
|
|
||||||
|
echo "Using Route Table $ROUTE_TABLE_ID"
|
||||||
|
|
||||||
|
SEC_GROUP_ID=$($AWS_CMD describe-security-groups | get_sec_group_id)
|
||||||
|
|
||||||
|
if [ -z "$SEC_GROUP_ID" ]; then
|
||||||
|
echo "Creating security group."
|
||||||
|
SEC_GROUP_ID=$($AWS_CMD create-security-group --group-name kubernetes-sec-group --description kubernetes-sec-group --vpc-id $VPC_ID | json_val '["GroupId"]')
|
||||||
|
$AWS_CMD authorize-security-group-ingress --group-id $SEC_GROUP_ID --protocol -1 --port all --cidr 0.0.0.0/0 > $LOG
|
||||||
|
fi
|
||||||
|
|
||||||
(
|
(
|
||||||
# We pipe this to the ami as a startup script in the user-data field. Requires a compatible ami
|
# We pipe this to the ami as a startup script in the user-data field. Requires a compatible ami
|
||||||
|
@ -250,8 +285,8 @@ function kube-up {
|
||||||
echo "cd /var/cache/kubernetes-install"
|
echo "cd /var/cache/kubernetes-install"
|
||||||
echo "readonly MASTER_NAME='${MASTER_NAME}'"
|
echo "readonly MASTER_NAME='${MASTER_NAME}'"
|
||||||
echo "readonly NODE_INSTANCE_PREFIX='${INSTANCE_PREFIX}-minion'"
|
echo "readonly NODE_INSTANCE_PREFIX='${INSTANCE_PREFIX}-minion'"
|
||||||
echo "readonly SERVER_BINARY_TAR_URL='https://s3-${ZONE}.amazonaws.com/${SERVER_BINARY_TAR_URL}'"
|
echo "readonly SERVER_BINARY_TAR_URL='https://s3.amazonaws.com/${SERVER_BINARY_TAR_URL}'"
|
||||||
echo "readonly SALT_TAR_URL='https://s3-${ZONE}.amazonaws.com/${SALT_TAR_URL}'"
|
echo "readonly SALT_TAR_URL='https://s3.amazonaws.com/${SALT_TAR_URL}'"
|
||||||
echo "readonly AWS_ZONE='${ZONE}'"
|
echo "readonly AWS_ZONE='${ZONE}'"
|
||||||
echo "readonly MASTER_HTPASSWD='${htpasswd}'"
|
echo "readonly MASTER_HTPASSWD='${htpasswd}'"
|
||||||
echo "readonly PORTAL_NET='${PORTAL_NET}'"
|
echo "readonly PORTAL_NET='${PORTAL_NET}'"
|
||||||
|
@ -263,9 +298,10 @@ function kube-up {
|
||||||
grep -v "^#" "${KUBE_ROOT}/cluster/aws/templates/salt-master.sh"
|
grep -v "^#" "${KUBE_ROOT}/cluster/aws/templates/salt-master.sh"
|
||||||
) > "${KUBE_TEMP}/master-start.sh"
|
) > "${KUBE_TEMP}/master-start.sh"
|
||||||
|
|
||||||
|
echo "Starting Master"
|
||||||
master_id=$($AWS_CMD run-instances \
|
master_id=$($AWS_CMD run-instances \
|
||||||
--image-id $IMAGE \
|
--image-id $IMAGE \
|
||||||
|
--iam-instance-profile Name=$IAM_PROFILE \
|
||||||
--instance-type $MASTER_SIZE \
|
--instance-type $MASTER_SIZE \
|
||||||
--subnet-id $SUBNET_ID \
|
--subnet-id $SUBNET_ID \
|
||||||
--private-ip-address 172.20.0.9 \
|
--private-ip-address 172.20.0.9 \
|
||||||
|
@ -274,11 +310,12 @@ function kube-up {
|
||||||
--associate-public-ip-address \
|
--associate-public-ip-address \
|
||||||
--user-data file://${KUBE_TEMP}/master-start.sh | json_val '["Instances"][0]["InstanceId"]')
|
--user-data file://${KUBE_TEMP}/master-start.sh | json_val '["Instances"][0]["InstanceId"]')
|
||||||
sleep 3
|
sleep 3
|
||||||
$AWS_CMD create-tags --resources $master_id --tags Key=Name,Value=$MASTER_NAME > /dev/null
|
$AWS_CMD create-tags --resources $master_id --tags Key=Name,Value=$MASTER_NAME > $LOG
|
||||||
sleep 3
|
sleep 3
|
||||||
$AWS_CMD create-tags --resources $master_id --tags Key=Role,Value=$MASTER_TAG > /dev/null
|
$AWS_CMD create-tags --resources $master_id --tags Key=Role,Value=$MASTER_TAG > $LOG
|
||||||
|
|
||||||
for (( i=0; i<${#MINION_NAMES[@]}; i++)); do
|
for (( i=0; i<${#MINION_NAMES[@]}; i++)); do
|
||||||
|
echo "Starting Minion (${MINION_NAMES[$i]})"
|
||||||
(
|
(
|
||||||
# We pipe this to the ami as a startup script in the user-data field. Requires a compatible ami
|
# We pipe this to the ami as a startup script in the user-data field. Requires a compatible ami
|
||||||
echo "#! /bin/bash"
|
echo "#! /bin/bash"
|
||||||
|
@ -288,6 +325,7 @@ function kube-up {
|
||||||
) > "${KUBE_TEMP}/minion-start-${i}.sh"
|
) > "${KUBE_TEMP}/minion-start-${i}.sh"
|
||||||
minion_id=$($AWS_CMD run-instances \
|
minion_id=$($AWS_CMD run-instances \
|
||||||
--image-id $IMAGE \
|
--image-id $IMAGE \
|
||||||
|
--iam-instance-profile Name=$IAM_PROFILE \
|
||||||
--instance-type $MINION_SIZE \
|
--instance-type $MINION_SIZE \
|
||||||
--subnet-id $SUBNET_ID \
|
--subnet-id $SUBNET_ID \
|
||||||
--private-ip-address 172.20.0.1${i} \
|
--private-ip-address 172.20.0.1${i} \
|
||||||
|
@ -298,7 +336,7 @@ function kube-up {
|
||||||
sleep 3
|
sleep 3
|
||||||
n=0
|
n=0
|
||||||
until [ $n -ge 5 ]; do
|
until [ $n -ge 5 ]; do
|
||||||
$AWS_CMD create-tags --resources $minion_id --tags Key=Name,Value=${MINION_NAMES[$i]} > /dev/null && break
|
$AWS_CMD create-tags --resources $minion_id --tags Key=Name,Value=${MINION_NAMES[$i]} > $LOG && break
|
||||||
n=$[$n+1]
|
n=$[$n+1]
|
||||||
sleep 15
|
sleep 15
|
||||||
done
|
done
|
||||||
|
@ -306,13 +344,13 @@ function kube-up {
|
||||||
sleep 3
|
sleep 3
|
||||||
n=0
|
n=0
|
||||||
until [ $n -ge 5 ]; do
|
until [ $n -ge 5 ]; do
|
||||||
$AWS_CMD create-tags --resources $minion_id --tags Key=Role,Value=$MINION_TAG > /dev/null && break
|
$AWS_CMD create-tags --resources $minion_id --tags Key=Role,Value=$MINION_TAG > $LOG && break
|
||||||
n=$[$n+1]
|
n=$[$n+1]
|
||||||
sleep 15
|
sleep 15
|
||||||
done
|
done
|
||||||
|
|
||||||
sleep 3
|
sleep 3
|
||||||
$AWS_CMD modify-instance-attribute --instance-id $minion_id --source-dest-check '{"Value": false}' > /dev/null
|
$AWS_CMD modify-instance-attribute --instance-id $minion_id --source-dest-check '{"Value": false}' > $LOG
|
||||||
|
|
||||||
# We are not able to add a route to the instance until that instance is in "running" state.
|
# We are not able to add a route to the instance until that instance is in "running" state.
|
||||||
# This is quite an ugly solution to this problem. In Bash 4 we could use assoc. arrays to do this for
|
# This is quite an ugly solution to this problem. In Bash 4 we could use assoc. arrays to do this for
|
||||||
|
@ -322,7 +360,7 @@ function kube-up {
|
||||||
if [[ "$instance_state" == "" ]]; then
|
if [[ "$instance_state" == "" ]]; then
|
||||||
echo "Minion ${MINION_NAMES[$i]} running"
|
echo "Minion ${MINION_NAMES[$i]} running"
|
||||||
sleep 10
|
sleep 10
|
||||||
$AWS_CMD create-route --route-table-id $ROUTE_TABLE_ID --destination-cidr-block ${MINION_IP_RANGES[$i]} --instance-id $minion_id > /dev/null
|
$AWS_CMD create-route --route-table-id $ROUTE_TABLE_ID --destination-cidr-block ${MINION_IP_RANGES[$i]} --instance-id $minion_id > $LOG
|
||||||
break
|
break
|
||||||
else
|
else
|
||||||
echo "Waiting for minion ${MINION_NAMES[$i]} to spawn"
|
echo "Waiting for minion ${MINION_NAMES[$i]} to spawn"
|
||||||
|
@ -341,8 +379,8 @@ function kube-up {
|
||||||
exit 2
|
exit 2
|
||||||
fi
|
fi
|
||||||
|
|
||||||
detect-master > /dev/null
|
detect-master > $LOG
|
||||||
detect-minions > /dev/null
|
detect-minions > $LOG
|
||||||
|
|
||||||
# Wait 3 minutes for cluster to come up. We hit it with a "highstate" after that to
|
# Wait 3 minutes for cluster to come up. We hit it with a "highstate" after that to
|
||||||
# make sure that everything is well configured.
|
# make sure that everything is well configured.
|
||||||
|
@ -353,7 +391,7 @@ function kube-up {
|
||||||
sleep 10
|
sleep 10
|
||||||
done
|
done
|
||||||
echo "Re-running salt highstate"
|
echo "Re-running salt highstate"
|
||||||
ssh -oStrictHostKeyChecking=no -i ~/.ssh/kube_aws_rsa ubuntu@${KUBE_MASTER_IP} sudo salt '*' state.highstate > /dev/null
|
ssh -oStrictHostKeyChecking=no -i ~/.ssh/kube_aws_rsa ubuntu@${KUBE_MASTER_IP} sudo salt '*' state.highstate > $LOG
|
||||||
|
|
||||||
echo "Waiting for cluster initialization."
|
echo "Waiting for cluster initialization."
|
||||||
echo
|
echo
|
||||||
|
@ -363,7 +401,7 @@ function kube-up {
|
||||||
echo
|
echo
|
||||||
|
|
||||||
until $(curl --insecure --user ${KUBE_USER}:${KUBE_PASSWORD} --max-time 5 \
|
until $(curl --insecure --user ${KUBE_USER}:${KUBE_PASSWORD} --max-time 5 \
|
||||||
--fail --output /dev/null --silent https://${KUBE_MASTER_IP}/api/v1beta1/pods); do
|
--fail --output $LOG --silent https://${KUBE_MASTER_IP}/api/v1beta1/pods); do
|
||||||
printf "."
|
printf "."
|
||||||
sleep 2
|
sleep 2
|
||||||
done
|
done
|
||||||
|
@ -379,7 +417,7 @@ function kube-up {
|
||||||
# Basic sanity checking
|
# Basic sanity checking
|
||||||
for i in ${KUBE_MINION_IP_ADDRESSES[@]}; do
|
for i in ${KUBE_MINION_IP_ADDRESSES[@]}; do
|
||||||
# Make sure docker is installed
|
# Make sure docker is installed
|
||||||
ssh -oStrictHostKeyChecking=no ubuntu@$i -i ~/.ssh/kube_aws_rsa which docker > /dev/null 2>&1
|
ssh -oStrictHostKeyChecking=no ubuntu@$i -i ~/.ssh/kube_aws_rsa which docker > $LOG 2>&1
|
||||||
if [ "$?" != "0" ]; then
|
if [ "$?" != "0" ]; then
|
||||||
echo "Docker failed to install on $i. Your cluster is unlikely to work correctly."
|
echo "Docker failed to install on $i. Your cluster is unlikely to work correctly."
|
||||||
echo "Please run ./cluster/aws/kube-down.sh and re-create the cluster. (sorry!)"
|
echo "Please run ./cluster/aws/kube-down.sh and re-create the cluster. (sorry!)"
|
||||||
|
@ -401,9 +439,9 @@ function kube-up {
|
||||||
# config file. Distribute the same way the htpasswd is done.
|
# config file. Distribute the same way the htpasswd is done.
|
||||||
(
|
(
|
||||||
umask 077
|
umask 077
|
||||||
ssh -oStrictHostKeyChecking=no -i ~/.ssh/kube_aws_rsa ubuntu@${KUBE_MASTER_IP} sudo cat /srv/kubernetes/kubecfg.crt >"${HOME}/${kube_cert}" 2>/dev/null
|
ssh -oStrictHostKeyChecking=no -i ~/.ssh/kube_aws_rsa ubuntu@${KUBE_MASTER_IP} sudo cat /srv/kubernetes/kubecfg.crt >"${HOME}/${kube_cert}" 2>$LOG
|
||||||
ssh -oStrictHostKeyChecking=no -i ~/.ssh/kube_aws_rsa ubuntu@${KUBE_MASTER_IP} sudo cat /srv/kubernetes/kubecfg.key >"${HOME}/${kube_key}" 2>/dev/null
|
ssh -oStrictHostKeyChecking=no -i ~/.ssh/kube_aws_rsa ubuntu@${KUBE_MASTER_IP} sudo cat /srv/kubernetes/kubecfg.key >"${HOME}/${kube_key}" 2>$LOG
|
||||||
ssh -oStrictHostKeyChecking=no -i ~/.ssh/kube_aws_rsa ubuntu@${KUBE_MASTER_IP} sudo cat /srv/kubernetes/ca.crt >"${HOME}/${ca_cert}" 2>/dev/null
|
ssh -oStrictHostKeyChecking=no -i ~/.ssh/kube_aws_rsa ubuntu@${KUBE_MASTER_IP} sudo cat /srv/kubernetes/ca.crt >"${HOME}/${ca_cert}" 2>$LOG
|
||||||
|
|
||||||
cat << EOF > ~/.kubernetes_auth
|
cat << EOF > ~/.kubernetes_auth
|
||||||
{
|
{
|
||||||
|
@ -423,7 +461,7 @@ EOF
|
||||||
function kube-down {
|
function kube-down {
|
||||||
AWS_CMD="aws --output json ec2"
|
AWS_CMD="aws --output json ec2"
|
||||||
instance_ids=$($AWS_CMD describe-instances | get_instance_ids)
|
instance_ids=$($AWS_CMD describe-instances | get_instance_ids)
|
||||||
$AWS_CMD terminate-instances --instance-ids $instance_ids > /dev/null
|
$AWS_CMD terminate-instances --instance-ids $instance_ids > $LOG
|
||||||
echo "Waiting for instances deleted"
|
echo "Waiting for instances deleted"
|
||||||
while true; do
|
while true; do
|
||||||
instance_states=$($AWS_CMD describe-instances --instance-ids $instance_ids | expect_instance_states terminated)
|
instance_states=$($AWS_CMD describe-instances --instance-ids $instance_ids | expect_instance_states terminated)
|
||||||
|
@ -444,10 +482,10 @@ function kube-down {
|
||||||
route_table_id=$($AWS_CMD describe-route-tables | get_route_table_id $vpc_id)
|
route_table_id=$($AWS_CMD describe-route-tables | get_route_table_id $vpc_id)
|
||||||
sec_group_id=$($AWS_CMD describe-security-groups | get_sec_group_id)
|
sec_group_id=$($AWS_CMD describe-security-groups | get_sec_group_id)
|
||||||
|
|
||||||
$AWS_CMD delete-subnet --subnet-id $subnet_id > /dev/null
|
$AWS_CMD delete-subnet --subnet-id $subnet_id > $LOG
|
||||||
$AWS_CMD detach-internet-gateway --internet-gateway-id $igw_id --vpc-id $vpc_id > /dev/null
|
$AWS_CMD detach-internet-gateway --internet-gateway-id $igw_id --vpc-id $vpc_id > $LOG
|
||||||
$AWS_CMD delete-internet-gateway --internet-gateway-id $igw_id > /dev/null
|
$AWS_CMD delete-internet-gateway --internet-gateway-id $igw_id > $LOG
|
||||||
$AWS_CMD delete-security-group --group-id $sec_group_id > /dev/null
|
$AWS_CMD delete-security-group --group-id $sec_group_id > $LOG
|
||||||
$AWS_CMD delete-route --route-table-id $route_table_id --destination-cidr-block 0.0.0.0/0 > /dev/null
|
$AWS_CMD delete-route --route-table-id $route_table_id --destination-cidr-block 0.0.0.0/0 > $LOG
|
||||||
$AWS_CMD delete-vpc --vpc-id $vpc_id > /dev/null
|
$AWS_CMD delete-vpc --vpc-id $vpc_id > $LOG
|
||||||
}
|
}
|
||||||
|
|
|
@ -22,8 +22,8 @@
|
||||||
{% set machines = "-machines=" + ','.join(salt['mine.get']('roles:kubernetes-pool', 'network.ip_addrs', expr_form='grain').keys()) -%}
|
{% set machines = "-machines=" + ','.join(salt['mine.get']('roles:kubernetes-pool', 'network.ip_addrs', expr_form='grain').keys()) -%}
|
||||||
{% endif -%}
|
{% endif -%}
|
||||||
{% if grains.cloud == 'aws' -%}
|
{% if grains.cloud == 'aws' -%}
|
||||||
# TODO: turn on AWS as a cloud provider. Need to figure out AWS role-based auth.
|
{% set cloud_provider = "-cloud_provider=aws" -%}
|
||||||
{% set cloud_provider = "" -%}
|
{% set cloud_config = "-cloud_config=/etc/aws.conf" -%}
|
||||||
{% set minion_regexp = "" -%}
|
{% set minion_regexp = "" -%}
|
||||||
{% set machines = "-machines " + ','.join(salt['mine.get']('roles:kubernetes-pool', 'network.ip_addrs', expr_form='grain').keys()) -%}
|
{% set machines = "-machines " + ','.join(salt['mine.get']('roles:kubernetes-pool', 'network.ip_addrs', expr_form='grain').keys()) -%}
|
||||||
{% endif -%}
|
{% endif -%}
|
||||||
|
@ -46,4 +46,8 @@
|
||||||
{% endif -%}
|
{% endif -%}
|
||||||
{% endif -%}
|
{% endif -%}
|
||||||
|
|
||||||
|
<<<<<<< HEAD
|
||||||
DAEMON_ARGS="{{daemon_args}} {{master}} {{machines}} {{ minion_regexp }} {{ cloud_provider }} {{pillar['log_level']}}"
|
DAEMON_ARGS="{{daemon_args}} {{master}} {{machines}} {{ minion_regexp }} {{ cloud_provider }} {{pillar['log_level']}}"
|
||||||
|
=======
|
||||||
|
DAEMON_ARGS="{{daemon_args}} {{master}} {{machines}} {{ minion_regexp }} {{ cloud_provider }} {{ cloud_config }}"
|
||||||
|
>>>>>>> 83f3982eee4d36962919d4f55a41ec1c2111d01a
|
||||||
|
|
Loading…
Reference in New Issue