
the firewall rules created in e2e tests. This allows the teardown code to run without needing to inspect the managed instance group for the cluster (which may no longer exist) and should make e2e teardown much more resilient.
321 lines
9.7 KiB
Bash
Executable File
321 lines
9.7 KiB
Bash
Executable File
#!/bin/bash
|
|
|
|
# Copyright 2014 The Kubernetes Authors All rights reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
# A library of helper functions and constant for the local config.
|
|
|
|
# Use the config file specified in $KUBE_CONFIG_FILE, or default to
|
|
# config-default.sh.
|
|
|
|
KUBE_PROMPT_FOR_UPDATE=y
|
|
KUBE_SKIP_UPDATE=${KUBE_SKIP_UPDATE-"n"}
|
|
KUBE_ROOT=$(dirname "${BASH_SOURCE}")/../..
|
|
source "${KUBE_ROOT}/cluster/gke/${KUBE_CONFIG_FILE:-config-default.sh}"
|
|
|
|
# Perform preparations required to run e2e tests
|
|
#
|
|
# Assumed vars:
|
|
# GCLOUD
|
|
function prepare-e2e() {
|
|
echo "... in gke:prepare-e2e()" >&2
|
|
|
|
# Ensure GCLOUD is set to some gcloud binary.
|
|
if [[ -z "${GCLOUD:-}" ]]; then
|
|
echo "GCLOUD environment variable is not set. It should be your gcloud binary. " >&2
|
|
echo "A sane default is probably \$ export GCLOUD=gcloud" >&2
|
|
exit 1
|
|
fi
|
|
}
|
|
|
|
|
|
# Use the gcloud defaults to find the project. If it is already set in the
|
|
# environment then go with that.
|
|
#
|
|
# Assumed vars:
|
|
# GCLOUD
|
|
# Vars set:
|
|
# PROJECT
|
|
function detect-project() {
|
|
echo "... in gke:detect-project()" >&2
|
|
if [[ -z "${PROJECT:-}" ]]; then
|
|
export PROJECT=$("${GCLOUD}" config list project | tail -n 1 | cut -f 3 -d ' ')
|
|
echo "... Using project: ${PROJECT}" >&2
|
|
fi
|
|
if [[ -z "${PROJECT:-}" ]]; then
|
|
echo "Could not detect Google Cloud Platform project. Set the default project using " >&2
|
|
echo "'gcloud config set project <PROJECT>'" >&2
|
|
exit 1
|
|
fi
|
|
}
|
|
|
|
# Execute prior to running tests to build a release if required for env.
|
|
function test-build-release() {
|
|
echo "... in gke:test-build-release()" >&2
|
|
# We currently use the Kubernetes version that GKE supports (not testing
|
|
# bleeding-edge builds).
|
|
}
|
|
|
|
# Verify needed binaries exist.
|
|
function verify-prereqs() {
|
|
echo "... in gke:verify-prereqs()" >&2
|
|
if ! which gcloud >/dev/null; then
|
|
local resp
|
|
if [[ "${KUBE_PROMPT_FOR_UPDATE}" == "y" ]]; then
|
|
echo "Can't find gcloud in PATH. Do you wish to install the Google Cloud SDK? [Y/n]"
|
|
read resp
|
|
else
|
|
resp="y"
|
|
fi
|
|
if [[ "${resp}" != "n" && "${resp}" != "N" ]]; then
|
|
curl https://sdk.cloud.google.com | bash
|
|
fi
|
|
if ! which gcloud >/dev/null; then
|
|
echo "Can't find gcloud in PATH, please fix and retry. The Google Cloud "
|
|
echo "SDK can be downloaded from https://cloud.google.com/sdk/."
|
|
exit 1
|
|
fi
|
|
fi
|
|
if [[ "${KUBE_SKIP_UPDATE}" == "y" ]]; then
|
|
return
|
|
fi
|
|
# update and install components as needed
|
|
if [[ "${KUBE_PROMPT_FOR_UPDATE}" != "y" ]]; then
|
|
gcloud_prompt="-q"
|
|
fi
|
|
local sudo_prefix=""
|
|
if [ ! -w $(dirname `which gcloud`) ]; then
|
|
sudo_prefix="sudo"
|
|
fi
|
|
${sudo_prefix} gcloud ${gcloud_prompt:-} components update preview || true
|
|
${sudo_prefix} gcloud ${gcloud_prompt:-} components update ${CMD_GROUP:-} || true
|
|
${sudo_prefix} gcloud ${gcloud_prompt:-} components update kubectl|| true
|
|
${sudo_prefix} gcloud ${gcloud_prompt:-} components update || true
|
|
}
|
|
|
|
# Instantiate a kubernetes cluster
|
|
#
|
|
# Assumed vars:
|
|
# GCLOUD
|
|
# CLUSTER_NAME
|
|
# ZONE
|
|
# CLUSTER_API_VERSION (optional)
|
|
# NUM_MINIONS
|
|
# MINION_SCOPES
|
|
# MACHINE_TYPE
|
|
function kube-up() {
|
|
echo "... in gke:kube-up()" >&2
|
|
detect-project >&2
|
|
|
|
# Make the specified network if we need to.
|
|
if ! "${GCLOUD}" compute networks --project "${PROJECT}" describe "${NETWORK}" &>/dev/null; then
|
|
echo "Creating new network: ${NETWORK}" >&2
|
|
"${GCLOUD}" compute networks create "${NETWORK}" --project="${PROJECT}" --range "${NETWORK_RANGE}"
|
|
else
|
|
echo "... Using network: ${NETWORK}" >&2
|
|
fi
|
|
|
|
# Allow SSH on all nodes in the network. This doesn't actually check whether
|
|
# such a rule exists, only whether we've created this exact rule.
|
|
if ! "${GCLOUD}" compute firewall-rules --project "${PROJECT}" describe "${FIREWALL_SSH}" &>/dev/null; then
|
|
echo "Creating new firewall for SSH: ${FIREWALL_SSH}" >&2
|
|
"${GCLOUD}" compute firewall-rules create "${FIREWALL_SSH}" \
|
|
--allow="tcp:22" \
|
|
--network="${NETWORK}" \
|
|
--project="${PROJECT}" \
|
|
--source-ranges="0.0.0.0/0"
|
|
else
|
|
echo "... Using firewall-rule: ${FIREWALL_SSH}" >&2
|
|
fi
|
|
|
|
local create_args=(
|
|
"--zone=${ZONE}"
|
|
"--project=${PROJECT}"
|
|
"--num-nodes=${NUM_MINIONS}"
|
|
"--network=${NETWORK}"
|
|
"--scopes=${MINION_SCOPES}"
|
|
"--cluster-version=${CLUSTER_API_VERSION}"
|
|
"--machine-type=${MACHINE_TYPE}"
|
|
)
|
|
|
|
# Bring up the cluster.
|
|
"${GCLOUD}" ${CMD_GROUP:-} container clusters create "${CLUSTER_NAME}" "${create_args[@]}"
|
|
}
|
|
|
|
# Execute prior to running tests to initialize required structure. This is
|
|
# called from hack/e2e-go only when running -up (it is run after kube-up, so
|
|
# the cluster already exists at this point).
|
|
#
|
|
# Assumed vars:
|
|
# CLUSTER_NAME
|
|
# GCLOUD
|
|
# ZONE
|
|
# Vars set:
|
|
# MINION_TAG
|
|
function test-setup() {
|
|
echo "... in gke:test-setup()" >&2
|
|
# Detect the project into $PROJECT if it isn't set
|
|
detect-project >&2
|
|
detect-minions >&2
|
|
|
|
# At this point, CLUSTER_NAME should have been used, so its value is final.
|
|
MINION_TAG=$($GCLOUD compute instances describe ${MINION_NAMES[0]} --project="${PROJECT}" --zone="${ZONE}" | grep -o "gke-${CLUSTER_NAME}-.\{8\}-node" | head -1)
|
|
OLD_MINION_TAG="k8s-${CLUSTER_NAME}-node"
|
|
|
|
# Open up port 80 & 8080 so common containers on minions can be reached.
|
|
"${GCLOUD}" compute firewall-rules create \
|
|
"${CLUSTER_NAME}-http-alt" \
|
|
--allow tcp:80,tcp:8080 \
|
|
--project "${PROJECT}" \
|
|
--target-tags "${MINION_TAG},${OLD_MINION_TAG}" \
|
|
--network="${NETWORK}"
|
|
|
|
"${GCLOUD}" compute firewall-rules create \
|
|
"${CLUSTER_NAME}-nodeports" \
|
|
--allow tcp:30000-32767,udp:30000-32767 \
|
|
--project "${PROJECT}" \
|
|
--target-tags "${MINION_TAG},${OLD_MINION_TAG}" \
|
|
--network="${NETWORK}"
|
|
}
|
|
|
|
# Detect the IP for the master. Note that on GKE, we don't know the name of the
|
|
# master, so KUBE_MASTER is not set.
|
|
#
|
|
# Assumed vars:
|
|
# ZONE
|
|
# CLUSTER_NAME
|
|
# Vars set:
|
|
# KUBE_MASTER_IP
|
|
function detect-master() {
|
|
echo "... in gke:detect-master()" >&2
|
|
detect-project >&2
|
|
KUBE_MASTER_IP=$("${GCLOUD}" ${CMD_GROUP:-} container clusters describe \
|
|
--project="${PROJECT}" --zone="${ZONE}" "${CLUSTER_NAME}" \
|
|
| grep endpoint | cut -f 2 -d ' ')
|
|
}
|
|
|
|
# Assumed vars:
|
|
# none
|
|
# Vars set:
|
|
# MINION_NAMES
|
|
function detect-minions() {
|
|
echo "... in gke:detect-minions()" >&2
|
|
detect-minion-names
|
|
}
|
|
|
|
# Detect minions created in the minion group
|
|
#
|
|
# Assumed vars:
|
|
# none
|
|
# Vars set:
|
|
# MINION_NAMES
|
|
function detect-minion-names {
|
|
echo "... in gke:detect-minion-names()" >&2
|
|
detect-project
|
|
detect-node-instance-group
|
|
MINION_NAMES=($(gcloud compute instance-groups managed list-instances \
|
|
"${NODE_INSTANCE_GROUP}" --zone "${ZONE}" --project "${PROJECT}" \
|
|
--format=yaml | grep instance: | cut -d ' ' -f 2))
|
|
|
|
echo "MINION_NAMES=${MINION_NAMES[*]}"
|
|
}
|
|
|
|
# Detect instance group name generated by gke
|
|
#
|
|
# Assumed vars:
|
|
# GCLOUD
|
|
# PROJECT
|
|
# ZONE
|
|
# CLUSTER_NAME
|
|
# Vars set:
|
|
# NODE_INSTANCE_GROUP
|
|
function detect-node-instance-group {
|
|
echo "... in gke:detect-node-instance-group()" >&2
|
|
NODE_INSTANCE_GROUP=$("${GCLOUD}" ${CMD_GROUP:-} container clusters describe \
|
|
--project="${PROJECT}" --zone="${ZONE}" "${CLUSTER_NAME}" \
|
|
| grep instanceGroupManagers | cut -d '/' -f 11)
|
|
}
|
|
|
|
# SSH to a node by name ($1) and run a command ($2).
|
|
#
|
|
# Assumed vars:
|
|
# GCLOUD
|
|
# ZONE
|
|
function ssh-to-node() {
|
|
echo "... in gke:ssh-to-node()" >&2
|
|
detect-project >&2
|
|
|
|
local node="$1"
|
|
local cmd="$2"
|
|
# Loop until we can successfully ssh into the box
|
|
for try in $(seq 1 5); do
|
|
if gcloud compute ssh --ssh-flag="-o LogLevel=quiet" --project "${PROJECT}" --zone="${ZONE}" "${node}" --command "echo test > /dev/null"; then
|
|
break
|
|
fi
|
|
sleep 5
|
|
done
|
|
# Then actually try the command.
|
|
gcloud compute ssh --ssh-flag="-o LogLevel=quiet" --project "${PROJECT}" --zone="${ZONE}" "${node}" --command "${cmd}"
|
|
}
|
|
|
|
# Restart the kube-proxy on a node ($1)
|
|
function restart-kube-proxy() {
|
|
echo "... in gke:restart-kube-proxy()" >&2
|
|
ssh-to-node "$1" "sudo /etc/init.d/kube-proxy restart"
|
|
}
|
|
|
|
# Restart the kube-proxy on master ($1)
|
|
function restart-apiserver() {
|
|
echo "... in gke:restart-apiserver()" >&2
|
|
ssh-to-node "$1" "sudo docker ps | grep /kube-apiserver | cut -d ' ' -f 1 | xargs sudo docker kill"
|
|
}
|
|
|
|
# Execute after running tests to perform any required clean-up. This is called
|
|
# from hack/e2e-test.sh. This calls kube-down, so the cluster still exists when
|
|
# this is called.
|
|
#
|
|
# Assumed vars:
|
|
# CLUSTER_NAME
|
|
# GCLOUD
|
|
# KUBE_ROOT
|
|
# ZONE
|
|
function test-teardown() {
|
|
echo "... in gke:test-teardown()" >&2
|
|
|
|
detect-project >&2
|
|
|
|
# First, remove anything we did with test-setup (currently, the firewall).
|
|
# NOTE: Keep in sync with names above in test-setup.
|
|
"${GCLOUD}" compute firewall-rules delete "${CLUSTER_NAME}-http-alt" \
|
|
--project="${PROJECT}" || true
|
|
"${GCLOUD}" compute firewall-rules delete "${CLUSTER_NAME}-nodeports" \
|
|
--project="${PROJECT}" || true
|
|
|
|
# Then actually turn down the cluster.
|
|
"${KUBE_ROOT}/cluster/kube-down.sh"
|
|
}
|
|
|
|
# Actually take down the cluster. This is called from test-teardown.
|
|
#
|
|
# Assumed vars:
|
|
# GCLOUD
|
|
# ZONE
|
|
# CLUSTER_NAME
|
|
function kube-down() {
|
|
echo "... in gke:kube-down()" >&2
|
|
detect-project >&2
|
|
"${GCLOUD}" ${CMD_GROUP:-} container clusters delete --project="${PROJECT}" \
|
|
--zone="${ZONE}" "${CLUSTER_NAME}" --quiet
|
|
}
|