
Right now, our doc generation scripts notice if you make changes and don't regen docs, don't include new docs, etc. But they miss it if your changes should have REMOVED a doc. Both kubectl-apiversion and kubectl-clusterinfo should have been removed, but weren't. This patch starts tracking all generated doc files and will cause problems if files should be removed and aren't.
149 lines
3.6 KiB
Bash
149 lines
3.6 KiB
Bash
#!/bin/bash
|
|
|
|
# Copyright 2014 Google Inc. All rights reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
kube::util::sortable_date() {
|
|
date "+%Y%m%d-%H%M%S"
|
|
}
|
|
|
|
kube::util::wait_for_url() {
|
|
local url=$1
|
|
local prefix=${2:-}
|
|
local wait=${3:-0.2}
|
|
local times=${4:-10}
|
|
|
|
which curl >/dev/null || {
|
|
kube::log::usage "curl must be installed"
|
|
exit 1
|
|
}
|
|
|
|
local i
|
|
|
|
for i in $(seq 1 $times); do
|
|
local out
|
|
if out=$(curl -fs $url 2>/dev/null); then
|
|
kube::log::status ${prefix}${out}
|
|
return 0
|
|
fi
|
|
sleep $wait
|
|
done
|
|
kube::log::error "Timed out waiting for ${url}"
|
|
return 1
|
|
}
|
|
|
|
# Create a temp dir that'll be deleted at the end of this bash session.
|
|
#
|
|
# Vars set:
|
|
# KUBE_TEMP
|
|
kube::util::ensure-temp-dir() {
|
|
if [[ -z ${KUBE_TEMP-} ]]; then
|
|
KUBE_TEMP=$(mktemp -d -t kubernetes.XXXXXX)
|
|
fi
|
|
}
|
|
|
|
# This figures out the host platform without relying on golang. We need this as
|
|
# we don't want a golang install to be a prerequisite to building yet we need
|
|
# this info to figure out where the final binaries are placed.
|
|
kube::util::host_platform() {
|
|
local host_os
|
|
local host_arch
|
|
case "$(uname -s)" in
|
|
Darwin)
|
|
host_os=darwin
|
|
;;
|
|
Linux)
|
|
host_os=linux
|
|
;;
|
|
*)
|
|
kube::log::error "Unsupported host OS. Must be Linux or Mac OS X."
|
|
exit 1
|
|
;;
|
|
esac
|
|
|
|
case "$(uname -m)" in
|
|
x86_64*)
|
|
host_arch=amd64
|
|
;;
|
|
i?86_64*)
|
|
host_arch=amd64
|
|
;;
|
|
amd64*)
|
|
host_arch=amd64
|
|
;;
|
|
arm*)
|
|
host_arch=arm
|
|
;;
|
|
i?86*)
|
|
host_arch=x86
|
|
;;
|
|
*)
|
|
kube::log::error "Unsupported host arch. Must be x86_64, 386 or arm."
|
|
exit 1
|
|
;;
|
|
esac
|
|
echo "${host_os}/${host_arch}"
|
|
}
|
|
|
|
kube::util::find-binary() {
|
|
local lookfor="${1}"
|
|
local host_platform="$(kube::util::host_platform)"
|
|
local locations=(
|
|
"${KUBE_ROOT}/_output/dockerized/bin/${host_platform}/${lookfor}"
|
|
"${KUBE_ROOT}/_output/local/bin/${host_platform}/${lookfor}"
|
|
"${KUBE_ROOT}/platforms/${host_platform}/${lookfor}"
|
|
)
|
|
local bin=$( (ls -t "${locations[@]}" 2>/dev/null || true) | head -1 )
|
|
echo -n "${bin}"
|
|
}
|
|
|
|
# Wait for background jobs to finish. Return with
|
|
# an error status if any of the jobs failed.
|
|
kube::util::wait-for-jobs() {
|
|
local fail=0
|
|
local job
|
|
for job in $(jobs -p); do
|
|
wait "${job}" || fail=$((fail + 1))
|
|
done
|
|
return ${fail}
|
|
}
|
|
|
|
# takes a binary to run $1 and then copies the results to $2
|
|
kube::util::gen-doc() {
|
|
local cmd="$1"
|
|
local dest="$2"
|
|
|
|
# remove all old generated file from the destination
|
|
for file in $(cat "${dest}/.files_generated" 2>/dev/null); do
|
|
set +e
|
|
rm "${dest}/${file}"
|
|
set -e
|
|
done
|
|
|
|
# We do this in a tmpdir in case the dest has other non-autogenned files
|
|
# We don't want to include them in the list of gen'd files
|
|
local tmpdir="${KUBE_ROOT}/doc_tmp"
|
|
mkdir "${tmpdir}"
|
|
# generate the new files
|
|
${cmd} "${tmpdir}"
|
|
# create the list of generated files
|
|
ls "${tmpdir}" | sort > "${tmpdir}/.files_generated"
|
|
# put the new generated file into the destination
|
|
find "${tmpdir}" -exec rsync -pt {} "${dest}" \; >/dev/null
|
|
#cleanup
|
|
rm -rf "${tmpdir}"
|
|
}
|
|
|
|
# ex: ts=2 sw=2 et filetype=sh
|