SSH_AUTHORIZED_KEY=$(cat ${HOME}/.ssh/id_rsa.pub)
# Use ! instead of usual / to avoid escaping / in
# SSH_AUTHORIZED_KEY
- sed -e 's!sshAuthorizedKey: .*!sshAuthorizedKey: '"${SSH_AUTHORIZED_KEY}"'!' ${SCRIPTDIR}/clusters-values.yaml >${BUILDDIR}/clusters-values.yaml
+ sed -e 's!sshAuthorizedKey: .*!sshAuthorizedKey: '"${SSH_AUTHORIZED_KEY}"'!' ${SCRIPTDIR}/cluster-e2etest-values.yaml >${BUILDDIR}/cluster-e2etest-values.yaml
+}
+
+function release_name {
+ local -r values_path=$1
+ name=$(basename ${values_path})
+ echo ${name%-values.yaml}
}
function deploy {
- helm -n metal3 install machines ${SCRIPTDIR}/../../machines --create-namespace -f ${BUILDDIR}/machines-values.yaml
- helm -n metal3 install clusters ${SCRIPTDIR}/../../clusters --create-namespace -f ${BUILDDIR}/clusters-values.yaml
+ for values in build/site/vm/machine-*-values.yaml; do
+ helm -n metal3 install $(release_name ${values}) ${SCRIPTDIR}/../../machine --create-namespace -f ${values}
+ done
+ helm -n metal3 install cluster-e2etest ${SCRIPTDIR}/../../cluster --create-namespace -f ${BUILDDIR}/cluster-e2etest-values.yaml
}
function clean {
- helm -n metal3 uninstall clusters
- helm -n metal3 uninstall machines
+ helm -n metal3 uninstall cluster-e2etest
+ for values in build/site/vm/machine-*-values.yaml; do
+ helm -n metal3 uninstall $(release_name ${values})
+ done
}
function is_cluster_ready {
[[ $(kubectl -n metal3 get cluster e2etest -o jsonpath='{.status.conditions[?(@.type=="Ready")].status}') == "True" ]]
}
-function are_kustomizations_ready {
- [[ $(kubectl --kubeconfig=${BUILDDIR}/e2etest-admin.conf get Kustomization -n flux-system -o jsonpath='{range .items[*]}{.status.conditions[?(@.type=="Ready")].status}{"\n"}{end}' | grep -c -v True) == 0 ]]
-}
-
-function are_helmreleases_ready {
- [[ $(kubectl --kubeconfig=${BUILDDIR}/e2etest-admin.conf get HelmRelease -n flux-system -o jsonpath='{range .items[*]}{.status.conditions[?(@.type=="Ready")].status}{"\n"}{end}' | grep -c -v True) == 0 ]]
+function is_control_plane_ready {
+ [[ $(kubectl --kubeconfig=${BUILDDIR}/e2etest-admin.conf get nodes -l node-role.kubernetes.io/control-plane -o jsonpath='{range .items[*]}{.status.conditions[?(@.type=="Ready")].status}{"\n"}{end}' | grep -c -v True) == 0 ]]
}
function wait_for_all_ready {
wait_for is_cluster_ready
clusterctl -n metal3 get kubeconfig e2etest >${BUILDDIR}/e2etest-admin.conf
chmod 600 ${BUILDDIR}/e2etest-admin.conf
- # TODO The following checks are not ideal: resources created by
- # operators aren't detected here, but this is the best that can be
- # currently done
- WAIT_FOR_INTERVAL=30s
- wait_for are_kustomizations_ready
- wait_for are_helmreleases_ready
+ wait_for is_control_plane_ready
}
case $1 in