X-Git-Url: https://gerrit.akraino.org/r/gitweb?a=blobdiff_plain;f=deploy%2Fsite%2Fvm%2Fvm.sh;h=fbe0890f02c45e4174f5f741632005d1e79a943f;hb=41776ab3743c491e4ff4c31e7a2ea48abe6451a2;hp=3f3b3fbb2677122d53d7dbd2f3865e35457e3d3d;hpb=790dc20efb05a3113b97800926a80753e9a89e98;p=icn.git diff --git a/deploy/site/vm/vm.sh b/deploy/site/vm/vm.sh index 3f3b3fb..fbe0890 100755 --- a/deploy/site/vm/vm.sh +++ b/deploy/site/vm/vm.sh @@ -12,38 +12,7 @@ mkdir -p ${BUILDDIR} SITE_REPO=${SITE_REPO:-"https://gerrit.akraino.org/r/icn"} SITE_BRANCH=${SITE_BRANCH:-"master"} -SITE_PATH=${SITE_PATH:-"deploy/site/vm"} - -FLUX_SOPS_KEY_NAME=${FLUX_SOPS_KEY_NAME:-"icn-site-vm"} -FLUX_SOPS_PRIVATE_KEY="${SCRIPTDIR}/../secrets/sops.asc" - -# !!!NOTE!!! THE KEYS USED BELOW ARE FOR TEST PURPOSES ONLY. DO NOT -# USE THESE OUTSIDE OF THIS ICN VIRTUAL TEST ENVIRONMENT. -function build_source { - # First decrypt the existing site YAML, otherwise we'll be - # attempting to encrypt it twice below - if [[ -f ${FLUX_SOPS_PRIVATE_KEY} ]]; then - gpg --import ${FLUX_SOPS_PRIVATE_KEY} - sops_decrypt ${SCRIPTDIR}/site.yaml - fi - - # Generate user password and authorized key in site YAML - # To login to guest, ssh -i ${SCRIPTDIR}/id_rsa - HASHED_PASSWORD=$(mkpasswd --method=SHA-512 --rounds 10000 "mypasswd") - sed -i -e 's!hashedPassword: .*!hashedPassword: '"${HASHED_PASSWORD}"'!' ${SCRIPTDIR}/site.yaml - ssh-keygen -t rsa -N "" -f ${SCRIPTDIR}/id_rsa <<${FLUX_SOPS_PRIVATE_KEY} -} +SITE_PATH=${SITE_PATH:-"deploy/site/vm/deployment"} function deploy { gpg --import ${FLUX_SOPS_PRIVATE_KEY} @@ -55,14 +24,14 @@ function clean { } function is_cluster_ready { - [[ $(kubectl -n metal3 get cluster icn -o jsonpath='{.status.conditions[?(@.type=="Ready")].status}') == "True" ]] + [[ $(kubectl -n ${SITE_NAMESPACE} get cluster icn -o jsonpath='{.status.conditions[?(@.type=="Ready")].status}') == "True" ]] } function is_control_plane_ready { # Checking the Cluster resource status is not sufficient, it # reports the control plane as ready before the nodes forming the # control plane are ready - local -r replicas=$(kubectl -n metal3 get kubeadmcontrolplane icn -o jsonpath='{.spec.replicas}') + local -r replicas=$(kubectl -n ${SITE_NAMESPACE} get kubeadmcontrolplane icn -o jsonpath='{.spec.replicas}') [[ $(kubectl --kubeconfig=${BUILDDIR}/icn-admin.conf get nodes -l node-role.kubernetes.io/control-plane -o jsonpath='{range .items[*]}{.status.conditions[?(@.type=="Ready")].status}{"\n"}{end}' | grep -c True) == ${replicas} ]] } @@ -80,11 +49,11 @@ function insert_control_plane_network_identity_into_ssh_config { chmod 600 ${HOME}/.ssh/config # Add the entry for the control plane network, host value in ssh # config is a wildcard - endpoint=$(helm -n metal3 get values -a cluster-icn | awk '/controlPlaneEndpoint:/ {print $2}') - prefix=$(helm -n metal3 get values -a cluster-icn | awk '/controlPlanePrefix:/ {print $2}') + endpoint=$(helm -n ${SITE_NAMESPACE} get values -a cluster-icn | awk '/controlPlaneEndpoint:/ {print $2}') + prefix=$(helm -n ${SITE_NAMESPACE} get values -a cluster-icn | awk '/controlPlanePrefix:/ {print $2}') host=$(ipcalc ${endpoint}/${prefix} | awk '/Network:/ {sub(/\.0.*/,".*"); print $2}') if [[ $(grep -c "Host ${host}" ${HOME}/.ssh/config) != 0 ]]; then - sed -i -e '/Host '"${host}"'/,+1 d' ${HOME}/.ssh/config + sed -i -e '/Host '"${host}"'/,+3 d' ${HOME}/.ssh/config fi cat <>${HOME}/.ssh/config Host ${host} @@ -94,6 +63,8 @@ Host ${host} EOF # Add the identity to authorized keys on this host to enable ssh # logins via its control plane address + authorized_key=$(cat ${SCRIPTDIR}/id_rsa.pub) + sed -i -e '\!'"${authorized_key}"'!d' ${HOME}/.ssh/authorized_keys cat ${SCRIPTDIR}/id_rsa.pub >> ~/.ssh/authorized_keys } @@ -101,22 +72,31 @@ function wait_for_all_ready { WAIT_FOR_INTERVAL=60s WAIT_FOR_TRIES=30 wait_for is_cluster_ready - clusterctl -n metal3 get kubeconfig icn >${BUILDDIR}/icn-admin.conf + clusterctl -n ${SITE_NAMESPACE} get kubeconfig icn >${BUILDDIR}/icn-admin.conf chmod 600 ${BUILDDIR}/icn-admin.conf wait_for is_control_plane_ready insert_control_plane_network_identity_into_ssh_config } +function is_cluster_deleted { + ! kubectl -n ${SITE_NAMESPACE} get cluster icn +} + +function wait_for_all_deleted { + WAIT_FOR_INTERVAL=60s + WAIT_FOR_TRIES=30 + wait_for is_cluster_deleted +} + case $1 in - "build-source") build_source ;; "clean") clean ;; "deploy") deploy ;; "wait") wait_for_all_ready ;; + "wait-clean") wait_for_all_deleted ;; *) cat <