X-Git-Url: https://gerrit.akraino.org/r/gitweb?a=blobdiff_plain;f=deploy%2Fsite%2Fvm%2Fvm.sh;h=477c4934a6cc37e9e735d7bc75e8b276e5ff87c9;hb=f26e5f2ec686975fd8274f47fd9917264a2d92de;hp=841644b154e88345a1ba666c15f18031c33bd81d;hpb=43dbb5a2353606d1801f08dcf3e2fd06399843fb;p=icn.git diff --git a/deploy/site/vm/vm.sh b/deploy/site/vm/vm.sh index 841644b..477c493 100755 --- a/deploy/site/vm/vm.sh +++ b/deploy/site/vm/vm.sh @@ -10,6 +10,10 @@ source $SCRIPTDIR/../common.sh BUILDDIR=${SCRIPTDIR/deploy/build} mkdir -p ${BUILDDIR} +SITE_REPO=${SITE_REPO:-"https://gerrit.akraino.org/r/icn"} +SITE_BRANCH=${SITE_BRANCH:-"master"} +SITE_PATH=${SITE_PATH:-"deploy/site/vm"} + FLUX_SOPS_KEY_NAME=${FLUX_SOPS_KEY_NAME:-"icn-site-vm"} # !!!NOTE!!! THE KEYS USED BELOW ARE FOR TEST PURPOSES ONLY. DO NOT @@ -42,23 +46,23 @@ function build_source { function deploy { gpg --import ${SCRIPTDIR}/sops.asc - flux_create_site https://gerrit.akraino.org/r/icn master deploy/site/vm ${FLUX_SOPS_KEY_NAME} + flux_create_site ${SITE_REPO} ${SITE_BRANCH} ${SITE_PATH} ${FLUX_SOPS_KEY_NAME} } function clean { - kubectl -n flux-system delete kustomization icn-master-site-vm + kubectl -n flux-system delete kustomization $(flux_site_kustomization_name ${SITE_REPO} ${SITE_BRANCH} ${SITE_PATH}) } function is_cluster_ready { - [[ $(kubectl -n metal3 get cluster e2etest -o jsonpath='{.status.conditions[?(@.type=="Ready")].status}') == "True" ]] + [[ $(kubectl -n metal3 get cluster icn -o jsonpath='{.status.conditions[?(@.type=="Ready")].status}') == "True" ]] } function is_control_plane_ready { # Checking the Cluster resource status is not sufficient, it # reports the control plane as ready before the nodes forming the # control plane are ready - local -r replicas=$(kubectl -n metal3 get kubeadmcontrolplane e2etest -o jsonpath='{.spec.replicas}') - [[ $(kubectl --kubeconfig=${BUILDDIR}/e2etest-admin.conf get nodes -l node-role.kubernetes.io/control-plane -o jsonpath='{range .items[*]}{.status.conditions[?(@.type=="Ready")].status}{"\n"}{end}' | grep -c True) == ${replicas} ]] + local -r replicas=$(kubectl -n metal3 get kubeadmcontrolplane icn -o jsonpath='{.spec.replicas}') + [[ $(kubectl --kubeconfig=${BUILDDIR}/icn-admin.conf get nodes -l node-role.kubernetes.io/control-plane -o jsonpath='{range .items[*]}{.status.conditions[?(@.type=="Ready")].status}{"\n"}{end}' | grep -c True) == ${replicas} ]] } function insert_control_plane_network_identity_into_ssh_config { @@ -71,8 +75,8 @@ function insert_control_plane_network_identity_into_ssh_config { chmod 600 ${HOME}/.ssh/config # Add the entry for the control plane network, host value in ssh # config is a wildcard - endpoint=$(helm -n metal3 get values -a cluster-e2etest | awk '/controlPlaneEndpoint:/ {print $2}') - prefix=$(helm -n metal3 get values -a cluster-e2etest | awk '/controlPlanePrefix:/ {print $2}') + endpoint=$(helm -n metal3 get values -a cluster-icn | awk '/controlPlaneEndpoint:/ {print $2}') + prefix=$(helm -n metal3 get values -a cluster-icn | awk '/controlPlanePrefix:/ {print $2}') host=$(ipcalc ${endpoint}/${prefix} | awk '/Network:/ {sub(/\.0.*/,".*"); print $2}') if [[ $(grep -c "Host ${host}" ${HOME}/.ssh/config) != 0 ]]; then sed -i -e '/Host '"${host}"'/,+1 d' ${HOME}/.ssh/config @@ -87,8 +91,8 @@ function wait_for_all_ready { WAIT_FOR_INTERVAL=60s WAIT_FOR_TRIES=30 wait_for is_cluster_ready - clusterctl -n metal3 get kubeconfig e2etest >${BUILDDIR}/e2etest-admin.conf - chmod 600 ${BUILDDIR}/e2etest-admin.conf + clusterctl -n metal3 get kubeconfig icn >${BUILDDIR}/icn-admin.conf + chmod 600 ${BUILDDIR}/icn-admin.conf wait_for is_control_plane_ready insert_control_plane_network_identity_into_ssh_config }