summary |
shortlog |
log |
commit | commitdiff |
review |
tree
raw |
patch |
inline | side by side (from parent 1:
818ebd9)
This compute cluster definition includes all possible ICN addons.
Signed-off-by: Todd Malsbary <todd.malsbary@intel.com>
Change-Id: I957e3a6dbdd459a8037f50f0313a65a652023eb1
35 files changed:
flux:
url: https://gerrit.akraino.org/r/icn
branch: master
flux:
url: https://gerrit.akraino.org/r/icn
branch: master
- path: ./deploy/site/cluster-e2etest
+ path: ./deploy/site/cluster-icn
```
A brief overview of the values is below. Refer to the machine and
```
A brief overview of the values is below. Refer to the machine and
mkdir -p ${BUILDDIR}
function register_emco_controllers {
mkdir -p ${BUILDDIR}
function register_emco_controllers {
- local -r cluster_name=${CLUSTER_NAME:-e2etest}
+ local -r cluster_name=${CLUSTER_NAME:-icn}
local -r host=$(kubectl -n metal3 get cluster/${cluster_name} -o jsonpath='{.spec.controlPlaneEndpoint.host}')
cat <<EOF >${BUILDDIR}/${cluster_name}-config.yaml
orchestrator:
local -r host=$(kubectl -n metal3 get cluster/${cluster_name} -o jsonpath='{.spec.controlPlaneEndpoint.host}')
cat <<EOF >${BUILDDIR}/${cluster_name}-config.yaml
orchestrator:
}
function unregister_emco_controllers {
}
function unregister_emco_controllers {
- local -r cluster_name=${CLUSTER_NAME:-e2etest}
+ local -r cluster_name=${CLUSTER_NAME:-icn}
emcoctl --config ${BUILDDIR}/${cluster_name}-config.yaml delete -f ${BUILDDIR}/${cluster_name}-controllers.yaml
}
function test_addons {
# Create a temporary kubeconfig file for the tests
emcoctl --config ${BUILDDIR}/${cluster_name}-config.yaml delete -f ${BUILDDIR}/${cluster_name}-controllers.yaml
}
function test_addons {
# Create a temporary kubeconfig file for the tests
- local -r cluster_name=${CLUSTER_NAME:-e2etest}
+ local -r cluster_name=${CLUSTER_NAME:-icn}
local -r cluster_kubeconfig="${BUILDDIR}/${cluster_name}.conf"
clusterctl -n metal3 get kubeconfig ${cluster_name} >${cluster_kubeconfig}
local -r cluster_kubeconfig="${BUILDDIR}/${cluster_name}.conf"
clusterctl -n metal3 get kubeconfig ${cluster_name} >${cluster_kubeconfig}
Usage: $(basename $0) COMMAND
The "test" command looks for the CLUSTER_NAME variable in the
Usage: $(basename $0) COMMAND
The "test" command looks for the CLUSTER_NAME variable in the
-environment (default: "e2etest"). This should be the name of the
+environment (default: "icn"). This should be the name of the
Cluster resource to execute the tests in.
Commands:
Cluster resource to execute the tests in.
Commands:
branch: master
# path is the repository to the resources to be applied to the
# cluster.
branch: master
# path is the repository to the resources to be applied to the
# cluster.
- path: ./deploy/site/e2etest
+ path: ./deploy/site/cluster-icn
# containerRuntime may be containerd or docker.
containerRuntime: containerd
# containerRuntime may be containerd or docker.
containerRuntime: containerd
}
function is_kata_deployed {
}
function is_kata_deployed {
- local -r cluster_name=${CLUSTER_NAME:-e2etest}
+ local -r cluster_name=${CLUSTER_NAME:-icn}
local -r cluster_kubeconfig="${BUILDDIR}/${cluster_name}.conf"
kubectl --kubeconfig=${cluster_kubeconfig} get runtimeclass/kata-qemu
}
function test_kata {
# Create a temporary kubeconfig file for the tests
local -r cluster_kubeconfig="${BUILDDIR}/${cluster_name}.conf"
kubectl --kubeconfig=${cluster_kubeconfig} get runtimeclass/kata-qemu
}
function test_kata {
# Create a temporary kubeconfig file for the tests
- local -r cluster_name=${CLUSTER_NAME:-e2etest}
+ local -r cluster_name=${CLUSTER_NAME:-icn}
local -r cluster_kubeconfig="${BUILDDIR}/${cluster_name}.conf"
clusterctl -n metal3 get kubeconfig ${cluster_name} >${cluster_kubeconfig}
local -r cluster_kubeconfig="${BUILDDIR}/${cluster_name}.conf"
clusterctl -n metal3 get kubeconfig ${cluster_name} >${cluster_kubeconfig}
Usage: $(basename $0) COMMAND
The "test" command looks for the CLUSTER_NAME variable in the
Usage: $(basename $0) COMMAND
The "test" command looks for the CLUSTER_NAME variable in the
-environment (default: "e2etest"). This should be the name of the
+environment (default: "icn"). This should be the name of the
Cluster resource to execute the tests in.
Commands:
Cluster resource to execute the tests in.
Commands:
apiVersion: kustomize.toolkit.fluxcd.io/v1beta2
kind: Kustomization
metadata:
apiVersion: kustomize.toolkit.fluxcd.io/v1beta2
kind: Kustomization
metadata:
- name: e2etest-flux-sync
namespace: flux-system
spec:
interval: 10m0s
namespace: flux-system
spec:
interval: 10m0s
- path: ./deploy/site/cluster-e2etest
+ path: ./deploy/site/cluster-icn
prune: true
sourceRef:
kind: GitRepository
prune: true
sourceRef:
kind: GitRepository
}
function is_cluster_ready {
}
function is_cluster_ready {
- [[ $(kubectl -n metal3 get cluster e2etest -o jsonpath='{.status.conditions[?(@.type=="Ready")].status}') == "True" ]]
+ [[ $(kubectl -n metal3 get cluster icn -o jsonpath='{.status.conditions[?(@.type=="Ready")].status}') == "True" ]]
}
function is_control_plane_ready {
# Checking the Cluster resource status is not sufficient, it
# reports the control plane as ready before the nodes forming the
# control plane are ready
}
function is_control_plane_ready {
# Checking the Cluster resource status is not sufficient, it
# reports the control plane as ready before the nodes forming the
# control plane are ready
- local -r replicas=$(kubectl -n metal3 get kubeadmcontrolplane e2etest -o jsonpath='{.spec.replicas}')
- [[ $(kubectl --kubeconfig=${BUILDDIR}/e2etest-admin.conf get nodes -l node-role.kubernetes.io/control-plane -o jsonpath='{range .items[*]}{.status.conditions[?(@.type=="Ready")].status}{"\n"}{end}' | grep -c True) == ${replicas} ]]
+ local -r replicas=$(kubectl -n metal3 get kubeadmcontrolplane icn -o jsonpath='{.spec.replicas}')
+ [[ $(kubectl --kubeconfig=${BUILDDIR}/icn-admin.conf get nodes -l node-role.kubernetes.io/control-plane -o jsonpath='{range .items[*]}{.status.conditions[?(@.type=="Ready")].status}{"\n"}{end}' | grep -c True) == ${replicas} ]]
}
function insert_control_plane_network_identity_into_ssh_config {
}
function insert_control_plane_network_identity_into_ssh_config {
chmod 600 ${HOME}/.ssh/config
# Add the entry for the control plane network, host value in ssh
# config is a wildcard
chmod 600 ${HOME}/.ssh/config
# Add the entry for the control plane network, host value in ssh
# config is a wildcard
- endpoint=$(helm -n metal3 get values -a cluster-e2etest | awk '/controlPlaneEndpoint:/ {print $2}')
- prefix=$(helm -n metal3 get values -a cluster-e2etest | awk '/controlPlanePrefix:/ {print $2}')
+ endpoint=$(helm -n metal3 get values -a cluster-icn | awk '/controlPlaneEndpoint:/ {print $2}')
+ prefix=$(helm -n metal3 get values -a cluster-icn | awk '/controlPlanePrefix:/ {print $2}')
host=$(ipcalc ${endpoint}/${prefix} | awk '/Network:/ {sub(/\.0.*/,".*"); print $2}')
if [[ $(grep -c "Host ${host}" ${HOME}/.ssh/config) != 0 ]]; then
sed -i -e '/Host '"${host}"'/,+1 d' ${HOME}/.ssh/config
host=$(ipcalc ${endpoint}/${prefix} | awk '/Network:/ {sub(/\.0.*/,".*"); print $2}')
if [[ $(grep -c "Host ${host}" ${HOME}/.ssh/config) != 0 ]]; then
sed -i -e '/Host '"${host}"'/,+1 d' ${HOME}/.ssh/config
WAIT_FOR_INTERVAL=60s
WAIT_FOR_TRIES=30
wait_for is_cluster_ready
WAIT_FOR_INTERVAL=60s
WAIT_FOR_TRIES=30
wait_for is_cluster_ready
- clusterctl -n metal3 get kubeconfig e2etest >${BUILDDIR}/e2etest-admin.conf
- chmod 600 ${BUILDDIR}/e2etest-admin.conf
+ clusterctl -n metal3 get kubeconfig icn >${BUILDDIR}/icn-admin.conf
+ chmod 600 ${BUILDDIR}/icn-admin.conf
wait_for is_control_plane_ready
insert_control_plane_network_identity_into_ssh_config
}
wait_for is_control_plane_ready
insert_control_plane_network_identity_into_ssh_config
}
apiVersion: helm.toolkit.fluxcd.io/v2beta1
kind: HelmRelease
metadata:
apiVersion: helm.toolkit.fluxcd.io/v2beta1
kind: HelmRelease
metadata:
namespace: metal3
spec:
interval: 5m
namespace: metal3
spec:
interval: 5m
name: icn
interval: 1m
values:
name: icn
interval: 1m
values:
clusterLabels:
site: pod11
controlPlaneEndpoint: 10.10.110.23
clusterLabels:
site: pod11
controlPlaneEndpoint: 10.10.110.23
flux:
url: https://gerrit.akraino.org/r/icn
branch: master
flux:
url: https://gerrit.akraino.org/r/icn
branch: master
- path: ./deploy/site/cluster-e2etest
+ path: ./deploy/site/cluster-icn
sops:
kms: []
gcp_kms: []
sops:
kms: []
gcp_kms: []
apiVersion: helm.toolkit.fluxcd.io/v2beta1
kind: HelmRelease
metadata:
apiVersion: helm.toolkit.fluxcd.io/v2beta1
kind: HelmRelease
metadata:
namespace: metal3
spec:
interval: 5m
namespace: metal3
spec:
interval: 5m
name: icn
interval: 1m
values:
name: icn
interval: 1m
values:
clusterLabels:
site: vm
keepalived:
clusterLabels:
site: vm
keepalived:
flux:
url: https://gerrit.akraino.org/r/icn
branch: master
flux:
url: https://gerrit.akraino.org/r/icn
branch: master
- path: ./deploy/site/cluster-e2etest
+ path: ./deploy/site/cluster-icn
sops:
kms: []
gcp_kms: []
sops:
kms: []
gcp_kms: []
}
function is_cluster_ready {
}
function is_cluster_ready {
- [[ $(kubectl -n metal3 get cluster e2etest -o jsonpath='{.status.conditions[?(@.type=="Ready")].status}') == "True" ]]
+ [[ $(kubectl -n metal3 get cluster icn -o jsonpath='{.status.conditions[?(@.type=="Ready")].status}') == "True" ]]
}
function is_control_plane_ready {
# Checking the Cluster resource status is not sufficient, it
# reports the control plane as ready before the nodes forming the
# control plane are ready
}
function is_control_plane_ready {
# Checking the Cluster resource status is not sufficient, it
# reports the control plane as ready before the nodes forming the
# control plane are ready
- local -r replicas=$(kubectl -n metal3 get kubeadmcontrolplane e2etest -o jsonpath='{.spec.replicas}')
- [[ $(kubectl --kubeconfig=${BUILDDIR}/e2etest-admin.conf get nodes -l node-role.kubernetes.io/control-plane -o jsonpath='{range .items[*]}{.status.conditions[?(@.type=="Ready")].status}{"\n"}{end}' | grep -c True) == ${replicas} ]]
+ local -r replicas=$(kubectl -n metal3 get kubeadmcontrolplane icn -o jsonpath='{.spec.replicas}')
+ [[ $(kubectl --kubeconfig=${BUILDDIR}/icn-admin.conf get nodes -l node-role.kubernetes.io/control-plane -o jsonpath='{range .items[*]}{.status.conditions[?(@.type=="Ready")].status}{"\n"}{end}' | grep -c True) == ${replicas} ]]
}
function insert_control_plane_network_identity_into_ssh_config {
}
function insert_control_plane_network_identity_into_ssh_config {
chmod 600 ${HOME}/.ssh/config
# Add the entry for the control plane network, host value in ssh
# config is a wildcard
chmod 600 ${HOME}/.ssh/config
# Add the entry for the control plane network, host value in ssh
# config is a wildcard
- endpoint=$(helm -n metal3 get values -a cluster-e2etest | awk '/controlPlaneEndpoint:/ {print $2}')
- prefix=$(helm -n metal3 get values -a cluster-e2etest | awk '/controlPlanePrefix:/ {print $2}')
+ endpoint=$(helm -n metal3 get values -a cluster-icn | awk '/controlPlaneEndpoint:/ {print $2}')
+ prefix=$(helm -n metal3 get values -a cluster-icn | awk '/controlPlanePrefix:/ {print $2}')
host=$(ipcalc ${endpoint}/${prefix} | awk '/Network:/ {sub(/\.0.*/,".*"); print $2}')
if [[ $(grep -c "Host ${host}" ${HOME}/.ssh/config) != 0 ]]; then
sed -i -e '/Host '"${host}"'/,+1 d' ${HOME}/.ssh/config
host=$(ipcalc ${endpoint}/${prefix} | awk '/Network:/ {sub(/\.0.*/,".*"); print $2}')
if [[ $(grep -c "Host ${host}" ${HOME}/.ssh/config) != 0 ]]; then
sed -i -e '/Host '"${host}"'/,+1 d' ${HOME}/.ssh/config
WAIT_FOR_INTERVAL=60s
WAIT_FOR_TRIES=30
wait_for is_cluster_ready
WAIT_FOR_INTERVAL=60s
WAIT_FOR_TRIES=30
wait_for is_cluster_ready
- clusterctl -n metal3 get kubeconfig e2etest >${BUILDDIR}/e2etest-admin.conf
- chmod 600 ${BUILDDIR}/e2etest-admin.conf
+ clusterctl -n metal3 get kubeconfig icn >${BUILDDIR}/icn-admin.conf
+ chmod 600 ${BUILDDIR}/icn-admin.conf
wait_for is_control_plane_ready
insert_control_plane_network_identity_into_ssh_config
}
wait_for is_control_plane_ready
insert_control_plane_network_identity_into_ssh_config
}
#flux:
# url: https://gerrit.akraino.org/r/icn
# branch: master
#flux:
# url: https://gerrit.akraino.org/r/icn
# branch: master
-# path: ./deploy/site/cluster-e2etest
+# path: ./deploy/site/cluster-icn