Install Multi-Primary
Follow this guide to install the Istio control plane on both cluster1 and cluster2, making each a primary cluster. Both clusters reside on the network1 network, meaning there is direct connectivity between the pods in both clusters.
Before proceeding, be sure to complete the steps under before you begin.
In this configuration, each control plane observes the API Servers in both clusters for endpoints.
Service workloads communicate directly (pod-to-pod) across cluster boundaries.
Configure cluster1 as a primary
Create the istioctl configuration for cluster1:
Install Istio as primary in cluster1 using istioctl and the IstioOperator API.
$ cat <<EOF > cluster1.yaml apiVersion: install.istio.io/v1alpha1 kind: IstioOperator spec: values: global: meshID: mesh1 multiCluster: clusterName: cluster1 network: network1 EOFApply the configuration to cluster1:
$ istioctl install --context="${CTX_CLUSTER1}" -f cluster1.yamlInstall Istio as primary in cluster1 using the following Helm commands:
Install the base chart in cluster1:
$ helm install istio-base istio/base -n istio-system --kube-context "${CTX_CLUSTER1}"Then, install the istiod chart in cluster1 with the following multi-cluster settings:
$ helm install istiod istio/istiod -n istio-system --kube-context "${CTX_CLUSTER1}" --set global.meshID=mesh1 --set global.multiCluster.clusterName=cluster1 --set global.network=network1Configure cluster2 as a primary
Create the istioctl configuration for cluster2:
Install Istio as primary in cluster2 using istioctl and the IstioOperator API.
$ cat <<EOF > cluster2.yaml apiVersion: install.istio.io/v1alpha1 kind: IstioOperator spec: values: global: meshID: mesh1 multiCluster: clusterName: cluster2 network: network1 EOFApply the configuration to cluster2:
$ istioctl install --context="${CTX_CLUSTER2}" -f cluster2.yamlInstall Istio as primary in cluster2 using the following Helm commands:
Install the base chart in cluster2:
$ helm install istio-base istio/base -n istio-system --kube-context "${CTX_CLUSTER2}"Then, install the istiod chart in cluster2 with the following multi-cluster settings:
$ helm install istiod istio/istiod -n istio-system --kube-context "${CTX_CLUSTER2}" --set global.meshID=mesh1 --set global.multiCluster.clusterName=cluster2 --set global.network=network1Enable Endpoint Discovery
Install a remote secret in cluster2 that provides access to cluster1’s API server.
$ istioctl create-remote-secret \ --context="${CTX_CLUSTER1}" \ --name=cluster1 | \ kubectl apply -f - --context="${CTX_CLUSTER2}"Install a remote secret in cluster1 that provides access to cluster2’s API server.
$ istioctl create-remote-secret \ --context="${CTX_CLUSTER2}" \ --name=cluster2 | \ kubectl apply -f - --context="${CTX_CLUSTER1}"Congratulations! You successfully installed an Istio mesh across multiple primary clusters!
Next Steps
You can now verify the installation.
Cleanup
Uninstall Istio from both cluster1 and cluster2 using the same mechanism you installed Istio with (istioctl or Helm).
Uninstall Istio in cluster1:
$ istioctl uninstall --context="${CTX_CLUSTER1}" -y --purge $ kubectl delete ns istio-system --context="${CTX_CLUSTER1}"Uninstall Istio in cluster2:
$ istioctl uninstall --context="${CTX_CLUSTER2}" -y --purge $ kubectl delete ns istio-system --context="${CTX_CLUSTER2}"Delete Istio Helm installation from cluster1:
$ helm delete istiod -n istio-system --kube-context "${CTX_CLUSTER1}" $ helm delete istio-base -n istio-system --kube-context "${CTX_CLUSTER1}"Delete the istio-system namespace from cluster1:
$ kubectl delete ns istio-system --context="${CTX_CLUSTER1}"Delete Istio Helm installation from cluster2:
$ helm delete istiod -n istio-system --kube-context "${CTX_CLUSTER2}" $ helm delete istio-base -n istio-system --kube-context "${CTX_CLUSTER2}"Delete the istio-system namespace from cluster2:
$ kubectl delete ns istio-system --context="${CTX_CLUSTER2}"(Optional) Delete CRDs installed by Istio:
Deleting CRDs permanently removes any Istio resources you have created in your clusters. Delete Istio CRDs installed in your clusters by running:
$ kubectl get crd -oname --context "${CTX_CLUSTER1}" | grep --color=never 'istio.io' | xargs kubectl delete --context "${CTX_CLUSTER1}" $ kubectl get crd -oname --context "${CTX_CLUSTER2}" | grep --color=never 'istio.io' | xargs kubectl delete --context "${CTX_CLUSTER2}"