Deploy an Equinix Metal based regional cluster with private networking

Before you deploy an additional regional Equinix Metal based cluster with private networking, complete the prerequisite steps described in Prerequisites.

To deploy an Equinix Metal based regional cluster with private networking:

  1. Log in to the bootstrap node running Ubuntu 20.04 that is configured as described in Prerequisites. Properly connect this node to the regional cluster VLAN.

  2. Prepare the bootstrap script:

    1. Download and run the Container Cloud bootstrap script:

      apt install wget
      chmod 0755
    2. Change the directory to the kaas-bootstrap folder created by the script.

  3. Obtain your license file that will be required during the bootstrap:

    1. Create a user account at

    2. Log in to your account and download the mirantis.lic license file.

    3. Save the license file as mirantis.lic under the kaas-bootstrap directory on the bootstrap node.

    4. Verify that mirantis.lic contains the exact Container Cloud license previously downloaded from by decoding the license JWT token, for example, using

      Example of a valid decoded Container Cloud license data with the mandatory license field:

          "exp": 1652304773,
          "iat": 1636669973,
          "sub": "demo",
          "license": {
              "dev": false,
              "limits": {
                  "clusters": 10,
                  "workers_per_cluster": 10
              "openstack": null


      The MKE license does not apply to mirantis.lic. For details about MKE license, see MKE documentation.

  4. Using the Equinix Metal console, obtain the project ID and the user-level API Key of the Equinix Metal project to be used for the Container Cloud deployment:

    1. Log in to the Equinix Metal console.

    2. Select the project that you want to use for the Container Cloud deployment.

    3. In Project Settings > General, capture your Project ID.

    4. In Profile Settings > Personal API Keys, capture the existing user-level API Key or create a new one:

      1. In Profile Settings > Personal API Keys, click Add New Key.

      2. Fill in the Description and select the Read/Write permissions.

      3. Click Add Key.

  5. Prepare the Equinix Metal configuration:

    1. Change the directory to kaas-bootstrap.

    2. In templates/equinixmetalv2/equinix-config.yaml.template, modify spec:projectID and spec:apiToken:value using the values obtained in the previous steps. For example:

        projectID: g98sd6f8-dc7s-8273-v8s7-d9v7395nd91
          value: Bi3m9c7qjYBD3UgsnSCSsqs2bYkbK
    3. In templates/equinixmetalv2/cluster.yaml.template:

      1. Modify the default configuration of the Equinix Metal facility depending on the previously prepared capacity settings as described in Prerequisites:

            # ...
            facility: am6
      2. Add projectSSHKeys that is the list of the Equinix Metal project SSH key names to be attached to cluster machines. These keys are required for access to the Equinix Metal out-of-band console Serial Over SSH (SOS) to debug provisioning failures. We recommend adding at least one project SSH key per cluster.

        Example of the project SSH keys configuration:

            # ...
            - <projectSSHKeyName>

        To create an SSH key in an Equinix Metal project:

        1. Log in to the Equinix Metal console.

        2. Select the project that you want to use for the Container Cloud deployment.

        3. In the Project Settings tab, select Project SSH Keys and click Add New Key.

        4. Enter the Key Name and Public Key values and click Add.

      3. Modify network parameters as required by your infrastructure:

            # ...
              vlanId: SET_EQUINIX_VLAN_ID
              loadBalancerHost: SET_LB_HOST
              cidr: SET_EQUINIX_NETWORK_CIDR
              gateway: SET_EQUINIX_NETWORK_GATEWAY




        ID of the VLAN created in the corresponding Equinix Metal Metro that the seed node and cluster nodes should be attached to.


        IP address to use for the MKE and Kubernetes API endpoints of the cluster.


        List of IP ranges in the format to use for Kubernetes LoadBalancer services. For example, on a management cluster, these services include the Container Cloud web UI and Keycloak. This list should include at least 12 addresses for a management or regional cluster and 5 addresses for a managed cluster.


        Network address in CIDR notation. For example,


        IP address of a gateway attached to this VLAN that provides the necessary external connectivity.


        List of IP ranges in the format. IP addresses from these ranges will be allocated to nodes that boot from DHCP during the provisioning process. Should include at least one address for each machine in the cluster.


        List of IP ranges in the format. IP addresses from these ranges will be allocated as permanent addresses of machines in this cluster. Should include at least one address for each machine in the cluster.


        Optional. List of IP ranges in the format. IP addresses from these ranges will not be allocated as permanent addresses of machines in this cluster.


        List of IP addresses of DNS servers that should be configured on machines. These servers must be accessible through the gateway from the provided VLAN. Required unless a proxy server is used.

    4. Add the following parameters to the bootstrap.env file:




      Name of the bridge that will be used to provide PXE services to provision machines during bootstrap.


      IP address that will be used for PXE services. Will be assigned to the KAAS_BM_PXE_BRIDGE bridge. Must be part of the cidr parameter.


      Number of bits in the network address KAAS_BM_PXE_IP. Must match the CIDR suffix in the cidr parameter.


      IP range in the format that will be used for Kubernetes LoadBalancer services in the bootstrap cluster.

      Example of the PXE parameters in bootstrap.env:

    5. Optional. In templates/equinixmetalv2/machines.yaml.template, modify the default configuration of the Equinix Metal machine type. The minimal required type is c3.small.x86.


      Mirantis highly recommends using the c3.small.x86 machine type for the control plane machines deployed with private network to prevent hardware issues with incorrect BIOS boot order.

          # ...
          machineType: c3.small.x86

      Also, modify other parameters as required.

  6. Optional if servers from the Ubuntu NTP pool (* are accessible from the VLAN where the regional cluster is being provisioned. Otherwise, this step is mandatory.

    Configure the regional NTP server parameters to be applied to all machines of regional and managed clusters in the specified region.

    In templates/equinixmetalv2/cluster.yaml.template, add the ntp:servers section with the list of required server names:

              - helmReleases:
                - name: equinix-provider
                provider: equinixmetalv2
  7. If you require all Internet access to go through a proxy server, in bootstrap.env, add the following environment variables to bootstrap the regional cluster using proxy:



    • NO_PROXY


    Example snippet:

    export HTTP_PROXY=
    export HTTPS_PROXY=
    export NO_PROXY=,registry.internal.lan
    export PROXY_CA_CERTIFICATE_PATH="/home/ubuntu/.mitmproxy/mitmproxy-ca-cert.cer"

    The following formats of variables are accepted:

    Proxy configuration data



    • - for anonymous access.

    • - for restricted access.


    Comma-separated list of IP addresses or domain names.


    Optional. Path to the proxy CA certificate for man-in-the-middle (MITM) proxies. Must be placed on the bootstrap node to be trusted. For details, see Install a CA certificate for a MITM proxy on a bootstrap node.


    If you require Internet access to go through a MITM proxy, ensure that the proxy has streaming enabled as described in Enable streaming for MITM.


    • This parameter is generally available for the OpenStack, bare metal, Equinix Metal with private networking, AWS, and vSphere providers.

    • For MOSK-based deployments, the parameter is generally available since MOSK 22.4.

    • For Azure and Equinix Metal with public networking, the feature is not supported.

    For implementation details, see Proxy and cache support.

    For the list of Mirantis resources and IP addresses to be accessible from the Container Cloud clusters, see Requirements for an Equinix Metal based cluster.

  8. Export the following parameters:

    export KUBECONFIG=<pathToMgmtClusterKubeconfig>
    export REGIONAL_CLUSTER_NAME=<newRegionalClusterName>
    export REGION=<NewRegionName>
    export SSH_KEY_NAME=<newRegionalClusterSshKeyName>

    Substitute the parameters enclosed in angle brackets with the corresponding values of your cluster.


    The REGION and REGIONAL_CLUSTER_NAME parameters values must contain only lowercase alphanumeric characters, hyphens, or periods.

  9. Re-verify that the selected Equinix Metal facility for the regional cluster bootstrap is still available and has enough capacity:



    Depending on your metal CLI version, naming of flags may vary. To verify naming of flags available for your metal CLI version, run metal capacity check --help.

    In the system response, if the value in the AVAILABILITY section has changed from true to false, find an available facility and update the previously configured facility field in cluster.yaml.template.

    For details about the verification procedure, see Verify the capacity of the Equinix Metal facility.

  10. Run the regional cluster bootstrap script:

    ./ deploy_regional


    When the bootstrap is complete, obtain and save in a secure location the kubeconfig-<regionalClusterName> file located in the same directory as the bootstrap script. This file contains the admin credentials for the regional cluster.

    The workflow of the regional cluster bootstrap script




    Prepare the bootstrap cluster for the new regional cluster.


    Load the updated Container Cloud CRDs for Credentials, Cluster, and Machines with information about the new regional cluster to the management cluster.


    Connect to each machine of the management cluster through SSH.


    Wait for the Machines and Cluster objects of the new regional cluster to be ready on the management cluster.


    Load the following objects to the new regional cluster: Secret with the management cluster kubeconfig and ClusterRole for the Container Cloud provider.


    Forward the bootstrap cluster endpoint to helm-controller.


    Wait for all CRDs to be available and verify the objects created using these CRDs.


    Pivot the cluster API stack to the regional cluster.


    Switch the LCM Agent from the bootstrap cluster to the regional one.


    Wait for the Container Cloud components to start on the regional cluster.

  11. Establish connection to the cluster private network:

    1. Install sshuttle.

    2. Obtain the cluster CIDR from the cluster specification:

      kubectl --kubeconfig <clusterKubeconfig> \
      get cluster <clusterName> -n <clusterProjectName> \
      -o jsonpath='{}'
    3. Obtain the public IP address of the related Equinix Metal router:

      1. Log in to the Equinix Metal console of the related project.

      2. In the list of servers, capture the IP address of the related Equinix Metal router server listed in the IPV4 ADDRESS column.

    4. Establish connection to the cluster private network from your local machine:

      sshuttle <clusterCIDR> -r ubuntu@<routerPublicIP> --ssh-cmd 'ssh -i <pathToRouterSSHKey>'

    Now, you can access the Keycloak, StackLight, and Container Cloud web UIs.

Now, you can proceed with deploying the managed clusters of supported provider types as described in Create and operate an Equinix Metal based managed cluster with private networking.


To decrease network traffic cost and not to complicate the network infrastructure, you must deploy managed clusters in the same region as the regional cluster to have both clusters deployed in the same metro.

For example, if you have a management cluster with region-one in Frankfurt and a regional cluster with region-two in Silicon Valley, create all Frankfurt-based managed clusters in region-one and all Silicon Valley based managed clusters in region-two.