Cluster update known issues

This section lists the cluster update known issues with workarounds for the Mirantis OpenStack for Kubernetes release 23.2.2.

[34342] etcd pods failure

During the update, you may encounter the issue that causes a failure of the etcd pods due to the simultaneous deployment of several pods on a single node.

The workaround is to remove the PVC for one etcd pod.

[35111] openstack-operator-ensure-resources job stuck in CrashLoopBackOff

During MOSK update to either 23.2 major release or any patch release of the 23.2 release series, the openstack-operator-ensure-resources job may get stuck in the CrashLoopBackOff state with the following error:

Traceback (most recent call last):
File "/usr/local/bin/osctl-ensure-shared-resources", line 8, in <module>
File "/usr/local/lib/python3.8/dist-packages/openstack_controller/cli/", line 61, in main
File "/usr/local/lib/python3.8/dist-packages/pykube/", line 165, in update
  self.patch(self.obj, subresource=subresource)
File "/usr/local/lib/python3.8/dist-packages/pykube/", line 157, in patch
File "/usr/local/lib/python3.8/dist-packages/pykube/", line 444, in raise_for_status
  raise HTTPError(resp.status_code, payload["message"])
pykube.exceptions.HTTPError: "" is invalid: spec.preserveUnknownFields: Invalid value: true: must be false in order to use defaults in the schema

As a workaround, delete the CRD from your cluster:

kubectl delete crd

[37012] Masakari failure during update

While updating your cluster, the Instance High Availability service (OpenStack Masakari) may not work as expected.

As a workaround, temporarily disable the service by removing instance-ha from the service list in the OpenStackDeployment custom resource.