OpenStack known issues¶
This section lists the OpenStack known issues with workarounds for the Mirantis OpenStack for Kubernetes release 22.5.
[29501] Cinder periodic database cleanup resets the state of volumes
[25594] Security groups shared through RBAC cannot be used to create instances
[29501] Cinder periodic database cleanup resets the state of volumes¶
Due to an issue in the database auto-cleanup job for the Block Storage service
(OpenStack Cinder), the state of volumes that are attached to instances gets
reset every time the job runs. The instances can still write and read block
storage data, however, volume objects appear in the OpenStack API as
not attached
causing confusion.
The workaround is to temporarily disable the job until the issue is fixed and execute the script below to restore the affected instances.
To disable the job, update the OpenStackDeployment
custom resource as
follows:
kind: OpenStackDeployment
spec:
features:
database:
cleanup:
cinder:
enabled: false
To restore the affected instances:
Obtain one of the Nova API pods:
nova_api_pod=$(kubectl -n openstack get pod -l application=nova,component=os-api --no-headers | head -n1 | awk '{print $1}')
Download the
restore_volume_attachments.py
script to your local environment.Note
The provided script does not fix the Cinder database clean-up job and is only intended to restore the functionality of the affected instances. Therefore, leave the job disabled.
Copy the script to the Nova API pod:
kubectl -n openstack cp restore_volume_attachments.py $nova_api_pod:tmp/
Run the script in the dry-run mode to only list affected instances and volumes:
kubectl -n openstack exec -ti $nova_api_pod -- python /tmp/restore_volume_attachments.py --dry-run
Run the script to restore the volume attachments:
kubectl -n openstack exec -ti $nova_api_pod -- python /tmp/restore_volume_attachments.py
[25124] MPLSoGRE encapsulation has limited throughput¶
Multiprotocol Label Switching over Generic Routing Encapsulation (MPLSoGRE) provides limited throughput while sending data between VMs up to 38 Mbps, as per Mirantis tests.
As a workaround, switch the encapsulation type to VXLAN in the
OpenStackDeployment
custom resource:
spec:
services:
networking:
neutron:
values:
conf:
bagpipe_bgp:
dataplane_driver_ipvpn:
mpls_over_gre: "False"
vxlan_encap: "True"