OpenStack known issues¶
This section lists the OpenStack known issues with workarounds for the Mirantis OpenStack for Kubernetes release 22.2.
[19065] Octavia load balancers lose Amphora VMs after failover
[6912] Octavia load balancers may not work properly with DVR
[22930] Octavia load balancers provisioning gets stuck¶
Octavia load balancers provisioning_status
may get stuck in the
ERROR
, PENDING_UPDATE
, PENDING_CREATE
, or PENDING_DELETE
state.
Occasionally, the listeners or pools associated with these load balancers may
also get stuck in the same state.
Workaround:
For administrative users that have access to the keystone-client pod:
Log in to a
keystone-client
pod.Delete the affected load balancer:
openstack loadbalancer delete <load_balancer_id> --force
For non-administrative users, access the Octavia API directly and delete the affected load balancer using the
"force": true
argument in the delete request:Access the Octavia API.
Obtain the token:
TOKEN=$(openstack token issue -f value -c id)
Obtain the endpoint:
ENDPOINT=$(openstack version show --service load-balancer --interface public --status CURRENT -f value -c Endpoint)
Delete the affected load balancers:
curl -H "X-Auth-Token: $TOKEN" -d '{"force": true}' -X DELETE $ENDPOINT/loadbalancers/<load_balancer_id>
[19065] Octavia load balancers lose Amphora VMs after failover¶
If an Amphora VM does not respond or responds too long to heartbeat requests, the Octavia load balancer automatically initiates a failover process after 60 seconds of unsuccessful attempts. Long responses of an Amphora VM may be caused by various events, such as a high load on the OpenStack compute node that hosts the Amphora VM, network issues, system service updates, and so on. After a failover, the Amphora VMs may be missing in the completed Octavia load balancer.
Workaround:
If your deployment is already affected, manually restore the work of the load balancer by recreating the Amphora VM:
Define the load balancer ID:
openstack loadbalancer amphora list --column loadbalancer_id --format value --status ERROR
Start the load balancer failover:
openstack loadbalancer failover <Load balancer ID>
To avoid an automatic failover start that may cause the issue, set the
heartbeat_timeout
parameter in theOpenStackDeployment
CR to a large value in seconds. The default is 60 seconds. For example:spec: services: load-balancer: octavia: values: conf: octavia: health_manager: heartbeat_timeout: 31536000
[6912] Octavia load balancers may not work properly with DVR¶
When Neutron is deployed in the DVR mode, Octavia load balancers may not work correctly. The symptoms include both failure to properly balance traffic and failure to perform an amphora failover.