[Bug 2081742] Re: cinder powermax driver hangs when creating multiple image volumes

Andre Ruiz 2081742 at bugs.launchpad.net
Mon Dec 9 19:53:59 UTC 2024


Yes, please send the invite.

-- 
You received this bug notification because you are a member of Ubuntu
OpenStack, which is subscribed to cinder in Ubuntu.
https://bugs.launchpad.net/bugs/2081742

Title:
  cinder powermax driver hangs when creating multiple image volumes

Status in OpenStack Cinder Charm:
  In Progress
Status in Cinder:
  In Progress
Status in cinder package in Ubuntu:
  Incomplete
Status in cinder source package in Jammy:
  Incomplete

Bug description:
  
  This bug is specific for the PowerMax driver in cinder and technically should be filed over the charm cinder-powermax but that charm has not been onboarded yet so filing in generic cinder for now.

  Note: this driver also covers the VMAX line, and the storage in
  quesiton is a VMAX-250F (all flash) using FC HBAs.

  Note2: A Unisphere appliance has been installed in the customer
  environment and seems to be working well (the driver talks to
  unisphere and not directly to the storage). So far, there is no reason
  to believe the Unisphere intallation has any issues -- it is also used
  by other systems without problems. I already checked resource
  utilization (cpu, mem, etc.) in the VM where Unisphere is running and
  it seems ok.

  The installation is Charmed Openstack, with Ubuntu Jammy and Openstack
  Yoga. Particular releases of the cinder charm is yoga/stable rev 699
  and cinder package version is 2:20.3.1-0ubuntu1.4.

  Cinder.conf has been configured (via a subordinate cinder driver
  charm) with a backend like this:

  =================================8<----------------------------------------
  [cinder-vmax]
  volume_driver = cinder.volume.drivers.dell_emc.powermax.fc.PowerMaxFCDriver
  volume_backend_name = vmax_fc
  san_ip = vmplnx5287.<redacted>
  san_login = openstack
  san_password = <redacted>
  powermax_array = 0005978xxxxx
  powermax_port_groups = [OS-FC-PG]
  powermax_service_level = Diamond
  powermax_srp = SRP_1
  retries = 800
  u4p_failover_autofailback = True
  use_multipath_for_image_xfer = True
  vmax_workload = NONE
  =================================8<----------------------------------------

  A type has been created in openstack to use this backend.

  Notes about the setup:

  - FC (fiber channel) HBAs seem to be working fine. As a test, customer can create volumes in the storage system and present them to the hosts without issues (Ubuntu finds them and you can format/mount/use).
  - Cinder driver seems to be communicating fine with Unisphere, you can see on the logs that it logs in and does issues all the api calls without problems.
  - Cinder-volume is running on the baremetals because of the FC cards (the service is disabled in the main cinder charm and a separate charm just for cinder-volume is added to the hosts).
  - If you try to create volumes inside openstack using "volume create --type vmax_fc" it works, volumes get created on the storage side without errors (you can check that with the storage management software).
  - If you try to create a batch of volumes (say 10 volumes) at the same time, they all get created without issues.
  - Those volumes created above can be assigned to VMs and used normally.
  - If you try to create a volume based on an image (volume create --type vmax_fc --image xxxx"), it does not matter if you just create the volume by itself or if is created as part of a VM/instance creation, it works. Volume gets created, it gets mounted somewhere using cinder-volume somewhere, the image is written in the volume.
  - The image creation is relatively fast, takes about 20 to 30 seconds (the whole process of creating the volume, presenting it to a host, downloading the image from glance, writing the image to it). The image in question is a ubuntu jammy image in QCOW2 format that has about 600MB. (I know about performance issues with big images, especially when both image and volume are in ceph RAW is preferred -- but in this case image is not in ceph so download needs to happen anyway and also image is small and it does happen fast).

  The issue happens when I try to create a batch (i.e. 10) volumes based
  on images. Most volumes, if not all, get stuck and never finish. It is
  not a matter of being slow, they just block forever. First few images
  block in "downloading" state and the rest block in "creating" state
  while waiting for slots in the queue.

  The problem seems related only to 1) being based on an image and 2)
  being a batch.

  The customer reports that he had this same issue when testing
  Kolla/Ansible Openstack before installing Charmed Openstack -- so it
  may be cinder code and not the charms.

To manage notifications about this bug go to:
https://bugs.launchpad.net/charm-cinder/+bug/2081742/+subscriptions




More information about the Ubuntu-openstack-bugs mailing list