Skip to content

SOLIDFIRE: "SolidFire" plugin doesn't work for ROOT volumes with VMware (6.5) #3598

@andrijapanicsb

Description

@andrijapanicsb
ISSUE TYPE
  • Bug Report
COMPONENT NAME
SolidFire plugin ('SolidFire" as opposed to "SolidFire Shared")
CLOUDSTACK VERSION
4.13 (master atm), but also observed by another community member on 4.11.3
CONFIGURATION
OS / ENVIRONMENT

VMware 6.5 tested

SUMMARY

Adding SolidFire Primary Storage via SolidFire plugin for VMware 6.5 fails with errors in mounting Datastore in ESXi hosts and raises the specific error in mgmt logs.

STEPS TO REPRODUCE

in vCenter, add iSCSI Software adapter to each ESXi hosts, configure proper network binding to vSwitchXXX for the iSCSI traffic (i.e. VLAN XXX so that ESXi can communicate with the SVIP on that VLAN), then add SolidFire (Zone wide (same problem with cluster-wide), protocol "Custom, provider "SolidFire", "Managed" box ticked and proper ULR - adding SF as Primary Storage is successful.

Try to spin a VM - that is when the things fail after a minutes or so.

Observing vCenter, the following thing happen (also check the screenshot)

  • Static iSCSI target is added to ESXI hosts
  • Rescanning HBAs
  • Creating datastore same size as the volume/template itself with the name ending in "Centos-5.3-x64" or similar (name of the template)
  • Deploying OVF template
  • Unregistering VM
  • Moving files around
  • unmounting VMFS
  • Removing iSCSI static targets
  • Rescan HBA
  • Again adding iSCSI static targets ???
  • Rescan HBAs
  • Rescan VMFS
  • RENAME datastore (from the template-alike name to the root-volume-alike name, ending with ROOT-XXX.YY)
    (... probably NOW problem happens ???)))
  • unmount datastore
  • remove iSCSI targets.

The error from the ACS is:
message: Datastore '-iqn.2010-01.com.solidfire:hl1k.root-32.29-0' is not accessible. No connected and accessible host is attached to this datastore

The problem is - this datastore (in it's latest, renamed state, still attached to ESXi hosts) - is unmounted, but can't be removed, NOR can I mount it - If I try to manually mount it, I get the vCenter message of "Operation failed, diagnostics report: Unable to find volume uuid[5d7abd9a-273aa9d5-bffe-1e00d4010711] lvm [snap-329aa3ea-5d7abd01-a5c83210-c87c-1e00d4010711] devices"

Screenshot from vCenter attached - note that the last 2 entries (on top) are my attempt to manually mount an existing SF datastore. - i.e. there are zero failures on vCenter side while ACS is doing it's job - something is failing on ACS side.

image

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions