-
Notifications
You must be signed in to change notification settings - Fork 1.2k
Description
ISSUE TYPE
- Bug Report
COMPONENT NAME
SolidFire plugin ('SolidFire" as opposed to "SolidFire Shared")
CLOUDSTACK VERSION
4.13 (master atm), but also observed by another community member on 4.11.3
CONFIGURATION
OS / ENVIRONMENT
VMware 6.5 tested
SUMMARY
Adding SolidFire Primary Storage via SolidFire plugin for VMware 6.5 fails with errors in mounting Datastore in ESXi hosts and raises the specific error in mgmt logs.
STEPS TO REPRODUCE
in vCenter, add iSCSI Software adapter to each ESXi hosts, configure proper network binding to vSwitchXXX for the iSCSI traffic (i.e. VLAN XXX so that ESXi can communicate with the SVIP on that VLAN), then add SolidFire (Zone wide (same problem with cluster-wide), protocol "Custom, provider "SolidFire", "Managed" box ticked and proper ULR - adding SF as Primary Storage is successful.
Try to spin a VM - that is when the things fail after a minutes or so.
Observing vCenter, the following thing happen (also check the screenshot)
- Static iSCSI target is added to ESXI hosts
- Rescanning HBAs
- Creating datastore same size as the volume/template itself with the name ending in "Centos-5.3-x64" or similar (name of the template)
- Deploying OVF template
- Unregistering VM
- Moving files around
- unmounting VMFS
- Removing iSCSI static targets
- Rescan HBA
- Again adding iSCSI static targets ???
- Rescan HBAs
- Rescan VMFS
- RENAME datastore (from the template-alike name to the root-volume-alike name, ending with ROOT-XXX.YY)
(... probably NOW problem happens ???))) - unmount datastore
- remove iSCSI targets.
The error from the ACS is:
message: Datastore '-iqn.2010-01.com.solidfire:hl1k.root-32.29-0' is not accessible. No connected and accessible host is attached to this datastore
The problem is - this datastore (in it's latest, renamed state, still attached to ESXi hosts) - is unmounted, but can't be removed, NOR can I mount it - If I try to manually mount it, I get the vCenter message of "Operation failed, diagnostics report: Unable to find volume uuid[5d7abd9a-273aa9d5-bffe-1e00d4010711] lvm [snap-329aa3ea-5d7abd01-a5c83210-c87c-1e00d4010711] devices"
Screenshot from vCenter attached - note that the last 2 entries (on top) are my attempt to manually mount an existing SF datastore. - i.e. there are zero failures on vCenter side while ACS is doing it's job - something is failing on ACS side.
