Join GitHub today
0.6.5.6 - I/O timeout during disk spin up #4638
see #3856 for further details. Bug is still present:
i'm still experiencing this bug with 0.6.5.6 on Ubuntu 16.04. with same chipset (SAS2008).
This is what i did:
i created a ext3 filesystem on the device and sent it to sleep. Then started the fily copy again. Result: No messages in dmesg.
I also checked the original file with the copied one - they are identical. So this bug has to do with ZFS and is not closed.Any Ideas? Need further Information @behlendorf ?
I also see this error for my pool. I also only see this error in combination with my zfs hdds in the syslog. It always happens, when my hdds have to wake up after a spindown (127). There are no errors in the hdd smart informations.
My hardware specs:
My HBA is pci passed through to the kvm. The mpt3sas modules are blacklisted on the host system.
Host-OS - "Proxmox":
Guest-OS (KVM) - Openmediavault3.0.41:
As you can see, I also use the proxmox kernel in the kvm.
I use the following zfs packages which have a dependency to the openmediavault-plugin "openmediavault-zfs":
If you need other informations, please tell me what you need.
Thanks and greetings Hoppel
OK, I tried another thing. I use 8x4TB WD Red hdds behind my lsi sas3008 controller. I read that there is a tool to deactivate the automatic spindown in the hdds firmware.
So I downloaded the "Idle3-tools" to my openmediavault (debian jessie) kvm.
The default value for my disks was:
So I decided to deactivate the default spindown with the following command for all 8 disks:
I power cycled the server completely, started again and had a look at the result with the following command:
At this stage of configuration I don't have any issues/errors in the syslog while opening a samba-share with a zfs file system as a basis.
After that I configured my "/etc/hdparm.conf" with the openmediavault webui the following way:
This way openmediavault has the privileges to spindown the disk after 20 minutes.
Now I see the following on the command line:
20 minutes later I see the following on the command line:
So we see that the spindown controlled by openmediavault works fine. Now I opened a file from one of my samba-shares with zfs as a file system. I can see the disks spinning up with hdparm and I see the following messages in the logfile again:
complete syslog: http://pastebin.com/9A300u3R
So that didn't help at all and i brought it back to default values:
What do you think about this?
A last check can be to clone the kvm to bare metal and to check the whole thing again. Maybe it has something to do with kvm or with pci passthrough. But for this I need some time.
referenced this issue
Oct 10, 2016
referenced this issue
Oct 10, 2016
I also encountered this issues.
There might be problem between zfs and "mpt3sas" driver.
Is it possible for you to reduce your spin up time in your controller bios? Maybe it's possible for you to stagger spin up two or three disks at a time. This should be possible if your controller is flashed to "it mode" and if your psu is powerful enough.
For me it's not possible to check this at the moment, because I use a beta firmware from supermicro, where the option to spin up some disks at a time is not available.
How many disks do you use behind your sas2008 controller for zfs? How long do you have to wait until all disks got up?
Same problem on Z87 Extreme11/ac -> 22 x SATA3 (16 x SAS3 12.0 Gb/s + 6 x SATA3 6.0 Gb/s) from LSI SAS 3008 Controller+ 3X24R Expander
OS: Ubuntu 18.04 dev
ZFS hangs on spin-up of SATA HDDs. So I assume It's a problem between LSI controller driver and ZFS.
I'll try BIOS updates, let's see if it fix the problem
UPDATE: I've updated MB BIOS and flash SAS Contoller to IT mode with a newest available FW from 9300 card. This did not help with the disk spin-up problem. funny enough it's not only the ZFS freezes but hddtemp and smartctl too. This issue might be related not to ZFS but to misbehavior of mpt3sas itself.
Please let me know if you found any solution or workarounds to the freezing disk on spin-up?
I have the same issue with SAS drives.
@d-helios one ugly hack to paper over the issue is to tweak
(tweak as appropriate for you).
You probably do not need the modprobe as the module will normally be loaded by this time, but for testing (systemctl stop/start w/ rmmod) it is needed.