Gentoo Forums
Gentoo Forums
Gentoo Forums
Quick Search: in
Hard disk problem: "Logical unit not ready"
View unanswered posts
View posts from last 24 hours

 
Reply to topic    Gentoo Forums Forum Index Kernel & Hardware
View previous topic :: View next topic  
Author Message
antu456
n00b
n00b


Joined: 11 Sep 2013
Posts: 6

PostPosted: Wed Sep 11, 2013 9:01 pm    Post subject: Hard disk problem: "Logical unit not ready" Reply with quote

Hey,

On one of my gentoo systems I regularly get errors in dmesg which I'm a bit worried about.

The system has 7 disks in a software RAID6 (mdraid), which go to standby after a few minutes of inactivity to save energy. When I access the filesystem on the raid while the disks are in standby mode I sometimes (but not always) get these messages in dmesg:
Code:

[1221060.356007] sd 0:0:1:0: [sdb] Unhandled error code
[1221060.356013] sd 0:0:1:0: [sdb] 
[1221060.356015] Result: hostbyte=DID_OK driverbyte=DRIVER_OK
[1221060.356018] sd 0:0:1:0: [sdb] CDB:
[1221060.356019] Read(16): 88 00 00 00 00 00 1b 66 7e 00 00 00 00 08 00 00
[1221060.356031] end_request: I/O error, dev sdb, sector 459701760
[1221060.356050] sd 0:0:1:0: [sdb] Device not ready
[1221060.356051] sd 0:0:1:0: [sdb] 
[1221060.356052] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE
[1221060.356053] sd 0:0:1:0: [sdb] 
[1221060.356054] Sense Key : Not Ready [current]
[1221060.356055] sd 0:0:1:0: [sdb] 
[1221060.356057] Add. Sense: Logical unit not ready, initializing command required
[1221060.356058] sd 0:0:1:0: [sdb] CDB:
[1221060.356058] Read(16): 88 00 00 00 00 00 1b 66 7e 08 00 00 00 08 00 00
[1221060.356063] end_request: I/O error, dev sdb, sector 459701768
[1221060.356066] sd 0:0:1:0: [sdb] Device not ready
[1221060.356066] sd 0:0:1:0: [sdb] 
[1221060.356067] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE
[1221060.356068] sd 0:0:1:0: [sdb] 
[1221060.356069] Sense Key : Not Ready [current]
[1221060.356070] sd 0:0:1:0: [sdb] 
[1221060.356070] Add. Sense: Logical unit not ready, initializing command required
[1221060.356071] sd 0:0:1:0: [sdb] CDB:
[1221060.356072] Read(16): 88 00 00 00 00 00 1b 66 7e 10 00 00 00 08 00 00
[1221060.356076] end_request: I/O error, dev sdb, sector 459701776
[1221060.356080] sd 0:0:1:0: [sdb] Device not ready
[1221060.356081] sd 0:0:1:0: [sdb] 
[1221060.356082] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE
[1221060.356082] sd 0:0:1:0: [sdb] 
[1221060.356083] Sense Key : Not Ready [current]
[1221060.356084] sd 0:0:1:0: [sdb] 
[1221060.356085] Add. Sense: Logical unit not ready, initializing command required
[1221060.356086] sd 0:0:1:0: [sdb] CDB:
[1221060.356086] Read(16): 88 00 00 00 00 00 1b 66 7e 18 00 00 00 08 00 00
[1221060.356090] end_request: I/O error, dev sdb, sector 459701784
[1221060.356093] sd 0:0:1:0: [sdb] Device not ready
[1221060.356094] sd 0:0:1:0: [sdb] 
[1221060.356095] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE
[1221060.356096] sd 0:0:1:0: [sdb] 
[1221060.356096] Sense Key : Not Ready [current]
[1221060.356097] sd 0:0:1:0: [sdb] 
[1221060.356098] Add. Sense: Logical unit not ready, initializing command required
[1221060.356099] sd 0:0:1:0: [sdb] CDB:
[1221060.356099] Read(16): 88 00 00 00 00 00 1b 66 7e 20 00 00 00 08 00 00
[1221060.356104] end_request: I/O error, dev sdb, sector 459701792
[1221060.356107] sd 0:0:1:0: [sdb] Device not ready
[1221060.356108] sd 0:0:1:0: [sdb] 
[1221060.356108] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE
[1221060.356109] sd 0:0:1:0: [sdb] 
[1221060.356110] Sense Key : Not Ready [current]
[1221060.356111] sd 0:0:1:0: [sdb] 
[1221060.356112] Add. Sense: Logical unit not ready, initializing command required
[1221060.356113] sd 0:0:1:0: [sdb] CDB:
[1221060.356113] Read(16): 88 00 00 00 00 00 1b 66 7e 28 00 00 00 08 00 00
[1221060.356117] end_request: I/O error, dev sdb, sector 459701800

But after a few seconds I can access/use the filesystem normally, it doesn't have any errors/problems. The messages are always the same, just the sector number and the affected disks are different (but only sda, sdb and sdc are affected, the other disks never make these messages).

Hardware is:
Mainboard: Supermicro X9SCM-F
Disks: 7x HGST Deskstar IDK 4TB (0S03356)
HBA: IBM ServeRAID M1015 (flashed to IT mode so it just passes the disks through)

Some of the disks (I think 3) are connected to the mainboard, some to the M1015.

Should I be worried about this or can I ignore these messages? Could this have to do with Staggered Spin up? What do I need to do to fix this?
Back to top
View user's profile Send private message
NeddySeagoon
Administrator
Administrator


Joined: 05 Jul 2003
Posts: 32098
Location: 56N 3W

PostPosted: Wed Sep 11, 2013 10:14 pm    Post subject: Reply with quote

antu456,

Welcome to Gentoo.

Yes, you should be worried and you should have current validated backups.

At face value, you have at least one hardware problem. Several problems at a time are rare, so to keep the options down, we'll work on the idea that its a single problem.

Look for one thing all the affected disks have in common. HDD controller? PSU?
Is you PSU good enough to support 7 HDD?
Do you have more than 2 drives on the same power cable from the PSU?

As you allow the drives to spin down, which is a very bad thing for drive life, it might be a spin up issue too, depending on how spin up control is implemented.
SCSI allows drives to be spun up when they are addressed. This feature does not have to be used and I'm unsure if it got into the SATA spec. The makes for slow spin up times of a group of drives but almost eliminates the start up current spike caused by, in your case, 7 drives trying to spin up at the same time.
The spin up current is about 5x the normal drive motor run current. Can your PSU provide the +12 current to start your drives all at the same time?

There are no solutions there - just some things to look at at.
_________________
Regards,

NeddySeagoon

Computer users fall into two groups:-
those that do backups
those that have never had a hard drive fail.
Back to top
View user's profile Send private message
antu456
n00b
n00b


Joined: 11 Sep 2013
Posts: 6

PostPosted: Thu Sep 12, 2013 1:01 am    Post subject: Reply with quote

Thanks for your answer! :-)

The disks are inside 2 Cremax ICY DOCK MB455SPF cages (I have 8 hard disks total but the first isn't part of the raid. Additionally there is a SSD for the system but it's not inside one of the cages), 5 disks in the first, 3 in the second one. They use the same PSU (the system has only one), but different cables if I remember correctly.

The PSU (Enermax Platimax 500 Watt ATX 2.3) should be able to support 7 disks spinning up. It's a 500W PSU, when in idle the system uses around 40-50W and when disks are spinning up it's 150-200W for a few seconds, then 90-100W while they are running.

I know that spinning disks up/down often is bad for the drives but they are only needed 2-3 times a day, so they just spin up a few times per day which should be fine (according to HGST they can be spun up 600.000 times). According to SMART, the spin up time/load cycle count are fine (for sda):

Code:
  3 Spin_Up_Time            0x0007   126   126   024    Pre-fail  Always       -       611 (Average 615)
192 Power-Off_Retract_Count 0x0032   100   100   000    Old_age   Always       -       490
193 Load_Cycle_Count        0x0012   100   100   000    Old_age   Always       -       490


I've looked what the affected disks (sda, sdb, and sdc) have in common and found this:

Code:
# ls /sys/block/sd* -lah
lrwxrwxrwx 1 root root 0 29. Aug 00:24 /sys/block/sda -> ../devices/pci0000:00/0000:00:01.0/0000:01:00.0/host0/port-0:0/end_device-0:0/target0:0:0/0:0:0:0/block/sda
lrwxrwxrwx 1 root root 0 29. Aug 00:24 /sys/block/sdb -> ../devices/pci0000:00/0000:00:01.0/0000:01:00.0/host0/port-0:1/end_device-0:1/target0:0:1/0:0:1:0/block/sdb
lrwxrwxrwx 1 root root 0 29. Aug 00:24 /sys/block/sdc -> ../devices/pci0000:00/0000:00:01.0/0000:01:00.0/host0/port-0:2/end_device-0:2/target0:0:2/0:0:2:0/block/sdc
lrwxrwxrwx 1 root root 0 29. Aug 00:24 /sys/block/sdd -> ../devices/pci0000:00/0000:00:1f.2/ata1/host1/target1:0:0/1:0:0:0/block/sdd
lrwxrwxrwx 1 root root 0 29. Aug 00:24 /sys/block/sde -> ../devices/pci0000:00/0000:00:1f.2/ata2/host2/target2:0:0/2:0:0:0/block/sde
lrwxrwxrwx 1 root root 0 29. Aug 00:24 /sys/block/sdf -> ../devices/pci0000:00/0000:00:1f.2/ata3/host3/target3:0:0/3:0:0:0/block/sdf
lrwxrwxrwx 1 root root 0 29. Aug 00:24 /sys/block/sdg -> ../devices/pci0000:00/0000:00:1f.2/ata4/host4/target4:0:0/4:0:0:0/block/sdg
lrwxrwxrwx 1 root root 0 29. Aug 00:24 /sys/block/sdh -> ../devices/pci0000:00/0000:00:1f.2/ata5/host5/target5:0:0/5:0:0:0/block/sdh
lrwxrwxrwx 1 root root 0 29. Aug 00:24 /sys/block/sdi -> ../devices/pci0000:00/0000:00:1f.2/ata6/host6/target6:0:0/6:0:0:0/block/sdi


The 3 affected disks use the same controller, the disks using the other controller are not affected.

Code:
# lspci
00:00.0 Host bridge: Intel Corporation Xeon E3-1200 v2/Ivy Bridge DRAM Controller (rev 09)
00:01.0 PCI bridge: Intel Corporation Xeon E3-1200 v2/3rd Gen Core processor PCI Express Root Port (rev 09)
00:19.0 Ethernet controller: Intel Corporation 82579LM Gigabit Network Connection (rev 05)
00:1a.0 USB controller: Intel Corporation 6 Series/C200 Series Chipset Family USB Enhanced Host Controller #2 (rev 05)
00:1c.0 PCI bridge: Intel Corporation 6 Series/C200 Series Chipset Family PCI Express Root Port 1 (rev b5)
00:1c.4 PCI bridge: Intel Corporation 6 Series/C200 Series Chipset Family PCI Express Root Port 5 (rev b5)
00:1d.0 USB controller: Intel Corporation 6 Series/C200 Series Chipset Family USB Enhanced Host Controller #1 (rev 05)
00:1e.0 PCI bridge: Intel Corporation 82801 PCI Bridge (rev a5)
00:1f.0 ISA bridge: Intel Corporation C204 Chipset Family LPC Controller (rev 05)
00:1f.2 SATA controller: Intel Corporation 6 Series/C200 Series Chipset Family SATA AHCI Controller (rev 05)
00:1f.3 SMBus: Intel Corporation 6 Series/C200 Series Chipset Family SMBus Controller (rev 05)
01:00.0 Serial Attached SCSI controller: LSI Logic / Symbios Logic SAS2008 PCI-Express Fusion-MPT SAS-2 [Falcon] (rev 03)
03:00.0 Ethernet controller: Intel Corporation 82574L Gigabit Network Connection
04:03.0 VGA compatible controller: Matrox Electronics Systems Ltd. MGA G200eW WPCM450 (rev 0a)


So it looks like the controller could be the problem.

I found a post on the linux kernel mailing list where someone has a very similar problem here. Note that he has the same controller (they have different names but are the same hardware), and is also using SATA disks. The only difference is that I don't get errors by mdraid. Later in that thread someone posted a patch which was included in the kernel (I'm using 3.9.9-gentoo, and the patch is included there, I checked). It's this patch here. But it doesn't seem to be the same problem I have (my disks do spin up and there are no errors beside those posted above).

Not sending the disks to standby isn't a solution for me (because that would be a big waste of energy, they are only needed 2-3 hours per day when backups are made/I access some big files on the NAS).

Could this be a bug in the driver (I'll update to the latest stable kernel later to check if that fixes the problem)? Or a configuration problem? What to do next?

Edit: There is a firmware update for the controller, I'll install the new firmware tomorrow and see if that helps. :-)
Back to top
View user's profile Send private message
NeddySeagoon
Administrator
Administrator


Joined: 05 Jul 2003
Posts: 32098
Location: 56N 3W

PostPosted: Thu Sep 12, 2013 12:46 pm    Post subject: Reply with quote

antu456,

You have done your research very well.

Yes, it could be a bug in the firmware, the driver or even the kernel raid code, not checking that all members of the set are on line before it tries a disk access.
That 'all' is tricky - a raid6 needs only n-2 drives to provide access and you don't want to prevent access if you have sufficient drives.

Can you move one or more of the affected drives to the other controller and see if the moved drives are no longer affected?
_________________
Regards,

NeddySeagoon

Computer users fall into two groups:-
those that do backups
those that have never had a hard drive fail.
Back to top
View user's profile Send private message
antu456
n00b
n00b


Joined: 11 Sep 2013
Posts: 6

PostPosted: Thu Sep 12, 2013 8:34 pm    Post subject: Reply with quote

Unfortunately it's not possible to move the affected drives to the other controller as there are no more free SATA ports on it (there are only 6).

I have updated the BIOS/firmware of the controller to the latest versions.
I updated the kernel to version 3.10.7-gentoo.

I've gone through the controller's BIOS and changed the settings:

  • Changed Report Device Missing Delay and IO Device Missing Delay from 0 to 15 seconds.
  • Changed IO Timeout [...] values from 10 to 20 seconds.

Screenshots:
http://i.imgur.com/AuP607p.png
http://i.imgur.com/NVA5iBQ.png

And then I checked again, but the error is still occuring.

Next thing I found in the dmesg output is this:

Code:
[    1.026520] mpt2sas version 14.100.00.00 loaded
[    1.042013] mpt2sas0: 64 BIT PCI BUS DMA ADDRESSING SUPPORTED, total mem (8144424 kB)
[    1.058775] mpt2sas 0000:01:00.0: irq 41 for MSI/MSI-X
[    1.058787] mpt2sas0-msix0: PCI-MSI-X enabled: IRQ 41
[    1.067429] mpt2sas0: iomem(0x00000000df600000), mapped(0xffffc90000070000), size(16384)
[    1.083939] mpt2sas0: ioport(0x000000000000e000), size(256)
[    1.329563] mpt2sas0: sending message unit reset !!
[    1.342888] mpt2sas0: message unit reset: SUCCESS
[    1.476110] mpt2sas0: Allocated physical memory: size(7418 kB)
[    1.483214] mpt2sas0: Current Controller Queue Depth(3307), Max Controller Queue Depth(3432)
[    1.498147] mpt2sas0: Scatter Gather Elements per IO(128)
[    1.699724] mpt2sas0: LSISAS2008: FWVersion(17.00.01.00), ChipRevision(0x03), BiosVersion(07.33.00.00)
[    1.716001] mpt2sas0: Protocol=(Initiator,Target), Capabilities=(TLR,EEDP,Snapshot Buffer,Diag Trace Buffer,Task Set Full,NCQ)
[    1.736901] mpt2sas0: sending port enable !!
[    1.742405] mpt2sas0: host_add: handle(0x0001), sas_addr(0x500605b00604f420), phys(8)
[    1.749451] mpt2sas0: port enable: SUCCESS


The driver version is 14.100.00.00 but in the current Linux git tree there is a newer version (16), I'll try the latest git kernel now.

Apart from that, I'm running out of ideas. :(
Back to top
View user's profile Send private message
antu456
n00b
n00b


Joined: 11 Sep 2013
Posts: 6

PostPosted: Sat Jun 14, 2014 2:31 pm    Post subject: Reply with quote

I tried some things to solve the problem:

  • Replaced the cage of the affected disks with a Supermicro CSE-M35T-1.
  • Updated to gentoo-sources-3.15.0-r1.
  • Updated the firmware of the controller to the latest version (19), BIOS update was not possible because the checksum of the new BIOS was invalid (I tried downloading it again, but always get the checksum error).
  • Switched off power management for the controller(s).
    Code:
    echo '' > /sys/class/scsi_host/hostX/link_power_management_policy # For every host.
    echo 'on' > /sys/bus/pci/devices/0000:01:00.0/power/control

  • I changed the controller settings and set:

    • Changed Report Device Missing Delay and IO Device Missing Delay to 60 seconds.
    • Changed IO Timeout [...] values to 120 seconds.

  • I changed the device timeout for all disks to 120 seconds.
    Code:
    for i in /sys/block/sd?/device/timeout; do echo 120 > "$i"; done


But the problem is still there.
Quote:

[ 8207.757438] sd 0:0:3:0: [sdd] Unhandled error code
[ 8207.757444] sd 0:0:3:0: [sdd]
[ 8207.757446] Result: hostbyte=DID_OK driverbyte=DRIVER_OK
[ 8207.757448] sd 0:0:3:0: [sdd] CDB:
[ 8207.757450] Read(16): 88 00 00 00 00 00 15 99 ad 08 00 00 00 08 00 00
[ 8207.757462] end_request: I/O error, dev sdd, sector 362392840
[ 8207.757486] sd 0:0:3:0: [sdd] Device not ready
[ 8207.757488] sd 0:0:3:0: [sdd]
[ 8207.757490] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE
[ 8207.757492] sd 0:0:3:0: [sdd]
[ 8207.757493] Sense Key : Not Ready [current]
[ 8207.757496] sd 0:0:3:0: [sdd]
[ 8207.757499] Add. Sense: Logical unit not ready, initializing command required
[ 8207.757501] sd 0:0:3:0: [sdd] CDB:
[ 8207.757502] Read(16): 88 00 00 00 00 00 15 99 ad 10 00 00 00 08 00 00
[ 8207.757511] end_request: I/O error, dev sdd, sector 362392848
[ 8207.757517] sd 0:0:3:0: [sdd] Device not ready
[ 8207.757519] sd 0:0:3:0: [sdd]
[ 8207.757520] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE
[ 8207.757522] sd 0:0:3:0: [sdd]
[ 8207.757523] Sense Key : Not Ready [current]
[ 8207.757526] sd 0:0:3:0: [sdd]
[ 8207.757527] Add. Sense: Logical unit not ready, initializing command required
[ 8207.757529] sd 0:0:3:0: [sdd] CDB:
[ 8207.757530] Read(16): 88 00 00 00 00 00 15 99 ad 18 00 00 00 08 00 00
[ 8207.757539] end_request: I/O error, dev sdd, sector 362392856
[ 8207.757546] sd 0:0:3:0: [sdd] Device not ready
[ 8207.757548] sd 0:0:3:0: [sdd]
[ 8207.757549] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE
[ 8207.757551] sd 0:0:3:0: [sdd]
[ 8207.757552] Sense Key : Not Ready [current]
[ 8207.757554] sd 0:0:3:0: [sdd]
[ 8207.757556] Add. Sense: Logical unit not ready, initializing command required
[ 8207.757558] sd 0:0:3:0: [sdd] CDB:
[ 8207.757559] Read(16): 88 00 00 00 00 00 15 99 ad 20 00 00 00 08 00 00
[ 8207.757567] end_request: I/O error, dev sdd, sector 362392864
[ 8207.757574] sd 0:0:3:0: [sdd] Device not ready
[ 8207.757575] sd 0:0:3:0: [sdd]
[ 8207.757577] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE
[ 8207.757578] sd 0:0:3:0: [sdd]
[ 8207.757579] Sense Key : Not Ready [current]
[ 8207.757582] sd 0:0:3:0: [sdd]
[ 8207.757583] Add. Sense: Logical unit not ready, initializing command required
[ 8207.757585] sd 0:0:3:0: [sdd] CDB:
[ 8207.757586] Read(16): 88 00 00 00 00 00 15 99 ad 28 00 00 00 08 00 00
[ 8207.757595] end_request: I/O error, dev sdd, sector 362392872
[ 8207.757601] sd 0:0:3:0: [sdd] Device not ready
[ 8207.757603] sd 0:0:3:0: [sdd]
[ 8207.757605] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE
[ 8207.757606] sd 0:0:3:0: [sdd]
[ 8207.757607] Sense Key : Not Ready [current]
[ 8207.757610] sd 0:0:3:0: [sdd]
[ 8207.757611] Add. Sense: Logical unit not ready, initializing command required
[ 8207.757613] sd 0:0:3:0: [sdd] CDB:
[ 8207.757614] Read(16): 88 00 00 00 00 00 15 99 ad 30 00 00 00 08 00 00
[ 8207.757623] end_request: I/O error, dev sdd, sector 362392880
[ 8234.946594] sd 0:0:2:0: [sdc] Unhandled error code
[ 8234.946600] sd 0:0:2:0: [sdc]
[ 8234.946602] Result: hostbyte=DID_OK driverbyte=DRIVER_OK
[ 8234.946604] sd 0:0:2:0: [sdc] CDB:
[ 8234.946606] Read(16): 88 00 00 00 00 00 44 19 ad 08 00 00 00 08 00 00
[ 8234.946617] end_request: I/O error, dev sdc, sector 1142533384
[ 8234.946634] sd 0:0:2:0: [sdc] Device not ready
[ 8234.946637] sd 0:0:2:0: [sdc]
[ 8234.946638] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE
[ 8234.946640] sd 0:0:2:0: [sdc]
[ 8234.946641] Sense Key : Not Ready [current]
[ 8234.946644] sd 0:0:2:0: [sdc]
[ 8234.946646] Add. Sense: Logical unit not ready, initializing command required
[ 8234.946648] sd 0:0:2:0: [sdc] CDB:
[ 8234.946650] Read(16): 88 00 00 00 00 00 44 19 ad 10 00 00 00 08 00 00
[ 8234.946659] end_request: I/O error, dev sdc, sector 1142533392
[ 8234.946664] sd 0:0:2:0: [sdc] Device not ready
[ 8234.946666] sd 0:0:2:0: [sdc]
[ 8234.946668] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE
[ 8234.946669] sd 0:0:2:0: [sdc]
[ 8234.946670] Sense Key : Not Ready [current]
[ 8234.946673] sd 0:0:2:0: [sdc]
[ 8234.946674] Add. Sense: Logical unit not ready, initializing command required
[ 8234.946676] sd 0:0:2:0: [sdc] CDB:
[ 8234.946677] Read(16): 88 00 00 00 00 00 44 19 ad 18 00 00 00 08 00 00
[ 8234.946686] end_request: I/O error, dev sdc, sector 1142533400


Somebody has another idea what I could try?

I want the disks to suspend after ~8 minutes, and I use this command to configure the standby settings of the disks:
Code:
hdparm -B 127 -S 100 <Path to disk>

Do I have to configure something else like LVM/MDRAID/Linux so that it recognizes that the disks could be in standby and waits until they wake up before sending commands?

I can reproduce the problem by:
Code:

hdparm -y /dev/<Disk> # Send disk to standby.
# Wait a little...
ls -lah /mnt/<Path on RAID>
Back to top
View user's profile Send private message
skywalker67
n00b
n00b


Joined: 23 Jul 2014
Posts: 1

PostPosted: Sat Jul 26, 2014 8:15 am    Post subject: Reply with quote

Hi antu456, I am experiencing the same problem as you on LSI 9211-8i and 64-bit Debian with kernel from testing, current version 3.14-1.

Like in your case the disk actually wakes up and I get one hostbyte=DID_OK driverbyte=DRIVER_OK and couple of hostbyte=DID_OK driverbyte=DRIVER_SENSE, in dmesg I can see end_request: I/O error but no error on higher level. I am setting the standby timeout on drives with hdparm -S 150 and sometimes I am waking them up with the same command. (I am not going to specifics why I do that, it is probably not important.) I am yet not sure if the problem occurs with standard I/O from system or when I am waking up the drive with -S 150 command.

The problem occurs only sometimes. I have this controller only couple of weeks and I have experienced it fewer than 10 times. Drives are going to standby and waking up couple of times a day. Until now the errors were always on same drive on last port of the card, marked as slot 7 in dmesg / kernel messages. I did not yet test reconnecting drives to other ports. What is your setup? Which exact controller do you have, on which drive does it happen and how often?
Back to top
View user's profile Send private message
antu456
n00b
n00b


Joined: 11 Sep 2013
Posts: 6

PostPosted: Sat Oct 18, 2014 12:32 pm    Post subject: Reply with quote

Unfortunately the problem persists.

I've now updated to gentoo-sources-3.17.1 but still get the errors. I'm just ignoring them as they seem to not cause any problems/corruptions.
Back to top
View user's profile Send private message
frostschutz
Advocate
Advocate


Joined: 22 Feb 2005
Posts: 2458
Location: Germany

PostPosted: Sat Oct 18, 2014 1:48 pm    Post subject: Reply with quote

I'm using a 7 disk raid5 and sending disks to standby works fine for me. Nothing in dmesg. It even does "staggered spinup", when I try to access the HDD filesystems. So it waits for the first disk to wait and return data, then proceeds to the next disk and so on... (with 7 disks this makes the wakeup process kinda slow). No raid controller involved though, all onboard. So I assume this is an issue with your controller(s)...?
_________________
Linux-User @ VServer (Debian), Desktop (Gentoo), Netbook (Ubuntu), Router (OpenWRT), PDA (Cacko), Smartphone (Android)
Back to top
View user's profile Send private message
Display posts from previous:   
Reply to topic    Gentoo Forums Forum Index Kernel & Hardware All times are GMT
Page 1 of 1

 
Jump to:  
You cannot post new topics in this forum
You cannot reply to topics in this forum
You cannot edit your posts in this forum
You cannot delete your posts in this forum
You cannot vote in polls in this forum