[ome-users] Storage issues

Benjamin Schmid Benjamin.Schmid at fau.de
Tue Sep 4 19:49:35 BST 2018


Dear all,

This question is not really related to OMERO, but maybe some of you have 
come across this before:

We have a Thecus storage system (N16000pro) that's configured as a RAID 
6 and connected via iSCSI to a machine that runs an OMERO server. The 
Thecus system provides two LUNs (it's 2 because the maximum size of a 
LUN is 16 TB). They show up on the server as 2 partitions, /dev/sdc1 and 
/dev/sdd1. LVM2 is used to combine the 2 partitions into one logical 
volume (/dev/vg0/lv0).

When I tried to reboot the server today, the logical volume wasn't 
mounted and syslog shows lots of scary error messages:
---
Sep  4 13:50:43 romulus kernel: [   20.964830] sd 8:0:0:1: [sdc] tag#0 
FAILED Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE
Sep  4 13:50:43 romulus kernel: [   20.964842] sd 8:0:0:1: [sdc] tag#0 
Sense Key : Not Ready [current]
Sep  4 13:50:43 romulus kernel: [   20.964847] sd 8:0:0:1: [sdc] tag#0 
Add. Sense: Logical unit communication failure
Sep  4 13:50:43 romulus kernel: [   20.964852] sd 8:0:0:1: [sdc] tag#0 
CDB: Write(10) 2a 00 b8 04 06 01 00 00 01 00
Sep  4 13:50:43 romulus kernel: [   20.964856] blk_update_request: I/O 
error, dev sdc, sector 24698171400
Sep  4 13:50:43 romulus kernel: [   20.964861] Buffer I/O error on dev 
dm-0, logical block 7378305025, lost async page write
Sep  4 13:50:43 romulus kernel: [   20.966575] JBD2: recovery failed
Sep  4 13:50:43 romulus kernel: [   20.966578] EXT4-fs (dm-0): error 
loading journal
Sep  4 13:50:43 romulus mount[1690]: mount: /dev/mapper/vg0-lv0: can't 
read superblock
Sep  4 13:50:43 romulus systemd[1]: srv.mount: Mount process exited, 
code=exited status=32
Sep  4 13:50:43 romulus systemd[1]: Failed to mount /srv.
---

I checked the filesystem with fsck, and eventually let fsck everything 
it suggested. I could mount the volume again, but since then I'm getting 
lots of errors (dmesg):
[ 2822.725945] blk_update_request: I/O error, dev sdc, sector 24358432768
[ 2822.731819] sd 8:0:0:1: [sdc] tag#19 FAILED Result: hostbyte=DID_OK 
driverbyte=DRIVER_SENSE
[ 2822.731829] sd 8:0:0:1: [sdc] tag#19 Sense Key : Not Ready [current]
[ 2822.731846] sd 8:0:0:1: [sdc] tag#19 Add. Sense: Logical unit 
communication failure
[ 2822.731853] sd 8:0:0:1: [sdc] tag#19 CDB: Write(10) 2a 00 b5 84 06 00 
00 00 10 00

I'm now wondering what goes wrong. In the Thecus OS, I run the SMART 
utility on each hard-drive, and all seem to be fine. Is it on the RAID 
level where the error occurs (instead of the HDD level)? Or is it the 
LVM2 level? Is there anything I can do from here?

I'm more than grateful for any ideas or suggestions.

Again, please aplogize this off-topic question. If you know any better 
place where to put it, please let me know.

Best wishes,
Benjamin



More information about the ome-users mailing list