The individual LUN looks good but the controller is showing amber, which
is confusing us. However other LUN's going through that controller are
mounting fine.
-Paul Edmon-
On 7/20/2022 3:08 PM, Colin Faber wrote:
raid check?
On Wed, Jul 20, 2022, 12:41 PM Paul Edmon <ped...@cfa.harvard.edu> wrote:
[root@holylfs02oss06 ~]# mount -t ldiskfs /dev/mapper/mpathd
/mnt/holylfs2-OST001f
mount: wrong fs type, bad option, bad superblock on
/dev/mapper/mpathd,
missing codepage or helper program, or other error
In some cases useful info is found in syslog - try
dmesg | tail or so.
e2fsck did not look good:
[root@holylfs02oss06 ~]# less OST001f.out
ext2fs_check_desc: Corrupt group descriptor: bad block for block
bitmap
e2fsck: Group descriptors look bad... trying backup blocks...
MMP interval is 10 seconds and total wait time is 42 seconds.
Please wait...
Superblock needs_recovery flag is clear, but journal has data.
Recovery flag not set in backup superblock, so running journal anyway.
Clear journal? no
Block bitmap for group 8128 is not in group. (block
3518518062363072290)
Relocate? no
Inode bitmap for group 8128 is not in group. (block
12235298632209565410)
Relocate? no
Inode table for group 8128 is not in group. (block
17751685088477790304)
WARNING: SEVERE DATA LOSS POSSIBLE.
Relocate? no
Block bitmap for group 8129 is not in group. (block
2193744380193356980)
Relocate? no
Inode bitmap for group 8129 is not in group. (block
4102707059848926418)
Relocate? no
It continues at length like that.
-Paul Edmon-
On 7/20/2022 2:31 PM, Colin Faber wrote:
Can you mount the target directly with -t ldiskfs ?
Also what does e2fsck report?
On Wed, Jul 20, 2022, 11:48 AM Paul Edmon via lustre-discuss
<lustre-discuss@lists.lustre.org> wrote:
We have a filesystem that we have running Lustre 2.10.4 in HA
mode using
IML. One of our OST's had some disk failures and after
reconstruction
of the RAID set it won't remount but gives:
[root@holylfs02oss06 ~]# mount -t lustre /dev/mapper/mpathd
/mnt/holylfs2-OST001f
Failed to initialize ZFS library: 256
mount.lustre: missing option mgsnode=<nid>
The weird thing is that we didn't build this with ZFS, the
devices are
all ldiskfs. We suspect some of the data is corrupt on the
disk but we
were wondering if anyone had seen this error before and if
there was a
solution.
-Paul Edmon-
_______________________________________________
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
_______________________________________________
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org