Re: [lustre-discuss] one ost down

2019-11-15 Thread Jongwoo Han
Strange, one drive failure in a raidset is not affected by rebuilding
another raidset. All OSTs have their own raid and they are independent.
And AFAIK, DDN is preconfigured to use RAID-6, so it will endure up to 2
drive or enclosure failure.
Have you tried replacing failed drive? If not, you better do first.
If it happens again after rebuild, another thing to try is to shut down the
OSSes by completely powerfing off entire shelves.

Regards,
Jongwoo Han

2019년 11월 15일 (금) 오후 6:01, Einar Næss Jensen 님이
작성:

>
> Hello dear lustre community.
>
>
> We have a lustre file system, where one ost is having problems.
>
> The underlying diskarray, an old sfa10k from DDN (without support), have
> one raidset with ca 1300 bad blocks. The bad blocks came about when one
> disk in the raid failed while another drive in other raidset was rebuilding.
>
>
> Now.
>
> The ost is offline, and the file system seems useable for new files, while
> old files on the corresponding ost is generating lots of kernel messages on
> the OSS.
>
> Quotainformation is not available though.
>
>
> Questions:
>
> May I assume that for new files, everything is fine, since they are not
> using the inactive device anyway?
>
> I tried to run e2fschk on the ost unmounted, while jobs were still running
> on the filesystem, and for a few minutes it seemd like this was working, as
> the filesystem seemed to come back complete afterwards. After a few minutes
> the ost failed again, though.
>
>
> Any pointers on how to rebuild/fix the ost and get it back is very much
> appreciated.
>
>
> Also how to regenerate the quotainformation, which is currently
> unavailable would help. With or without the troublesome OST.
>
>
>
>
> Best Regards
>
> Einar Næss Jensen (on flight to Denver)
>
>
>
> --
> Einar Næss Jensen
> NTNU HPC Section
> Norwegian University of Science and Technoloy
> Address: Høgskoleringen 7i
>  N-7491 Trondheim, NORWAY
> tlf: +47 90990249
> email:   einar.nass.jen...@ntnu.no
> ___
> lustre-discuss mailing list
> lustre-discuss@lists.lustre.org
> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>


-- 
Jongwoo Han
+82-505-227-6108
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org


Re: [lustre-discuss] one ost down

2019-11-15 Thread Andreas Dilger
If the HDD has enough bad sectors that it is reporting errors to user space 
then it means that all of the remapping sectors are already consumed will 
typically continue to have more errors in the future. It should be replaced 
rather than continuing to be used. 

I would agree with Marek that making a "dd" copy is safest, as it will avoid 
further errors appearing, and allows you to restore from the backup if 
something goes wrong with your repair attempt. 

Cheers, Andreas

> On Nov 15, 2019, at 05:56, Marek Magryś  wrote:
> 
> Hi Einar,
> 
> You can to run ddrescue from the OSS server to create a block-level
> backup of the OST. ddrescue should create a fairly decent backup from
> what is available on the block device. I would suggest to run such
> backup before you play with clearing the bad blocks, in worst case you
> will be able to restore the OST to the current state.
> 
> Cheers,
> Marek
> 
>  Original Message 
>> Hi Einar,
>> 
>>  
>> 
>> As for the OST in bad shape, if you have not cleared the bad blocks on
>> the storage system you’ll keep having IO errors when your server tries
>> to access these blocks, that’s kind of a protection mechanism and lots
>> of IO errors might give you many issues. The procedure to clean them up
>> is a bit of storage and filesystem surgery. I would suggest, this high
>> level view plan:
>> 
>>  
>> 
>>  * Obtain the bad blocks from the storage system (offset, size, etc…)
>>  * Map them to filesystem blocks: watch out, the storage system speaks
>>probably and for old systems about 512bytes blocks and the
>>filesystem blocks are 4KB, so you need to map storage blocks to
>>filesystem blocks
>>  * Clear the bad blocks on the storage system, each storage system has
>>their own commands to clear those. You’ll probably no longer have IO
>>errors accessing these sectors after clearing the bad blocks
>>  * Optional, zero the bad storage blocks with dd (and just these bad
>>blocks of course) to ignore the “trash” there might be on these blocks
>>  * Find out with debugfs which files are affected
>>  * Run e2fsck on the device
>> 
>>  
>> 
>> As I said, surgery, so if you really care about what you have on that
>> device try to do a block level backup before… But the minimum for sure
>> is that you need to clear the bad blocks, otherwise you get IO access
>> error on the device.
>> 
>>  
>> 
>> Regards,
>> 
>>  
>> 
>> Diego
>> 
>>  
>> 
>>  
>> 
>> *From: *lustre-discuss  on
>> behalf of Einar Næss Jensen 
>> *Date: *Friday, 15 November 2019 at 10:01
>> *To: *"lustre-discuss@lists.lustre.org" 
>> *Subject: *[lustre-discuss] one ost down
>> 
>>  
>> 
>>  
>> 
>> Hello dear lustre community.
>> 
>>  
>> 
>> We have a lustre file system, where one ost is having problems.
>> 
>> The underlying diskarray, an old sfa10k from DDN (without support), have
>> one raidset with ca 1300 bad blocks. The bad blocks came about when one
>> disk in the raid failed while another drive in other raidset was rebuilding.
>> 
>>  
>> 
>> Now.
>> 
>> The ost is offline, and the file system seems useable for new files,
>> while old files on the corresponding ost is generating lots of kernel
>> messages on the OSS.
>> 
>> Quotainformation is not available though.
>> 
>>  
>> 
>> Questions: 
>> 
>> May I assume that for new files, everything is fine, since they are not
>> using the inactive device anyway?
>> 
>> I tried to run e2fschk on the ost unmounted, while jobs were still
>> running on the filesystem, and for a few minutes it seemd like this was
>> working, as the filesystem seemed to come back complete afterwards.
>> After a few minutes the ost failed again, though.
>> 
>>  
>> 
>> Any pointers on how to rebuild/fix the ost and get it back is very much
>> appreciated. 
>> 
>>  
>> 
>> Also how to regenerate the quotainformation, which is currently
>> unavailable would help. With or without the troublesome OST.
>> 
>>  
>> 
>>  
>> 
>>  
>> 
>> Best Regards
>> 
>> Einar Næss Jensen (on flight to Denver)
>> 
>>  
>> 
>>  
>> 
>> -- 
>> Einar Næss Jensen
>> NTNU HPC Section
>> Norwegian University of Science and Technoloy
>> Address: Høgskoleringen 7i
>>  N-7491 Trondheim, NORWAY
>> tlf: +47 90990249
>> email:   einar.nass.jen...@ntnu.no
>> 
>> 
>> ___
>> lustre-discuss mailing list
>> lustre-discuss@lists.lustre.org
>> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>> 
> ___
> lustre-discuss mailing list
> lustre-discuss@lists.lustre.org
> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org


Re: [lustre-discuss] one ost down

2019-11-15 Thread Marek Magryś
Hi Einar,

You can to run ddrescue from the OSS server to create a block-level
backup of the OST. ddrescue should create a fairly decent backup from
what is available on the block device. I would suggest to run such
backup before you play with clearing the bad blocks, in worst case you
will be able to restore the OST to the current state.

Cheers,
Marek

 Original Message 
> Hi Einar,
> 
>  
> 
> As for the OST in bad shape, if you have not cleared the bad blocks on
> the storage system you’ll keep having IO errors when your server tries
> to access these blocks, that’s kind of a protection mechanism and lots
> of IO errors might give you many issues. The procedure to clean them up
> is a bit of storage and filesystem surgery. I would suggest, this high
> level view plan:
> 
>  
> 
>   * Obtain the bad blocks from the storage system (offset, size, etc…)
>   * Map them to filesystem blocks: watch out, the storage system speaks
> probably and for old systems about 512bytes blocks and the
> filesystem blocks are 4KB, so you need to map storage blocks to
> filesystem blocks
>   * Clear the bad blocks on the storage system, each storage system has
> their own commands to clear those. You’ll probably no longer have IO
> errors accessing these sectors after clearing the bad blocks
>   * Optional, zero the bad storage blocks with dd (and just these bad
> blocks of course) to ignore the “trash” there might be on these blocks
>   * Find out with debugfs which files are affected
>   * Run e2fsck on the device
> 
>  
> 
> As I said, surgery, so if you really care about what you have on that
> device try to do a block level backup before… But the minimum for sure
> is that you need to clear the bad blocks, otherwise you get IO access
> error on the device.
> 
>  
> 
> Regards,
> 
>  
> 
> Diego
> 
>  
> 
>  
> 
> *From: *lustre-discuss  on
> behalf of Einar Næss Jensen 
> *Date: *Friday, 15 November 2019 at 10:01
> *To: *"lustre-discuss@lists.lustre.org" 
> *Subject: *[lustre-discuss] one ost down
> 
>  
> 
>  
> 
> Hello dear lustre community.
> 
>  
> 
> We have a lustre file system, where one ost is having problems.
> 
> The underlying diskarray, an old sfa10k from DDN (without support), have
> one raidset with ca 1300 bad blocks. The bad blocks came about when one
> disk in the raid failed while another drive in other raidset was rebuilding.
> 
>  
> 
> Now.
> 
> The ost is offline, and the file system seems useable for new files,
> while old files on the corresponding ost is generating lots of kernel
> messages on the OSS.
> 
> Quotainformation is not available though.
> 
>  
> 
> Questions: 
> 
> May I assume that for new files, everything is fine, since they are not
> using the inactive device anyway?
> 
> I tried to run e2fschk on the ost unmounted, while jobs were still
> running on the filesystem, and for a few minutes it seemd like this was
> working, as the filesystem seemed to come back complete afterwards.
> After a few minutes the ost failed again, though.
> 
>  
> 
> Any pointers on how to rebuild/fix the ost and get it back is very much
> appreciated. 
> 
>  
> 
> Also how to regenerate the quotainformation, which is currently
> unavailable would help. With or without the troublesome OST.
> 
>  
> 
>  
> 
>  
> 
> Best Regards
> 
> Einar Næss Jensen (on flight to Denver)
> 
>  
> 
>  
> 
> -- 
> Einar Næss Jensen
> NTNU HPC Section
> Norwegian University of Science and Technoloy
> Address: Høgskoleringen 7i
>  N-7491 Trondheim, NORWAY
> tlf: +47 90990249
> email:   einar.nass.jen...@ntnu.no
> 
> 
> ___
> lustre-discuss mailing list
> lustre-discuss@lists.lustre.org
> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
> 
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org


Re: [lustre-discuss] one ost down

2019-11-15 Thread Moreno Diego (ID SIS)
Hi Einar,

As for the OST in bad shape, if you have not cleared the bad blocks on the 
storage system you’ll keep having IO errors when your server tries to access 
these blocks, that’s kind of a protection mechanism and lots of IO errors might 
give you many issues. The procedure to clean them up is a bit of storage and 
filesystem surgery. I would suggest, this high level view plan:


  *   Obtain the bad blocks from the storage system (offset, size, etc…)
  *   Map them to filesystem blocks: watch out, the storage system speaks 
probably and for old systems about 512bytes blocks and the filesystem blocks 
are 4KB, so you need to map storage blocks to filesystem blocks
  *   Clear the bad blocks on the storage system, each storage system has their 
own commands to clear those. You’ll probably no longer have IO errors accessing 
these sectors after clearing the bad blocks
  *   Optional, zero the bad storage blocks with dd (and just these bad blocks 
of course) to ignore the “trash” there might be on these blocks
  *   Find out with debugfs which files are affected
  *   Run e2fsck on the device

As I said, surgery, so if you really care about what you have on that device 
try to do a block level backup before… But the minimum for sure is that you 
need to clear the bad blocks, otherwise you get IO access error on the device.

Regards,

Diego


From: lustre-discuss  on behalf of 
Einar Næss Jensen 
Date: Friday, 15 November 2019 at 10:01
To: "lustre-discuss@lists.lustre.org" 
Subject: [lustre-discuss] one ost down




Hello dear lustre community.



We have a lustre file system, where one ost is having problems.

The underlying diskarray, an old sfa10k from DDN (without support), have one 
raidset with ca 1300 bad blocks. The bad blocks came about when one disk in the 
raid failed while another drive in other raidset was rebuilding.



Now.

The ost is offline, and the file system seems useable for new files, while old 
files on the corresponding ost is generating lots of kernel messages on the OSS.

Quotainformation is not available though.



Questions:

May I assume that for new files, everything is fine, since they are not using 
the inactive device anyway?

I tried to run e2fschk on the ost unmounted, while jobs were still running on 
the filesystem, and for a few minutes it seemd like this was working, as the 
filesystem seemed to come back complete afterwards. After a few minutes the ost 
failed again, though.



Any pointers on how to rebuild/fix the ost and get it back is very much 
appreciated.



Also how to regenerate the quotainformation, which is currently unavailable 
would help. With or without the troublesome OST.







Best Regards

Einar Næss Jensen (on flight to Denver)




--
Einar Næss Jensen
NTNU HPC Section
Norwegian University of Science and Technoloy
Address: Høgskoleringen 7i
 N-7491 Trondheim, NORWAY
tlf: +47 90990249
email:   einar.nass.jen...@ntnu.no
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org


[lustre-discuss] one ost down

2019-11-15 Thread Einar Næss Jensen

Hello dear lustre community.


We have a lustre file system, where one ost is having problems.

The underlying diskarray, an old sfa10k from DDN (without support), have one 
raidset with ca 1300 bad blocks. The bad blocks came about when one disk in the 
raid failed while another drive in other raidset was rebuilding.


Now.

The ost is offline, and the file system seems useable for new files, while old 
files on the corresponding ost is generating lots of kernel messages on the OSS.

Quotainformation is not available though.


Questions:

May I assume that for new files, everything is fine, since they are not using 
the inactive device anyway?

I tried to run e2fschk on the ost unmounted, while jobs were still running on 
the filesystem, and for a few minutes it seemd like this was working, as the 
filesystem seemed to come back complete afterwards. After a few minutes the ost 
failed again, though.


Any pointers on how to rebuild/fix the ost and get it back is very much 
appreciated.


Also how to regenerate the quotainformation, which is currently unavailable 
would help. With or without the troublesome OST.




Best Regards

Einar Næss Jensen (on flight to Denver)



--
Einar Næss Jensen
NTNU HPC Section
Norwegian University of Science and Technoloy
Address: Høgskoleringen 7i
 N-7491 Trondheim, NORWAY
tlf: +47 90990249
email:   einar.nass.jen...@ntnu.no
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org