Re: [lustre-discuss] one ost down
Strange, one drive failure in a raidset is not affected by rebuilding another raidset. All OSTs have their own raid and they are independent. And AFAIK, DDN is preconfigured to use RAID-6, so it will endure up to 2 drive or enclosure failure. Have you tried replacing failed drive? If not, you better do first. If it happens again after rebuild, another thing to try is to shut down the OSSes by completely powerfing off entire shelves. Regards, Jongwoo Han 2019년 11월 15일 (금) 오후 6:01, Einar Næss Jensen 님이 작성: > > Hello dear lustre community. > > > We have a lustre file system, where one ost is having problems. > > The underlying diskarray, an old sfa10k from DDN (without support), have > one raidset with ca 1300 bad blocks. The bad blocks came about when one > disk in the raid failed while another drive in other raidset was rebuilding. > > > Now. > > The ost is offline, and the file system seems useable for new files, while > old files on the corresponding ost is generating lots of kernel messages on > the OSS. > > Quotainformation is not available though. > > > Questions: > > May I assume that for new files, everything is fine, since they are not > using the inactive device anyway? > > I tried to run e2fschk on the ost unmounted, while jobs were still running > on the filesystem, and for a few minutes it seemd like this was working, as > the filesystem seemed to come back complete afterwards. After a few minutes > the ost failed again, though. > > > Any pointers on how to rebuild/fix the ost and get it back is very much > appreciated. > > > Also how to regenerate the quotainformation, which is currently > unavailable would help. With or without the troublesome OST. > > > > > Best Regards > > Einar Næss Jensen (on flight to Denver) > > > > -- > Einar Næss Jensen > NTNU HPC Section > Norwegian University of Science and Technoloy > Address: Høgskoleringen 7i > N-7491 Trondheim, NORWAY > tlf: +47 90990249 > email: einar.nass.jen...@ntnu.no > ___ > lustre-discuss mailing list > lustre-discuss@lists.lustre.org > http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org > -- Jongwoo Han +82-505-227-6108 ___ lustre-discuss mailing list lustre-discuss@lists.lustre.org http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
Re: [lustre-discuss] one ost down
If the HDD has enough bad sectors that it is reporting errors to user space then it means that all of the remapping sectors are already consumed will typically continue to have more errors in the future. It should be replaced rather than continuing to be used. I would agree with Marek that making a "dd" copy is safest, as it will avoid further errors appearing, and allows you to restore from the backup if something goes wrong with your repair attempt. Cheers, Andreas > On Nov 15, 2019, at 05:56, Marek Magryś wrote: > > Hi Einar, > > You can to run ddrescue from the OSS server to create a block-level > backup of the OST. ddrescue should create a fairly decent backup from > what is available on the block device. I would suggest to run such > backup before you play with clearing the bad blocks, in worst case you > will be able to restore the OST to the current state. > > Cheers, > Marek > > Original Message >> Hi Einar, >> >> >> >> As for the OST in bad shape, if you have not cleared the bad blocks on >> the storage system you’ll keep having IO errors when your server tries >> to access these blocks, that’s kind of a protection mechanism and lots >> of IO errors might give you many issues. The procedure to clean them up >> is a bit of storage and filesystem surgery. I would suggest, this high >> level view plan: >> >> >> >> * Obtain the bad blocks from the storage system (offset, size, etc…) >> * Map them to filesystem blocks: watch out, the storage system speaks >>probably and for old systems about 512bytes blocks and the >>filesystem blocks are 4KB, so you need to map storage blocks to >>filesystem blocks >> * Clear the bad blocks on the storage system, each storage system has >>their own commands to clear those. You’ll probably no longer have IO >>errors accessing these sectors after clearing the bad blocks >> * Optional, zero the bad storage blocks with dd (and just these bad >>blocks of course) to ignore the “trash” there might be on these blocks >> * Find out with debugfs which files are affected >> * Run e2fsck on the device >> >> >> >> As I said, surgery, so if you really care about what you have on that >> device try to do a block level backup before… But the minimum for sure >> is that you need to clear the bad blocks, otherwise you get IO access >> error on the device. >> >> >> >> Regards, >> >> >> >> Diego >> >> >> >> >> >> *From: *lustre-discuss on >> behalf of Einar Næss Jensen >> *Date: *Friday, 15 November 2019 at 10:01 >> *To: *"lustre-discuss@lists.lustre.org" >> *Subject: *[lustre-discuss] one ost down >> >> >> >> >> >> Hello dear lustre community. >> >> >> >> We have a lustre file system, where one ost is having problems. >> >> The underlying diskarray, an old sfa10k from DDN (without support), have >> one raidset with ca 1300 bad blocks. The bad blocks came about when one >> disk in the raid failed while another drive in other raidset was rebuilding. >> >> >> >> Now. >> >> The ost is offline, and the file system seems useable for new files, >> while old files on the corresponding ost is generating lots of kernel >> messages on the OSS. >> >> Quotainformation is not available though. >> >> >> >> Questions: >> >> May I assume that for new files, everything is fine, since they are not >> using the inactive device anyway? >> >> I tried to run e2fschk on the ost unmounted, while jobs were still >> running on the filesystem, and for a few minutes it seemd like this was >> working, as the filesystem seemed to come back complete afterwards. >> After a few minutes the ost failed again, though. >> >> >> >> Any pointers on how to rebuild/fix the ost and get it back is very much >> appreciated. >> >> >> >> Also how to regenerate the quotainformation, which is currently >> unavailable would help. With or without the troublesome OST. >> >> >> >> >> >> >> >> Best Regards >> >> Einar Næss Jensen (on flight to Denver) >> >> >> >> >> >> -- >> Einar Næss Jensen >> NTNU HPC Section >> Norwegian University of Science and Technoloy >> Address: Høgskoleringen 7i >> N-7491 Trondheim, NORWAY >> tlf: +47 90990249 >> email: einar.nass.jen...@ntnu.no >> >> >> ___ >> lustre-discuss mailing list >> lustre-discuss@lists.lustre.org >> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org >> > ___ > lustre-discuss mailing list > lustre-discuss@lists.lustre.org > http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org ___ lustre-discuss mailing list lustre-discuss@lists.lustre.org http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
Re: [lustre-discuss] one ost down
Hi Einar, You can to run ddrescue from the OSS server to create a block-level backup of the OST. ddrescue should create a fairly decent backup from what is available on the block device. I would suggest to run such backup before you play with clearing the bad blocks, in worst case you will be able to restore the OST to the current state. Cheers, Marek Original Message > Hi Einar, > > > > As for the OST in bad shape, if you have not cleared the bad blocks on > the storage system you’ll keep having IO errors when your server tries > to access these blocks, that’s kind of a protection mechanism and lots > of IO errors might give you many issues. The procedure to clean them up > is a bit of storage and filesystem surgery. I would suggest, this high > level view plan: > > > > * Obtain the bad blocks from the storage system (offset, size, etc…) > * Map them to filesystem blocks: watch out, the storage system speaks > probably and for old systems about 512bytes blocks and the > filesystem blocks are 4KB, so you need to map storage blocks to > filesystem blocks > * Clear the bad blocks on the storage system, each storage system has > their own commands to clear those. You’ll probably no longer have IO > errors accessing these sectors after clearing the bad blocks > * Optional, zero the bad storage blocks with dd (and just these bad > blocks of course) to ignore the “trash” there might be on these blocks > * Find out with debugfs which files are affected > * Run e2fsck on the device > > > > As I said, surgery, so if you really care about what you have on that > device try to do a block level backup before… But the minimum for sure > is that you need to clear the bad blocks, otherwise you get IO access > error on the device. > > > > Regards, > > > > Diego > > > > > > *From: *lustre-discuss on > behalf of Einar Næss Jensen > *Date: *Friday, 15 November 2019 at 10:01 > *To: *"lustre-discuss@lists.lustre.org" > *Subject: *[lustre-discuss] one ost down > > > > > > Hello dear lustre community. > > > > We have a lustre file system, where one ost is having problems. > > The underlying diskarray, an old sfa10k from DDN (without support), have > one raidset with ca 1300 bad blocks. The bad blocks came about when one > disk in the raid failed while another drive in other raidset was rebuilding. > > > > Now. > > The ost is offline, and the file system seems useable for new files, > while old files on the corresponding ost is generating lots of kernel > messages on the OSS. > > Quotainformation is not available though. > > > > Questions: > > May I assume that for new files, everything is fine, since they are not > using the inactive device anyway? > > I tried to run e2fschk on the ost unmounted, while jobs were still > running on the filesystem, and for a few minutes it seemd like this was > working, as the filesystem seemed to come back complete afterwards. > After a few minutes the ost failed again, though. > > > > Any pointers on how to rebuild/fix the ost and get it back is very much > appreciated. > > > > Also how to regenerate the quotainformation, which is currently > unavailable would help. With or without the troublesome OST. > > > > > > > > Best Regards > > Einar Næss Jensen (on flight to Denver) > > > > > > -- > Einar Næss Jensen > NTNU HPC Section > Norwegian University of Science and Technoloy > Address: Høgskoleringen 7i > N-7491 Trondheim, NORWAY > tlf: +47 90990249 > email: einar.nass.jen...@ntnu.no > > > ___ > lustre-discuss mailing list > lustre-discuss@lists.lustre.org > http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org > ___ lustre-discuss mailing list lustre-discuss@lists.lustre.org http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
Re: [lustre-discuss] one ost down
Hi Einar, As for the OST in bad shape, if you have not cleared the bad blocks on the storage system you’ll keep having IO errors when your server tries to access these blocks, that’s kind of a protection mechanism and lots of IO errors might give you many issues. The procedure to clean them up is a bit of storage and filesystem surgery. I would suggest, this high level view plan: * Obtain the bad blocks from the storage system (offset, size, etc…) * Map them to filesystem blocks: watch out, the storage system speaks probably and for old systems about 512bytes blocks and the filesystem blocks are 4KB, so you need to map storage blocks to filesystem blocks * Clear the bad blocks on the storage system, each storage system has their own commands to clear those. You’ll probably no longer have IO errors accessing these sectors after clearing the bad blocks * Optional, zero the bad storage blocks with dd (and just these bad blocks of course) to ignore the “trash” there might be on these blocks * Find out with debugfs which files are affected * Run e2fsck on the device As I said, surgery, so if you really care about what you have on that device try to do a block level backup before… But the minimum for sure is that you need to clear the bad blocks, otherwise you get IO access error on the device. Regards, Diego From: lustre-discuss on behalf of Einar Næss Jensen Date: Friday, 15 November 2019 at 10:01 To: "lustre-discuss@lists.lustre.org" Subject: [lustre-discuss] one ost down Hello dear lustre community. We have a lustre file system, where one ost is having problems. The underlying diskarray, an old sfa10k from DDN (without support), have one raidset with ca 1300 bad blocks. The bad blocks came about when one disk in the raid failed while another drive in other raidset was rebuilding. Now. The ost is offline, and the file system seems useable for new files, while old files on the corresponding ost is generating lots of kernel messages on the OSS. Quotainformation is not available though. Questions: May I assume that for new files, everything is fine, since they are not using the inactive device anyway? I tried to run e2fschk on the ost unmounted, while jobs were still running on the filesystem, and for a few minutes it seemd like this was working, as the filesystem seemed to come back complete afterwards. After a few minutes the ost failed again, though. Any pointers on how to rebuild/fix the ost and get it back is very much appreciated. Also how to regenerate the quotainformation, which is currently unavailable would help. With or without the troublesome OST. Best Regards Einar Næss Jensen (on flight to Denver) -- Einar Næss Jensen NTNU HPC Section Norwegian University of Science and Technoloy Address: Høgskoleringen 7i N-7491 Trondheim, NORWAY tlf: +47 90990249 email: einar.nass.jen...@ntnu.no ___ lustre-discuss mailing list lustre-discuss@lists.lustre.org http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
[lustre-discuss] one ost down
Hello dear lustre community. We have a lustre file system, where one ost is having problems. The underlying diskarray, an old sfa10k from DDN (without support), have one raidset with ca 1300 bad blocks. The bad blocks came about when one disk in the raid failed while another drive in other raidset was rebuilding. Now. The ost is offline, and the file system seems useable for new files, while old files on the corresponding ost is generating lots of kernel messages on the OSS. Quotainformation is not available though. Questions: May I assume that for new files, everything is fine, since they are not using the inactive device anyway? I tried to run e2fschk on the ost unmounted, while jobs were still running on the filesystem, and for a few minutes it seemd like this was working, as the filesystem seemed to come back complete afterwards. After a few minutes the ost failed again, though. Any pointers on how to rebuild/fix the ost and get it back is very much appreciated. Also how to regenerate the quotainformation, which is currently unavailable would help. With or without the troublesome OST. Best Regards Einar Næss Jensen (on flight to Denver) -- Einar Næss Jensen NTNU HPC Section Norwegian University of Science and Technoloy Address: Høgskoleringen 7i N-7491 Trondheim, NORWAY tlf: +47 90990249 email: einar.nass.jen...@ntnu.no ___ lustre-discuss mailing list lustre-discuss@lists.lustre.org http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org