Hi David, To follow up on this I had a 4th drive fail (out of 12) and have opted to order the below disks as a replacement, I have an ongoing case with Intel via the supplier - Will report back anything useful - But I am going to avoid the Intel s4600 2TB SSD's for the moment.
1.92TB Samsung SM863a 2.5" Enterprise SSD, SATA3 6Gb/s, 2-bit MLC V-NAND Regards Sean Redmond On Wed, Jan 10, 2018 at 11:08 PM, Sean Redmond <sean.redmo...@gmail.com> wrote: > Hi David, > > Thanks for your email, they are connected inside Dell R730XD (2.5 inch 24 > disk model) in None RAID mode via a perc RAID card. > > The version of ceph is Jewel with kernel 4.13.X and ubuntu 16.04. > > Thanks for your feedback on the HGST disks. > > Thanks > > On Wed, Jan 10, 2018 at 10:55 PM, David Herselman <d...@syrex.co> wrote: > >> Hi Sean, >> >> >> >> No, Intel’s feedback has been… Pathetic… I have yet to receive anything >> more than a request to ‘sign’ a non-disclosure agreement, to obtain beta >> firmware. No official answer as to whether or not one can logically unlock >> the drives, no answer to my question whether or not Intel publish serial >> numbers anywhere pertaining to recalled batches and no information >> pertaining to whether or not firmware updates would address any known >> issues. >> >> >> >> This with us being an accredited Intel Gold partner… >> >> >> >> >> >> We’ve returned the lot and ended up with 9/12 of the drives failing in >> the same manner. The replaced drives, which had different serial number >> ranges, also failed. Very frustrating is that the drives fail in a way that >> result in unbootable servers, unless one adds ‘rootdelay=240’ to the kernel. >> >> >> >> >> >> I would be interested to know what platform your drives were in and >> whether or not they were connected to a RAID module/card. >> >> >> >> PS: After much searching we’ve decided to order the NVMe conversion kit >> and have ordered HGST UltraStar SN200 2.5 inch SFF drives with a 3 DWPD >> rating. >> >> >> >> >> >> Regards >> >> David Herselman >> >> >> >> *From:* Sean Redmond [mailto:sean.redmo...@gmail.com] >> *Sent:* Thursday, 11 January 2018 12:45 AM >> *To:* David Herselman <d...@syrex.co> >> *Cc:* Christian Balzer <ch...@gol.com>; ceph-users@lists.ceph.com >> >> *Subject:* Re: [ceph-users] Many concurrent drive failures - How do I >> activate pgs? >> >> >> >> Hi, >> >> >> >> I have a case where 3 out to 12 of these Intel S4600 2TB model failed >> within a matter of days after being burn-in tested then placed into >> production. >> >> >> >> I am interested to know, did you every get any further feedback from the >> vendor on your issue? >> >> >> >> Thanks >> >> >> >> On Thu, Dec 21, 2017 at 1:38 PM, David Herselman <d...@syrex.co> wrote: >> >> Hi, >> >> I assume this can only be a physical manufacturing flaw or a firmware >> bug? Do Intel publish advisories on recalled equipment? Should others be >> concerned about using Intel DC S4600 SSD drives? Could this be an >> electrical issue on the Hot Swap Backplane or BMC firmware issue? Either >> way, all pure Intel... >> >> The hole is only 1.3 GB (4 MB x 339 objects) but perfectly striped >> through images, file systems are subsequently severely damaged. >> >> Is it possible to get Ceph to read in partial data shards? It would >> provide between 25-75% more yield... >> >> >> Is there anything wrong with how we've proceeded thus far? Would be nice >> to reference examples of using ceph-objectstore-tool but documentation is >> virtually non-existent. >> >> We used another SSD drive to simulate bringing all the SSDs back online. >> We carved up the drive to provide equal partitions to essentially simulate >> the original SSDs: >> # Partition a drive to provide 12 x 150GB partitions, eg: >> sdd 8:48 0 1.8T 0 disk >> |-sdd1 8:49 0 140G 0 part >> |-sdd2 8:50 0 140G 0 part >> |-sdd3 8:51 0 140G 0 part >> |-sdd4 8:52 0 140G 0 part >> |-sdd5 8:53 0 140G 0 part >> |-sdd6 8:54 0 140G 0 part >> |-sdd7 8:55 0 140G 0 part >> |-sdd8 8:56 0 140G 0 part >> |-sdd9 8:57 0 140G 0 part >> |-sdd10 8:58 0 140G 0 part >> |-sdd11 8:59 0 140G 0 part >> +-sdd12 8:60 0 140G 0 part >> >> >> Pre-requisites: >> ceph osd set noout; >> apt-get install uuid-runtime; >> >> >> for ID in `seq 24 35`; do >> UUID=`uuidgen`; >> OSD_SECRET=`ceph-authtool --gen-print-key`; >> DEVICE='/dev/sdd'$[$ID-23]; # 24-23 = /dev/sdd1, 35-23 = /dev/sdd12 >> echo "{\"cephx_secret\": \"$OSD_SECRET\"}" | ceph osd new $UUID $ID >> -i - -n client.bootstrap-osd -k /var/lib/ceph/bootstrap-osd/ceph.keyring; >> mkdir /var/lib/ceph/osd/ceph-$ID; >> mkfs.xfs $DEVICE; >> mount $DEVICE /var/lib/ceph/osd/ceph-$ID; >> ceph-authtool --create-keyring /var/lib/ceph/osd/ceph-$ID/keyring >> --name osd.$ID --add-key $OSD_SECRET; >> ceph-osd -i $ID --mkfs --osd-uuid $UUID; >> chown -R ceph:ceph /var/lib/ceph/osd/ceph-$ID; >> systemctl enable ceph-osd@$ID; >> systemctl start ceph-osd@$ID; >> done >> >> >> Once up we imported previous exports of empty head files in to 'real' >> OSDs: >> kvm5b: >> systemctl stop ceph-osd@8; >> ceph-objectstore-tool --op import --pgid 7.4s0 --data-path >> /var/lib/ceph/osd/ceph-8 --journal-path /var/lib/ceph/osd/ceph-8/journal >> --file /var/lib/vz/template/ssd_recovery/osd8_7.4s0.export; >> chown ceph:ceph -R /var/lib/ceph/osd/ceph-8; >> systemctl start ceph-osd@8; >> kvm5f: >> systemctl stop ceph-osd@23; >> ceph-objectstore-tool --op import --pgid 7.fs0 --data-path >> /var/lib/ceph/osd/ceph-23 --journal-path /var/lib/ceph/osd/ceph-23/journal >> --file /var/lib/vz/template/ssd_recovery/osd23_7.fs0.export; >> chown ceph:ceph -R /var/lib/ceph/osd/ceph-23; >> systemctl start ceph-osd@23; >> >> >> Bulk import previously exported objects: >> cd /var/lib/vz/template/ssd_recovery; >> for FILE in `ls -1A osd*_*.export | grep -Pv '^osd(8|23)_'`; do >> OSD=`echo $FILE | perl -pe 's/^osd(\d+).*/\1/'`; >> PGID=`echo $FILE | perl -pe 's/^osd\d+_(.*?).export/\1/g'`; >> echo -e "systemctl stop ceph-osd@$OSD\t ceph-objectstore-tool --op >> import --pgid $PGID --data-path /var/lib/ceph/osd/ceph-$OSD --journal-path >> /var/lib/ceph/osd/ceph-$OSD/journal --file /var/lib/vz/template/ssd_recov >> ery/osd"$OSD"_$PGID.export"; >> done | sort >> >> Sample output (this will wrap): >> systemctl stop ceph-osd@27 ceph-objectstore-tool --op import >> --pgid 7.4s3 --data-path /var/lib/ceph/osd/ceph-27 --journal-path >> /var/lib/ceph/osd/ceph-27/journal --file /var/lib/vz/template/ssd_recov >> ery/osd27_7.4s3.export >> systemctl stop ceph-osd@27 ceph-objectstore-tool --op import >> --pgid 7.fs5 --data-path /var/lib/ceph/osd/ceph-27 --journal-path >> /var/lib/ceph/osd/ceph-27/journal --file /var/lib/vz/template/ssd_recov >> ery/osd27_7.fs5.export >> systemctl stop ceph-osd@30 ceph-objectstore-tool --op import >> --pgid 7.fs4 --data-path /var/lib/ceph/osd/ceph-30 --journal-path >> /var/lib/ceph/osd/ceph-30/journal --file /var/lib/vz/template/ssd_recov >> ery/osd30_7.fs4.export >> systemctl stop ceph-osd@31 ceph-objectstore-tool --op import >> --pgid 7.4s2 --data-path /var/lib/ceph/osd/ceph-31 --journal-path >> /var/lib/ceph/osd/ceph-31/journal --file /var/lib/vz/template/ssd_recov >> ery/osd31_7.4s2.export >> systemctl stop ceph-osd@32 ceph-objectstore-tool --op import >> --pgid 7.4s4 --data-path /var/lib/ceph/osd/ceph-32 --journal-path >> /var/lib/ceph/osd/ceph-32/journal --file /var/lib/vz/template/ssd_recov >> ery/osd32_7.4s4.export >> systemctl stop ceph-osd@32 ceph-objectstore-tool --op import >> --pgid 7.fs2 --data-path /var/lib/ceph/osd/ceph-32 --journal-path >> /var/lib/ceph/osd/ceph-32/journal --file /var/lib/vz/template/ssd_recov >> ery/osd32_7.fs2.export >> systemctl stop ceph-osd@34 ceph-objectstore-tool --op import >> --pgid 7.4s5 --data-path /var/lib/ceph/osd/ceph-34 --journal-path >> /var/lib/ceph/osd/ceph-34/journal --file /var/lib/vz/template/ssd_recov >> ery/osd34_7.4s5.export >> systemctl stop ceph-osd@34 ceph-objectstore-tool --op import >> --pgid 7.fs1 --data-path /var/lib/ceph/osd/ceph-34 --journal-path >> /var/lib/ceph/osd/ceph-34/journal --file /var/lib/vz/template/ssd_recov >> ery/osd34_7.fs1.export >> >> >> Reset permissions and then started the OSDs: >> for OSD in 27 30 31 32 34; do >> chown -R ceph:ceph /var/lib/ceph/osd/ceph-$OSD; >> systemctl start ceph-osd@$OSD; >> done >> >> >> Then finally started all the OSDs... Now to hope that Intel have a way of >> accessing drives that are in a 'disable logical state'. >> >> >> >> The imports succeed, herewith a link to the output after running an >> import for placement group 7.4s2 on OSD 31: >> https://drive.google.com/open?id=1-Jo1jmrWrGLO2OgflacGPlEf2p32Y4hn >> >> Sample snippet: >> Write 1#7:fffcd2ec:::rbd_data.4.be8e9974b0dc51.0000000000002869:head# >> snapset 0=[]:{} >> Write 1#7:fffd4823:::rbd_data.4.ba24ef2ae8944a.000000000000a2b0:head# >> snapset 0=[]:{} >> Write 1#7:fffd6fb6:::benchmark_data_kvm5b_20945_object14722:head# >> snapset 0=[]:{} >> Write 1#7:ffffa069:::rbd_data.4.ba24ef2ae8944a.000000000000aea9:head# >> snapset 0=[]:{} >> Import successful >> >> >> Data does get written, I can tell by the size of the FileStore mount >> points: >> [root@kvm5b ssd_recovery]# df -h | grep -P 'ceph-(27|30|31|32|34)$' >> /dev/sdd4 140G 5.2G 135G 4% /var/lib/ceph/osd/ceph-27 >> /dev/sdd7 140G 14G 127G 10% /var/lib/ceph/osd/ceph-30 >> /dev/sdd8 140G 14G 127G 10% /var/lib/ceph/osd/ceph-31 >> /dev/sdd9 140G 22G 119G 16% /var/lib/ceph/osd/ceph-32 >> /dev/sdd11 140G 22G 119G 16% /var/lib/ceph/osd/ceph-34 >> >> >> How do I tell Ceph to read these object shards? >> >> >> >> PS: It's probably a good idea to reweight the OSDs to 0 before starting >> again. This should prevent data flowing on to them, if they are not in a >> different device class or other crush selection ruleset. Ie: >> for OSD in `seq 24 35`; do >> ceph osd crush reweight osd.$OSD 0; >> done >> >> >> Regards >> David Herselman >> >> -----Original Message----- >> >> From: David Herselman >> Sent: Thursday, 21 December 2017 3:49 AM >> To: 'Christian Balzer' <ch...@gol.com>; ceph-users@lists.ceph.com >> Subject: RE: [ceph-users] Many concurrent drive failures - How do I >> activate pgs? >> >> Hi Christian, >> >> Thanks for taking the time, I haven't been contacted by anyone yet but >> managed to get the down placement groups cleared by exporting 7.4s0 and >> 7.fs0 and then marking them as complete on the surviving OSDs: >> kvm5c: >> ceph-objectstore-tool --op export --pgid 7.4s0 --data-path >> /var/lib/ceph/osd/ceph-8 --journal-path /var/lib/ceph/osd/ceph-8/journal >> --file /var/lib/vz/template/ssd_recovery/osd8_7.4s0.export; >> ceph-objectstore-tool --op mark-complete --data-path >> /var/lib/ceph/osd/ceph-8 --journal-path /var/lib/ceph/osd/ceph-8/journal >> --pgid 7.4s0; >> kvm5f: >> ceph-objectstore-tool --op export --pgid 7.fs0 --data-path >> /var/lib/ceph/osd/ceph-23 --journal-path /var/lib/ceph/osd/ceph-23/journal >> --file /var/lib/vz/template/ssd_recovery/osd23_7.fs0.export; >> ceph-objectstore-tool --op mark-complete --data-path >> /var/lib/ceph/osd/ceph-23 --journal-path /var/lib/ceph/osd/ceph-23/journal >> --pgid 7.fs0; >> >> This would presumably simply punch holes in the RBD images but at least >> we can copy them out of that pool and hope that Intel can somehow unlock >> the drives for us to then export/import objects. >> >> >> To answer your questions though, we have 6 near identical Intel Wildcat >> Pass 1U servers and have Proxmox loaded on them. Proxmox uses a Debian 9 >> base with the Ubuntu kernel, for which they apply cherry picked kernel >> patches (eg Intel NIC driver updates, vhost perf regression and mem-leak >> fixes, etc): >> >> kvm5a: >> Intel R1208WTTGSR System (serial: BQWS55091014) >> Intel S2600WTTR Motherboard (serial: BQWL54950385, BIOS ID: >> SE5C610.86B.01.01.0021.032120170601) >> 2 x Intel Xeon E5-2640v4 2.4GHz (HT disabled) >> 24 x Micron 8GB DDR4 2133MHz (24 x 18ASF1G72PZ-2G1B1) >> Intel AXX10GBNIA I/O Module >> kvm5b: >> Intel R1208WTTGS System (serial: BQWS53890178) >> Intel S2600WTT Motherboard (serial: BQWL52550359, BIOS ID: >> SE5C610.86B.01.01.0021.032120170601) >> 2 x Intel Xeon E5-2640v4 2.4GHz (HT enabled) >> 4 x Micron 64GB DDR4 2400MHz LR-DIMM (4 x 72ASS8G72LZ-2G3B2) >> Intel AXX10GBNIA I/O Module >> kvm5c: >> Intel R1208WT2GS System (serial: BQWS50490279) >> Intel S2600WT2 Motherboard (serial: BQWL44650203, BIOS ID: >> SE5C610.86B.01.01.0021.032120170601) >> 2 x Intel Xeon E5-2640v3 2.6GHz (HT enabled) >> 4 x Micron 64GB DDR4 2400MHz LR-DIMM (4 x 72ASS8G72LZ-2G3B2) >> Intel AXX10GBNIA I/O Module >> kvm5d: >> Intel R1208WTTGSR System (serial: BQWS62291318) >> Intel S2600WTTR Motherboard (serial: BQWL61855187, BIOS ID: >> SE5C610.86B.01.01.0021.032120170601) >> 2 x Intel Xeon E5-2640v4 2.4GHz (HT enabled) >> 4 x Micron 64GB DDR4 2400MHz LR-DIMM (4 x 72ASS8G72LZ-2G3B2) >> Intel AXX10GBNIA I/O Module >> kvm5e: >> Intel R1208WTTGSR System (serial: BQWS64290162) >> Intel S2600WTTR Motherboard (serial: BQWL63953066, BIOS ID: >> SE5C610.86B.01.01.0021.032120170601) >> 2 x Intel Xeon E5-2640v4 2.4GHz (HT enabled) >> 4 x Micron 64GB DDR4 2400MHz LR-DIMM (4 x 72ASS8G72LZ-2G3B2) >> Intel AXX10GBNIA I/O Module >> kvm5f: >> Intel R1208WTTGSR System (serial: BQWS71790632) >> Intel S2600WTTR Motherboard (serial: BQWL71050622, BIOS ID: >> SE5C610.86B.01.01.0021.032120170601) >> 2 x Intel Xeon E5-2640v4 2.4GHz (HT enabled) >> 4 x Micron 64GB DDR4 2400MHz LR-DIMM (4 x 72ASS8G72LZ-2G3B2) >> Intel AXX10GBNIA I/O Module >> >> Summary: >> * 5b has an Intel S2600WTT, 5c has an Intel S2600WT2, all others have >> S2600WTTR Motherboards >> * 5a has ECC Registered Dual Rank DDR DIMMs, all others have ECC >> LoadReduced-DIMMs >> * 5c has an Intel X540-AT2 10 GbE adapter as the on-board NICs are only >> 1 GbE >> >> >> Each system has identical discs: >> * 2 x 480 GB Intel SSD DC S3610 (SSDSC2BX480G4) - partitioned as >> software RAID1 OS volume and Ceph FileStore journals (spinners) >> * 4 x 2 TB Seagate discs (ST2000NX0243) - Ceph FileStore OSDs (journals >> in S3610 partitions) >> * 2 x 1.9 TB Intel SSD DC S4600 (SSDSC2KG019T7) - Ceph BlueStore OSDs >> (problematic) >> >> >> Additional information: >> * All drives are directly attached to the on-board AHCI SATA >> controllers, via the standard 2.5 inch drive chassis hot-swap bays. >> * We added 12 x 1.9 TB SSD DC S4600 drives last week Thursday, 2 in >> each system's slots 7 & 8 >> * Systems have been operating with existing Intel SSD DC S3610 and 2 TB >> Seagate discs for over a year; we added the most recent node (kvm5f) on the >> 23rd of November. >> * 6 of the 12 Intel SSD DC S4600 drives failed in less than 100 hours. >> * They work perfectly until they suddenly stop responding and are >> thereafter, even with us physically shutting down the server and powering >> it back up again, completely inaccessible. Intel diagnostic tool reports >> 'logically locked'. >> >> >> Drive failures appear random to me: >> kvm5a - bay 7 - offline - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, >> SerialNo=BTYM739208851P9DGN >> kvm5a - bay 8 - online - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, >> SerialNo=PHYM727602TM1P9DGN >> kvm5b - bay 7 - offline - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, >> SerialNo=PHYM7276031E1P9DGN >> kvm5b - bay 8 - online - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, >> SerialNo=BTYM7392087W1P9DGN >> kvm5c - bay 7 - offline - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, >> SerialNo=BTYM739200ZJ1P9DGN >> kvm5c - bay 8 - offline - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, >> SerialNo=BTYM7392088B1P9DGN >> kvm5d - bay 7 - offline - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, >> SerialNo=BTYM738604Y11P9DGN >> kvm5d - bay 8 - online - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, >> SerialNo=PHYM727603181P9DGN >> kvm5e - bay 7 - online - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, >> SerialNo=BTYM7392013B1P9DGN >> kvm5e - bay 8 - offline - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, >> SerialNo=BTYM7392087E1P9DGN >> kvm5f - bay 7 - online - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, >> SerialNo=BTYM739208721P9DGN >> kvm5f - bay 8 - online - Model=INTEL SSDSC2KG019T7, FwRev=SCV10100, >> SerialNo=BTYM739208C41P9DGN >> >> >> Intel SSD Data Center Tool reports: >> C:\isdct>isdct.exe show -intelssd >> >> - Intel SSD DC S4600 Series PHYM7276031E1P9DGN - >> >> Bootloader : Property not found >> DevicePath : \\\\.\\PHYSICALDRIVE1 >> DeviceStatus : Selected drive is in a disable logical state. >> Firmware : SCV10100 >> FirmwareUpdateAvailable : Please contact Intel Customer Support for >> further assistance at the following website: >> http://www.intel.com/go/ssdsupport. >> Index : 0 >> ModelNumber : INTEL SSDSC2KG019T7 >> ProductFamily : Intel SSD DC S4600 Series SerialNumber : >> PHYM7276031E1P9DGN >> >> >> C:\isdct>isdct show -a -intelssd 0 >> >> - Intel SSD DC S4600 Series PHYM7276031E1P9DGN - >> >> AccessibleMaxAddressSupported : True >> AggregationThreshold : Selected drive is in a disable logical state. >> AggregationTime : Selected drive is in a disable logical state. >> ArbitrationBurst : Selected drive is in a disable logical state. >> BusType : 11 >> CoalescingDisable : Selected drive is in a disable logical state. >> ControllerCompatibleIDs : PCI\\VEN_8086&DEV_8C02&REV_05P >> CI\\VEN_8086&DEV_8C02PCI\\VEN_8086&CC_010601PCI\\VEN_8086& >> CC_0106PCI\\VEN_8086PCI\\CC_010601PCI\\CC_0106 >> ControllerDescription : @mshdc.inf,%pci\\cc_010601.devicedesc%;Standard >> SATA AHCI Controller ControllerID : PCI\\VEN_8086&DEV_8C02&SUBSYS_ >> 78461462&REV_05\\3&11583659&0&FA >> ControllerIDEMode : False >> ControllerManufacturer : @mshdc.inf,%ms-ahci%;Standard SATA AHCI >> Controller ControllerService : storahci DIPMEnabled : False DIPMSupported : >> False DevicePath : \\\\.\\PHYSICALDRIVE1 DeviceStatus : Selected drive >> is in a disable logical state. >> DigitalFenceSupported : False >> DownloadMicrocodePossible : True >> DriverDescription : Standard SATA AHCI Controller DriverMajorVersion : 10 >> DriverManufacturer : Standard SATA AHCI Controller DriverMinorVersion : 0 >> DriverVersion : 10.0.16299.98 DynamicMMIOEnabled : The selected drive does >> not support this feature. >> EnduranceAnalyzer : Selected drive is in a disable logical state. >> ErrorString : *BAD_CONTEXT_2020 F4 >> Firmware : SCV10100 >> FirmwareUpdateAvailable : Please contact Intel Customer Support for >> further assistance at the following website: >> http://www.intel.com/go/ssdsupport. >> HDD : False >> HighPriorityWeightArbitration : Selected drive is in a disable logical >> state. >> IEEE1667Supported : False >> IOCompletionQueuesRequested : Selected drive is in a disable logical >> state. >> IOSubmissionQueuesRequested : Selected drive is in a disable logical >> state. >> Index : 0 >> Intel : True >> IntelGen3SATA : True >> IntelNVMe : False >> InterruptVector : Selected drive is in a disable logical state. >> IsDualPort : False >> LatencyTrackingEnabled : Selected drive is in a disable logical state. >> LowPriorityWeightArbitration : Selected drive is in a disable logical >> state. >> Lun : 0 >> MaximumLBA : 3750748847 >> MediumPriorityWeightArbitration : Selected drive is in a disable logical >> state. >> ModelNumber : INTEL SSDSC2KG019T7 >> NVMePowerState : Selected drive is in a disable logical state. >> NativeMaxLBA : Selected drive is in a disable logical state. >> OEM : Generic >> OpalState : Selected drive is in a disable logical state. >> PLITestTimeInterval : Selected drive is in a disable logical state. >> PNPString : SCSI\\DISK&VEN_INTEL&PROD_SSDSC2KG019T7\\4&2BE6C224&0&010000 >> PathID : 1 >> PhySpeed : Selected drive is in a disable logical state. >> PhysicalSectorSize : Selected drive is in a disable logical state. >> PhysicalSize : 1920383410176 >> PowerGovernorAveragePower : Selected drive is in a disable logical state. >> PowerGovernorBurstPower : Selected drive is in a disable logical state. >> PowerGovernorMode : Selected drive is in a disable logical state. >> Product : Youngsville >> ProductFamily : Intel SSD DC S4600 Series ProductProtocol : ATA >> ReadErrorRecoveryTimer : Selected drive is in a disable logical state. >> RemoteSecureEraseSupported : False >> SCSIPortNumber : 0 >> SMARTEnabled : True >> SMARTHealthCriticalWarningsConfiguration : Selected drive is in a >> disable logical state. >> SMARTSelfTestSupported : True >> SMBusAddress : Selected drive is in a disable logical state. >> SSCEnabled : False >> SanitizeBlockEraseSupported : False >> SanitizeCryptoScrambleSupported : True >> SanitizeSupported : True >> SataGen1 : True >> SataGen2 : True >> SataGen3 : True >> SataNegotiatedSpeed : Unknown >> SectorSize : 512 >> SecurityEnabled : False >> SecurityFrozen : False >> SecurityLocked : False >> SecuritySupported : False >> SerialNumber : PHYM7276031E1P9DGN >> TCGSupported : False >> TargetID : 0 >> TempThreshold : Selected drive is in a disable logical state. >> TemperatureLoggingInterval : Selected drive is in a disable logical state. >> TimeLimitedErrorRecovery : Selected drive is in a disable logical state. >> TrimSize : 4 >> TrimSupported : True >> VolatileWriteCacheEnabled : Selected drive is in a disable logical state. >> WWID : 3959312879584368077 >> WriteAtomicityDisableNormal : Selected drive is in a disable logical >> state. >> WriteCacheEnabled : True >> WriteCacheReorderingStateEnabled : Selected drive is in a disable >> logical state. >> WriteCacheState : Selected drive is in a disable logical state. >> WriteCacheSupported : True >> WriteErrorRecoveryTimer : Selected drive is in a disable logical state. >> >> >> >> SMART information is inaccessible, overall status is failed. Herewith the >> stats from a partner disc which was still working when the others failed: >> Device Model: INTEL SSDSC2KG019T7 >> Serial Number: PHYM727602TM1P9DGN >> LU WWN Device Id: 5 5cd2e4 14e1636bb >> Firmware Version: SCV10100 >> User Capacity: 1,920,383,410,176 bytes [1.92 TB] >> Sector Sizes: 512 bytes logical, 4096 bytes physical >> Rotation Rate: Solid State Device >> Form Factor: 2.5 inches >> Device is: Not in smartctl database [for details use: -P showall] >> ATA Version is: ACS-3 T13/2161-D revision 5 >> SATA Version is: SATA 3.2, 6.0 Gb/s (current: 6.0 Gb/s) >> Local Time is: Mon Dec 18 19:33:51 2017 SAST >> SMART support is: Available - device has SMART capability. >> SMART support is: Enabled >> >> ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE >> UPDATED WHEN_FAILED RAW_VALUE >> 5 Reallocated_Sector_Ct 0x0032 100 100 000 Old_age Always >> - 0 >> 9 Power_On_Hours 0x0032 100 100 000 Old_age Always >> - 98 >> 12 Power_Cycle_Count 0x0032 100 100 000 Old_age Always >> - 3 >> 170 Unknown_Attribute 0x0033 100 100 010 Pre-fail Always >> - 0 >> 171 Unknown_Attribute 0x0032 100 100 000 Old_age Always >> - 1 >> 172 Unknown_Attribute 0x0032 100 100 000 Old_age Always >> - 0 >> 174 Unknown_Attribute 0x0032 100 100 000 Old_age Always >> - 0 >> 175 Program_Fail_Count_Chip 0x0033 100 100 010 Pre-fail Always >> - 17567121432 >> 183 Runtime_Bad_Block 0x0032 100 100 000 Old_age Always >> - 0 >> 184 End-to-End_Error 0x0033 100 100 090 Pre-fail Always >> - 0 >> 187 Reported_Uncorrect 0x0032 100 100 000 Old_age Always >> - 0 >> 190 Airflow_Temperature_Cel 0x0022 077 076 000 Old_age Always >> - 23 (Min/Max 17/29) >> 192 Power-Off_Retract_Count 0x0032 100 100 000 Old_age Always >> - 0 >> 194 Temperature_Celsius 0x0022 100 100 000 Old_age Always >> - 23 >> 197 Current_Pending_Sector 0x0012 100 100 000 Old_age Always >> - 0 >> 199 UDMA_CRC_Error_Count 0x003e 100 100 000 Old_age Always >> - 0 >> 225 Unknown_SSD_Attribute 0x0032 100 100 000 Old_age Always >> - 14195 >> 226 Unknown_SSD_Attribute 0x0032 100 100 000 Old_age Always >> - 0 >> 227 Unknown_SSD_Attribute 0x0032 100 100 000 Old_age Always >> - 42 >> 228 Power-off_Retract_Count 0x0032 100 100 000 Old_age Always >> - 5905 >> 232 Available_Reservd_Space 0x0033 100 100 010 Pre-fail Always >> - 0 >> 233 Media_Wearout_Indicator 0x0032 100 100 000 Old_age Always >> - 0 >> 234 Unknown_Attribute 0x0032 100 100 000 Old_age Always >> - 0 >> 241 Total_LBAs_Written 0x0032 100 100 000 Old_age Always >> - 14195 >> 242 Total_LBAs_Read 0x0032 100 100 000 Old_age Always >> - 10422 >> 243 Unknown_Attribute 0x0032 100 100 000 Old_age Always >> - 41906 >> >> >> Media wear out : 0% used >> LBAs written: 14195 >> Power on hours: <100 >> Power cycle count: once at the factory, once at our offices to check if >> there was newer firmware (there wasn't) and once when we restarted the node >> to see if it could then access a failed drive. >> >> >> Regards >> David Herselman >> >> >> -----Original Message----- >> From: Christian Balzer [mailto:ch...@gol.com] >> Sent: Thursday, 21 December 2017 3:24 AM >> To: ceph-users@lists.ceph.com >> Cc: David Herselman <d...@syrex.co> >> Subject: Re: [ceph-users] Many concurrent drive failures - How do I >> activate pgs? >> >> Hello, >> >> first off, I don't have anything to add to your conclusions of the >> current status, alas there are at least 2 folks here on the ML making a >> living from Ceph disaster recovery, so I hope you have been contacted >> already. >> >> Now once your data is safe or you have a moment, I and others here would >> probably be quite interested in some more details, see inline below. >> >> On Wed, 20 Dec 2017 22:25:23 +0000 David Herselman wrote: >> >> [snip] >> > >> > We've happily been running a 6 node cluster with 4 x FileStore HDDs per >> node (journals on SSD partitions) for over a year and recently upgraded all >> nodes to Debian 9, Ceph Luminous 12.2.2 and kernel 4.13.8. We ordered 12 x >> Intel DC S4600 SSDs which arrived last week so we added two per node on >> Thursday evening and brought them up as BlueStore OSDs. We had proactively >> updated our existing pools to reference only devices classed as 'hdd', so >> that we could move select images over to ssd replicated and erasure coded >> pools. >> > >> Could you tell us more about that cluster, as in HW, how are the SSDs >> connected and FW version of the controller if applicable. >> >> Kernel 4.13.8 suggests that this is a handrolled, upstream kernel. >> While not necessarily related I'll note that as far as Debian kernels >> (which are very lightly if at all patched) are concerned, nothing beyond >> 4.9 has been working to my satisfaction. >> 4.11 still worked, but 4.12 crash-reboot-looped on all my Supermicro X10 >> machines (quite a varied selection). >> The current 4.13.13 backport boots on some of those machines, but still >> throws errors with the EDAC devices, which works fine with 4.9. >> >> 4.14 is known to happily destroy data if used with bcache and even if one >> doesn't use that it should give you pause. >> >> > We were pretty diligent and downloaded Intel's Firmware Update Tool and >> validated that each new drive had the latest available firmware before >> installing them in the nodes. We did numerous benchmarks on Friday and >> eventually moved some images over to the new storage pools. Everything was >> working perfectly and extensive tests on Sunday showed excellent >> performance. Sunday night one of the new SSDs died and Ceph replicated and >> redistributed data accordingly, then another failed in the early hours of >> Monday morning and Ceph did what it needed to. >> > >> > We had the two failed drives replaced by 11am and Ceph was up to >> 2/4918587 objects degraded (0.000%) when a third drive failed. At this >> point we updated the crush maps for the rbd_ssd and ec_ssd pools and set >> the device class to 'hdd', to essentially evacuate everything off the SSDs. >> Other SSDs then failed at 3:22pm, 4:19pm, 5:49pm and 5:50pm. We've >> ultimately lost half the Intel S4600 drives, which are all completely >> inaccessible. Our status at 11:42pm Monday night was: 1/1398478 objects >> unfound (0.000%) and 339/4633062 objects degraded (0.007%). >> > >> The relevant logs when and how those SSDs failed would be interesting. >> Was the distribution of the failed SSDs random among the cluster? >> Are you running smartd and did it have something to say? >> >> Completely inaccessible sounds a lot like the infamous "self-bricking" of >> Intel SSDs when they discover something isn't right, or they don't like the >> color scheme of the server inside (^.^). >> >> I'm using quite a lot of Intel SSDs and had only one "fatal" incident. >> A DC S3700 detected that its powercap had failed, but of course kept >> working fine. Until a reboot was need, when it promptly bricked itself, >> data inaccessible, SMART reporting barely that something was there. >> >> So one wonders what caused your SSDs to get their knickers in such a >> twist. >> Are the survivors showing any unusual signs in their SMART output? >> >> Of course what your vendor/Intel will have to say will also be of >> interest. ^o^ >> >> Regards, >> >> Christian >> -- >> Christian Balzer Network/Systems Engineer >> ch...@gol.com Rakuten Communications >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> >> > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com