On Tue, Mar 6, 2018 at 5:26 PM, Marco Baldini - H.S. Amiata <
mbald...@hsamiata.it> wrote:

> Hi
>
> I monitor dmesg in each of the 3 nodes, no hardware issue reported. And
> the problem happens with various different OSDs in different nodes, for me
> it is clear it's not an hardware problem.
>

If you have osd_debug set to 25 or greater when you run the deep scrub you
should get more information about the nature of the read error in the
ReplicatedBackend::be_deep_scrub() function (assuming this is a replicated
pool).

This may create large logs so watch they don't exhaust storage.

> Thanks for reply
>
>
>
> Il 05/03/2018 21:45, Vladimir Prokofev ha scritto:
>
> > always solved by ceph pg repair <PG>
> That doesn't necessarily means that there's no hardware issue. In my case
> repair also worked fine and returned cluster to OK state every time, but in
> time faulty disk fail another scrub operation, and this repeated multiple
> times before we replaced that disk.
> One last thing to look into is dmesg at your OSD nodes. If there's a
> hardware read error it will be logged in dmesg.
>
> 2018-03-05 18:26 GMT+03:00 Marco Baldini - H.S. Amiata <
> mbald...@hsamiata.it>:
>
>> Hi and thanks for reply
>>
>> The OSDs are all healthy, in fact after a ceph pg repair <PG> the ceph
>> health is back to OK and in the OSD log I see  <PG> repair ok, 0 fixed
>>
>> The SMART data of the 3 OSDs seems fine
>>
>> *OSD.5*
>>
>> # ceph-disk list | grep osd.5
>>  /dev/sdd1 ceph data, active, cluster ceph, osd.5, block /dev/sdd2
>>
>> # smartctl -a /dev/sdd
>> smartctl 6.6 2016-05-31 r4324 [x86_64-linux-4.13.13-6-pve] (local build)
>> Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
>>
>> === START OF INFORMATION SECTION ===
>> Model Family:     Seagate Barracuda 7200.14 (AF)
>> Device Model:     ST1000DM003-1SB10C
>> Serial Number:    Z9A1MA1V
>> LU WWN Device Id: 5 000c50 090c7028b
>> Firmware Version: CC43
>> User Capacity:    1,000,204,886,016 bytes [1.00 TB]
>> Sector Sizes:     512 bytes logical, 4096 bytes physical
>> Rotation Rate:    7200 rpm
>> Form Factor:      3.5 inches
>> Device is:        In smartctl database [for details use: -P show]
>> ATA Version is:   ATA8-ACS T13/1699-D revision 4
>> SATA Version is:  SATA 3.0, 6.0 Gb/s (current: 6.0 Gb/s)
>> Local Time is:    Mon Mar  5 16:17:22 2018 CET
>> SMART support is: Available - device has SMART capability.
>> SMART support is: Enabled
>>
>> === START OF READ SMART DATA SECTION ===
>> SMART overall-health self-assessment test result: PASSED
>>
>> General SMART Values:
>> Offline data collection status:  (0x82)      Offline data collection activity
>>                                      was completed without error.
>>                                      Auto Offline Data Collection: Enabled.
>> Self-test execution status:      (   0)      The previous self-test routine 
>> completed
>>                                      without error or no self-test has ever
>>                                      been run.
>> Total time to complete Offline
>> data collection:             (    0) seconds.
>> Offline data collection
>> capabilities:                         (0x7b) SMART execute Offline immediate.
>>                                      Auto Offline data collection on/off 
>> support.
>>                                      Suspend Offline collection upon new
>>                                      command.
>>                                      Offline surface scan supported.
>>                                      Self-test supported.
>>                                      Conveyance Self-test supported.
>>                                      Selective Self-test supported.
>> SMART capabilities:            (0x0003)      Saves SMART data before entering
>>                                      power-saving mode.
>>                                      Supports SMART auto save timer.
>> Error logging capability:        (0x01)      Error logging supported.
>>                                      General Purpose Logging supported.
>> Short self-test routine
>> recommended polling time:     (   1) minutes.
>> Extended self-test routine
>> recommended polling time:     ( 109) minutes.
>> Conveyance self-test routine
>> recommended polling time:     (   2) minutes.
>> SCT capabilities:           (0x1085) SCT Status supported.
>>
>> SMART Attributes Data Structure revision number: 10
>> Vendor Specific SMART Attributes with Thresholds:
>> ID# ATTRIBUTE_NAME          FLAG     VALUE WORST THRESH TYPE      UPDATED  
>> WHEN_FAILED RAW_VALUE
>>   1 Raw_Read_Error_Rate     0x000f   082   063   006    Pre-fail  Always     
>>   -       193297722
>>   3 Spin_Up_Time            0x0003   097   097   000    Pre-fail  Always     
>>   -       0
>>   4 Start_Stop_Count        0x0032   100   100   020    Old_age   Always     
>>   -       60
>>   5 Reallocated_Sector_Ct   0x0033   100   100   010    Pre-fail  Always     
>>   -       0
>>   7 Seek_Error_Rate         0x000f   091   060   045    Pre-fail  Always     
>>   -       1451132477
>>   9 Power_On_Hours          0x0032   085   085   000    Old_age   Always     
>>   -       13283
>>  10 Spin_Retry_Count        0x0013   100   100   097    Pre-fail  Always     
>>   -       0
>>  12 Power_Cycle_Count       0x0032   100   100   020    Old_age   Always     
>>   -       61
>> 183 Runtime_Bad_Block       0x0032   100   100   000    Old_age   Always     
>>   -       0
>> 184 End-to-End_Error        0x0032   100   100   099    Old_age   Always     
>>   -       0
>> 187 Reported_Uncorrect      0x0032   100   100   000    Old_age   Always     
>>   -       0
>> 188 Command_Timeout         0x0032   100   100   000    Old_age   Always     
>>   -       0 0 0
>> 189 High_Fly_Writes         0x003a   086   086   000    Old_age   Always     
>>   -       14
>> 190 Airflow_Temperature_Cel 0x0022   071   055   040    Old_age   Always     
>>   -       29 (Min/Max 23/32)
>> 193 Load_Cycle_Count        0x0032   100   100   000    Old_age   Always     
>>   -       607
>> 194 Temperature_Celsius     0x0022   029   014   000    Old_age   Always     
>>   -       29 (0 14 0 0 0)
>> 195 Hardware_ECC_Recovered  0x001a   004   001   000    Old_age   Always     
>>   -       193297722
>> 197 Current_Pending_Sector  0x0012   100   100   000    Old_age   Always     
>>   -       0
>> 198 Offline_Uncorrectable   0x0010   100   100   000    Old_age   Offline    
>>   -       0
>> 199 UDMA_CRC_Error_Count    0x003e   200   200   000    Old_age   Always     
>>   -       0
>> 240 Head_Flying_Hours       0x0000   100   253   000    Old_age   Offline    
>>   -       13211h+23m+08.363s
>> 241 Total_LBAs_Written      0x0000   100   253   000    Old_age   Offline    
>>   -       53042120064
>> 242 Total_LBAs_Read         0x0000   100   253   000    Old_age   Offline    
>>   -       170788993187
>>
>> *OSD.4*
>>
>> # ceph-disk list | grep osd.4
>>  /dev/sdc1 ceph data, active, cluster ceph, osd.4, block /dev/sdc2
>>
>> # smartctl -a /dev/sdc
>> smartctl 6.6 2016-05-31 r4324 [x86_64-linux-4.13.13-6-pve] (local build)
>> Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
>>
>> === START OF INFORMATION SECTION ===
>> Model Family:     Seagate Barracuda 7200.14 (AF)
>> Device Model:     ST1000DM003-1SB10C
>> Serial Number:    Z9A1M1BW
>> LU WWN Device Id: 5 000c50 090c78d27
>> Firmware Version: CC43
>> User Capacity:    1,000,204,886,016 bytes [1.00 TB]
>> Sector Sizes:     512 bytes logical, 4096 bytes physical
>> Rotation Rate:    7200 rpm
>> Form Factor:      3.5 inches
>> Device is:        In smartctl database [for details use: -P show]
>> ATA Version is:   ATA8-ACS T13/1699-D revision 4
>> SATA Version is:  SATA 3.0, 6.0 Gb/s (current: 6.0 Gb/s)
>> Local Time is:    Mon Mar  5 16:20:46 2018 CET
>> SMART support is: Available - device has SMART capability.
>> SMART support is: Enabled
>>
>> === START OF READ SMART DATA SECTION ===
>> SMART overall-health self-assessment test result: PASSED
>>
>> General SMART Values:
>> Offline data collection status:  (0x82)      Offline data collection activity
>>                                      was completed without error.
>>                                      Auto Offline Data Collection: Enabled.
>> Self-test execution status:      (   0)      The previous self-test routine 
>> completed
>>                                      without error or no self-test has ever
>>                                      been run.
>> Total time to complete Offline
>> data collection:             (    0) seconds.
>> Offline data collection
>> capabilities:                         (0x7b) SMART execute Offline immediate.
>>                                      Auto Offline data collection on/off 
>> support.
>>                                      Suspend Offline collection upon new
>>                                      command.
>>                                      Offline surface scan supported.
>>                                      Self-test supported.
>>                                      Conveyance Self-test supported.
>>                                      Selective Self-test supported.
>> SMART capabilities:            (0x0003)      Saves SMART data before entering
>>                                      power-saving mode.
>>                                      Supports SMART auto save timer.
>> Error logging capability:        (0x01)      Error logging supported.
>>                                      General Purpose Logging supported.
>> Short self-test routine
>> recommended polling time:     (   1) minutes.
>> Extended self-test routine
>> recommended polling time:     ( 109) minutes.
>> Conveyance self-test routine
>> recommended polling time:     (   2) minutes.
>> SCT capabilities:           (0x1085) SCT Status supported.
>>
>> SMART Attributes Data Structure revision number: 10
>> Vendor Specific SMART Attributes with Thresholds:
>> ID# ATTRIBUTE_NAME          FLAG     VALUE WORST THRESH TYPE      UPDATED  
>> WHEN_FAILED RAW_VALUE
>>   1 Raw_Read_Error_Rate     0x000f   082   063   006    Pre-fail  Always     
>>   -       194906537
>>   3 Spin_Up_Time            0x0003   097   097   000    Pre-fail  Always     
>>   -       0
>>   4 Start_Stop_Count        0x0032   100   100   020    Old_age   Always     
>>   -       64
>>   5 Reallocated_Sector_Ct   0x0033   100   100   010    Pre-fail  Always     
>>   -       0
>>   7 Seek_Error_Rate         0x000f   091   060   045    Pre-fail  Always     
>>   -       1485899434
>>   9 Power_On_Hours          0x0032   085   085   000    Old_age   Always     
>>   -       13390
>>  10 Spin_Retry_Count        0x0013   100   100   097    Pre-fail  Always     
>>   -       0
>>  12 Power_Cycle_Count       0x0032   100   100   020    Old_age   Always     
>>   -       65
>> 183 Runtime_Bad_Block       0x0032   100   100   000    Old_age   Always     
>>   -       0
>> 184 End-to-End_Error        0x0032   100   100   099    Old_age   Always     
>>   -       0
>> 187 Reported_Uncorrect      0x0032   100   100   000    Old_age   Always     
>>   -       0
>> 188 Command_Timeout         0x0032   100   100   000    Old_age   Always     
>>   -       0 0 0
>> 189 High_Fly_Writes         0x003a   095   095   000    Old_age   Always     
>>   -       5
>> 190 Airflow_Temperature_Cel 0x0022   074   051   040    Old_age   Always     
>>   -       26 (Min/Max 19/29)
>> 193 Load_Cycle_Count        0x0032   100   100   000    Old_age   Always     
>>   -       616
>> 194 Temperature_Celsius     0x0022   026   014   000    Old_age   Always     
>>   -       26 (0 14 0 0 0)
>> 195 Hardware_ECC_Recovered  0x001a   004   001   000    Old_age   Always     
>>   -       194906537
>> 197 Current_Pending_Sector  0x0012   100   100   000    Old_age   Always     
>>   -       0
>> 198 Offline_Uncorrectable   0x0010   100   100   000    Old_age   Offline    
>>   -       0
>> 199 UDMA_CRC_Error_Count    0x003e   200   200   000    Old_age   Always     
>>   -       0
>> 240 Head_Flying_Hours       0x0000   100   253   000    Old_age   Offline    
>>   -       13315h+20m+30.974s
>> 241 Total_LBAs_Written      0x0000   100   253   000    Old_age   Offline    
>>   -       52137467719
>> 242 Total_LBAs_Read         0x0000   100   253   000    Old_age   Offline    
>>   -       177227508503
>>
>>
>> *OSD.8*
>>
>> # ceph-disk list | grep osd.8
>>  /dev/sda1 ceph data, active, cluster ceph, osd.8, block /dev/sda2
>>
>> # smartctl -a /dev/sda
>> smartctl 6.6 2016-05-31 r4324 [x86_64-linux-4.13.13-6-pve] (local build)
>> Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
>>
>> === START OF INFORMATION SECTION ===
>> Model Family:     Seagate Barracuda 7200.14 (AF)
>> Device Model:     ST1000DM003-1SB10C
>> Serial Number:    Z9A2BEF2
>> LU WWN Device Id: 5 000c50 0910f5427
>> Firmware Version: CC43
>> User Capacity:    1,000,203,804,160 bytes [1.00 TB]
>> Sector Sizes:     512 bytes logical, 4096 bytes physical
>> Rotation Rate:    7200 rpm
>> Form Factor:      3.5 inches
>> Device is:        In smartctl database [for details use: -P show]
>> ATA Version is:   ATA8-ACS T13/1699-D revision 4
>> SATA Version is:  SATA 3.0, 6.0 Gb/s (current: 6.0 Gb/s)
>> Local Time is:    Mon Mar  5 16:22:47 2018 CET
>> SMART support is: Available - device has SMART capability.
>> SMART support is: Enabled
>>
>> === START OF READ SMART DATA SECTION ===
>> SMART overall-health self-assessment test result: PASSED
>>
>> General SMART Values:
>> Offline data collection status:  (0x82)      Offline data collection activity
>>                                      was completed without error.
>>                                      Auto Offline Data Collection: Enabled.
>> Self-test execution status:      (   0)      The previous self-test routine 
>> completed
>>                                      without error or no self-test has ever
>>                                      been run.
>> Total time to complete Offline
>> data collection:             (    0) seconds.
>> Offline data collection
>> capabilities:                         (0x7b) SMART execute Offline immediate.
>>                                      Auto Offline data collection on/off 
>> support.
>>                                      Suspend Offline collection upon new
>>                                      command.
>>                                      Offline surface scan supported.
>>                                      Self-test supported.
>>                                      Conveyance Self-test supported.
>>                                      Selective Self-test supported.
>> SMART capabilities:            (0x0003)      Saves SMART data before entering
>>                                      power-saving mode.
>>                                      Supports SMART auto save timer.
>> Error logging capability:        (0x01)      Error logging supported.
>>                                      General Purpose Logging supported.
>> Short self-test routine
>> recommended polling time:     (   1) minutes.
>> Extended self-test routine
>> recommended polling time:     ( 110) minutes.
>> Conveyance self-test routine
>> recommended polling time:     (   2) minutes.
>> SCT capabilities:           (0x1085) SCT Status supported.
>>
>> SMART Attributes Data Structure revision number: 10
>> Vendor Specific SMART Attributes with Thresholds:
>> ID# ATTRIBUTE_NAME          FLAG     VALUE WORST THRESH TYPE      UPDATED  
>> WHEN_FAILED RAW_VALUE
>>   1 Raw_Read_Error_Rate     0x000f   083   063   006    Pre-fail  Always     
>>   -       224621855
>>   3 Spin_Up_Time            0x0003   097   097   000    Pre-fail  Always     
>>   -       0
>>   4 Start_Stop_Count        0x0032   100   100   020    Old_age   Always     
>>   -       275
>>   5 Reallocated_Sector_Ct   0x0033   100   100   010    Pre-fail  Always     
>>   -       0
>>   7 Seek_Error_Rate         0x000f   081   060   045    Pre-fail  Always     
>>   -       149383284
>>   9 Power_On_Hours          0x0032   093   093   000    Old_age   Always     
>>   -       6210
>>  10 Spin_Retry_Count        0x0013   100   100   097    Pre-fail  Always     
>>   -       0
>>  12 Power_Cycle_Count       0x0032   100   100   020    Old_age   Always     
>>   -       265
>> 183 Runtime_Bad_Block       0x0032   100   100   000    Old_age   Always     
>>   -       0
>> 184 End-to-End_Error        0x0032   100   100   099    Old_age   Always     
>>   -       0
>> 187 Reported_Uncorrect      0x0032   100   100   000    Old_age   Always     
>>   -       0
>> 188 Command_Timeout         0x0032   100   100   000    Old_age   Always     
>>   -       0 0 0
>> 189 High_Fly_Writes         0x003a   098   098   000    Old_age   Always     
>>   -       2
>> 190 Airflow_Temperature_Cel 0x0022   069   058   040    Old_age   Always     
>>   -       31 (Min/Max 21/35)
>> 193 Load_Cycle_Count        0x0032   100   100   000    Old_age   Always     
>>   -       516
>> 194 Temperature_Celsius     0x0022   031   017   000    Old_age   Always     
>>   -       31 (0 17 0 0 0)
>> 195 Hardware_ECC_Recovered  0x001a   005   001   000    Old_age   Always     
>>   -       224621855
>> 197 Current_Pending_Sector  0x0012   100   100   000    Old_age   Always     
>>   -       0
>> 198 Offline_Uncorrectable   0x0010   100   100   000    Old_age   Offline    
>>   -       0
>> 199 UDMA_CRC_Error_Count    0x003e   200   200   000    Old_age   Always     
>>   -       0
>> 240 Head_Flying_Hours       0x0000   100   253   000    Old_age   Offline    
>>   -       6154h+03m+35.126s
>> 241 Total_LBAs_Written      0x0000   100   253   000    Old_age   Offline    
>>   -       24333847321
>> 242 Total_LBAs_Read         0x0000   100   253   000    Old_age   Offline    
>>   -       50261005553
>>
>>
>>
>> However it's not only these 3 OSD to have PG with errors, these are onlyl
>> the most recent, in the last 3 months I had often OSD_SCRUB_ERRORS in
>> various OSDs, always solved by ceph pg repair <PG>, I don't think it's an
>> hardware issue.
>>
>>
>>
>>
>>
>> Il 05/03/2018 13:40, Vladimir Prokofev ha scritto:
>>
>> > candidate had a read error
>> speaks for itself - while scrubbing it coudn't read data.
>> I had similar issue, and it was just OSD dying - errors and relocated
>> sectors in SMART, just replaced the disk. But in your case it seems that
>> errors are on different OSDs? Are your OSDs all healthy?
>> You can use this command to see some details.
>> rados list-inconsistent-obj <pg.id> --format=json-pretty
>> pg.id is the PG that's reporting as inconsistent. My guess is that
>> you'll see read errors in this output, with OSD number that encountered
>> error. After that you have to check that OSD health - SMART details, etc.
>> Not always it's the disk itself that causing problems - for example we
>> had read errors because of a faulty backplane interface in a server;
>> changing the chassis resolved this issue.
>>
>>
>> 2018-03-05 14:21 GMT+03:00 Marco Baldini - H.S. Amiata <
>> mbald...@hsamiata.it>:
>>
>>> Hi
>>>
>>> After some days with debug_osd 5/5 I found [ERR] in different days,
>>> different PGs, different OSDs, different hosts. This is what I get in the
>>> OSD logs:
>>>
>>> *OSD.5 (host 3)*
>>> 2018-03-01 20:30:02.702269 7fdf4d515700  2 osd.5 pg_epoch: 16486 pg[9.1c( v 
>>> 16486'51798 (16431'50251,16486'51798] local-lis/les=16474/16475 n=3629 
>>> ec=1477/1477 lis/c 16474/16474 les/c/f 16475/16477/0 16474/16474/16474) 
>>> [5,6] r=0 lpr=16474 crt=16486'51798 lcod 16486'51797 mlcod 16486'51797 
>>> active+clean+scrubbing+deep] 9.1c shard 6: soid 
>>> 9:3b157c56:::rbd_data.1526386b8b4567.0000000000001761:head candidate had a 
>>> read error
>>> 2018-03-01 20:30:02.702278 7fdf4d515700 -1 log_channel(cluster) log [ERR] : 
>>> 9.1c shard 6: soid 
>>> 9:3b157c56:::rbd_data.1526386b8b4567.0000000000001761:head candidate had a 
>>> read error
>>>
>>> *
>>> OSD.4 (host 3)*
>>> 2018-02-28 00:03:33.458558 7f112cf76700 -1 log_channel(cluster) log [ERR] : 
>>> 13.65 shard 2: soid 
>>> 13:a719ecdf:::rbd_data.5f65056b8b4567.000000000000f8eb:head candidate had a 
>>> read error
>>>
>>> *OSD.8 (host 2)*
>>> 2018-02-27 23:55:15.100084 7f4dd0816700 -1 log_channel(cluster) log [ERR] : 
>>> 14.31 shard 1: soid 
>>> 14:8cc6cd37:::rbd_data.30b15b6b8b4567.00000000000081a1:head candidate had a 
>>> read error
>>>
>>> I don't know what this error is meaning, and as always a ceph pg repair
>>> fixes it. I don't think this is normal.
>>>
>>> Ideas?
>>>
>>> Thanks
>>>
>>> Il 28/02/2018 14:48, Marco Baldini - H.S. Amiata ha scritto:
>>>
>>> Hi
>>>
>>> I read the bugtracker issue and it seems a lot like my problem, even if
>>> I can't check the reported checksum because I don't have it in my logs,
>>> perhaps it's because of debug osd = 0/0 in ceph.conf
>>>
>>> I just raised the OSD log level
>>>
>>> ceph tell osd.* injectargs --debug-osd 5/5
>>>
>>> I'll check OSD logs in the next days...
>>>
>>> Thanks
>>>
>>>
>>>
>>> Il 28/02/2018 11:59, Paul Emmerich ha scritto:
>>>
>>> Hi,
>>>
>>> might be http://tracker.ceph.com/issues/22464
>>>
>>> Can you check the OSD log file to see if the reported checksum
>>> is 0x6706be76?
>>>
>>>
>>> Paul
>>>
>>> Am 28.02.2018 um 11:43 schrieb Marco Baldini - H.S. Amiata <
>>> mbald...@hsamiata.it>:
>>>
>>> Hello
>>>
>>> I have a little ceph cluster with 3 nodes, each with 3x1TB HDD and
>>> 1x240GB SSD. I created this cluster after Luminous release, so all OSDs are
>>> Bluestore. In my crush map I have two rules, one targeting the SSDs and one
>>> targeting the HDDs. I have 4 pools, one using the SSD rule and the others
>>> using the HDD rule, three pools are size=3 min_size=2, one is size=2
>>> min_size=1 (this one have content that it's ok to lose)
>>>
>>> In the last 3 month I'm having a strange random problem. I planned my
>>> osd scrubs during the night (osd scrub begin hour = 20, osd scrub end hour
>>> = 7) when office is closed so there is low impact on the users. Some
>>> mornings, when I ceph the cluster health, I find:
>>>
>>> HEALTH_ERR X scrub errors; Possible data damage: Y pgs inconsistent
>>> OSD_SCRUB_ERRORS X scrub errors
>>> PG_DAMAGED Possible data damage: Y pg inconsistent
>>>
>>> X and Y sometimes are 1, sometimes 2.
>>>
>>> I issue a ceph health detail, check the damaged PGs, and run a ceph pg
>>> repair for the damaged PGs, I get
>>>
>>> instructing pg PG on osd.N to repair
>>>
>>> PG are different, OSD that have to repair PG is different, even the node
>>> hosting the OSD is different, I made a list of all PGs and OSDs. This
>>> morning is the most recent case:
>>>
>>> > ceph health detail
>>> HEALTH_ERR 2 scrub errors; Possible data damage: 2 pgs inconsistent
>>> OSD_SCRUB_ERRORS 2 scrub errors
>>> PG_DAMAGED Possible data damage: 2 pgs inconsistent
>>> pg 13.65 is active+clean+inconsistent, acting [4,2,6]
>>> pg 14.31 is active+clean+inconsistent, acting [8,3,1]
>>>
>>> > ceph pg repair 13.65
>>> instructing pg 13.65 on osd.4 to repair
>>>
>>> (node-2)> tail /var/log/ceph/ceph-osd.4.log
>>> 2018-02-28 08:38:47.593447 7f112cf76700  0 log_channel(cluster) log [DBG] : 
>>> 13.65 repair starts
>>> 2018-02-28 08:39:37.573342 7f112cf76700  0 log_channel(cluster) log [DBG] : 
>>> 13.65 repair ok, 0 fixed
>>>
>>> > ceph pg repair 14.31
>>> instructing pg 14.31 on osd.8 to repair
>>>
>>> (node-3)> tail /var/log/ceph/ceph-osd.8.log
>>> 2018-02-28 08:52:37.297490 7f4dd0816700  0 log_channel(cluster) log [DBG] : 
>>> 14.31 repair starts
>>> 2018-02-28 08:53:00.704020 7f4dd0816700  0 log_channel(cluster) log [DBG] : 
>>> 14.31 repair ok, 0 fixed
>>>
>>>
>>>
>>> I made a list of when I got OSD_SCRUB_ERRORS, what PG and what OSD had
>>> to repair PG. Date is dd/mm/yyyy
>>>
>>> 21/12/2017   --  pg 14.29 is active+clean+inconsistent, acting [6,2,4]
>>>
>>> 18/01/2018   --  pg 14.5a is active+clean+inconsistent, acting [6,4,1]
>>>
>>> 22/01/2018   --  pg 9.3a is active+clean+inconsistent, acting [2,7]
>>>
>>> 29/01/2018   --  pg 13.3e is active+clean+inconsistent, acting [4,6,1]
>>>                  instructing pg 13.3e on osd.4 to repair
>>>
>>> 07/02/2018   --  pg 13.7e is active+clean+inconsistent, acting [8,2,5]
>>>                  instructing pg 13.7e on osd.8 to repair
>>>
>>> 09/02/2018   --  pg 13.30 is active+clean+inconsistent, acting [7,3,2]
>>>                  instructing pg 13.30 on osd.7 to repair
>>>
>>> 15/02/2018   --  pg 9.35 is active+clean+inconsistent, acting [1,8]
>>>                  instructing pg 9.35 on osd.1 to repair
>>>
>>>                  pg 13.3e is active+clean+inconsistent, acting [4,6,1]
>>>                  instructing pg 13.3e on osd.4 to repair
>>>
>>> 17/02/2018   --  pg 9.2d is active+clean+inconsistent, acting [7,5]
>>>                  instructing pg 9.2d on osd.7 to repair
>>>
>>> 22/02/2018   --  pg 9.24 is active+clean+inconsistent, acting [5,8]
>>>                  instructing pg 9.24 on osd.5 to repair
>>>
>>> 28/02/2018   --  pg 13.65 is active+clean+inconsistent, acting [4,2,6]
>>>                  instructing pg 13.65 on osd.4 to repair
>>>
>>>                  pg 14.31 is active+clean+inconsistent, acting [8,3,1]
>>>                  instructing pg 14.31 on osd.8 to repair
>>>
>>>
>>>
>>>
>>> If can be useful, my ceph.conf is here:
>>>
>>> [global]
>>> auth client required = none
>>> auth cluster required = none
>>> auth service required = none
>>> fsid = 24d5d6bc-0943-4345-b44e-46c19099004b
>>> cluster network = 10.10.10.0/24
>>> public network = 10.10.10.0/24
>>> keyring = /etc/pve/priv/$cluster.$name.keyring
>>> mon allow pool delete = true
>>> osd journal size = 5120
>>> osd pool default min size = 2
>>> osd pool default size = 3
>>> bluestore_block_db_size = 64424509440
>>>
>>> debug asok = 0/0
>>> debug auth = 0/0
>>> debug buffer = 0/0
>>> debug client = 0/0
>>> debug context = 0/0
>>> debug crush = 0/0
>>> debug filer = 0/0
>>> debug filestore = 0/0
>>> debug finisher = 0/0
>>> debug heartbeatmap = 0/0
>>> debug journal = 0/0
>>> debug journaler = 0/0
>>> debug lockdep = 0/0
>>> debug mds = 0/0
>>> debug mds balancer = 0/0
>>> debug mds locker = 0/0
>>> debug mds log = 0/0
>>> debug mds log expire = 0/0
>>> debug mds migrator = 0/0
>>> debug mon = 0/0
>>> debug monc = 0/0
>>> debug ms = 0/0
>>> debug objclass = 0/0
>>> debug objectcacher = 0/0
>>> debug objecter = 0/0
>>> debug optracker = 0/0
>>> debug osd = 0/0
>>> debug paxos = 0/0
>>> debug perfcounter = 0/0
>>> debug rados = 0/0
>>> debug rbd = 0/0
>>> debug rgw = 0/0
>>> debug throttle = 0/0
>>> debug timer = 0/0
>>> debug tp = 0/0
>>>
>>>
>>> [osd]
>>> keyring = /var/lib/ceph/osd/ceph-$id/keyring
>>> osd max backfills = 1
>>> osd recovery max active = 1
>>>
>>> osd scrub begin hour = 20
>>> osd scrub end hour = 7
>>> osd scrub during recovery = false
>>> osd scrub load threshold = 0.3
>>>
>>> [client]
>>> rbd cache = true
>>> rbd cache size = 268435456      # 256MB
>>> rbd cache max dirty = 201326592    # 192MB
>>> rbd cache max dirty age = 2
>>> rbd cache target dirty = 33554432    # 32MB
>>> rbd cache writethrough until flush = true
>>>
>>>
>>> #[mgr]
>>> #debug_mgr = 20
>>>
>>>
>>> [mon.pve-hs-main]
>>> host = pve-hs-main
>>> mon addr = 10.10.10.251:6789
>>>
>>> [mon.pve-hs-2]
>>> host = pve-hs-2
>>> mon addr = 10.10.10.252:6789
>>>
>>> [mon.pve-hs-3]
>>> host = pve-hs-3
>>> mon addr = 10.10.10.253:6789
>>>
>>>  My ceph versions:
>>>
>>> {
>>>     "mon": {
>>>         "ceph version 12.2.2 (215dd7151453fae88e6f968c975b6ce309d42dcf) 
>>> luminous (stable)": 3
>>>     },
>>>     "mgr": {
>>>         "ceph version 12.2.2 (215dd7151453fae88e6f968c975b6ce309d42dcf) 
>>> luminous (stable)": 3
>>>     },
>>>     "osd": {
>>>         "ceph version 12.2.2 (215dd7151453fae88e6f968c975b6ce309d42dcf) 
>>> luminous (stable)": 12
>>>     },
>>>     "mds": {},
>>>     "overall": {
>>>         "ceph version 12.2.2 (215dd7151453fae88e6f968c975b6ce309d42dcf) 
>>> luminous (stable)": 18
>>>     }
>>> }
>>>
>>>
>>>
>>>
>>> My ceph osd tree:
>>>
>>> ID CLASS WEIGHT  TYPE NAME            STATUS REWEIGHT PRI-AFF
>>> -1       8.93686 root default
>>> -6       2.94696     host pve-hs-2
>>>  3   hdd 0.90959         osd.3            up  1.00000 1.00000
>>>  4   hdd 0.90959         osd.4            up  1.00000 1.00000
>>>  5   hdd 0.90959         osd.5            up  1.00000 1.00000
>>> 10   ssd 0.21819         osd.10           up  1.00000 1.00000
>>> -3       2.86716     host pve-hs-3
>>>  6   hdd 0.85599         osd.6            up  1.00000 1.00000
>>>  7   hdd 0.85599         osd.7            up  1.00000 1.00000
>>>  8   hdd 0.93700         osd.8            up  1.00000 1.00000
>>> 11   ssd 0.21819         osd.11           up  1.00000 1.00000
>>> -7       3.12274     host pve-hs-main
>>>  0   hdd 0.96819         osd.0            up  1.00000 1.00000
>>>  1   hdd 0.96819         osd.1            up  1.00000 1.00000
>>>  2   hdd 0.96819         osd.2            up  1.00000 1.00000
>>>  9   ssd 0.21819         osd.9            up  1.00000 1.00000
>>>
>>>
>>> My pools:
>>>
>>> pool 9 'cephbackup' replicated size 2 min_size 1 crush_rule 1 object_hash 
>>> rjenkins pg_num 64 pgp_num 64 last_change 5665 flags hashpspool 
>>> stripe_width 0 application rbd
>>>         removed_snaps [1~3]
>>> pool 13 'cephwin' replicated size 3 min_size 2 crush_rule 1 object_hash 
>>> rjenkins pg_num 128 pgp_num 128 last_change 16454 flags hashpspool 
>>> stripe_width 0 application rbd
>>>         removed_snaps [1~5]
>>> pool 14 'cephnix' replicated size 3 min_size 2 crush_rule 1 object_hash 
>>> rjenkins pg_num 128 pgp_num 128 last_change 16482 flags hashpspool 
>>> stripe_width 0 application rbd
>>>         removed_snaps [1~227]
>>> pool 17 'cephssd' replicated size 3 min_size 2 crush_rule 2 object_hash 
>>> rjenkins pg_num 64 pgp_num 64 last_change 8601 flags hashpspool 
>>> stripe_width 0 application rbd
>>>         removed_snaps [1~3]
>>>
>>>
>>>
>>> I can't understand where the problem comes from, I don't think it's
>>> hardware, if I had a failed disk, then I should have problems always on the
>>> same OSD. Any ideas
>>>
>>> Thanks
>>>
>>>
>>>
>>> --
>>> *Marco Baldini*
>>> *H.S. Amiata Srl*
>>> Ufficio:   0577-779396
>>> Cellulare:   335-8765169
>>> WEB:   www.hsamiata.it
>>> EMAIL:   mbald...@hsamiata.it
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>> --
>>> Mit freundlichen Grüßen / Best Regards
>>> Paul Emmerich
>>>
>>> croit GmbH
>>> Freseniusstr. 31h
>>> 81247 München
>>> www.croit.io
>>> Tel: +49 89 1896585 90 <+49%2089%20189658590>
>>>
>>> Geschäftsführer: Martin Verges
>>> Handelsregister: Amtsgericht München
>>> USt-IdNr: DE310638492
>>>
>>>
>>> --
>>> *Marco Baldini*
>>> *H.S. Amiata Srl*
>>> Ufficio:   0577-779396
>>> Cellulare:   335-8765169
>>> WEB:   www.hsamiata.it
>>> EMAIL:   mbald...@hsamiata.it
>>>
>>>
>>> _______________________________________________
>>> ceph-users mailing 
>>> listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>> --
>>> *Marco Baldini*
>>> *H.S. Amiata Srl*
>>> Ufficio:   0577-779396
>>> Cellulare:   335-8765169
>>> WEB:   www.hsamiata.it
>>> EMAIL:   mbald...@hsamiata.it
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>
>>
>> _______________________________________________
>> ceph-users mailing 
>> listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>> --
>> *Marco Baldini*
>> *H.S. Amiata Srl*
>> Ufficio:   0577-779396
>> Cellulare:   335-8765169
>> WEB:   www.hsamiata.it
>> EMAIL:   mbald...@hsamiata.it
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>
>
> _______________________________________________
> ceph-users mailing 
> listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
> --
> *Marco Baldini*
> *H.S. Amiata Srl*
> Ufficio:   0577-779396
> Cellulare:   335-8765169
> WEB:   www.hsamiata.it
> EMAIL:   mbald...@hsamiata.it
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>


-- 
Cheers,
Brad
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to