Re: [ceph-users] Major ceph disaster

2019-05-23 Thread Alexandre Marangone
The PGs will stay active+recovery_wait+degraded until you solve the unfound
objects issue.
You can follow this doc to look at which objects are unfound[1]  and if no
other recourse mark them lost

[1]
http://docs.ceph.com/docs/master/rados/troubleshooting/troubleshooting-pg/#unfound-objects
.

On Thu, May 23, 2019 at 5:47 AM Kevin Flöh  wrote:

> thank you for this idea, it has improved the situation. Nevertheless,
> there are still 2 PGs in recovery_wait. ceph -s gives me:
>
>cluster:
>  id: 23e72372-0d44-4cad-b24f-3641b14b86f4
>  health: HEALTH_WARN
>  3/125481112 objects unfound (0.000%)
>  Degraded data redundancy: 3/497011315 objects degraded
> (0.000%), 2 pgs degraded
>
>services:
>  mon: 3 daemons, quorum ceph-node03,ceph-node01,ceph-node02
>  mgr: ceph-node01(active), standbys: ceph-node01.etp.kit.edu
>  mds: cephfs-1/1/1 up  {0=ceph-node03.etp.kit.edu=up:active}, 3
> up:standby
>  osd: 96 osds: 96 up, 96 in
>
>data:
>  pools:   2 pools, 4096 pgs
>  objects: 125.48M objects, 259TiB
>  usage:   370TiB used, 154TiB / 524TiB avail
>  pgs: 3/497011315 objects degraded (0.000%)
>   3/125481112 objects unfound (0.000%)
>   4083 active+clean
>   10   active+clean+scrubbing+deep
>   2active+recovery_wait+degraded
>   1active+clean+scrubbing
>
>io:
>  client:   318KiB/s rd, 77.0KiB/s wr, 190op/s rd, 0op/s wr
>
>
> and ceph health detail:
>
> HEALTH_WARN 3/125481112 objects unfound (0.000%); Degraded data
> redundancy: 3/497011315 objects degraded (0.000%), 2 p
> gs degraded
> OBJECT_UNFOUND 3/125481112 objects unfound (0.000%)
>  pg 1.24c has 1 unfound objects
>  pg 1.779 has 2 unfound objects
> PG_DEGRADED Degraded data redundancy: 3/497011315 objects degraded
> (0.000%), 2 pgs degraded
>  pg 1.24c is active+recovery_wait+degraded, acting [32,4,61,36], 1
> unfound
>  pg 1.779 is active+recovery_wait+degraded, acting [50,4,77,62], 2
> unfound
>
>
> also the status changed form HEALTH_ERR to HEALTH_WARN. We also did ceph
> osd down for all OSDs of the degraded PGs. Do you have any further
> suggestions on how to proceed?
>
> On 23.05.19 11:08 vorm., Dan van der Ster wrote:
> > I think those osds (1, 11, 21, 32, ...) need a little kick to re-peer
> > their degraded PGs.
> >
> > Open a window with `watch ceph -s`, then in another window slowly do
> >
> >  ceph osd down 1
> >  # then wait a minute or so for that osd.1 to re-peer fully.
> >  ceph osd down 11
> >  ...
> >
> > Continue that for each of the osds with stuck requests, or until there
> > are no more recovery_wait/degraded PGs.
> >
> > After each `ceph osd down...`, you should expect to see several PGs
> > re-peer, and then ideally the slow requests will disappear and the
> > degraded PGs will become active+clean.
> > If anything else happens, you should stop and let us know.
> >
> >
> > -- dan
> >
> > On Thu, May 23, 2019 at 10:59 AM Kevin Flöh  wrote:
> >> This is the current status of ceph:
> >>
> >>
> >> cluster:
> >>   id: 23e72372-0d44-4cad-b24f-3641b14b86f4
> >>   health: HEALTH_ERR
> >>   9/125481144 objects unfound (0.000%)
> >>   Degraded data redundancy: 9/497011417 objects degraded
> >> (0.000%), 7 pgs degraded
> >>   9 stuck requests are blocked > 4096 sec. Implicated osds
> >> 1,11,21,32,43,50,65
> >>
> >> services:
> >>   mon: 3 daemons, quorum ceph-node03,ceph-node01,ceph-node02
> >>   mgr: ceph-node01(active), standbys: ceph-node01.etp.kit.edu
> >>   mds: cephfs-1/1/1 up  {0=ceph-node03.etp.kit.edu=up:active}, 3
> >> up:standby
> >>   osd: 96 osds: 96 up, 96 in
> >>
> >> data:
> >>   pools:   2 pools, 4096 pgs
> >>   objects: 125.48M objects, 259TiB
> >>   usage:   370TiB used, 154TiB / 524TiB avail
> >>   pgs: 9/497011417 objects degraded (0.000%)
> >>9/125481144 objects unfound (0.000%)
> >>4078 active+clean
> >>11   active+clean+scrubbing+deep
> >>7active+recovery_wait+degraded
> >>
> >> io:
> >>   client:   211KiB/s rd, 46.0KiB/s wr, 158op/s rd, 0op/s wr
> >>
> >> On 23.05.19 10:54 vorm., Dan van der Ster wrote:
> >>> What's the full ceph status?
> >>> Normally recovery_wait just means that the relevant osd's are busy
> >>> recovering/backfilling another PG.
> >>>
> >>> On Thu, May 23, 2019 at 10:53 AM Kevin Flöh 
> wrote:
>  Hi,
> 
>  we have set the PGs to recover and now they are stuck in
> active+recovery_wait+degraded and instructing them to deep-scrub does not
> change anything. Hence, the rados report is empty. Is there a way to stop
> the recovery wait to start the deep-scrub and get the output? I guess the
> recovery_wait might be caused by missing objects. Do we need to delete them
> first to get the recovery going?
> 
>  Kevin
> 
>  On 

Re: [ceph-users] How to just delete PGs stuck incomplete on EC pool

2019-03-02 Thread Alexandre Marangone
If you have no way to recover the drives, you can try to reboot the OSDs
with `osd_find_best_info_ignore_history_les = true` (revert it afterwards),
you'll lose data. If after this, the PGs are down, you can mark the OSDs
blocking the PGs from become active lost.

On Sat, Mar 2, 2019 at 6:08 AM Daniel K  wrote:

> They all just started having read errors. Bus resets. Slow reads. Which is
> one of the reasons the cluster didn't recover fast enough to compensate.
>
> I tried to be mindful of the drive type and specifically avoided the
> larger capacity Seagates that are SMR. Used 1 SM863 for every 6 drives for
> the WAL.
>
> Not sure why they failed. The data isn't critical at this point, just need
> to get the cluster back to normal.
>
> On Sat, Mar 2, 2019, 9:00 AM  wrote:
>
>> Did they break, or did something went wronng trying to replace them?
>>
>> Jespe
>>
>>
>>
>> Sent from myMail for iOS
>>
>>
>> Saturday, 2 March 2019, 14.34 +0100 from Daniel K :
>>
>> I bought the wrong drives trying to be cheap. They were 2TB WD Blue
>> 5400rpm 2.5 inch laptop drives.
>>
>> They've been replace now with HGST 10K 1.8TB SAS drives.
>>
>>
>>
>> On Sat, Mar 2, 2019, 12:04 AM  wrote:
>>
>>
>>
>> Saturday, 2 March 2019, 04.20 +0100 from satha...@gmail.com <
>> satha...@gmail.com>:
>>
>> 56 OSD, 6-node 12.2.5 cluster on Proxmox
>>
>> We had multiple drives fail(about 30%) within a few days of each other,
>> likely faster than the cluster could recover.
>>
>>
>> Hov did so many drives break?
>>
>> Jesper
>>
>> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Ceph on Solaris / Illumos

2015-04-15 Thread Alexandre Marangone
The LX branded zones might be a way to run OSDs on Illumos:
https://wiki.smartos.org/display/DOC/LX+Branded+Zones

For fun, I tried a month or so ago, managed to have a quorum. OSDs
wouldn't start, I didn't look further as far as debugging. I'll give
it a go when I have more time.

On Wed, Apr 15, 2015 at 7:04 AM, Mark Nelson  wrote:
>
>
> On 04/15/2015 08:16 AM, Jake Young wrote:
>>
>> Has anyone compiled ceph (either osd or client) on a Solaris based OS?
>>
>> The thread on ZFS support for osd got me thinking about using solaris as
>> an osd server. It would have much better ZFS performance and I wonder if
>> the osd performance without a journal would be 2x better.
>
>
> Doubt it.  You may be able to do a little better, but you have to pay the
> piper some how.  If you clone from journal you will introduce fragmentation.
> If you throw the journal away you'll suffer for everything but very large
> writes unless you throw safety away.  I think if we are going to generally
> beat filestore (not just for optimal benchmarking tests!) it's going to take
> some very careful cleverness. Thankfully Sage is very clever and is working
> on it in newstore. Even there, filestore has been proving difficult to beat
> for writes.
>
>>
>> A second thought I had was using the Comstar iscsi / fcoe target
>> software that is part of Solaris. Has anyone done anything with a ceph
>> rbd client for Solaris based OSs?
>
>
> No idea!
>
>>
>> Jake
>>
>>
>> ___
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com