tough luck, hope everything comes up ok afterwards. What models on the SSD?

/Josef
On 17 Apr 2015 20:05, "Andrija Panic" <andrija.pa...@gmail.com> wrote:

> SSD died that hosted journals for 6 OSDs - 2 x SSD died, so 12 OSDs are
> down, and rebalancing is about finish... after which I need to fix the OSDs.
>
> On 17 April 2015 at 19:01, Josef Johansson <jo...@oderland.se> wrote:
>
>> Hi,
>>
>> Did 6 other OSDs go down when re-adding?
>>
>> /Josef
>>
>> On 17 Apr 2015, at 18:49, Andrija Panic <andrija.pa...@gmail.com> wrote:
>>
>> 12 osds down - I expect less work with removing and adding osd?
>> On Apr 17, 2015 6:35 PM, "Krzysztof Nowicki" <
>> krzysztof.a.nowi...@gmail.com> wrote:
>>
>>> Why not just wipe out the OSD filesystem, run ceph-osd --mkfs with the
>>> existing OSD UUID, copy the keyring and let it populate itself?
>>>
>>> pt., 17 kwi 2015 o 18:31 użytkownik Andrija Panic <
>>> andrija.pa...@gmail.com> napisał:
>>>
>>>> Thx guys, thats what I will be doing at the end.
>>>>
>>>> Cheers
>>>> On Apr 17, 2015 6:24 PM, "Robert LeBlanc" <rob...@leblancnet.us> wrote:
>>>>
>>>>> Delete and re-add all six OSDs.
>>>>>
>>>>> On Fri, Apr 17, 2015 at 3:36 AM, Andrija Panic <
>>>>> andrija.pa...@gmail.com> wrote:
>>>>>
>>>>>> Hi guys,
>>>>>>
>>>>>> I have 1 SSD that hosted 6 OSD's Journals, that is dead, so 6 OSD
>>>>>> down, ceph rebalanced etc.
>>>>>>
>>>>>> Now I have new SSD inside, and I will partition it etc - but would
>>>>>> like to know, how to proceed now, with the journal recreation for those 6
>>>>>> OSDs that are down now.
>>>>>>
>>>>>> Should I flush journal (where to, journals doesnt still exist...?),
>>>>>> or just recreate journal from scratch (making symboliv links again: ln -s
>>>>>> /dev/$DISK$PART /var/lib/ceph/osd/ceph-$ID/journal) and starting OSDs.
>>>>>>
>>>>>> I expect the folowing procedure, but would like confirmation please:
>>>>>>
>>>>>> rm /var/lib/ceph/osd/ceph-$ID/journal -f (sym link)
>>>>>> ln -s /dev/SDAxxx /var/lib/ceph/osd/ceph-$ID/journal
>>>>>> ceph-osd -i $ID --mkjournal
>>>>>> ll /var/lib/ceph/osd/ceph-$ID/journal
>>>>>> service ceph start osd.$ID
>>>>>>
>>>>>> Any thought greatly appreciated !
>>>>>>
>>>>>> Thanks,
>>>>>>
>>>>>> --
>>>>>>
>>>>>> Andrija Panić
>>>>>>
>>>>>> _______________________________________________
>>>>>> ceph-users mailing list
>>>>>> ceph-users@lists.ceph.com
>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>>
>>>>>>
>>>>>  _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@lists.ceph.com
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>
>>>  _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>>
>
>
> --
>
> Andrija Panić
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to