>> On Thu, 2 Dec 2010, Christian Brunner wrote:
>>> We have simulated the simultanious crash of multiple osds in our
>>> environment. After starting all the cosd again, we have the following
>>> situation:
>>>
>>> 2010-12-02 16:18:33.944436    pg v724432: 3712 pgs: 1 active, 3605
>>> active+clean, 1 crashed+peering, 46 down+peering, 56
>>> crashed+down+peering, 3 active+clean+inconsistent; 177 GB data, 365 GB
>>> used, 83437 GB / 83834 GB avail; 1/93704 degraded (0.001%)
>>>
>>> When I set of an "rbd rm" command for one of our rbd volumes, it seems
>>> to hit the the "crashed+down+peering" pg. After that the command is
>>> stuck.
>>
>> The pg isn't active, so any IO will hang until peering completes.  What
>> version of the code are you running?  If it's something from unstable
>> from the last couple of weeks it's probably related to problems there;
>> please upgrade and restart the osds.  If it's the latest and greatest
>> 'rc', we should look at the logs to see what's going on!
>
> We are running 0.23 - I will upgrade to the latest 'rc' tomorrow.

Upgrading to the latest rc version worked well. Everything is working
again and all, except one pg are "active+clean". However there is one
pg marked as "active+clean+inconsistent".

What can I do about an inconsistent group? I general a short
description of the possible pg states would be helpful.

Thanks,
Christian
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to