Hi,
Anyways its not suggested to have a single node, but if you are having and
you must (may be for testing purposes) you can include :

"osd crush chooseleaf type = 0" in global section of ceph.conf and restart
all ceph services, to have all pgs in active+clean state.

Thanks and Regards
Ashish Chandra
Cloud Engineer, Reliance Jio


On Thu, Mar 13, 2014 at 2:35 PM, Robert van Leeuwen <
robert.vanleeu...@spilgames.com> wrote:

>  > The question is that I cannot understand why the status of the PGS is
> always STUCK UNCLEAN. As I
> > see it, the status should be ACTIVE+CLEAN.
>
> It looks like you have one physical node.
> If you have a pool with a replication count of 2 (default) I think it wil
> try to spread the data across 2 failure domains by default.
> My guess is the default crush map will see a node as a single failure
> domain by default.
> So, edit the crushmap to allow this or add a second node.
>
> Cheers,
> Robert van Leeuwen
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to