So it was a PG problem. I added a couple of OSD per host, reconfigured the CRUSH map and the cluster began to work properly.
Thanks Giuseppe 2015-04-14 19:02 GMT+02:00 Saverio Proto <ziopr...@gmail.com>: > No error message. You just finish the RAM memory and you blow up the > cluster because of too many PGs. > > Saverio > > 2015-04-14 18:52 GMT+02:00 Giuseppe Civitella < > giuseppe.civite...@gmail.com>: > > Hi Saverio, > > > > I first made a test on my test staging lab where I have only 4 OSD. > > On my mon servers (which run other services) I have 16BG RAM, 15GB used > but > > 5 cached. On the OSD servers I have 3GB RAM, 3GB used but 2 cached. > > "ceph -s" tells me nothing about PGs, shouldn't I get an error message > from > > its output? > > > > Thanks > > Giuseppe > > > > 2015-04-14 18:20 GMT+02:00 Saverio Proto <ziopr...@gmail.com>: > >> > >> You only have 4 OSDs ? > >> How much RAM per server ? > >> I think you have already too many PG. Check your RAM usage. > >> > >> Check on Ceph wiki guidelines to dimension the correct number of PGs. > >> Remeber that everytime to create a new pool you add PGs into the > >> system. > >> > >> Saverio > >> > >> > >> 2015-04-14 17:58 GMT+02:00 Giuseppe Civitella > >> <giuseppe.civite...@gmail.com>: > >> > Hi all, > >> > > >> > I've been following this tutorial to realize my setup: > >> > > >> > > http://www.sebastien-han.fr/blog/2014/08/25/ceph-mix-sata-and-ssd-within-the-same-box/ > >> > > >> > I got this CRUSH map from my test lab: > >> > http://paste.openstack.org/show/203887/ > >> > > >> > then I modified the map and uploaded it. This is the final version: > >> > http://paste.openstack.org/show/203888/ > >> > > >> > When applied the new CRUSH map, after some rebalancing, I get this > >> > health > >> > status: > >> > [-> avalon1 root@controller001 Ceph <-] # ceph -s > >> > cluster af09420b-4032-415e-93fc-6b60e9db064e > >> > health HEALTH_WARN crush map has legacy tunables; > mon.controller001 > >> > low > >> > disk space; clock skew detected on mon.controller002 > >> > monmap e1: 3 mons at > >> > > >> > {controller001= > 10.235.24.127:6789/0,controller002=10.235.24.128:6789/0,controller003=10.235.24.129:6789/0 > }, > >> > election epoch 314, quorum 0,1,2 > >> > controller001,controller002,controller003 > >> > osdmap e3092: 4 osds: 4 up, 4 in > >> > pgmap v785873: 576 pgs, 6 pools, 71548 MB data, 18095 objects > >> > 8842 MB used, 271 GB / 279 GB avail > >> > 576 active+clean > >> > > >> > and this osd tree: > >> > [-> avalon1 root@controller001 Ceph <-] # ceph osd tree > >> > # id weight type name up/down reweight > >> > -8 2 root sed > >> > -5 1 host ceph001-sed > >> > 2 1 osd.2 up 1 > >> > -7 1 host ceph002-sed > >> > 3 1 osd.3 up 1 > >> > -1 2 root default > >> > -4 1 host ceph001-sata > >> > 0 1 osd.0 up 1 > >> > -6 1 host ceph002-sata > >> > 1 1 osd.1 up 1 > >> > > >> > which seems not a bad situation. The problem rise when I try to > create a > >> > new > >> > pool, the command "ceph osd pool create sed 128 128" gets stuck. It > >> > never > >> > ends. And I noticed that my Cinder installation is not able to create > >> > volumes anymore. > >> > I've been looking in the logs for errors and found nothing. > >> > Any hint about how to proceed to restore my ceph cluster? > >> > Is there something wrong with the steps I take to update the CRUSH > map? > >> > Is > >> > the problem related to Emperor? > >> > > >> > Regards, > >> > Giuseppe > >> > > >> > > >> > > >> > > >> > 2015-04-13 18:26 GMT+02:00 Giuseppe Civitella > >> > <giuseppe.civite...@gmail.com>: > >> >> > >> >> Hi all, > >> >> > >> >> I've got a Ceph cluster which serves volumes to a Cinder > installation. > >> >> It > >> >> runs Emperor. > >> >> I'd like to be able to replace some of the disks with OPAL disks and > >> >> create a new pool which uses exclusively the latter kind of disk. I'd > >> >> like > >> >> to have a "traditional" pool and a "secure" one coexisting on the > same > >> >> ceph > >> >> host. I'd then use Cinder multi backend feature to serve them. > >> >> My question is: how is it possible to realize such a setup? How can I > >> >> bind > >> >> a pool to certain OSDs? > >> >> > >> >> Thanks > >> >> Giuseppe > >> > > >> > > >> > > >> > _______________________________________________ > >> > ceph-users mailing list > >> > ceph-users@lists.ceph.com > >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> > > > > > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com