Hi! I have tried, but situation not changed significantly:
# ceph -w cluster e90dfd37-98d1-45bb-a847-8590a5ed8e71 health HEALTH_WARN 192 pgs stuck inactive; 192 pgs stuck unclean; 2/2 in osds are down monmap e1: 1 mons at {host1=172.17.0.4:6789/0}, election epoch 1, quorum 0 host1 osdmap e9: 2 osds: 0 up, 2 in pgmap v10: 192 pgs, 3 pools, 0 bytes data, 0 objects 0 kB used, 0 kB / 0 kB avail 192 creating 2014-02-16 17:25:29.872538 mon.0 [INF] osdmap e9: 2 osds: 0 up, 2 in # ceph osd tree # id weight type name up/down reweight -1 2 root default -2 2 host host1 0 1 osd.0 down 1 1 1 osd.1 down 1 # ceph health HEALTH_WARN 192 pgs stuck inactive; 192 pgs stuck unclean; 2/2 in osds are down ps showed both osd daemons running. Pavel. 17 февр. 2014 г., в 1:50, Karan Singh <karan.si...@csc.fi> написал(а): > Hi Pavel > > Try to add at least 1 more OSD ( bare minimum ) and set pool replication to 2 > after that. > For osd.0 try , # ceph osd in osd.0 , once the osd is IN , try to bring > up osd.0 services up > > > Finally your both the OSD should be IN and UP , so that your cluster can > store data. > > Regards > Karan > > > On 16 Feb 2014, at 20:06, Pavel V. Kaygorodov <pa...@inasan.ru> wrote: > >> Hi, All! >> >> I am trying to setup ceph from scratch, without dedicated drive, with one >> mon and one osd. >> After all, I see following output of ceph osd tree: >> >> # id weight type name up/down reweight >> -1 1 root default >> -2 1 host host1 >> 0 1 osd.0 down 0 >> >> ceph -w: >> >> cluster e90dfd37-98d1-45bb-a847-8590a5ed8e71 >> health HEALTH_WARN 192 pgs stuck inactive; 192 pgs stuck unclean >> monmap e1: 1 mons at {host1=172.17.0.4:6789/0}, election epoch 1, quorum >> 0 host1 >> osdmap e5: 1 osds: 0 up, 0 in >> pgmap v6: 192 pgs, 3 pools, 0 bytes data, 0 objects >> 0 kB used, 0 kB / 0 kB avail >> 192 creating >> >> 2014-02-16 13:27:30.095938 mon.0 [INF] osdmap e5: 1 osds: 0 up, 0 in >> >> What can be wrong? >> I see working daemons, and nothing bad in log files. >> >> //////////////////////////////////// >> >> How to reproduce: >> I have cloned and compiled sources on debian/jessie: >> >> git clone --recursive -b v0.75 https://github.com/ceph/ceph.git >> cd /ceph/ && ./autogen.sh && ./configure && make && make install >> >> Everything seems ok. >> >> I have created ceph.conf: >> >> [global] >> >> fsid = e90dfd37-98d1-45bb-a847-8590a5ed8e71 >> mon initial members = host1 >> >> auth cluster required = cephx >> auth service required = cephx >> auth client required = cephx >> >> keyring = /data/ceph.client.admin.keyring >> >> osd pool default size = 1 >> osd pool default min size = 1 >> osd pool default pg num = 333 >> osd pool default pgp num = 333 >> osd crush chooseleaf type = 0 >> osd journal size = 1000 >> >> filestore xattr use omap = true >> >> ;journal dio = false >> ;journal aio = false >> >> mon addr = ceph.dkctl >> mon host = ceph.dkctl >> >> log file = /data/logs/ceph.log >> >> [mon] >> mon data = /data/mon0 >> keyring = /data/ceph.mon.keyring >> log file = /data/logs/mon0.log >> >> [osd.0] >> osd host = host1 >> osd data = /data/osd0 >> osd journal = /data/osd0.journal >> log file = /data/logs/osd0.log >> keyring = /data/ceph.osd0.keyring >> >> /////////////////////////////// >> >> I have initialized mon and osd using following script: >> >> /usr/local/bin/ceph-authtool --create-keyring /data/ceph.mon.keyring >> --gen-key -n mon. --cap mon 'allow *' >> /usr/local/bin/ceph-authtool --create-keyring >> /data/ceph.client.admin.keyring --gen-key -n client.admin --set-uid=0 --cap >> mon 'allow *' --cap osd 'allow *' --cap mds 'allow' >> /usr/local/bin/ceph-authtool /data/ceph.mon.keyring --import-keyring >> /data/ceph.client.admin.keyring >> /usr/local/bin/monmaptool --create --add host1 `grep ceph /etc/hosts | awk >> '{print $1}'` --fsid de90dfd37-98d1-45bb-a847-8590a5ed8e71 /data/monmap >> /usr/local/bin/ceph-mon --mkfs -i host1 --monmap /data/monmap --keyring >> /data/ceph.mon.keyring >> /usr/local/bin/ceph-mon -c /ceph.conf --public-addr `grep ceph /etc/hosts | >> awk '{print $1}'` -i host1 >> /usr/local/bin/ceph osd create e90dfd37-98d1-45bb-a847-8590a5ed8e71 >> /usr/local/bin/ceph osd create e90dfd37-98d1-45bb-a847-8590a5ed8e71 >> /usr/local/bin/ceph-osd -i 0 --mkfs --mkkey >> /usr/local/bin/ceph auth add osd.0 osd 'allow *' mon 'allow rwx' -i >> /data/ceph.osd0.keyring >> /usr/local/bin/ceph osd crush add-bucket host1 host >> /usr/local/bin/ceph osd crush move host1 root=default >> /usr/local/bin/ceph osd crush add osd.0 1.0 host=host1 >> >> //////////////////////////////////////////////////// >> >> Script ouptut seems to be ok: >> >> creating /data/ceph.mon.keyring >> creating /data/ceph.client.admin.keyring >> importing contents of /data/ceph.client.admin.keyring into >> /data/ceph.mon.keyring >> /usr/local/bin/monmaptool: monmap file /data/monmap >> /usr/local/bin/monmaptool: set fsid to e90dfd37-98d1-45bb-a847-8590a5ed8e71 >> /usr/local/bin/monmaptool: writing epoch 0 to /data/monmap (1 monitors) >> /usr/local/bin/ceph-mon: set fsid to e90dfd37-98d1-45bb-a847-8590a5ed8e71 >> /usr/local/bin/ceph-mon: created monfs at /data/mon0 for mon.host1 >> 0 >> 2014-02-16 13:24:37.833469 7f5ef61747c0 -1 journal FileJournal::_open: >> disabling aio for non-block journal. Use journal_force_aio to force use of >> aio anyway >> 2014-02-16 13:24:37.941111 7f5ef61747c0 -1 journal FileJournal::_open: >> disabling aio for non-block journal. Use journal_force_aio to force use of >> aio anyway >> 2014-02-16 13:24:37.948704 7f5ef61747c0 -1 filestore(/data/osd0) could not >> find 23c2fcde/osd_superblock/0//-1 in index: (2) No such file or directory >> 2014-02-16 13:24:38.054345 7f5ef61747c0 -1 created object store /data/osd0 >> journal /data/osd0.journal for osd.0 fsid >> e90dfd37-98d1-45bb-a847-8590a5ed8e71 >> 2014-02-16 13:24:38.054427 7f5ef61747c0 -1 auth: error reading file: >> /data/ceph.osd0.keyring: can't open /data/ceph.osd0.keyring: (2) No such >> file or directory >> 2014-02-16 13:24:38.054529 7f5ef61747c0 -1 created new key in keyring >> /data/ceph.osd0.keyring >> added key for osd.0 >> added bucket host1 type host to crush map >> moved item id -2 name 'host1' to location {root=default} in crush map >> add item id 0 name 'osd.0' weight 1 at location {host=host1} to crush map >> >> /////////////////////////////////////////////////////////////////////// >> >> I am started daemons, using commands: >> >> /usr/local/bin/ceph-mon -c /ceph.conf --public-addr `grep ceph /etc/hosts | >> awk '{print $1}'` -i host1 --debug_ms 1000 >> /usr/local/bin/ceph-osd -c /ceph.conf -i 0 --debug_ms 1000 >> >> ///////////////////////////////////////////////////////////////////// >> >> Daemons are alive, log files show evidences of some communications between >> mon and osd: >> >> mon0.log: >> >> 2014-02-16 13:27:30.399455 7f02699e0700 10 -- 172.17.0.4:6789/0 >> >> 172.17.0.4:6800/20 pipe(0x358a500 sd=21 :6789 s=2 pgs=1 cs=1 l=1 >> c=0x34d1a20).write_ack 8 >> 2014-02-16 13:27:30.399463 7f02699e0700 10 -- 172.17.0.4:6789/0 >> >> 172.17.0.4:6800/20 pipe(0x358a500 sd=21 :6789 s=2 pgs=1 cs=1 l=1 >> c=0x34d1a20).writer: state = open policy.server=1 >> 2014-02-16 13:27:30.399467 7f02699e0700 20 -- 172.17.0.4:6789/0 >> >> 172.17.0.4:6800/20 pipe(0x358a500 sd=21 :6789 s=2 pgs=1 cs=1 l=1 >> c=0x34d1a20).writer sleeping >> 2014-02-16 13:27:33.367087 7f02705a8700 20 -- 172.17.0.4:6789/0 >> >> 172.17.0.4:6800/20 pipe(0x358a500 sd=21 :6789 s=2 pgs=1 cs=1 l=1 >> c=0x34d1a20).reader got KEEPALIVE >> 2014-02-16 13:27:33.367133 7f02705a8700 20 -- 172.17.0.4:6789/0 >> >> 172.17.0.4:6800/20 pipe(0x358a500 sd=21 :6789 s=2 pgs=1 cs=1 l=1 >> c=0x34d1a20).reader reading tag... >> >> osd0.log: >> >> 2014-02-16 13:27:53.368482 7f340bfb3700 20 -- 172.17.0.4:6800/20 >> send_keepalive con 0x2381a20, have pipe. >> 2014-02-16 13:27:53.368702 7f3420ef7700 10 -- 172.17.0.4:6800/20 >> >> 172.17.0.4:6789/0 pipe(0x2388000 sd=25 :38871 s=2 pgs=1 cs=1 l=1 >> c=0x2381a20).writer: state = open policy.server=0 >> 2014-02-16 13:27:53.368806 7f3420ef7700 10 -- 172.17.0.4:6800/20 >> >> 172.17.0.4:6789/0 pipe(0x2388000 sd=25 :38871 s=2 pgs=1 cs=1 l=1 >> c=0x2381a20).write_keepalive >> 2014-02-16 13:27:53.369000 7f3420ef7700 10 -- 172.17.0.4:6800/20 >> >> 172.17.0.4:6789/0 pipe(0x2388000 sd=25 :38871 s=2 pgs=1 cs=1 l=1 >> c=0x2381a20).writer: state = open policy.server=0 >> 2014-02-16 13:27:53.369070 7f3420ef7700 20 -- 172.17.0.4:6800/20 >> >> 172.17.0.4:6789/0 pipe(0x2388000 sd=25 :38871 s=2 pgs=1 cs=1 l=1 >> c=0x2381a20).writer sleeping >> >> With best regards, >> Pavel. >> >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com