Sure, http://www.sebastien-han.fr/blog/2012/06/10/introducing-ceph-to-openstack/
On Wed, Jun 5, 2013 at 11:38 AM, Явор Маринов <ymari...@neterra.net> wrote: > I've managed to start and mount the cluster by completely starting the > process from scratch. Other thing that i'm searching for is any > documentation how to add another node (or hard drives) on a running cluster > without affecting the mount point and the running service. Can you point me > for this? > > > > > On 06/05/2013 11:20 AM, Igor Laskovy wrote: > > >and I'm unable to mount the cluster with the following command: > >root@ceph1:/mnt# mount -t ceph 192.168.2.170:6789:/ /mnt > > So, what it says? > > I'm also recommend to you start from my russian doc > http://habrahabr.ru/post/179823 > > > On Tue, Jun 4, 2013 at 4:22 PM, Явор Маринов <ymari...@neterra.net> wrote: > >> That's the exact documentation which i'm using the directory on ceph2 >> is created, and the service is starting without any problems on both nodes. >> However the health of the cluster is getting WARN and i was able to mount >> the cluster >> >> >> >> >> On 06/04/2013 03:43 PM, Andrei Mikhailovsky wrote: >> >> Yavor, >> >> I would highly recommend taking a look at the quick install guide: >> http://ceph.com/docs/next/start/quick-start/ >> >> As per the guide, you need to precreate the directories prior to starting >> ceph. >> >> Andrei >> ------------------------------ >> *From: *"Явор Маринов" <ymari...@neterra.net> <ymari...@neterra.net> >> *To: *ceph-users@lists.ceph.com >> *Sent: *Tuesday, 4 June, 2013 11:03:52 AM >> *Subject: *[ceph-users] ceph configuration >> >> >> Hello, >> >> I'm new to the Ceph mailing list, and I need some advices for our >> testing cluster. I have 2 servers with x2 hard disks. On the first >> server i configured monitor and OSD, and on the second server only OSD. >> The configuration looks like as follows: >> >> [mon.a] >> >> host = ceph1 >> mon addr = 192.168.2.170:6789 >> >> [osd.0] >> host = ceph1 >> addr = 192.168.2.170 >> devs = /dev/sdb >> >> [osd.1] >> host = ceph2 >> addr = 192.168.2.114 >> devs = /dev/sdb >> >> Once i initiate 'service ceph -a start' i keep getting the following >> error: >> >> Mounting xfs on ceph2:/var/lib/ceph/osd/ceph-1 >> df: `/var/lib/ceph/osd/ceph-1/.': No such file or directory >> >> and I'm unable to mount the cluster with the following command: >> root@ceph1:/mnt# mount -t ceph 192.168.2.170:6789:/ /mnt >> >> Also executing 'ceph health' i'm getting this response: >> HEALTH_WARN 143 pgs degraded; 576 pgs stuck unclean; recovery 15/122 >> degraded (12.295%) >> >> This is fresh install and there aren't any nodes which are added/removed. >> >> Any help will be much appreciated. >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> > > > -- > Igor Laskovy > facebook.com/igor.laskovy > studiogrizzly.com > > > -- Igor Laskovy facebook.com/igor.laskovy studiogrizzly.com
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com