Sure,
http://www.sebastien-han.fr/blog/2012/06/10/introducing-ceph-to-openstack/


On Wed, Jun 5, 2013 at 11:38 AM, Явор Маринов <ymari...@neterra.net> wrote:

>  I've managed to start and mount the cluster by completely starting the
> process from scratch. Other thing that i'm searching for is any
> documentation how to add another node (or hard drives) on a running cluster
> without affecting the mount point and the running service. Can you point me
> for this?
>
>
>
>
>  On 06/05/2013 11:20 AM, Igor Laskovy wrote:
>
> >and I'm unable to mount the cluster with the following command:
> >root@ceph1:/mnt# mount -t ceph 192.168.2.170:6789:/ /mnt
>
>  So, what it says?
>
>  I'm also recommend to you start from my russian doc
> http://habrahabr.ru/post/179823
>
>
> On Tue, Jun 4, 2013 at 4:22 PM, Явор Маринов <ymari...@neterra.net> wrote:
>
>>  That's the exact documentation which i'm using the directory on ceph2
>> is created, and the service is starting without any problems on both nodes.
>> However the health of the cluster is getting WARN and i was able to mount
>> the cluster
>>
>>
>>
>>
>>  On 06/04/2013 03:43 PM, Andrei Mikhailovsky wrote:
>>
>> Yavor,
>>
>> I would highly recommend taking a look at the quick install guide:
>> http://ceph.com/docs/next/start/quick-start/
>>
>> As per the guide, you need to precreate the directories prior to starting
>> ceph.
>>
>> Andrei
>> ------------------------------
>> *From: *"Явор Маринов" <ymari...@neterra.net> <ymari...@neterra.net>
>> *To: *ceph-users@lists.ceph.com
>> *Sent: *Tuesday, 4 June, 2013 11:03:52 AM
>> *Subject: *[ceph-users] ceph configuration
>>
>>
>> Hello,
>>
>> I'm new to the Ceph mailing list, and I need some advices for our
>> testing cluster. I have 2 servers with x2 hard disks. On the first
>> server i configured monitor and OSD, and on the second server only OSD.
>> The configuration looks like as follows:
>>
>> [mon.a]
>>
>>          host = ceph1
>>          mon addr = 192.168.2.170:6789
>>
>> [osd.0]
>>          host = ceph1
>>          addr = 192.168.2.170
>>          devs = /dev/sdb
>>
>> [osd.1]
>>          host = ceph2
>>          addr = 192.168.2.114
>>          devs = /dev/sdb
>>
>> Once i initiate 'service ceph -a start' i keep getting the following
>> error:
>>
>> Mounting xfs on ceph2:/var/lib/ceph/osd/ceph-1
>> df: `/var/lib/ceph/osd/ceph-1/.': No such file or directory
>>
>> and I'm unable to mount the cluster with the following command:
>> root@ceph1:/mnt# mount -t ceph 192.168.2.170:6789:/ /mnt
>>
>> Also executing 'ceph health' i'm getting this response:
>> HEALTH_WARN 143 pgs degraded; 576 pgs stuck unclean; recovery 15/122
>> degraded (12.295%)
>>
>> This is fresh install and there aren't any nodes which are added/removed.
>>
>> Any help will be much appreciated.
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>
>
>  --
> Igor Laskovy
> facebook.com/igor.laskovy
> studiogrizzly.com
>
>
>


-- 
Igor Laskovy
facebook.com/igor.laskovy
studiogrizzly.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to