You're right. I checked the other servers, and noticed the server that has
problem is just one. the others are all right.
I can't figure what is different. All servers are copied from one VM. I
don't know why that happended.

I'll install all again from the scratch, and will report on this thread
after.



I've been watching my every step since 1995.
http://kirrie.pe.kr


2013/9/3 Joao Eduardo Luis <joao.l...@inktank.com>

> On 09/03/2013 02:02 AM, 이주헌 wrote:
>
>> Hi all.
>>
>> I have 1 MDS and 3 OSDs. I installed them via ceph-deploy. (dumpling
>> 0.67.2 version)
>> At first, It works perfectly. But, after I reboot one of OSD, ceph-mon
>> launched on port 6800 not 6789.
>>
>
> This has been a recurrent issue I've been completely unable to reproduce
> so far.
>
> Are you able to reproduce this reliably?
>
> Could you share the steps you took leading you to this state?
>
>   -Joao
>
>
>> This is a result of 'ceph -s'
>> ---
>>    cluster c59d13fd-c4c9-4cd0-b2ed-**b654428b3171
>>     health HEALTH_WARN 1 mons down, quorum 0,1,2
>> ceph-mds,ceph-osd0,ceph-osd1
>>     monmap e1: 4 mons at
>> {ceph-mds=192.168.13.135:6789/**0,ceph-osd0=192.168.13.136:**
>> 6789/0,ceph-osd1=192.168.13.**137:6789/0,ceph-osd2=192.168.**
>> 13.138:6789/0<http://192.168.13.135:6789/0,ceph-osd0=192.168.13.136:6789/0,ceph-osd1=192.168.13.137:6789/0,ceph-osd2=192.168.13.138:6789/0>
>> <http://192.168.13.135:6789/0,**ceph-osd0=192.168.13.136:6789/**
>> 0,ceph-osd1=192.168.13.137:**6789/0,ceph-osd2=192.168.13.**138:6789/0<http://192.168.13.135:6789/0,ceph-osd0=192.168.13.136:6789/0,ceph-osd1=192.168.13.137:6789/0,ceph-osd2=192.168.13.138:6789/0>
>> >},
>>
>> election epoch 206, quorum 0,1,2 ceph-mds,ceph-osd0,ceph-osd1
>>     osdmap e22: 4 osds: 2 up, 2 in
>>      pgmap v67: 192 pgs: 192 active+clean; 145 MB data, 2414 MB used,
>> 18043 MB / 20458 MB avail
>>     mdsmap e4: 1/1/1 up {0=ceph-mds=up:active}
>> ---
>> 1 mons down -> It's running on 6800.
>>
>> This is a /etc/ceph/ceph.conf that is created automatically by
>> ceph-deploy.
>> ---
>> [global]
>> fsid = c59d13fd-c4c9-4cd0-b2ed-**b654428b3171
>> mon_initial_members = ceph-mds, ceph-osd0, ceph-osd1, ceph-osd2
>> mon_host = 192.168.13.135,192.168.13.136,**192.168.13.137,192.168.13.138
>> auth_supported = cephx
>> osd_journal_size = 1024
>> filestore_xattr_use_omap = true
>> ---
>> According to my understanding, ceph-mon's default port is 6789.
>> Why does it run on 6800 instead of 6789?
>> Restarting ceph-mon has a same result.
>>
>> Sorry for my poor english. I don't write or speak english fluently.
>>
>>
>> ______________________________**_________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/**listinfo.cgi/ceph-users-ceph.**com<http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>>
>>
>
> --
> Joao Eduardo Luis
> Software Engineer | http://inktank.com | http://ceph.com
> ______________________________**_________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/**listinfo.cgi/ceph-users-ceph.**com<http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to