I've not defined cluster IPs for each OSD server but only the whole subnet.
Should I define each IP for each OSD ? This is not wrote on docs and
could be tricky to do this in big environments with hundreds of nodes

2014-04-24 20:04 GMT+02:00 McNamara, Bradley <bradley.mcnam...@seattle.gov>:
> Do you have all of the cluster IP's defined in the host file on each OSD 
> server?  As I understand it, the mon's do not use a cluster network, only the 
> OSD servers.
>
> -----Original Message-----
> From: ceph-users-boun...@lists.ceph.com 
> [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Gandalf Corvotempesta
> Sent: Thursday, April 24, 2014 8:54 AM
> To: ceph-users@lists.ceph.com
> Subject: [ceph-users] cluster_network ignored
>
> I'm trying to configure a small ceph cluster with both public and cluster 
> networks.
> This is my conf:
>
> [global]
>   public_network = 192.168.0/24
>   cluster_network = 10.0.0.0/24
>   auth cluster required = cephx
>   auth service required = cephx
>   auth client required = cephx
>   fsid = 004baba0-74dc-4429-84ec-1e376fb7bcad
>   osd pool default pg num = 8192
>   osd pool default pgp num = 8192
>   osd pool default size = 3
>
> [mon]
>   mon osd down out interval = 600
>   mon osd mon down reporters = 7
>   [mon.osd1]
>     host = osd1
>     mon addr = 192.168.0.1
>   [mon.osd2]
>     host = osd2
>     mon addr = 192.168.0.2
>   [mon.osd3]
>     host = osd3
>     mon addr = 192.168.0.3
>
> [osd]
>   osd mkfs type = xfs
>   osd journal size = 16384
>   osd mon heartbeat interval = 30
>   filestore merge threshold = 40
>   filestore split multiple = 8
>   osd op threads = 8
>   osd recovery max active = 5
>   osd max backfills = 2
>   osd recovery op priority = 2
>
>
> on each node I have bond0 bound to 192.168.0.x and bond1 bound to 10.0.0.x 
> When ceph is doing recovery, I can see replication through bond0 (public 
> interface) and nothing via bond1 (cluster interface)
>
> Should I configure anything else ?
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to