Hello,

I have a cluster CEPH (10.2.5-1trusty) I use the various
possibilities: 

-Block 

- Object 

- CephFS


root@ih-par1-cld1-ceph-01:~# cat /etc/ceph/ceph.conf
[....]
mon_host =
10.4.0.1, 10.4.0.3, 10.4.0.5
[....]
public_network =
10.4.0.0/24
cluster_network = 192.168.33.0/24
[....] 

I have dedicated
servers for the storage Block, Object and other servers for CephFS (Full
SSD): 

root@ih-par1-cld1-ceph-01:~# ceph osd tree
ID WEIGHT TYPE NAME
UP/DOWN REWEIGHT PRIMARY-AFFINITY 
 -6 2.79593 root ssdforcephfs 
 -7
0.46599 host ih-prd-cephfs-02 
 32 0.23299 osd.32 up 1.00000 1.00000 

33 0.23299 osd.33 up 1.00000 1.00000 
 -8 0.46599 host ih-prd-cephfs-03

 34 0.23299 osd.34 up 1.00000 1.00000 
 35 0.23299 osd.35 up 1.00000
1.00000 
 -9 0.46599 host ih-prd-cephfs-05 
 36 0.23299 osd.36 up
1.00000 1.00000 
 37 0.23299 osd.37 up 1.00000 1.00000 
-10 0.46599 host
ih-prd-cephfs-01 
 38 0.23299 osd.38 up 1.00000 1.00000 
 39 0.23299
osd.39 up 1.00000 1.00000 
-11 0.46599 host ih-prd-cephfs-04 
 40
0.23299 osd.40 up 1.00000 1.00000 
 41 0.23299 osd.41 up 1.00000 1.00000

-12 0.46599 host ih-prd-cephfs-07 
 42 0.23299 osd.42 up 1.00000
1.00000 
 43 0.23299 osd.43 up 1.00000 1.00000 
 -1 116.47998 root
default 
 -2 43.67999 host ih-par1-cld1-ceph-01 
 0 3.64000 osd.0 up
1.00000 1.00000 
 2 3.64000 osd.2 up 1.00000 1.00000 
 6 3.64000 osd.6
up 1.00000 1.00000 
 8 3.64000 osd.8 up 1.00000 1.00000 
 15 3.64000
osd.15 up 1.00000 1.00000 
 16 3.64000 osd.16 up 1.00000 1.00000 
 19
3.64000 osd.19 up 1.00000 1.00000 
 22 3.64000 osd.22 up 1.00000 1.00000

 24 3.64000 osd.24 up 1.00000 1.00000 
 26 3.64000 osd.26 up 1.00000
1.00000 
 28 3.64000 osd.28 up 1.00000 1.00000 
 30 3.64000 osd.30 up
1.00000 1.00000 
 -3 43.67999 host ih-par1-cld1-ceph-03 
 1 3.64000
osd.1 up 1.00000 1.00000 
 3 3.64000 osd.3 up 1.00000 1.00000 
 5
3.64000 osd.5 up 1.00000 1.00000 
 7 3.64000 osd.7 up 1.00000 1.00000 

13 3.64000 osd.13 up 1.00000 1.00000 
 4 3.64000 osd.4 up 1.00000
1.00000 
 20 3.64000 osd.20 up 1.00000 1.00000 
 23 3.64000 osd.23 up
1.00000 1.00000 
 25 3.64000 osd.25 up 1.00000 1.00000 
 27 3.64000
osd.27 up 1.00000 1.00000 
 29 3.64000 osd.29 up 1.00000 1.00000 
 31
3.64000 osd.31 up 1.00000 1.00000 
 -5 29.12000 host
ih-par1-cld1-ceph-05 
 9 3.64000 osd.9 up 1.00000 1.00000 
 10 3.64000
osd.10 up 1.00000 1.00000 
 11 3.64000 osd.11 up 1.00000 1.00000 
 12
3.64000 osd.12 up 1.00000 1.00000 
 14 3.64000 osd.14 up 1.00000 1.00000

 17 3.64000 osd.17 up 1.00000 1.00000 
 18 3.64000 osd.18 up 1.00000
1.00000 
 21 3.64000 osd.21 up 1.00000 1.00000 

I use OpenNebula for
the use of RDB on the public network: 10.4.0.0/16.

I shall like
separating the RDB, RGW network and CephFS... I have my customers CephFS
who can accèder to all the network RBD, the hypervisors
OpenNebula

Example:

Customer A (CephFS, path: /client1) = > Reaches at
present all the network 10.4.0.0/16
Customer B (CephFS, path: /client2)
= > Reaches at present all the network 10.4.0.0/16

How is it possible
to separate networks: RBD, RGW and have multiple access networks for
CephFS?

I hope to have been clear:/

Thank you 
 
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to