Hi ALL, Anyone can help with this issue would be much appreciated.
I have created an image on one client and mounted it on both 2 client I have setup. When I write data on one client, I cannot access the data on another client, what could be causing this issue? root@nodeB:/mnt# ceph osd tree ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY -1 1.81738 root default -2 0.90869 host nodeB 0 0.90869 osd.0 up 1.00000 1.00000 -3 0.90869 host nodeC 1 0.90869 osd.1 up 1.00000 1.00000 cluster_master@nodeC:/mnt$ ceph osd dump | grep data pool 1 'data' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 128 pgp_num 128 last_change 17 flags hashpspool stripe_width 0 cluster_master@nodeC:/mnt$ cat decompiled-crush-map.txt # begin crush map tunable choose_local_tries 0 tunable choose_local_fallback_tries 0 tunable choose_total_tries 50 tunable chooseleaf_descend_once 1 tunable chooseleaf_vary_r 1 tunable straw_calc_version 1 # devices device 0 osd.0 device 1 osd.1 # types type 0 osd type 1 host type 2 chassis type 3 rack type 4 row type 5 pdu type 6 pod type 7 room type 8 datacenter type 9 region type 10 root # buckets host nodeB { id -2 # do not change unnecessarily # weight 0.909 alg straw hash 0 # rjenkins1 item osd.0 weight 0.909 } host nodeC { id -3 # do not change unnecessarily # weight 0.909 alg straw hash 0 # rjenkins1 item osd.1 weight 0.909 } root default { id -1 # do not change unnecessarily # weight 1.817 alg straw hash 0 # rjenkins1 item nodeB weight 0.909 item nodeC weight 0.909 } # rules rule replicated_ruleset { ruleset 0 type replicated min_size 1 max_size 10 step take default step chooseleaf firstn 0 type host step emit } # end crush map
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com