Hello,

On Mon, 13 Mar 2017 21:32:45 +0000 James Okken wrote:

> Hi all,
> 
> I have a 3 storage node openstack setup using CEPH.
> I believe that means I have 3 OSDs, as each storage node has a one of 3 fiber 
> channel storage locations mounted.

You use "believe" a lot, so I'm assuming you're quite new and unfamiliar
with Ceph.
Reading docs and google are your friends.

> The storage media behind each node is actually single 7TB HP fiber channel 
> MSA array.
> The best performance configuration for the hard drives in the MSA just 
> happened to be 3x 2.3TB RAID10's. And that matched nicely to the 
> 3xStorageNode/OSD of the CEPH setup.

A pretty unusual approach, not that others (including me) have done similar
things.
Having just 3 OSDs is iffy, since there are corner cases where Ceph may
not be able to distribute PGs (using default parameters) with such a small
pool of OSDs. 

> I believe my replication factor is 3.
> 
You answered that yourself in the dump, a
"ceph osd pool ls detail"
would have been more elegant and yielded the same information.

You also have a min_size of 1, which can be problematic (search the
archives), but with a really small cluster that may still be advantageous.
Lastly, since your OSDs are RAIDs and thus very reliable, a replication of
2 is feasible.

> My question is how much total CEPH storage does this allow me? Only 2.3TB? or 
> does the way CEPH duplicates data enable more than 1/3 of the storage?

3 means 3, so 2.3TB. Note that Ceph is spare, so that can help quite a bit.

> A follow up question would be what is the best way to tell, thru CEPH, the 
> space used and space free? Thanks!!
> 
Well, what do you use with *ix, "df", don'tcha?
"ceph df {detail}"

Christian

> root@node-1:/var/log# ceph osd tree
> ID WEIGHT  TYPE NAME        UP/DOWN REWEIGHT PRIMARY-AFFINITY
> -1 6.53998 root default
> -5 2.17999     host node-28
> 3 2.17999         osd.3         up  1.00000          1.00000
> -6 2.17999     host node-30
> 4 2.17999         osd.4         up  1.00000          1.00000
> -7 2.17999     host node-31
> 5 2.17999         osd.5         up  1.00000          1.00000
> 0       0 osd.0               down        0          1.00000
> 1       0 osd.1               down        0          1.00000
> 2       0 osd.2               down        0          1.00000
> 
> 
> 
> ##
> root@node-1:/var/log# ceph osd lspools
> 0 rbd,2 volumes,3 backups,4 .rgw.root,5 .rgw.control,6 .rgw,7 .rgw.gc,8 
> .users.uid,9 .users,10 compute,11 images,
> 
> 
> 
> ##
> root@node-1:/var/log# ceph osd dump
> epoch 216
> fsid d06d61b0-1cd0-4e1a-ac20-67972d0e1fde
> created 2016-10-11 14:15:05.638099
> modified 2017-03-09 14:45:01.030678
> flags
> pool 0 'rbd' replicated size 3 min_size 1 crush_ruleset 0 object_hash 
> rjenkins pg_num 64 pgp_num 64 last_change 1 flags hashpspool stripe_width 0
> pool 2 'volumes' replicated size 3 min_size 1 crush_ruleset 0 object_hash 
> rjenkins pg_num 64 pgp_num 64 last_change 130 flags hashpspool stripe_width 0
>         removed_snaps [1~5]
> pool 3 'backups' replicated size 3 min_size 1 crush_ruleset 0 object_hash 
> rjenkins pg_num 64 pgp_num 64 last_change 14 flags hashpspool stripe_width 0
> pool 4 '.rgw.root' replicated size 3 min_size 1 crush_ruleset 0 object_hash 
> rjenkins pg_num 64 pgp_num 64 last_change 16 flags hashpspool stripe_width 0
> pool 5 '.rgw.control' replicated size 3 min_size 1 crush_ruleset 0 
> object_hash rjenkins pg_num 64 pgp_num 64 last_change 18 owner 
> 18446744073709551615 flags hashpspool stripe_width 0
> pool 6 '.rgw' replicated size 3 min_size 1 crush_ruleset 0 object_hash 
> rjenkins pg_num 64 pgp_num 64 last_change 20 owner 18446744073709551615 flags 
> hashpspool stripe_width 0
> pool 7 '.rgw.gc' replicated size 3 min_size 1 crush_ruleset 0 object_hash 
> rjenkins pg_num 64 pgp_num 64 last_change 21 flags hashpspool stripe_width 0
> pool 8 '.users.uid' replicated size 3 min_size 1 crush_ruleset 0 object_hash 
> rjenkins pg_num 64 pgp_num 64 last_change 22 owner 18446744073709551615 flags 
> hashpspool stripe_width 0
> pool 9 '.users' replicated size 3 min_size 1 crush_ruleset 0 object_hash 
> rjenkins pg_num 64 pgp_num 64 last_change 24 flags hashpspool stripe_width 0
> pool 10 'compute' replicated size 3 min_size 1 crush_ruleset 0 object_hash 
> rjenkins pg_num 64 pgp_num 64 last_change 216 flags hashpspool stripe_width 0
>         removed_snaps [1~37]
> pool 11 'images' replicated size 3 min_size 1 crush_ruleset 0 object_hash 
> rjenkins pg_num 64 pgp_num 64 last_change 189 flags hashpspool stripe_width 0
>         removed_snaps [1~3,5~8,f~4,14~2,18~2,1c~1,1e~1]
> max_osd 6
> osd.0 down out weight 0 up_from 48 up_thru 50 down_at 52 last_clean_interval 
> [44,45) 192.168.0.9:6800/4485 192.168.1.4:6800/4485 192.168.1.4:6801/4485 
> 192.168.0.9:6801/4485 exists,new
> osd.1 down out weight 0 up_from 10 up_thru 48 down_at 50 last_clean_interval 
> [5,8) 192.168.0.7:6800/60912 192.168.1.6:6801/60912 192.168.1.6:6802/60912 
> 192.168.0.7:6801/60912 exists,new
> osd.2 down out weight 0 up_from 10 up_thru 48 down_at 50 last_clean_interval 
> [5,8) 192.168.0.6:6800/61013 192.168.1.7:6800/61013 192.168.1.7:6801/61013 
> 192.168.0.6:6801/61013 exists,new
> osd.3 up   in  weight 1 up_from 192 up_thru 201 down_at 190 
> last_clean_interval [83,191) 192.168.0.9:6800/2634194 
> 192.168.1.7:6802/3634194 192.168.1.7:6803/3634194 192.168.0.9:6802/3634194 
> exists,up 28b02052-3196-4203-bec8-ac83a69fcbc5
> osd.4 up   in  weight 1 up_from 196 up_thru 201 down_at 194 
> last_clean_interval [80,195) 192.168.0.7:6800/2629319 
> 192.168.1.6:6802/3629319 192.168.1.6:6803/3629319 192.168.0.7:6802/3629319 
> exists,up 124b58e6-1e38-4246-8838-cfc3b88e8a5a
> osd.5 up   in  weight 1 up_from 201 up_thru 201 down_at 199 
> last_clean_interval [134,200) 192.168.0.6:6800/5494 192.168.1.4:6802/1005494 
> 192.168.1.4:6803/1005494 192.168.0.6:6802/1005494 exists,up 
> ddfca14e-e6f6-4c48-aa8f-0ebfc765d32f
> root@node-1:/var/log#
> 
> 
> James Okken
> Lab Manager
> Dialogic Research Inc.
> 4 Gatehall Drive
> Parsippany
> NJ 07054
> USA
> 
> Tel:       973 967 5179
> Email:   james.ok...@dialogic.com<mailto:james.ok...@dialogic.com>
> Web:    www.dialogic.com<http://www.dialogic.com/> - The Network Fuel 
> Company<http://www.dialogic.com/en/landing/itw.aspx>
> 
> This e-mail is intended only for the named recipient(s) and may contain 
> information that is privileged, confidential and/or exempt from disclosure 
> under applicable law. No waiver of privilege, confidence or otherwise is 
> intended by virtue of communication via the internet. Any unauthorized use, 
> dissemination or copying is strictly prohibited. If you have received this 
> e-mail in error, or are not named as a recipient, please immediately notify 
> the sender and destroy all copies of this e-mail.
> 


-- 
Christian Balzer        Network/Systems Engineer                
ch...@gol.com           Global OnLine Japan/Rakuten Communications
http://www.gol.com/
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to