ceph osd tree please :)

2014-10-29 12:03 GMT+03:00 Vickie CH <mika.leaf...@gmail.com>:

> Dear all,
> Thanks for the reply.
> Pool replicated size is 2. Because the replicated size parameter already
> write into ceph.conf before deploy.
> Because not familiar crush map.  I will according Mark's information to do
> a test that change the crush map to see the result.
>
> -----------ceph.conf------------------
> [global]
> fsid = c404ded6-4086-4f0b-b479-
> 89bc018af954
> mon_initial_members = storage0
> mon_host = 192.168.1.10
> auth_cluster_required = cephx
> auth_service_required = cephx
> auth_client_required = cephx
> filestore_xattr_use_omap = true
>
> *osd_pool_default_size = 2osd_pool_default_min_size = 1*
> osd_pool_default_pg_num = 128
> osd_journal_size = 2048
> osd_pool_default_pgp_num = 128
> osd_mkfs_type = xfs
> -------------------------------------------
>
> ----------------------ceph osd dump result -----------------------------
> pool 0 'data' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 14 flags hashpspool
> crash_replay_interval 45 stripe_width 0
> pool 1 'metadata' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 15 flags hashpspool stripe_width 0
> pool 2 'rbd' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 16 flags hashpspool stripe_width 0
> max_osd 2
>
> ------------------------------------------------------------------------------
>
> Best wishes,
> Mika
>
> Best wishes,
> Mika
>
> 2014-10-29 16:56 GMT+08:00 Mark Kirkwood <mark.kirkw...@catalyst.net.nz>:
>
>> That is not my experience:
>>
>> $ ceph -v
>> ceph version 0.86-579-g06a73c3 (06a73c39169f2f332dec760f56d3ec20455b1646)
>>
>> $ cat /etc/ceph/ceph.conf
>> [global]
>> ...
>> osd pool default size = 2
>>
>> $ ceph osd dump|grep size
>> pool 2 'hot' replicated size 2 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 128 pgp_num 128 last_change 47 flags
>> hashpspool,incomplete_clones tier_of 1 cache_mode writeback target_bytes
>> 2000000000 hit_set bloom{false_positive_probability: 0.05, target_size:
>> 0, seed: 0} 3600s x1 stripe_width 0
>> pool 10 '.rgw.root' replicated size 2 min_size 1 crush_ruleset 0
>> object_hash rjenkins pg_num 8 pgp_num 8 last_change 102 owner
>> 18446744073709551615 flags hashpspool stripe_width 0
>> pool 11 '.rgw.control' replicated size 2 min_size 1 crush_ruleset 0
>> object_hash rjenkins pg_num 8 pgp_num 8 last_change 104 owner
>> 18446744073709551615 flags hashpspool stripe_width 0
>> pool 12 '.rgw' replicated size 2 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 8 pgp_num 8 last_change 106 owner 18446744073709551615
>> flags hashpspool stripe_width 0
>> pool 13 '.rgw.gc' replicated size 2 min_size 1 crush_ruleset 0
>> object_hash rjenkins pg_num 8 pgp_num 8 last_change 107 owner
>> 18446744073709551615 flags hashpspool stripe_width 0
>> pool 14 '.users.uid' replicated size 2 min_size 1 crush_ruleset 0
>> object_hash rjenkins pg_num 8 pgp_num 8 last_change 108 owner
>> 18446744073709551615 flags hashpspool stripe_width 0
>> pool 15 '.rgw.buckets.index' replicated size 2 min_size 1 crush_ruleset 0
>> object_hash rjenkins pg_num 8 pgp_num 8 last_change 110 owner
>> 18446744073709551615 flags hashpspool stripe_width 0
>> pool 16 '.rgw.buckets' replicated size 2 min_size 1 crush_ruleset 0
>> object_hash rjenkins pg_num 8 pgp_num 8 last_change 112 owner
>> 18446744073709551615 flags hashpspool stripe_width 0
>> pool 17 'rbd' replicated size 2 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 1024 pgp_num 1024 last_change 186 flags hashpspool
>> stripe_width 0
>>
>>
>>
>>
>>
>>
>> On 29/10/14 21:46, Irek Fasikhov wrote:
>>
>>> Hi.
>>> This parameter does not apply to pools by default.
>>> ceph osd dump | grep pool. see size=?
>>>
>>>
>>> 2014-10-29 11:40 GMT+03:00 Vickie CH <mika.leaf...@gmail.com
>>> <mailto:mika.leaf...@gmail.com>>:
>>>
>>>     Der Irek:
>>>
>>>     Thanks for your reply.
>>>     Even already set "osd_pool_default_size = 2" the cluster still need
>>>     3 different hosts right?
>>>     Is this default number can be changed by user and write into
>>>     ceph.conf before deploy?
>>>
>>>
>>>     Best wishes,
>>>     Mika
>>>
>>>     2014-10-29 16:29 GMT+08:00 Irek Fasikhov <malm...@gmail.com
>>>     <mailto:malm...@gmail.com>>:
>>>
>>>         Hi.
>>>
>>>         Because the disc requires three different hosts, the default
>>>         number of replications 3.
>>>
>>>         2014-10-29 10:56 GMT+03:00 Vickie CH <mika.leaf...@gmail.com
>>>         <mailto:mika.leaf...@gmail.com>>:
>>>
>>>
>>>             Hi all,
>>>                    Try to use two OSDs to create a cluster. After the
>>>             deply finished, I found the health status is "88
>>>             active+degraded" "104 active+remapped". Before use 2 osds to
>>>             create cluster the result is ok. I'm confuse why this
>>>             situation happened. Do I need to set crush map to fix this
>>>             problem?
>>>
>>>
>>>             ----------ceph.conf---------------------------------
>>>             [global]
>>>             fsid = c404ded6-4086-4f0b-b479-89bc018af954
>>>             mon_initial_members = storage0
>>>             mon_host = 192.168.1.10
>>>             auth_cluster_required = cephx
>>>             auth_service_required = cephx
>>>             auth_client_required = cephx
>>>             filestore_xattr_use_omap = true
>>>             osd_pool_default_size = 2
>>>             osd_pool_default_min_size = 1
>>>             osd_pool_default_pg_num = 128
>>>             osd_journal_size = 2048
>>>             osd_pool_default_pgp_num = 128
>>>             osd_mkfs_type = xfs
>>>             ---------------------------------------------------------
>>>
>>>             -----------ceph -s-----------------------------------
>>>             cluster c404ded6-4086-4f0b-b479-89bc018af954
>>>                   health HEALTH_WARN 88 pgs degraded; 192 pgs stuck
>>> unclean
>>>                   monmap e1: 1 mons at {storage0=192.168.10.10:6789/0
>>>             <http://192.168.10.10:6789/0>}, election epoch 2, quorum 0
>>>             storage0
>>>                   osdmap e20: 2 osds: 2 up, 2 in
>>>                    pgmap v45: 192 pgs, 3 pools, 0 bytes data, 0 objects
>>>                          79752 kB used, 1858 GB / 1858 GB avail
>>>                                88 active+degraded
>>>                               104 active+remapped
>>>             --------------------------------------------------------
>>>
>>>
>>>             Best wishes,
>>>             Mika
>>>
>>>             _______________________________________________
>>>             ceph-users mailing list
>>>             ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
>>>             http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>>
>>>
>>>         --
>>>         С уважением, Фасихов Ирек Нургаязович
>>>         Моб.: +79229045757
>>>
>>>
>>>
>>>
>>>
>>> --
>>> С уважением, Фасихов Ирек Нургаязович
>>> Моб.: +79229045757
>>>
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>
>


-- 
С уважением, Фасихов Ирек Нургаязович
Моб.: +79229045757
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to