Hi,

Your settings for
osd_max_backfills = 10
osd_recovery_max_active = 10

Are above default and are too high. These limits are per OSD, so you may have a single disk doing 10 backfills ( actually 20 since in and out ) at the same time.

Try dynamic limit it:

ceph tell osd.* injectargs '--osd_max_backfills 1'
ceph tell osd.* injectargs '--osd-recovery-max-active 1'
ceph tell osd.* injectargs '--osd_recovery_sleep 0.2'

Even if the commands return a required reboot message, they do take effect, you can double check by reading from a specific local OSD ( example 1):
ceph daemon osd.1 config get osd_max_backfills

You can also assign less weights the newely added OSDs for example 0.5 of their value to reduce data movement, then re-bump it when the cluster balanced.

/Maged


On 12/01/2019 04:56, Subhachandra Chandra wrote:
Hi,

    We have a cluster with 9 hosts and 540 HDDs using Bluestore and containerized OSDs running luminous 12.2.4. While trying to add new nodes, the cluster collapsed as it could not keep up with establishing enough tcp connections. We fixed sysctl to be able to handle more connections and also recycle tw sockets faster. Currently, as we are trying to restart the cluster by bringing up a few OSDs at a time, some of the OSDs get very busy after around 360 of them come up. iostats show that the busy OSDs are constantly reading from the Bluestore partition. The number of busy OSDs per node vary and norecover is set with no active clients. OSD logs don't show anything other than cephx: verify_authorizer errors which happen on both busy and idle OSDs and doesn't seem to be related to drive reads.

  How can we figure out why the OSDs are busy reading from the drives? If it is some kind of recovery, is there a way to track progress? Output of ceph -s and logs from a busy and idle OSD are copied below.

Thanks
Chandra

Uptime stats with load averages show variance across the9 older nodes.

02:43:44 up 19:21,0 users,load average: 0.88, 1.03, 1.06

02:43:44 up7:58,0 users,load average: 16.91, 13.49, 12.43

02:43:44 up 1 day, 14 min,0 users,load average: 7.67, 6.70, 6.35

02:43:45 up7:01,0 users,load average: 84.40, 84.20, 83.73

02:43:45 up6:40,1 user,load average: 17.08, 17.40, 20.05

02:43:45 up 19:46,0 users,load average: 15.58, 11.93, 11.44

02:43:45 up 20:39,0 users,load average: 7.88, 6.50, 5.69

02:43:46 up 1 day,1:20,0 users,load average: 5.03, 3.81, 3.49

02:43:46 up 1 day, 58 min,0 users,load average: 0.62, 1.00, 1.38


Ceph Config

--------------

[global]

cluster network = 192.168.13.0/24 <http://192.168.13.0/24>

fsid = <>

mon host = 172.16.13.101,172.16.13.102,172.16.13.103

mon initial members = ctrl1,ctrl2,ctrl3

mon_max_pg_per_osd = 750

mon_osd_backfillfull_ratio = 0.92

mon_osd_down_out_interval = 900

mon_osd_full_ratio = 0.95

mon_osd_nearfull_ratio = 0.85

osd_crush_chooseleaf_type = 3

osd_heartbeat_grace = 900

mon_osd_laggy_max_interval = 900

osd_max_pg_per_osd_hard_ratio = 1.0

public network = 172.16.13.0/24 <http://172.16.13.0/24>


[mon]

mon_compact_on_start = true


[osd]

osd_deep_scrub_interval = 2419200

osd_deep_scrub_stride = 4194304

osd_max_backfills = 10

osd_max_object_size = 276824064

osd_max_scrubs = 1

osd_max_write_size = 264

osd_pool_erasure_code_stripe_unit = 2097152

osd_recovery_max_active = 10

osd_heartbeat_interval = 15


Data nodes Sysctl params

-----------------------------

fs.aio-max-nr=1048576

kernel.pid_max=4194303

kernel.threads-max=2097152

net.core.netdev_max_backlog=65536

net.core.optmem_max=1048576

net.core.rmem_max=8388608

net.core.rmem_default=8388608

net.core.somaxconn=2048

net.core.wmem_max=8388608

net.core.wmem_default=8388608

vm.max_map_count=524288

vm.min_free_kbytes=262144


net.ipv4.tcp_tw_reuse=1

net.ipv4.tcp_max_syn_backlog=16384

net.ipv4.tcp_fin_timeout=10

net.ipv4.tcp_slow_start_after_idle=0



Ceph -s output

---------------


root@ctrl1:/# ceph -s

cluster:

id: 06126476-6deb-4baa-b7ca-50f5ccfacb68

health: HEALTH_ERR

noout,nobackfill,norebalance,norecover,noscrub,nodeep-scrub flag(s) set

704 osds down

9 hosts (540 osds) down

71 nearfull osd(s)

2 pool(s) nearfull

780664/74163111 objects misplaced (1.053%)

7724/8242239 objects unfound (0.094%)

396 PGs pending on creation

Reduced data availability: 32597 pgs inactive, 29764 pgs down, 820 pgs peering, 74 pgs incomplete, 1 pg stale

Degraded data redundancy: 679158/74163111 objects degraded (0.916%), 1250 pgs degraded, 1106 pgs undersized

33 slow requests are blocked > 32 sec

9 stuck requests are blocked > 4096 sec

mons ctrl1,ctrl2,ctrl3 are using a lot of disk space

services:

mon: 3 daemons, quorum ctrl1,ctrl2,ctrl3

mgr: ctrl1(active), standbys: ctrl2, ctrl3

osd: 1080 osds: 376 up, 1080 in; 1963 remapped pgs

flags noout,nobackfill,norebalance,norecover,noscrub,nodeep-scrub

data:

pools: 2 pools, 33280 pgs

objects: 8049k objects, 2073 TB

usage: 2277 TB used, 458 TB / 2736 TB avail

pgs: 3.585% pgs unknown

94.363% pgs not active

679158/74163111 objects degraded (0.916%)

780664/74163111 objects misplaced (1.053%)

7724/8242239 objects unfound (0.094%)

29754 down

1193unknown

535 peering

496 activating+undersized+degraded+remapped

284 remapped+peering

258 active+undersized+degraded+remapped

161 activating+degraded+remapped

143 active+recovering+undersized+degraded+remapped

89active+undersized+degraded

76active+clean+remapped

71incomplete

48active+undersized+remapped

46undersized+degraded+peered

34active+recovering+degraded+remapped

26active+clean

21activating+remapped

13activating+undersized+degraded

10down+remapped

9 active+recovery_wait+undersized+degraded+remapped

4 activating

3 remapped+incomplete

2 activating+undersized+remapped

1 stale+peering

1 undersized+peered

1 undersized+remapped+peered

1 activating+degraded

root@ctrl1:/# ceph version

ceph version 12.2.4 (52085d5249a80c5f5121a76d6288429f35e4e77b) luminous (stable)


osd.86 - busy
-------------------

2019-01-12 02:30:02.582363 7f00ea0f47000 osd.86 213711 No AuthAuthorizeHandler found for protocol 0

2019-01-12 02:30:02.582383 7f00ea0f47000 -- 192.168.13.5:6806/351612 <http://192.168.13.5:6806/351612> >> 192.168.13.8:6868/770039 <http://192.168.13.8:6868/770039> conn(0x55c4a8d89000 :6806 s=STATE_ACCEPTING_WAIT_CONNECT_MSG_AUTH pgs=0 cs=0 l=0).handle_connect_msg: got bad authorizer

2019-01-12 02:30:04.005541 7f00ea8f57000 auth: could not find secret_id=7544

2019-01-12 02:30:04.005554 7f00ea8f57000 cephx: verify_authorizer could not get service secret for service osd secret_id=7544

2019-01-12 02:30:04.005557 7f00ea8f57000 -- 192.168.13.5:6806/351612 <http://192.168.13.5:6806/351612> >> 192.168.13.3:6836/405613 <http://192.168.13.3:6836/405613> conn(0x55c4f3617800 :6806 s=STATE_ACCEPTING_WAIT_CONNECT_MSG_AUTH pgs=0 cs=0 l=0).handle_connect_msg: got bad authorizer

2019-01-12 02:30:07.910816 7f00e98f37000 auth: could not find secret_id=7550

2019-01-12 02:30:07.910864 7f00e98f37000 cephx: verify_authorizer could not get service secret for service osd secret_id=7550

2019-01-12 02:30:07.910884 7f00e98f37000 -- 192.168.13.5:6806/351612 <http://192.168.13.5:6806/351612> >> 192.168.13.8:6824/767660 <http://192.168.13.8:6824/767660> conn(0x55c4ec462800 :6806 s=STATE_ACCEPTING_WAIT_CONNECT_MSG_AUTH pgs=0 cs=0 l=0).handle_connect_msg: got bad authorizer

2019-01-12 02:30:16.982636 7f00ea8f57000 osd.86 213711 No AuthAuthorizeHandler found for protocol 0

2019-01-12 02:30:16.982640 7f00ea8f57000 -- 192.168.13.5:6806/351612 <http://192.168.13.5:6806/351612> >> 192.168.13.6:6816/349322 <http://192.168.13.6:6816/349322> conn(0x55c4aee6f000 :6806 s=STATE_ACCEPTING_WAIT_CONNECT_MSG_AUTH pgs=0 cs=0 l=0).handle_connect_msg: got bad authorizer


OSD.132 - idle

---------------

2019-01-12 02:31:38.370478 7fa5084507000 auth: could not find secret_id=7551

2019-01-12 02:31:38.370487 7fa5084507000 cephx: verify_authorizer could not get service secret for service osd secret_id=7551

2019-01-12 02:31:38.370489 7fa5084507000 -- 192.168.13.5:6854/356774 <http://192.168.13.5:6854/356774> >> 192.168.13.8:6872/1201589 <http://192.168.13.8:6872/1201589> conn(0x563b3e46f000 :6854 s=STATE_ACCEPTING_WAIT_CONNECT_MSG_AUTH pgs=0 cs=0 l=0).handle_connect_msg: got bad authorizer

2019-01-12 02:31:41.121603 7fa5094527000 auth: could not find secret_id=7544

2019-01-12 02:31:41.121672 7fa5094527000 cephx: verify_authorizer could not get service secret for service osd secret_id=7544

2019-01-12 02:31:41.121707 7fa5094527000 -- 192.168.13.5:6854/356774 <http://192.168.13.5:6854/356774> >> 192.168.13.9:6808/515991 <http://192.168.13.9:6808/515991> conn(0x563b53997800 :6854 s=STATE_ACCEPTING_WAIT_CONNECT_MSG_AUTH pgs=0 cs=0 l=0).handle_connect_msg: got bad authorizer


This email message, including attachments, may contain private, proprietary, or privileged information and is the confidential information and/or property of GRAIL, Inc., and is for the sole use of the intended recipient(s). Any unauthorized review, use, disclosure or distribution is strictly prohibited. If you are not the intended recipient, please contact the sender by reply email and destroy all copies of the original message.


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

--
Maged Mokhtar
CEO PetaSAN
4 Emad El Deen Kamel
Cairo 11371, Egypt
www.petasan.org
+201006979931
skype: maged.mokhtar

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to