It's nice to hear that. You can also decrease the osd ram usage from
4gb to 2gb. If you have enough spare ram go for it.
Good luck.

Lee <lqui...@gmail.com>, 6 Oca 2022 Per, 00:46 tarihinde şunu yazdı:
>
> I'm not rushing,
>
> I have found the issue, Im am getting OOM errors as the OSD boots, basically 
> is starts to process the PG's and then the node runs out of memory and the 
> daemon kill's
>
> 2022-01-05 20:09:08 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:01.024+0000 7fce3c6bc700 10 osd.51 24448261 tick
> 2022-01-05 20:09:10 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:01.060+0000 7fce3b441700 10 osd.51 24448261 
> tick_without_osd_lock
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:02.268+0000 7fce3c6bc700 10 osd.51 24448261 do_waiters -- 
> start
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:09.544+0000 7fce3c6bc700 10 osd.51 24448261 do_waiters -- 
> finish
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:10.260+0000 7fce1e407700  5 osd.51 24448261 heartbeat 
> osd_stat(store_statfs(0x2258948000/0x40000000/0x3a38800000, data 
> 0x17919fd8c4/0x179feb4000, compress 0x0/0x0/0x0, omap 0xc9773, meta 
> 0x3ff3688d), peers [] op hist [])
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:13.060+0000 7fce3c6bc700 20 osd.51 24448261 tick 
> last_purged_snaps_scrub 2022-01-04T22:29:39.121925+0000 next 
> 2022-01-05T22:29:39.121925+0000
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:13.104+0000 7fce1e407700 20 osd.51 24448261 
> check_full_status cur ratio 0.410072, physical ratio 0.410072, new state none
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:13.108+0000 7fce34c34700 20 
> bluestore(/var/lib/ceph/osd/ceph-51) deferred_try_submit 0 osrs, 0 txcs
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:13.108+0000 7fce34c34700  5 
> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size: 134217728 
> kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864 meta_used: 75234 
> data_alloc: 67108864 data_used: 0
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:13.160+0000 7fce34c34700 20 
> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size: 134217728 
> kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864 meta_used: 75234 
> data_alloc: 67108864 data_used: 0
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:13.216+0000 7fce34c34700 20 
> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size: 134217728 
> kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864 meta_used: 75234 
> data_alloc: 67108864 data_used: 0
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:13.264+0000 7fce34c34700 20 
> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size: 134217728 
> kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864 meta_used: 75234 
> data_alloc: 67108864 data_used: 0
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:13.400+0000 7fce34c34700 20 
> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size: 134217728 
> kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864 meta_used: 75234 
> data_alloc: 67108864 data_used: 0
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:13.536+0000 7fce34c34700 20 
> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size: 134217728 
> kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864 meta_used: 75234 
> data_alloc: 67108864 data_used: 0
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:13.640+0000 7fce34c34700 20 
> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size: 134217728 
> kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864 meta_used: 75234 
> data_alloc: 67108864 data_used: 0
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:13.644+0000 7fce1e407700  5 osd.51 24448261 heartbeat 
> osd_stat(store_statfs(0x2258948000/0x40000000/0x3a38800000, data 
> 0x17919fd8c4/0x179feb4000, compress 0x0/0x0/0x0, omap 0xc9773, meta 
> 0x3ff3688d), peers [] op hist [])
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:13.712+0000 7fce34c34700 20 
> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size: 134217728 
> kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864 meta_used: 75234 
> data_alloc: 67108864 data_used: 0
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:13.688+0000 7fce1e407700 20 osd.51 24448261 
> check_full_status cur ratio 0.410072, physical ratio 0.410072, new state none
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:13.480+0000 7fce3b441700 20 
> bluestore(/var/lib/ceph/osd/ceph-51) statfs 
> store_statfs(0x2258948000/0x40000000/0x3a38800000, data 
> 0x17919fd8c4/0x179feb4000, compress 0x0/0x0/0x0, omap 0xc9773, meta 
> 0x3ff3688d)
> 2022-01-05 20:09:13 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:13.844+0000 7fce34c34700 20 
> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size: 134217728 
> kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864 meta_used: 75234 
> data_alloc: 67108864 data_used: 0
> 2022-01-05 20:09:14 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:14.016+0000 7fce34c34700 20 
> bluestore.MempoolThread(0x55f42e762a98) _resize_shards cache_size: 134217728 
> kv_alloc: 67108864 kv_used: 67082912 meta_alloc: 67108864 meta_used: 75234 
> data_alloc: 67108864 data_used: 0
> 2022-01-05 20:09:38 bb-ceph-enc-rm63-osd03-31 osd.51 
> 2022-01-05T20:09:14.104+0000 7fce3c6bc700 10 osd.51 24448261 tick
> 2022-01-05 20:10:37 bb-ceph-enc-rm63-osd03-31 init.scope ceph-osd@51.service: 
> Main process exited, code=killed, status=9/KILL
> 2022-01-05 20:10:37 bb-ceph-enc-rm63-osd03-31 init.scope ceph-osd@51.service: 
> Failed with result 'signal'.
> 2022-01-05 20:10:47 bb-ceph-enc-rm63-osd03-31 init.scope ceph-osd@51.service: 
> Scheduled restart job, restart counter is at 1.
> 2022-01-05 20:10:47 bb-ceph-enc-rm63-osd03-31 init.scope Stopped Ceph object 
> storage daemon osd.51.
>
> I have just increased the RAM physically in one of the node's removed the 
> other OSD's physically for now and managed to get one of the 3 down to come 
> up. Just stepping through each at the moment.
>
> Regards
>
> Lee
>
> On Wed, 5 Jan 2022 at 21:10, mhnx <morphinwith...@gmail.com> wrote:
>>
>> First of all, do not rush into bad decisions.
>> Production is down and you wanna make it online but you should fix the
>> problem and be sure first. If a second crash occurs in a healing state
>> you will lose metadata.
>> You don't need to debug first!
>>
>> You didn't mention your cluster status and we don't know what you have.
>> We need some information;
>> 1- ceph -s
>> 2- ceph health detail
>> 3- ceph df
>> 4- tail /var/log/ceph/ceph-osd{crashed osd number}.log -n 1000
>>
>>
>>
>> Lee <lqui...@gmail.com>, 5 Oca 2022 Çar, 23:14 tarihinde şunu yazdı:
>> >
>> > Looking for some help as this is production effecting..
>> >
>> > We run a 3 Node cluster with a mix of 5xSSD,15xSATA and 5xSAS in each node.
>> > Running 15.2.15. All using DB/WAL on NVME SSD except the SSD's
>> >
>> > Earlier today I increased the PG num from 32 to 128 on one of our pools,
>> > due to the status complaining. pretty normally really. 2-3 mins in I
>> > watched in horror as SSD based OSD's crashed on all 3 nodes, refusing to
>> > restart.
>> >
>> > I've set debug_bluefs and bluestore to 20 it will get so far and then the
>> > daemon fails.
>> >
>> > 2022-01-05 19:39:23 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:23.335+0000 7f2794383700 20
>> > bluestore(/var/lib/ceph/osd/ceph-51) deferred_try_submit 0 osrs, 0 txcs
>> > 2022-01-05 19:39:23 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:23.335+0000 7f2794383700  5
>> > bluestore.MempoolThread(0x560433f0aa98) _resize_shards cache_size:
>> > 134217728 kv_alloc: 67108864 kv_used: 67075728 meta_alloc: 67108864
>> > meta_used: 75234 data_alloc: 67108864 data_used: 0
>> > 2022-01-05 19:39:23 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:23.387+0000 7f2794383700 20
>> > bluestore.MempoolThread(0x560433f0aa98) _resize_shards cache_size:
>> > 134217728 kv_alloc: 67108864 kv_used: 67075728 meta_alloc: 67108864
>> > meta_used: 75234 data_alloc: 67108864 data_used: 0
>> > 2022-01-05 19:39:23 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:23.467+0000 7f2794383700 20
>> > bluestore.MempoolThread(0x560433f0aa98) _resize_shards cache_size:
>> > 134217728 kv_alloc: 67108864 kv_used: 67075728 meta_alloc: 67108864
>> > meta_used: 75234 data_alloc: 67108864 data_used: 0
>> > 2022-01-05 19:39:24 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:23.979+0000 7f2794383700 20
>> > bluestore.MempoolThread(0x560433f0aa98) _resize_shards cache_size:
>> > 134217728 kv_alloc: 67108864 kv_used: 67075728 meta_alloc: 67108864
>> > meta_used: 75234 data_alloc: 67108864 data_used: 0
>> > 2022-01-05 19:39:24 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:24.167+0000 7f2794383700 20
>> > bluestore.MempoolThread(0x560433f0aa98) _resize_shards cache_size:
>> > 134217728 kv_alloc: 67108864 kv_used: 67075728 meta_alloc: 67108864
>> > meta_used: 75234 data_alloc: 67108864 data_used: 0
>> > 2022-01-05 19:39:24 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:24.271+0000 7f2794383700 20
>> > bluestore.MempoolThread(0x560433f0aa98) _resize_shards cache_size:
>> > 134217728 kv_alloc: 67108864 kv_used: 67075728 meta_alloc: 67108864
>> > meta_used: 75234 data_alloc: 67108864 data_used: 0
>> > 2022-01-05 19:39:24 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:24.327+0000 7f2794383700 20
>> > bluestore.MempoolThread(0x560433f0aa98) _resize_shards cache_size:
>> > 134217728 kv_alloc: 67108864 kv_used: 67075728 meta_alloc: 67108864
>> > meta_used: 75234 data_alloc: 67108864 data_used: 0
>> > 2022-01-05 19:39:32 bb-ceph-enc-rm63-osd03-31 init.scope 
>> > ceph-osd@51.service:
>> > Main process exited, code=killed, status=9/KILL
>> > 2022-01-05 19:39:32 bb-ceph-enc-rm63-osd03-31 init.scope 
>> > ceph-osd@51.service:
>> > Failed with result 'signal'.
>> > 2022-01-05 19:39:42 bb-ceph-enc-rm63-osd03-31 init.scope 
>> > ceph-osd@51.service:
>> > Scheduled restart job, restart counter is at 1.
>> >
>> > I've run
>> > ceph-bluestore-tool bluefs-bdev-sizes --path /var/lib/ceph/osd/ceph-51
>> > inferring bluefs devices from bluestore path
>> > 1 : device size 0x3a38800000 : own 0x[1bf2200000~254300000] = 0x254300000 :
>> > using 0x3fd10000(1021 MiB) : bluestore has 0x1d83400000(118 GiB) available
>> >
>> > Also fsck and repair all seems to be ok.
>> >
>> > The normal log looks like
>> >
>> > 2022-01-05 19:39:42 bb-ceph-enc-rm63-osd03-31 init.scope Starting Ceph
>> > object storage daemon osd.51...
>> > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:46.467+0000 7fca32943e00  0 set uid:gid to 64045:64045
>> > (ceph:ceph)
>> > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:46.467+0000 7fca32943e00  0 ceph version 15.2.15
>> > (2dfb18841cfecc2f7eb7eb2afd65986ca4d95985) octopus (stable), process
>> > ceph-osd, pid 139577
>> > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:46.467+0000 7fca32943e00  0 pidfile_write: ignore empty
>> > --pid-file
>> > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bdev create path
>> > /var/lib/ceph/osd/ceph-51/block type kernel
>> > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bdev(0x55b4b234e000
>> > /var/lib/ceph/osd/ceph-51/block) open path /var/lib/ceph/osd/ceph-51/block
>> > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bdev(0x55b4b234e000
>> > /var/lib/ceph/osd/ceph-51/block) open size 250056015872 (0x3a38800000, 233
>> > GiB) block_size 4096 (4 KiB) non-rotational discard not supported
>> > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:46.491+0000 7fca32943e00  1
>> > bluestore(/var/lib/ceph/osd/ceph-51) _set_cache_sizes cache_size 1073741824
>> > meta 0.4 kv 0.4 data 0.2
>> > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bdev create path
>> > /var/lib/ceph/osd/ceph-51/block type kernel
>> > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bdev(0x55b4b234e380
>> > /var/lib/ceph/osd/ceph-51/block) open path /var/lib/ceph/osd/ceph-51/block
>> > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bdev(0x55b4b234e380
>> > /var/lib/ceph/osd/ceph-51/block) open size 250056015872 (0x3a38800000, 233
>> > GiB) block_size 4096 (4 KiB) non-rotational discard not supported
>> > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bluefs add_block_device bdev 1
>> > path /var/lib/ceph/osd/ceph-51/block size 233 GiB
>> > 2022-01-05 19:39:46 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:46.491+0000 7fca32943e00  1 bdev(0x55b4b234e380
>> > /var/lib/ceph/osd/ceph-51/block) close
>> > 2022-01-05 19:39:47 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:47.067+0000 7fca32943e00  0 starting osd.51 osd_data
>> > /var/lib/ceph/osd/ceph-51 /var/lib/ceph/osd/ceph-51/journal
>> > 2022-01-05 19:39:47 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:47.159+0000 7fca32943e00  0 load: jerasure load: lrc load:
>> > isa
>> > 2022-01-05 19:39:47 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:47.159+0000 7fca32943e00  1 bdev create path
>> > /var/lib/ceph/osd/ceph-51/block type kernel
>> > 2022-01-05 19:39:47 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:47.159+0000 7fca32943e00  1 bdev(0x55b4b234e000
>> > /var/lib/ceph/osd/ceph-51/block) open path /var/lib/ceph/osd/ceph-51/block
>> > 2022-01-05 19:39:47 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:47.163+0000 7fca32943e00  1 bdev(0x55b4b234e000
>> > /var/lib/ceph/osd/ceph-51/block) open size 250056015872 (0x3a38800000, 233
>> > GiB) block_size 4096 (4 KiB) non-rotational discard not supported
>> > 2022-01-05 19:39:47 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:47.163+0000 7fca32943e00  1
>> > bluestore(/var/lib/ceph/osd/ceph-51) _set_cache_sizes cache_size 1073741824
>> > meta 0.4 kv 0.4 data 0.2
>> > 2022-01-05 19:39:47 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:47.163+0000 7fca32943e00  1 bdev(0x55b4b234e000
>> > /var/lib/ceph/osd/ceph-51/block) close
>> > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:48.619+0000 7fca32943e00  1
>> > bluestore(/var/lib/ceph/osd/ceph-51) _open_alloc loaded 138 GiB in 276582
>> > extents available 129 GiB
>> > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:48.619+0000 7fca32943e00  1 bluefs umount
>> > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:48.619+0000 7fca32943e00  1 bdev(0x55b4b234e380
>> > /var/lib/ceph/osd/ceph-51/block) close
>> > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:48.803+0000 7fca32943e00  1 bdev create path
>> > /var/lib/ceph/osd/ceph-51/block type kernel
>> > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:48.803+0000 7fca32943e00  1 bdev(0x55b4b234e380
>> > /var/lib/ceph/osd/ceph-51/block) open path /var/lib/ceph/osd/ceph-51/block
>> > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:48.803+0000 7fca32943e00  1 bdev(0x55b4b234e380
>> > /var/lib/ceph/osd/ceph-51/block) open size 250056015872 (0x3a38800000, 233
>> > GiB) block_size 4096 (4 KiB) non-rotational discard not supported
>> > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:48.803+0000 7fca32943e00  1 bluefs add_block_device bdev 1
>> > path /var/lib/ceph/osd/ceph-51/block size 233 GiB
>> > 2022-01-05 19:39:48 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:48.803+0000 7fca32943e00  1 bluefs mount
>> > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:49.087+0000 7fca32943e00  1
>> > bluestore(/var/lib/ceph/osd/ceph-51) _open_db opened rocksdb path db
>> > options
>> > compression=kNoCompression,max_write_buffer_number=4,min_write_buffer_number_to_merge=1,recycle_log_file_num=4,write_buffer_size=268435456,writable_file_max_buffer_size=0,compaction_readahead_size=2097152,max_background_compactions=2
>> > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:49.087+0000 7fca32943e00  1
>> > bluestore(/var/lib/ceph/osd/ceph-51) _upgrade_super from 4, latest 4
>> > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:49.087+0000 7fca32943e00  1
>> > bluestore(/var/lib/ceph/osd/ceph-51) _upgrade_super done
>> > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:49.131+0000 7fca32943e00  0
>> >  /build/ceph-15.2.15/src/cls/cephfs/cls_cephfs.cc:198: loading cephfs
>> > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:49.131+0000 7fca32943e00  0
>> >  /build/ceph-15.2.15/src/cls/hello/cls_hello.cc:312: loading cls_hello
>> > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:49.135+0000 7fca32943e00  0 _get_class not permitted to
>> > load kvs
>> > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:49.171+0000 7fca32943e00  0 _get_class not permitted to
>> > load lua
>> > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:49.207+0000 7fca32943e00  0 _get_class not permitted to
>> > load queue
>> > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:49.319+0000 7fca32943e00  0 _get_class not permitted to
>> > load sdk
>> > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:49.319+0000 7fca32943e00  0 osd.51 24448261 crush map has
>> > features 288514051259236352, adjusting msgr requires for clients
>> > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:49.319+0000 7fca32943e00  0 osd.51 24448261 crush map has
>> > features 288514051259236352 was 8705, adjusting msgr requires for mons
>> > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:49.319+0000 7fca32943e00  0 osd.51 24448261 crush map has
>> > features 3314933000852226048, adjusting msgr requires for osds
>> > 2022-01-05 19:39:49 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:39:49.319+0000 7fca32943e00  1 osd.51 24448261
>> > check_osdmap_features require_osd_release unknown -> octopus
>> > 2022-01-05 19:41:25 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:41:24.999+0000 7fca32943e00  0 osd.51 24448261 load_pgs
>> > opened 66 pgs
>> > 2022-01-05 19:41:25 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:41:25.071+0000 7fca32943e00 -1 osd.51 24448261
>> > log_to_monitors {default=true}
>> > 2022-01-05 19:41:25 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:41:25.071+0000 7fca32943e00 -1 osd.51 24448261
>> > log_to_monitors {default=true}
>> > 2022-01-05 19:42:16 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:42:16.631+0000 7fca32943e00  0 osd.51 24448261 done with
>> > init, starting boot process
>> > 2022-01-05 19:42:16 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:42:16.631+0000 7fca32943e00  1 osd.51 24448261 start_boot
>> > 2022-01-05 19:42:16 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:42:16.635+0000 7fca14615700  1 osd.51 pg_epoch: 24448130
>> > pg[44.17( v 24448128'27126321 (24447767'27121032,24448128'27126321]
>> > local-lis/les=24447864/24447865 n=2356 ec=4550661/4550661
>> > lis/c=24447864/24447864 les/c/f=24447865/24447865/22709931 sis=24448130)
>> > [51,48,15] r=0 lpr=24448130 pi=[24447864,24448130)/1 crt=24448128'27126321
>> > lcod 0'0 mlcod 0'0 unknown mbc={}] start_peering_interval up [51,48,15] ->
>> > [51,48,15], acting [51,48,15] -> [51,48,15], acting_primary 51 -> 51,
>> > up_primary 51 -> 51, role 0 -> 0, features acting 4540138292840890367
>> > upacting 4540138292840890367
>> > 2022-01-05 19:42:16 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:42:16.635+0000 7fca13613700  1 osd.51 pg_epoch: 24448130
>> > pg[44.1( v 24448129'31648690 (24447777'31643388,24448129'31648690]
>> > local-lis/les=24447865/24447866 n=2314 ec=4550661/4550661
>> > lis/c=24447865/24447865 les/c/f=24447866/24447866/22709931 sis=24448130)
>> > [51,15,5] r=0 lpr=24448130 pi=[24447865,24448130)/1 crt=24448129'31648690
>> > lcod 0'0 mlcod 0'0 unknown mbc={}] start_peering_interval up [51,15,5] ->
>> > [51,15,5], acting [51,15,5] -> [51,15,5], acting_primary 51 -> 51,
>> > up_primary 51 -> 51, role 0 -> 0, features acting 4540138292840890367
>> > upacting 4540138292840890367
>> > 2022-01-05 19:42:16 bb-ceph-enc-rm63-osd03-31 osd.51
>> > 2022-01-05T19:42:16.635+0000 7fca15617700  1 osd.51 pg_epoch: 24448130
>> > pg[44.15( v 24448129'37939392 (24447777'37936883,24448129'37939392]
>> > local-lis/les=24448118/24448119 n=2350 ec=4550661/4550661
>> > lis/c=24448118/24448118 les/c/f=24448119/24448119/22709931 sis=24448130)
>> > [5,14,51] r=2 lpr=24448130 pi=[24448118,24448130)/1 crt=24448129'37939392
>> > lcod 0'0 mlcod 0'0 unknown mbc={}] start_peering_interval up [5,14,51] ->
>> > [5,14,51], acting [5,14,51] -> [5,14,51], acting_primary 5 -> 5, up_primary
>> > 5 -> 5, role 2 -> 2, features acting 4540138292840890367 upacting
>> > 4540138292840890367
>> > 2022-01-05 19:42:51 bb-ceph-enc-rm63-osd03-31 init.scope 
>> > ceph-osd@51.service:
>> > Main process exited, code=killed, status=9/KILL
>> > 2022-01-05 19:42:51 bb-ceph-enc-rm63-osd03-31 init.scope 
>> > ceph-osd@51.service:
>> > Failed with result 'signal'.
>> > 2022-01-05 19:43:01 bb-ceph-enc-rm63-osd03-31 init.scope 
>> > ceph-osd@51.service:
>> > Scheduled restart job, restart counter is at 2.
>> >
>> >
>> > The problem I have this has basically taken the production and metadata SSD
>> > pool's down fully and all 3 copies are offline. And I cannot find a way to
>> > find out what is causing these to crash.
>> >
>> > Kind Regards
>> >
>> > Lee
>> > _______________________________________________
>> > ceph-users mailing list -- ceph-users@ceph.io
>> > To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to