[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-12 Thread José H . Freidhof
Hi Igor the reason why i tested differekt rocksdb options is that was having a really bad write performance with the default settings (30-60mb/s) on the cluster... actually i have 200mb/s read and 180mb/s write performance now i dont now which of the booth settings are the good ones Another

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-12 Thread José H . Freidhof
Hi Igor, Thx for checking the logs.. but what the hell is going on here? :-) Yes its true i tested the and created the osd´s with three different rockdb options. I can not understand why the osd dont have the same rockdb option, because i have created ALL OSDs new after set and test those

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-12 Thread José H . Freidhof
Hello Igor "Does single OSD startup (after if's experiencing "unable to allocate) takes 20 mins as well?" A: YES Here the example log of the startup and recovery of a problematic osd. https://paste.ubuntu.com/p/2WVJbg7cBy/ Here the example log of a problematic osd

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-11 Thread Igor Fedotov
hmm... so it looks like RocksDB still doesn't perform WAL cleanup during regular operation but applies it on OSD startup Does single OSD startup (after if's experiencing "unable to allocate) takes 20 mins as well? Could you please share OSD log containing both that long startup and

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-08 Thread José H . Freidhof
Hi Igor, "And was osd.2 redeployed AFTER settings had been reset to defaults ?" A: YES "Anything particular about current cluster use cases?" A: we are using it temporary as a iscsi target for a vmware esxi cluster with 6 hosts. We created two 10tb iscsi images/luns for vmware, because the other

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-08 Thread Igor Fedotov
And was osd.2 redeployed AFTER settings had been reset to defaults ? Anything particular about current cluster use cases? E.g. is it a sort of regular usage (with load flukes and peak) or may be some permanently running stress load testing. The latter might tend to hold the resources and e.g.

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-07 Thread José H . Freidhof
Hi Igor, yes the same problem is on osd.2 we have 3 OSD Nodes... Each Node has 20 Bluestore OSDs ... in total we have 60 OSDs i checked right now one node... and 15 of 20 OSDs have this problem and error in the log. the settings that you have complained some emails ago .. i have reverted them

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-07 Thread José H . Freidhof
Hi Igor, yesss Am Do., 7. Okt. 2021 um 19:24 Uhr schrieb Igor Fedotov < igor.fedo...@croit.io>: > Hey Jose, > > so you applied default settings and compacted osd.8, right? > > This helped for a short time and now it's back to the same state, right? > > > Igor > > > On 10/7/2021 12:46 PM, José

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-07 Thread Igor Fedotov
And does redeployed osd.2 expose the same issue (or at least DB/WAL disbalance) again? Were settings reverted to defaults for it as well? Thanks Igor On 10/7/2021 12:46 PM, José H. Freidhof wrote: Good morning, i checked today the osd.8 and the log shows again the same error bluefs

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-07 Thread Igor Fedotov
Hey Jose, so you applied default settings and compacted osd.8, right? This helped for a short time and now it's back to the same state, right? Igor On 10/7/2021 12:46 PM, José H. Freidhof wrote: Good morning, i checked today the osd.8 and the log shows again the same error bluefs

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-07 Thread José H . Freidhof
Good morning, i checked today the osd.8 and the log shows again the same error bluefs _allocate unable to allocate 0x10 on bdev 0, allocator name bluefs-wal, allocator type hybrid, capacity 0xb4000, block size 0x10, free 0xff000, fragmentation 0, allocated 0x0 any idea why that could

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-06 Thread José H . Freidhof
Hi Igor, today i repaired one osd node and all osd´s on the node, creating them new again after that i waited for the rebalance/recovery process and the cluster was healthy after some hours.. i notices that the osd.2 does not have any more this error in the log. but i noticed it now on the

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-06 Thread Igor Fedotov
On 10/6/2021 4:25 PM, José H. Freidhof wrote: hi, no risk no fun  okay I have reset the settings you mentioned to standard. what you exactly mean with taking offline the osd? ceph orch daemon stop osd.2? or mark down? "daemon stop" is enough. You  might want to set noout flag before that

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-06 Thread José H . Freidhof
hi, no risk no fun  okay I have reset the settings you mentioned to standard. what you exactly mean with taking offline the osd? ceph orch daemon stop osd.2? or mark down? for the command which path i use? you mean: bluestore-kv /var/lib/ceph/$fsid/osd.2 compact??? Igor Fedotov schrieb am

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-06 Thread Igor Fedotov
On 10/6/2021 2:16 PM, José H. Freidhof wrote: Hi Igor, yes i have some osd settings set :-) here are my ceph config dump. those settings are from a redhat document for bluestore devices maybe it is that setting causing this problem? "advanced mon_compact_on_trimfalse"??? OMG!!! No -

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-06 Thread José H . Freidhof
Hi Igor, yes i have some osd settings set :-) here are my ceph config dump. those settings are from a redhat document for bluestore devices maybe it is that setting causing this problem? "advanced mon_compact_on_trimfalse"??? i will test it this afternoon... at the moment are everything

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-06 Thread Igor Fedotov
Jose, In fact 48GB is a way too much for WAL drive - usually the write ahead log tend to be 2-4 GBs. But in your case it's ~150GB, while DB itself is very small (146MB!!!): WAL         45 GiB      111 GiB     0 B         0 B         0 B       154 GiB     2400 DB          0 B         164

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-06 Thread José H . Freidhof
Hello Igor, yes the volume is nvme wal partitions for the bluestore devicegroups are only 48gb each on each osd node are 1 nvme with 1tb splitted in 20 lvs with 48gb (WAL) on each osd node are 4 ssd with 1tb splitted in 5 lvs with 175gb (rock.db) on each osd node are 20 hdd with 5.5tb with 1

[ceph-users] Re: bluefs _allocate unable to allocate

2021-10-06 Thread Igor Fedotov
Hey Jose, it looks like your WAL volume is out of space which looks weird given its capacity = 48Gb. Could you please share the output of the following commands: ceph daemon osd.N bluestore bluefs device info ceph daemon osd.N bluefs stats Thanks, Igor On 10/6/2021 12:24 PM, José H.