[ceph-users] log_latency slow operation observed for submit_transact, latency = 22.644258499s

2024-03-22 Thread Torkil Svensgaard
Good morning, Cephadm Reef 18.2.1. We recently added 4 hosts and changed a failure domain from host to datacenter which is the reason for the large misplaced percentage. We were seeing some pretty crazy spikes in "OSD Read Latencies" and "OSD Write Latencies" on the dashboard. Most of the ti

[ceph-users] Re: Are we logging IRC channels?

2024-03-22 Thread Alvaro Soto
Should we bring to life this again? On Tue, Mar 19, 2024, 8:14 PM Mark Nelson wrote: > A long time ago Wido used to have a bot logging IRC afaik, but I think > that's been gone for some time. > > > Mark > > > On 3/19/24 19:36, Alvaro Soto wrote: > > Hi Community!!! > > Are we logging IRC channel

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Bandelow, Gunnar
Hi Michael, i think yesterday i found the culprit in my case. After inspecting "ceph pg dump" and especially the column "last_scrub_duration". I found, that any PG without proper scrubbing was located on one of three OSDs (and all these OSDs share the same SSD for their DB). I put them on "out" a

[ceph-users] Re: log_latency slow operation observed for submit_transact, latency = 22.644258499s

2024-03-22 Thread Igor Fedotov
Hi Torkil, highly likely you're facing a well known issue with RocksDB performance drop after bulk data removal. The latter might occur at source OSDs after PG migration completion. You might want to use DB compaction (preferably offline one using ceph-kvstore-tool) to get OSD out of this "d

[ceph-users] Re: log_latency slow operation observed for submit_transact, latency = 22.644258499s

2024-03-22 Thread Torkil Svensgaard
On 22-03-2024 08:38, Igor Fedotov wrote: Hi Torkil, Hi Igor highly likely you're facing a well known issue with RocksDB performance drop after bulk data removal. The latter might occur at source OSDs after PG migration completion. Aha, thanks. You might want to use DB compaction (prefera

[ceph-users] Re: node-exporter error

2024-03-22 Thread Eugen Block
Hi, what does your node-exporter spec look like? ceph orch ls node-exporter --export If other node-exporter daemons are running in the cluster, what's the difference between them? Do they all have the same container image? ceph config get mgr mgr/cephadm/container_image_node_exporter and c

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Michel Jouvin
Hi, As I said in my initial message, I'd in mind to do exactly the same as I identified in my initial analysis that all the PGs with this problem where sharing one OSD (but only 20 PGs had the problem over ~200 hosted by the OSD). But as I don't feel I'm in an urgent situation, I was wonderin

[ceph-users] Re: log_latency slow operation observed for submit_transact, latency = 22.644258499s

2024-03-22 Thread Alexander E. Patrakov
Hello Torkil, The easiest way (in my opinion) to perform offline compaction is a bit different than what Igor suggested. We had a prior off-list conversation indicating that the results would be equivalent. 1. ceph config set osd osd_compact_on_start true 2. Restart the OSD that you want to compa

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Pierre Riteau
Hello Michel, It might be worth mentioning that the next releases of Reef and Quincy should increase the default value of osd_max_scrubs from 1 to 3. See the Reef pull request: https://github.com/ceph/ceph/pull/55173 You could try increasing this configuration setting if you haven't already, but n

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Michel Jouvin
Pierre, Yes, as mentioned in my initial email, I checked the OSD state and found nothing wrong either in the OSD logs or in the system logs (SMART errors). Thanks for the advice of increasing osd_max_scrubs, I may try it, but I doubt it is a contention problem because it really only affects a

[ceph-users] Ceph fs understand usage

2024-03-22 Thread Marcus
Hi all, I have setup a test cluster with 3 servers, Everything has default values with a replication of 3. I have created one volume called gds-common and the data pool has been configured with compression lz4 and compression_mode aggressive. I have copied 71TB data to this volume but I can no

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Frédéric Nass
Hello Michel, Pierre also suggested checking the performance of this OSD's device(s) which can be done by running a ceph tell osd.x bench. One think I can think of is how the scrubbing speed of this very OSD could be influenced by mclock sheduling, would the max iops capacity calculated by thi

[ceph-users] High OSD commit_latency after kernel upgrade

2024-03-22 Thread Özkan Göksu
Hello! After upgrading "5.15.0-84-generic" to "5.15.0-100-generic" (Ubuntu 22.04.2 LTS) , commit latency started acting weird with "CT4000MX500SSD" drives. osd commit_latency(ms) apply_latency(ms) 36 867867 373045 3045 38

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Michel Jouvin
Hi Frédéric, I think you raise the right point, sorry if I misunderstood Pierre's suggestion to look at OSD performances. Just before reading your email, I was implementing Pierre's suggestion for max_osd_scrubs and I saw the osd_mclock_max_capacity_iops_hdd for a few OSDs (I guess those with

[ceph-users] Re: High OSD commit_latency after kernel upgrade

2024-03-22 Thread Anthony D'Atri
https://askubuntu.com/questions/1454997/how-to-stop-sys-from-changing-usb-ssd-provisioning-mode-from-unmap-to-full-in-ub How to stop sys from changing USB SSD provisioning_mode from unmap to full in Ubuntu 22.04? askubuntu.com ? > On Mar 22, 2024, at 09:36, Özkan Göksu wrote: > > Hello! > >

[ceph-users] Re: High OSD commit_latency after kernel upgrade

2024-03-22 Thread Özkan Göksu
Hello Anthony, thank you for the answer. While researching I also found out this type of issues but the thing I did not understand is in the same server the OS drives "SAMSUNG MZ7WD480" is all good. root@sd-01:~# lsblk -D NAME DISC-ALN DISC-GRAN DISC-MAX

[ceph-users] Re: High OSD commit_latency after kernel upgrade

2024-03-22 Thread Özkan Göksu
Hello again. In ceph recommendations I found this: https://docs.ceph.com/en/quincy/start/hardware-recommendations/ WRITE CACHES Enterprise SSDs and HDDs normally include power loss protection features which ensure data durability when power is lost while operating, and use multi-level caches to

[ceph-users] Re: High OSD commit_latency after kernel upgrade

2024-03-22 Thread Anthony D'Atri
Maybe because the Crucial units are detected as client drives? But also look at the device paths and the output of whatever "disklist" is. Your boot drives are SATA and the others are SAS which seems even more likely to be a factor. > On Mar 22, 2024, at 10:42, Özkan Göksu wrote: > > Hello A

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Frédéric Nass
Michel,   Glad to know that was it.   I was wondering when would per OSD osd_mclock_max_capacity_iops_hdd value be set in cluster's config database since I don't have any set in my lab. Turns out the per OSD osd_mclock_max_capacity_iops_hdd is only set when the calculated value is belo

[ceph-users] Re: log_latency slow operation observed for submit_transact, latency = 22.644258499s

2024-03-22 Thread Joshua Baergen
Personally, I don't think the compaction is actually required. Reef has compact-on-iteration enabled, which should take care of this automatically. We see this sort of delay pretty often during PG cleaning, at the end of a PG being cleaned, when the PG has a high count of objects, whether or not OS

[ceph-users] Re: High OSD commit_latency after kernel upgrade

2024-03-22 Thread Özkan Göksu
After I set these 2 udev rules: root@sd-02:~# cat /etc/udev/rules.d/98-ceph-provisioning-mode.rules ACTION=="add", SUBSYSTEM=="scsi_disk", ATTR{provisioning_mode}:="unmap" root@sd-02:~# cat /etc/udev/rules.d/99-ceph-write-through.rules ACTION=="add", SUBSYSTEM=="scsi_disk", ATTR{cache_type}:="wri

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Michel Jouvin
Hi, The attempt to rerun the bench was not really a success. I got the following messages: - Mar 22 14:48:36 idr-osd2 ceph-osd[326854]: osd.29 83873 maybe_override_max_osd_capacity_for_qos osd bench result - bandwidth (MiB/sec): 10.910 iops: 2792.876 elapsed_sec: 1.074 Mar 22 14:48:36 i

[ceph-users] Re: [ext] Re: cephadm auto disk preparation and OSD installation incomplete

2024-03-22 Thread Kuhring, Mathias
Hey Eugen, Thank you for the quick reply. The 5 missing disks on the one host were completely installed after I fully cleaned them up as I described. So, seems a smaller number of disks can make it. Regarding the other host with 40 disks: Failing the MGR didn't have any effect. There are nor er

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Michel Jouvin
Frédéric, We arrived at the same conclusions! I agree that an insane low value would be a good addition: the idea would be that the benchmark emits a warning about the value but the it will not put a value lower than the minimum defined. I don't have a precise idea of the possible bad side ef

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Anthony D'Atri
Perhaps emitting an extremely low value could have value for identifying a compromised drive? > On Mar 22, 2024, at 12:49, Michel Jouvin > wrote: > > Frédéric, > > We arrived at the same conclusions! I agree that an insane low value would be > a good addition: the idea would be that the benc

[ceph-users] Re: 18.8.2: osd_mclock_iops_capacity_threshold_hdd untypical values

2024-03-22 Thread Michel Jouvin
Follow-up, changing the title to the real topic... I did more tests on my OSDs using "ceph tell osd.x bench..." as advised by https://docs.ceph.com/en/quincy/rados/configuration/mclock-config-ref/#benchmarking-test-steps-using-osd-bench (exact impact of "cache drop" is not clear/visible based

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Kai Stian Olstad
On Fri, Mar 22, 2024 at 04:29:21PM +0100, Frédéric Nass wrote: A/ these incredibly low values were calculated a while back with an unmature version of the code or under some specific hardware conditions and you can hope this won't happen again The OSD run bench and update osd_mclock_max_capac

[ceph-users] Re: [ext] Re: cephadm auto disk preparation and OSD installation incomplete

2024-03-22 Thread Kuhring, Mathias
I'm afraid the parameter mgr mgr/cephadm/default_cephadm_command_timeout is buggy. Once not on default anymore, MGR is preparing the parameter a bit (e.g. substracting 5 secs) And there making it float, but cephadm is not having it (not even if I try the default 900 myself): [WRN] CEPHADM_APPLY

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Frédéric Nass
Michel,   Log says that osd.29 is providing 2792 '4k' iops at 10.910 MiB/s. These figures suggest that a controller write-back cache is in use along the IO path. Is that right?   Since 2792 is above 500, osd_mclock_max_capacity_iops_hdd falls back to 315 and OSD is suggesting runni

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Frédéric Nass
  > The OSD run bench and update osd_mclock_max_capacity_iops_{hdd,ssd} every > time the OSD is started. > If you check the OSD log you'll see it does the bench.   Are you sure about the update on every start? Does the update happen only if the benchmark result is < 500 iops?   Loo

[ceph-users] Re: Reef (18.2): Some PG not scrubbed/deep scrubbed for 1 month

2024-03-22 Thread Kai Stian Olstad
On Fri, Mar 22, 2024 at 06:51:44PM +0100, Frédéric Nass wrote: The OSD run bench and update osd_mclock_max_capacity_iops_{hdd,ssd} every time the OSD is started. If you check the OSD log you'll see it does the bench.   Are you sure about the update on every start? Does the update happen only

[ceph-users] Re: Call for Interest: Managed SMB Protocol Support

2024-03-22 Thread Alexander E. Patrakov
Hi John, > A few major features we have planned include: > * Standalone servers (internally defined users/groups) No concerns here > * Active Directory Domain Member Servers In the second case, what is the plan regarding UID mapping? Is NFS coexistence planned, or a concurrent mount of the same

[ceph-users] Reset health.

2024-03-22 Thread Albert Shih
Hi, Very basic question : 2 days ago I reboot all the cluster. Everything work fine. But I'm guessing during the shutdown 4 osd was mark as crash [WRN] RECENT_CRASH: 4 daemons have recently crashed osd.381 crashed on host cthulhu5 at 2024-03-20T18:33:12.017102Z osd.379 crashed on host ct

[ceph-users] How you manage log

2024-03-22 Thread Albert Shih
Hi, With our small cluster (11 nodes) I notice ceph log a lot Beside to keep that somewhere «just in case», is they are anything to check regularly in the log (in prevention of more serious problem) ? Or can we trust «ceph health» and use the log only for debug. Regards -- Albert SHIH 🦫 🐸 Fr

[ceph-users] Re: Reset health.

2024-03-22 Thread Murilo Morais
You can use the `ceph crash` interface to view/archive recent crashes. [1] To list recent crashes: ceph crash ls-new To get information about a particular crash: ceph crash info To silence a crash: ceph crash archive To silence all active crashes: ceph crash archive-all [1] https://docs.ceph.c

[ceph-users] Ceph versus Zabbix: failure: no data sent

2024-03-22 Thread John Jasen
If the documentation is to be believed, it's just install the zabbix sender, then; ceph mgr module enable zabbix ceph zabbix config-set zabbix_host my-zabbix-server (Optional) Set the identifier to the fsid. And poof. I should now have a discovered entity on my zabbix server to add templates to

[ceph-users] Re: Are we logging IRC channels?

2024-03-22 Thread Mark Nelson
Sure! I think Wido just did it all unofficially, but afaik we've lost all of those records now. I don't know if Wido still reads the mailing list but he might be able to chime in. There was a ton of knowledge in the irc channel back in the day. With slack, it feels like a lot of discussions

[ceph-users] Re: Are we logging IRC channels?

2024-03-22 Thread Gregory Farnum
I put it on the list for the next CLT. :) (though I imagine it will move to the infrastructure meeting from there.) On Fri, Mar 22, 2024 at 4:42 PM Mark Nelson wrote: > Sure! I think Wido just did it all unofficially, but afaik we've lost > all of those records now. I don't know if Wido still

[ceph-users] Laptop Losing Connectivity To CephFS On Sleep/Hibernation

2024-03-22 Thread duluxoz
Hi All, I'm looking for some help/advice to solve the issue outlined in the heading. I'm running CepfFS (name: cephfs) on a Ceph Reef (v18.2.2 - latest update) cluster, connecting from a laptop running Rocky Linux v9.3 (latest update) with KDE v5 (latest update). I've set up the laptop to co

[ceph-users] Mounting A RBD Via Kernal Modules

2024-03-22 Thread duluxoz
Hi All, I'm trying to mount a Ceph Reef (v18.2.2 - latest version) RBD Image as a 2nd HDD on a Rocky Linux v9.3 (latest version) host. The EC pool has been created and initialised and the image has been created. The ceph-common package has been installed on the host. The correct keyring ha