[ceph-users] Re: OSDs spam log with scrub starts

2023-08-31 Thread Zakhar Kirpichenko
This is happening to our 16.2.14 cluster as well. I'm not sure whether this was happening before the upgrade to 16.2.14. /Z On Thu, 31 Aug 2023, 17:49 Adrien Georget, wrote: > Hello, > > On our 16.2.14 CephFS cluster, all OSDs are spamming logs with messages > like "log_channel(cluster) log

[ceph-users] Re: Pacific 16.2.14 debian Incomplete

2023-08-30 Thread Zakhar Kirpichenko
It looks much better, at least for Ubuntu focal, thanks! /Z On Thu, 31 Aug 2023 at 03:48, Yuri Weinstein wrote: > We redeployed all packages again. > > Please confirm that the issue is resolved. > > Thank you for your help and patience! > > On Wed, Aug 30, 2023 at 3:44 

[ceph-users] Re: Pacific 16.2.14 debian Incomplete

2023-08-30 Thread Zakhar Kirpichenko
Now the release email comes and the repositories are still missing packages. What a mess. /Z On Wed, 30 Aug 2023 at 19:27, Yuri Weinstein wrote: > 16.2.14 has not been released yet. > > Please don't do any upgrades before we send an announcement email. > > TIA > > On Wed, Aug 30, 2023 at 8:45 

[ceph-users] Re: Pacific 16.2.14 debian Incomplete

2023-08-30 Thread Zakhar Kirpichenko
Hi, Please note that some packages have been pushed for Ubuntu focal as well, but the repo is incomplete. I think it would be good if such things could be avoided in the future. /Z On Wed, 30 Aug 2023 at 19:27, Yuri Weinstein wrote: > 16.2.14 has not been released yet. > > Please don't do any

[ceph-users] Re: librbd 4k read/write?

2023-08-10 Thread Zakhar Kirpichenko
Hi, You can use the following formula to roughly calculate the IOPS you can get from a cluster: (Drive_IOPS * Number_of_Drives * 0.75) / Cluster_Size. For example, for 60 10K rpm SAS drives each capable of 200 4K IOPS and a replicated pool with size 3: (~200 * 60 * 0.75) / 3 = ~3000 IOPS with

[ceph-users] Re: 16.2.13: ERROR:ceph-crash:directory /var/lib/ceph/crash/posted does not exist; please create

2023-06-07 Thread Zakhar Kirpichenko
to 167:167, which in my system is the user ID crash process uses inside the crash container. /Z On Mon, 5 Jun 2023 at 11:24, Zakhar Kirpichenko wrote: > Any other thoughts on this, please? Should I file a bug report? > > /Z > > On Fri, 2 Jun 2023 at 06:11, Zakhar Kirpichenko wro

[ceph-users] Re: 16.2.13: ERROR:ceph-crash:directory /var/lib/ceph/crash/posted does not exist; please create

2023-06-05 Thread Zakhar Kirpichenko
Any other thoughts on this, please? Should I file a bug report? /Z On Fri, 2 Jun 2023 at 06:11, Zakhar Kirpichenko wrote: > Thanks, Josh. The cluster is managed by cephadm. > > On Thu, 1 Jun 2023, 23:07 Josh Baergen, wrote: > >> Hi Zakhar, >> >> I'm going to

[ceph-users] Re: 16.2.13: ERROR:ceph-crash:directory /var/lib/ceph/crash/posted does not exist; please create

2023-06-01 Thread Zakhar Kirpichenko
e the directory permissions, assuming that you manage > the directories yourself. If this is managed by cephadm or something like > that, then that seems like some sort of missing migration in the upgrade. > > Josh > > On Thu, Jun 1, 2023 at 12:34 PM Zakhar Kirpichenko > wrote

[ceph-users] 16.2.13: ERROR:ceph-crash:directory /var/lib/ceph/crash/posted does not exist; please create

2023-06-01 Thread Zakhar Kirpichenko
Hi, I'm having an issue with crash daemons on Pacific 16.2.13 hosts. ceph-crash throws the following error on all hosts: ERROR:ceph-crash:directory /var/lib/ceph/crash/posted does not exist; please create ERROR:ceph-crash:directory /var/lib/ceph/crash/posted does not exist; please create

[ceph-users] Re: v16.2.13 Pacific released

2023-05-09 Thread Zakhar Kirpichenko
Thanks! An upgrade from 16.2.12 on Ubuntu 20.04 LTS went smoothly. /Z On Wed, 10 May 2023 at 00:45, Yuri Weinstein wrote: > We're happy to announce the 13th backport release in the Pacific series. > > https://ceph.io/en/news/blog/2023/v16-2-13-pacific-released/ > > Notable Changes >

[ceph-users] Re: quincy 17.2.6 - write performance continuously slowing down until OSD restart needed

2023-05-09 Thread Zakhar Kirpichenko
he perf counters and leave the cluster > run for 30 mins and collect perf counters again. > > - wait 24 hours and collect the counters one more time > > - share all four counters snapshots. > > > Thanks, > > Igor > > On 5/8/2023 11:31 PM, Zakhar Kirpichenko wrote: &g

[ceph-users] Re: quincy 17.2.6 - write performance continuously slowing down until OSD restart needed

2023-05-08 Thread Zakhar Kirpichenko
Don't mean to hijack the thread, but I may be observing something similar with 16.2.12: OSD performance noticeably peaks after OSD restart and then gradually reduces over 10-14 days, while commit and apply latencies increase across the board. Non-default settings are:

[ceph-users] Re: Ceph 16.2.12, particular OSD shows higher latency than others

2023-04-27 Thread Zakhar Kirpichenko
Periodically we observe OSD perf drop due to degraded RocksDB > performance, particularly after bulk data removal/migration.. Compaction > is quite helpful in this case. > > > Thanks, > > Igor > > > > On 26/04/2023 20:22, Zakhar Kirpichenko wrote: > > Hi, >

[ceph-users] Re: Ceph 16.2.12, particular OSD shows higher latency than others

2023-04-27 Thread Zakhar Kirpichenko
-primary and others would take over. If the new > primary OSDs show increased latencies as well there might be something > else going on. > > Zitat von Zakhar Kirpichenko : > > > Thanks, Eugen. I very much appreciate your time and replies. > > > > It's a hybrid OSD

[ceph-users] Re: Ceph 16.2.12, particular OSD shows higher latency than others

2023-04-27 Thread Zakhar Kirpichenko
at OSD (ceph pg > ls-by-osd )? Have you tried to restart and/or compact the OSD and > see if anything improves? > You could set its primary-affinity to 0, or the worst case rebuild > that OSD. And there are no smart errors or anything in dmesg reported > about this disk? > > Zi

[ceph-users] Re: Ceph 16.2.12, particular OSD shows higher latency than others

2023-04-27 Thread Zakhar Kirpichenko
at": "2023-04-27T07:37:35.046036+", > "age": 54.80501619997, > "duration": 0.5819846869995, > ... > > The output contains the PG (so you know which pool is involved) and > the duration of the operation, not su

[ceph-users] Re: Ceph 16.2.12, particular OSD shows higher latency than others

2023-04-26 Thread Zakhar Kirpichenko
f 13:f6c9079e:::rbd_data.eed629ecc1f946.001c:head [stat,write 4018176~4096] snapc 0=[] ondisk+write+known_if_redirected e118835)", "initiated_at": "2023-04-26T07:00:58.299270+", There's a lot more information there ofc. I also tried to `dump_o

[ceph-users] Ceph 16.2.12, bluestore cache doesn't seem to be used much

2023-04-26 Thread Zakhar Kirpichenko
Hi, I have a Ceph 16.2.12 cluster with hybrid OSDs (HDD block storage, DB/WAL on NVME). All OSD settings are default except, cache-related settings are as follows: osd.14dev bluestore_cache_autotune true osd.14dev bluestore_cache_size_hdd

[ceph-users] Ceph 16.2.12, particular OSD shows higher latency than others

2023-04-26 Thread Zakhar Kirpichenko
Hi, I have a Ceph 16.2.12 cluster with uniform hardware, same drive make/model, etc. A particular OSD is showing higher latency than usual in `ceph osd perf`, usually mid to high tens of milliseconds while other OSDs show low single digits, although its drive's I/O stats don't look different from

[ceph-users] Re: cephadm upgrade 16.2.10 to 16.2.11: osds crash and get stuck restarting

2023-01-26 Thread Zakhar Kirpichenko
ultiple times after the reboot. /Z On Thu, 26 Jan 2023 at 05:41, Konstantin Shalygin wrote: > Hi Zakhar, > > On 26 Jan 2023, at 08:33, Zakhar Kirpichenko wrote: > > Jan 25 23:07:53 ceph01 bash[2553123]: > > /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AV

[ceph-users] cephadm upgrade 16.2.10 to 16.2.11: osds crash and get stuck restarting

2023-01-25 Thread Zakhar Kirpichenko
Hi, Attempted to upgrade 16.2.10 to 16.2.11, 2 OSDs out of many started crashing in a loop on the very 1st host: Jan 25 23:07:53 ceph01 bash[2553123]: Uptime(secs): 0.0 total, 0.0 interval Jan 25 23:07:53 ceph01 bash[2553123]: Flush(GB): cumulative 0.000, interval 0.000 Jan 25 23:07:53 ceph01

[ceph-users] Re: Upgrade Ceph 16.2.10 to 17.2.x for Openstack RBD storage

2022-12-05 Thread Zakhar Kirpichenko
erver and client versions is unsupported and may lead to anomalous behavior." As pointed out by a kind soul on reddit. /Z On Mon, 5 Dec 2022 at 06:01, Zakhar Kirpichenko wrote: > Hi! > > I'm planning to upgrade our Ceph cluster from Pacific (16.2.10) to Quincy > (17.2.x). The cluster

[ceph-users] Upgrade Ceph 16.2.10 to 17.2.x for Openstack RBD storage

2022-12-04 Thread Zakhar Kirpichenko
Hi! I'm planning to upgrade our Ceph cluster from Pacific (16.2.10) to Quincy (17.2.x). The cluster is used for Openstack block storage (RBD), Openstack version is Wallaby built on Ubuntu 20.04. Is anyone using Ceph Quincy (17.2.x) with Openstack Wallaby? If you are, please let me know if you've

[ceph-users] Re: Upgrade 16.2.10 to 17.2.x: any caveats?

2022-11-24 Thread Zakhar Kirpichenko
Thanks, Stefan. I've read this very well. My question is whether there's anything not covered by the available documentation that we should be aware of. /Z On Fri, 25 Nov 2022 at 09:11, Stefan Kooman wrote: > On 11/24/22 18:53, Zakhar Kirpichenko wrote: > > Hi! > > > > I

[ceph-users] Re: 16.2.10: ceph osd perf always shows high latency for a specific OSD

2022-10-07 Thread Zakhar Kirpichenko
how much snapshots in chain for this image. > More than 10 snaps for one image can increase client ops latency to tens > milliseconds even for NVMe drives, that usually operates at usec or 1-2msec > > > k > Sent from my iPhone > > > On 7 Oct 2022, at 14:35, Zakhar Kirpichen

[ceph-users] Re: 16.2.10: ceph osd perf always shows high latency for a specific OSD

2022-10-07 Thread Zakhar Kirpichenko
t; > Sent from my iPhone > > > On 7 Oct 2022, at 07:33, Zakhar Kirpichenko wrote: > > > > Anyone, please? > > ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] 16.2.10: ceph osd perf always shows high latency for a specific OSD

2022-10-06 Thread Zakhar Kirpichenko
Hi, I'm having a peculiar "issue" in my cluster, which I'm not sure whether it's real: a particular OSD always shows significant latency in `ceph osd perf` report, an order of magnitude higher than any other OSD. I traced this OSD to a particular drive in a particular host. OSD logs don't look

[ceph-users] Re: pacific doesn't defer small writes for pre-pacific hdd osds

2022-07-14 Thread Zakhar Kirpichenko
Many thanks, Dan. Much appreciated! /Z On Thu, 14 Jul 2022 at 08:43, Dan van der Ster wrote: > Yes, that is correct. No need to restart the osds. > > .. Dan > > > On Thu., Jul. 14, 2022, 07:04 Zakhar Kirpichenko, > wrote: > >> Hi! >> >> M

[ceph-users] Re: pacific doesn't defer small writes for pre-pacific hdd osds

2022-07-13 Thread Zakhar Kirpichenko
Hi! My apologies for butting in. Please confirm that bluestore_prefer_deferred_size_hdd is a runtime option, which doesn't require OSDs to be stopped or rebuilt? Best regards, Zakhar On Tue, 12 Jul 2022 at 14:46, Dan van der Ster wrote: > Hi Igor, > > Thank you for the reply and information.

[ceph-users] Re: Ceph PGs stuck inactive after rebuild node

2022-04-06 Thread Zakhar Kirpichenko
> > I remember this has been discussed a couple of times in the past on > > this list, but I'm wondering if this still happens in newer releases. > > I assume there's no way of preventing that, so we'll probably go with > > the safe approach on the next node. It's a pr

[ceph-users] Re: Ceph PGs stuck inactive after rebuild node

2022-04-06 Thread Zakhar Kirpichenko
Hi Eugen, Can you please elaborate on what you mean by "restarting the primary PG"? Best regards, Zakhar On Wed, Apr 6, 2022 at 5:15 PM Eugen Block wrote: > Update: Restarting the primary PG helped to bring the PGs back to > active state. Consider this thread closed. > > > Zitat von Eugen

[ceph-users] Re: Cephadm is stable or not in product?

2022-03-08 Thread Zakhar Kirpichenko
Hi! I run cephadm-based 16.2.x cluster in production. It's been mostly fine, but not without quirks. Hope this helps. /Z On Tue, Mar 8, 2022 at 6:17 AM norman.kern wrote: > Dear Ceph folks, > > Anyone is using cephadm in product(Version: Pacific)? I found several bugs > on it and > I really

[ceph-users] Re: cephadm: update fewer OSDs at a time?

2022-02-14 Thread Zakhar Kirpichenko
you have some pools with size = 2 and min_size = 2? > > Regards, > Eugen > > > Zitat von Zakhar Kirpichenko : > > > Hi! > > > > Sometimes when we upgrade our cephadm-managed 16.2.x cluster, cephadm > > decides that it's safe to upgrade a bunch of OSDs at a ti

[ceph-users] cephadm: update fewer OSDs at a time?

2022-02-13 Thread Zakhar Kirpichenko
Hi! Sometimes when we upgrade our cephadm-managed 16.2.x cluster, cephadm decides that it's safe to upgrade a bunch of OSDs at a time, as a result sometimes RBD-backed Openstack VMs appear to get I/O stalls and read-only filesystems. Is there a way to make cephadm upgrade fewer OSDs at a time, or

[ceph-users] Ceph 16.2.7 + cephadm, how to reduce logging and trim existing logs?

2022-01-27 Thread Zakhar Kirpichenko
Hi! I have a Ceph cluster version 16.2.7 built with cephadm. Ceph logs in /var/log/ceph/ all rotate nicely, but container logs in /var/lib/docker/containers/ keep growing. For example, a log of a monitor container has reached almost 3 GB since December 2021: 09:18 [root@ceph02

[ceph-users] Re: Best practice to consider journal disk for CEPH OSD

2021-12-05 Thread Zakhar Kirpichenko
e block db. > > Currently we are planning to install new OSD with SATA-SSD. so block-db > will not be required. > > we will use device-class and different replicated_rule > > $ceph osd crush set-device-class SSD osd.10 > $ceph osd crush rule create-replicated replicated_rule_SSD default hos

[ceph-users] Re: Best practice to consider journal disk for CEPH OSD

2021-12-05 Thread Zakhar Kirpichenko
Hi! If you use SSDs for OSDs, there's no real benefit from putting DB/WAL on a separate drive. Best regards, Z On Sun, Dec 5, 2021 at 10:15 AM Md. Hejbul Tawhid MUNNA wrote: > Hi, > > We are running openstack cloud with backend ceph storage. Currently we have > only HDD storage in our ceph

[ceph-users] Re: Pacific: parallel PG reads?

2021-11-11 Thread Zakhar Kirpichenko
; > Weiwen Hu > > > > 从 Windows 版邮件 <https://go.microsoft.com/fwlink/?LinkId=550986>发送 > > > > *发件人: *Zakhar Kirpichenko > *发送时间: *2021年11月11日 20:54 > *收件人: *ceph-users > *主题: *[ceph-users] Pacific: parallel PG reads? > > > > Hi, > > I'm sti

[ceph-users] Re: Pacific: parallel PG reads?

2021-11-11 Thread Zakhar Kirpichenko
Hi, On Thu, Nov 11, 2021 at 3:26 PM Janne Johansson wrote: > Den tors 11 nov. 2021 kl 13:54 skrev Zakhar Kirpichenko >: > > I'm still trying to combat really bad read performance from HDD-backed > > replicated pools, which is under 100 MB/s most of the time with 1 thread >

[ceph-users] Pacific: parallel PG reads?

2021-11-11 Thread Zakhar Kirpichenko
Hi, I'm still trying to combat really bad read performance from HDD-backed replicated pools, which is under 100 MB/s most of the time with 1 thread and QD=1. I don't quite understand why the reads are that slow, i.e. much slower than a single HDD, but do understand that Ceph clients read a PG

[ceph-users] Pacific PG count questions

2021-11-08 Thread Zakhar Kirpichenko
Hi, I have a 6-node cluster running Pacific 16.2.6 with 54 x 10TB HDD and 12 x 6.4 TB NVME drives. By default, the autoscaler appears to scale down each pool to 32 PGs, which causes a very uneven data distribution and somewhat lower performance. Although I knew that previously there was a neat

[ceph-users] Re: Optimal Erasure Code profile?

2021-11-05 Thread Zakhar Kirpichenko
oda Services Co., Ltd. > e: istvan.sz...@agoda.com > --- > > -----Original Message- > From: Zakhar Kirpichenko > Sent: Friday, November 5, 2021 6:45 PM > To: ceph-users > Subject: [ceph-users] Re: Optimal Erasure Code

[ceph-users] Re: Optimal Erasure Code profile?

2021-11-05 Thread Zakhar Kirpichenko
the simple Host failure domain on 6 aktive servers and “k=4, m=2” > erasure coding. > But I don't advise you to do that! > > > Best > Sebastian > > > > On 05.11.2021, at 06:14, Zakhar Kirpichenko wrote: > > > > Hi! > > > > I've got a CEPH 16.

[ceph-users] Re: Are setting 'ceph auth caps' and/or adding a cache pool I/O-disruptive operations?

2021-11-05 Thread Zakhar Kirpichenko
more clear that adding a cache tier isn't a good idea. /Z On Fri, Nov 5, 2021 at 7:55 AM Anthony D'Atri wrote: > Cache tiers are kind of deprecated, they’re finicky and easy to run into > trouble with. Suggest avoiding. > > > On Nov 4, 2021, at 10:42 PM, Zakhar Kirpichenko > wrote: &g

[ceph-users] Stale monitoring alerts in UI

2021-11-05 Thread Zakhar Kirpichenko
Hi, I seem to have some stale monitoring alerts in my Mgr UI, which do not want to go away. For example (I'm also attaching an image for your convenience): MTU Mismatch: Node ceph04 has a different MTU size (9000) than the median value on device storage-int. The alerts appears to be active, but

[ceph-users] Are setting 'ceph auth caps' and/or adding a cache pool I/O-disruptive operations?

2021-11-04 Thread Zakhar Kirpichenko
Hi, I'm trying to figure out if setting auth caps and/or adding a cache pool are I/O-disruptive operations, i.e. if caps reset to 'none' for a brief moment or client I/O momentarily stops for other reasons. For example, I had the following auth setting in my 16.2.x cluster: client.cinder

[ceph-users] Optimal Erasure Code profile?

2021-11-04 Thread Zakhar Kirpichenko
Hi! I've got a CEPH 16.2.6 cluster, the hardware is 6 x Supermicro SSG-6029P nodes, each equipped with: 2 x Intel(R) Xeon(R) Gold 5220R CPUs 384 GB RAM 2 x boot drives 2 x 1.6 TB enterprise NVME drives (DB/WAL) 2 x 6.4 TB enterprise drives (storage tier) 9 x 9TB HDDs (storage tier) 2 x Intel

[ceph-users] Re: Best way to add multiple nodes to a cluster?

2021-11-02 Thread Zakhar Kirpichenko
> From: Etienne Menguy > Sent: Tuesday, November 2, 2021 3:17 PM > To: Zakhar Kirpichenko > Cc: ceph-users > Subject: [ceph-users] Re: Best way to add multiple nodes to a cluster? > > Email received from the internet. If in doubt, don't click any link nor > open any atta

[ceph-users] Best way to add multiple nodes to a cluster?

2021-11-02 Thread Zakhar Kirpichenko
Hi! I have a 3-node 16.2.6 cluster with 33 OSDs, and plan to add another 3 nodes of the same configuration to it. What is the best way to add the new nodes and OSDs so that I can avoid a massive rebalance and performance hit until all new nodes and OSDs are in place and operational? I would very

[ceph-users] Re: Ceph performance optimization with SSDs

2021-10-22 Thread Zakhar Kirpichenko
Hi, It is my experience with Bluestore that even if you put DB/WAL on a very fast SSD (NVME in my case), the OSD will still refuse to write faster than the storage device (HDD) can write on average. This is a bummer because the behavior is very different from Filestore, which could buffer write

[ceph-users] Re: Can CEPH RBD devices be assigned to virtual machines in pre-allocation mode?

2021-10-22 Thread Zakhar Kirpichenko
Hi! Although I'm not sure why you'd want to do it, there's a thick provision option for RBD create, i.e. it is possible: $ rbd help create | grep thick --thick-provision fully allocate storage and zero image Note that fully allocating and zeroing the image means writing data to every

[ceph-users] Re: about rbd and database

2021-10-21 Thread Zakhar Kirpichenko
Hi, RBD is a block device and is suitable for running any application, but its performance depends on many factors and it has significant overheads. What you need to determine is whether RBD on your particular hardware with your particular settings provides satisfactory random I/O performance and

[ceph-users] Re: OSD Crashes in 16.2.6

2021-10-12 Thread Zakhar Kirpichenko
he risk of an in place upgrade. > > > > On Tue, Oct 12, 2021 at 2:22 PM Igor Fedotov > wrote: > >> Zakhar, >> >> could you please point me to the similar reports at Proxmox forum? >> >> Curious what's the Ceph release mentioned there... >>

[ceph-users] Re: OSD Crashes in 16.2.6

2021-10-12 Thread Zakhar Kirpichenko
x > forum posts as well. I'm not sure if going to the 5.4 kernel will create > any other challenges for us, as we're using dual port mellanox connectx-6 > 200G nics in the hosts, but it is definitely something we can try. > > Marco > > On Tue, Oct 12, 2021 at 1:53 PM Zakhar

[ceph-users] Re: OSD Crashes in 16.2.6

2021-10-12 Thread Zakhar Kirpichenko
Hi, This could be kernel-related, as I've seen similar reports in Proxmox forum. Specifically, 5.11.x with Ceph seems to be hitting kernel NULL pointer dereference. Perhaps a newer kernel would help. If not, I'm running 16.2.6 with kernel 5.4.x without any issues. Best regards, Z On Tue, Oct

[ceph-users] Re: CEPH 16.2.x: disappointing I/O performance

2021-10-06 Thread Zakhar Kirpichenko
Hi, I tried experimenting with RBD striping feature: rbd image 'volume-bd873c3f-c8c7-4270-81f8-951f65fc860c': size 50 GiB in 12800 objects order 22 (4 MiB objects) snapshot_count: 0 id: 187607b9ebf28a block_name_prefix: rbd_data.187607b9ebf28a format: 2 features: layering, striping,

[ceph-users] Re: CEPH 16.2.x: disappointing I/O performance

2021-10-06 Thread Zakhar Kirpichenko
These are valid points, thank you for the input! /Z On Wed, Oct 6, 2021 at 11:39 AM Stefan Kooman wrote: > On 10/6/21 09:23, Zakhar Kirpichenko wrote: > > Hi, > > > > Indeed, that's a lot of CPU and RAM, the idea was to put sufficient > > resources in case we want to

[ceph-users] Re: CEPH 16.2.x: disappointing I/O performance

2021-10-06 Thread Zakhar Kirpichenko
, 2021 at 11:16 AM Janne Johansson wrote: > Den ons 6 okt. 2021 kl 10:11 skrev Zakhar Kirpichenko : > > > > I've initially disabled power-saving features, which nicely improved the > > network latency. > > > > Btw, the first interesting find: I enabled 'rbd_balance

[ceph-users] Re: CEPH 16.2.x: disappointing I/O performance

2021-10-06 Thread Zakhar Kirpichenko
I've initially disabled power-saving features, which nicely improved the network latency. Btw, the first interesting find: I enabled 'rbd_balance_parent_reads' on the clients, and single-thread reads now scale much better, I routinely get similar readings from a single disk doing 4k reads with 1

[ceph-users] Re: CEPH 16.2.x: disappointing I/O performance

2021-10-06 Thread Zakhar Kirpichenko
:06, Zakhar Kirpichenko wrote: > > Hi, > > > > I built a CEPH 16.2.x cluster with relatively fast and modern hardware, > and > > its performance is kind of disappointing. I would very much appreciate an > > advice and/or pointers :-) > > > > The hardware

[ceph-users] Re: CEPH 16.2.x: disappointing I/O performance

2021-10-06 Thread Zakhar Kirpichenko
disabling the HDD write cache can > actually improve latency quite substantially (e.g. > https://ceph-users.ceph.narkive.com/UU9QMu9W/disabling-write-cache-on-sata-hdds-reduces-write-latency-7-times) > - might be worth a try > > > On Wed, 6 Oct 2021 at 10:07, Zakhar Kirpichenko w

[ceph-users] Re: CEPH 16.2.x: disappointing I/O performance

2021-10-06 Thread Zakhar Kirpichenko
Actually recent versions of virtio-blk support discard as well. In general, I found that virtio-scsi isn't worth it. I appreciate your input though, thanks! /Z On Wed, 6 Oct 2021, 09:03 Janne Johansson, wrote: > Den ons 6 okt. 2021 kl 06:37 skrev Zakhar Kirpichenko : > > Got it. I d

[ceph-users] Re: CEPH 16.2.x: disappointing I/O performance

2021-10-05 Thread Zakhar Kirpichenko
ume / attachment / VM, or enables the client-side admin socket, that > sort of throttling can be very visually apparent. > > > > On Oct 5, 2021, at 8:35 PM, Zakhar Kirpichenko wrote: > > > > Hi! > > > > The clients are KVM VMs, there's QEMU/libvirt impact for sure. I

[ceph-users] Re: CEPH 16.2.x: disappointing I/O performance

2021-10-05 Thread Zakhar Kirpichenko
ndreds of PGs. But QD=1 and small block sizes are going to > limit your throughput. > > What are your clients? Are they bare metal? Are they VMs? If they’re > VMs, do you have QEMU/libvirt throttling in play? I see that a lot. > > > On Oct 5, 2021, at 2:06 PM, Zakhar Kirp

[ceph-users] Re: CEPH 16.2.x: disappointing I/O performance

2021-10-05 Thread Zakhar Kirpichenko
y math 7k IOPS at 4k should only be >> 27MiB/sec - not sure how the 120MiB/sec was achieved >> The read benchmark seems in line with 13k IOPS at 4k making around >> 52MiB/sec bandwidth which again is expected. >> >> >> On Wed, 6 Oct 2021 at 04:08, Zakhar Kirpichenko

[ceph-users] Re: CEPH 16.2.x: disappointing I/O performance

2021-10-05 Thread Zakhar Kirpichenko
eved > The read benchmark seems in line with 13k IOPS at 4k making around > 52MiB/sec bandwidth which again is expected. > > > On Wed, 6 Oct 2021 at 04:08, Zakhar Kirpichenko wrote: > >> Hi, >> >> I built a CEPH 16.2.x cluster with relatively fast and

[ceph-users] Re: *****SPAM***** Re: CEPH 16.2.x: disappointing I/O performance

2021-10-05 Thread Zakhar Kirpichenko
Aren't all writes to bluestore turned into sequential writes? /Z On Tue, 5 Oct 2021, 20:05 Marc, wrote: > > Hi Zakhar, > > > using 16 threads) are not. Literally every storage device in my setup > > can read and write at least 200+ MB/s sequentially, so I'm trying to > > find an explanation

[ceph-users] Re: CEPH 16.2.x: disappointing I/O performance

2021-10-05 Thread Zakhar Kirpichenko
Hi Marc, Many thanks for your comment! As I mentioned, rados bench results are more or less acceptable and explainable. RBD clients writing at ~120 MB/st tops (regardless of the number of threads or block size btw) and reading ~50 MB/s in a single thread (I managed to read over 500 MB/s using 16

[ceph-users] CEPH 16.2.x: disappointing I/O performance

2021-10-05 Thread Zakhar Kirpichenko
Hi, I built a CEPH 16.2.x cluster with relatively fast and modern hardware, and its performance is kind of disappointing. I would very much appreciate an advice and/or pointers :-) The hardware is 3 x Supermicro SSG-6029P nodes, each equipped with: 2 x Intel(R) Xeon(R) Gold 5220R CPUs 384 GB

[ceph-users] Re: Adding cache tier to an existing objectstore cluster possible?

2021-09-20 Thread Zakhar Kirpichenko
our VMs was almost impossible. But this > is an RBD cache so I can't tell how the other protocols perform with a > cache tier. > > > Zitat von Zakhar Kirpichenko : > > > Hi, > > > > You can arbitrarily add or remove the cache tier, there's no problem with > &

[ceph-users] Re: Adding cache tier to an existing objectstore cluster possible?

2021-09-19 Thread Zakhar Kirpichenko
Hi, You can arbitrarily add or remove the cache tier, there's no problem with that. The problem is that cache tier doesn't work well, I tried it in front of replicated and EC-pools with very mixed results: when it worked there wasn't as much of a speed/latency benefit as one would expect from

<    1   2