[ceph-users] Re: Fwd: BlueFS spillover yet again

2020-02-05 Thread Janne Johansson
Den ons 5 feb. 2020 kl 17:27 skrev Vladimir Prokofev : > Thank you for the insight. > > If you're using the default options for rocksdb, then the size of L3 will > be 25GB > Where this number comes from? Any documentation I can read? > I want to have a better understanding on how DB size is

[ceph-users] Re: osd is immidietly down and uses CPU full.

2020-02-05 Thread 西宮牧人
Hi, everyone. The problem was solved. A PG's epoch of active osds was dfferent from one of acting osds. When I removed head and TEMP directory of the PG from active osds, blocked request vanished. Thanks to your kindness. 2020年2月6日(木) 4:19 : > What do you guys Think about > > Ceph osd Set

[ceph-users] Re: Understanding Bluestore performance characteristics

2020-02-05 Thread Stefan Kooman
Quoting vita...@yourcmc.ru (vita...@yourcmc.ru): > SSD (block.db) partition contains object metadata in RocksDB so it probably > loads the metadata before modifying objects (if it's not in cache yet). Also > it sometimes performs compaction which also results in disk reads and > writes. There are

[ceph-users] Re: Problem with OSD - stuck in CPU loop after rbd snapshot mount

2020-02-05 Thread Jan Pekař - Imatic
Hi, because load to my OSD's stopped approximately after 12 hours and everything returned to normal, I did another test. I have chosen another rbd image and some old snapshot and I tried to mount it - now without cloning it. rbd-nbd map succeeded, but situation from my first test repeated,

[ceph-users] Re: Fwd: BlueFS spillover yet again

2020-02-05 Thread Wido den Hollander
On 2/5/20 2:21 PM, Vladimir Prokofev wrote: > Cluster upgraded from 12.2.12 to 14.2.5. All went smooth, except BlueFS > spillover warning. > We create OSDs with ceph-deploy, command goes like this: > ceph-deploy osd create --bluestore --data /dev/sdf --block-db /dev/sdb5 > --block-wal /dev/sdb6

[ceph-users] Re: osd is immidietly down and uses CPU full.

2020-02-05 Thread ceph
What do you guys Think about Ceph osd Set noout/Down And See if the osd will become healthy? Another idea which is in my mind is to remove the sayed osd from the Cluster... As long as the other osds on the same node dont have an issue i guess the disk has a Problem... Just my 2 Cents -

[ceph-users] Re: recovery_unfound

2020-02-05 Thread Chad William Seys
Hi Jake, In case it's relevant we are on Nautilus 14.2.6, not Mimic. Yeah, my guess is that it is multiversion. Also, my scenario simply should not have lost any data, so don't kick yourself too hard. This command supposedly check cephfs for damaged files. It came back with nothing for

[ceph-users] Re: Strange performance drop and low oss performance

2020-02-05 Thread Marc Roos
I didn't have such drop in performance testing 'rados bench 360 write -p rbd' on 3x replicated (slow)hdd pool. Sort of near the average, sometimes drops to 90. But I guess the test hits than an osd that is scrubbing and being used by other processes. -Original Message- Sent: 05

[ceph-users] Re: Fwd: BlueFS spillover yet again

2020-02-05 Thread Vladimir Prokofev
Thank you for the insight. > If you're using the default options for rocksdb, then the size of L3 will be 25GB Where this number comes from? Any documentation I can read? I want to have a better understanding on how DB size is calculated. ср, 5 февр. 2020 г. в 18:53, Moreno, Orlando : > Hi

[ceph-users] Re: Strange performance drop and low oss performance

2020-02-05 Thread quexian da
Thanks for your valuable answer about write cache! For object gateway, the performance is got by `swift-bench -t 64` which uses 64 threads concurrently. Will the radosgw and http overhead be so significant (94.5MB/s to 26MB/s for cluster1) when multiple threads are used? Thanks in advance! On

[ceph-users] Re: recovery_unfound

2020-02-05 Thread Jake Grimmett
Hi Chad, In case it's relevant we are on Nautilus 14.2.6, not Mimic. I've followed Paul's advice and issued a "ceph osd down XXX" command for the primary osd in each affected pg. I've also tried doing a systemctl restart for several of the primary osd's, again with no apparent effect.

[ceph-users] Re: Fwd: BlueFS spillover yet again

2020-02-05 Thread Moreno, Orlando
Hi Vladimir, If you're using the default options for rocksdb, then the size of L3 will be 25GB. Since your block-db is only 20GB and L3 can only be filled if the entire level's size is available, bluefs will begin spillover. Like Igor said, having 30GB+ is recommended if you want to host up to

[ceph-users] Re: Strange performance drop and low oss performance

2020-02-05 Thread Janne Johansson
Den ons 5 feb. 2020 kl 16:19 skrev quexian da : > Thanks for your valuable answer! > Is the write cache specific to ceph? Could you please provide some links > to the documentation about the write cache? Thanks! > > It is all the possible caches used by ceph, by the device driver, the filesystem

[ceph-users] Re: Strange performance drop and low oss performance

2020-02-05 Thread quexian da
Thanks for your valuable answer! Is the write cache specific to ceph? Could you please provide some links to the documentation about the write cache? Thanks! Do you have any idea about the slow oss speed? Is it normal that the write performance of object gateway is slower than that of rados

[ceph-users] Re: Strange performance drop and low oss performance

2020-02-05 Thread Janne Johansson
Den ons 5 feb. 2020 kl 11:14 skrev quexian da : > Hello, > > I'm a beginner on ceph. I set up three ceph clusters on google cloud. > Cluster1 has three nodes and each node has three disks. Cluster2 has three > nodes and each node has two disks. Cluster3 has five nodes and each node > has five

[ceph-users] Re: Fwd: BlueFS spillover yet again

2020-02-05 Thread Igor Fedotov
Hi Vladimir, there were a plenty of discussions/recommendations around db volume size selection here. In short it's advised to have DB volume of 30 - 64GB for most of use cases. Thanks, Igor On 2/5/2020 4:21 PM, Vladimir Prokofev wrote: Cluster upgraded from 12.2.12 to 14.2.5. All went

[ceph-users] Fwd: BlueFS spillover yet again

2020-02-05 Thread Vladimir Prokofev
Cluster upgraded from 12.2.12 to 14.2.5. All went smooth, except BlueFS spillover warning. We create OSDs with ceph-deploy, command goes like this: ceph-deploy osd create --bluestore --data /dev/sdf --block-db /dev/sdb5 --block-wal /dev/sdb6 ceph-osd3 where block-db and block-wal are SSD

[ceph-users] Re: slow using ISCSI - Help-me

2020-02-05 Thread Gesiel Galvão Bernardes
Hi, Do you have any suggestions on where I can look? Regards, Gesiel Em dom., 2 de fev. de 2020 às 00:37, Gesiel Galvão Bernardes < gesiel.bernar...@gmail.com> escreveu: > Hi, > > Just now was possible continue this. Below is the information required. > Thanks advance, > > Gesiel > > Em seg.,

[ceph-users] Re: [Ceph-community] HEALTH_WARN - daemons have recently crashed

2020-02-05 Thread Sage Weil
[Moving this to ceph-users@ceph.io] This looks like https://tracker.ceph.com/issues/43365, which *looks* like it is an issue with the standard libraries in ubuntu 18.04. One user said: "After upgrading our monitor Ubuntu 18.04 packages (apt-get upgrade) with the 5.3.0-26-generic kernel, it

[ceph-users] Re: Understanding Bluestore performance characteristics

2020-02-05 Thread vitalif
Hi, This helped to disable deferred writes in my case: bluestore_min_alloc_size=4096 bluestore_prefer_deferred_size=0 bluestore_prefer_deferred_size_ssd=0 If you already deployed your OSDs with min_alloc_size=4K then you don't need to redeploy them again. Hi Vitality, I completely

[ceph-users] Re: OSDs crashing

2020-02-05 Thread Raymond Clotfelter
I have found that if I set norecovery then I can get almost all OSDs to come up and stay up, but the moment I unset norecovery so that the cluster can heal itself, tons of OSDs go offline again. The OSD host servers have plenty of available RAM, they are not maxing on CPU or I/O near as I can

[ceph-users] Re: Bluestore cache parameter precedence

2020-02-05 Thread Igor Fedotov
Hi Boris, not sure I understand why do you think it needs correction. IMO _use_rotational_settings() function does what you're suggesting plus it has some additional logic to be able to enforce specific device type. The latter is intended primarily for testing/debugging and enabled via

[ceph-users] Re: Understanding Bluestore performance characteristics

2020-02-05 Thread Bradley Kite
Hi Vitality, I completely destroyed the test cluster and re-deployed it after changing these settings but it did not make a difference - there are still a high number of deferred writes. Regards -- Brad. On Wed, 5 Feb 2020 at 10:55, wrote: > min_alloc_size can't be changed after formatting an

[ceph-users] Re: Understanding Bluestore performance characteristics

2020-02-05 Thread vitalif
min_alloc_size can't be changed after formatting an OSD, and yes, bluestore defers all writes that are < min_alloc_size. And default min_alloc_size_ssd is 16KB. ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to

[ceph-users] Re: Cephalocon Seoul is canceled

2020-02-05 Thread Romit Misra
Hi Sage, Just read the news on Cancellation of Cephlacon 2020, although the site is still status quo/ Double checking that we can proceed with the cancellation of logistics for South Korea Thanks Romit On Tue, Feb 4, 2020 at 11:02 PM wrote: > Send ceph-users mailing list submissions to >

[ceph-users] Strange performance drop and low oss performance

2020-02-05 Thread quexian da
Hello, I'm a beginner on ceph. I set up three ceph clusters on google cloud. Cluster1 has three nodes and each node has three disks. Cluster2 has three nodes and each node has two disks. Cluster3 has five nodes and each node has five disks. All disks are HDD. Disk speed shown by `dd if=/dev/zero

[ceph-users] Re: osd_memory_target ignored

2020-02-05 Thread Stefan Kooman
Quoting Frank Schilder (fr...@dtu.dk): > Dear Stefan, > > is it possible that there is a mistake in the documentation or a bug? Out of > curiosity, I restarted one of these OSDs and the memory usage starts going up: > > ceph 881203 15.4 4.0 6201580 5344764 ? Sl 09:18 6:38 >

[ceph-users] Re: osd_memory_target ignored

2020-02-05 Thread Frank Schilder
Dear Stefan, is it possible that there is a mistake in the documentation or a bug? Out of curiosity, I restarted one of these OSDs and the memory usage starts going up: ceph 881203 15.4 4.0 6201580 5344764 ? Sl 09:18 6:38 /usr/bin/ceph-osd --cluster ceph -f -i 243 --setuser ceph

[ceph-users] Re: osd_memory_target ignored

2020-02-05 Thread Frank Schilder
Hi Stefan, its all at the defaults it seems: [root@gnosis ~]# ceph config get osd.243 bluestore_cache_size 0 [root@gnosis ~]# ceph config get osd.243 bluestore_cache_size_ssd 3221225472 I explicitly removed the old settings with commands like ceph config rm osd.243 bluestore_cache_size Best