[ceph-users] Re: benchmark Ceph

2020-09-14 Thread rainning
What is your Ceph version? From the test results you posted, your environment's performance is okay in regard of your setup. But there are definitely many things that can be tuned to get you better number. I normally use top, iostat, pidstat, vmstat, dstat, iperf3, blktrace, netmon, ceph admin

[ceph-users] Re: Nautilus: rbd image stuck unaccessible after VM restart

2020-09-14 Thread Cashapp Failed
The cash app is otherwise called square cash, which is a companion to a friend app that permits the clients to transfer assets by connecting their financial balance. It is fundamentally a versatile app-based stage for installment administrations. It is an a lot less difficult app when contrasted

[ceph-users] Re: benchmark Ceph

2020-09-14 Thread Tony Liu
Here is the test inside VM. # fio --name=test --ioengine=libaio --numjobs=1 --runtime=30 \ --direct=1 --size=2G --end_fsync=1 \ --rw=read --bs=4K --iodepth=1 test: (groupid=0, jobs=1): err= 0: pid=14615: Mon Sep 14 21:50:55 2020 read: IOPS=320

[ceph-users] Re: Disk consume for CephFS

2020-09-14 Thread tri
I suggest trying the rsync --sparse option. Typically, qcow2 files (tend to be large) are sparse files. Without the sparse option, the files expand in their destination. September 14, 2020 6:15 PM, fotof...@gmail.com wrote: > Hello. > > I'm using the Nautilus Ceph version for some huge folder

[ceph-users] Re: benchmark Ceph

2020-09-14 Thread rainning
Can you post the fio results with the ioengine using libaio? From what you posted, it seems to me that the read test hit cache. And the write performance was not good, the latency was too high (~35.4ms) while the numjobs and iodepth both were 1. Did you monitor system stat on both side (VM/Compu

[ceph-users] benchmark Ceph

2020-09-14 Thread Tony Liu
Hi, I have a 3-OSD-node Ceph cluster with 1 480GB SSD and 8 x 2TB 12Gpbs SAS HDD on each node, to provide storage to a OpenStack cluster. Both public and cluster networks are 2x10G. WAL and DB of each OSD is on SSD and they share the same 60GB partition. I run fio with different combinations of o

[ceph-users] Re: Disk consume for CephFS

2020-09-14 Thread Nathan Fish
What about hardlinks, are there any of those? Are there lots of directories or tiny (<4k) files? Also, size=2 is not very safe. You want size=3, min_size=2 if you are doing replication. On Mon, Sep 14, 2020 at 6:15 PM wrote: > > Hello. > > I'm using the Nautilus Ceph version for some huge folder

[ceph-users] Re: Syncing cephfs from Ceph to Ceph

2020-09-14 Thread Stefan Kooman
On 2020-09-09 15:51, Eugen Block wrote: > Hi Simon, > >> What about the idea of creating the cluster over two data centers? >> Would it be possible to modify the crush map, so one pool gets >> replicated over those two data centers and if one fails, the other one >> would still be functional? > >

[ceph-users] Re: New pool with SSD OSDs

2020-09-14 Thread Stefan Kooman
On 2020-09-14 17:51, Tecnologia Charne.Net wrote: > Thanks Stefan! > Compiling crush map by hand on production cluster makes me sweat > but we like to take risks, don't we? I the crushtool says it's OK, I guess it's OK ;-). But yeah, that's the most powerful operation one can perform on a clus

[ceph-users] Disk consume for CephFS

2020-09-14 Thread fotofors
Hello. I'm using the Nautilus Ceph version for some huge folder with approximately 1.7TB of files.I created the filesystem and started to copy files via rsync. However, I've had to stop the process, because Ceph shows me that the new size of the folder is almost 6TB. I double checked the repli

[ceph-users] Re: Unable to start mds when creating cephfs volume with erasure encoding data pool

2020-09-14 Thread Patrick Donnelly
On Sun, Sep 13, 2020 at 1:26 PM wrote: > > Hi all, > > I'm using ceph Octopus version and deployed it using cephadm. The ceph > documentation provides 2 ways for creating a new cephfs volume: > > 1. via "ceph fs volume create ..." - I can use this and it works fine with > the MDS automatically

[ceph-users] Re: New pool with SSD OSDs

2020-09-14 Thread Stefan Kooman
On 2020-09-14 16:09, André Gemünd wrote: > Same happened to us two weeks ago using nautilus, although we added the rules > and storage classes. I think this post [1] from Wido den Hollander might be useful information. That way you can avoid data movement if data is already on hdd. Gr. Stefan

[ceph-users] Re: Nautilus Scrub and deep-Scrub execution order

2020-09-14 Thread Robin H. Johnson
On Mon, Sep 14, 2020 at 11:40:22AM -, Johannes L wrote: > Hello Ceph-Users > > after upgrading one of our clusters to Nautilus we noticed the x pgs not > scrubbed/deep-scrubbed in time warnings. > Through some digging we found out that it seems like the scrubbing takes > place at random and

[ceph-users] Re: Issues with the ceph-bluestore-tool during cluster upgrade from Mimic to Nautilus

2020-09-14 Thread Jean-Philippe Méthot
Alright, here’s the full log file. Jean-Philippe Méthot Senior Openstack system administrator Administrateur système Openstack sénior PlanetHoster inc. 4414-4416 Louis B Mayer Laval, QC, H7P 0G1, Canada TEL : +1.514.802.1644 - Poste : 2644 FAX : +1.514.612.0678 CA/US : 1.855.774.4678 FR : 01 7

[ceph-users] Re: ceph pgs inconsistent, always the same checksum

2020-09-14 Thread Welby McRoberts
Hi Igor We'll take a look at disabling swap on the nodes and see if that improves the situation. Having checked across all osds we're not seeing bluestore_reads_with_retries as anything other than a zero value. We get the error anywhere from 3 - 10 occurrences of the error a week, but it's usuall

[ceph-users] Re: New pool with SSD OSDs

2020-09-14 Thread Tecnologia Charne.Net
Thanks Stefan! Compiling crush map by hand on production cluster makes me sweat but we like to take risks, don't we? El 14/9/20 a las 11:48, Stefan Kooman escribió: On 2020-09-14 16:09, André Gemünd wrote: Same happened to us two weeks ago using nautilus, although we added the rules and s

[ceph-users] Re: New pool with SSD OSDs

2020-09-14 Thread André Gemünd
Oh, yes, I think this would have helped indeed! Thanks for pointing it out. Greetings André - Am 14. Sep 2020 um 16:48 schrieb Stefan Kooman ste...@bit.nl: > On 2020-09-14 16:09, André Gemünd wrote: >> Same happened to us two weeks ago using nautilus, although we added the rules >> and sto

[ceph-users] Re: Choosing suitable SSD for Ceph cluster

2020-09-14 Thread Seena Fallah
Yes I'm planning to use only 60% - 70% of my disks and pools like buckets.index doesn't grow too much and don't need much space! I'm just trying to make this pool faster because I see it sometimes needs 1Milion iops and I think NVME is a good option for this pool. But finding a good datacenter NVME

[ceph-users] Re: Choosing suitable SSD for Ceph cluster

2020-09-14 Thread Martin Verges
Hello, Please keep in mind that you can have significant operational problems if you choose too small OSDs. Sometimes your OSDs require >40G for osdmaps/pgmaps/... and the smaller you OSD, the more likely it will be a problem as Ceph is totally unable to deal with full disks and break apart. -- M

[ceph-users] Re: New pool with SSD OSDs

2020-09-14 Thread Tecnologia Charne.Net
Exactly! I created a replicated-hdd rule and set it to an existing small pool without any changes on OSDs (all HDD) and PGs starts migration... It seems like new rules forces migrations... El 14/9/20 a las 11:09, André Gemünd escribió: Same happened to us two weeks ago using nautilus, although

[ceph-users] Re: Issues with the ceph-bluestore-tool during cluster upgrade from Mimic to Nautilus

2020-09-14 Thread Igor Fedotov
Thanks! Now got the root cause. The fix is on its way... Meanwhile you might want to try to workaround the issue via setting "bluestore_hybrid_alloc_mem_cap" to 0 or using different allocator, e.g. avl for bluestore_allocator (and optionally for bluefs_allocator too). Hope this helps, Igor

[ceph-users] ceph-container: docker restart, mon's unable to join

2020-09-14 Thread Stefan Kooman
Hi, In an attempt to get a (test) Mimic cluster running on Ubuntu 20.04 we are using docker with ceph-container images (ceph/daemon:latest-mimic). Deploying monitors and mgrs works fine. If however a monitor container gets stopped and started (i.e. docker restart) two out of three (with exception

[ceph-users] Re: New pool with SSD OSDs

2020-09-14 Thread André Gemünd
Same happened to us two weeks ago using nautilus, although we added the rules and storage classes. - Am 14. Sep 2020 um 16:02 schrieb Marc Roos m.r...@f1-outsourcing.eu: > I did the same, 1 or 2 years ago, creating a replicated_ruleset_hdd and > replicated_ruleset_ssd. Eventhough I did not

[ceph-users] Re: New pool with SSD OSDs

2020-09-14 Thread Marc Roos
I did the same, 1 or 2 years ago, creating a replicated_ruleset_hdd and replicated_ruleset_ssd. Eventhough I did not have any ssd's on any of the nodes at that time, adding this hdd type criteria made pg's migrate. I thought it was strange that this happens on a hdd only cluster, so I mention

[ceph-users] Re: Choosing suitable SSD for Ceph cluster

2020-09-14 Thread response
https://www.kingston.com/unitedkingdom/en/ssd/dc1000b-data-center-boot-ssd look good for your purpose. - Original Message - From: "Seena Fallah" To: "Виталий Филиппов" Cc: "Anthony D'Atri" , "ceph-users" Sent: Monday, September 14, 2020 2:47:14 PM Subject: [ceph-users] Re: Choosing

[ceph-users] Re: Choosing suitable SSD for Ceph cluster

2020-09-14 Thread Seena Fallah
Thanks for the sheet. I need a low space disk for my use case (around 240GB). Do you have any suggestions with M.2 and capacitors? On Mon, Sep 14, 2020 at 6:11 PM wrote: > There's also Micron 7300 Pro/Max. Please benchmark it like described here > https://docs.google.com/spreadsheets/d/1E9-eXjzs

[ceph-users] Re: Choosing suitable SSD for Ceph cluster

2020-09-14 Thread vitalif
There's also Micron 7300 Pro/Max. Please benchmark it like described here https://docs.google.com/spreadsheets/d/1E9-eXjzsKboiCCX-0u0r5fAjjufLKayaut_FOPxYZjc/edit (https://docs.google.com/spreadsheets/d/1E9-eXjzsKboiCCX-0u0r5fAjjufLKayaut_FOPxYZjc/edit) and send me the results if you get one :)

[ceph-users] Re: Choosing suitable SSD for Ceph cluster

2020-09-14 Thread vitalif
Samsung PM983 M.2 I want to have a separate disk for buckets index pool and all of my server bays are full and I should use m2 storage devices. Also the bucket index doesn't need much space so I plan to have a 6x device with replica 3 for it. Each disk could be 240GB to not waste space but ther

[ceph-users] New pool with SSD OSDs

2020-09-14 Thread Tecnologia Charne.Net
Hello! We have a Ceph cluster with 30 HDD 4 TB in 6 hosts, only for RBD. Now, we're receiving other 6 servers with 6 SSD 2 TB each and we want to create a separate pool for RBD on SSD, and let unused and backup volumes stays in HDD. I have some questions: As I am only using "replicated_ru

[ceph-users] Re: ceph rbox test on passive compressed pool

2020-09-14 Thread Marc Roos
> mail/b875f40571f1545ff43052412a8e mtime 2020-09-06 > 16:25:53.00, > size 63580 > mail/e87c120b19f1545ff43052412a8e mtime 2020-09-06 > 16:24:25.00, > size 525 Hi David, How is this going? To me this looks more like deduplication than compression. This

[ceph-users] Re: ceph rbox test on passive compressed pool

2020-09-14 Thread Marc Roos
> mail/b875f40571f1545ff43052412a8e mtime 2020-09-06 > 16:25:53.00, > size 63580 > mail/e87c120b19f1545ff43052412a8e mtime 2020-09-06 > 16:24:25.00, > size 525 Hi David, How is this going. To me this looks more like deduplication than compression. This

[ceph-users] Re: virtual machines crashes after upgrade to octopus

2020-09-14 Thread Jason Dillaman
On Mon, Sep 14, 2020 at 5:13 AM Lomayani S. Laizer wrote: > > Hello, > Last week i got time to try debug crashes of these vms > > Below log includes rados debug which i left last time > > https://storage.habari.co.tz/index.php/s/AQEJ7tQS7epC4Zn > > I have observed the following with these settin

[ceph-users] Nautilus Scrub and deep-Scrub execution order

2020-09-14 Thread Johannes L
Hello Ceph-Users after upgrading one of our clusters to Nautilus we noticed the x pgs not scrubbed/deep-scrubbed in time warnings. Through some digging we found out that it seems like the scrubbing takes place at random and doesn't take the age of the last scrub/deep-scrub into consideration. I

[ceph-users] Re: ceph pgs inconsistent, always the same checksum

2020-09-14 Thread Igor Fedotov
Hi David, you might want to try to disable swap for your nodes. Look like there is some implicit correlation between such read errors and enabled swapping. Also wondering whether you can observe non-zero values for "bluestore_reads_with_retries" performance counters over your OSDs. How wide-

[ceph-users] Re: Issues with the ceph-bluestore-tool during cluster upgrade from Mimic to Nautilus

2020-09-14 Thread Igor Fedotov
Well, I can see duplicate admin socket command registration/de-registration (and the second de-registration asserts) but don't understand how this could happen. Would you share the full log, please? Thanks, Igor On 9/11/2020 7:26 PM, Jean-Philippe Méthot wrote: Here’s the out file, as requ

[ceph-users] Re: ceph-osd performance on ram disk

2020-09-14 Thread George Shuklin
On 11/09/2020 17:44, Mark Nelson wrote: On 9/11/20 4:15 AM, George Shuklin wrote: On 10/09/2020 19:37, Mark Nelson wrote: On 9/10/20 11:03 AM, George Shuklin wrote: ... Are there any knobs to tweak to see higher performance for ceph-osd? I'm pretty sure it's not any kind of leveling, GC or

[ceph-users] Orchestrator & ceph osd purge

2020-09-14 Thread Robert Sander
Hi, is it correct that when using the orchestrator to deploy and manage a cluster you should not use "ceph osd purge" any more as the orchestrator then is not able to find the OSD for the "ceph orch osd rm" operation? Regards -- Robert Sander Heinlein Support GmbH Schwedter Str. 8/9b, 10119 Berl

[ceph-users] Re: Is it possible to assign osd id numbers?

2020-09-14 Thread George Shuklin
On 11/09/2020 22:43, Shain Miley wrote: Thank you for your answer below. I'm not looking to reuse them as much as I am trying to control what unused number is actually used. For example if I have 20 osds and 2 have failed...when I replace a disk in one server I don't want it to automatically

[ceph-users] virtual machines crashes after upgrade to octopus

2020-09-14 Thread Lomayani S. Laizer
Hello, Last week i got time to try debug crashes of these vms Below log includes rados debug which i left last time https://storage.habari.co.tz/index.php/s/AQEJ7tQS7epC4Zn I have observed the following with these settings in openstack and ceph disk_cachemodes="network=writeback" is set in op