Re: [ceph-users] Remove rbd image after interrupt of deletion command

2019-06-11 Thread Igor Podlesny
On Tue, 11 Jun 2019 at 14:46, Sakirnth Nagarasa wrote: > On 6/7/19 3:35 PM, Jason Dillaman wrote: [...] > > Can you run "rbd rm --log-to-stderr=true --debug-rbd=20 > > ${POOLNAME}/${IMAGE}" and provide the logs via pastebin.com? > > > >> Cheers, > >> Sakirnth > > It is not necessary anymore the

Re: [ceph-users] balancer module makes OSD distribution worse

2019-06-08 Thread Igor Podlesny
On Thu, 6 Jun 2019 at 03:01, Josh Haft wrote: > > Hi everyone, > > On my 13.2.5 cluster, I recently enabled the ceph balancer module in > crush-compat mode. Why did you choose compat mode? Don't you want to try another one instead? -- End of message. Next message?

Re: [ceph-users] Can I limit OSD memory usage?

2019-06-08 Thread Igor Podlesny
On Sat, 8 Jun 2019 at 04:35, Sergei Genchev wrote: > > Hi, > My OSD processes are constantly getting killed by OOM killer. My > cluster has 5 servers, each with 18 spinning disks, running 18 OSD > daemons in 48GB of memory. > I was trying to limit OSD cache, according to >

[ceph-users] Any CEPH's iSCSI gateway users?

2019-05-28 Thread Igor Podlesny
What is your experience? Does it make sense to use it -- is it solid enough or beta quality rather (both in terms of stability and performance)? I've read it was more or less packaged to work with RHEL. Does it hold true still? What's the best way to install it on, say, CentOS or Debian/Ubuntu?

Re: [ceph-users] Massive TCP connection on radosgw

2019-05-22 Thread Igor Podlesny
On Wed, 22 May 2019 at 20:32, Torben Hørup wrote: > > Which states are all these connections in ? > > ss -tn That set of the args won't display anything but ESTAB-lished conn-s.. One typically needs `-atn` instead. -- End of message. Next message?

Re: [ceph-users] Nautilus, k+m erasure coding a profile vs size+min_size

2019-05-21 Thread Igor Podlesny
On Tue, 21 May 2019 at 19:32, Yoann Moulin wrote: > > >> I am doing some tests with Nautilus and cephfs on erasure coding pool. [...] > > http://lists.ceph.com/pipermail/ceph-users-ceph.com/2019-May/034867.html > > Oh thanks, I missed that thread, make sense. I agree with some comment that > it

Re: [ceph-users] Tip for erasure code profile?

2019-05-03 Thread Igor Podlesny
On Fri, 3 May 2019 at 22:46, Robert Sander wrote: > The cluster spans 2 rooms ... > The failure domain would be the room level ... > Is that even possible with erasure coding? Sure deal but you'd need slightly more rooms then. For e. g., minimal EC(2, 1) means (2 + 1) rooms. -- End of message.

Re: [ceph-users] Unexplainable high memory usage OSD with BlueStore

2019-05-03 Thread Igor Podlesny
On Fri, 3 May 2019 at 21:39, Mark Nelson wrote: [...] > > [osd] > > ... > > bluestore_allocator = bitmap > > bluefs_allocator = bitmap > > > > I would restart the nodes one by one and see, what happens. > > If you are using 12.2.11 you likely still have the old bitmap allocator Would those

Re: [ceph-users] Unexplainable high memory usage OSD with BlueStore

2019-05-03 Thread Igor Podlesny
On Fri, 3 May 2019 at 13:38, Denny Fuchs wrote: [...] > If I understand correct: I should try to set bitmap allocator That's among one of the options I mentioned. Another one was to try using jemalloc (re-read my emails). > [osd] > ... > bluestore_allocator = bitmap > bluefs_allocator = bitmap

Re: [ceph-users] Unexplainable high memory usage OSD with BlueStore

2019-05-02 Thread Igor Podlesny
On Fri, 3 May 2019 at 05:12, Mark Nelson wrote: [...] > > -- https://www.kernel.org/doc/Documentation/vm/transhuge.txt > > Why are you quoting the description for the madvise setting when that's > clearly not what was set in the case I just showed you? Similarly why(?) are you telling us it must

Re: [ceph-users] Unexplainable high memory usage OSD with BlueStore

2019-05-02 Thread Igor Podlesny
On Fri, 3 May 2019 at 01:29, Mark Nelson wrote: > On 5/2/19 11:46 AM, Igor Podlesny wrote: > > On Thu, 2 May 2019 at 05:02, Mark Nelson wrote: > > [...] > >> FWIW, if you still have an OSD up with tcmalloc, it's probably worth > >> looking at the heap stats to se

Re: [ceph-users] Unexplainable high memory usage OSD with BlueStore

2019-05-02 Thread Igor Podlesny
On Thu, 2 May 2019 at 05:02, Mark Nelson wrote: [...] > FWIW, if you still have an OSD up with tcmalloc, it's probably worth > looking at the heap stats to see how much memory tcmalloc thinks it's > allocated vs how much RSS memory is being used by the process. It's > quite possible that there

Re: [ceph-users] Unexplainable high memory usage OSD with BlueStore

2019-05-01 Thread Igor Podlesny
On Tue, 30 Apr 2019 at 20:56, Igor Podlesny wrote: > On Tue, 30 Apr 2019 at 19:10, Denny Fuchs wrote: > [..] > > Any suggestions ? > > -- Try different allocator. Ah, BTW, except memory allocator there's another option: recently backported bitmap allocator. Igor Fedoto

Re: [ceph-users] Data distribution question

2019-04-30 Thread Igor Podlesny
On Wed, 1 May 2019 at 01:58, Dan van der Ster wrote: > On Tue, Apr 30, 2019 at 8:26 PM Igor Podlesny wrote: [...] > All of the clients need to be luminous our newer: > > # ceph osd set-require-min-compat-client luminous > > You need to enable the module: > > # ceph mg

Re: [ceph-users] Data distribution question

2019-04-30 Thread Igor Podlesny
On Wed, 1 May 2019 at 01:26, Igor Podlesny wrote: > On Wed, 1 May 2019 at 01:01, Dan van der Ster wrote: > >> > The upmap balancer in v12.2.12 works really well... Perfectly uniform on > >> > our clusters. > >> > >> mode upmap ? > > > > y

Re: [ceph-users] Data distribution question

2019-04-30 Thread Igor Podlesny
On Wed, 1 May 2019 at 01:26, Jack wrote: > If those pools are useless, you can: > - drop them As Dan pointed out it's unlikely of having any effect. The thing is imbalance is a "property" of a pool, I'd suppose that most often -- is the most loaded one (or of a few most loaded ones). Not that

Re: [ceph-users] Data distribution question

2019-04-30 Thread Igor Podlesny
On Wed, 1 May 2019 at 01:01, Dan van der Ster wrote: >> > The upmap balancer in v12.2.12 works really well... Perfectly uniform on >> > our clusters. >> >> mode upmap ? > > yes, mgr balancer, mode upmap. I see. Was it a matter of just: 1) ceph balancer mode upmap 2) ceph balancer on or were

Re: [ceph-users] Data distribution question

2019-04-30 Thread Igor Podlesny
On Wed, 1 May 2019 at 00:24, Dan van der Ster wrote: > > The upmap balancer in v12.2.12 works really well... Perfectly uniform on our > clusters. > > .. Dan mode upmap ? -- End of message. Next message? ___ ceph-users mailing list

Re: [ceph-users] Unexplainable high memory usage OSD with BlueStore

2019-04-30 Thread Igor Podlesny
On Tue, 30 Apr 2019 at 19:10, Denny Fuchs wrote: [..] > Any suggestions ? -- Try different allocator. In Proxmox 4 they by default had this in /etc/default/ceph {{ ## use jemalloc instead of tcmalloc # # jemalloc is generally faster for small IO workloads and when # ceph-osd is backed by SSDs.

Re: [ceph-users] Need some advice about Pools and Erasure Coding

2019-04-30 Thread Igor Podlesny
On Tue, 30 Apr 2019 at 19:11, Adrien Gillard wrote: > On Tue, Apr 30, 2019 at 10:06 AM Igor Podlesny wrote: > > > > On Tue, 30 Apr 2019 at 04:13, Adrien Gillard > wrote: > > > I would add that the use of cache tiering, though still possible, is > not recommend

Re: [ceph-users] BlueStore bitmap allocator under Luminous and Mimic

2019-04-30 Thread Igor Podlesny
On Mon, 15 Apr 2019 at 19:40, Wido den Hollander wrote: > > Hi, > > With the release of 12.2.12 the bitmap allocator for BlueStore is now > available under Mimic and Luminous. > > [osd] > bluestore_allocator = bitmap > bluefs_allocator = bitmap Hi! Have you tried this? :) -- End of message.

Re: [ceph-users] Need some advice about Pools and Erasure Coding

2019-04-30 Thread Igor Podlesny
On Tue, 30 Apr 2019 at 04:13, Adrien Gillard wrote: > I would add that the use of cache tiering, though still possible, is not > recommended It lacks references. CEPH docs I gave links to didn't say so. > comes with its own challenges. It's challenging for some to not over-quote when

Re: [ceph-users] Need some advice about Pools and Erasure Coding

2019-04-29 Thread Igor Podlesny
On Mon, 29 Apr 2019 at 16:19, Rainer Krienke wrote: [...] > - Do I still (nautilus) need two pools for EC based RBD images, one EC > data pool and a second replicated pool for metadatata? The answer is given at

Re: [ceph-users] Need some advice about Pools and Erasure Coding

2019-04-29 Thread Igor Podlesny
On Mon, 29 Apr 2019 at 16:37, Burkhard Linke wrote: > On 4/29/19 11:19 AM, Rainer Krienke wrote: [...] > > - I also thought about the different k+m settings for a EC pool, for > > example k=4, m=2 compared to k=8 and m=2. Both settings allow for two > > OSDs to fail without any data loss, but I

Re: [ceph-users] Is it possible to get list of all the PGs assigned to an OSD?

2019-04-29 Thread Igor Podlesny
On Mon, 29 Apr 2019 at 15:13, Eugen Block wrote: > > Sure there is: > > ceph pg ls-by-osd Thank you Eugen, I overlooked it somehow :) -- End of message. Next message? ___ ceph-users mailing list ceph-users@lists.ceph.com

[ceph-users] Is it possible to get list of all the PGs assigned to an OSD?

2019-04-29 Thread Igor Podlesny
Or is there no direct way to accomplish that? What workarounds can be used then? -- End of message. Next message? ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[ceph-users] Does ceph osd reweight-by-xxx work correctly if OSDs aren't of same size?

2019-04-29 Thread Igor Podlesny
Say, some nodes have OSDs that are 1.5 times bigger, than other nodes have, meanwhile weights of all the nodes in question is almost equal (due having different number of OSDs obviously) -- End of message. Next message? ___ ceph-users mailing list

Re: [ceph-users] How does CEPH calculates PGs per OSD for erasure coded (EC) pools?

2019-04-28 Thread Igor Podlesny
On Sun, 28 Apr 2019 at 16:14, Paul Emmerich wrote: > Use k+m for PG calculation, that value also shows up as "erasure size" > in ceph osd pool ls detail So does it mean that for PG calculation those 2 pools are equivalent: 1) EC(4, 2) 2) replicated, size 6 ? Sounds weird to be honest.

[ceph-users] How does CEPH calculates PGs per OSD for erasure coded (EC) pools?

2019-04-28 Thread Igor Podlesny
For replicated pools (w/o rounding to nearest power of two) overall PGs number is calculated so: Pools_PGs = 100 * (OSDs / Pool_Size), where 100 -- target number of PGs per single OSD related to that pool, Pool_Size -- factor showing how much raw storage would in fact be used to

[ceph-users] Were fixed CephFS lock ups when it's running on nodes with OSDs?

2019-04-20 Thread Igor Podlesny
I remember seeing reports in regards but it's being a while now. Can anyone tell? -- End of message. Next message? ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Re: [ceph-users] Does "ceph df" use "bogus" copies factor instead of (k, m) for erasure coded pool?

2019-04-16 Thread Igor Podlesny
On Tue, 16 Apr 2019 at 17:05, Paul Emmerich wrote: > > No, the problem is that a storage system should never tell a client > that it has written data if it cannot guarantee that the data is still > there if one device fails. [...] Ah, now I got your point. Anyways, it should be users' choice

Re: [ceph-users] Does "ceph df" use "bogus" copies factor instead of (k, m) for erasure coded pool?

2019-04-16 Thread Igor Podlesny
On Tue, 16 Apr 2019 at 16:52, Paul Emmerich wrote: > On Tue, Apr 16, 2019 at 11:50 AM Igor Podlesny wrote: > > On Tue, 16 Apr 2019 at 14:46, Paul Emmerich wrote: [...] > > Looked at it, didn't see any explanation of your point of view. If > > there're 2 active data i

Re: [ceph-users] Does "ceph df" use "bogus" copies factor instead of (k, m) for erasure coded pool?

2019-04-16 Thread Igor Podlesny
On Tue, 16 Apr 2019 at 14:46, Paul Emmerich wrote: > Sorry, I just realized I didn't answer your original question. [...] No problemo. -- I've figured out the answer to my own question earlier anyways. And actually gave a hint today

Re: [ceph-users] 'Missing' capacity

2019-04-15 Thread Igor Podlesny
On Tue, 16 Apr 2019 at 06:43, Mark Schouten wrote: [...] > So where is the rest of the free space? :X Makes sense to see: sudo ceph osd df tree -- End of message. Next message? ___ ceph-users mailing list ceph-users@lists.ceph.com

Re: [ceph-users] Does "ceph df" use "bogus" copies factor instead of (k, m) for erasure coded pool?

2019-04-12 Thread Igor Podlesny
And as to min_size choice -- since you've replied exactly to that part of mine message only. On Sat, 13 Apr 2019 at 06:54, Paul Emmerich wrote: > On Fri, Apr 12, 2019 at 9:30 PM Igor Podlesny wrote: > > For e. g., an EC pool with default profile (2, 1) has bogus "sizing"

Re: [ceph-users] Does "ceph df" use "bogus" copies factor instead of (k, m) for erasure coded pool?

2019-04-12 Thread Igor Podlesny
On Sat, 13 Apr 2019 at 06:54, Paul Emmerich wrote: > > Please don't use an EC pool with 2+1, that configuration makes no sense. That's too much of an irony given that (2, 1) is default EC profile, described in CEPH documentation in addition. > min_size 3 is the default for that pool, yes. That

[ceph-users] Does "ceph df" use "bogus" copies factor instead of (k, m) for erasure coded pool?

2019-04-12 Thread Igor Podlesny
For e. g., an EC pool with default profile (2, 1) has bogus "sizing" params (size=3, min_size=3). Min. size 3 is wrong as far as I know and it's been fixed in fresh releases (but not in Luminous). But besides that it looks like pool usage isn't calculated according to EC overhead but as if it was

[ceph-users] Does Bluestore backed OSD detect bit rot immediately when reading or only when scrubbed?

2019-04-01 Thread Igor Podlesny
It's wide known that some filesystems (well, ok -- 2 of them: ZFS and Btrfs) detect bit rot on any read request, although, of course an admin can initiate "whole platter" scrubbing. Before Bluestore CEPH could provide only "on demand" detection. I don't take into consideration imaginary setups