On 2/22/21 11:06 AM, Joao Martins wrote:
> On 2/20/21 1:18 AM, Dan Williams wrote:
>> On Tue, Dec 8, 2020 at 9:32 AM Joao Martins <joao.m.mart...@oracle.com>
>> wrote:
>>>
>>> The link above describes it quite nicely, but the idea is to reuse tail
>>> page vmemmap areas, particular the area which only describes tail pages.
>>> So a vmemmap page describes 64 struct pages, and the first page for a given
>>> ZONE_DEVICE vmemmap would contain the head page and 63 tail pages. The
>>> second
>>> vmemmap page would contain only tail pages, and that's what gets reused
>>> across
>>> the rest of the subsection/section. The bigger the page size, the bigger the
>>> savings (2M hpage -> save 6 vmemmap pages; 1G hpage -> save 4094 vmemmap
>>> pages).
>>>
>>> In terms of savings, per 1Tb of memory, the struct page cost would go down
>>> with compound pagemap:
>>>
>>> * with 2M pages we lose 4G instead of 16G (0.39% instead of 1.5% of total
>>> memory)
>>> * with 1G pages we lose 8MB instead of 16G (0.0007% instead of 1.5% of
>>> total memory)
>>
>> Nice!
>>
>
> I failed to mention this in the cover letter but I should say that with this
> trick we will
> need to build the vmemmap page tables with basepages for 2M align, as opposed
> to hugepages
> in the vmemmap page tables (as you probably could tell from the patches).
Also to be clear: by "we will need to build the vmemmap page tables with
basepages for 2M
align" I strictly refer to the ZONE_DEVICE range we are mapping the struct
pages. It's not
the enterity of the vmemmap!
> This means that
> we have to allocate a PMD page, and that costs 2GB per 1Tb (as opposed to
> 4M). This is
> fixable for 1G align by reusing PMD pages (albeit I haven't done that in this
> RFC series).
>
> The footprint reduction is still big, so to iterate the numbers above (and I
> will fix this
> in the v2 cover letter):
>
> * with 2M pages we lose 4G instead of 16G (0.39% instead of 1.5% of total
> memory)
> * with 1G pages we lose 8MB instead of 16G (0.0007% instead of 1.5% of total
> memory)
>
> For vmemmap page tables, we need to use base pages for 2M pages. So taking
> that into
> account, in this RFC series:
>
> * with 2M pages we lose 6G instead of 16G (0.586% instead of 1.5% of total
> memory)
> * with 1G pages we lose ~2GB instead of 16G (0.19% instead of 1.5% of total
> memory)
>
> For 1G align, we are able to reuse vmemmap PMDs that only point to tail
> pages, so
> ultimately we can get the page table overhead from 2GB to 12MB:
>
> * with 1G pages we lose 20MB instead of 16G (0.0019% instead of 1.5% of total
> memory)
>
>>>
>>> The RDMA patch (patch 8/9) is to demonstrate the improvement for an existing
>>> user. For unpin_user_pages() we have an additional test to demonstrate the
>>> improvement. The test performs MR reg/unreg continuously and measuring its
>>> rate for a given period. So essentially ib_mem_get and ib_mem_release being
>>> stress tested which at the end of day means: pin_user_pages_longterm() and
>>> unpin_user_pages() for a scatterlist:
>>>
>>> Before:
>>> 159 rounds in 5.027 sec: 31617.923 usec / round (device-dax)
>>> 466 rounds in 5.009 sec: 10748.456 usec / round (hugetlbfs)
>>>
>>> After:
>>> 305 rounds in 5.010 sec: 16426.047 usec / round (device-dax)
>>> 1073 rounds in 5.004 sec: 4663.622 usec / round (hugetlbfs)
>>
>> Why does hugetlbfs get faster for a ZONE_DEVICE change? Might answer
>> that question myself when I get to patch 8.
>>
> Because the unpinning improvements aren't ZONE_DEVICE specific.
>
> FWIW, I moved those two offending patches outside of this series:
>
>
> https://lore.kernel.org/linux-mm/20210212130843.13865-1-joao.m.mart...@oracle.com/
>
>>>
>>> Patch 9: Improves {pin,get}_user_pages() and its longterm counterpart. It
>>> is very experimental, and I imported most of follow_hugetlb_page(), except
>>> that we do the same trick as gup-fast. In doing the patch I feel this
>>> batching
>>> should live in follow_page_mask() and having that being changed to return a
>>> set
>>> of pages/something-else when walking over PMD/PUDs for THP / devmap pages.
>>> This
>>> patch then brings the previous test of mr reg/unreg (above) on parity
>>> between device-dax and hugetlbfs.
>>>
>>> Some of the patches are a little fresh/WIP (specially patch 3 and 9) and we
>>> are
>>> still running tests. Hence the RFC, asking for comments and general
>>> direction
>>> of the work before continuing.
>>
>> Will go look at the code, but I don't see anything scary conceptually
>> here. The fact that pfn_to_page() does not need to change is among the
>> most compelling features of this approach.
>>
> Glad to hear that :D
>
_______________________________________________
Linux-nvdimm mailing list -- linux-nvdimm@lists.01.org
To unsubscribe send an email to linux-nvdimm-le...@lists.01.org