On Tue, Dec 22, 2020 at 04:40:32AM -0800, Nadav Amit wrote: > > On Dec 21, 2020, at 1:24 PM, Yu Zhao <yuz...@google.com> wrote: > > > > On Mon, Dec 21, 2020 at 12:26:22PM -0800, Linus Torvalds wrote: > >> On Mon, Dec 21, 2020 at 12:23 PM Nadav Amit <nadav.a...@gmail.com> wrote: > >>> Using mmap_write_lock() was my initial fix and there was a strong pushback > >>> on this approach due to its potential impact on performance. > >> > >> From whom? > >> > >> Somebody who doesn't understand that correctness is more important > >> than performance? And that userfaultfd is not the most important part > >> of the system? > >> > >> The fact is, userfaultfd is CLEARLY BUGGY. > >> > >> Linus > > > > Fair enough. > > > > Nadav, for your patch (you might want to update the commit message). > > > > Reviewed-by: Yu Zhao <yuz...@google.com> > > > > While we are all here, there is also clear_soft_dirty() that could > > use a similar fix… > > Just an update as for why I have still not sent v2: I fixed > clear_soft_dirty(), created a reproducer, and the reproducer kept failing. > > So after some debugging, it appears that clear_refs_write() does not flush > the TLB. It indeed calls tlb_finish_mmu() but since 0758cd830494 > ("asm-generic/tlb: avoid potential double flush”), tlb_finish_mmu() does not > flush the TLB since there is clear_refs_write() does not call to > __tlb_adjust_range() (unless there are nested TLBs are pending).
Sorry Nadav, I assumed you knew this existing problem fixed by: https://patchwork.kernel.org/project/linux-mm/cover/20201210121110.10094-1-w...@kernel.org/ > So I have a patch for this issue too: arguably the tlb_gather interface is > not the right one for clear_refs_write() that does not clear PTEs but > changes them. > > Yet, sadly, my reproducer keeps falling (less frequently, but still). So I > will keep debugging to see what goes wrong. I will send v2 once I figure out > what the heck is wrong in the code or my reproducer. > > For the reference, here is my reproducer: Thanks. This would be helpful in case any other breakages happen in the future. > -- >8 -- > > #define _GNU_SOURCE > #include <sys/types.h> > #include <sys/stat.h> > #include <sys/mman.h> > #include <unistd.h> > #include <stdio.h> > #include <stdlib.h> > #include <assert.h> > #include <fcntl.h> > #include <string.h> > #include <threads.h> > #include <stdatomic.h> > > #define PAGE_SIZE (4096) > #define TLB_SIZE (2000) > #define N_PAGES (300000) > #define ITERATIONS (100) > #define N_THREADS (2) > > static int stop; > static char *m; > > static int writer(void *argp) > { > unsigned long t_idx = (unsigned long)argp; > int i, cnt = 0; > > while (!atomic_load(&stop)) { > cnt++; > atomic_fetch_add((atomic_int *)m, 1); > > /* > * First thread only accesses the page to have it cached in the > * TLB. > */ > if (t_idx == 0) > continue; > > /* > * Other threads access enough entries to cause eviction from > * the TLB and trigger #PF upon the next access (before the TLB > * flush of clear_ref actually takes place). > */ > for (i = 1; i < TLB_SIZE; i++) { > if (atomic_load((atomic_int *)(m + PAGE_SIZE * i))) { > fprintf(stderr, "unexpected error\n"); > exit(1); > } > } > } > return cnt; > } > > /* > * Runs mlock/munlock in the background to raise the page-count of the page > and > * force copying instead of reusing the page. > */ > static int do_mlock(void *argp) > { > while (!atomic_load(&stop)) { > if (mlock(m, PAGE_SIZE) || munlock(m, PAGE_SIZE)) { > perror("mlock/munlock"); > exit(1); > } > } > return 0; > } > > int main(void) > { > int r, cnt, fd, total = 0; > long i; > thrd_t thr[N_THREADS]; > thrd_t mlock_thr[N_THREADS]; > > fd = open("/proc/self/clear_refs", O_WRONLY, 0666); > if (fd < 0) { > perror("open"); > exit(1); > } > > /* > * Have large memory for clear_ref, so there would be some time between > * the unmap and the actual deferred flush. > */ > m = mmap(NULL, PAGE_SIZE * N_PAGES, PROT_READ|PROT_WRITE, > MAP_PRIVATE|MAP_ANONYMOUS|MAP_POPULATE, -1, 0); > if (m == MAP_FAILED) { > perror("mmap"); > exit(1); > } > > for (i = 0; i < N_THREADS; i++) { > r = thrd_create(&thr[i], writer, (void *)i); > assert(r == thrd_success); > } > > for (i = 0; i < N_THREADS; i++) { > r = thrd_create(&mlock_thr[i], do_mlock, (void *)i); > assert(r == thrd_success); > } > > for (i = 0; i < ITERATIONS; i++) { > for (i = 0; i < ITERATIONS; i++) { > r = pwrite(fd, "4", 1, 0); > if (r < 0) { > perror("pwrite"); > exit(1); > } > } > } > > atomic_store(&stop, 1); > > for (i = 0; i < N_THREADS; i++) { > r = thrd_join(mlock_thr[i], NULL); > assert(r == thrd_success); > } > > for (i = 0; i < N_THREADS; i++) { > r = thrd_join(thr[i], &cnt); > assert(r == thrd_success); > total += cnt; > } > > r = atomic_load((atomic_int *)(m)); > if (r != total) { > fprintf(stderr, "failed - expected=%d actual=%d\n", total, r); > exit(-1); > } > > fprintf(stderr, "ok\n"); > return 0; > }