On Fri, Mar 14, 2014 at 12:17:05PM +0100, Peter Zijlstra wrote: > On Thu, Mar 13, 2014 at 04:03:52PM -0400, Don Zickus wrote: > > Hi Peter, > > > > So we found another corner case with MMAP2 interface. I don't think it is > > a big hurdle to overcome, just wanted a suggestion. > > > > Joe ran specjbb2013 (which creates about 10,000 java threads across 9 > > processes) and our c2c tool turned up some cacheline collision data on > > libjvm.so. This didn't make sense because you shouldn't be able to write > > to a shared library. > > > > Even worse, our tool said it affected all the java process and a majority > > of the threads. Which again didn't make sense because this shared library > > should be local to each pid's memory. > > > > Anyway, what we determined is that the shared library had mmap data that > > was non-zero (because it was backed by a file, libjvm.so). So the > > assumption was if the major, minor, inode and inode generation numbers > > were non-zero, this memory segment was shared across processes. > > > > So perf setup its map files for the mmap area and then started sampling data > > addresses. A few hundred HITMs were to a virtual address that fell into > > the libjvm.so memory segment (which was assumed to be mmap'd across > > processes). > > > > Coalescing all the data suggested that multiple pids/tids were contending > > for a cacheline in a shared library. > > > > After talking with Larry Woodman, we realized when you write to a 'data' or > > 'bss' segment of a shared library, you incur a COW fault that maps to an > > anonymous page in the pid's memory. However, perf doesn't see this. > > > > So when all the tids start writing to this 'data' or 'bss' segment they > > generate HITMs within their pid (which is fine). However the tool thinks > > it affects other pids (which is not fine). > > > > My question is, how can our tool determine if a virtual address is private > > to a pid or not? Originally it had to have a zero for maj, min, ino, and > > ino gen. But for file map'd libraries this doesn't always work because we > > don't see COW faults in perf (and we may not want too :-) ). > > > > Is there another technique we can use? Perhaps during the reading of > > /proc/<pid>/maps, if the protection is marked 'p' for private, we just tell > > the sort algorithm to sort locally to the process but a 's' for shared can > > be sorted globally based on data addresses? > > > > Or something else that tells us that a virtual address has changed its > > mapping? Thoughts? > > Very good indeed; we're missing the protection and flags bits. > > How about something like the below; with that you can solve your problem > by looking at mmap2.flags & MAP_PRIVATE.
Yes this seemed to work. I attached a slight update to your patch (one that compiles :-) ). And I will reply to this thread with the tool changes I used to verify this (in case I did that piece wrong). Cheers, Don diff --git a/include/uapi/linux/perf_event.h b/include/uapi/linux/perf_event.h index 853bc1c..2ed502f 100644 --- a/include/uapi/linux/perf_event.h +++ b/include/uapi/linux/perf_event.h @@ -699,6 +699,7 @@ enum perf_event_type { * u32 min; * u64 ino; * u64 ino_generation; + * u32 prot, flags; * char filename[]; * struct sample_id sample_id; * }; diff --git a/kernel/events/core.c b/kernel/events/core.c index 81919fe..ace46f8 100644 --- a/kernel/events/core.c +++ b/kernel/events/core.c @@ -39,6 +39,7 @@ #include <linux/hw_breakpoint.h> #include <linux/mm_types.h> #include <linux/cgroup.h> +#include <linux/mman.h> #include "internal.h" @@ -5100,6 +5101,7 @@ struct perf_mmap_event { int maj, min; u64 ino; u64 ino_generation; + u32 prot, flags; struct { struct perf_event_header header; @@ -5141,6 +5143,8 @@ static void perf_event_mmap_output(struct perf_event *event, mmap_event->event_id.header.size += sizeof(mmap_event->min); mmap_event->event_id.header.size += sizeof(mmap_event->ino); mmap_event->event_id.header.size += sizeof(mmap_event->ino_generation); + mmap_event->event_id.header.size += sizeof(mmap_event->prot); + mmap_event->event_id.header.size += sizeof(mmap_event->flags); } perf_event_header__init_id(&mmap_event->event_id.header, &sample, event); @@ -5159,6 +5163,8 @@ static void perf_event_mmap_output(struct perf_event *event, perf_output_put(&handle, mmap_event->min); perf_output_put(&handle, mmap_event->ino); perf_output_put(&handle, mmap_event->ino_generation); + perf_output_put(&handle, mmap_event->prot); + perf_output_put(&handle, mmap_event->flags); } __output_copy(&handle, mmap_event->file_name, @@ -5177,6 +5183,7 @@ static void perf_event_mmap_event(struct perf_mmap_event *mmap_event) struct file *file = vma->vm_file; int maj = 0, min = 0; u64 ino = 0, gen = 0; + u32 prot = 0, flags = 0; unsigned int size; char tmp[16]; char *buf = NULL; @@ -5207,6 +5214,28 @@ static void perf_event_mmap_event(struct perf_mmap_event *mmap_event) gen = inode->i_generation; maj = MAJOR(dev); min = MINOR(dev); + + if (vma->vm_flags & VM_READ) + prot |= PROT_READ; + if (vma->vm_flags & VM_WRITE) + prot |= PROT_WRITE; + if (vma->vm_flags & VM_EXEC) + prot |= PROT_EXEC; + + if (vma->vm_flags & VM_MAYSHARE) + flags = MAP_SHARED; + else + flags = MAP_PRIVATE; + + if (vma->vm_flags & VM_DENYWRITE) + flags |= MAP_DENYWRITE; + if (vma->vm_flags & VM_MAYEXEC) + flags |= MAP_EXECUTABLE; + if (vma->vm_flags & VM_LOCKED) + flags |= MAP_LOCKED; + if (vma->vm_flags & VM_HUGETLB) + flags |= MAP_HUGETLB; + goto got_name; } else { name = (char *)arch_vma_name(vma); @@ -5247,6 +5276,8 @@ got_name: mmap_event->min = min; mmap_event->ino = ino; mmap_event->ino_generation = gen; + mmap_event->prot = prot; + mmap_event->flags = flags; if (!(vma->vm_flags & VM_EXEC)) mmap_event->event_id.header.misc |= PERF_RECORD_MISC_MMAP_DATA; @@ -5287,6 +5318,8 @@ void perf_event_mmap(struct vm_area_struct *vma) /* .min (attr_mmap2 only) */ /* .ino (attr_mmap2 only) */ /* .ino_generation (attr_mmap2 only) */ + /* .prot (attr_mmap2 only) */ + /* .flags (attr_mmap2 only) */ }; perf_event_mmap_event(&mmap_event); -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/