From: Jiri Olsa <jo...@kernel.org>

Move prev/start/end from tools/perf's mmap to libperf's perf_mmap struct.

Committer notes:

Add linux/types.h as we use u64.

Signed-off-by: Jiri Olsa <jo...@kernel.org>
Cc: Alexander Shishkin <alexander.shish...@linux.intel.com>
Cc: Michael Petlan <mpet...@redhat.com>
Cc: Namhyung Kim <namhy...@kernel.org>
Cc: Peter Zijlstra <a.p.zijls...@chello.nl>
Link: http://lore.kernel.org/lkml/20190913132355.21634-16-jo...@kernel.org
Signed-off-by: Arnaldo Carvalho de Melo <a...@redhat.com>
---
 tools/perf/lib/include/internal/mmap.h |  4 +++
 tools/perf/util/mmap.c                 | 50 +++++++++++++-------------
 tools/perf/util/mmap.h                 |  3 --
 3 files changed, 29 insertions(+), 28 deletions(-)

diff --git a/tools/perf/lib/include/internal/mmap.h 
b/tools/perf/lib/include/internal/mmap.h
index e6fb850ba47a..ebf5b93754fd 100644
--- a/tools/perf/lib/include/internal/mmap.h
+++ b/tools/perf/lib/include/internal/mmap.h
@@ -3,6 +3,7 @@
 #define __LIBPERF_INTERNAL_MMAP_H
 
 #include <linux/refcount.h>
+#include <linux/types.h>
 
 /**
  * struct perf_mmap - perf's ring buffer mmap details
@@ -15,6 +16,9 @@ struct perf_mmap {
        int              fd;
        int              cpu;
        refcount_t       refcnt;
+       u64              prev;
+       u64              start;
+       u64              end;
 };
 
 #endif /* __LIBPERF_INTERNAL_MMAP_H */
diff --git a/tools/perf/util/mmap.c b/tools/perf/util/mmap.c
index d6406d216cfe..6ce70ff005cb 100644
--- a/tools/perf/util/mmap.c
+++ b/tools/perf/util/mmap.c
@@ -94,19 +94,19 @@ union perf_event *perf_mmap__read_event(struct mmap *map)
 
        /* non-overwirte doesn't pause the ringbuffer */
        if (!map->overwrite)
-               map->end = perf_mmap__read_head(map);
+               map->core.end = perf_mmap__read_head(map);
 
-       event = perf_mmap__read(map, &map->start, map->end);
+       event = perf_mmap__read(map, &map->core.start, map->core.end);
 
        if (!map->overwrite)
-               map->prev = map->start;
+               map->core.prev = map->core.start;
 
        return event;
 }
 
 static bool perf_mmap__empty(struct mmap *map)
 {
-       return perf_mmap__read_head(map) == map->prev && 
!map->auxtrace_mmap.base;
+       return perf_mmap__read_head(map) == map->core.prev && 
!map->auxtrace_mmap.base;
 }
 
 void perf_mmap__get(struct mmap *map)
@@ -125,7 +125,7 @@ void perf_mmap__put(struct mmap *map)
 void perf_mmap__consume(struct mmap *map)
 {
        if (!map->overwrite) {
-               u64 old = map->prev;
+               u64 old = map->core.prev;
 
                perf_mmap__write_tail(map, old);
        }
@@ -368,7 +368,7 @@ int perf_mmap__mmap(struct mmap *map, struct mmap_params 
*mp, int fd, int cpu)
         * perf_evlist__filter_pollfd().
         */
        refcount_set(&map->core.refcnt, 2);
-       map->prev = 0;
+       map->core.prev = 0;
        map->core.mask = mp->mask;
        map->core.base = mmap(NULL, perf_mmap__mmap_len(map), mp->prot,
                         MAP_SHARED, fd, 0);
@@ -443,22 +443,22 @@ static int overwrite_rb_find_range(void *buf, int mask, 
u64 *start, u64 *end)
 static int __perf_mmap__read_init(struct mmap *md)
 {
        u64 head = perf_mmap__read_head(md);
-       u64 old = md->prev;
+       u64 old = md->core.prev;
        unsigned char *data = md->core.base + page_size;
        unsigned long size;
 
-       md->start = md->overwrite ? head : old;
-       md->end = md->overwrite ? old : head;
+       md->core.start = md->overwrite ? head : old;
+       md->core.end = md->overwrite ? old : head;
 
-       if ((md->end - md->start) < md->flush)
+       if ((md->core.end - md->core.start) < md->flush)
                return -EAGAIN;
 
-       size = md->end - md->start;
+       size = md->core.end - md->core.start;
        if (size > (unsigned long)(md->core.mask) + 1) {
                if (!md->overwrite) {
                        WARN_ONCE(1, "failed to keep up with mmap data. (warn 
only once)\n");
 
-                       md->prev = head;
+                       md->core.prev = head;
                        perf_mmap__consume(md);
                        return -EAGAIN;
                }
@@ -467,7 +467,7 @@ static int __perf_mmap__read_init(struct mmap *md)
                 * Backward ring buffer is full. We still have a chance to read
                 * most of data from it.
                 */
-               if (overwrite_rb_find_range(data, md->core.mask, &md->start, 
&md->end))
+               if (overwrite_rb_find_range(data, md->core.mask, 
&md->core.start, &md->core.end))
                        return -EINVAL;
        }
 
@@ -498,12 +498,12 @@ int perf_mmap__push(struct mmap *md, void *to,
        if (rc < 0)
                return (rc == -EAGAIN) ? 1 : -1;
 
-       size = md->end - md->start;
+       size = md->core.end - md->core.start;
 
-       if ((md->start & md->core.mask) + size != (md->end & md->core.mask)) {
-               buf = &data[md->start & md->core.mask];
-               size = md->core.mask + 1 - (md->start & md->core.mask);
-               md->start += size;
+       if ((md->core.start & md->core.mask) + size != (md->core.end & 
md->core.mask)) {
+               buf = &data[md->core.start & md->core.mask];
+               size = md->core.mask + 1 - (md->core.start & md->core.mask);
+               md->core.start += size;
 
                if (push(md, to, buf, size) < 0) {
                        rc = -1;
@@ -511,16 +511,16 @@ int perf_mmap__push(struct mmap *md, void *to,
                }
        }
 
-       buf = &data[md->start & md->core.mask];
-       size = md->end - md->start;
-       md->start += size;
+       buf = &data[md->core.start & md->core.mask];
+       size = md->core.end - md->core.start;
+       md->core.start += size;
 
        if (push(md, to, buf, size) < 0) {
                rc = -1;
                goto out;
        }
 
-       md->prev = head;
+       md->core.prev = head;
        perf_mmap__consume(md);
 out:
        return rc;
@@ -529,8 +529,8 @@ int perf_mmap__push(struct mmap *md, void *to,
 /*
  * Mandatory for overwrite mode
  * The direction of overwrite mode is backward.
- * The last perf_mmap__read() will set tail to map->prev.
- * Need to correct the map->prev to head which is the end of next read.
+ * The last perf_mmap__read() will set tail to map->core.prev.
+ * Need to correct the map->core.prev to head which is the end of next read.
  */
 void perf_mmap__read_done(struct mmap *map)
 {
@@ -540,5 +540,5 @@ void perf_mmap__read_done(struct mmap *map)
        if (!refcount_read(&map->core.refcnt))
                return;
 
-       map->prev = perf_mmap__read_head(map);
+       map->core.prev = perf_mmap__read_head(map);
 }
diff --git a/tools/perf/util/mmap.h b/tools/perf/util/mmap.h
index 5febd22fbe2e..a3dd53f2bfb8 100644
--- a/tools/perf/util/mmap.h
+++ b/tools/perf/util/mmap.h
@@ -22,9 +22,6 @@ struct aiocb;
  */
 struct mmap {
        struct perf_mmap        core;
-       u64              prev;
-       u64              start;
-       u64              end;
        bool             overwrite;
        struct auxtrace_mmap auxtrace_mmap;
        char             event_copy[PERF_SAMPLE_MAX_SIZE] __aligned(8);
-- 
2.21.0

Reply via email to