[PATCH 08/10] perf: Optimize the !vmalloc backed buffer

From: Peter Zijlstra
Date: Fri May 21 2010 - 05:11:56 EST


Reduce code and data by using the knowledge that for !PERF_USE_VMALLOC
data_order is always 0.

Signed-off-by: Peter Zijlstra <a.p.zijlstra@xxxxxxxxx>
---
include/linux/perf_event.h | 2 +-
kernel/perf_event.c | 41 ++++++++++++++++++++++++++---------------
2 files changed, 27 insertions(+), 16 deletions(-)

Index: linux-2.6/include/linux/perf_event.h
===================================================================
--- linux-2.6.orig/include/linux/perf_event.h
+++ linux-2.6/include/linux/perf_event.h
@@ -588,8 +588,8 @@ struct perf_mmap_data {
struct rcu_head rcu_head;
#ifdef CONFIG_PERF_USE_VMALLOC
struct work_struct work;
+ int page_order; /* allocation order */
#endif
- int data_order; /* allocation order */
int nr_pages; /* nr of data pages */
int writable; /* are we writable */
int nr_locked; /* nr pages mlocked */
Index: linux-2.6/kernel/perf_event.c
===================================================================
--- linux-2.6.orig/kernel/perf_event.c
+++ linux-2.6/kernel/perf_event.c
@@ -2297,11 +2297,6 @@ unlock:
rcu_read_unlock();
}

-static unsigned long perf_data_size(struct perf_mmap_data *data)
-{
- return data->nr_pages << (PAGE_SHIFT + data->data_order);
-}
-
#ifndef CONFIG_PERF_USE_VMALLOC

/*
@@ -2359,7 +2354,6 @@ perf_mmap_data_alloc(struct perf_event *
goto fail_data_pages;
}

- data->data_order = 0;
data->nr_pages = nr_pages;

return data;
@@ -2395,6 +2389,11 @@ static void perf_mmap_data_free(struct p
kfree(data);
}

+static inline int page_order(struct perf_mmap_data *data)
+{
+ return 0;
+}
+
#else

/*
@@ -2403,10 +2402,15 @@ static void perf_mmap_data_free(struct p
* Required for architectures that have d-cache aliasing issues.
*/

+static inline int page_order(struct perf_mmap_data *data)
+{
+ return data->page_order;
+}
+
static struct page *
perf_mmap_to_page(struct perf_mmap_data *data, unsigned long pgoff)
{
- if (pgoff > (1UL << data->data_order))
+ if (pgoff > (1UL << page_order(data)))
return NULL;

return vmalloc_to_page((void *)data->user_page + pgoff * PAGE_SIZE);
@@ -2426,7 +2430,7 @@ static void perf_mmap_data_free_work(str
int i, nr;

data = container_of(work, struct perf_mmap_data, work);
- nr = 1 << data->data_order;
+ nr = 1 << page_order(data);

base = data->user_page;
for (i = 0; i < nr + 1; i++)
@@ -2465,7 +2469,7 @@ perf_mmap_data_alloc(struct perf_event *

data->user_page = all_buf;
data->data_pages[0] = all_buf + PAGE_SIZE;
- data->data_order = ilog2(nr_pages);
+ data->page_order = ilog2(nr_pages);
data->nr_pages = 1;

return data;
@@ -2479,6 +2483,11 @@ fail:

#endif

+static unsigned long perf_data_size(struct perf_mmap_data *data)
+{
+ return data->nr_pages << (PAGE_SHIFT + page_order(data));
+}
+
static int perf_mmap_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
{
struct perf_event *event = vma->vm_file->private_data;
@@ -2978,10 +2987,12 @@ void perf_output_copy(struct perf_output
handle->addr += size;
handle->size -= size;
if (!handle->size) {
+ struct perf_mmap_data *data = handle->data;
+
handle->page++;
- handle->page &= handle->data->nr_pages - 1;
- handle->addr = handle->data->data_pages[handle->page];
- handle->size = PAGE_SIZE << handle->data->data_order;
+ handle->page &= data->nr_pages - 1;
+ handle->addr = data->data_pages[handle->page];
+ handle->size = PAGE_SIZE << page_order(data);
}
} while (len);
}
@@ -3049,12 +3060,12 @@ int perf_output_begin(struct perf_output
if (head - local_read(&data->wakeup) > data->watermark)
local_add(data->watermark, &data->wakeup);

- handle->page = handle->offset >> (PAGE_SHIFT + data->data_order);
+ handle->page = handle->offset >> (PAGE_SHIFT + page_order(data));
handle->page &= data->nr_pages - 1;
- handle->size = handle->offset & ((PAGE_SIZE << data->data_order) - 1);
+ handle->size = handle->offset & ((PAGE_SIZE << page_order(data)) - 1);
handle->addr = data->data_pages[handle->page];
handle->addr += handle->size;
- handle->size = (PAGE_SIZE << data->data_order) - handle->size;
+ handle->size = (PAGE_SIZE << page_order(data)) - handle->size;

if (have_lost) {
lost_event.header.type = PERF_RECORD_LOST;


--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/