Skip to content

Commit

Permalink
ftrace: limit trace entries
Browse files Browse the repository at this point in the history
Currently there is no protection from the root user to use up all of
memory for trace buffers. If the root user allocates too many entries,
the OOM killer might start kill off all tasks.

This patch adds an algorith to check the following condition:

 pages_requested > (freeable_memory + current_trace_buffer_pages) / 4

If the above is met then the allocation fails. The above prevents more
than 1/4th of freeable memory from being used by trace buffers.

To determine the freeable_memory, I made determine_dirtyable_memory in
mm/page-writeback.c global.

Special thanks goes to Peter Zijlstra for suggesting the above calculation.

Signed-off-by: Steven Rostedt <[email protected]>
Signed-off-by: Ingo Molnar <[email protected]>
Signed-off-by: Thomas Gleixner <[email protected]>
  • Loading branch information
rostedt authored and KAGA-KOKO committed May 23, 2008
1 parent 6c6c279 commit 3eefae9
Show file tree
Hide file tree
Showing 3 changed files with 47 additions and 3 deletions.
2 changes: 2 additions & 0 deletions include/linux/writeback.h
Original file line number Diff line number Diff line change
Expand Up @@ -105,6 +105,8 @@ extern int vm_highmem_is_dirtyable;
extern int block_dump;
extern int laptop_mode;

extern unsigned long determine_dirtyable_memory(void);

extern int dirty_ratio_handler(struct ctl_table *table, int write,
struct file *filp, void __user *buffer, size_t *lenp,
loff_t *ppos);
Expand Down
38 changes: 38 additions & 0 deletions kernel/trace/trace.c
Original file line number Diff line number Diff line change
Expand Up @@ -27,6 +27,7 @@
#include <linux/poll.h>
#include <linux/gfp.h>
#include <linux/fs.h>
#include <linux/writeback.h>

#include <linux/stacktrace.h>

Expand All @@ -51,6 +52,8 @@ static int trace_free_page(void);

static int tracing_disabled = 1;

static unsigned long tracing_pages_allocated;

long
ns2usecs(cycle_t nsec)
{
Expand Down Expand Up @@ -2591,12 +2594,41 @@ tracing_entries_write(struct file *filp, const char __user *ubuf,
}

if (val > global_trace.entries) {
long pages_requested;
unsigned long freeable_pages;

/* make sure we have enough memory before mapping */
pages_requested =
(val + (ENTRIES_PER_PAGE-1)) / ENTRIES_PER_PAGE;

/* account for each buffer (and max_tr) */
pages_requested *= tracing_nr_buffers * 2;

/* Check for overflow */
if (pages_requested < 0) {
cnt = -ENOMEM;
goto out;
}

freeable_pages = determine_dirtyable_memory();

/* we only allow to request 1/4 of useable memory */
if (pages_requested >
((freeable_pages + tracing_pages_allocated) / 4)) {
cnt = -ENOMEM;
goto out;
}

while (global_trace.entries < val) {
if (trace_alloc_page()) {
cnt = -ENOMEM;
goto out;
}
/* double check that we don't go over the known pages */
if (tracing_pages_allocated > pages_requested)
break;
}

} else {
/* include the number of entries in val (inc of page entries) */
while (global_trace.entries > val + (ENTRIES_PER_PAGE - 1))
Expand Down Expand Up @@ -2776,6 +2808,7 @@ static int trace_alloc_page(void)
struct page *page, *tmp;
LIST_HEAD(pages);
void *array;
unsigned pages_allocated = 0;
int i;

/* first allocate a page for each CPU */
Expand All @@ -2787,6 +2820,7 @@ static int trace_alloc_page(void)
goto free_pages;
}

pages_allocated++;
page = virt_to_page(array);
list_add(&page->lru, &pages);

Expand All @@ -2798,6 +2832,7 @@ static int trace_alloc_page(void)
"for trace buffer!\n");
goto free_pages;
}
pages_allocated++;
page = virt_to_page(array);
list_add(&page->lru, &pages);
#endif
Expand All @@ -2819,6 +2854,7 @@ static int trace_alloc_page(void)
SetPageLRU(page);
#endif
}
tracing_pages_allocated += pages_allocated;
global_trace.entries += ENTRIES_PER_PAGE;

return 0;
Expand Down Expand Up @@ -2853,6 +2889,8 @@ static int trace_free_page(void)
page = list_entry(p, struct page, lru);
ClearPageLRU(page);
list_del(&page->lru);
tracing_pages_allocated--;
tracing_pages_allocated--;
__free_page(page);

tracing_reset(data);
Expand Down
10 changes: 7 additions & 3 deletions mm/page-writeback.c
Original file line number Diff line number Diff line change
Expand Up @@ -126,8 +126,6 @@ static void background_writeout(unsigned long _min_pages);
static struct prop_descriptor vm_completions;
static struct prop_descriptor vm_dirties;

static unsigned long determine_dirtyable_memory(void);

/*
* couple the period to the dirty_ratio:
*
Expand Down Expand Up @@ -347,7 +345,13 @@ static unsigned long highmem_dirtyable_memory(unsigned long total)
#endif
}

static unsigned long determine_dirtyable_memory(void)
/**
* determine_dirtyable_memory - amount of memory that may be used
*
* Returns the numebr of pages that can currently be freed and used
* by the kernel for direct mappings.
*/
unsigned long determine_dirtyable_memory(void)
{
unsigned long x;

Expand Down

0 comments on commit 3eefae9

Please sign in to comment.