From 46a753b7615c040771ab1a323a8398c3988069c0 Mon Sep 17 00:00:00 2001 From: Christoph Lameter Date: Mon, 7 Jan 2008 23:20:26 -0800 Subject: [PATCH] --- yaml --- r: 82918 b: refs/heads/master c: f61396aed90acb033952531c522d1010f87e24f4 h: refs/heads/master v: v3 --- [refs] | 2 +- trunk/mm/slub.c | 26 +++++++++++++------------- 2 files changed, 14 insertions(+), 14 deletions(-) diff --git a/[refs] b/[refs] index fb30d991e40f..874e8f8b7a18 100644 --- a/[refs] +++ b/[refs] @@ -1,2 +1,2 @@ --- -refs/heads/master: 151c602f79cb9154c3f3d83223cae355af463d6f +refs/heads/master: f61396aed90acb033952531c522d1010f87e24f4 diff --git a/trunk/mm/slub.c b/trunk/mm/slub.c index 65bf21dc996a..9aa12b54ad1b 100644 --- a/trunk/mm/slub.c +++ b/trunk/mm/slub.c @@ -2607,6 +2607,19 @@ void kfree(const void *x) } EXPORT_SYMBOL(kfree); +static unsigned long count_partial(struct kmem_cache_node *n) +{ + unsigned long flags; + unsigned long x = 0; + struct page *page; + + spin_lock_irqsave(&n->list_lock, flags); + list_for_each_entry(page, &n->partial, lru) + x += page->inuse; + spin_unlock_irqrestore(&n->list_lock, flags); + return x; +} + /* * kmem_cache_shrink removes empty slabs from the partial lists and sorts * the remaining slabs by the number of items in use. The slabs with the @@ -3078,19 +3091,6 @@ void *__kmalloc_node_track_caller(size_t size, gfp_t gfpflags, return slab_alloc(s, gfpflags, node, caller); } -static unsigned long count_partial(struct kmem_cache_node *n) -{ - unsigned long flags; - unsigned long x = 0; - struct page *page; - - spin_lock_irqsave(&n->list_lock, flags); - list_for_each_entry(page, &n->partial, lru) - x += page->inuse; - spin_unlock_irqrestore(&n->list_lock, flags); - return x; -} - #if defined(CONFIG_SYSFS) && defined(CONFIG_SLUB_DEBUG) static int validate_slab(struct kmem_cache *s, struct page *page, unsigned long *map)