From 2f4d9cf851476e1ac9bc28af47cf56a498a819f9 Mon Sep 17 00:00:00 2001 From: Christoph Lameter Date: Fri, 21 Dec 2007 14:37:37 -0800 Subject: [PATCH] --- yaml --- r: 75118 b: refs/heads/master c: 76be895001f2b0bee42a7685e942d3e08d5dd46c h: refs/heads/master v: v3 --- [refs] | 2 +- trunk/mm/slub.c | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/[refs] b/[refs] index 45410d3da617..8cc752625019 100644 --- a/[refs] +++ b/[refs] @@ -1,2 +1,2 @@ --- -refs/heads/master: ea67db4cdbbf7f4e74150e71da0984e25121f500 +refs/heads/master: 76be895001f2b0bee42a7685e942d3e08d5dd46c diff --git a/trunk/mm/slub.c b/trunk/mm/slub.c index b9f37cb0f2e6..3655ad359f03 100644 --- a/trunk/mm/slub.c +++ b/trunk/mm/slub.c @@ -172,7 +172,7 @@ static inline void ClearSlabDebug(struct page *page) * Mininum number of partial slabs. These will be left on the partial * lists even if they are empty. kmem_cache_shrink may reclaim them. */ -#define MIN_PARTIAL 2 +#define MIN_PARTIAL 5 /* * Maximum number of desirable partial slabs. @@ -1613,7 +1613,7 @@ static void __slab_free(struct kmem_cache *s, struct page *page, * then add it. */ if (unlikely(!prior)) - add_partial(get_node(s, page_to_nid(page)), page); + add_partial_tail(get_node(s, page_to_nid(page)), page); out_unlock: slab_unlock(page);