From 97901bf4729d397a636261b14f9ac7a2dbe9d823 Mon Sep 17 00:00:00 2001 From: Catalin Marinas Date: Thu, 11 Jun 2009 13:23:17 +0100 Subject: [PATCH] --- yaml --- r: 147735 b: refs/heads/master c: 4374e616d28e65265a5b433ceece275449f3d2e3 h: refs/heads/master i: 147733: 514ba7ac66fb29deae2bd906946233e9c60e4f8d 147731: 067871ecc757f88eb8d08e0f4e7c95b58579654c 147727: 6acb80f87ea3451afacbffe7c734bd4c5f6c5254 v: v3 --- [refs] | 2 +- trunk/mm/slob.c | 7 +++++++ 2 files changed, 8 insertions(+), 1 deletion(-) diff --git a/[refs] b/[refs] index 8c8887558b02..bda11d9a46d3 100644 --- a/[refs] +++ b/[refs] @@ -1,2 +1,2 @@ --- -refs/heads/master: d5cff635290aec9ad7e6ee546aa4fae895361cbb +refs/heads/master: 4374e616d28e65265a5b433ceece275449f3d2e3 diff --git a/trunk/mm/slob.c b/trunk/mm/slob.c index 9b1737b0787b..12f261499925 100644 --- a/trunk/mm/slob.c +++ b/trunk/mm/slob.c @@ -67,6 +67,7 @@ #include #include #include +#include #include /* @@ -509,6 +510,7 @@ void *__kmalloc_node(size_t size, gfp_t gfp, int node) size, PAGE_SIZE << order, gfp, node); } + kmemleak_alloc(ret, size, 1, gfp); return ret; } EXPORT_SYMBOL(__kmalloc_node); @@ -521,6 +523,7 @@ void kfree(const void *block) if (unlikely(ZERO_OR_NULL_PTR(block))) return; + kmemleak_free(block); sp = slob_page(block); if (is_slob_page(sp)) { @@ -584,12 +587,14 @@ struct kmem_cache *kmem_cache_create(const char *name, size_t size, } else if (flags & SLAB_PANIC) panic("Cannot create slab cache %s\n", name); + kmemleak_alloc(c, sizeof(struct kmem_cache), 1, GFP_KERNEL); return c; } EXPORT_SYMBOL(kmem_cache_create); void kmem_cache_destroy(struct kmem_cache *c) { + kmemleak_free(c); slob_free(c, sizeof(struct kmem_cache)); } EXPORT_SYMBOL(kmem_cache_destroy); @@ -613,6 +618,7 @@ void *kmem_cache_alloc_node(struct kmem_cache *c, gfp_t flags, int node) if (c->ctor) c->ctor(b); + kmemleak_alloc_recursive(b, c->size, 1, c->flags, flags); return b; } EXPORT_SYMBOL(kmem_cache_alloc_node); @@ -635,6 +641,7 @@ static void kmem_rcu_free(struct rcu_head *head) void kmem_cache_free(struct kmem_cache *c, void *b) { + kmemleak_free_recursive(b, c->flags); if (unlikely(c->flags & SLAB_DESTROY_BY_RCU)) { struct slob_rcu *slob_rcu; slob_rcu = b + (c->size - sizeof(struct slob_rcu));