Skip to content

Commit

Permalink
slub: provide /proc/slabinfo
Browse files Browse the repository at this point in the history
This adds a read-only /proc/slabinfo file on SLUB, that makes slabtop work.

[ mingo@elte.hu: build fix. ]

Cc: Andi Kleen <andi@firstfloor.org>
Cc: Christoph Lameter <clameter@sgi.com>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Signed-off-by: Pekka Enberg <penberg@cs.helsinki.fi>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
  • Loading branch information
Pekka J Enberg authored and Linus Torvalds committed Jan 1, 2008
1 parent 476c6c1 commit 57ed3ed
Show file tree
Hide file tree
Showing 2 changed files with 94 additions and 13 deletions.
2 changes: 2 additions & 0 deletions include/linux/slub_def.h
Original file line number Diff line number Diff line change
Expand Up @@ -200,4 +200,6 @@ static __always_inline void *kmalloc_node(size_t size, gfp_t flags, int node)
}
#endif

extern const struct seq_operations slabinfo_op;

#endif /* _LINUX_SLUB_DEF_H */
105 changes: 92 additions & 13 deletions mm/slub.c
Original file line number Diff line number Diff line change
Expand Up @@ -3076,6 +3076,19 @@ void *__kmalloc_node_track_caller(size_t size, gfp_t gfpflags,
return slab_alloc(s, gfpflags, node, caller);
}

static unsigned long count_partial(struct kmem_cache_node *n)
{
unsigned long flags;
unsigned long x = 0;
struct page *page;

spin_lock_irqsave(&n->list_lock, flags);
list_for_each_entry(page, &n->partial, lru)
x += page->inuse;
spin_unlock_irqrestore(&n->list_lock, flags);
return x;
}

#if defined(CONFIG_SYSFS) && defined(CONFIG_SLUB_DEBUG)
static int validate_slab(struct kmem_cache *s, struct page *page,
unsigned long *map)
Expand Down Expand Up @@ -3458,19 +3471,6 @@ static int list_locations(struct kmem_cache *s, char *buf,
return n;
}

static unsigned long count_partial(struct kmem_cache_node *n)
{
unsigned long flags;
unsigned long x = 0;
struct page *page;

spin_lock_irqsave(&n->list_lock, flags);
list_for_each_entry(page, &n->partial, lru)
x += page->inuse;
spin_unlock_irqrestore(&n->list_lock, flags);
return x;
}

enum slab_stat_type {
SL_FULL,
SL_PARTIAL,
Expand Down Expand Up @@ -4123,3 +4123,82 @@ static int __init slab_sysfs_init(void)

__initcall(slab_sysfs_init);
#endif

/*
* The /proc/slabinfo ABI
*/
#ifdef CONFIG_PROC_FS

static void print_slabinfo_header(struct seq_file *m)
{
seq_puts(m, "slabinfo - version: 2.1\n");
seq_puts(m, "# name <active_objs> <num_objs> <objsize> "
"<objperslab> <pagesperslab>");
seq_puts(m, " : tunables <limit> <batchcount> <sharedfactor>");
seq_puts(m, " : slabdata <active_slabs> <num_slabs> <sharedavail>");
seq_putc(m, '\n');
}

static void *s_start(struct seq_file *m, loff_t *pos)
{
loff_t n = *pos;

down_read(&slub_lock);
if (!n)
print_slabinfo_header(m);

return seq_list_start(&slab_caches, *pos);
}

static void *s_next(struct seq_file *m, void *p, loff_t *pos)
{
return seq_list_next(p, &slab_caches, pos);
}

static void s_stop(struct seq_file *m, void *p)
{
up_read(&slub_lock);
}

static int s_show(struct seq_file *m, void *p)
{
unsigned long nr_partials = 0;
unsigned long nr_slabs = 0;
unsigned long nr_inuse = 0;
unsigned long nr_objs;
struct kmem_cache *s;
int node;

s = list_entry(p, struct kmem_cache, list);

for_each_online_node(node) {
struct kmem_cache_node *n = get_node(s, node);

if (!n)
continue;

nr_partials += n->nr_partial;
nr_slabs += atomic_long_read(&n->nr_slabs);
nr_inuse += count_partial(n);
}

nr_objs = nr_slabs * s->objects;
nr_inuse += (nr_slabs - nr_partials) * s->objects;

seq_printf(m, "%-17s %6lu %6lu %6u %4u %4d", s->name, nr_inuse,
nr_objs, s->size, s->objects, (1 << s->order));
seq_printf(m, " : tunables %4u %4u %4u", 0, 0, 0);
seq_printf(m, " : slabdata %6lu %6lu %6lu", nr_slabs, nr_slabs,
0UL);
seq_putc(m, '\n');
return 0;
}

const struct seq_operations slabinfo_op = {
.start = s_start,
.next = s_next,
.stop = s_stop,
.show = s_show,
};

#endif /* CONFIG_PROC_FS */

0 comments on commit 57ed3ed

Please sign in to comment.