diff --git a/[refs] b/[refs] index 57884014c49c..8c69a02b3daf 100644 --- a/[refs] +++ b/[refs] @@ -1,2 +1,2 @@ --- -refs/heads/master: ffadd4d0feb5376c82dc3a4104731b7ce2794edc +refs/heads/master: 51735a7ca67531267a27b57e5fe20f7815192f9c diff --git a/trunk/include/linux/slub_def.h b/trunk/include/linux/slub_def.h index 986e09dcfd8f..e217a7a68ea7 100644 --- a/trunk/include/linux/slub_def.h +++ b/trunk/include/linux/slub_def.h @@ -129,9 +129,9 @@ struct kmem_cache { * This should be dropped to PAGE_SIZE / 2 once the page allocator * "fastpath" becomes competitive with the slab allocator fastpaths. */ -#define SLUB_MAX_SIZE (PAGE_SIZE) +#define SLUB_MAX_SIZE (2 * PAGE_SIZE) -#define SLUB_PAGE_SHIFT (PAGE_SHIFT + 1) +#define SLUB_PAGE_SHIFT (PAGE_SHIFT + 2) /* * We keep the general caches in an array of slab caches that are used for