summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorChristoph Lameter <cl@linux.com>2012-09-05 01:38:33 +0200
committerPekka Enberg <penberg@kernel.org>2012-09-05 11:00:35 +0200
commit7c9adf5a5471647f392169ef19d3e81dcfa76045 (patch)
tree4e32d1875e45eafd4c4c9f2caedcf3419479c0b8
parentmm/slab_common: Improve error handling in kmem_cache_create (diff)
downloadlinux-7c9adf5a5471647f392169ef19d3e81dcfa76045.tar.xz
linux-7c9adf5a5471647f392169ef19d3e81dcfa76045.zip
mm/sl[aou]b: Move list_add() to slab_common.c
Move the code to append the new kmem_cache to the list of slab caches to the kmem_cache_create code in the shared code. This is possible now since the acquisition of the mutex was moved into kmem_cache_create(). Acked-by: David Rientjes <rientjes@google.com> Reviewed-by: Glauber Costa <glommer@parallels.com> Reviewed-by: Joonsoo Kim <js1304@gmail.com> Signed-off-by: Christoph Lameter <cl@linux.com> Signed-off-by: Pekka Enberg <penberg@kernel.org>
-rw-r--r--mm/slab.c7
-rw-r--r--mm/slab_common.c7
-rw-r--r--mm/slob.c4
-rw-r--r--mm/slub.c2
4 files changed, 16 insertions, 4 deletions
diff --git a/mm/slab.c b/mm/slab.c
index 3b4587bb7b19..a69903168497 100644
--- a/mm/slab.c
+++ b/mm/slab.c
@@ -1680,6 +1680,7 @@ void __init kmem_cache_init(void)
ARCH_KMALLOC_FLAGS|SLAB_PANIC,
NULL);
+ list_add(&sizes[INDEX_AC].cs_cachep->list, &slab_caches);
if (INDEX_AC != INDEX_L3) {
sizes[INDEX_L3].cs_cachep =
__kmem_cache_create(names[INDEX_L3].name,
@@ -1687,6 +1688,7 @@ void __init kmem_cache_init(void)
ARCH_KMALLOC_MINALIGN,
ARCH_KMALLOC_FLAGS|SLAB_PANIC,
NULL);
+ list_add(&sizes[INDEX_L3].cs_cachep->list, &slab_caches);
}
slab_early_init = 0;
@@ -1705,6 +1707,7 @@ void __init kmem_cache_init(void)
ARCH_KMALLOC_MINALIGN,
ARCH_KMALLOC_FLAGS|SLAB_PANIC,
NULL);
+ list_add(&sizes->cs_cachep->list, &slab_caches);
}
#ifdef CONFIG_ZONE_DMA
sizes->cs_dmacachep = __kmem_cache_create(
@@ -1714,6 +1717,7 @@ void __init kmem_cache_init(void)
ARCH_KMALLOC_FLAGS|SLAB_CACHE_DMA|
SLAB_PANIC,
NULL);
+ list_add(&sizes->cs_dmacachep->list, &slab_caches);
#endif
sizes++;
names++;
@@ -2583,6 +2587,7 @@ __kmem_cache_create (const char *name, size_t size, size_t align,
}
cachep->ctor = ctor;
cachep->name = name;
+ cachep->refcount = 1;
if (setup_cpu_cache(cachep, gfp)) {
__kmem_cache_destroy(cachep);
@@ -2599,8 +2604,6 @@ __kmem_cache_create (const char *name, size_t size, size_t align,
slab_set_debugobj_lock_classes(cachep);
}
- /* cache setup completed, link it into the list */
- list_add(&cachep->list, &slab_caches);
return cachep;
}
diff --git a/mm/slab_common.c b/mm/slab_common.c
index fe8dc943c285..5190a7cd02bd 100644
--- a/mm/slab_common.c
+++ b/mm/slab_common.c
@@ -111,6 +111,13 @@ struct kmem_cache *kmem_cache_create(const char *name, size_t size, size_t align
if (!s)
err = -ENOSYS; /* Until __kmem_cache_create returns code */
+ /*
+ * Check if the slab has actually been created and if it was a
+ * real instatiation. Aliases do not belong on the list
+ */
+ if (s && s->refcount == 1)
+ list_add(&s->list, &slab_caches);
+
out_locked:
mutex_unlock(&slab_mutex);
put_online_cpus();
diff --git a/mm/slob.c b/mm/slob.c
index 45d4ca79933a..5225d28f2694 100644
--- a/mm/slob.c
+++ b/mm/slob.c
@@ -540,6 +540,10 @@ struct kmem_cache *__kmem_cache_create(const char *name, size_t size,
void kmem_cache_destroy(struct kmem_cache *c)
{
+ mutex_lock(&slab_mutex);
+ list_del(&c->list);
+ mutex_unlock(&slab_mutex);
+
kmemleak_free(c);
if (c->flags & SLAB_DESTROY_BY_RCU)
rcu_barrier();
diff --git a/mm/slub.c b/mm/slub.c
index c66908983218..24aa362edef7 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -3975,7 +3975,6 @@ struct kmem_cache *__kmem_cache_create(const char *name, size_t size,
size, align, flags, ctor)) {
int r;
- list_add(&s->list, &slab_caches);
mutex_unlock(&slab_mutex);
r = sysfs_slab_add(s);
mutex_lock(&slab_mutex);
@@ -3983,7 +3982,6 @@ struct kmem_cache *__kmem_cache_create(const char *name, size_t size,
if (!r)
return s;
- list_del(&s->list);
kmem_cache_close(s);
}
kmem_cache_free(kmem_cache, s);