ARCH_KMALLOC_FLAGS|SLAB_PANIC,
NULL);
+ list_add(&sizes[INDEX_AC].cs_cachep->list, &slab_caches);
if (INDEX_AC != INDEX_L3) {
sizes[INDEX_L3].cs_cachep =
__kmem_cache_create(names[INDEX_L3].name,
ARCH_KMALLOC_MINALIGN,
ARCH_KMALLOC_FLAGS|SLAB_PANIC,
NULL);
+ list_add(&sizes[INDEX_L3].cs_cachep->list, &slab_caches);
}
slab_early_init = 0;
ARCH_KMALLOC_MINALIGN,
ARCH_KMALLOC_FLAGS|SLAB_PANIC,
NULL);
+ list_add(&sizes->cs_cachep->list, &slab_caches);
}
#ifdef CONFIG_ZONE_DMA
sizes->cs_dmacachep = __kmem_cache_create(
ARCH_KMALLOC_FLAGS|SLAB_CACHE_DMA|
SLAB_PANIC,
NULL);
+ list_add(&sizes->cs_dmacachep->list, &slab_caches);
#endif
sizes++;
names++;
}
cachep->ctor = ctor;
cachep->name = name;
+ cachep->refcount = 1;
if (setup_cpu_cache(cachep, gfp)) {
__kmem_cache_destroy(cachep);
slab_set_debugobj_lock_classes(cachep);
}
- /* cache setup completed, link it into the list */
- list_add(&cachep->list, &slab_caches);
return cachep;
}
if (!s)
err = -ENOSYS; /* Until __kmem_cache_create returns code */
+ /*
+ * Check if the slab has actually been created and if it was a
+ * real instatiation. Aliases do not belong on the list
+ */
+ if (s && s->refcount == 1)
+ list_add(&s->list, &slab_caches);
+
out_locked:
mutex_unlock(&slab_mutex);
put_online_cpus();
void kmem_cache_destroy(struct kmem_cache *c)
{
+ mutex_lock(&slab_mutex);
+ list_del(&c->list);
+ mutex_unlock(&slab_mutex);
+
kmemleak_free(c);
if (c->flags & SLAB_DESTROY_BY_RCU)
rcu_barrier();
size, align, flags, ctor)) {
int r;
- list_add(&s->list, &slab_caches);
mutex_unlock(&slab_mutex);
r = sysfs_slab_add(s);
mutex_lock(&slab_mutex);
if (!r)
return s;
- list_del(&s->list);
kmem_cache_close(s);
}
kmem_cache_free(kmem_cache, s);