diff options
author | Christoph Lameter <cl@linux.com> | 2013-02-05 19:45:23 +0100 |
---|---|---|
committer | Pekka Enberg <penberg@kernel.org> | 2013-02-06 19:36:14 +0100 |
commit | db8450673458e724229e91050ac9a92253b01234 (patch) | |
tree | 32e380798c28ca23b09855ac92fffd3400b64536 /mm | |
parent | slab: Handle ARCH_DMA_MINALIGN correctly (diff) | |
download | linux-db8450673458e724229e91050ac9a92253b01234.tar.xz linux-db8450673458e724229e91050ac9a92253b01234.zip |
slab: Fixup CONFIG_PAGE_ALLOC/DEBUG_SLAB_LEAK sections
Variables were not properly converted and the conversion caused
a naming conflict.
Signed-off-by: Tetsuo Handa <penguin-kernel@I-love.SAKURA.ne.jp>
Signed-off-by: Christoph Lameter <cl@linux.com>
Signed-off-by: Pekka Enberg <penberg@kernel.org>
Diffstat (limited to 'mm')
-rw-r--r-- | mm/slab.c | 29 |
1 files changed, 15 insertions, 14 deletions
diff --git a/mm/slab.c b/mm/slab.c index 17f859614546..910df6326831 100644 --- a/mm/slab.c +++ b/mm/slab.c @@ -2306,8 +2306,9 @@ __kmem_cache_create (struct kmem_cache *cachep, unsigned long flags) } #if FORCED_DEBUG && defined(CONFIG_DEBUG_PAGEALLOC) if (size >= kmalloc_size(INDEX_NODE + 1) - && cachep->object_size > cache_line_size() && ALIGN(size, align) < PAGE_SIZE) { - cachep->obj_offset += PAGE_SIZE - ALIGN(size, align); + && cachep->object_size > cache_line_size() + && ALIGN(size, cachep->align) < PAGE_SIZE) { + cachep->obj_offset += PAGE_SIZE - ALIGN(size, cachep->align); size = PAGE_SIZE; } #endif @@ -4377,7 +4378,7 @@ static int leaks_show(struct seq_file *m, void *p) struct slab *slabp; struct kmem_cache_node *n; const char *name; - unsigned long *n = m->private; + unsigned long *x = m->private; int node; int i; @@ -4388,7 +4389,7 @@ static int leaks_show(struct seq_file *m, void *p) /* OK, we can do it */ - n[1] = 0; + x[1] = 0; for_each_online_node(node) { n = cachep->node[node]; @@ -4399,32 +4400,32 @@ static int leaks_show(struct seq_file *m, void *p) spin_lock_irq(&n->list_lock); list_for_each_entry(slabp, &n->slabs_full, list) - handle_slab(n, cachep, slabp); + handle_slab(x, cachep, slabp); list_for_each_entry(slabp, &n->slabs_partial, list) - handle_slab(n, cachep, slabp); + handle_slab(x, cachep, slabp); spin_unlock_irq(&n->list_lock); } name = cachep->name; - if (n[0] == n[1]) { + if (x[0] == x[1]) { /* Increase the buffer size */ mutex_unlock(&slab_mutex); - m->private = kzalloc(n[0] * 4 * sizeof(unsigned long), GFP_KERNEL); + m->private = kzalloc(x[0] * 4 * sizeof(unsigned long), GFP_KERNEL); if (!m->private) { /* Too bad, we are really out */ - m->private = n; + m->private = x; mutex_lock(&slab_mutex); return -ENOMEM; } - *(unsigned long *)m->private = n[0] * 2; - kfree(n); + *(unsigned long *)m->private = x[0] * 2; + kfree(x); mutex_lock(&slab_mutex); /* Now make sure this entry will be retried */ m->count = m->size; return 0; } - for (i = 0; i < n[1]; i++) { - seq_printf(m, "%s: %lu ", name, n[2*i+3]); - show_symbol(m, n[2*i+2]); + for (i = 0; i < x[1]; i++) { + seq_printf(m, "%s: %lu ", name, x[2*i+3]); + show_symbol(m, x[2*i+2]); seq_putc(m, '\n'); } |