summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorNeilBrown <neilb@suse.com>2019-04-02 01:07:45 +0200
committerDavid S. Miller <davem@davemloft.net>2019-04-08 04:12:12 +0200
commit7a41c294c1463100fdc82a356e22e36bbaa6b0f9 (patch)
tree136016cd26613f78f0c47d1d477eef9ac84061c3
parentMerge branch 'net-hsr-improvements-and-bug-fixes' (diff)
downloadlinux-7a41c294c1463100fdc82a356e22e36bbaa6b0f9.tar.xz
linux-7a41c294c1463100fdc82a356e22e36bbaa6b0f9.zip
rhashtable: use cmpxchg() in nested_table_alloc()
nested_table_alloc() relies on the fact that there is at most one spinlock allocated for every slot in the top level nested table, so it is not possible for two threads to try to allocate the same table at the same time. This assumption is a little fragile (it is not explicit) and is unnecessary as cmpxchg() can be used instead. A future patch will replace the spinlocks by per-bucket bitlocks, and then we won't be able to protect the slot pointer with a spinlock. So replace rcu_assign_pointer() with cmpxchg() - which has equivalent barrier properties. If it the cmp fails, free the table that was just allocated. Signed-off-by: NeilBrown <neilb@suse.com> Signed-off-by: David S. Miller <davem@davemloft.net>
-rw-r--r--lib/rhashtable.c8
1 files changed, 5 insertions, 3 deletions
diff --git a/lib/rhashtable.c b/lib/rhashtable.c
index 811d51b7cb86..6c4f5c8e9baa 100644
--- a/lib/rhashtable.c
+++ b/lib/rhashtable.c
@@ -131,9 +131,11 @@ static union nested_table *nested_table_alloc(struct rhashtable *ht,
INIT_RHT_NULLS_HEAD(ntbl[i].bucket);
}
- rcu_assign_pointer(*prev, ntbl);
-
- return ntbl;
+ if (cmpxchg(prev, NULL, ntbl) == NULL)
+ return ntbl;
+ /* Raced with another thread. */
+ kfree(ntbl);
+ return rcu_dereference(*prev);
}
static struct bucket_table *nested_bucket_table_alloc(struct rhashtable *ht,