aboutsummaryrefslogtreecommitdiffstats
path: root/include/linux/rculist_bl.h
diff options
context:
space:
mode:
authorNick Piggin <npiggin@kernel.dk>2011-01-14 02:36:43 +0000
committerNick Piggin <npiggin@hera.kernel.org>2011-01-14 02:36:43 +0000
commit2c6755988afc003a0332406a134fb6a1626f9b28 (patch)
tree84eea829cab7b28585000f6fd7fe0dc176b746c5 /include/linux/rculist_bl.h
parent90dbb77ba48dddb87445d238e84cd137cf97dd98 (diff)
downloadkernel_samsung_tuna-2c6755988afc003a0332406a134fb6a1626f9b28.zip
kernel_samsung_tuna-2c6755988afc003a0332406a134fb6a1626f9b28.tar.gz
kernel_samsung_tuna-2c6755988afc003a0332406a134fb6a1626f9b28.tar.bz2
fs: hlist UP debug fixup
Po-Yu Chuang <ratbert.chuang@gmail.com> noticed that hlist_bl_set_first could crash on a UP system when LIST_BL_LOCKMASK is 0, because LIST_BL_BUG_ON(!((unsigned long)h->first & LIST_BL_LOCKMASK)); always evaulates to true. Fix the expression, and also avoid a dependency between bit spinlock implementation and list bl code (list code shouldn't know anything except that bit 0 is set when adding and removing elements). Eventually if a good use case comes up, we might use this list to store 1 or more arbitrary bits of data, so it really shouldn't be tied to locking either, but for now they are helpful for debugging. Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Diffstat (limited to 'include/linux/rculist_bl.h')
-rw-r--r--include/linux/rculist_bl.h3
1 files changed, 2 insertions, 1 deletions
diff --git a/include/linux/rculist_bl.h b/include/linux/rculist_bl.h
index b872b49..cf1244f 100644
--- a/include/linux/rculist_bl.h
+++ b/include/linux/rculist_bl.h
@@ -11,7 +11,8 @@ static inline void hlist_bl_set_first_rcu(struct hlist_bl_head *h,
struct hlist_bl_node *n)
{
LIST_BL_BUG_ON((unsigned long)n & LIST_BL_LOCKMASK);
- LIST_BL_BUG_ON(!((unsigned long)h->first & LIST_BL_LOCKMASK));
+ LIST_BL_BUG_ON(((unsigned long)h->first & LIST_BL_LOCKMASK) !=
+ LIST_BL_LOCKMASK);
rcu_assign_pointer(h->first,
(struct hlist_bl_node *)((unsigned long)n | LIST_BL_LOCKMASK));
}