mirror of
https://gitee.com/bianbu-linux/linux-6.6
synced 2025-07-08 00:23:23 -04:00
seqlock: seqcount_t: Implement all read APIs as statement expressions
The sequence counters read APIs are implemented as CPP macros, so they can take either seqcount_t or any of the seqcount_LOCKNAME_t variants. Such macros then get *directly* transformed to internal C functions that only take plain seqcount_t. Further commits need access to seqcount_LOCKNAME_t inside of the actual read APIs code. Thus transform all of the seqcount read APIs to pure GCC statement expressions instead. This will not break type-safety: all of the transformed APIs resolve to a _Generic() selection that does not have a "default" case. This will also not affect the transformed APIs readability: previously added kernel-doc above all of seqlock.h functions makes the expectations quite clear for call-site developers. Signed-off-by: Ahmed S. Darwish <a.darwish@linutronix.de> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Link: https://lkml.kernel.org/r/20200904153231.11994-4-a.darwish@linutronix.de
This commit is contained in:
parent
5cdd25572a
commit
52ac39e5db
1 changed files with 45 additions and 49 deletions
|
@ -184,6 +184,12 @@ __seqprop_##lockname##_ptr(seqcount_##lockname##_t *s) \
|
||||||
return &s->seqcount; \
|
return &s->seqcount; \
|
||||||
} \
|
} \
|
||||||
\
|
\
|
||||||
|
static __always_inline unsigned \
|
||||||
|
__seqprop_##lockname##_sequence(const seqcount_##lockname##_t *s) \
|
||||||
|
{ \
|
||||||
|
return READ_ONCE(s->seqcount.sequence); \
|
||||||
|
} \
|
||||||
|
\
|
||||||
static __always_inline bool \
|
static __always_inline bool \
|
||||||
__seqprop_##lockname##_preemptible(const seqcount_##lockname##_t *s) \
|
__seqprop_##lockname##_preemptible(const seqcount_##lockname##_t *s) \
|
||||||
{ \
|
{ \
|
||||||
|
@ -205,6 +211,11 @@ static inline seqcount_t *__seqprop_ptr(seqcount_t *s)
|
||||||
return s;
|
return s;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline unsigned __seqprop_sequence(const seqcount_t *s)
|
||||||
|
{
|
||||||
|
return READ_ONCE(s->sequence);
|
||||||
|
}
|
||||||
|
|
||||||
static inline bool __seqprop_preemptible(const seqcount_t *s)
|
static inline bool __seqprop_preemptible(const seqcount_t *s)
|
||||||
{
|
{
|
||||||
return false;
|
return false;
|
||||||
|
@ -250,6 +261,7 @@ SEQCOUNT_LOCKNAME(ww_mutex, struct ww_mutex, true, &s->lock->base)
|
||||||
__seqprop_case((s), ww_mutex, prop))
|
__seqprop_case((s), ww_mutex, prop))
|
||||||
|
|
||||||
#define __seqcount_ptr(s) __seqprop(s, ptr)
|
#define __seqcount_ptr(s) __seqprop(s, ptr)
|
||||||
|
#define __seqcount_sequence(s) __seqprop(s, sequence)
|
||||||
#define __seqcount_lock_preemptible(s) __seqprop(s, preemptible)
|
#define __seqcount_lock_preemptible(s) __seqprop(s, preemptible)
|
||||||
#define __seqcount_assert_lock_held(s) __seqprop(s, assert)
|
#define __seqcount_assert_lock_held(s) __seqprop(s, assert)
|
||||||
|
|
||||||
|
@ -268,21 +280,15 @@ SEQCOUNT_LOCKNAME(ww_mutex, struct ww_mutex, true, &s->lock->base)
|
||||||
* Return: count to be passed to read_seqcount_retry()
|
* Return: count to be passed to read_seqcount_retry()
|
||||||
*/
|
*/
|
||||||
#define __read_seqcount_begin(s) \
|
#define __read_seqcount_begin(s) \
|
||||||
__read_seqcount_t_begin(__seqcount_ptr(s))
|
({ \
|
||||||
|
unsigned seq; \
|
||||||
static inline unsigned __read_seqcount_t_begin(const seqcount_t *s)
|
\
|
||||||
{
|
while ((seq = __seqcount_sequence(s)) & 1) \
|
||||||
unsigned ret;
|
cpu_relax(); \
|
||||||
|
\
|
||||||
repeat:
|
kcsan_atomic_next(KCSAN_SEQLOCK_REGION_MAX); \
|
||||||
ret = READ_ONCE(s->sequence);
|
seq; \
|
||||||
if (unlikely(ret & 1)) {
|
})
|
||||||
cpu_relax();
|
|
||||||
goto repeat;
|
|
||||||
}
|
|
||||||
kcsan_atomic_next(KCSAN_SEQLOCK_REGION_MAX);
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* raw_read_seqcount_begin() - begin a seqcount_t read section w/o lockdep
|
* raw_read_seqcount_begin() - begin a seqcount_t read section w/o lockdep
|
||||||
|
@ -291,14 +297,12 @@ repeat:
|
||||||
* Return: count to be passed to read_seqcount_retry()
|
* Return: count to be passed to read_seqcount_retry()
|
||||||
*/
|
*/
|
||||||
#define raw_read_seqcount_begin(s) \
|
#define raw_read_seqcount_begin(s) \
|
||||||
raw_read_seqcount_t_begin(__seqcount_ptr(s))
|
({ \
|
||||||
|
unsigned seq = __read_seqcount_begin(s); \
|
||||||
static inline unsigned raw_read_seqcount_t_begin(const seqcount_t *s)
|
\
|
||||||
{
|
smp_rmb(); \
|
||||||
unsigned ret = __read_seqcount_t_begin(s);
|
seq; \
|
||||||
smp_rmb();
|
})
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* read_seqcount_begin() - begin a seqcount_t read critical section
|
* read_seqcount_begin() - begin a seqcount_t read critical section
|
||||||
|
@ -307,13 +311,10 @@ static inline unsigned raw_read_seqcount_t_begin(const seqcount_t *s)
|
||||||
* Return: count to be passed to read_seqcount_retry()
|
* Return: count to be passed to read_seqcount_retry()
|
||||||
*/
|
*/
|
||||||
#define read_seqcount_begin(s) \
|
#define read_seqcount_begin(s) \
|
||||||
read_seqcount_t_begin(__seqcount_ptr(s))
|
({ \
|
||||||
|
seqcount_lockdep_reader_access(__seqcount_ptr(s)); \
|
||||||
static inline unsigned read_seqcount_t_begin(const seqcount_t *s)
|
raw_read_seqcount_begin(s); \
|
||||||
{
|
})
|
||||||
seqcount_lockdep_reader_access(s);
|
|
||||||
return raw_read_seqcount_t_begin(s);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* raw_read_seqcount() - read the raw seqcount_t counter value
|
* raw_read_seqcount() - read the raw seqcount_t counter value
|
||||||
|
@ -327,15 +328,13 @@ static inline unsigned read_seqcount_t_begin(const seqcount_t *s)
|
||||||
* Return: count to be passed to read_seqcount_retry()
|
* Return: count to be passed to read_seqcount_retry()
|
||||||
*/
|
*/
|
||||||
#define raw_read_seqcount(s) \
|
#define raw_read_seqcount(s) \
|
||||||
raw_read_seqcount_t(__seqcount_ptr(s))
|
({ \
|
||||||
|
unsigned seq = __seqcount_sequence(s); \
|
||||||
static inline unsigned raw_read_seqcount_t(const seqcount_t *s)
|
\
|
||||||
{
|
smp_rmb(); \
|
||||||
unsigned ret = READ_ONCE(s->sequence);
|
kcsan_atomic_next(KCSAN_SEQLOCK_REGION_MAX); \
|
||||||
smp_rmb();
|
seq; \
|
||||||
kcsan_atomic_next(KCSAN_SEQLOCK_REGION_MAX);
|
})
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* raw_seqcount_begin() - begin a seqcount_t read critical section w/o
|
* raw_seqcount_begin() - begin a seqcount_t read critical section w/o
|
||||||
|
@ -355,16 +354,13 @@ static inline unsigned raw_read_seqcount_t(const seqcount_t *s)
|
||||||
* Return: count to be passed to read_seqcount_retry()
|
* Return: count to be passed to read_seqcount_retry()
|
||||||
*/
|
*/
|
||||||
#define raw_seqcount_begin(s) \
|
#define raw_seqcount_begin(s) \
|
||||||
raw_seqcount_t_begin(__seqcount_ptr(s))
|
({ \
|
||||||
|
/* \
|
||||||
static inline unsigned raw_seqcount_t_begin(const seqcount_t *s)
|
* If the counter is odd, let read_seqcount_retry() fail \
|
||||||
{
|
* by decrementing the counter. \
|
||||||
/*
|
*/ \
|
||||||
* If the counter is odd, let read_seqcount_retry() fail
|
raw_read_seqcount(s) & ~1; \
|
||||||
* by decrementing the counter.
|
})
|
||||||
*/
|
|
||||||
return raw_read_seqcount_t(s) & ~1;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* __read_seqcount_retry() - end a seqcount_t read section w/o barrier
|
* __read_seqcount_retry() - end a seqcount_t read section w/o barrier
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue