context_tracking: Convert state to atomic_t

Context tracking's state and dynticks counter are going to be merged
in a single field so that both updates can happen atomically and at the
same time. Prepare for that with converting the state into an atomic_t.

[ paulmck: Apply kernel test robot feedback. ]

Signed-off-by: Frederic Weisbecker <frederic@kernel.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Neeraj Upadhyay <quic_neeraju@quicinc.com>
Cc: Uladzislau Rezki <uladzislau.rezki@sony.com>
Cc: Joel Fernandes <joel@joelfernandes.org>
Cc: Boqun Feng <boqun.feng@gmail.com>
Cc: Nicolas Saenz Julienne <nsaenz@kernel.org>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Xiongfeng Wang <wangxiongfeng2@huawei.com>
Cc: Yu Liao <liaoyu15@huawei.com>
Cc: Phil Auld <pauld@redhat.com>
Cc: Paul Gortmaker<paul.gortmaker@windriver.com>
Cc: Alex Belits <abelits@marvell.com>
Signed-off-by: Paul E. McKenney <paulmck@kernel.org>
Reviewed-by: Nicolas Saenz Julienne <nsaenzju@redhat.com>
Tested-by: Nicolas Saenz Julienne <nsaenzju@redhat.com>
This commit is contained in:
Frederic Weisbecker 2022-06-08 16:40:35 +02:00 committed by Paul E. McKenney
parent c33ef43a35
commit 171476775d
5 changed files with 159 additions and 92 deletions

View File

@ -56,7 +56,7 @@ static inline enum ctx_state exception_enter(void)
!context_tracking_enabled()) !context_tracking_enabled())
return 0; return 0;
prev_ctx = this_cpu_read(context_tracking.state); prev_ctx = __ct_state();
if (prev_ctx != CONTEXT_KERNEL) if (prev_ctx != CONTEXT_KERNEL)
ct_user_exit(prev_ctx); ct_user_exit(prev_ctx);
@ -86,33 +86,21 @@ static __always_inline void context_tracking_guest_exit(void)
__ct_user_exit(CONTEXT_GUEST); __ct_user_exit(CONTEXT_GUEST);
} }
/** #define CT_WARN_ON(cond) WARN_ON(context_tracking_enabled() && (cond))
* ct_state() - return the current context tracking state if known
*
* Returns the current cpu's context tracking state if context tracking
* is enabled. If context tracking is disabled, returns
* CONTEXT_DISABLED. This should be used primarily for debugging.
*/
static __always_inline enum ctx_state ct_state(void)
{
return context_tracking_enabled() ?
this_cpu_read(context_tracking.state) : CONTEXT_DISABLED;
}
#else #else
static inline void user_enter(void) { } static inline void user_enter(void) { }
static inline void user_exit(void) { } static inline void user_exit(void) { }
static inline void user_enter_irqoff(void) { } static inline void user_enter_irqoff(void) { }
static inline void user_exit_irqoff(void) { } static inline void user_exit_irqoff(void) { }
static inline enum ctx_state exception_enter(void) { return 0; } static inline int exception_enter(void) { return 0; }
static inline void exception_exit(enum ctx_state prev_ctx) { } static inline void exception_exit(enum ctx_state prev_ctx) { }
static inline enum ctx_state ct_state(void) { return CONTEXT_DISABLED; } static inline int ct_state(void) { return -1; }
static __always_inline bool context_tracking_guest_enter(void) { return false; } static __always_inline bool context_tracking_guest_enter(void) { return false; }
static inline void context_tracking_guest_exit(void) { } static inline void context_tracking_guest_exit(void) { }
#define CT_WARN_ON(cond) do { } while (0)
#endif /* !CONFIG_CONTEXT_TRACKING_USER */ #endif /* !CONFIG_CONTEXT_TRACKING_USER */
#define CT_WARN_ON(cond) WARN_ON(context_tracking_enabled() && (cond))
#ifdef CONFIG_CONTEXT_TRACKING_USER_FORCE #ifdef CONFIG_CONTEXT_TRACKING_USER_FORCE
extern void context_tracking_init(void); extern void context_tracking_init(void);
#else #else
@ -130,16 +118,16 @@ extern void ct_idle_exit(void);
*/ */
static __always_inline bool rcu_dynticks_curr_cpu_in_eqs(void) static __always_inline bool rcu_dynticks_curr_cpu_in_eqs(void)
{ {
return !(arch_atomic_read(this_cpu_ptr(&context_tracking.dynticks)) & 0x1); return !(arch_atomic_read(this_cpu_ptr(&context_tracking.state)) & RCU_DYNTICKS_IDX);
} }
/* /*
* Increment the current CPU's context_tracking structure's ->dynticks field * Increment the current CPU's context_tracking structure's ->state field
* with ordering. Return the new value. * with ordering. Return the new value.
*/ */
static __always_inline unsigned long rcu_dynticks_inc(int incby) static __always_inline unsigned long ct_state_inc(int incby)
{ {
return arch_atomic_add_return(incby, this_cpu_ptr(&context_tracking.dynticks)); return arch_atomic_add_return(incby, this_cpu_ptr(&context_tracking.state));
} }
#else #else

View File

@ -6,16 +6,24 @@
#include <linux/static_key.h> #include <linux/static_key.h>
#include <linux/context_tracking_irq.h> #include <linux/context_tracking_irq.h>
enum ctx_state {
CONTEXT_DISABLED = -1, /* returned by ct_state() if unknown */
CONTEXT_KERNEL = 0,
CONTEXT_USER,
CONTEXT_GUEST,
};
/* Offset to allow distinguishing irq vs. task-based idle entry/exit. */ /* Offset to allow distinguishing irq vs. task-based idle entry/exit. */
#define DYNTICK_IRQ_NONIDLE ((LONG_MAX / 2) + 1) #define DYNTICK_IRQ_NONIDLE ((LONG_MAX / 2) + 1)
enum ctx_state {
CONTEXT_DISABLED = -1, /* returned by ct_state() if unknown */
CONTEXT_KERNEL = 0,
CONTEXT_IDLE = 1,
CONTEXT_USER = 2,
CONTEXT_GUEST = 3,
CONTEXT_MAX = 4,
};
/* Even value for idle, else odd. */
#define RCU_DYNTICKS_IDX CONTEXT_MAX
#define CT_STATE_MASK (CONTEXT_MAX - 1)
#define CT_DYNTICKS_MASK (~CT_STATE_MASK)
struct context_tracking { struct context_tracking {
#ifdef CONFIG_CONTEXT_TRACKING_USER #ifdef CONFIG_CONTEXT_TRACKING_USER
/* /*
@ -26,10 +34,11 @@ struct context_tracking {
*/ */
bool active; bool active;
int recursion; int recursion;
enum ctx_state state; #endif
#ifdef CONFIG_CONTEXT_TRACKING
atomic_t state;
#endif #endif
#ifdef CONFIG_CONTEXT_TRACKING_IDLE #ifdef CONFIG_CONTEXT_TRACKING_IDLE
atomic_t dynticks; /* Even value for idle, else odd. */
long dynticks_nesting; /* Track process nesting level. */ long dynticks_nesting; /* Track process nesting level. */
long dynticks_nmi_nesting; /* Track irq/NMI nesting level. */ long dynticks_nmi_nesting; /* Track irq/NMI nesting level. */
#endif #endif
@ -37,26 +46,31 @@ struct context_tracking {
#ifdef CONFIG_CONTEXT_TRACKING #ifdef CONFIG_CONTEXT_TRACKING
DECLARE_PER_CPU(struct context_tracking, context_tracking); DECLARE_PER_CPU(struct context_tracking, context_tracking);
static __always_inline int __ct_state(void)
{
return atomic_read(this_cpu_ptr(&context_tracking.state)) & CT_STATE_MASK;
}
#endif #endif
#ifdef CONFIG_CONTEXT_TRACKING_IDLE #ifdef CONFIG_CONTEXT_TRACKING_IDLE
static __always_inline int ct_dynticks(void) static __always_inline int ct_dynticks(void)
{ {
return atomic_read(this_cpu_ptr(&context_tracking.dynticks)); return atomic_read(this_cpu_ptr(&context_tracking.state)) & CT_DYNTICKS_MASK;
} }
static __always_inline int ct_dynticks_cpu(int cpu) static __always_inline int ct_dynticks_cpu(int cpu)
{ {
struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu); struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu);
return atomic_read(&ct->dynticks); return atomic_read(&ct->state) & CT_DYNTICKS_MASK;
} }
static __always_inline int ct_dynticks_cpu_acquire(int cpu) static __always_inline int ct_dynticks_cpu_acquire(int cpu)
{ {
struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu); struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu);
return atomic_read_acquire(&ct->dynticks); return atomic_read_acquire(&ct->state) & CT_DYNTICKS_MASK;
} }
static __always_inline long ct_dynticks_nesting(void) static __always_inline long ct_dynticks_nesting(void)
@ -102,6 +116,27 @@ static inline bool context_tracking_enabled_this_cpu(void)
return context_tracking_enabled() && __this_cpu_read(context_tracking.active); return context_tracking_enabled() && __this_cpu_read(context_tracking.active);
} }
/**
* ct_state() - return the current context tracking state if known
*
* Returns the current cpu's context tracking state if context tracking
* is enabled. If context tracking is disabled, returns
* CONTEXT_DISABLED. This should be used primarily for debugging.
*/
static __always_inline int ct_state(void)
{
int ret;
if (!context_tracking_enabled())
return CONTEXT_DISABLED;
preempt_disable();
ret = __ct_state();
preempt_enable();
return ret;
}
#else #else
static __always_inline bool context_tracking_enabled(void) { return false; } static __always_inline bool context_tracking_enabled(void) { return false; }
static __always_inline bool context_tracking_enabled_cpu(int cpu) { return false; } static __always_inline bool context_tracking_enabled_cpu(int cpu) { return false; }

View File

@ -28,8 +28,8 @@ DEFINE_PER_CPU(struct context_tracking, context_tracking) = {
#ifdef CONFIG_CONTEXT_TRACKING_IDLE #ifdef CONFIG_CONTEXT_TRACKING_IDLE
.dynticks_nesting = 1, .dynticks_nesting = 1,
.dynticks_nmi_nesting = DYNTICK_IRQ_NONIDLE, .dynticks_nmi_nesting = DYNTICK_IRQ_NONIDLE,
.dynticks = ATOMIC_INIT(1),
#endif #endif
.state = ATOMIC_INIT(RCU_DYNTICKS_IDX),
}; };
EXPORT_SYMBOL_GPL(context_tracking); EXPORT_SYMBOL_GPL(context_tracking);
@ -76,7 +76,7 @@ static __always_inline void rcu_dynticks_task_trace_exit(void)
* RCU is watching prior to the call to this function and is no longer * RCU is watching prior to the call to this function and is no longer
* watching upon return. * watching upon return.
*/ */
static noinstr void rcu_dynticks_eqs_enter(void) static noinstr void ct_kernel_exit_state(int offset)
{ {
int seq; int seq;
@ -86,9 +86,9 @@ static noinstr void rcu_dynticks_eqs_enter(void)
* next idle sojourn. * next idle sojourn.
*/ */
rcu_dynticks_task_trace_enter(); // Before ->dynticks update! rcu_dynticks_task_trace_enter(); // Before ->dynticks update!
seq = rcu_dynticks_inc(1); seq = ct_state_inc(offset);
// RCU is no longer watching. Better be in extended quiescent state! // RCU is no longer watching. Better be in extended quiescent state!
WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && (seq & 0x1)); WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && (seq & RCU_DYNTICKS_IDX));
} }
/* /*
@ -96,7 +96,7 @@ static noinstr void rcu_dynticks_eqs_enter(void)
* called from an extended quiescent state, that is, RCU is not watching * called from an extended quiescent state, that is, RCU is not watching
* prior to the call to this function and is watching upon return. * prior to the call to this function and is watching upon return.
*/ */
static noinstr void rcu_dynticks_eqs_exit(void) static noinstr void ct_kernel_enter_state(int offset)
{ {
int seq; int seq;
@ -105,10 +105,10 @@ static noinstr void rcu_dynticks_eqs_exit(void)
* and we also must force ordering with the next RCU read-side * and we also must force ordering with the next RCU read-side
* critical section. * critical section.
*/ */
seq = rcu_dynticks_inc(1); seq = ct_state_inc(offset);
// RCU is now watching. Better not be in an extended quiescent state! // RCU is now watching. Better not be in an extended quiescent state!
rcu_dynticks_task_trace_exit(); // After ->dynticks update! rcu_dynticks_task_trace_exit(); // After ->dynticks update!
WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !(seq & 0x1)); WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !(seq & RCU_DYNTICKS_IDX));
} }
/* /*
@ -119,7 +119,7 @@ static noinstr void rcu_dynticks_eqs_exit(void)
* the possibility of usermode upcalls having messed up our count * the possibility of usermode upcalls having messed up our count
* of interrupt nesting level during the prior busy period. * of interrupt nesting level during the prior busy period.
*/ */
static void noinstr rcu_eqs_enter(bool user) static void noinstr ct_kernel_exit(bool user, int offset)
{ {
struct context_tracking *ct = this_cpu_ptr(&context_tracking); struct context_tracking *ct = this_cpu_ptr(&context_tracking);
@ -139,13 +139,13 @@ static void noinstr rcu_eqs_enter(bool user)
WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !user && !is_idle_task(current)); WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !user && !is_idle_task(current));
rcu_preempt_deferred_qs(current); rcu_preempt_deferred_qs(current);
// instrumentation for the noinstr rcu_dynticks_eqs_enter() // instrumentation for the noinstr ct_kernel_exit_state()
instrument_atomic_write(&ct->dynticks, sizeof(ct->dynticks)); instrument_atomic_write(&ct->state, sizeof(ct->state));
instrumentation_end(); instrumentation_end();
WRITE_ONCE(ct->dynticks_nesting, 0); /* Avoid irq-access tearing. */ WRITE_ONCE(ct->dynticks_nesting, 0); /* Avoid irq-access tearing. */
// RCU is watching here ... // RCU is watching here ...
rcu_dynticks_eqs_enter(); ct_kernel_exit_state(offset);
// ... but is no longer watching here. // ... but is no longer watching here.
rcu_dynticks_task_enter(); rcu_dynticks_task_enter();
} }
@ -158,7 +158,7 @@ static void noinstr rcu_eqs_enter(bool user)
* allow for the possibility of usermode upcalls messing up our count of * allow for the possibility of usermode upcalls messing up our count of
* interrupt nesting level during the busy period that is just now starting. * interrupt nesting level during the busy period that is just now starting.
*/ */
static void noinstr rcu_eqs_exit(bool user) static void noinstr ct_kernel_enter(bool user, int offset)
{ {
struct context_tracking *ct = this_cpu_ptr(&context_tracking); struct context_tracking *ct = this_cpu_ptr(&context_tracking);
long oldval; long oldval;
@ -173,12 +173,12 @@ static void noinstr rcu_eqs_exit(bool user)
} }
rcu_dynticks_task_exit(); rcu_dynticks_task_exit();
// RCU is not watching here ... // RCU is not watching here ...
rcu_dynticks_eqs_exit(); ct_kernel_enter_state(offset);
// ... but is watching here. // ... but is watching here.
instrumentation_begin(); instrumentation_begin();
// instrumentation for the noinstr rcu_dynticks_eqs_exit() // instrumentation for the noinstr ct_kernel_enter_state()
instrument_atomic_write(&ct->dynticks, sizeof(ct->dynticks)); instrument_atomic_write(&ct->state, sizeof(ct->state));
trace_rcu_dyntick(TPS("End"), ct_dynticks_nesting(), 1, ct_dynticks()); trace_rcu_dyntick(TPS("End"), ct_dynticks_nesting(), 1, ct_dynticks());
WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !user && !is_idle_task(current)); WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !user && !is_idle_task(current));
@ -192,7 +192,7 @@ static void noinstr rcu_eqs_exit(bool user)
* ct_nmi_exit - inform RCU of exit from NMI context * ct_nmi_exit - inform RCU of exit from NMI context
* *
* If we are returning from the outermost NMI handler that interrupted an * If we are returning from the outermost NMI handler that interrupted an
* RCU-idle period, update ct->dynticks and ct->dynticks_nmi_nesting * RCU-idle period, update ct->state and ct->dynticks_nmi_nesting
* to let the RCU grace-period handling know that the CPU is back to * to let the RCU grace-period handling know that the CPU is back to
* being RCU-idle. * being RCU-idle.
* *
@ -229,12 +229,12 @@ void noinstr ct_nmi_exit(void)
trace_rcu_dyntick(TPS("Startirq"), ct_dynticks_nmi_nesting(), 0, ct_dynticks()); trace_rcu_dyntick(TPS("Startirq"), ct_dynticks_nmi_nesting(), 0, ct_dynticks());
WRITE_ONCE(ct->dynticks_nmi_nesting, 0); /* Avoid store tearing. */ WRITE_ONCE(ct->dynticks_nmi_nesting, 0); /* Avoid store tearing. */
// instrumentation for the noinstr rcu_dynticks_eqs_enter() // instrumentation for the noinstr ct_kernel_exit_state()
instrument_atomic_write(&ct->dynticks, sizeof(ct->dynticks)); instrument_atomic_write(&ct->state, sizeof(ct->state));
instrumentation_end(); instrumentation_end();
// RCU is watching here ... // RCU is watching here ...
rcu_dynticks_eqs_enter(); ct_kernel_exit_state(RCU_DYNTICKS_IDX);
// ... but is no longer watching here. // ... but is no longer watching here.
if (!in_nmi()) if (!in_nmi())
@ -244,7 +244,7 @@ void noinstr ct_nmi_exit(void)
/** /**
* ct_nmi_enter - inform RCU of entry to NMI context * ct_nmi_enter - inform RCU of entry to NMI context
* *
* If the CPU was idle from RCU's viewpoint, update ct->dynticks and * If the CPU was idle from RCU's viewpoint, update ct->state and
* ct->dynticks_nmi_nesting to let the RCU grace-period handling know * ct->dynticks_nmi_nesting to let the RCU grace-period handling know
* that the CPU is active. This implementation permits nested NMIs, as * that the CPU is active. This implementation permits nested NMIs, as
* long as the nesting level does not overflow an int. (You will probably * long as the nesting level does not overflow an int. (You will probably
@ -275,14 +275,14 @@ void noinstr ct_nmi_enter(void)
rcu_dynticks_task_exit(); rcu_dynticks_task_exit();
// RCU is not watching here ... // RCU is not watching here ...
rcu_dynticks_eqs_exit(); ct_kernel_enter_state(RCU_DYNTICKS_IDX);
// ... but is watching here. // ... but is watching here.
instrumentation_begin(); instrumentation_begin();
// instrumentation for the noinstr rcu_dynticks_curr_cpu_in_eqs() // instrumentation for the noinstr rcu_dynticks_curr_cpu_in_eqs()
instrument_atomic_read(&ct->dynticks, sizeof(ct->dynticks)); instrument_atomic_read(&ct->state, sizeof(ct->state));
// instrumentation for the noinstr rcu_dynticks_eqs_exit() // instrumentation for the noinstr ct_kernel_enter_state()
instrument_atomic_write(&ct->dynticks, sizeof(ct->dynticks)); instrument_atomic_write(&ct->state, sizeof(ct->state));
incby = 1; incby = 1;
} else if (!in_nmi()) { } else if (!in_nmi()) {
@ -315,7 +315,7 @@ void noinstr ct_nmi_enter(void)
void noinstr ct_idle_enter(void) void noinstr ct_idle_enter(void)
{ {
WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !raw_irqs_disabled()); WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !raw_irqs_disabled());
rcu_eqs_enter(false); ct_kernel_exit(false, RCU_DYNTICKS_IDX + CONTEXT_IDLE);
} }
EXPORT_SYMBOL_GPL(ct_idle_enter); EXPORT_SYMBOL_GPL(ct_idle_enter);
@ -333,7 +333,7 @@ void noinstr ct_idle_exit(void)
unsigned long flags; unsigned long flags;
raw_local_irq_save(flags); raw_local_irq_save(flags);
rcu_eqs_exit(false); ct_kernel_enter(false, RCU_DYNTICKS_IDX - CONTEXT_IDLE);
raw_local_irq_restore(flags); raw_local_irq_restore(flags);
} }
EXPORT_SYMBOL_GPL(ct_idle_exit); EXPORT_SYMBOL_GPL(ct_idle_exit);
@ -421,8 +421,8 @@ void ct_irq_exit_irqson(void)
local_irq_restore(flags); local_irq_restore(flags);
} }
#else #else
static __always_inline void rcu_eqs_enter(bool user) { } static __always_inline void ct_kernel_exit(bool user, int offset) { }
static __always_inline void rcu_eqs_exit(bool user) { } static __always_inline void ct_kernel_enter(bool user, int offset) { }
#endif /* #ifdef CONFIG_CONTEXT_TRACKING_IDLE */ #endif /* #ifdef CONFIG_CONTEXT_TRACKING_IDLE */
#ifdef CONFIG_CONTEXT_TRACKING_USER #ifdef CONFIG_CONTEXT_TRACKING_USER
@ -463,6 +463,7 @@ static __always_inline void context_tracking_recursion_exit(void)
*/ */
void noinstr __ct_user_enter(enum ctx_state state) void noinstr __ct_user_enter(enum ctx_state state)
{ {
struct context_tracking *ct = this_cpu_ptr(&context_tracking);
lockdep_assert_irqs_disabled(); lockdep_assert_irqs_disabled();
/* Kernel threads aren't supposed to go to userspace */ /* Kernel threads aren't supposed to go to userspace */
@ -471,8 +472,8 @@ void noinstr __ct_user_enter(enum ctx_state state)
if (!context_tracking_recursion_enter()) if (!context_tracking_recursion_enter())
return; return;
if ( __this_cpu_read(context_tracking.state) != state) { if (__ct_state() != state) {
if (__this_cpu_read(context_tracking.active)) { if (ct->active) {
/* /*
* At this stage, only low level arch entry code remains and * At this stage, only low level arch entry code remains and
* then we'll run in userspace. We can assume there won't be * then we'll run in userspace. We can assume there won't be
@ -492,28 +493,49 @@ void noinstr __ct_user_enter(enum ctx_state state)
* that will fire and reschedule once we resume in user/guest mode. * that will fire and reschedule once we resume in user/guest mode.
*/ */
rcu_irq_work_resched(); rcu_irq_work_resched();
/* /*
* Enter RCU idle mode right before resuming userspace. No use of RCU * Enter RCU idle mode right before resuming userspace. No use of RCU
* is permitted between this call and rcu_eqs_exit(). This way the * is permitted between this call and rcu_eqs_exit(). This way the
* CPU doesn't need to maintain the tick for RCU maintenance purposes * CPU doesn't need to maintain the tick for RCU maintenance purposes
* when the CPU runs in userspace. * when the CPU runs in userspace.
*/ */
rcu_eqs_enter(true); ct_kernel_exit(true, RCU_DYNTICKS_IDX + state);
/*
* Special case if we only track user <-> kernel transitions for tickless
* cputime accounting but we don't support RCU extended quiescent state.
* In this we case we don't care about any concurrency/ordering.
*/
if (!IS_ENABLED(CONFIG_CONTEXT_TRACKING_IDLE))
atomic_set(&ct->state, state);
} else {
/*
* Even if context tracking is disabled on this CPU, because it's outside
* the full dynticks mask for example, we still have to keep track of the
* context transitions and states to prevent inconsistency on those of
* other CPUs.
* If a task triggers an exception in userspace, sleep on the exception
* handler and then migrate to another CPU, that new CPU must know where
* the exception returns by the time we call exception_exit().
* This information can only be provided by the previous CPU when it called
* exception_enter().
* OTOH we can spare the calls to vtime and RCU when context_tracking.active
* is false because we know that CPU is not tickless.
*/
if (!IS_ENABLED(CONFIG_CONTEXT_TRACKING_IDLE)) {
/* Tracking for vtime only, no concurrent RCU EQS accounting */
atomic_set(&ct->state, state);
} else {
/*
* Tracking for vtime and RCU EQS. Make sure we don't race
* with NMIs. OTOH we don't care about ordering here since
* RCU only requires RCU_DYNTICKS_IDX increments to be fully
* ordered.
*/
atomic_add(state, &ct->state);
}
} }
/*
* Even if context tracking is disabled on this CPU, because it's outside
* the full dynticks mask for example, we still have to keep track of the
* context transitions and states to prevent inconsistency on those of
* other CPUs.
* If a task triggers an exception in userspace, sleep on the exception
* handler and then migrate to another CPU, that new CPU must know where
* the exception returns by the time we call exception_exit().
* This information can only be provided by the previous CPU when it called
* exception_enter().
* OTOH we can spare the calls to vtime and RCU when context_tracking.active
* is false because we know that CPU is not tickless.
*/
__this_cpu_write(context_tracking.state, state);
} }
context_tracking_recursion_exit(); context_tracking_recursion_exit();
} }
@ -581,24 +603,47 @@ NOKPROBE_SYMBOL(user_enter_callable);
*/ */
void noinstr __ct_user_exit(enum ctx_state state) void noinstr __ct_user_exit(enum ctx_state state)
{ {
struct context_tracking *ct = this_cpu_ptr(&context_tracking);
if (!context_tracking_recursion_enter()) if (!context_tracking_recursion_enter())
return; return;
if (__this_cpu_read(context_tracking.state) == state) { if (__ct_state() == state) {
if (__this_cpu_read(context_tracking.active)) { if (ct->active) {
/* /*
* Exit RCU idle mode while entering the kernel because it can * Exit RCU idle mode while entering the kernel because it can
* run a RCU read side critical section anytime. * run a RCU read side critical section anytime.
*/ */
rcu_eqs_exit(true); ct_kernel_enter(true, RCU_DYNTICKS_IDX - state);
if (state == CONTEXT_USER) { if (state == CONTEXT_USER) {
instrumentation_begin(); instrumentation_begin();
vtime_user_exit(current); vtime_user_exit(current);
trace_user_exit(0); trace_user_exit(0);
instrumentation_end(); instrumentation_end();
} }
/*
* Special case if we only track user <-> kernel transitions for tickless
* cputime accounting but we don't support RCU extended quiescent state.
* In this we case we don't care about any concurrency/ordering.
*/
if (!IS_ENABLED(CONFIG_CONTEXT_TRACKING_IDLE))
atomic_set(&ct->state, CONTEXT_KERNEL);
} else {
if (!IS_ENABLED(CONFIG_CONTEXT_TRACKING_IDLE)) {
/* Tracking for vtime only, no concurrent RCU EQS accounting */
atomic_set(&ct->state, CONTEXT_KERNEL);
} else {
/*
* Tracking for vtime and RCU EQS. Make sure we don't race
* with NMIs. OTOH we don't care about ordering here since
* RCU only requires RCU_DYNTICKS_IDX increments to be fully
* ordered.
*/
atomic_sub(state, &ct->state);
}
} }
__this_cpu_write(context_tracking.state, CONTEXT_KERNEL);
} }
context_tracking_recursion_exit(); context_tracking_recursion_exit();
} }

View File

@ -272,9 +272,9 @@ void rcu_softirq_qs(void)
*/ */
static void rcu_dynticks_eqs_online(void) static void rcu_dynticks_eqs_online(void)
{ {
if (ct_dynticks() & 0x1) if (ct_dynticks() & RCU_DYNTICKS_IDX)
return; return;
rcu_dynticks_inc(1); ct_state_inc(RCU_DYNTICKS_IDX);
} }
/* /*
@ -293,7 +293,7 @@ static int rcu_dynticks_snap(int cpu)
*/ */
static bool rcu_dynticks_in_eqs(int snap) static bool rcu_dynticks_in_eqs(int snap)
{ {
return !(snap & 0x1); return !(snap & RCU_DYNTICKS_IDX);
} }
/* Return true if the specified CPU is currently idle from an RCU viewpoint. */ /* Return true if the specified CPU is currently idle from an RCU viewpoint. */
@ -321,8 +321,7 @@ bool rcu_dynticks_zero_in_eqs(int cpu, int *vp)
int snap; int snap;
// If not quiescent, force back to earlier extended quiescent state. // If not quiescent, force back to earlier extended quiescent state.
snap = ct_dynticks_cpu(cpu) & ~0x1; snap = ct_dynticks_cpu(cpu) & ~RCU_DYNTICKS_IDX;
smp_rmb(); // Order ->dynticks and *vp reads. smp_rmb(); // Order ->dynticks and *vp reads.
if (READ_ONCE(*vp)) if (READ_ONCE(*vp))
return false; // Non-zero, so report failure; return false; // Non-zero, so report failure;
@ -348,9 +347,9 @@ notrace void rcu_momentary_dyntick_idle(void)
int seq; int seq;
raw_cpu_write(rcu_data.rcu_need_heavy_qs, false); raw_cpu_write(rcu_data.rcu_need_heavy_qs, false);
seq = rcu_dynticks_inc(2); seq = ct_state_inc(2 * RCU_DYNTICKS_IDX);
/* It is illegal to call this from idle state. */ /* It is illegal to call this from idle state. */
WARN_ON_ONCE(!(seq & 0x1)); WARN_ON_ONCE(!(seq & RCU_DYNTICKS_IDX));
rcu_preempt_deferred_qs(current); rcu_preempt_deferred_qs(current);
} }
EXPORT_SYMBOL_GPL(rcu_momentary_dyntick_idle); EXPORT_SYMBOL_GPL(rcu_momentary_dyntick_idle);

View File

@ -469,7 +469,7 @@ static void print_cpu_stall_info(int cpu)
rcuc_starved = rcu_is_rcuc_kthread_starving(rdp, &j); rcuc_starved = rcu_is_rcuc_kthread_starving(rdp, &j);
if (rcuc_starved) if (rcuc_starved)
sprintf(buf, " rcuc=%ld jiffies(starved)", j); sprintf(buf, " rcuc=%ld jiffies(starved)", j);
pr_err("\t%d-%c%c%c%c: (%lu %s) idle=%03x/%ld/%#lx softirq=%u/%u fqs=%ld%s%s\n", pr_err("\t%d-%c%c%c%c: (%lu %s) idle=%04x/%ld/%#lx softirq=%u/%u fqs=%ld%s%s\n",
cpu, cpu,
"O."[!!cpu_online(cpu)], "O."[!!cpu_online(cpu)],
"o."[!!(rdp->grpmask & rdp->mynode->qsmaskinit)], "o."[!!(rdp->grpmask & rdp->mynode->qsmaskinit)],
@ -478,7 +478,7 @@ static void print_cpu_stall_info(int cpu)
rdp->rcu_iw_pending ? (int)min(delta, 9UL) + '0' : rdp->rcu_iw_pending ? (int)min(delta, 9UL) + '0' :
"!."[!delta], "!."[!delta],
ticks_value, ticks_title, ticks_value, ticks_title,
rcu_dynticks_snap(cpu) & 0xfff, rcu_dynticks_snap(cpu) & 0xffff,
ct_dynticks_nesting_cpu(cpu), ct_dynticks_nmi_nesting_cpu(cpu), ct_dynticks_nesting_cpu(cpu), ct_dynticks_nmi_nesting_cpu(cpu),
rdp->softirq_snap, kstat_softirqs_cpu(RCU_SOFTIRQ, cpu), rdp->softirq_snap, kstat_softirqs_cpu(RCU_SOFTIRQ, cpu),
data_race(rcu_state.n_force_qs) - rcu_state.n_force_qs_gpstart, data_race(rcu_state.n_force_qs) - rcu_state.n_force_qs_gpstart,