mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-01-24 01:09:38 -05:00
rcu: Make exit_rcu() more precise and consolidate
When running preemptible RCU, if a task exits in an RCU read-side critical section having blocked within that same RCU read-side critical section, the task must be removed from the list of tasks blocking a grace period (perhaps the current grace period, perhaps the next grace period, depending on timing). The exit() path invokes exit_rcu() to do this cleanup. However, the current implementation of exit_rcu() needlessly does the cleanup even if the task did not block within the current RCU read-side critical section, which wastes time and needlessly increases the size of the state space. Fix this by only doing the cleanup if the current task is actually on the list of tasks blocking some grace period. While we are at it, consolidate the two identical exit_rcu() functions into a single function. Signed-off-by: Paul E. McKenney <paul.mckenney@linaro.org> Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com> Tested-by: Linus Torvalds <torvalds@linux-foundation.org> Conflicts: kernel/rcupdate.c
This commit is contained in:
parent
616c310e83
commit
9dd8fb16c3
6 changed files with 29 additions and 49 deletions
|
@ -191,6 +191,7 @@ extern void rcu_idle_enter(void);
|
||||||
extern void rcu_idle_exit(void);
|
extern void rcu_idle_exit(void);
|
||||||
extern void rcu_irq_enter(void);
|
extern void rcu_irq_enter(void);
|
||||||
extern void rcu_irq_exit(void);
|
extern void rcu_irq_exit(void);
|
||||||
|
extern void exit_rcu(void);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* RCU_NONIDLE - Indicate idle-loop code that needs RCU readers
|
* RCU_NONIDLE - Indicate idle-loop code that needs RCU readers
|
||||||
|
|
|
@ -87,10 +87,6 @@ static inline void kfree_call_rcu(struct rcu_head *head,
|
||||||
|
|
||||||
#ifdef CONFIG_TINY_RCU
|
#ifdef CONFIG_TINY_RCU
|
||||||
|
|
||||||
static inline void exit_rcu(void)
|
|
||||||
{
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline int rcu_needs_cpu(int cpu)
|
static inline int rcu_needs_cpu(int cpu)
|
||||||
{
|
{
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -98,7 +94,6 @@ static inline int rcu_needs_cpu(int cpu)
|
||||||
|
|
||||||
#else /* #ifdef CONFIG_TINY_RCU */
|
#else /* #ifdef CONFIG_TINY_RCU */
|
||||||
|
|
||||||
extern void exit_rcu(void);
|
|
||||||
int rcu_preempt_needs_cpu(void);
|
int rcu_preempt_needs_cpu(void);
|
||||||
|
|
||||||
static inline int rcu_needs_cpu(int cpu)
|
static inline int rcu_needs_cpu(int cpu)
|
||||||
|
|
|
@ -45,18 +45,6 @@ static inline void rcu_virt_note_context_switch(int cpu)
|
||||||
rcu_note_context_switch(cpu);
|
rcu_note_context_switch(cpu);
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_TREE_PREEMPT_RCU
|
|
||||||
|
|
||||||
extern void exit_rcu(void);
|
|
||||||
|
|
||||||
#else /* #ifdef CONFIG_TREE_PREEMPT_RCU */
|
|
||||||
|
|
||||||
static inline void exit_rcu(void)
|
|
||||||
{
|
|
||||||
}
|
|
||||||
|
|
||||||
#endif /* #else #ifdef CONFIG_TREE_PREEMPT_RCU */
|
|
||||||
|
|
||||||
extern void synchronize_rcu_bh(void);
|
extern void synchronize_rcu_bh(void);
|
||||||
extern void synchronize_sched_expedited(void);
|
extern void synchronize_sched_expedited(void);
|
||||||
extern void synchronize_rcu_expedited(void);
|
extern void synchronize_rcu_expedited(void);
|
||||||
|
|
|
@ -51,6 +51,34 @@
|
||||||
|
|
||||||
#include "rcu.h"
|
#include "rcu.h"
|
||||||
|
|
||||||
|
#ifdef CONFIG_PREEMPT_RCU
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Check for a task exiting while in a preemptible-RCU read-side
|
||||||
|
* critical section, clean up if so. No need to issue warnings,
|
||||||
|
* as debug_check_no_locks_held() already does this if lockdep
|
||||||
|
* is enabled.
|
||||||
|
*/
|
||||||
|
void exit_rcu(void)
|
||||||
|
{
|
||||||
|
struct task_struct *t = current;
|
||||||
|
|
||||||
|
if (likely(list_empty(¤t->rcu_node_entry)))
|
||||||
|
return;
|
||||||
|
t->rcu_read_lock_nesting = 1;
|
||||||
|
barrier();
|
||||||
|
t->rcu_read_unlock_special = RCU_READ_UNLOCK_BLOCKED;
|
||||||
|
__rcu_read_unlock();
|
||||||
|
}
|
||||||
|
|
||||||
|
#else /* #ifdef CONFIG_PREEMPT_RCU */
|
||||||
|
|
||||||
|
void exit_rcu(void)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
|
#endif /* #else #ifdef CONFIG_PREEMPT_RCU */
|
||||||
|
|
||||||
#ifdef CONFIG_DEBUG_LOCK_ALLOC
|
#ifdef CONFIG_DEBUG_LOCK_ALLOC
|
||||||
static struct lock_class_key rcu_lock_key;
|
static struct lock_class_key rcu_lock_key;
|
||||||
struct lockdep_map rcu_lock_map =
|
struct lockdep_map rcu_lock_map =
|
||||||
|
|
|
@ -851,22 +851,6 @@ int rcu_preempt_needs_cpu(void)
|
||||||
return rcu_preempt_ctrlblk.rcb.rcucblist != NULL;
|
return rcu_preempt_ctrlblk.rcb.rcucblist != NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
|
||||||
* Check for a task exiting while in a preemptible -RCU read-side
|
|
||||||
* critical section, clean up if so. No need to issue warnings,
|
|
||||||
* as debug_check_no_locks_held() already does this if lockdep
|
|
||||||
* is enabled.
|
|
||||||
*/
|
|
||||||
void exit_rcu(void)
|
|
||||||
{
|
|
||||||
struct task_struct *t = current;
|
|
||||||
|
|
||||||
if (t->rcu_read_lock_nesting == 0)
|
|
||||||
return;
|
|
||||||
t->rcu_read_lock_nesting = 1;
|
|
||||||
__rcu_read_unlock();
|
|
||||||
}
|
|
||||||
|
|
||||||
#else /* #ifdef CONFIG_TINY_PREEMPT_RCU */
|
#else /* #ifdef CONFIG_TINY_PREEMPT_RCU */
|
||||||
|
|
||||||
#ifdef CONFIG_RCU_TRACE
|
#ifdef CONFIG_RCU_TRACE
|
||||||
|
|
|
@ -969,22 +969,6 @@ static void __init __rcu_init_preempt(void)
|
||||||
rcu_init_one(&rcu_preempt_state, &rcu_preempt_data);
|
rcu_init_one(&rcu_preempt_state, &rcu_preempt_data);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
|
||||||
* Check for a task exiting while in a preemptible-RCU read-side
|
|
||||||
* critical section, clean up if so. No need to issue warnings,
|
|
||||||
* as debug_check_no_locks_held() already does this if lockdep
|
|
||||||
* is enabled.
|
|
||||||
*/
|
|
||||||
void exit_rcu(void)
|
|
||||||
{
|
|
||||||
struct task_struct *t = current;
|
|
||||||
|
|
||||||
if (t->rcu_read_lock_nesting == 0)
|
|
||||||
return;
|
|
||||||
t->rcu_read_lock_nesting = 1;
|
|
||||||
__rcu_read_unlock();
|
|
||||||
}
|
|
||||||
|
|
||||||
#else /* #ifdef CONFIG_TREE_PREEMPT_RCU */
|
#else /* #ifdef CONFIG_TREE_PREEMPT_RCU */
|
||||||
|
|
||||||
static struct rcu_state *rcu_state = &rcu_sched_state;
|
static struct rcu_state *rcu_state = &rcu_sched_state;
|
||||||
|
|
Loading…
Add table
Reference in a new issue