[PATCH v2 15/45] rcu: Use get/put_online_cpus_atomic() to prevent CPU offline
Paul E. McKenney
paulmck at linux.vnet.ibm.com
Wed Jun 26 08:00:26 EST 2013
On Wed, Jun 26, 2013 at 01:57:55AM +0530, Srivatsa S. Bhat wrote:
> Once stop_machine() is gone from the CPU offline path, we won't be able
> to depend on disabling preemption to prevent CPUs from going offline
> from under us.
>
> In RCU code, rcu_implicit_dynticks_qs() checks if a CPU is offline,
> while being protected by a spinlock. Use the get/put_online_cpus_atomic()
> APIs to prevent CPUs from going offline, while invoking from atomic context.
I am not completely sure that this is needed. Here is a (quite possibly
flawed) argument for its not being needed:
o rcu_gp_init() holds off CPU-hotplug operations during
grace-period initialization. Therefore, RCU will avoid
looking for quiescent states from CPUs that were offline
(and thus in an extended quiescent state) at the beginning
of the grace period.
o If force_qs_rnp() is looking for a quiescent state from
a given CPU, and if it senses that CPU as being offline,
then even without synchronization we know that the CPU
was offline some time during the current grace period.
After all, it was online at the beginning of the grace
period (otherwise, we would not be looking at it at all),
and our later sampling of its state must have therefore
happened after the start of the grace period. Given that
the grace period has not yet ended, it also has to happened
before the end of the grace period.
o Therefore, we should be able to sample the offline state
without synchronization.
Possible flaws in this argument: memory ordering, oddnesses in
the sampling and updates of the cpumask recording which CPUs are
online, and so on.
Thoughts?
Thanx, Paul
> Cc: Dipankar Sarma <dipankar at in.ibm.com>
> Cc: "Paul E. McKenney" <paulmck at linux.vnet.ibm.com>
> Signed-off-by: Srivatsa S. Bhat <srivatsa.bhat at linux.vnet.ibm.com>
> ---
>
> kernel/rcutree.c | 4 ++++
> 1 file changed, 4 insertions(+)
>
> diff --git a/kernel/rcutree.c b/kernel/rcutree.c
> index cf3adc6..caeed1a 100644
> --- a/kernel/rcutree.c
> +++ b/kernel/rcutree.c
> @@ -2107,6 +2107,8 @@ static void force_qs_rnp(struct rcu_state *rsp, int (*f)(struct rcu_data *))
> rcu_initiate_boost(rnp, flags); /* releases rnp->lock */
> continue;
> }
> +
> + get_online_cpus_atomic();
> cpu = rnp->grplo;
> bit = 1;
> for (; cpu <= rnp->grphi; cpu++, bit <<= 1) {
> @@ -2114,6 +2116,8 @@ static void force_qs_rnp(struct rcu_state *rsp, int (*f)(struct rcu_data *))
> f(per_cpu_ptr(rsp->rda, cpu)))
> mask |= bit;
> }
> + put_online_cpus_atomic();
> +
> if (mask != 0) {
>
> /* rcu_report_qs_rnp() releases rnp->lock. */
>
More information about the Linuxppc-dev
mailing list