|
@@ -1350,7 +1350,7 @@ rcu_send_cbs_to_orphanage(int cpu, struct rcu_state *rsp,
|
|
|
rsp->qlen += rdp->qlen;
|
|
|
rdp->n_cbs_orphaned += rdp->qlen;
|
|
|
rdp->qlen_lazy = 0;
|
|
|
- rdp->qlen = 0;
|
|
|
+ ACCESS_ONCE(rdp->qlen) = 0;
|
|
|
}
|
|
|
|
|
|
/*
|
|
@@ -1600,7 +1600,7 @@ static void rcu_do_batch(struct rcu_state *rsp, struct rcu_data *rdp)
|
|
|
}
|
|
|
smp_mb(); /* List handling before counting for rcu_barrier(). */
|
|
|
rdp->qlen_lazy -= count_lazy;
|
|
|
- rdp->qlen -= count;
|
|
|
+ ACCESS_ONCE(rdp->qlen) -= count;
|
|
|
rdp->n_cbs_invoked += count;
|
|
|
|
|
|
/* Reinstate batch limit if we have worked down the excess. */
|
|
@@ -1889,7 +1889,7 @@ __call_rcu(struct rcu_head *head, void (*func)(struct rcu_head *rcu),
|
|
|
rdp = this_cpu_ptr(rsp->rda);
|
|
|
|
|
|
/* Add the callback to our list. */
|
|
|
- rdp->qlen++;
|
|
|
+ ACCESS_ONCE(rdp->qlen)++;
|
|
|
if (lazy)
|
|
|
rdp->qlen_lazy++;
|
|
|
else
|
|
@@ -2423,7 +2423,7 @@ rcu_boot_init_percpu_data(int cpu, struct rcu_state *rsp)
|
|
|
rdp->grpmask = 1UL << (cpu - rdp->mynode->grplo);
|
|
|
init_callback_list(rdp);
|
|
|
rdp->qlen_lazy = 0;
|
|
|
- rdp->qlen = 0;
|
|
|
+ ACCESS_ONCE(rdp->qlen) = 0;
|
|
|
rdp->dynticks = &per_cpu(rcu_dynticks, cpu);
|
|
|
WARN_ON_ONCE(rdp->dynticks->dynticks_nesting != DYNTICK_TASK_EXIT_IDLE);
|
|
|
WARN_ON_ONCE(atomic_read(&rdp->dynticks->dynticks) != 1);
|