|
@@ -65,6 +65,8 @@ static int irqreader = 1; /* RCU readers from irq (timers). */
|
|
|
static int fqs_duration; /* Duration of bursts (us), 0 to disable. */
|
|
|
static int fqs_holdoff; /* Hold time within burst (us). */
|
|
|
static int fqs_stutter = 3; /* Wait time between bursts (s). */
|
|
|
+static bool gp_exp; /* Use expedited GP wait primitives. */
|
|
|
+static bool gp_normal; /* Use normal GP wait primitives. */
|
|
|
static int n_barrier_cbs; /* Number of callbacks to test RCU barriers. */
|
|
|
static int object_debug; /* Test object-debug double call_rcu()?. */
|
|
|
static int onoff_interval; /* Wait time between CPU hotplugs, 0=disable. */
|
|
@@ -99,6 +101,10 @@ module_param(fqs_holdoff, int, 0444);
|
|
|
MODULE_PARM_DESC(fqs_holdoff, "Holdoff time within fqs bursts (us)");
|
|
|
module_param(fqs_stutter, int, 0444);
|
|
|
MODULE_PARM_DESC(fqs_stutter, "Wait time between fqs bursts (s)");
|
|
|
+module_param(gp_normal, bool, 0444);
|
|
|
+MODULE_PARM_DESC(gp_normal, "Use normal (non-expedited) GP wait primitives");
|
|
|
+module_param(gp_exp, bool, 0444);
|
|
|
+MODULE_PARM_DESC(gp_exp, "Use expedited GP wait primitives");
|
|
|
module_param(n_barrier_cbs, int, 0444);
|
|
|
MODULE_PARM_DESC(n_barrier_cbs, "# of callbacks/kthreads for barrier testing");
|
|
|
module_param(object_debug, int, 0444);
|
|
@@ -363,6 +369,7 @@ struct rcu_torture_ops {
|
|
|
int (*completed)(void);
|
|
|
void (*deferred_free)(struct rcu_torture *p);
|
|
|
void (*sync)(void);
|
|
|
+ void (*exp_sync)(void);
|
|
|
void (*call)(struct rcu_head *head, void (*func)(struct rcu_head *rcu));
|
|
|
void (*cb_barrier)(void);
|
|
|
void (*fqs)(void);
|
|
@@ -446,81 +453,27 @@ static void rcu_torture_deferred_free(struct rcu_torture *p)
|
|
|
call_rcu(&p->rtort_rcu, rcu_torture_cb);
|
|
|
}
|
|
|
|
|
|
-static struct rcu_torture_ops rcu_ops = {
|
|
|
- .init = NULL,
|
|
|
- .readlock = rcu_torture_read_lock,
|
|
|
- .read_delay = rcu_read_delay,
|
|
|
- .readunlock = rcu_torture_read_unlock,
|
|
|
- .completed = rcu_torture_completed,
|
|
|
- .deferred_free = rcu_torture_deferred_free,
|
|
|
- .sync = synchronize_rcu,
|
|
|
- .call = call_rcu,
|
|
|
- .cb_barrier = rcu_barrier,
|
|
|
- .fqs = rcu_force_quiescent_state,
|
|
|
- .stats = NULL,
|
|
|
- .irq_capable = 1,
|
|
|
- .can_boost = rcu_can_boost(),
|
|
|
- .name = "rcu"
|
|
|
-};
|
|
|
-
|
|
|
-static void rcu_sync_torture_deferred_free(struct rcu_torture *p)
|
|
|
-{
|
|
|
- int i;
|
|
|
- struct rcu_torture *rp;
|
|
|
- struct rcu_torture *rp1;
|
|
|
-
|
|
|
- cur_ops->sync();
|
|
|
- list_add(&p->rtort_free, &rcu_torture_removed);
|
|
|
- list_for_each_entry_safe(rp, rp1, &rcu_torture_removed, rtort_free) {
|
|
|
- i = rp->rtort_pipe_count;
|
|
|
- if (i > RCU_TORTURE_PIPE_LEN)
|
|
|
- i = RCU_TORTURE_PIPE_LEN;
|
|
|
- atomic_inc(&rcu_torture_wcount[i]);
|
|
|
- if (++rp->rtort_pipe_count >= RCU_TORTURE_PIPE_LEN) {
|
|
|
- rp->rtort_mbtest = 0;
|
|
|
- list_del(&rp->rtort_free);
|
|
|
- rcu_torture_free(rp);
|
|
|
- }
|
|
|
- }
|
|
|
-}
|
|
|
-
|
|
|
static void rcu_sync_torture_init(void)
|
|
|
{
|
|
|
INIT_LIST_HEAD(&rcu_torture_removed);
|
|
|
}
|
|
|
|
|
|
-static struct rcu_torture_ops rcu_sync_ops = {
|
|
|
+static struct rcu_torture_ops rcu_ops = {
|
|
|
.init = rcu_sync_torture_init,
|
|
|
.readlock = rcu_torture_read_lock,
|
|
|
.read_delay = rcu_read_delay,
|
|
|
.readunlock = rcu_torture_read_unlock,
|
|
|
.completed = rcu_torture_completed,
|
|
|
- .deferred_free = rcu_sync_torture_deferred_free,
|
|
|
+ .deferred_free = rcu_torture_deferred_free,
|
|
|
.sync = synchronize_rcu,
|
|
|
- .call = NULL,
|
|
|
- .cb_barrier = NULL,
|
|
|
- .fqs = rcu_force_quiescent_state,
|
|
|
- .stats = NULL,
|
|
|
- .irq_capable = 1,
|
|
|
- .can_boost = rcu_can_boost(),
|
|
|
- .name = "rcu_sync"
|
|
|
-};
|
|
|
-
|
|
|
-static struct rcu_torture_ops rcu_expedited_ops = {
|
|
|
- .init = rcu_sync_torture_init,
|
|
|
- .readlock = rcu_torture_read_lock,
|
|
|
- .read_delay = rcu_read_delay, /* just reuse rcu's version. */
|
|
|
- .readunlock = rcu_torture_read_unlock,
|
|
|
- .completed = rcu_no_completed,
|
|
|
- .deferred_free = rcu_sync_torture_deferred_free,
|
|
|
- .sync = synchronize_rcu_expedited,
|
|
|
- .call = NULL,
|
|
|
- .cb_barrier = NULL,
|
|
|
+ .exp_sync = synchronize_rcu_expedited,
|
|
|
+ .call = call_rcu,
|
|
|
+ .cb_barrier = rcu_barrier,
|
|
|
.fqs = rcu_force_quiescent_state,
|
|
|
.stats = NULL,
|
|
|
.irq_capable = 1,
|
|
|
.can_boost = rcu_can_boost(),
|
|
|
- .name = "rcu_expedited"
|
|
|
+ .name = "rcu"
|
|
|
};
|
|
|
|
|
|
/*
|
|
@@ -549,13 +502,14 @@ static void rcu_bh_torture_deferred_free(struct rcu_torture *p)
|
|
|
}
|
|
|
|
|
|
static struct rcu_torture_ops rcu_bh_ops = {
|
|
|
- .init = NULL,
|
|
|
+ .init = rcu_sync_torture_init,
|
|
|
.readlock = rcu_bh_torture_read_lock,
|
|
|
.read_delay = rcu_read_delay, /* just reuse rcu's version. */
|
|
|
.readunlock = rcu_bh_torture_read_unlock,
|
|
|
.completed = rcu_bh_torture_completed,
|
|
|
.deferred_free = rcu_bh_torture_deferred_free,
|
|
|
.sync = synchronize_rcu_bh,
|
|
|
+ .exp_sync = synchronize_rcu_bh_expedited,
|
|
|
.call = call_rcu_bh,
|
|
|
.cb_barrier = rcu_barrier_bh,
|
|
|
.fqs = rcu_bh_force_quiescent_state,
|
|
@@ -564,38 +518,6 @@ static struct rcu_torture_ops rcu_bh_ops = {
|
|
|
.name = "rcu_bh"
|
|
|
};
|
|
|
|
|
|
-static struct rcu_torture_ops rcu_bh_sync_ops = {
|
|
|
- .init = rcu_sync_torture_init,
|
|
|
- .readlock = rcu_bh_torture_read_lock,
|
|
|
- .read_delay = rcu_read_delay, /* just reuse rcu's version. */
|
|
|
- .readunlock = rcu_bh_torture_read_unlock,
|
|
|
- .completed = rcu_bh_torture_completed,
|
|
|
- .deferred_free = rcu_sync_torture_deferred_free,
|
|
|
- .sync = synchronize_rcu_bh,
|
|
|
- .call = NULL,
|
|
|
- .cb_barrier = NULL,
|
|
|
- .fqs = rcu_bh_force_quiescent_state,
|
|
|
- .stats = NULL,
|
|
|
- .irq_capable = 1,
|
|
|
- .name = "rcu_bh_sync"
|
|
|
-};
|
|
|
-
|
|
|
-static struct rcu_torture_ops rcu_bh_expedited_ops = {
|
|
|
- .init = rcu_sync_torture_init,
|
|
|
- .readlock = rcu_bh_torture_read_lock,
|
|
|
- .read_delay = rcu_read_delay, /* just reuse rcu's version. */
|
|
|
- .readunlock = rcu_bh_torture_read_unlock,
|
|
|
- .completed = rcu_bh_torture_completed,
|
|
|
- .deferred_free = rcu_sync_torture_deferred_free,
|
|
|
- .sync = synchronize_rcu_bh_expedited,
|
|
|
- .call = NULL,
|
|
|
- .cb_barrier = NULL,
|
|
|
- .fqs = rcu_bh_force_quiescent_state,
|
|
|
- .stats = NULL,
|
|
|
- .irq_capable = 1,
|
|
|
- .name = "rcu_bh_expedited"
|
|
|
-};
|
|
|
-
|
|
|
/*
|
|
|
* Definitions for srcu torture testing.
|
|
|
*/
|
|
@@ -670,6 +592,11 @@ static int srcu_torture_stats(char *page)
|
|
|
return cnt;
|
|
|
}
|
|
|
|
|
|
+static void srcu_torture_synchronize_expedited(void)
|
|
|
+{
|
|
|
+ synchronize_srcu_expedited(&srcu_ctl);
|
|
|
+}
|
|
|
+
|
|
|
static struct rcu_torture_ops srcu_ops = {
|
|
|
.init = rcu_sync_torture_init,
|
|
|
.readlock = srcu_torture_read_lock,
|
|
@@ -678,45 +605,13 @@ static struct rcu_torture_ops srcu_ops = {
|
|
|
.completed = srcu_torture_completed,
|
|
|
.deferred_free = srcu_torture_deferred_free,
|
|
|
.sync = srcu_torture_synchronize,
|
|
|
+ .exp_sync = srcu_torture_synchronize_expedited,
|
|
|
.call = srcu_torture_call,
|
|
|
.cb_barrier = srcu_torture_barrier,
|
|
|
.stats = srcu_torture_stats,
|
|
|
.name = "srcu"
|
|
|
};
|
|
|
|
|
|
-static struct rcu_torture_ops srcu_sync_ops = {
|
|
|
- .init = rcu_sync_torture_init,
|
|
|
- .readlock = srcu_torture_read_lock,
|
|
|
- .read_delay = srcu_read_delay,
|
|
|
- .readunlock = srcu_torture_read_unlock,
|
|
|
- .completed = srcu_torture_completed,
|
|
|
- .deferred_free = rcu_sync_torture_deferred_free,
|
|
|
- .sync = srcu_torture_synchronize,
|
|
|
- .call = NULL,
|
|
|
- .cb_barrier = NULL,
|
|
|
- .stats = srcu_torture_stats,
|
|
|
- .name = "srcu_sync"
|
|
|
-};
|
|
|
-
|
|
|
-static void srcu_torture_synchronize_expedited(void)
|
|
|
-{
|
|
|
- synchronize_srcu_expedited(&srcu_ctl);
|
|
|
-}
|
|
|
-
|
|
|
-static struct rcu_torture_ops srcu_expedited_ops = {
|
|
|
- .init = rcu_sync_torture_init,
|
|
|
- .readlock = srcu_torture_read_lock,
|
|
|
- .read_delay = srcu_read_delay,
|
|
|
- .readunlock = srcu_torture_read_unlock,
|
|
|
- .completed = srcu_torture_completed,
|
|
|
- .deferred_free = rcu_sync_torture_deferred_free,
|
|
|
- .sync = srcu_torture_synchronize_expedited,
|
|
|
- .call = NULL,
|
|
|
- .cb_barrier = NULL,
|
|
|
- .stats = srcu_torture_stats,
|
|
|
- .name = "srcu_expedited"
|
|
|
-};
|
|
|
-
|
|
|
/*
|
|
|
* Definitions for sched torture testing.
|
|
|
*/
|
|
@@ -745,6 +640,8 @@ static struct rcu_torture_ops sched_ops = {
|
|
|
.completed = rcu_no_completed,
|
|
|
.deferred_free = rcu_sched_torture_deferred_free,
|
|
|
.sync = synchronize_sched,
|
|
|
+ .exp_sync = synchronize_sched_expedited,
|
|
|
+ .call = call_rcu_sched,
|
|
|
.cb_barrier = rcu_barrier_sched,
|
|
|
.fqs = rcu_sched_force_quiescent_state,
|
|
|
.stats = NULL,
|
|
@@ -752,35 +649,6 @@ static struct rcu_torture_ops sched_ops = {
|
|
|
.name = "sched"
|
|
|
};
|
|
|
|
|
|
-static struct rcu_torture_ops sched_sync_ops = {
|
|
|
- .init = rcu_sync_torture_init,
|
|
|
- .readlock = sched_torture_read_lock,
|
|
|
- .read_delay = rcu_read_delay, /* just reuse rcu's version. */
|
|
|
- .readunlock = sched_torture_read_unlock,
|
|
|
- .completed = rcu_no_completed,
|
|
|
- .deferred_free = rcu_sync_torture_deferred_free,
|
|
|
- .sync = synchronize_sched,
|
|
|
- .cb_barrier = NULL,
|
|
|
- .fqs = rcu_sched_force_quiescent_state,
|
|
|
- .stats = NULL,
|
|
|
- .name = "sched_sync"
|
|
|
-};
|
|
|
-
|
|
|
-static struct rcu_torture_ops sched_expedited_ops = {
|
|
|
- .init = rcu_sync_torture_init,
|
|
|
- .readlock = sched_torture_read_lock,
|
|
|
- .read_delay = rcu_read_delay, /* just reuse rcu's version. */
|
|
|
- .readunlock = sched_torture_read_unlock,
|
|
|
- .completed = rcu_no_completed,
|
|
|
- .deferred_free = rcu_sync_torture_deferred_free,
|
|
|
- .sync = synchronize_sched_expedited,
|
|
|
- .cb_barrier = NULL,
|
|
|
- .fqs = rcu_sched_force_quiescent_state,
|
|
|
- .stats = NULL,
|
|
|
- .irq_capable = 1,
|
|
|
- .name = "sched_expedited"
|
|
|
-};
|
|
|
-
|
|
|
/*
|
|
|
* RCU torture priority-boost testing. Runs one real-time thread per
|
|
|
* CPU for moderate bursts, repeatedly registering RCU callbacks and
|
|
@@ -930,9 +798,11 @@ rcu_torture_fqs(void *arg)
|
|
|
static int
|
|
|
rcu_torture_writer(void *arg)
|
|
|
{
|
|
|
+ bool exp;
|
|
|
int i;
|
|
|
long oldbatch = rcu_batches_completed();
|
|
|
struct rcu_torture *rp;
|
|
|
+ struct rcu_torture *rp1;
|
|
|
struct rcu_torture *old_rp;
|
|
|
static DEFINE_RCU_RANDOM(rand);
|
|
|
|
|
@@ -957,7 +827,31 @@ rcu_torture_writer(void *arg)
|
|
|
i = RCU_TORTURE_PIPE_LEN;
|
|
|
atomic_inc(&rcu_torture_wcount[i]);
|
|
|
old_rp->rtort_pipe_count++;
|
|
|
- cur_ops->deferred_free(old_rp);
|
|
|
+ if (gp_normal == gp_exp)
|
|
|
+ exp = !!(rcu_random(&rand) & 0x80);
|
|
|
+ else
|
|
|
+ exp = gp_exp;
|
|
|
+ if (!exp) {
|
|
|
+ cur_ops->deferred_free(old_rp);
|
|
|
+ } else {
|
|
|
+ cur_ops->exp_sync();
|
|
|
+ list_add(&old_rp->rtort_free,
|
|
|
+ &rcu_torture_removed);
|
|
|
+ list_for_each_entry_safe(rp, rp1,
|
|
|
+ &rcu_torture_removed,
|
|
|
+ rtort_free) {
|
|
|
+ i = rp->rtort_pipe_count;
|
|
|
+ if (i > RCU_TORTURE_PIPE_LEN)
|
|
|
+ i = RCU_TORTURE_PIPE_LEN;
|
|
|
+ atomic_inc(&rcu_torture_wcount[i]);
|
|
|
+ if (++rp->rtort_pipe_count >=
|
|
|
+ RCU_TORTURE_PIPE_LEN) {
|
|
|
+ rp->rtort_mbtest = 0;
|
|
|
+ list_del(&rp->rtort_free);
|
|
|
+ rcu_torture_free(rp);
|
|
|
+ }
|
|
|
+ }
|
|
|
+ }
|
|
|
}
|
|
|
rcutorture_record_progress(++rcu_torture_current_version);
|
|
|
oldbatch = cur_ops->completed();
|
|
@@ -986,10 +880,18 @@ rcu_torture_fakewriter(void *arg)
|
|
|
schedule_timeout_uninterruptible(1 + rcu_random(&rand)%10);
|
|
|
udelay(rcu_random(&rand) & 0x3ff);
|
|
|
if (cur_ops->cb_barrier != NULL &&
|
|
|
- rcu_random(&rand) % (nfakewriters * 8) == 0)
|
|
|
+ rcu_random(&rand) % (nfakewriters * 8) == 0) {
|
|
|
cur_ops->cb_barrier();
|
|
|
- else
|
|
|
+ } else if (gp_normal == gp_exp) {
|
|
|
+ if (rcu_random(&rand) & 0x80)
|
|
|
+ cur_ops->sync();
|
|
|
+ else
|
|
|
+ cur_ops->exp_sync();
|
|
|
+ } else if (gp_normal) {
|
|
|
cur_ops->sync();
|
|
|
+ } else {
|
|
|
+ cur_ops->exp_sync();
|
|
|
+ }
|
|
|
rcu_stutter_wait("rcu_torture_fakewriter");
|
|
|
} while (!kthread_should_stop() && fullstop == FULLSTOP_DONTSTOP);
|
|
|
|
|
@@ -2000,11 +1902,9 @@ rcu_torture_init(void)
|
|
|
int cpu;
|
|
|
int firsterr = 0;
|
|
|
int retval;
|
|
|
- static struct rcu_torture_ops *torture_ops[] =
|
|
|
- { &rcu_ops, &rcu_sync_ops, &rcu_expedited_ops,
|
|
|
- &rcu_bh_ops, &rcu_bh_sync_ops, &rcu_bh_expedited_ops,
|
|
|
- &srcu_ops, &srcu_sync_ops, &srcu_expedited_ops,
|
|
|
- &sched_ops, &sched_sync_ops, &sched_expedited_ops, };
|
|
|
+ static struct rcu_torture_ops *torture_ops[] = {
|
|
|
+ &rcu_ops, &rcu_bh_ops, &srcu_ops, &sched_ops,
|
|
|
+ };
|
|
|
|
|
|
mutex_lock(&fullstop_mutex);
|
|
|
|