|
@@ -65,9 +65,6 @@ static DEFINE_PER_CPU(unsigned long, cfq_ioc_count);
|
|
static struct completion *ioc_gone;
|
|
static struct completion *ioc_gone;
|
|
static DEFINE_SPINLOCK(ioc_gone_lock);
|
|
static DEFINE_SPINLOCK(ioc_gone_lock);
|
|
|
|
|
|
-static DEFINE_SPINLOCK(cic_index_lock);
|
|
|
|
-static DEFINE_IDA(cic_index_ida);
|
|
|
|
-
|
|
|
|
#define CFQ_PRIO_LISTS IOPRIO_BE_NR
|
|
#define CFQ_PRIO_LISTS IOPRIO_BE_NR
|
|
#define cfq_class_idle(cfqq) ((cfqq)->ioprio_class == IOPRIO_CLASS_IDLE)
|
|
#define cfq_class_idle(cfqq) ((cfqq)->ioprio_class == IOPRIO_CLASS_IDLE)
|
|
#define cfq_class_rt(cfqq) ((cfqq)->ioprio_class == IOPRIO_CLASS_RT)
|
|
#define cfq_class_rt(cfqq) ((cfqq)->ioprio_class == IOPRIO_CLASS_RT)
|
|
@@ -290,7 +287,6 @@ struct cfq_data {
|
|
unsigned int cfq_group_idle;
|
|
unsigned int cfq_group_idle;
|
|
unsigned int cfq_latency;
|
|
unsigned int cfq_latency;
|
|
|
|
|
|
- unsigned int cic_index;
|
|
|
|
struct list_head cic_list;
|
|
struct list_head cic_list;
|
|
|
|
|
|
/*
|
|
/*
|
|
@@ -484,7 +480,7 @@ static inline void cic_set_cfqq(struct cfq_io_context *cic,
|
|
|
|
|
|
static inline void *cfqd_dead_key(struct cfq_data *cfqd)
|
|
static inline void *cfqd_dead_key(struct cfq_data *cfqd)
|
|
{
|
|
{
|
|
- return (void *)(cfqd->cic_index << CIC_DEAD_INDEX_SHIFT | CIC_DEAD_KEY);
|
|
|
|
|
|
+ return (void *)(cfqd->queue->id << CIC_DEAD_INDEX_SHIFT | CIC_DEAD_KEY);
|
|
}
|
|
}
|
|
|
|
|
|
static inline struct cfq_data *cic_to_cfqd(struct cfq_io_context *cic)
|
|
static inline struct cfq_data *cic_to_cfqd(struct cfq_io_context *cic)
|
|
@@ -3105,7 +3101,7 @@ cfq_drop_dead_cic(struct cfq_data *cfqd, struct io_context *ioc,
|
|
BUG_ON(rcu_dereference_check(ioc->ioc_data,
|
|
BUG_ON(rcu_dereference_check(ioc->ioc_data,
|
|
lockdep_is_held(&ioc->lock)) == cic);
|
|
lockdep_is_held(&ioc->lock)) == cic);
|
|
|
|
|
|
- radix_tree_delete(&ioc->radix_root, cfqd->cic_index);
|
|
|
|
|
|
+ radix_tree_delete(&ioc->radix_root, cfqd->queue->id);
|
|
hlist_del_rcu(&cic->cic_list);
|
|
hlist_del_rcu(&cic->cic_list);
|
|
spin_unlock_irqrestore(&ioc->lock, flags);
|
|
spin_unlock_irqrestore(&ioc->lock, flags);
|
|
|
|
|
|
@@ -3133,7 +3129,7 @@ cfq_cic_lookup(struct cfq_data *cfqd, struct io_context *ioc)
|
|
}
|
|
}
|
|
|
|
|
|
do {
|
|
do {
|
|
- cic = radix_tree_lookup(&ioc->radix_root, cfqd->cic_index);
|
|
|
|
|
|
+ cic = radix_tree_lookup(&ioc->radix_root, cfqd->queue->id);
|
|
rcu_read_unlock();
|
|
rcu_read_unlock();
|
|
if (!cic)
|
|
if (!cic)
|
|
break;
|
|
break;
|
|
@@ -3169,8 +3165,7 @@ static int cfq_cic_link(struct cfq_data *cfqd, struct io_context *ioc,
|
|
cic->key = cfqd;
|
|
cic->key = cfqd;
|
|
|
|
|
|
spin_lock_irqsave(&ioc->lock, flags);
|
|
spin_lock_irqsave(&ioc->lock, flags);
|
|
- ret = radix_tree_insert(&ioc->radix_root,
|
|
|
|
- cfqd->cic_index, cic);
|
|
|
|
|
|
+ ret = radix_tree_insert(&ioc->radix_root, cfqd->queue->id, cic);
|
|
if (!ret)
|
|
if (!ret)
|
|
hlist_add_head_rcu(&cic->cic_list, &ioc->cic_list);
|
|
hlist_add_head_rcu(&cic->cic_list, &ioc->cic_list);
|
|
spin_unlock_irqrestore(&ioc->lock, flags);
|
|
spin_unlock_irqrestore(&ioc->lock, flags);
|
|
@@ -3944,10 +3939,6 @@ static void cfq_exit_queue(struct elevator_queue *e)
|
|
|
|
|
|
cfq_shutdown_timer_wq(cfqd);
|
|
cfq_shutdown_timer_wq(cfqd);
|
|
|
|
|
|
- spin_lock(&cic_index_lock);
|
|
|
|
- ida_remove(&cic_index_ida, cfqd->cic_index);
|
|
|
|
- spin_unlock(&cic_index_lock);
|
|
|
|
-
|
|
|
|
/*
|
|
/*
|
|
* Wait for cfqg->blkg->key accessors to exit their grace periods.
|
|
* Wait for cfqg->blkg->key accessors to exit their grace periods.
|
|
* Do this wait only if there are other unlinked groups out
|
|
* Do this wait only if there are other unlinked groups out
|
|
@@ -3969,24 +3960,6 @@ static void cfq_exit_queue(struct elevator_queue *e)
|
|
kfree(cfqd);
|
|
kfree(cfqd);
|
|
}
|
|
}
|
|
|
|
|
|
-static int cfq_alloc_cic_index(void)
|
|
|
|
-{
|
|
|
|
- int index, error;
|
|
|
|
-
|
|
|
|
- do {
|
|
|
|
- if (!ida_pre_get(&cic_index_ida, GFP_KERNEL))
|
|
|
|
- return -ENOMEM;
|
|
|
|
-
|
|
|
|
- spin_lock(&cic_index_lock);
|
|
|
|
- error = ida_get_new(&cic_index_ida, &index);
|
|
|
|
- spin_unlock(&cic_index_lock);
|
|
|
|
- if (error && error != -EAGAIN)
|
|
|
|
- return error;
|
|
|
|
- } while (error);
|
|
|
|
-
|
|
|
|
- return index;
|
|
|
|
-}
|
|
|
|
-
|
|
|
|
static void *cfq_init_queue(struct request_queue *q)
|
|
static void *cfq_init_queue(struct request_queue *q)
|
|
{
|
|
{
|
|
struct cfq_data *cfqd;
|
|
struct cfq_data *cfqd;
|
|
@@ -3994,23 +3967,9 @@ static void *cfq_init_queue(struct request_queue *q)
|
|
struct cfq_group *cfqg;
|
|
struct cfq_group *cfqg;
|
|
struct cfq_rb_root *st;
|
|
struct cfq_rb_root *st;
|
|
|
|
|
|
- i = cfq_alloc_cic_index();
|
|
|
|
- if (i < 0)
|
|
|
|
- return NULL;
|
|
|
|
-
|
|
|
|
cfqd = kmalloc_node(sizeof(*cfqd), GFP_KERNEL | __GFP_ZERO, q->node);
|
|
cfqd = kmalloc_node(sizeof(*cfqd), GFP_KERNEL | __GFP_ZERO, q->node);
|
|
- if (!cfqd) {
|
|
|
|
- spin_lock(&cic_index_lock);
|
|
|
|
- ida_remove(&cic_index_ida, i);
|
|
|
|
- spin_unlock(&cic_index_lock);
|
|
|
|
|
|
+ if (!cfqd)
|
|
return NULL;
|
|
return NULL;
|
|
- }
|
|
|
|
-
|
|
|
|
- /*
|
|
|
|
- * Don't need take queue_lock in the routine, since we are
|
|
|
|
- * initializing the ioscheduler, and nobody is using cfqd
|
|
|
|
- */
|
|
|
|
- cfqd->cic_index = i;
|
|
|
|
|
|
|
|
/* Init root service tree */
|
|
/* Init root service tree */
|
|
cfqd->grp_service_tree = CFQ_RB_ROOT;
|
|
cfqd->grp_service_tree = CFQ_RB_ROOT;
|
|
@@ -4294,7 +4253,6 @@ static void __exit cfq_exit(void)
|
|
*/
|
|
*/
|
|
if (elv_ioc_count_read(cfq_ioc_count))
|
|
if (elv_ioc_count_read(cfq_ioc_count))
|
|
wait_for_completion(&all_gone);
|
|
wait_for_completion(&all_gone);
|
|
- ida_destroy(&cic_index_ida);
|
|
|
|
cfq_slab_kill();
|
|
cfq_slab_kill();
|
|
}
|
|
}
|
|
|
|
|