|
@@ -38,8 +38,6 @@
|
|
#include "ib.h"
|
|
#include "ib.h"
|
|
#include "xlist.h"
|
|
#include "xlist.h"
|
|
|
|
|
|
-static struct workqueue_struct *rds_ib_fmr_wq;
|
|
|
|
-
|
|
|
|
static DEFINE_PER_CPU(unsigned long, clean_list_grace);
|
|
static DEFINE_PER_CPU(unsigned long, clean_list_grace);
|
|
#define CLEAN_LIST_BUSY_BIT 0
|
|
#define CLEAN_LIST_BUSY_BIT 0
|
|
|
|
|
|
@@ -307,7 +305,7 @@ static struct rds_ib_mr *rds_ib_alloc_fmr(struct rds_ib_device *rds_ibdev)
|
|
int err = 0, iter = 0;
|
|
int err = 0, iter = 0;
|
|
|
|
|
|
if (atomic_read(&pool->dirty_count) >= pool->max_items / 10)
|
|
if (atomic_read(&pool->dirty_count) >= pool->max_items / 10)
|
|
- queue_delayed_work(rds_ib_fmr_wq, &pool->flush_worker, 10);
|
|
|
|
|
|
+ schedule_delayed_work(&pool->flush_worker, 10);
|
|
|
|
|
|
while (1) {
|
|
while (1) {
|
|
ibmr = rds_ib_reuse_fmr(pool);
|
|
ibmr = rds_ib_reuse_fmr(pool);
|
|
@@ -696,24 +694,6 @@ out_nolock:
|
|
return ret;
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
|
|
-int rds_ib_fmr_init(void)
|
|
|
|
-{
|
|
|
|
- rds_ib_fmr_wq = create_workqueue("rds_fmr_flushd");
|
|
|
|
- if (!rds_ib_fmr_wq)
|
|
|
|
- return -ENOMEM;
|
|
|
|
- return 0;
|
|
|
|
-}
|
|
|
|
-
|
|
|
|
-/*
|
|
|
|
- * By the time this is called all the IB devices should have been torn down and
|
|
|
|
- * had their pools freed. As each pool is freed its work struct is waited on,
|
|
|
|
- * so the pool flushing work queue should be idle by the time we get here.
|
|
|
|
- */
|
|
|
|
-void rds_ib_fmr_exit(void)
|
|
|
|
-{
|
|
|
|
- destroy_workqueue(rds_ib_fmr_wq);
|
|
|
|
-}
|
|
|
|
-
|
|
|
|
static void rds_ib_mr_pool_flush_worker(struct work_struct *work)
|
|
static void rds_ib_mr_pool_flush_worker(struct work_struct *work)
|
|
{
|
|
{
|
|
struct rds_ib_mr_pool *pool = container_of(work, struct rds_ib_mr_pool, flush_worker.work);
|
|
struct rds_ib_mr_pool *pool = container_of(work, struct rds_ib_mr_pool, flush_worker.work);
|
|
@@ -741,7 +721,7 @@ void rds_ib_free_mr(void *trans_private, int invalidate)
|
|
/* If we've pinned too many pages, request a flush */
|
|
/* If we've pinned too many pages, request a flush */
|
|
if (atomic_read(&pool->free_pinned) >= pool->max_free_pinned ||
|
|
if (atomic_read(&pool->free_pinned) >= pool->max_free_pinned ||
|
|
atomic_read(&pool->dirty_count) >= pool->max_items / 10)
|
|
atomic_read(&pool->dirty_count) >= pool->max_items / 10)
|
|
- queue_delayed_work(rds_ib_fmr_wq, &pool->flush_worker, 10);
|
|
|
|
|
|
+ schedule_delayed_work(&pool->flush_worker, 10);
|
|
|
|
|
|
if (invalidate) {
|
|
if (invalidate) {
|
|
if (likely(!in_interrupt())) {
|
|
if (likely(!in_interrupt())) {
|
|
@@ -749,8 +729,7 @@ void rds_ib_free_mr(void *trans_private, int invalidate)
|
|
} else {
|
|
} else {
|
|
/* We get here if the user created a MR marked
|
|
/* We get here if the user created a MR marked
|
|
* as use_once and invalidate at the same time. */
|
|
* as use_once and invalidate at the same time. */
|
|
- queue_delayed_work(rds_ib_fmr_wq,
|
|
|
|
- &pool->flush_worker, 10);
|
|
|
|
|
|
+ schedule_delayed_work(&pool->flush_worker, 10);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|