|
@@ -152,7 +152,7 @@ static void ib_fmr_batch_release(struct ib_fmr_pool *pool)
|
|
|
|
|
|
#ifdef DEBUG
|
|
|
if (fmr->ref_count !=0) {
|
|
|
- printk(KERN_WARNING PFX "Unmapping FMR 0x%08x with ref count %d",
|
|
|
+ printk(KERN_WARNING PFX "Unmapping FMR 0x%08x with ref count %d\n",
|
|
|
fmr, fmr->ref_count);
|
|
|
}
|
|
|
#endif
|
|
@@ -170,7 +170,7 @@ static void ib_fmr_batch_release(struct ib_fmr_pool *pool)
|
|
|
|
|
|
ret = ib_unmap_fmr(&fmr_list);
|
|
|
if (ret)
|
|
|
- printk(KERN_WARNING PFX "ib_unmap_fmr returned %d", ret);
|
|
|
+ printk(KERN_WARNING PFX "ib_unmap_fmr returned %d\n", ret);
|
|
|
|
|
|
spin_lock_irq(&pool->pool_lock);
|
|
|
list_splice(&unmap_list, &pool->free_list);
|
|
@@ -235,13 +235,13 @@ struct ib_fmr_pool *ib_create_fmr_pool(struct ib_pd *pd,
|
|
|
|
|
|
attr = kmalloc(sizeof *attr, GFP_KERNEL);
|
|
|
if (!attr) {
|
|
|
- printk(KERN_WARNING PFX "couldn't allocate device attr struct");
|
|
|
+ printk(KERN_WARNING PFX "couldn't allocate device attr struct\n");
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
}
|
|
|
|
|
|
ret = ib_query_device(device, attr);
|
|
|
if (ret) {
|
|
|
- printk(KERN_WARNING PFX "couldn't query device: %d", ret);
|
|
|
+ printk(KERN_WARNING PFX "couldn't query device: %d\n", ret);
|
|
|
kfree(attr);
|
|
|
return ERR_PTR(ret);
|
|
|
}
|
|
@@ -255,7 +255,7 @@ struct ib_fmr_pool *ib_create_fmr_pool(struct ib_pd *pd,
|
|
|
|
|
|
pool = kmalloc(sizeof *pool, GFP_KERNEL);
|
|
|
if (!pool) {
|
|
|
- printk(KERN_WARNING PFX "couldn't allocate pool struct");
|
|
|
+ printk(KERN_WARNING PFX "couldn't allocate pool struct\n");
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
}
|
|
|
|
|
@@ -272,7 +272,7 @@ struct ib_fmr_pool *ib_create_fmr_pool(struct ib_pd *pd,
|
|
|
kmalloc(IB_FMR_HASH_SIZE * sizeof *pool->cache_bucket,
|
|
|
GFP_KERNEL);
|
|
|
if (!pool->cache_bucket) {
|
|
|
- printk(KERN_WARNING PFX "Failed to allocate cache in pool");
|
|
|
+ printk(KERN_WARNING PFX "Failed to allocate cache in pool\n");
|
|
|
ret = -ENOMEM;
|
|
|
goto out_free_pool;
|
|
|
}
|
|
@@ -296,7 +296,7 @@ struct ib_fmr_pool *ib_create_fmr_pool(struct ib_pd *pd,
|
|
|
"ib_fmr(%s)",
|
|
|
device->name);
|
|
|
if (IS_ERR(pool->thread)) {
|
|
|
- printk(KERN_WARNING PFX "couldn't start cleanup thread");
|
|
|
+ printk(KERN_WARNING PFX "couldn't start cleanup thread\n");
|
|
|
ret = PTR_ERR(pool->thread);
|
|
|
goto out_free_pool;
|
|
|
}
|
|
@@ -314,7 +314,7 @@ struct ib_fmr_pool *ib_create_fmr_pool(struct ib_pd *pd,
|
|
|
GFP_KERNEL);
|
|
|
if (!fmr) {
|
|
|
printk(KERN_WARNING PFX "failed to allocate fmr "
|
|
|
- "struct for FMR %d", i);
|
|
|
+ "struct for FMR %d\n", i);
|
|
|
goto out_fail;
|
|
|
}
|
|
|
|
|
@@ -326,7 +326,7 @@ struct ib_fmr_pool *ib_create_fmr_pool(struct ib_pd *pd,
|
|
|
fmr->fmr = ib_alloc_fmr(pd, params->access, &fmr_attr);
|
|
|
if (IS_ERR(fmr->fmr)) {
|
|
|
printk(KERN_WARNING PFX "fmr_create failed "
|
|
|
- "for FMR %d", i);
|
|
|
+ "for FMR %d\n", i);
|
|
|
kfree(fmr);
|
|
|
goto out_fail;
|
|
|
}
|
|
@@ -381,7 +381,7 @@ void ib_destroy_fmr_pool(struct ib_fmr_pool *pool)
|
|
|
}
|
|
|
|
|
|
if (i < pool->pool_size)
|
|
|
- printk(KERN_WARNING PFX "pool still has %d regions registered",
|
|
|
+ printk(KERN_WARNING PFX "pool still has %d regions registered\n",
|
|
|
pool->pool_size - i);
|
|
|
|
|
|
kfree(pool->cache_bucket);
|
|
@@ -518,7 +518,7 @@ int ib_fmr_pool_unmap(struct ib_pool_fmr *fmr)
|
|
|
|
|
|
#ifdef DEBUG
|
|
|
if (fmr->ref_count < 0)
|
|
|
- printk(KERN_WARNING PFX "FMR %p has ref count %d < 0",
|
|
|
+ printk(KERN_WARNING PFX "FMR %p has ref count %d < 0\n",
|
|
|
fmr, fmr->ref_count);
|
|
|
#endif
|
|
|
|