|
@@ -961,7 +961,7 @@ static int __bio_copy_iov(struct bio *bio, struct bio_vec *iovecs,
|
|
|
int iov_idx = 0;
|
|
|
unsigned int iov_off = 0;
|
|
|
|
|
|
- __bio_for_each_segment(bvec, bio, i, 0) {
|
|
|
+ bio_for_each_segment_all(bvec, bio, i) {
|
|
|
char *bv_addr = page_address(bvec->bv_page);
|
|
|
unsigned int bv_len = iovecs[i].bv_len;
|
|
|
|
|
@@ -1143,7 +1143,7 @@ struct bio *bio_copy_user_iov(struct request_queue *q,
|
|
|
return bio;
|
|
|
cleanup:
|
|
|
if (!map_data)
|
|
|
- bio_for_each_segment(bvec, bio, i)
|
|
|
+ bio_for_each_segment_all(bvec, bio, i)
|
|
|
__free_page(bvec->bv_page);
|
|
|
|
|
|
bio_put(bio);
|
|
@@ -1357,7 +1357,7 @@ static void __bio_unmap_user(struct bio *bio)
|
|
|
/*
|
|
|
* make sure we dirty pages we wrote to
|
|
|
*/
|
|
|
- __bio_for_each_segment(bvec, bio, i, 0) {
|
|
|
+ bio_for_each_segment_all(bvec, bio, i) {
|
|
|
if (bio_data_dir(bio) == READ)
|
|
|
set_page_dirty_lock(bvec->bv_page);
|
|
|
|
|
@@ -1463,7 +1463,7 @@ static void bio_copy_kern_endio(struct bio *bio, int err)
|
|
|
int i;
|
|
|
char *p = bmd->sgvecs[0].iov_base;
|
|
|
|
|
|
- __bio_for_each_segment(bvec, bio, i, 0) {
|
|
|
+ bio_for_each_segment_all(bvec, bio, i) {
|
|
|
char *addr = page_address(bvec->bv_page);
|
|
|
int len = bmd->iovecs[i].bv_len;
|
|
|
|
|
@@ -1503,7 +1503,7 @@ struct bio *bio_copy_kern(struct request_queue *q, void *data, unsigned int len,
|
|
|
if (!reading) {
|
|
|
void *p = data;
|
|
|
|
|
|
- bio_for_each_segment(bvec, bio, i) {
|
|
|
+ bio_for_each_segment_all(bvec, bio, i) {
|
|
|
char *addr = page_address(bvec->bv_page);
|
|
|
|
|
|
memcpy(addr, p, bvec->bv_len);
|
|
@@ -1789,7 +1789,7 @@ sector_t bio_sector_offset(struct bio *bio, unsigned short index,
|
|
|
if (index >= bio->bi_idx)
|
|
|
index = bio->bi_vcnt - 1;
|
|
|
|
|
|
- __bio_for_each_segment(bv, bio, i, 0) {
|
|
|
+ bio_for_each_segment_all(bv, bio, i) {
|
|
|
if (i == index) {
|
|
|
if (offset > bv->bv_offset)
|
|
|
sectors += (offset - bv->bv_offset) / sector_sz;
|