|
@@ -834,6 +834,12 @@ struct bio *bio_copy_user_iov(struct request_queue *q,
|
|
end = (uaddr + iov[i].iov_len + PAGE_SIZE - 1) >> PAGE_SHIFT;
|
|
end = (uaddr + iov[i].iov_len + PAGE_SIZE - 1) >> PAGE_SHIFT;
|
|
start = uaddr >> PAGE_SHIFT;
|
|
start = uaddr >> PAGE_SHIFT;
|
|
|
|
|
|
|
|
+ /*
|
|
|
|
+ * Overflow, abort
|
|
|
|
+ */
|
|
|
|
+ if (end < start)
|
|
|
|
+ return ERR_PTR(-EINVAL);
|
|
|
|
+
|
|
nr_pages += end - start;
|
|
nr_pages += end - start;
|
|
len += iov[i].iov_len;
|
|
len += iov[i].iov_len;
|
|
}
|
|
}
|
|
@@ -962,6 +968,12 @@ static struct bio *__bio_map_user_iov(struct request_queue *q,
|
|
unsigned long end = (uaddr + len + PAGE_SIZE - 1) >> PAGE_SHIFT;
|
|
unsigned long end = (uaddr + len + PAGE_SIZE - 1) >> PAGE_SHIFT;
|
|
unsigned long start = uaddr >> PAGE_SHIFT;
|
|
unsigned long start = uaddr >> PAGE_SHIFT;
|
|
|
|
|
|
|
|
+ /*
|
|
|
|
+ * Overflow, abort
|
|
|
|
+ */
|
|
|
|
+ if (end < start)
|
|
|
|
+ return ERR_PTR(-EINVAL);
|
|
|
|
+
|
|
nr_pages += end - start;
|
|
nr_pages += end - start;
|
|
/*
|
|
/*
|
|
* buffer must be aligned to at least hardsector size for now
|
|
* buffer must be aligned to at least hardsector size for now
|
|
@@ -989,7 +1001,7 @@ static struct bio *__bio_map_user_iov(struct request_queue *q,
|
|
unsigned long start = uaddr >> PAGE_SHIFT;
|
|
unsigned long start = uaddr >> PAGE_SHIFT;
|
|
const int local_nr_pages = end - start;
|
|
const int local_nr_pages = end - start;
|
|
const int page_limit = cur_page + local_nr_pages;
|
|
const int page_limit = cur_page + local_nr_pages;
|
|
-
|
|
|
|
|
|
+
|
|
ret = get_user_pages_fast(uaddr, local_nr_pages,
|
|
ret = get_user_pages_fast(uaddr, local_nr_pages,
|
|
write_to_vm, &pages[cur_page]);
|
|
write_to_vm, &pages[cur_page]);
|
|
if (ret < local_nr_pages) {
|
|
if (ret < local_nr_pages) {
|