|
@@ -527,7 +527,7 @@ static inline void wake_up_page(struct page *page, int bit)
|
|
__wake_up_bit(page_waitqueue(page), &page->flags, bit);
|
|
__wake_up_bit(page_waitqueue(page), &page->flags, bit);
|
|
}
|
|
}
|
|
|
|
|
|
-void fastcall wait_on_page_bit(struct page *page, int bit_nr)
|
|
|
|
|
|
+void wait_on_page_bit(struct page *page, int bit_nr)
|
|
{
|
|
{
|
|
DEFINE_WAIT_BIT(wait, &page->flags, bit_nr);
|
|
DEFINE_WAIT_BIT(wait, &page->flags, bit_nr);
|
|
|
|
|
|
@@ -551,7 +551,7 @@ EXPORT_SYMBOL(wait_on_page_bit);
|
|
* the clear_bit and the read of the waitqueue (to avoid SMP races with a
|
|
* the clear_bit and the read of the waitqueue (to avoid SMP races with a
|
|
* parallel wait_on_page_locked()).
|
|
* parallel wait_on_page_locked()).
|
|
*/
|
|
*/
|
|
-void fastcall unlock_page(struct page *page)
|
|
|
|
|
|
+void unlock_page(struct page *page)
|
|
{
|
|
{
|
|
smp_mb__before_clear_bit();
|
|
smp_mb__before_clear_bit();
|
|
if (!TestClearPageLocked(page))
|
|
if (!TestClearPageLocked(page))
|
|
@@ -585,7 +585,7 @@ EXPORT_SYMBOL(end_page_writeback);
|
|
* chances are that on the second loop, the block layer's plug list is empty,
|
|
* chances are that on the second loop, the block layer's plug list is empty,
|
|
* so sync_page() will then return in state TASK_UNINTERRUPTIBLE.
|
|
* so sync_page() will then return in state TASK_UNINTERRUPTIBLE.
|
|
*/
|
|
*/
|
|
-void fastcall __lock_page(struct page *page)
|
|
|
|
|
|
+void __lock_page(struct page *page)
|
|
{
|
|
{
|
|
DEFINE_WAIT_BIT(wait, &page->flags, PG_locked);
|
|
DEFINE_WAIT_BIT(wait, &page->flags, PG_locked);
|
|
|
|
|
|
@@ -606,7 +606,7 @@ int fastcall __lock_page_killable(struct page *page)
|
|
* Variant of lock_page that does not require the caller to hold a reference
|
|
* Variant of lock_page that does not require the caller to hold a reference
|
|
* on the page's mapping.
|
|
* on the page's mapping.
|
|
*/
|
|
*/
|
|
-void fastcall __lock_page_nosync(struct page *page)
|
|
|
|
|
|
+void __lock_page_nosync(struct page *page)
|
|
{
|
|
{
|
|
DEFINE_WAIT_BIT(wait, &page->flags, PG_locked);
|
|
DEFINE_WAIT_BIT(wait, &page->flags, PG_locked);
|
|
__wait_on_bit_lock(page_waitqueue(page), &wait, __sleep_on_page_lock,
|
|
__wait_on_bit_lock(page_waitqueue(page), &wait, __sleep_on_page_lock,
|
|
@@ -1276,7 +1276,7 @@ asmlinkage ssize_t sys_readahead(int fd, loff_t offset, size_t count)
|
|
* This adds the requested page to the page cache if it isn't already there,
|
|
* This adds the requested page to the page cache if it isn't already there,
|
|
* and schedules an I/O to read in its contents from disk.
|
|
* and schedules an I/O to read in its contents from disk.
|
|
*/
|
|
*/
|
|
-static int fastcall page_cache_read(struct file * file, pgoff_t offset)
|
|
|
|
|
|
+static int page_cache_read(struct file *file, pgoff_t offset)
|
|
{
|
|
{
|
|
struct address_space *mapping = file->f_mapping;
|
|
struct address_space *mapping = file->f_mapping;
|
|
struct page *page;
|
|
struct page *page;
|