|
@@ -415,14 +415,8 @@ void global_dirty_limits(unsigned long *pbackground, unsigned long *pdirty)
|
|
|
|
|
|
if (vm_dirty_bytes)
|
|
|
dirty = DIV_ROUND_UP(vm_dirty_bytes, PAGE_SIZE);
|
|
|
- else {
|
|
|
- int dirty_ratio;
|
|
|
-
|
|
|
- dirty_ratio = vm_dirty_ratio;
|
|
|
- if (dirty_ratio < 5)
|
|
|
- dirty_ratio = 5;
|
|
|
- dirty = (dirty_ratio * available_memory) / 100;
|
|
|
- }
|
|
|
+ else
|
|
|
+ dirty = (vm_dirty_ratio * available_memory) / 100;
|
|
|
|
|
|
if (dirty_background_bytes)
|
|
|
background = DIV_ROUND_UP(dirty_background_bytes, PAGE_SIZE);
|
|
@@ -510,7 +504,7 @@ static void balance_dirty_pages(struct address_space *mapping,
|
|
|
* catch-up. This avoids (excessively) small writeouts
|
|
|
* when the bdi limits are ramping up.
|
|
|
*/
|
|
|
- if (nr_reclaimable + nr_writeback <
|
|
|
+ if (nr_reclaimable + nr_writeback <=
|
|
|
(background_thresh + dirty_thresh) / 2)
|
|
|
break;
|
|
|
|
|
@@ -542,8 +536,8 @@ static void balance_dirty_pages(struct address_space *mapping,
|
|
|
* the last resort safeguard.
|
|
|
*/
|
|
|
dirty_exceeded =
|
|
|
- (bdi_nr_reclaimable + bdi_nr_writeback >= bdi_thresh)
|
|
|
- || (nr_reclaimable + nr_writeback >= dirty_thresh);
|
|
|
+ (bdi_nr_reclaimable + bdi_nr_writeback > bdi_thresh)
|
|
|
+ || (nr_reclaimable + nr_writeback > dirty_thresh);
|
|
|
|
|
|
if (!dirty_exceeded)
|
|
|
break;
|