|
@@ -1,155 +0,0 @@
|
|
|
-/*
|
|
|
- * mm/thrash.c
|
|
|
- *
|
|
|
- * Copyright (C) 2004, Red Hat, Inc.
|
|
|
- * Copyright (C) 2004, Rik van Riel <riel@redhat.com>
|
|
|
- * Released under the GPL, see the file COPYING for details.
|
|
|
- *
|
|
|
- * Simple token based thrashing protection, using the algorithm
|
|
|
- * described in: http://www.cse.ohio-state.edu/hpcs/WWW/HTML/publications/abs05-1.html
|
|
|
- *
|
|
|
- * Sep 2006, Ashwin Chaugule <ashwin.chaugule@celunite.com>
|
|
|
- * Improved algorithm to pass token:
|
|
|
- * Each task has a priority which is incremented if it contended
|
|
|
- * for the token in an interval less than its previous attempt.
|
|
|
- * If the token is acquired, that task's priority is boosted to prevent
|
|
|
- * the token from bouncing around too often and to let the task make
|
|
|
- * some progress in its execution.
|
|
|
- */
|
|
|
-
|
|
|
-#include <linux/jiffies.h>
|
|
|
-#include <linux/mm.h>
|
|
|
-#include <linux/sched.h>
|
|
|
-#include <linux/swap.h>
|
|
|
-#include <linux/memcontrol.h>
|
|
|
-
|
|
|
-#include <trace/events/vmscan.h>
|
|
|
-
|
|
|
-#define TOKEN_AGING_INTERVAL (0xFF)
|
|
|
-
|
|
|
-static DEFINE_SPINLOCK(swap_token_lock);
|
|
|
-struct mm_struct *swap_token_mm;
|
|
|
-static struct mem_cgroup *swap_token_memcg;
|
|
|
-
|
|
|
-#ifdef CONFIG_CGROUP_MEM_RES_CTLR
|
|
|
-static struct mem_cgroup *swap_token_memcg_from_mm(struct mm_struct *mm)
|
|
|
-{
|
|
|
- struct mem_cgroup *memcg;
|
|
|
-
|
|
|
- memcg = try_get_mem_cgroup_from_mm(mm);
|
|
|
- if (memcg)
|
|
|
- css_put(mem_cgroup_css(memcg));
|
|
|
-
|
|
|
- return memcg;
|
|
|
-}
|
|
|
-#else
|
|
|
-static struct mem_cgroup *swap_token_memcg_from_mm(struct mm_struct *mm)
|
|
|
-{
|
|
|
- return NULL;
|
|
|
-}
|
|
|
-#endif
|
|
|
-
|
|
|
-void grab_swap_token(struct mm_struct *mm)
|
|
|
-{
|
|
|
- int current_interval;
|
|
|
- unsigned int old_prio = mm->token_priority;
|
|
|
- static unsigned int global_faults;
|
|
|
- static unsigned int last_aging;
|
|
|
-
|
|
|
- global_faults++;
|
|
|
-
|
|
|
- current_interval = global_faults - mm->faultstamp;
|
|
|
-
|
|
|
- if (!spin_trylock(&swap_token_lock))
|
|
|
- return;
|
|
|
-
|
|
|
- /* First come first served */
|
|
|
- if (!swap_token_mm)
|
|
|
- goto replace_token;
|
|
|
-
|
|
|
- /*
|
|
|
- * Usually, we don't need priority aging because long interval faults
|
|
|
- * makes priority decrease quickly. But there is one exception. If the
|
|
|
- * token owner task is sleeping, it never make long interval faults.
|
|
|
- * Thus, we need a priority aging mechanism instead. The requirements
|
|
|
- * of priority aging are
|
|
|
- * 1) An aging interval is reasonable enough long. Too short aging
|
|
|
- * interval makes quick swap token lost and decrease performance.
|
|
|
- * 2) The swap token owner task have to get priority aging even if
|
|
|
- * it's under sleep.
|
|
|
- */
|
|
|
- if ((global_faults - last_aging) > TOKEN_AGING_INTERVAL) {
|
|
|
- swap_token_mm->token_priority /= 2;
|
|
|
- last_aging = global_faults;
|
|
|
- }
|
|
|
-
|
|
|
- if (mm == swap_token_mm) {
|
|
|
- mm->token_priority += 2;
|
|
|
- goto update_priority;
|
|
|
- }
|
|
|
-
|
|
|
- if (current_interval < mm->last_interval)
|
|
|
- mm->token_priority++;
|
|
|
- else {
|
|
|
- if (likely(mm->token_priority > 0))
|
|
|
- mm->token_priority--;
|
|
|
- }
|
|
|
-
|
|
|
- /* Check if we deserve the token */
|
|
|
- if (mm->token_priority > swap_token_mm->token_priority)
|
|
|
- goto replace_token;
|
|
|
-
|
|
|
-update_priority:
|
|
|
- trace_update_swap_token_priority(mm, old_prio, swap_token_mm);
|
|
|
-
|
|
|
-out:
|
|
|
- mm->faultstamp = global_faults;
|
|
|
- mm->last_interval = current_interval;
|
|
|
- spin_unlock(&swap_token_lock);
|
|
|
- return;
|
|
|
-
|
|
|
-replace_token:
|
|
|
- mm->token_priority += 2;
|
|
|
- trace_replace_swap_token(swap_token_mm, mm);
|
|
|
- swap_token_mm = mm;
|
|
|
- swap_token_memcg = swap_token_memcg_from_mm(mm);
|
|
|
- last_aging = global_faults;
|
|
|
- goto out;
|
|
|
-}
|
|
|
-
|
|
|
-/* Called on process exit. */
|
|
|
-void __put_swap_token(struct mm_struct *mm)
|
|
|
-{
|
|
|
- spin_lock(&swap_token_lock);
|
|
|
- if (likely(mm == swap_token_mm)) {
|
|
|
- trace_put_swap_token(swap_token_mm);
|
|
|
- swap_token_mm = NULL;
|
|
|
- swap_token_memcg = NULL;
|
|
|
- }
|
|
|
- spin_unlock(&swap_token_lock);
|
|
|
-}
|
|
|
-
|
|
|
-static bool match_memcg(struct mem_cgroup *a, struct mem_cgroup *b)
|
|
|
-{
|
|
|
- if (!a)
|
|
|
- return true;
|
|
|
- if (!b)
|
|
|
- return true;
|
|
|
- if (a == b)
|
|
|
- return true;
|
|
|
- return false;
|
|
|
-}
|
|
|
-
|
|
|
-void disable_swap_token(struct mem_cgroup *memcg)
|
|
|
-{
|
|
|
- /* memcg reclaim don't disable unrelated mm token. */
|
|
|
- if (match_memcg(memcg, swap_token_memcg)) {
|
|
|
- spin_lock(&swap_token_lock);
|
|
|
- if (match_memcg(memcg, swap_token_memcg)) {
|
|
|
- trace_disable_swap_token(swap_token_mm);
|
|
|
- swap_token_mm = NULL;
|
|
|
- swap_token_memcg = NULL;
|
|
|
- }
|
|
|
- spin_unlock(&swap_token_lock);
|
|
|
- }
|
|
|
-}
|