|
@@ -748,37 +748,32 @@ static unsigned long mem_cgroup_nr_lru_pages(struct mem_cgroup *memcg,
|
|
return total;
|
|
return total;
|
|
}
|
|
}
|
|
|
|
|
|
-static bool __memcg_event_check(struct mem_cgroup *memcg, int target)
|
|
|
|
|
|
+static bool mem_cgroup_event_ratelimit(struct mem_cgroup *memcg,
|
|
|
|
+ enum mem_cgroup_events_target target)
|
|
{
|
|
{
|
|
unsigned long val, next;
|
|
unsigned long val, next;
|
|
|
|
|
|
val = __this_cpu_read(memcg->stat->events[MEM_CGROUP_EVENTS_COUNT]);
|
|
val = __this_cpu_read(memcg->stat->events[MEM_CGROUP_EVENTS_COUNT]);
|
|
next = __this_cpu_read(memcg->stat->targets[target]);
|
|
next = __this_cpu_read(memcg->stat->targets[target]);
|
|
/* from time_after() in jiffies.h */
|
|
/* from time_after() in jiffies.h */
|
|
- return ((long)next - (long)val < 0);
|
|
|
|
-}
|
|
|
|
-
|
|
|
|
-static void __mem_cgroup_target_update(struct mem_cgroup *memcg, int target)
|
|
|
|
-{
|
|
|
|
- unsigned long val, next;
|
|
|
|
-
|
|
|
|
- val = __this_cpu_read(memcg->stat->events[MEM_CGROUP_EVENTS_COUNT]);
|
|
|
|
-
|
|
|
|
- switch (target) {
|
|
|
|
- case MEM_CGROUP_TARGET_THRESH:
|
|
|
|
- next = val + THRESHOLDS_EVENTS_TARGET;
|
|
|
|
- break;
|
|
|
|
- case MEM_CGROUP_TARGET_SOFTLIMIT:
|
|
|
|
- next = val + SOFTLIMIT_EVENTS_TARGET;
|
|
|
|
- break;
|
|
|
|
- case MEM_CGROUP_TARGET_NUMAINFO:
|
|
|
|
- next = val + NUMAINFO_EVENTS_TARGET;
|
|
|
|
- break;
|
|
|
|
- default:
|
|
|
|
- return;
|
|
|
|
|
|
+ if ((long)next - (long)val < 0) {
|
|
|
|
+ switch (target) {
|
|
|
|
+ case MEM_CGROUP_TARGET_THRESH:
|
|
|
|
+ next = val + THRESHOLDS_EVENTS_TARGET;
|
|
|
|
+ break;
|
|
|
|
+ case MEM_CGROUP_TARGET_SOFTLIMIT:
|
|
|
|
+ next = val + SOFTLIMIT_EVENTS_TARGET;
|
|
|
|
+ break;
|
|
|
|
+ case MEM_CGROUP_TARGET_NUMAINFO:
|
|
|
|
+ next = val + NUMAINFO_EVENTS_TARGET;
|
|
|
|
+ break;
|
|
|
|
+ default:
|
|
|
|
+ break;
|
|
|
|
+ }
|
|
|
|
+ __this_cpu_write(memcg->stat->targets[target], next);
|
|
|
|
+ return true;
|
|
}
|
|
}
|
|
-
|
|
|
|
- __this_cpu_write(memcg->stat->targets[target], next);
|
|
|
|
|
|
+ return false;
|
|
}
|
|
}
|
|
|
|
|
|
/*
|
|
/*
|
|
@@ -789,25 +784,27 @@ static void memcg_check_events(struct mem_cgroup *memcg, struct page *page)
|
|
{
|
|
{
|
|
preempt_disable();
|
|
preempt_disable();
|
|
/* threshold event is triggered in finer grain than soft limit */
|
|
/* threshold event is triggered in finer grain than soft limit */
|
|
- if (unlikely(__memcg_event_check(memcg, MEM_CGROUP_TARGET_THRESH))) {
|
|
|
|
|
|
+ if (unlikely(mem_cgroup_event_ratelimit(memcg,
|
|
|
|
+ MEM_CGROUP_TARGET_THRESH))) {
|
|
|
|
+ bool do_softlimit, do_numainfo;
|
|
|
|
+
|
|
|
|
+ do_softlimit = mem_cgroup_event_ratelimit(memcg,
|
|
|
|
+ MEM_CGROUP_TARGET_SOFTLIMIT);
|
|
|
|
+#if MAX_NUMNODES > 1
|
|
|
|
+ do_numainfo = mem_cgroup_event_ratelimit(memcg,
|
|
|
|
+ MEM_CGROUP_TARGET_NUMAINFO);
|
|
|
|
+#endif
|
|
|
|
+ preempt_enable();
|
|
|
|
+
|
|
mem_cgroup_threshold(memcg);
|
|
mem_cgroup_threshold(memcg);
|
|
- __mem_cgroup_target_update(memcg, MEM_CGROUP_TARGET_THRESH);
|
|
|
|
- if (unlikely(__memcg_event_check(memcg,
|
|
|
|
- MEM_CGROUP_TARGET_SOFTLIMIT))) {
|
|
|
|
|
|
+ if (unlikely(do_softlimit))
|
|
mem_cgroup_update_tree(memcg, page);
|
|
mem_cgroup_update_tree(memcg, page);
|
|
- __mem_cgroup_target_update(memcg,
|
|
|
|
- MEM_CGROUP_TARGET_SOFTLIMIT);
|
|
|
|
- }
|
|
|
|
#if MAX_NUMNODES > 1
|
|
#if MAX_NUMNODES > 1
|
|
- if (unlikely(__memcg_event_check(memcg,
|
|
|
|
- MEM_CGROUP_TARGET_NUMAINFO))) {
|
|
|
|
|
|
+ if (unlikely(do_numainfo))
|
|
atomic_inc(&memcg->numainfo_events);
|
|
atomic_inc(&memcg->numainfo_events);
|
|
- __mem_cgroup_target_update(memcg,
|
|
|
|
- MEM_CGROUP_TARGET_NUMAINFO);
|
|
|
|
- }
|
|
|
|
#endif
|
|
#endif
|
|
- }
|
|
|
|
- preempt_enable();
|
|
|
|
|
|
+ } else
|
|
|
|
+ preempt_enable();
|
|
}
|
|
}
|
|
|
|
|
|
struct mem_cgroup *mem_cgroup_from_cont(struct cgroup *cont)
|
|
struct mem_cgroup *mem_cgroup_from_cont(struct cgroup *cont)
|