ext4: fix race between sync and completed io work
[linux-2.6-block.git] / mm / thrash.c
CommitLineData
1da177e4
LT
1/*
2 * mm/thrash.c
3 *
4 * Copyright (C) 2004, Red Hat, Inc.
5 * Copyright (C) 2004, Rik van Riel <riel@redhat.com>
6 * Released under the GPL, see the file COPYING for details.
7 *
8 * Simple token based thrashing protection, using the algorithm
e21c7ffd 9 * described in: http://www.cse.ohio-state.edu/hpcs/WWW/HTML/publications/abs05-1.html
7602bdf2
AC
10 *
11 * Sep 2006, Ashwin Chaugule <ashwin.chaugule@celunite.com>
12 * Improved algorithm to pass token:
13 * Each task has a priority which is incremented if it contended
14 * for the token in an interval less than its previous attempt.
15 * If the token is acquired, that task's priority is boosted to prevent
16 * the token from bouncing around too often and to let the task make
17 * some progress in its execution.
1da177e4 18 */
7602bdf2 19
1da177e4
LT
20#include <linux/jiffies.h>
21#include <linux/mm.h>
22#include <linux/sched.h>
23#include <linux/swap.h>
a433658c 24#include <linux/memcontrol.h>
1da177e4 25
83cd81a3
KM
26#include <trace/events/vmscan.h>
27
d7911ef3
KM
28#define TOKEN_AGING_INTERVAL (0xFF)
29
1da177e4 30static DEFINE_SPINLOCK(swap_token_lock);
7602bdf2 31struct mm_struct *swap_token_mm;
22d5368a 32static struct mem_cgroup *swap_token_memcg;
1da177e4 33
a433658c
KM
34#ifdef CONFIG_CGROUP_MEM_RES_CTLR
35static struct mem_cgroup *swap_token_memcg_from_mm(struct mm_struct *mm)
36{
37 struct mem_cgroup *memcg;
38
39 memcg = try_get_mem_cgroup_from_mm(mm);
40 if (memcg)
41 css_put(mem_cgroup_css(memcg));
42
43 return memcg;
44}
45#else
46static struct mem_cgroup *swap_token_memcg_from_mm(struct mm_struct *mm)
47{
48 return NULL;
49}
50#endif
51
a5c9b696 52void grab_swap_token(struct mm_struct *mm)
1da177e4 53{
7602bdf2 54 int current_interval;
83cd81a3 55 unsigned int old_prio = mm->token_priority;
53bb01f5
KM
56 static unsigned int global_faults;
57 static unsigned int last_aging;
1da177e4 58
7602bdf2 59 global_faults++;
1da177e4 60
a5c9b696 61 current_interval = global_faults - mm->faultstamp;
1da177e4 62
7602bdf2
AC
63 if (!spin_trylock(&swap_token_lock))
64 return;
1da177e4 65
7602bdf2 66 /* First come first served */
a433658c
KM
67 if (!swap_token_mm)
68 goto replace_token;
69
45ebb840
KM
70 /*
71 * Usually, we don't need priority aging because long interval faults
72 * makes priority decrease quickly. But there is one exception. If the
73 * token owner task is sleeping, it never make long interval faults.
74 * Thus, we need a priority aging mechanism instead. The requirements
75 * of priority aging are
76 * 1) An aging interval is reasonable enough long. Too short aging
77 * interval makes quick swap token lost and decrease performance.
78 * 2) The swap token owner task have to get priority aging even if
79 * it's under sleep.
80 */
d7911ef3
KM
81 if ((global_faults - last_aging) > TOKEN_AGING_INTERVAL) {
82 swap_token_mm->token_priority /= 2;
83 last_aging = global_faults;
84 }
85
a433658c
KM
86 if (mm == swap_token_mm) {
87 mm->token_priority += 2;
83cd81a3 88 goto update_priority;
7602bdf2 89 }
1da177e4 90
a433658c
KM
91 if (current_interval < mm->last_interval)
92 mm->token_priority++;
93 else {
94 if (likely(mm->token_priority > 0))
95 mm->token_priority--;
1da177e4 96 }
7602bdf2 97
a433658c
KM
98 /* Check if we deserve the token */
99 if (mm->token_priority > swap_token_mm->token_priority)
100 goto replace_token;
101
83cd81a3 102update_priority:
d7911ef3 103 trace_update_swap_token_priority(mm, old_prio, swap_token_mm);
83cd81a3 104
7602bdf2 105out:
a5c9b696
HD
106 mm->faultstamp = global_faults;
107 mm->last_interval = current_interval;
7602bdf2 108 spin_unlock(&swap_token_lock);
a433658c
KM
109 return;
110
111replace_token:
112 mm->token_priority += 2;
83cd81a3 113 trace_replace_swap_token(swap_token_mm, mm);
a433658c
KM
114 swap_token_mm = mm;
115 swap_token_memcg = swap_token_memcg_from_mm(mm);
d7911ef3 116 last_aging = global_faults;
a433658c 117 goto out;
1da177e4
LT
118}
119
120/* Called on process exit. */
121void __put_swap_token(struct mm_struct *mm)
122{
123 spin_lock(&swap_token_lock);
a433658c 124 if (likely(mm == swap_token_mm)) {
83cd81a3 125 trace_put_swap_token(swap_token_mm);
7602bdf2 126 swap_token_mm = NULL;
a433658c
KM
127 swap_token_memcg = NULL;
128 }
1da177e4
LT
129 spin_unlock(&swap_token_lock);
130}
a433658c
KM
131
132static bool match_memcg(struct mem_cgroup *a, struct mem_cgroup *b)
133{
134 if (!a)
135 return true;
136 if (!b)
137 return true;
138 if (a == b)
139 return true;
140 return false;
141}
142
143void disable_swap_token(struct mem_cgroup *memcg)
144{
145 /* memcg reclaim don't disable unrelated mm token. */
146 if (match_memcg(memcg, swap_token_memcg)) {
147 spin_lock(&swap_token_lock);
148 if (match_memcg(memcg, swap_token_memcg)) {
83cd81a3 149 trace_disable_swap_token(swap_token_mm);
a433658c
KM
150 swap_token_mm = NULL;
151 swap_token_memcg = NULL;
152 }
153 spin_unlock(&swap_token_lock);
154 }
155}