Merge branch 'parisc-4.17-1' of git://git.kernel.org/pub/scm/linux/kernel/git/deller...
[linux-2.6-block.git] / drivers / md / dm-bio-prison-v2.c
CommitLineData
742c8fdc
JT
1/*
2 * Copyright (C) 2012-2017 Red Hat, Inc.
3 *
4 * This file is released under the GPL.
5 */
6
7#include "dm.h"
8#include "dm-bio-prison-v2.h"
9
10#include <linux/spinlock.h>
11#include <linux/mempool.h>
12#include <linux/module.h>
13#include <linux/slab.h>
14#include <linux/rwsem.h>
15
16/*----------------------------------------------------------------*/
17
18#define MIN_CELLS 1024
19
20struct dm_bio_prison_v2 {
21 struct workqueue_struct *wq;
22
23 spinlock_t lock;
24 mempool_t *cell_pool;
25 struct rb_root cells;
26};
27
28static struct kmem_cache *_cell_cache;
29
30/*----------------------------------------------------------------*/
31
32/*
33 * @nr_cells should be the number of cells you want in use _concurrently_.
34 * Don't confuse it with the number of distinct keys.
35 */
36struct dm_bio_prison_v2 *dm_bio_prison_create_v2(struct workqueue_struct *wq)
37{
38 struct dm_bio_prison_v2 *prison = kmalloc(sizeof(*prison), GFP_KERNEL);
39
40 if (!prison)
41 return NULL;
42
43 prison->wq = wq;
44 spin_lock_init(&prison->lock);
45
46 prison->cell_pool = mempool_create_slab_pool(MIN_CELLS, _cell_cache);
47 if (!prison->cell_pool) {
48 kfree(prison);
49 return NULL;
50 }
51
52 prison->cells = RB_ROOT;
53
54 return prison;
55}
56EXPORT_SYMBOL_GPL(dm_bio_prison_create_v2);
57
58void dm_bio_prison_destroy_v2(struct dm_bio_prison_v2 *prison)
59{
60 mempool_destroy(prison->cell_pool);
61 kfree(prison);
62}
63EXPORT_SYMBOL_GPL(dm_bio_prison_destroy_v2);
64
65struct dm_bio_prison_cell_v2 *dm_bio_prison_alloc_cell_v2(struct dm_bio_prison_v2 *prison, gfp_t gfp)
66{
67 return mempool_alloc(prison->cell_pool, gfp);
68}
69EXPORT_SYMBOL_GPL(dm_bio_prison_alloc_cell_v2);
70
71void dm_bio_prison_free_cell_v2(struct dm_bio_prison_v2 *prison,
72 struct dm_bio_prison_cell_v2 *cell)
73{
74 mempool_free(cell, prison->cell_pool);
75}
76EXPORT_SYMBOL_GPL(dm_bio_prison_free_cell_v2);
77
78static void __setup_new_cell(struct dm_cell_key_v2 *key,
79 struct dm_bio_prison_cell_v2 *cell)
80{
81 memset(cell, 0, sizeof(*cell));
82 memcpy(&cell->key, key, sizeof(cell->key));
83 bio_list_init(&cell->bios);
84}
85
86static int cmp_keys(struct dm_cell_key_v2 *lhs,
87 struct dm_cell_key_v2 *rhs)
88{
89 if (lhs->virtual < rhs->virtual)
90 return -1;
91
92 if (lhs->virtual > rhs->virtual)
93 return 1;
94
95 if (lhs->dev < rhs->dev)
96 return -1;
97
98 if (lhs->dev > rhs->dev)
99 return 1;
100
101 if (lhs->block_end <= rhs->block_begin)
102 return -1;
103
104 if (lhs->block_begin >= rhs->block_end)
105 return 1;
106
107 return 0;
108}
109
110/*
111 * Returns true if node found, otherwise it inserts a new one.
112 */
113static bool __find_or_insert(struct dm_bio_prison_v2 *prison,
114 struct dm_cell_key_v2 *key,
115 struct dm_bio_prison_cell_v2 *cell_prealloc,
116 struct dm_bio_prison_cell_v2 **result)
117{
118 int r;
119 struct rb_node **new = &prison->cells.rb_node, *parent = NULL;
120
121 while (*new) {
122 struct dm_bio_prison_cell_v2 *cell =
6e333d0b 123 rb_entry(*new, struct dm_bio_prison_cell_v2, node);
742c8fdc
JT
124
125 r = cmp_keys(key, &cell->key);
126
127 parent = *new;
128 if (r < 0)
129 new = &((*new)->rb_left);
130
131 else if (r > 0)
132 new = &((*new)->rb_right);
133
134 else {
135 *result = cell;
136 return true;
137 }
138 }
139
140 __setup_new_cell(key, cell_prealloc);
141 *result = cell_prealloc;
142 rb_link_node(&cell_prealloc->node, parent, new);
143 rb_insert_color(&cell_prealloc->node, &prison->cells);
144
145 return false;
146}
147
148static bool __get(struct dm_bio_prison_v2 *prison,
149 struct dm_cell_key_v2 *key,
150 unsigned lock_level,
151 struct bio *inmate,
152 struct dm_bio_prison_cell_v2 *cell_prealloc,
153 struct dm_bio_prison_cell_v2 **cell)
154{
155 if (__find_or_insert(prison, key, cell_prealloc, cell)) {
156 if ((*cell)->exclusive_lock) {
157 if (lock_level <= (*cell)->exclusive_level) {
158 bio_list_add(&(*cell)->bios, inmate);
159 return false;
160 }
161 }
162
163 (*cell)->shared_count++;
164
165 } else
166 (*cell)->shared_count = 1;
167
168 return true;
169}
170
171bool dm_cell_get_v2(struct dm_bio_prison_v2 *prison,
172 struct dm_cell_key_v2 *key,
173 unsigned lock_level,
174 struct bio *inmate,
175 struct dm_bio_prison_cell_v2 *cell_prealloc,
176 struct dm_bio_prison_cell_v2 **cell_result)
177{
178 int r;
179 unsigned long flags;
180
181 spin_lock_irqsave(&prison->lock, flags);
182 r = __get(prison, key, lock_level, inmate, cell_prealloc, cell_result);
183 spin_unlock_irqrestore(&prison->lock, flags);
184
185 return r;
186}
187EXPORT_SYMBOL_GPL(dm_cell_get_v2);
188
189static bool __put(struct dm_bio_prison_v2 *prison,
190 struct dm_bio_prison_cell_v2 *cell)
191{
192 BUG_ON(!cell->shared_count);
193 cell->shared_count--;
194
195 // FIXME: shared locks granted above the lock level could starve this
196 if (!cell->shared_count) {
197 if (cell->exclusive_lock){
198 if (cell->quiesce_continuation) {
199 queue_work(prison->wq, cell->quiesce_continuation);
200 cell->quiesce_continuation = NULL;
201 }
202 } else {
203 rb_erase(&cell->node, &prison->cells);
204 return true;
205 }
206 }
207
208 return false;
209}
210
211bool dm_cell_put_v2(struct dm_bio_prison_v2 *prison,
212 struct dm_bio_prison_cell_v2 *cell)
213{
214 bool r;
215 unsigned long flags;
216
217 spin_lock_irqsave(&prison->lock, flags);
218 r = __put(prison, cell);
219 spin_unlock_irqrestore(&prison->lock, flags);
220
221 return r;
222}
223EXPORT_SYMBOL_GPL(dm_cell_put_v2);
224
225static int __lock(struct dm_bio_prison_v2 *prison,
226 struct dm_cell_key_v2 *key,
227 unsigned lock_level,
228 struct dm_bio_prison_cell_v2 *cell_prealloc,
229 struct dm_bio_prison_cell_v2 **cell_result)
230{
231 struct dm_bio_prison_cell_v2 *cell;
232
233 if (__find_or_insert(prison, key, cell_prealloc, &cell)) {
234 if (cell->exclusive_lock)
235 return -EBUSY;
236
237 cell->exclusive_lock = true;
238 cell->exclusive_level = lock_level;
239 *cell_result = cell;
240
241 // FIXME: we don't yet know what level these shared locks
242 // were taken at, so have to quiesce them all.
243 return cell->shared_count > 0;
244
245 } else {
246 cell = cell_prealloc;
247 cell->shared_count = 0;
248 cell->exclusive_lock = true;
249 cell->exclusive_level = lock_level;
250 *cell_result = cell;
251 }
252
253 return 0;
254}
255
256int dm_cell_lock_v2(struct dm_bio_prison_v2 *prison,
257 struct dm_cell_key_v2 *key,
258 unsigned lock_level,
259 struct dm_bio_prison_cell_v2 *cell_prealloc,
260 struct dm_bio_prison_cell_v2 **cell_result)
261{
262 int r;
263 unsigned long flags;
264
265 spin_lock_irqsave(&prison->lock, flags);
266 r = __lock(prison, key, lock_level, cell_prealloc, cell_result);
267 spin_unlock_irqrestore(&prison->lock, flags);
268
269 return r;
270}
271EXPORT_SYMBOL_GPL(dm_cell_lock_v2);
272
273static void __quiesce(struct dm_bio_prison_v2 *prison,
274 struct dm_bio_prison_cell_v2 *cell,
275 struct work_struct *continuation)
276{
277 if (!cell->shared_count)
278 queue_work(prison->wq, continuation);
279 else
280 cell->quiesce_continuation = continuation;
281}
282
283void dm_cell_quiesce_v2(struct dm_bio_prison_v2 *prison,
284 struct dm_bio_prison_cell_v2 *cell,
285 struct work_struct *continuation)
286{
287 unsigned long flags;
288
289 spin_lock_irqsave(&prison->lock, flags);
290 __quiesce(prison, cell, continuation);
291 spin_unlock_irqrestore(&prison->lock, flags);
292}
293EXPORT_SYMBOL_GPL(dm_cell_quiesce_v2);
294
295static int __promote(struct dm_bio_prison_v2 *prison,
296 struct dm_bio_prison_cell_v2 *cell,
297 unsigned new_lock_level)
298{
299 if (!cell->exclusive_lock)
300 return -EINVAL;
301
302 cell->exclusive_level = new_lock_level;
303 return cell->shared_count > 0;
304}
305
306int dm_cell_lock_promote_v2(struct dm_bio_prison_v2 *prison,
307 struct dm_bio_prison_cell_v2 *cell,
308 unsigned new_lock_level)
309{
310 int r;
311 unsigned long flags;
312
313 spin_lock_irqsave(&prison->lock, flags);
314 r = __promote(prison, cell, new_lock_level);
315 spin_unlock_irqrestore(&prison->lock, flags);
316
317 return r;
318}
319EXPORT_SYMBOL_GPL(dm_cell_lock_promote_v2);
320
321static bool __unlock(struct dm_bio_prison_v2 *prison,
322 struct dm_bio_prison_cell_v2 *cell,
323 struct bio_list *bios)
324{
325 BUG_ON(!cell->exclusive_lock);
326
327 bio_list_merge(bios, &cell->bios);
328 bio_list_init(&cell->bios);
329
330 if (cell->shared_count) {
331 cell->exclusive_lock = 0;
332 return false;
333 }
334
335 rb_erase(&cell->node, &prison->cells);
336 return true;
337}
338
339bool dm_cell_unlock_v2(struct dm_bio_prison_v2 *prison,
340 struct dm_bio_prison_cell_v2 *cell,
341 struct bio_list *bios)
342{
343 bool r;
344 unsigned long flags;
345
346 spin_lock_irqsave(&prison->lock, flags);
347 r = __unlock(prison, cell, bios);
348 spin_unlock_irqrestore(&prison->lock, flags);
349
350 return r;
351}
352EXPORT_SYMBOL_GPL(dm_cell_unlock_v2);
353
354/*----------------------------------------------------------------*/
355
356int __init dm_bio_prison_init_v2(void)
357{
358 _cell_cache = KMEM_CACHE(dm_bio_prison_cell_v2, 0);
359 if (!_cell_cache)
360 return -ENOMEM;
361
362 return 0;
363}
364
365void dm_bio_prison_exit_v2(void)
366{
367 kmem_cache_destroy(_cell_cache);
368 _cell_cache = NULL;
369}