cachefiles: Fix refcounting bug in backing-file read monitoring
[linux-2.6-block.git] / fs / fscache / object.c
CommitLineData
36c95590
DH
1/* FS-Cache object state machine handler
2 *
3 * Copyright (C) 2007 Red Hat, Inc. All Rights Reserved.
4 * Written by David Howells (dhowells@redhat.com)
5 *
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
10 *
11 * See Documentation/filesystems/caching/object.txt for a description of the
12 * object state machine and the in-kernel representations.
13 */
14
15#define FSCACHE_DEBUG_LEVEL COOKIE
16#include <linux/module.h>
ef778e7a 17#include <linux/slab.h>
caaef690 18#include <linux/prefetch.h>
36c95590
DH
19#include "internal.h"
20
caaef690
DH
21static const struct fscache_state *fscache_abort_initialisation(struct fscache_object *, int);
22static const struct fscache_state *fscache_kill_dependents(struct fscache_object *, int);
23static const struct fscache_state *fscache_drop_object(struct fscache_object *, int);
24static const struct fscache_state *fscache_initialise_object(struct fscache_object *, int);
25static const struct fscache_state *fscache_invalidate_object(struct fscache_object *, int);
26static const struct fscache_state *fscache_jumpstart_dependents(struct fscache_object *, int);
27static const struct fscache_state *fscache_kill_object(struct fscache_object *, int);
28static const struct fscache_state *fscache_lookup_failure(struct fscache_object *, int);
29static const struct fscache_state *fscache_look_up_object(struct fscache_object *, int);
30static const struct fscache_state *fscache_object_available(struct fscache_object *, int);
31static const struct fscache_state *fscache_parent_ready(struct fscache_object *, int);
32static const struct fscache_state *fscache_update_object(struct fscache_object *, int);
e26bfebd 33static const struct fscache_state *fscache_object_dead(struct fscache_object *, int);
caaef690
DH
34
35#define __STATE_NAME(n) fscache_osm_##n
36#define STATE(n) (&__STATE_NAME(n))
37
38/*
39 * Define a work state. Work states are execution states. No event processing
40 * is performed by them. The function attached to a work state returns a
41 * pointer indicating the next state to which the state machine should
42 * transition. Returning NO_TRANSIT repeats the current state, but goes back
43 * to the scheduler first.
44 */
45#define WORK_STATE(n, sn, f) \
46 const struct fscache_state __STATE_NAME(n) = { \
47 .name = #n, \
48 .short_name = sn, \
49 .work = f \
50 }
51
52/*
53 * Returns from work states.
54 */
55#define transit_to(state) ({ prefetch(&STATE(state)->work); STATE(state); })
56
57#define NO_TRANSIT ((struct fscache_state *)NULL)
58
59/*
60 * Define a wait state. Wait states are event processing states. No execution
61 * is performed by them. Wait states are just tables of "if event X occurs,
62 * clear it and transition to state Y". The dispatcher returns to the
63 * scheduler if none of the events in which the wait state has an interest are
64 * currently pending.
65 */
66#define WAIT_STATE(n, sn, ...) \
67 const struct fscache_state __STATE_NAME(n) = { \
68 .name = #n, \
69 .short_name = sn, \
70 .work = NULL, \
71 .transitions = { __VA_ARGS__, { 0, NULL } } \
72 }
73
74#define TRANSIT_TO(state, emask) \
75 { .events = (emask), .transit_to = STATE(state) }
76
77/*
78 * The object state machine.
79 */
80static WORK_STATE(INIT_OBJECT, "INIT", fscache_initialise_object);
81static WORK_STATE(PARENT_READY, "PRDY", fscache_parent_ready);
82static WORK_STATE(ABORT_INIT, "ABRT", fscache_abort_initialisation);
83static WORK_STATE(LOOK_UP_OBJECT, "LOOK", fscache_look_up_object);
84static WORK_STATE(CREATE_OBJECT, "CRTO", fscache_look_up_object);
85static WORK_STATE(OBJECT_AVAILABLE, "AVBL", fscache_object_available);
86static WORK_STATE(JUMPSTART_DEPS, "JUMP", fscache_jumpstart_dependents);
87
88static WORK_STATE(INVALIDATE_OBJECT, "INVL", fscache_invalidate_object);
89static WORK_STATE(UPDATE_OBJECT, "UPDT", fscache_update_object);
90
91static WORK_STATE(LOOKUP_FAILURE, "LCFL", fscache_lookup_failure);
92static WORK_STATE(KILL_OBJECT, "KILL", fscache_kill_object);
93static WORK_STATE(KILL_DEPENDENTS, "KDEP", fscache_kill_dependents);
94static WORK_STATE(DROP_OBJECT, "DROP", fscache_drop_object);
e26bfebd 95static WORK_STATE(OBJECT_DEAD, "DEAD", fscache_object_dead);
caaef690
DH
96
97static WAIT_STATE(WAIT_FOR_INIT, "?INI",
98 TRANSIT_TO(INIT_OBJECT, 1 << FSCACHE_OBJECT_EV_NEW_CHILD));
99
100static WAIT_STATE(WAIT_FOR_PARENT, "?PRN",
101 TRANSIT_TO(PARENT_READY, 1 << FSCACHE_OBJECT_EV_PARENT_READY));
102
103static WAIT_STATE(WAIT_FOR_CMD, "?CMD",
104 TRANSIT_TO(INVALIDATE_OBJECT, 1 << FSCACHE_OBJECT_EV_INVALIDATE),
105 TRANSIT_TO(UPDATE_OBJECT, 1 << FSCACHE_OBJECT_EV_UPDATE),
106 TRANSIT_TO(JUMPSTART_DEPS, 1 << FSCACHE_OBJECT_EV_NEW_CHILD));
107
108static WAIT_STATE(WAIT_FOR_CLEARANCE, "?CLR",
109 TRANSIT_TO(KILL_OBJECT, 1 << FSCACHE_OBJECT_EV_CLEARED));
110
111/*
112 * Out-of-band event transition tables. These are for handling unexpected
113 * events, such as an I/O error. If an OOB event occurs, the state machine
114 * clears and disables the event and forces a transition to the nominated work
115 * state (acurrently executing work states will complete first).
116 *
117 * In such a situation, object->state remembers the state the machine should
118 * have been in/gone to and returning NO_TRANSIT returns to that.
119 */
120static const struct fscache_transition fscache_osm_init_oob[] = {
121 TRANSIT_TO(ABORT_INIT,
122 (1 << FSCACHE_OBJECT_EV_ERROR) |
123 (1 << FSCACHE_OBJECT_EV_KILL)),
124 { 0, NULL }
36c95590 125};
caaef690
DH
126
127static const struct fscache_transition fscache_osm_lookup_oob[] = {
128 TRANSIT_TO(LOOKUP_FAILURE,
129 (1 << FSCACHE_OBJECT_EV_ERROR) |
130 (1 << FSCACHE_OBJECT_EV_KILL)),
131 { 0, NULL }
132};
133
134static const struct fscache_transition fscache_osm_run_oob[] = {
135 TRANSIT_TO(KILL_OBJECT,
136 (1 << FSCACHE_OBJECT_EV_ERROR) |
137 (1 << FSCACHE_OBJECT_EV_KILL)),
138 { 0, NULL }
440f0aff
DH
139};
140
a18feb55
DH
141static int fscache_get_object(struct fscache_object *,
142 enum fscache_obj_ref_trace);
143static void fscache_put_object(struct fscache_object *,
144 enum fscache_obj_ref_trace);
caaef690 145static bool fscache_enqueue_dependents(struct fscache_object *, int);
36c95590 146static void fscache_dequeue_object(struct fscache_object *);
402cb8dd 147static void fscache_update_aux_data(struct fscache_object *);
36c95590 148
36c95590
DH
149/*
150 * we need to notify the parent when an op completes that we had outstanding
151 * upon it
152 */
153static inline void fscache_done_parent_op(struct fscache_object *object)
154{
155 struct fscache_object *parent = object->parent;
156
157 _enter("OBJ%x {OBJ%x,%x}",
158 object->debug_id, parent->debug_id, parent->n_ops);
159
160 spin_lock_nested(&parent->lock, 1);
36c95590 161 parent->n_obj_ops--;
1362729b 162 parent->n_ops--;
36c95590
DH
163 if (parent->n_ops == 0)
164 fscache_raise_event(parent, FSCACHE_OBJECT_EV_CLEARED);
165 spin_unlock(&parent->lock);
166}
167
ef778e7a 168/*
caaef690 169 * Object state machine dispatcher.
ef778e7a 170 */
caaef690 171static void fscache_object_sm_dispatcher(struct fscache_object *object)
ef778e7a 172{
caaef690
DH
173 const struct fscache_transition *t;
174 const struct fscache_state *state, *new_state;
175 unsigned long events, event_mask;
a18feb55 176 bool oob;
caaef690 177 int event = -1;
36c95590
DH
178
179 ASSERT(object != NULL);
180
181 _enter("{OBJ%x,%s,%lx}",
caaef690
DH
182 object->debug_id, object->state->name, object->events);
183
184 event_mask = object->event_mask;
185restart:
186 object->event_mask = 0; /* Mask normal event handling */
187 state = object->state;
188restart_masked:
189 events = object->events;
190
191 /* Handle any out-of-band events (typically an error) */
192 if (events & object->oob_event_mask) {
193 _debug("{OBJ%x} oob %lx",
194 object->debug_id, events & object->oob_event_mask);
a18feb55 195 oob = true;
caaef690
DH
196 for (t = object->oob_table; t->events; t++) {
197 if (events & t->events) {
198 state = t->transit_to;
199 ASSERT(state->work != NULL);
200 event = fls(events & t->events) - 1;
201 __clear_bit(event, &object->oob_event_mask);
202 clear_bit(event, &object->events);
203 goto execute_work_state;
204 }
d461d26d 205 }
caaef690 206 }
a18feb55 207 oob = false;
36c95590 208
caaef690
DH
209 /* Wait states are just transition tables */
210 if (!state->work) {
211 if (events & event_mask) {
212 for (t = state->transitions; t->events; t++) {
213 if (events & t->events) {
214 new_state = t->transit_to;
215 event = fls(events & t->events) - 1;
a18feb55
DH
216 trace_fscache_osm(object, state,
217 true, false, event);
caaef690
DH
218 clear_bit(event, &object->events);
219 _debug("{OBJ%x} ev %d: %s -> %s",
220 object->debug_id, event,
221 state->name, new_state->name);
222 object->state = state = new_state;
223 goto execute_work_state;
224 }
225 }
36c95590 226
caaef690
DH
227 /* The event mask didn't include all the tabled bits */
228 BUG();
36c95590 229 }
caaef690
DH
230 /* Randomly woke up */
231 goto unmask_events;
36c95590
DH
232 }
233
caaef690
DH
234execute_work_state:
235 _debug("{OBJ%x} exec %s", object->debug_id, state->name);
36c95590 236
a18feb55 237 trace_fscache_osm(object, state, false, oob, event);
caaef690
DH
238 new_state = state->work(object, event);
239 event = -1;
240 if (new_state == NO_TRANSIT) {
241 _debug("{OBJ%x} %s notrans", object->debug_id, state->name);
e26bfebd
DH
242 if (unlikely(state == STATE(OBJECT_DEAD))) {
243 _leave(" [dead]");
244 return;
245 }
caaef690
DH
246 fscache_enqueue_object(object);
247 event_mask = object->oob_event_mask;
248 goto unmask_events;
36c95590
DH
249 }
250
caaef690
DH
251 _debug("{OBJ%x} %s -> %s",
252 object->debug_id, state->name, new_state->name);
253 object->state = state = new_state;
36c95590 254
caaef690 255 if (state->work) {
e26bfebd 256 if (unlikely(state == STATE(OBJECT_DEAD))) {
caaef690
DH
257 _leave(" [dead]");
258 return;
259 }
260 goto restart_masked;
261 }
36c95590 262
caaef690
DH
263 /* Transited to wait state */
264 event_mask = object->oob_event_mask;
265 for (t = state->transitions; t->events; t++)
266 event_mask |= t->events;
267
268unmask_events:
269 object->event_mask = event_mask;
270 smp_mb();
271 events = object->events;
272 if (events & event_mask)
273 goto restart;
274 _leave(" [msk %lx]", event_mask);
36c95590
DH
275}
276
277/*
278 * execute an object
279 */
610be24e 280static void fscache_object_work_func(struct work_struct *work)
36c95590
DH
281{
282 struct fscache_object *object =
283 container_of(work, struct fscache_object, work);
284 unsigned long start;
285
286 _enter("{OBJ%x}", object->debug_id);
287
36c95590 288 start = jiffies;
caaef690 289 fscache_object_sm_dispatcher(object);
36c95590 290 fscache_hist(fscache_objs_histogram, start);
a18feb55 291 fscache_put_object(object, fscache_obj_put_work);
36c95590 292}
610be24e
DH
293
294/**
295 * fscache_object_init - Initialise a cache object description
296 * @object: Object description
297 * @cookie: Cookie object will be attached to
298 * @cache: Cache in which backing object will be found
299 *
300 * Initialise a cache object description to its basic values.
301 *
302 * See Documentation/filesystems/caching/backend-api.txt for a complete
303 * description.
304 */
305void fscache_object_init(struct fscache_object *object,
306 struct fscache_cookie *cookie,
307 struct fscache_cache *cache)
308{
caaef690
DH
309 const struct fscache_transition *t;
310
610be24e
DH
311 atomic_inc(&cache->object_count);
312
caaef690
DH
313 object->state = STATE(WAIT_FOR_INIT);
314 object->oob_table = fscache_osm_init_oob;
315 object->flags = 1 << FSCACHE_OBJECT_IS_LIVE;
610be24e
DH
316 spin_lock_init(&object->lock);
317 INIT_LIST_HEAD(&object->cache_link);
318 INIT_HLIST_NODE(&object->cookie_link);
319 INIT_WORK(&object->work, fscache_object_work_func);
320 INIT_LIST_HEAD(&object->dependents);
321 INIT_LIST_HEAD(&object->dep_link);
322 INIT_LIST_HEAD(&object->pending_ops);
323 object->n_children = 0;
324 object->n_ops = object->n_in_progress = object->n_exclusive = 0;
caaef690 325 object->events = 0;
610be24e
DH
326 object->store_limit = 0;
327 object->store_limit_l = 0;
328 object->cache = cache;
329 object->cookie = cookie;
330 object->parent = NULL;
7026f192
DH
331#ifdef CONFIG_FSCACHE_OBJECT_LIST
332 RB_CLEAR_NODE(&object->objlist_link);
333#endif
caaef690
DH
334
335 object->oob_event_mask = 0;
336 for (t = object->oob_table; t->events; t++)
337 object->oob_event_mask |= t->events;
338 object->event_mask = object->oob_event_mask;
339 for (t = object->state->transitions; t->events; t++)
340 object->event_mask |= t->events;
610be24e
DH
341}
342EXPORT_SYMBOL(fscache_object_init);
440f0aff 343
f09b443d
DH
344/*
345 * Mark the object as no longer being live, making sure that we synchronise
346 * against op submission.
347 */
348static inline void fscache_mark_object_dead(struct fscache_object *object)
349{
350 spin_lock(&object->lock);
351 clear_bit(FSCACHE_OBJECT_IS_LIVE, &object->flags);
352 spin_unlock(&object->lock);
353}
354
caaef690
DH
355/*
356 * Abort object initialisation before we start it.
357 */
358static const struct fscache_state *fscache_abort_initialisation(struct fscache_object *object,
359 int event)
360{
caaef690
DH
361 _enter("{OBJ%x},%d", object->debug_id, event);
362
363 object->oob_event_mask = 0;
caaef690 364 fscache_dequeue_object(object);
caaef690
DH
365 return transit_to(KILL_OBJECT);
366}
367
36c95590
DH
368/*
369 * initialise an object
370 * - check the specified object's parent to see if we can make use of it
371 * immediately to do a creation
372 * - we may need to start the process of creating a parent and we need to wait
373 * for the parent's lookup and creation to complete if it's not there yet
36c95590 374 */
caaef690
DH
375static const struct fscache_state *fscache_initialise_object(struct fscache_object *object,
376 int event)
36c95590
DH
377{
378 struct fscache_object *parent;
caaef690 379 bool success;
36c95590 380
caaef690 381 _enter("{OBJ%x},%d", object->debug_id, event);
36c95590 382
caaef690 383 ASSERT(list_empty(&object->dep_link));
36c95590
DH
384
385 parent = object->parent;
386 if (!parent) {
caaef690 387 _leave(" [no parent]");
1362729b 388 return transit_to(DROP_OBJECT);
caaef690 389 }
36c95590 390
1362729b 391 _debug("parent: %s of:%lx", parent->state->name, parent->flags);
caaef690
DH
392
393 if (fscache_object_is_dying(parent)) {
394 _leave(" [bad parent]");
1362729b 395 return transit_to(DROP_OBJECT);
caaef690
DH
396 }
397
398 if (fscache_object_is_available(parent)) {
399 _leave(" [ready]");
400 return transit_to(PARENT_READY);
401 }
402
403 _debug("wait");
404
405 spin_lock(&parent->lock);
406 fscache_stat(&fscache_n_cop_grab_object);
407 success = false;
408 if (fscache_object_is_live(parent) &&
a18feb55 409 object->cache->ops->grab_object(object, fscache_obj_get_add_to_deps)) {
caaef690
DH
410 list_add(&object->dep_link, &parent->dependents);
411 success = true;
412 }
413 fscache_stat_d(&fscache_n_cop_grab_object);
414 spin_unlock(&parent->lock);
415 if (!success) {
416 _leave(" [grab failed]");
1362729b 417 return transit_to(DROP_OBJECT);
36c95590
DH
418 }
419
caaef690
DH
420 /* fscache_acquire_non_index_cookie() uses this
421 * to wake the chain up */
422 fscache_raise_event(parent, FSCACHE_OBJECT_EV_NEW_CHILD);
423 _leave(" [wait]");
424 return transit_to(WAIT_FOR_PARENT);
425}
426
427/*
428 * Once the parent object is ready, we should kick off our lookup op.
429 */
430static const struct fscache_state *fscache_parent_ready(struct fscache_object *object,
431 int event)
432{
433 struct fscache_object *parent = object->parent;
434
435 _enter("{OBJ%x},%d", object->debug_id, event);
436
437 ASSERT(parent != NULL);
438
439 spin_lock(&parent->lock);
440 parent->n_ops++;
441 parent->n_obj_ops++;
442 object->lookup_jif = jiffies;
443 spin_unlock(&parent->lock);
444
36c95590 445 _leave("");
caaef690 446 return transit_to(LOOK_UP_OBJECT);
36c95590
DH
447}
448
449/*
450 * look an object up in the cache from which it was allocated
451 * - we hold an "access lock" on the parent object, so the parent object cannot
452 * be withdrawn by either party till we've finished
36c95590 453 */
caaef690
DH
454static const struct fscache_state *fscache_look_up_object(struct fscache_object *object,
455 int event)
36c95590
DH
456{
457 struct fscache_cookie *cookie = object->cookie;
caaef690 458 struct fscache_object *parent = object->parent;
fee096de 459 int ret;
36c95590 460
caaef690
DH
461 _enter("{OBJ%x},%d", object->debug_id, event);
462
463 object->oob_table = fscache_osm_lookup_oob;
36c95590 464
36c95590
DH
465 ASSERT(parent != NULL);
466 ASSERTCMP(parent->n_ops, >, 0);
467 ASSERTCMP(parent->n_obj_ops, >, 0);
468
469 /* make sure the parent is still available */
493f7bc1 470 ASSERT(fscache_object_is_available(parent));
36c95590 471
493f7bc1 472 if (fscache_object_is_dying(parent) ||
1362729b
DH
473 test_bit(FSCACHE_IOERROR, &object->cache->flags) ||
474 !fscache_use_cookie(object)) {
caaef690
DH
475 _leave(" [unavailable]");
476 return transit_to(LOOKUP_FAILURE);
36c95590
DH
477 }
478
1362729b
DH
479 _debug("LOOKUP \"%s\" in \"%s\"",
480 cookie->def->name, object->cache->tag->name);
36c95590
DH
481
482 fscache_stat(&fscache_n_object_lookups);
52bd75fd 483 fscache_stat(&fscache_n_cop_lookup_object);
fee096de 484 ret = object->cache->ops->lookup_object(object);
52bd75fd 485 fscache_stat_d(&fscache_n_cop_lookup_object);
36c95590 486
1362729b 487 fscache_unuse_cookie(object);
36c95590 488
fee096de
DH
489 if (ret == -ETIMEDOUT) {
490 /* probably stuck behind another object, so move this one to
491 * the back of the queue */
492 fscache_stat(&fscache_n_object_lookups_timed_out);
caaef690
DH
493 _leave(" [timeout]");
494 return NO_TRANSIT;
fee096de
DH
495 }
496
caaef690
DH
497 if (ret < 0) {
498 _leave(" [error]");
499 return transit_to(LOOKUP_FAILURE);
500 }
501
502 _leave(" [ok]");
503 return transit_to(OBJECT_AVAILABLE);
36c95590
DH
504}
505
506/**
507 * fscache_object_lookup_negative - Note negative cookie lookup
508 * @object: Object pointing to cookie to mark
509 *
510 * Note negative lookup, permitting those waiting to read data from an already
511 * existing backing object to continue as there's no data for them to read.
512 */
513void fscache_object_lookup_negative(struct fscache_object *object)
514{
515 struct fscache_cookie *cookie = object->cookie;
516
caaef690 517 _enter("{OBJ%x,%s}", object->debug_id, object->state->name);
36c95590 518
caaef690 519 if (!test_and_set_bit(FSCACHE_OBJECT_IS_LOOKED_UP, &object->flags)) {
36c95590
DH
520 fscache_stat(&fscache_n_object_lookups_negative);
521
caaef690
DH
522 /* Allow write requests to begin stacking up and read requests to begin
523 * returning ENODATA.
524 */
36c95590 525 set_bit(FSCACHE_COOKIE_NO_DATA_YET, &cookie->flags);
94d30ae9 526 clear_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags);
36c95590
DH
527
528 _debug("wake up lookup %p", &cookie->flags);
caaef690 529 clear_bit_unlock(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags);
36c95590 530 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP);
36c95590 531 }
36c95590
DH
532 _leave("");
533}
534EXPORT_SYMBOL(fscache_object_lookup_negative);
535
536/**
537 * fscache_obtained_object - Note successful object lookup or creation
538 * @object: Object pointing to cookie to mark
539 *
540 * Note successful lookup and/or creation, permitting those waiting to write
541 * data to a backing object to continue.
542 *
543 * Note that after calling this, an object's cookie may be relinquished by the
544 * netfs, and so must be accessed with object lock held.
545 */
546void fscache_obtained_object(struct fscache_object *object)
547{
548 struct fscache_cookie *cookie = object->cookie;
549
caaef690 550 _enter("{OBJ%x,%s}", object->debug_id, object->state->name);
36c95590
DH
551
552 /* if we were still looking up, then we must have a positive lookup
553 * result, in which case there may be data available */
caaef690 554 if (!test_and_set_bit(FSCACHE_OBJECT_IS_LOOKED_UP, &object->flags)) {
36c95590
DH
555 fscache_stat(&fscache_n_object_lookups_positive);
556
caaef690
DH
557 /* We do (presumably) have data */
558 clear_bit_unlock(FSCACHE_COOKIE_NO_DATA_YET, &cookie->flags);
94d30ae9 559 clear_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags);
36c95590 560
caaef690
DH
561 /* Allow write requests to begin stacking up and read requests
562 * to begin shovelling data.
563 */
564 clear_bit_unlock(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags);
36c95590 565 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP);
36c95590 566 } else {
36c95590 567 fscache_stat(&fscache_n_object_created);
36c95590
DH
568 }
569
caaef690 570 set_bit(FSCACHE_OBJECT_IS_AVAILABLE, &object->flags);
36c95590
DH
571 _leave("");
572}
573EXPORT_SYMBOL(fscache_obtained_object);
574
575/*
576 * handle an object that has just become available
577 */
caaef690
DH
578static const struct fscache_state *fscache_object_available(struct fscache_object *object,
579 int event)
36c95590 580{
caaef690
DH
581 _enter("{OBJ%x},%d", object->debug_id, event);
582
583 object->oob_table = fscache_osm_run_oob;
36c95590
DH
584
585 spin_lock(&object->lock);
586
36c95590
DH
587 fscache_done_parent_op(object);
588 if (object->n_in_progress == 0) {
589 if (object->n_ops > 0) {
590 ASSERTCMP(object->n_ops, >=, object->n_obj_ops);
36c95590
DH
591 fscache_start_operations(object);
592 } else {
593 ASSERT(list_empty(&object->pending_ops));
594 }
595 }
596 spin_unlock(&object->lock);
597
52bd75fd 598 fscache_stat(&fscache_n_cop_lookup_complete);
36c95590 599 object->cache->ops->lookup_complete(object);
52bd75fd 600 fscache_stat_d(&fscache_n_cop_lookup_complete);
36c95590
DH
601
602 fscache_hist(fscache_obj_instantiate_histogram, object->lookup_jif);
603 fscache_stat(&fscache_n_object_avail);
604
605 _leave("");
caaef690 606 return transit_to(JUMPSTART_DEPS);
36c95590
DH
607}
608
609/*
caaef690 610 * Wake up this object's dependent objects now that we've become available.
36c95590 611 */
caaef690
DH
612static const struct fscache_state *fscache_jumpstart_dependents(struct fscache_object *object,
613 int event)
36c95590 614{
caaef690 615 _enter("{OBJ%x},%d", object->debug_id, event);
36c95590 616
caaef690
DH
617 if (!fscache_enqueue_dependents(object, FSCACHE_OBJECT_EV_PARENT_READY))
618 return NO_TRANSIT; /* Not finished; requeue */
619 return transit_to(WAIT_FOR_CMD);
620}
36c95590 621
caaef690
DH
622/*
623 * Handle lookup or creation failute.
624 */
625static const struct fscache_state *fscache_lookup_failure(struct fscache_object *object,
626 int event)
627{
628 struct fscache_cookie *cookie;
6897e3df 629
caaef690 630 _enter("{OBJ%x},%d", object->debug_id, event);
36c95590 631
caaef690 632 object->oob_event_mask = 0;
36c95590 633
caaef690
DH
634 fscache_stat(&fscache_n_cop_lookup_complete);
635 object->cache->ops->lookup_complete(object);
636 fscache_stat_d(&fscache_n_cop_lookup_complete);
36c95590 637
6515d1db
DH
638 set_bit(FSCACHE_OBJECT_KILLED_BY_CACHE, &object->flags);
639
caaef690
DH
640 cookie = object->cookie;
641 set_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags);
1362729b 642 if (test_and_clear_bit(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags))
caaef690 643 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP);
36c95590 644
caaef690
DH
645 fscache_done_parent_op(object);
646 return transit_to(KILL_OBJECT);
647}
648
649/*
650 * Wait for completion of all active operations on this object and the death of
651 * all child objects of this object.
652 */
653static const struct fscache_state *fscache_kill_object(struct fscache_object *object,
654 int event)
655{
656 _enter("{OBJ%x,%d,%d},%d",
657 object->debug_id, object->n_ops, object->n_children, event);
658
f09b443d 659 fscache_mark_object_dead(object);
1362729b 660 object->oob_event_mask = 0;
caaef690 661
6bdded59
DH
662 if (test_bit(FSCACHE_OBJECT_RETIRED, &object->flags)) {
663 /* Reject any new read/write ops and abort any that are pending. */
664 clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags);
665 fscache_cancel_all_ops(object);
666 }
667
caaef690
DH
668 if (list_empty(&object->dependents) &&
669 object->n_ops == 0 &&
670 object->n_children == 0)
1362729b 671 return transit_to(DROP_OBJECT);
caaef690 672
1362729b
DH
673 if (object->n_in_progress == 0) {
674 spin_lock(&object->lock);
675 if (object->n_ops > 0 && object->n_in_progress == 0)
676 fscache_start_operations(object);
677 spin_unlock(&object->lock);
678 }
caaef690
DH
679
680 if (!list_empty(&object->dependents))
681 return transit_to(KILL_DEPENDENTS);
682
683 return transit_to(WAIT_FOR_CLEARANCE);
36c95590
DH
684}
685
686/*
caaef690 687 * Kill dependent objects.
36c95590 688 */
caaef690
DH
689static const struct fscache_state *fscache_kill_dependents(struct fscache_object *object,
690 int event)
36c95590 691{
caaef690 692 _enter("{OBJ%x},%d", object->debug_id, event);
36c95590 693
caaef690
DH
694 if (!fscache_enqueue_dependents(object, FSCACHE_OBJECT_EV_KILL))
695 return NO_TRANSIT; /* Not finished */
696 return transit_to(WAIT_FOR_CLEARANCE);
36c95590
DH
697}
698
36c95590 699/*
caaef690
DH
700 * Drop an object's attachments
701 */
702static const struct fscache_state *fscache_drop_object(struct fscache_object *object,
703 int event)
36c95590 704{
caaef690 705 struct fscache_object *parent = object->parent;
1362729b 706 struct fscache_cookie *cookie = object->cookie;
caaef690 707 struct fscache_cache *cache = object->cache;
1362729b 708 bool awaken = false;
36c95590 709
caaef690 710 _enter("{OBJ%x,%d},%d", object->debug_id, object->n_children, event);
36c95590 711
1362729b
DH
712 ASSERT(cookie != NULL);
713 ASSERT(!hlist_unhashed(&object->cookie_link));
714
402cb8dd
DH
715 if (test_bit(FSCACHE_COOKIE_AUX_UPDATED, &cookie->flags)) {
716 _debug("final update");
717 fscache_update_aux_data(object);
718 }
719
1362729b
DH
720 /* Make sure the cookie no longer points here and that the netfs isn't
721 * waiting for us.
722 */
723 spin_lock(&cookie->lock);
724 hlist_del_init(&object->cookie_link);
94d30ae9
DH
725 if (hlist_empty(&cookie->backing_objects) &&
726 test_and_clear_bit(FSCACHE_COOKIE_INVALIDATING, &cookie->flags))
1362729b
DH
727 awaken = true;
728 spin_unlock(&cookie->lock);
729
730 if (awaken)
731 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_INVALIDATING);
caaef690
DH
732
733 /* Prevent a race with our last child, which has to signal EV_CLEARED
734 * before dropping our spinlock.
735 */
36c95590 736 spin_lock(&object->lock);
36c95590
DH
737 spin_unlock(&object->lock);
738
caaef690
DH
739 /* Discard from the cache's collection of objects */
740 spin_lock(&cache->object_list_lock);
741 list_del_init(&object->cache_link);
742 spin_unlock(&cache->object_list_lock);
743
744 fscache_stat(&fscache_n_cop_drop_object);
745 cache->ops->drop_object(object);
746 fscache_stat_d(&fscache_n_cop_drop_object);
747
748 /* The parent object wants to know when all it dependents have gone */
749 if (parent) {
750 _debug("release parent OBJ%x {%d}",
751 parent->debug_id, parent->n_children);
752
753 spin_lock(&parent->lock);
754 parent->n_children--;
755 if (parent->n_children == 0)
756 fscache_raise_event(parent, FSCACHE_OBJECT_EV_CLEARED);
757 spin_unlock(&parent->lock);
758 object->parent = NULL;
759 }
760
761 /* this just shifts the object release to the work processor */
a18feb55 762 fscache_put_object(object, fscache_obj_put_drop_obj);
caaef690 763 fscache_stat(&fscache_n_object_dead);
36c95590
DH
764
765 _leave("");
caaef690 766 return transit_to(OBJECT_DEAD);
36c95590
DH
767}
768
769/*
8b8edefa 770 * get a ref on an object
36c95590 771 */
a18feb55
DH
772static int fscache_get_object(struct fscache_object *object,
773 enum fscache_obj_ref_trace why)
36c95590 774{
52bd75fd 775 int ret;
36c95590 776
52bd75fd 777 fscache_stat(&fscache_n_cop_grab_object);
a18feb55 778 ret = object->cache->ops->grab_object(object, why) ? 0 : -EAGAIN;
52bd75fd
DH
779 fscache_stat_d(&fscache_n_cop_grab_object);
780 return ret;
36c95590
DH
781}
782
783/*
caaef690 784 * Discard a ref on an object
36c95590 785 */
a18feb55
DH
786static void fscache_put_object(struct fscache_object *object,
787 enum fscache_obj_ref_trace why)
36c95590 788{
52bd75fd 789 fscache_stat(&fscache_n_cop_put_object);
a18feb55 790 object->cache->ops->put_object(object, why);
52bd75fd 791 fscache_stat_d(&fscache_n_cop_put_object);
36c95590
DH
792}
793
1362729b
DH
794/**
795 * fscache_object_destroy - Note that a cache object is about to be destroyed
796 * @object: The object to be destroyed
797 *
798 * Note the imminent destruction and deallocation of a cache object record.
799 */
800void fscache_object_destroy(struct fscache_object *object)
801{
802 fscache_objlist_remove(object);
803
804 /* We can get rid of the cookie now */
a18feb55 805 fscache_cookie_put(object->cookie, fscache_cookie_put_object);
1362729b
DH
806 object->cookie = NULL;
807}
808EXPORT_SYMBOL(fscache_object_destroy);
809
36c95590
DH
810/*
811 * enqueue an object for metadata-type processing
812 */
813void fscache_enqueue_object(struct fscache_object *object)
814{
815 _enter("{OBJ%x}", object->debug_id);
816
a18feb55 817 if (fscache_get_object(object, fscache_obj_get_queue) >= 0) {
8b8edefa
TH
818 wait_queue_head_t *cong_wq =
819 &get_cpu_var(fscache_object_cong_wait);
820
821 if (queue_work(fscache_object_wq, &object->work)) {
822 if (fscache_object_congested())
823 wake_up(cong_wq);
824 } else
a18feb55 825 fscache_put_object(object, fscache_obj_put_queue);
8b8edefa
TH
826
827 put_cpu_var(fscache_object_cong_wait);
828 }
829}
830
831/**
832 * fscache_object_sleep_till_congested - Sleep until object wq is congested
caaef690 833 * @timeoutp: Scheduler sleep timeout
8b8edefa
TH
834 *
835 * Allow an object handler to sleep until the object workqueue is congested.
836 *
837 * The caller must set up a wake up event before calling this and must have set
838 * the appropriate sleep mode (such as TASK_UNINTERRUPTIBLE) and tested its own
839 * condition before calling this function as no test is made here.
840 *
841 * %true is returned if the object wq is congested, %false otherwise.
842 */
843bool fscache_object_sleep_till_congested(signed long *timeoutp)
844{
170d800a 845 wait_queue_head_t *cong_wq = this_cpu_ptr(&fscache_object_cong_wait);
8b8edefa
TH
846 DEFINE_WAIT(wait);
847
848 if (fscache_object_congested())
849 return true;
850
851 add_wait_queue_exclusive(cong_wq, &wait);
852 if (!fscache_object_congested())
853 *timeoutp = schedule_timeout(*timeoutp);
854 finish_wait(cong_wq, &wait);
855
856 return fscache_object_congested();
36c95590 857}
8b8edefa 858EXPORT_SYMBOL_GPL(fscache_object_sleep_till_congested);
36c95590
DH
859
860/*
caaef690
DH
861 * Enqueue the dependents of an object for metadata-type processing.
862 *
863 * If we don't manage to finish the list before the scheduler wants to run
864 * again then return false immediately. We return true if the list was
865 * cleared.
36c95590 866 */
caaef690 867static bool fscache_enqueue_dependents(struct fscache_object *object, int event)
36c95590
DH
868{
869 struct fscache_object *dep;
caaef690 870 bool ret = true;
36c95590
DH
871
872 _enter("{OBJ%x}", object->debug_id);
873
874 if (list_empty(&object->dependents))
caaef690 875 return true;
36c95590
DH
876
877 spin_lock(&object->lock);
878
879 while (!list_empty(&object->dependents)) {
880 dep = list_entry(object->dependents.next,
881 struct fscache_object, dep_link);
882 list_del_init(&dep->dep_link);
883
caaef690 884 fscache_raise_event(dep, event);
a18feb55 885 fscache_put_object(dep, fscache_obj_put_enq_dep);
36c95590 886
caaef690
DH
887 if (!list_empty(&object->dependents) && need_resched()) {
888 ret = false;
889 break;
890 }
36c95590
DH
891 }
892
893 spin_unlock(&object->lock);
caaef690 894 return ret;
36c95590
DH
895}
896
897/*
898 * remove an object from whatever queue it's waiting on
36c95590 899 */
caaef690 900static void fscache_dequeue_object(struct fscache_object *object)
36c95590
DH
901{
902 _enter("{OBJ%x}", object->debug_id);
903
904 if (!list_empty(&object->dep_link)) {
905 spin_lock(&object->parent->lock);
906 list_del_init(&object->dep_link);
907 spin_unlock(&object->parent->lock);
908 }
909
910 _leave("");
911}
912
913/**
914 * fscache_check_aux - Ask the netfs whether an object on disk is still valid
915 * @object: The object to ask about
916 * @data: The auxiliary data for the object
917 * @datalen: The size of the auxiliary data
918 *
1362729b
DH
919 * This function consults the netfs about the coherency state of an object.
920 * The caller must be holding a ref on cookie->n_active (held by
921 * fscache_look_up_object() on behalf of the cache backend during object lookup
922 * and creation).
36c95590
DH
923 */
924enum fscache_checkaux fscache_check_aux(struct fscache_object *object,
ee1235a9
DH
925 const void *data, uint16_t datalen,
926 loff_t object_size)
36c95590
DH
927{
928 enum fscache_checkaux result;
929
930 if (!object->cookie->def->check_aux) {
931 fscache_stat(&fscache_n_checkaux_none);
932 return FSCACHE_CHECKAUX_OKAY;
933 }
934
935 result = object->cookie->def->check_aux(object->cookie->netfs_data,
ee1235a9 936 data, datalen, object_size);
36c95590
DH
937 switch (result) {
938 /* entry okay as is */
939 case FSCACHE_CHECKAUX_OKAY:
940 fscache_stat(&fscache_n_checkaux_okay);
941 break;
942
943 /* entry requires update */
944 case FSCACHE_CHECKAUX_NEEDS_UPDATE:
945 fscache_stat(&fscache_n_checkaux_update);
946 break;
947
948 /* entry requires deletion */
949 case FSCACHE_CHECKAUX_OBSOLETE:
950 fscache_stat(&fscache_n_checkaux_obsolete);
951 break;
952
953 default:
954 BUG();
955 }
956
957 return result;
958}
959EXPORT_SYMBOL(fscache_check_aux);
ef778e7a
DH
960
961/*
962 * Asynchronously invalidate an object.
963 */
caaef690
DH
964static const struct fscache_state *_fscache_invalidate_object(struct fscache_object *object,
965 int event)
ef778e7a
DH
966{
967 struct fscache_operation *op;
968 struct fscache_cookie *cookie = object->cookie;
969
caaef690
DH
970 _enter("{OBJ%x},%d", object->debug_id, event);
971
1362729b
DH
972 /* We're going to need the cookie. If the cookie is not available then
973 * retire the object instead.
974 */
975 if (!fscache_use_cookie(object)) {
e5a95541 976 ASSERT(radix_tree_empty(&object->cookie->stores));
94d30ae9 977 set_bit(FSCACHE_OBJECT_RETIRED, &object->flags);
1362729b
DH
978 _leave(" [no cookie]");
979 return transit_to(KILL_OBJECT);
980 }
ef778e7a
DH
981
982 /* Reject any new read/write ops and abort any that are pending. */
983 fscache_invalidate_writes(cookie);
984 clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags);
985 fscache_cancel_all_ops(object);
986
987 /* Now we have to wait for in-progress reads and writes */
988 op = kzalloc(sizeof(*op), GFP_KERNEL);
1362729b
DH
989 if (!op)
990 goto nomem;
ef778e7a 991
08c2e3d0 992 fscache_operation_init(cookie, op, object->cache->ops->invalidate_object,
d3b97ca4 993 NULL, NULL);
1362729b
DH
994 op->flags = FSCACHE_OP_ASYNC |
995 (1 << FSCACHE_OP_EXCLUSIVE) |
996 (1 << FSCACHE_OP_UNUSE_COOKIE);
08c2e3d0 997 trace_fscache_page_op(cookie, NULL, op, fscache_page_op_invalidate);
ef778e7a
DH
998
999 spin_lock(&cookie->lock);
1000 if (fscache_submit_exclusive_op(object, op) < 0)
8d76349d 1001 goto submit_op_failed;
ef778e7a
DH
1002 spin_unlock(&cookie->lock);
1003 fscache_put_operation(op);
1004
1005 /* Once we've completed the invalidation, we know there will be no data
1006 * stored in the cache and thus we can reinstate the data-check-skip
1007 * optimisation.
1008 */
1009 set_bit(FSCACHE_COOKIE_NO_DATA_YET, &cookie->flags);
1010
1011 /* We can allow read and write requests to come in once again. They'll
1012 * queue up behind our exclusive invalidation operation.
1013 */
caaef690
DH
1014 if (test_and_clear_bit(FSCACHE_COOKIE_INVALIDATING, &cookie->flags))
1015 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_INVALIDATING);
1016 _leave(" [ok]");
1017 return transit_to(UPDATE_OBJECT);
8d76349d 1018
1362729b 1019nomem:
f09b443d 1020 fscache_mark_object_dead(object);
1362729b
DH
1021 fscache_unuse_cookie(object);
1022 _leave(" [ENOMEM]");
1023 return transit_to(KILL_OBJECT);
1024
8d76349d 1025submit_op_failed:
f09b443d 1026 fscache_mark_object_dead(object);
8d76349d 1027 spin_unlock(&cookie->lock);
920bce20 1028 fscache_unuse_cookie(object);
8d76349d 1029 kfree(op);
8d76349d 1030 _leave(" [EIO]");
caaef690
DH
1031 return transit_to(KILL_OBJECT);
1032}
1033
1034static const struct fscache_state *fscache_invalidate_object(struct fscache_object *object,
1035 int event)
1036{
1037 const struct fscache_state *s;
1038
1039 fscache_stat(&fscache_n_invalidates_run);
1040 fscache_stat(&fscache_n_cop_invalidate_object);
1041 s = _fscache_invalidate_object(object, event);
1042 fscache_stat_d(&fscache_n_cop_invalidate_object);
1043 return s;
1044}
1045
402cb8dd
DH
1046/*
1047 * Update auxiliary data.
1048 */
1049static void fscache_update_aux_data(struct fscache_object *object)
1050{
1051 fscache_stat(&fscache_n_updates_run);
1052 fscache_stat(&fscache_n_cop_update_object);
1053 object->cache->ops->update_object(object);
1054 fscache_stat_d(&fscache_n_cop_update_object);
1055}
1056
caaef690
DH
1057/*
1058 * Asynchronously update an object.
1059 */
1060static const struct fscache_state *fscache_update_object(struct fscache_object *object,
1061 int event)
1062{
1063 _enter("{OBJ%x},%d", object->debug_id, event);
1064
402cb8dd 1065 fscache_update_aux_data(object);
caaef690
DH
1066
1067 _leave("");
1068 return transit_to(WAIT_FOR_CMD);
ef778e7a 1069}
182d919b
DH
1070
1071/**
1072 * fscache_object_retrying_stale - Note retrying stale object
1073 * @object: The object that will be retried
1074 *
1075 * Note that an object lookup found an on-disk object that was adjudged to be
1076 * stale and has been deleted. The lookup will be retried.
1077 */
1078void fscache_object_retrying_stale(struct fscache_object *object)
1079{
1080 fscache_stat(&fscache_n_cache_no_space_reject);
1081}
1082EXPORT_SYMBOL(fscache_object_retrying_stale);
1083
1084/**
1085 * fscache_object_mark_killed - Note that an object was killed
1086 * @object: The object that was culled
1087 * @why: The reason the object was killed.
1088 *
1089 * Note that an object was killed. Returns true if the object was
1090 * already marked killed, false if it wasn't.
1091 */
1092void fscache_object_mark_killed(struct fscache_object *object,
1093 enum fscache_why_object_killed why)
1094{
1095 if (test_and_set_bit(FSCACHE_OBJECT_KILLED_BY_CACHE, &object->flags)) {
1096 pr_err("Error: Object already killed by cache [%s]\n",
1097 object->cache->identifier);
1098 return;
1099 }
1100
1101 switch (why) {
1102 case FSCACHE_OBJECT_NO_SPACE:
1103 fscache_stat(&fscache_n_cache_no_space_reject);
1104 break;
1105 case FSCACHE_OBJECT_IS_STALE:
1106 fscache_stat(&fscache_n_cache_stale_objects);
1107 break;
1108 case FSCACHE_OBJECT_WAS_RETIRED:
1109 fscache_stat(&fscache_n_cache_retired_objects);
1110 break;
1111 case FSCACHE_OBJECT_WAS_CULLED:
1112 fscache_stat(&fscache_n_cache_culled_objects);
1113 break;
1114 }
1115}
1116EXPORT_SYMBOL(fscache_object_mark_killed);
e26bfebd
DH
1117
1118/*
1119 * The object is dead. We can get here if an object gets queued by an event
1120 * that would lead to its death (such as EV_KILL) when the dispatcher is
1121 * already running (and so can be requeued) but hasn't yet cleared the event
1122 * mask.
1123 */
1124static const struct fscache_state *fscache_object_dead(struct fscache_object *object,
1125 int event)
1126{
1127 if (!test_and_set_bit(FSCACHE_OBJECT_RUN_AFTER_DEAD,
1128 &object->flags))
1129 return NO_TRANSIT;
1130
1131 WARN(true, "FS-Cache object redispatched after death");
1132 return NO_TRANSIT;
1133}