blob: 9edc920f651f3929f9ad4f27e061df728c424794 [file] [log] [blame]
David Howells36c955902009-04-03 16:42:38 +01001/* FS-Cache object state machine handler
2 *
3 * Copyright (C) 2007 Red Hat, Inc. All Rights Reserved.
4 * Written by David Howells (dhowells@redhat.com)
5 *
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
10 *
11 * See Documentation/filesystems/caching/object.txt for a description of the
12 * object state machine and the in-kernel representations.
13 */
14
15#define FSCACHE_DEBUG_LEVEL COOKIE
16#include <linux/module.h>
David Howellsef778e72012-12-20 21:52:36 +000017#include <linux/slab.h>
David Howellscaaef692013-05-10 19:50:26 +010018#include <linux/prefetch.h>
David Howells36c955902009-04-03 16:42:38 +010019#include "internal.h"
20
David Howellscaaef692013-05-10 19:50:26 +010021static const struct fscache_state *fscache_abort_initialisation(struct fscache_object *, int);
22static const struct fscache_state *fscache_kill_dependents(struct fscache_object *, int);
23static const struct fscache_state *fscache_drop_object(struct fscache_object *, int);
24static const struct fscache_state *fscache_initialise_object(struct fscache_object *, int);
25static const struct fscache_state *fscache_invalidate_object(struct fscache_object *, int);
26static const struct fscache_state *fscache_jumpstart_dependents(struct fscache_object *, int);
27static const struct fscache_state *fscache_kill_object(struct fscache_object *, int);
28static const struct fscache_state *fscache_lookup_failure(struct fscache_object *, int);
29static const struct fscache_state *fscache_look_up_object(struct fscache_object *, int);
30static const struct fscache_state *fscache_object_available(struct fscache_object *, int);
31static const struct fscache_state *fscache_parent_ready(struct fscache_object *, int);
32static const struct fscache_state *fscache_update_object(struct fscache_object *, int);
David Howellse26bfeb2017-01-31 09:45:28 +000033static const struct fscache_state *fscache_object_dead(struct fscache_object *, int);
David Howells36c955902009-04-03 16:42:38 +010034
David Howellscaaef692013-05-10 19:50:26 +010035#define __STATE_NAME(n) fscache_osm_##n
36#define STATE(n) (&__STATE_NAME(n))
37
38/*
39 * Define a work state. Work states are execution states. No event processing
40 * is performed by them. The function attached to a work state returns a
41 * pointer indicating the next state to which the state machine should
42 * transition. Returning NO_TRANSIT repeats the current state, but goes back
43 * to the scheduler first.
44 */
45#define WORK_STATE(n, sn, f) \
46 const struct fscache_state __STATE_NAME(n) = { \
47 .name = #n, \
48 .short_name = sn, \
49 .work = f \
50 }
51
52/*
53 * Returns from work states.
54 */
55#define transit_to(state) ({ prefetch(&STATE(state)->work); STATE(state); })
56
57#define NO_TRANSIT ((struct fscache_state *)NULL)
58
59/*
60 * Define a wait state. Wait states are event processing states. No execution
61 * is performed by them. Wait states are just tables of "if event X occurs,
62 * clear it and transition to state Y". The dispatcher returns to the
63 * scheduler if none of the events in which the wait state has an interest are
64 * currently pending.
65 */
66#define WAIT_STATE(n, sn, ...) \
67 const struct fscache_state __STATE_NAME(n) = { \
68 .name = #n, \
69 .short_name = sn, \
70 .work = NULL, \
71 .transitions = { __VA_ARGS__, { 0, NULL } } \
72 }
73
74#define TRANSIT_TO(state, emask) \
75 { .events = (emask), .transit_to = STATE(state) }
76
77/*
78 * The object state machine.
79 */
80static WORK_STATE(INIT_OBJECT, "INIT", fscache_initialise_object);
81static WORK_STATE(PARENT_READY, "PRDY", fscache_parent_ready);
82static WORK_STATE(ABORT_INIT, "ABRT", fscache_abort_initialisation);
83static WORK_STATE(LOOK_UP_OBJECT, "LOOK", fscache_look_up_object);
84static WORK_STATE(CREATE_OBJECT, "CRTO", fscache_look_up_object);
85static WORK_STATE(OBJECT_AVAILABLE, "AVBL", fscache_object_available);
86static WORK_STATE(JUMPSTART_DEPS, "JUMP", fscache_jumpstart_dependents);
87
88static WORK_STATE(INVALIDATE_OBJECT, "INVL", fscache_invalidate_object);
89static WORK_STATE(UPDATE_OBJECT, "UPDT", fscache_update_object);
90
91static WORK_STATE(LOOKUP_FAILURE, "LCFL", fscache_lookup_failure);
92static WORK_STATE(KILL_OBJECT, "KILL", fscache_kill_object);
93static WORK_STATE(KILL_DEPENDENTS, "KDEP", fscache_kill_dependents);
94static WORK_STATE(DROP_OBJECT, "DROP", fscache_drop_object);
David Howellse26bfeb2017-01-31 09:45:28 +000095static WORK_STATE(OBJECT_DEAD, "DEAD", fscache_object_dead);
David Howellscaaef692013-05-10 19:50:26 +010096
97static WAIT_STATE(WAIT_FOR_INIT, "?INI",
98 TRANSIT_TO(INIT_OBJECT, 1 << FSCACHE_OBJECT_EV_NEW_CHILD));
99
100static WAIT_STATE(WAIT_FOR_PARENT, "?PRN",
101 TRANSIT_TO(PARENT_READY, 1 << FSCACHE_OBJECT_EV_PARENT_READY));
102
103static WAIT_STATE(WAIT_FOR_CMD, "?CMD",
104 TRANSIT_TO(INVALIDATE_OBJECT, 1 << FSCACHE_OBJECT_EV_INVALIDATE),
105 TRANSIT_TO(UPDATE_OBJECT, 1 << FSCACHE_OBJECT_EV_UPDATE),
106 TRANSIT_TO(JUMPSTART_DEPS, 1 << FSCACHE_OBJECT_EV_NEW_CHILD));
107
108static WAIT_STATE(WAIT_FOR_CLEARANCE, "?CLR",
109 TRANSIT_TO(KILL_OBJECT, 1 << FSCACHE_OBJECT_EV_CLEARED));
110
111/*
112 * Out-of-band event transition tables. These are for handling unexpected
113 * events, such as an I/O error. If an OOB event occurs, the state machine
114 * clears and disables the event and forces a transition to the nominated work
115 * state (acurrently executing work states will complete first).
116 *
117 * In such a situation, object->state remembers the state the machine should
118 * have been in/gone to and returning NO_TRANSIT returns to that.
119 */
120static const struct fscache_transition fscache_osm_init_oob[] = {
121 TRANSIT_TO(ABORT_INIT,
122 (1 << FSCACHE_OBJECT_EV_ERROR) |
123 (1 << FSCACHE_OBJECT_EV_KILL)),
124 { 0, NULL }
125};
126
127static const struct fscache_transition fscache_osm_lookup_oob[] = {
128 TRANSIT_TO(LOOKUP_FAILURE,
129 (1 << FSCACHE_OBJECT_EV_ERROR) |
130 (1 << FSCACHE_OBJECT_EV_KILL)),
131 { 0, NULL }
132};
133
134static const struct fscache_transition fscache_osm_run_oob[] = {
135 TRANSIT_TO(KILL_OBJECT,
136 (1 << FSCACHE_OBJECT_EV_ERROR) |
137 (1 << FSCACHE_OBJECT_EV_KILL)),
138 { 0, NULL }
David Howells440f0af2009-11-19 18:11:01 +0000139};
140
David Howellsa18feb52018-04-04 13:41:27 +0100141static int fscache_get_object(struct fscache_object *,
142 enum fscache_obj_ref_trace);
143static void fscache_put_object(struct fscache_object *,
144 enum fscache_obj_ref_trace);
David Howellscaaef692013-05-10 19:50:26 +0100145static bool fscache_enqueue_dependents(struct fscache_object *, int);
David Howells36c955902009-04-03 16:42:38 +0100146static void fscache_dequeue_object(struct fscache_object *);
David Howells402cb8d2018-04-04 13:41:28 +0100147static void fscache_update_aux_data(struct fscache_object *);
David Howells36c955902009-04-03 16:42:38 +0100148
David Howells36c955902009-04-03 16:42:38 +0100149/*
150 * we need to notify the parent when an op completes that we had outstanding
151 * upon it
152 */
153static inline void fscache_done_parent_op(struct fscache_object *object)
154{
155 struct fscache_object *parent = object->parent;
156
157 _enter("OBJ%x {OBJ%x,%x}",
158 object->debug_id, parent->debug_id, parent->n_ops);
159
160 spin_lock_nested(&parent->lock, 1);
David Howells36c955902009-04-03 16:42:38 +0100161 parent->n_obj_ops--;
David Howells13627292013-05-10 19:50:26 +0100162 parent->n_ops--;
David Howells36c955902009-04-03 16:42:38 +0100163 if (parent->n_ops == 0)
164 fscache_raise_event(parent, FSCACHE_OBJECT_EV_CLEARED);
165 spin_unlock(&parent->lock);
166}
167
168/*
David Howellscaaef692013-05-10 19:50:26 +0100169 * Object state machine dispatcher.
David Howellsef778e72012-12-20 21:52:36 +0000170 */
David Howellscaaef692013-05-10 19:50:26 +0100171static void fscache_object_sm_dispatcher(struct fscache_object *object)
David Howellsef778e72012-12-20 21:52:36 +0000172{
David Howellscaaef692013-05-10 19:50:26 +0100173 const struct fscache_transition *t;
174 const struct fscache_state *state, *new_state;
175 unsigned long events, event_mask;
David Howellsa18feb52018-04-04 13:41:27 +0100176 bool oob;
David Howellscaaef692013-05-10 19:50:26 +0100177 int event = -1;
David Howells36c955902009-04-03 16:42:38 +0100178
179 ASSERT(object != NULL);
180
181 _enter("{OBJ%x,%s,%lx}",
David Howellscaaef692013-05-10 19:50:26 +0100182 object->debug_id, object->state->name, object->events);
David Howells36c955902009-04-03 16:42:38 +0100183
David Howellscaaef692013-05-10 19:50:26 +0100184 event_mask = object->event_mask;
185restart:
186 object->event_mask = 0; /* Mask normal event handling */
187 state = object->state;
188restart_masked:
189 events = object->events;
David Howells36c955902009-04-03 16:42:38 +0100190
David Howellscaaef692013-05-10 19:50:26 +0100191 /* Handle any out-of-band events (typically an error) */
192 if (events & object->oob_event_mask) {
193 _debug("{OBJ%x} oob %lx",
194 object->debug_id, events & object->oob_event_mask);
David Howellsa18feb52018-04-04 13:41:27 +0100195 oob = true;
David Howellscaaef692013-05-10 19:50:26 +0100196 for (t = object->oob_table; t->events; t++) {
197 if (events & t->events) {
198 state = t->transit_to;
199 ASSERT(state->work != NULL);
200 event = fls(events & t->events) - 1;
201 __clear_bit(event, &object->oob_event_mask);
202 clear_bit(event, &object->events);
203 goto execute_work_state;
204 }
David Howellsd461d262009-11-19 18:11:41 +0000205 }
David Howellscaaef692013-05-10 19:50:26 +0100206 }
David Howellsa18feb52018-04-04 13:41:27 +0100207 oob = false;
David Howells36c955902009-04-03 16:42:38 +0100208
David Howellscaaef692013-05-10 19:50:26 +0100209 /* Wait states are just transition tables */
210 if (!state->work) {
211 if (events & event_mask) {
212 for (t = state->transitions; t->events; t++) {
213 if (events & t->events) {
214 new_state = t->transit_to;
215 event = fls(events & t->events) - 1;
David Howellsa18feb52018-04-04 13:41:27 +0100216 trace_fscache_osm(object, state,
217 true, false, event);
David Howellscaaef692013-05-10 19:50:26 +0100218 clear_bit(event, &object->events);
219 _debug("{OBJ%x} ev %d: %s -> %s",
220 object->debug_id, event,
221 state->name, new_state->name);
222 object->state = state = new_state;
223 goto execute_work_state;
224 }
225 }
David Howells36c955902009-04-03 16:42:38 +0100226
David Howellscaaef692013-05-10 19:50:26 +0100227 /* The event mask didn't include all the tabled bits */
228 BUG();
David Howells36c955902009-04-03 16:42:38 +0100229 }
David Howellscaaef692013-05-10 19:50:26 +0100230 /* Randomly woke up */
231 goto unmask_events;
David Howells36c955902009-04-03 16:42:38 +0100232 }
233
David Howellscaaef692013-05-10 19:50:26 +0100234execute_work_state:
235 _debug("{OBJ%x} exec %s", object->debug_id, state->name);
236
David Howellsa18feb52018-04-04 13:41:27 +0100237 trace_fscache_osm(object, state, false, oob, event);
David Howellscaaef692013-05-10 19:50:26 +0100238 new_state = state->work(object, event);
239 event = -1;
240 if (new_state == NO_TRANSIT) {
241 _debug("{OBJ%x} %s notrans", object->debug_id, state->name);
David Howellse26bfeb2017-01-31 09:45:28 +0000242 if (unlikely(state == STATE(OBJECT_DEAD))) {
243 _leave(" [dead]");
244 return;
245 }
David Howellscaaef692013-05-10 19:50:26 +0100246 fscache_enqueue_object(object);
247 event_mask = object->oob_event_mask;
248 goto unmask_events;
David Howells36c955902009-04-03 16:42:38 +0100249 }
250
David Howellscaaef692013-05-10 19:50:26 +0100251 _debug("{OBJ%x} %s -> %s",
252 object->debug_id, state->name, new_state->name);
253 object->state = state = new_state;
254
255 if (state->work) {
David Howellse26bfeb2017-01-31 09:45:28 +0000256 if (unlikely(state == STATE(OBJECT_DEAD))) {
David Howellscaaef692013-05-10 19:50:26 +0100257 _leave(" [dead]");
258 return;
259 }
260 goto restart_masked;
David Howells36c955902009-04-03 16:42:38 +0100261 }
262
David Howellscaaef692013-05-10 19:50:26 +0100263 /* Transited to wait state */
264 event_mask = object->oob_event_mask;
265 for (t = state->transitions; t->events; t++)
266 event_mask |= t->events;
David Howells36c955902009-04-03 16:42:38 +0100267
David Howellscaaef692013-05-10 19:50:26 +0100268unmask_events:
269 object->event_mask = event_mask;
270 smp_mb();
271 events = object->events;
272 if (events & event_mask)
273 goto restart;
274 _leave(" [msk %lx]", event_mask);
David Howells36c955902009-04-03 16:42:38 +0100275}
276
277/*
278 * execute an object
279 */
David Howells610be242013-05-10 19:50:25 +0100280static void fscache_object_work_func(struct work_struct *work)
David Howells36c955902009-04-03 16:42:38 +0100281{
282 struct fscache_object *object =
283 container_of(work, struct fscache_object, work);
284 unsigned long start;
285
286 _enter("{OBJ%x}", object->debug_id);
287
David Howells36c955902009-04-03 16:42:38 +0100288 start = jiffies;
David Howellscaaef692013-05-10 19:50:26 +0100289 fscache_object_sm_dispatcher(object);
David Howells36c955902009-04-03 16:42:38 +0100290 fscache_hist(fscache_objs_histogram, start);
David Howellsa18feb52018-04-04 13:41:27 +0100291 fscache_put_object(object, fscache_obj_put_work);
David Howells36c955902009-04-03 16:42:38 +0100292}
David Howells610be242013-05-10 19:50:25 +0100293
294/**
295 * fscache_object_init - Initialise a cache object description
296 * @object: Object description
297 * @cookie: Cookie object will be attached to
298 * @cache: Cache in which backing object will be found
299 *
300 * Initialise a cache object description to its basic values.
301 *
302 * See Documentation/filesystems/caching/backend-api.txt for a complete
303 * description.
304 */
305void fscache_object_init(struct fscache_object *object,
306 struct fscache_cookie *cookie,
307 struct fscache_cache *cache)
308{
David Howellscaaef692013-05-10 19:50:26 +0100309 const struct fscache_transition *t;
310
David Howells610be242013-05-10 19:50:25 +0100311 atomic_inc(&cache->object_count);
312
David Howellscaaef692013-05-10 19:50:26 +0100313 object->state = STATE(WAIT_FOR_INIT);
314 object->oob_table = fscache_osm_init_oob;
315 object->flags = 1 << FSCACHE_OBJECT_IS_LIVE;
David Howells610be242013-05-10 19:50:25 +0100316 spin_lock_init(&object->lock);
317 INIT_LIST_HEAD(&object->cache_link);
318 INIT_HLIST_NODE(&object->cookie_link);
319 INIT_WORK(&object->work, fscache_object_work_func);
320 INIT_LIST_HEAD(&object->dependents);
321 INIT_LIST_HEAD(&object->dep_link);
322 INIT_LIST_HEAD(&object->pending_ops);
323 object->n_children = 0;
324 object->n_ops = object->n_in_progress = object->n_exclusive = 0;
David Howellscaaef692013-05-10 19:50:26 +0100325 object->events = 0;
David Howells610be242013-05-10 19:50:25 +0100326 object->store_limit = 0;
327 object->store_limit_l = 0;
328 object->cache = cache;
329 object->cookie = cookie;
Kiran Kumar Modukurif29507c2018-06-21 13:31:44 -0700330 fscache_cookie_get(cookie, fscache_cookie_get_attach_object);
David Howells610be242013-05-10 19:50:25 +0100331 object->parent = NULL;
David Howells7026f192014-02-17 15:01:47 +0000332#ifdef CONFIG_FSCACHE_OBJECT_LIST
333 RB_CLEAR_NODE(&object->objlist_link);
334#endif
David Howellscaaef692013-05-10 19:50:26 +0100335
336 object->oob_event_mask = 0;
337 for (t = object->oob_table; t->events; t++)
338 object->oob_event_mask |= t->events;
339 object->event_mask = object->oob_event_mask;
340 for (t = object->state->transitions; t->events; t++)
341 object->event_mask |= t->events;
David Howells610be242013-05-10 19:50:25 +0100342}
343EXPORT_SYMBOL(fscache_object_init);
David Howells440f0af2009-11-19 18:11:01 +0000344
345/*
David Howellsf09b4432015-02-24 10:05:28 +0000346 * Mark the object as no longer being live, making sure that we synchronise
347 * against op submission.
348 */
349static inline void fscache_mark_object_dead(struct fscache_object *object)
350{
351 spin_lock(&object->lock);
352 clear_bit(FSCACHE_OBJECT_IS_LIVE, &object->flags);
353 spin_unlock(&object->lock);
354}
355
356/*
David Howellscaaef692013-05-10 19:50:26 +0100357 * Abort object initialisation before we start it.
358 */
359static const struct fscache_state *fscache_abort_initialisation(struct fscache_object *object,
360 int event)
361{
David Howellscaaef692013-05-10 19:50:26 +0100362 _enter("{OBJ%x},%d", object->debug_id, event);
363
364 object->oob_event_mask = 0;
David Howellscaaef692013-05-10 19:50:26 +0100365 fscache_dequeue_object(object);
David Howellscaaef692013-05-10 19:50:26 +0100366 return transit_to(KILL_OBJECT);
367}
368
369/*
David Howells36c955902009-04-03 16:42:38 +0100370 * initialise an object
371 * - check the specified object's parent to see if we can make use of it
372 * immediately to do a creation
373 * - we may need to start the process of creating a parent and we need to wait
374 * for the parent's lookup and creation to complete if it's not there yet
David Howells36c955902009-04-03 16:42:38 +0100375 */
David Howellscaaef692013-05-10 19:50:26 +0100376static const struct fscache_state *fscache_initialise_object(struct fscache_object *object,
377 int event)
David Howells36c955902009-04-03 16:42:38 +0100378{
379 struct fscache_object *parent;
David Howellscaaef692013-05-10 19:50:26 +0100380 bool success;
David Howells36c955902009-04-03 16:42:38 +0100381
David Howellscaaef692013-05-10 19:50:26 +0100382 _enter("{OBJ%x},%d", object->debug_id, event);
David Howells36c955902009-04-03 16:42:38 +0100383
David Howellscaaef692013-05-10 19:50:26 +0100384 ASSERT(list_empty(&object->dep_link));
David Howells36c955902009-04-03 16:42:38 +0100385
386 parent = object->parent;
387 if (!parent) {
David Howellscaaef692013-05-10 19:50:26 +0100388 _leave(" [no parent]");
David Howells13627292013-05-10 19:50:26 +0100389 return transit_to(DROP_OBJECT);
David Howells36c955902009-04-03 16:42:38 +0100390 }
391
David Howells13627292013-05-10 19:50:26 +0100392 _debug("parent: %s of:%lx", parent->state->name, parent->flags);
David Howellscaaef692013-05-10 19:50:26 +0100393
394 if (fscache_object_is_dying(parent)) {
395 _leave(" [bad parent]");
David Howells13627292013-05-10 19:50:26 +0100396 return transit_to(DROP_OBJECT);
David Howellscaaef692013-05-10 19:50:26 +0100397 }
398
399 if (fscache_object_is_available(parent)) {
400 _leave(" [ready]");
401 return transit_to(PARENT_READY);
402 }
403
404 _debug("wait");
405
406 spin_lock(&parent->lock);
407 fscache_stat(&fscache_n_cop_grab_object);
408 success = false;
409 if (fscache_object_is_live(parent) &&
David Howellsa18feb52018-04-04 13:41:27 +0100410 object->cache->ops->grab_object(object, fscache_obj_get_add_to_deps)) {
David Howellscaaef692013-05-10 19:50:26 +0100411 list_add(&object->dep_link, &parent->dependents);
412 success = true;
413 }
414 fscache_stat_d(&fscache_n_cop_grab_object);
415 spin_unlock(&parent->lock);
416 if (!success) {
417 _leave(" [grab failed]");
David Howells13627292013-05-10 19:50:26 +0100418 return transit_to(DROP_OBJECT);
David Howellscaaef692013-05-10 19:50:26 +0100419 }
420
421 /* fscache_acquire_non_index_cookie() uses this
422 * to wake the chain up */
423 fscache_raise_event(parent, FSCACHE_OBJECT_EV_NEW_CHILD);
424 _leave(" [wait]");
425 return transit_to(WAIT_FOR_PARENT);
426}
427
428/*
429 * Once the parent object is ready, we should kick off our lookup op.
430 */
431static const struct fscache_state *fscache_parent_ready(struct fscache_object *object,
432 int event)
433{
434 struct fscache_object *parent = object->parent;
435
436 _enter("{OBJ%x},%d", object->debug_id, event);
437
438 ASSERT(parent != NULL);
439
440 spin_lock(&parent->lock);
441 parent->n_ops++;
442 parent->n_obj_ops++;
443 object->lookup_jif = jiffies;
444 spin_unlock(&parent->lock);
445
David Howells36c955902009-04-03 16:42:38 +0100446 _leave("");
David Howellscaaef692013-05-10 19:50:26 +0100447 return transit_to(LOOK_UP_OBJECT);
David Howells36c955902009-04-03 16:42:38 +0100448}
449
450/*
451 * look an object up in the cache from which it was allocated
452 * - we hold an "access lock" on the parent object, so the parent object cannot
453 * be withdrawn by either party till we've finished
David Howells36c955902009-04-03 16:42:38 +0100454 */
David Howellscaaef692013-05-10 19:50:26 +0100455static const struct fscache_state *fscache_look_up_object(struct fscache_object *object,
456 int event)
David Howells36c955902009-04-03 16:42:38 +0100457{
458 struct fscache_cookie *cookie = object->cookie;
David Howellscaaef692013-05-10 19:50:26 +0100459 struct fscache_object *parent = object->parent;
David Howellsfee096d2009-11-19 18:12:05 +0000460 int ret;
David Howells36c955902009-04-03 16:42:38 +0100461
David Howellscaaef692013-05-10 19:50:26 +0100462 _enter("{OBJ%x},%d", object->debug_id, event);
David Howells36c955902009-04-03 16:42:38 +0100463
David Howellscaaef692013-05-10 19:50:26 +0100464 object->oob_table = fscache_osm_lookup_oob;
465
David Howells36c955902009-04-03 16:42:38 +0100466 ASSERT(parent != NULL);
467 ASSERTCMP(parent->n_ops, >, 0);
468 ASSERTCMP(parent->n_obj_ops, >, 0);
469
470 /* make sure the parent is still available */
David Howells493f7bc2013-05-10 19:50:26 +0100471 ASSERT(fscache_object_is_available(parent));
David Howells36c955902009-04-03 16:42:38 +0100472
David Howells493f7bc2013-05-10 19:50:26 +0100473 if (fscache_object_is_dying(parent) ||
David Howells13627292013-05-10 19:50:26 +0100474 test_bit(FSCACHE_IOERROR, &object->cache->flags) ||
475 !fscache_use_cookie(object)) {
David Howellscaaef692013-05-10 19:50:26 +0100476 _leave(" [unavailable]");
477 return transit_to(LOOKUP_FAILURE);
David Howells36c955902009-04-03 16:42:38 +0100478 }
479
David Howells13627292013-05-10 19:50:26 +0100480 _debug("LOOKUP \"%s\" in \"%s\"",
481 cookie->def->name, object->cache->tag->name);
David Howells36c955902009-04-03 16:42:38 +0100482
483 fscache_stat(&fscache_n_object_lookups);
David Howells52bd75f2009-11-19 18:11:08 +0000484 fscache_stat(&fscache_n_cop_lookup_object);
David Howellsfee096d2009-11-19 18:12:05 +0000485 ret = object->cache->ops->lookup_object(object);
David Howells52bd75f2009-11-19 18:11:08 +0000486 fscache_stat_d(&fscache_n_cop_lookup_object);
David Howells36c955902009-04-03 16:42:38 +0100487
David Howells13627292013-05-10 19:50:26 +0100488 fscache_unuse_cookie(object);
David Howells36c955902009-04-03 16:42:38 +0100489
David Howellsfee096d2009-11-19 18:12:05 +0000490 if (ret == -ETIMEDOUT) {
491 /* probably stuck behind another object, so move this one to
492 * the back of the queue */
493 fscache_stat(&fscache_n_object_lookups_timed_out);
David Howellscaaef692013-05-10 19:50:26 +0100494 _leave(" [timeout]");
495 return NO_TRANSIT;
David Howellsfee096d2009-11-19 18:12:05 +0000496 }
497
David Howellscaaef692013-05-10 19:50:26 +0100498 if (ret < 0) {
499 _leave(" [error]");
500 return transit_to(LOOKUP_FAILURE);
501 }
502
503 _leave(" [ok]");
504 return transit_to(OBJECT_AVAILABLE);
David Howells36c955902009-04-03 16:42:38 +0100505}
506
507/**
508 * fscache_object_lookup_negative - Note negative cookie lookup
509 * @object: Object pointing to cookie to mark
510 *
511 * Note negative lookup, permitting those waiting to read data from an already
512 * existing backing object to continue as there's no data for them to read.
513 */
514void fscache_object_lookup_negative(struct fscache_object *object)
515{
516 struct fscache_cookie *cookie = object->cookie;
517
David Howellscaaef692013-05-10 19:50:26 +0100518 _enter("{OBJ%x,%s}", object->debug_id, object->state->name);
David Howells36c955902009-04-03 16:42:38 +0100519
David Howellscaaef692013-05-10 19:50:26 +0100520 if (!test_and_set_bit(FSCACHE_OBJECT_IS_LOOKED_UP, &object->flags)) {
David Howells36c955902009-04-03 16:42:38 +0100521 fscache_stat(&fscache_n_object_lookups_negative);
522
David Howellscaaef692013-05-10 19:50:26 +0100523 /* Allow write requests to begin stacking up and read requests to begin
524 * returning ENODATA.
525 */
David Howells36c955902009-04-03 16:42:38 +0100526 set_bit(FSCACHE_COOKIE_NO_DATA_YET, &cookie->flags);
David Howells94d30ae2013-09-21 00:09:31 +0100527 clear_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags);
David Howells36c955902009-04-03 16:42:38 +0100528
529 _debug("wake up lookup %p", &cookie->flags);
David Howellscaaef692013-05-10 19:50:26 +0100530 clear_bit_unlock(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags);
David Howells36c955902009-04-03 16:42:38 +0100531 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP);
David Howells36c955902009-04-03 16:42:38 +0100532 }
David Howells36c955902009-04-03 16:42:38 +0100533 _leave("");
534}
535EXPORT_SYMBOL(fscache_object_lookup_negative);
536
537/**
538 * fscache_obtained_object - Note successful object lookup or creation
539 * @object: Object pointing to cookie to mark
540 *
541 * Note successful lookup and/or creation, permitting those waiting to write
542 * data to a backing object to continue.
543 *
544 * Note that after calling this, an object's cookie may be relinquished by the
545 * netfs, and so must be accessed with object lock held.
546 */
547void fscache_obtained_object(struct fscache_object *object)
548{
549 struct fscache_cookie *cookie = object->cookie;
550
David Howellscaaef692013-05-10 19:50:26 +0100551 _enter("{OBJ%x,%s}", object->debug_id, object->state->name);
David Howells36c955902009-04-03 16:42:38 +0100552
553 /* if we were still looking up, then we must have a positive lookup
554 * result, in which case there may be data available */
David Howellscaaef692013-05-10 19:50:26 +0100555 if (!test_and_set_bit(FSCACHE_OBJECT_IS_LOOKED_UP, &object->flags)) {
David Howells36c955902009-04-03 16:42:38 +0100556 fscache_stat(&fscache_n_object_lookups_positive);
557
David Howellscaaef692013-05-10 19:50:26 +0100558 /* We do (presumably) have data */
559 clear_bit_unlock(FSCACHE_COOKIE_NO_DATA_YET, &cookie->flags);
David Howells94d30ae2013-09-21 00:09:31 +0100560 clear_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags);
David Howells36c955902009-04-03 16:42:38 +0100561
David Howellscaaef692013-05-10 19:50:26 +0100562 /* Allow write requests to begin stacking up and read requests
563 * to begin shovelling data.
564 */
565 clear_bit_unlock(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags);
David Howells36c955902009-04-03 16:42:38 +0100566 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP);
David Howells36c955902009-04-03 16:42:38 +0100567 } else {
David Howells36c955902009-04-03 16:42:38 +0100568 fscache_stat(&fscache_n_object_created);
David Howells36c955902009-04-03 16:42:38 +0100569 }
570
David Howellscaaef692013-05-10 19:50:26 +0100571 set_bit(FSCACHE_OBJECT_IS_AVAILABLE, &object->flags);
David Howells36c955902009-04-03 16:42:38 +0100572 _leave("");
573}
574EXPORT_SYMBOL(fscache_obtained_object);
575
576/*
577 * handle an object that has just become available
578 */
David Howellscaaef692013-05-10 19:50:26 +0100579static const struct fscache_state *fscache_object_available(struct fscache_object *object,
580 int event)
David Howells36c955902009-04-03 16:42:38 +0100581{
David Howellscaaef692013-05-10 19:50:26 +0100582 _enter("{OBJ%x},%d", object->debug_id, event);
583
584 object->oob_table = fscache_osm_run_oob;
David Howells36c955902009-04-03 16:42:38 +0100585
586 spin_lock(&object->lock);
587
David Howells36c955902009-04-03 16:42:38 +0100588 fscache_done_parent_op(object);
589 if (object->n_in_progress == 0) {
590 if (object->n_ops > 0) {
591 ASSERTCMP(object->n_ops, >=, object->n_obj_ops);
David Howells36c955902009-04-03 16:42:38 +0100592 fscache_start_operations(object);
593 } else {
594 ASSERT(list_empty(&object->pending_ops));
595 }
596 }
597 spin_unlock(&object->lock);
598
David Howells52bd75f2009-11-19 18:11:08 +0000599 fscache_stat(&fscache_n_cop_lookup_complete);
David Howells36c955902009-04-03 16:42:38 +0100600 object->cache->ops->lookup_complete(object);
David Howells52bd75f2009-11-19 18:11:08 +0000601 fscache_stat_d(&fscache_n_cop_lookup_complete);
David Howells36c955902009-04-03 16:42:38 +0100602
603 fscache_hist(fscache_obj_instantiate_histogram, object->lookup_jif);
604 fscache_stat(&fscache_n_object_avail);
605
606 _leave("");
David Howellscaaef692013-05-10 19:50:26 +0100607 return transit_to(JUMPSTART_DEPS);
David Howells36c955902009-04-03 16:42:38 +0100608}
609
610/*
David Howellscaaef692013-05-10 19:50:26 +0100611 * Wake up this object's dependent objects now that we've become available.
David Howells36c955902009-04-03 16:42:38 +0100612 */
David Howellscaaef692013-05-10 19:50:26 +0100613static const struct fscache_state *fscache_jumpstart_dependents(struct fscache_object *object,
614 int event)
615{
616 _enter("{OBJ%x},%d", object->debug_id, event);
617
618 if (!fscache_enqueue_dependents(object, FSCACHE_OBJECT_EV_PARENT_READY))
619 return NO_TRANSIT; /* Not finished; requeue */
620 return transit_to(WAIT_FOR_CMD);
621}
622
623/*
624 * Handle lookup or creation failute.
625 */
626static const struct fscache_state *fscache_lookup_failure(struct fscache_object *object,
627 int event)
628{
629 struct fscache_cookie *cookie;
David Howellscaaef692013-05-10 19:50:26 +0100630
631 _enter("{OBJ%x},%d", object->debug_id, event);
632
633 object->oob_event_mask = 0;
634
635 fscache_stat(&fscache_n_cop_lookup_complete);
636 object->cache->ops->lookup_complete(object);
637 fscache_stat_d(&fscache_n_cop_lookup_complete);
638
David Howells6515d1d2015-02-25 11:53:57 +0000639 set_bit(FSCACHE_OBJECT_KILLED_BY_CACHE, &object->flags);
640
David Howellscaaef692013-05-10 19:50:26 +0100641 cookie = object->cookie;
642 set_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags);
David Howells13627292013-05-10 19:50:26 +0100643 if (test_and_clear_bit(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags))
David Howellscaaef692013-05-10 19:50:26 +0100644 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP);
David Howellscaaef692013-05-10 19:50:26 +0100645
646 fscache_done_parent_op(object);
647 return transit_to(KILL_OBJECT);
648}
649
650/*
651 * Wait for completion of all active operations on this object and the death of
652 * all child objects of this object.
653 */
654static const struct fscache_state *fscache_kill_object(struct fscache_object *object,
655 int event)
656{
657 _enter("{OBJ%x,%d,%d},%d",
658 object->debug_id, object->n_ops, object->n_children, event);
659
David Howellsf09b4432015-02-24 10:05:28 +0000660 fscache_mark_object_dead(object);
David Howells13627292013-05-10 19:50:26 +0100661 object->oob_event_mask = 0;
David Howellscaaef692013-05-10 19:50:26 +0100662
David Howells6bdded52017-01-18 14:29:25 +0000663 if (test_bit(FSCACHE_OBJECT_RETIRED, &object->flags)) {
664 /* Reject any new read/write ops and abort any that are pending. */
665 clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags);
666 fscache_cancel_all_ops(object);
667 }
668
David Howellscaaef692013-05-10 19:50:26 +0100669 if (list_empty(&object->dependents) &&
670 object->n_ops == 0 &&
671 object->n_children == 0)
David Howells13627292013-05-10 19:50:26 +0100672 return transit_to(DROP_OBJECT);
David Howellscaaef692013-05-10 19:50:26 +0100673
David Howells13627292013-05-10 19:50:26 +0100674 if (object->n_in_progress == 0) {
675 spin_lock(&object->lock);
676 if (object->n_ops > 0 && object->n_in_progress == 0)
677 fscache_start_operations(object);
678 spin_unlock(&object->lock);
679 }
David Howellscaaef692013-05-10 19:50:26 +0100680
681 if (!list_empty(&object->dependents))
682 return transit_to(KILL_DEPENDENTS);
683
684 return transit_to(WAIT_FOR_CLEARANCE);
685}
686
687/*
688 * Kill dependent objects.
689 */
690static const struct fscache_state *fscache_kill_dependents(struct fscache_object *object,
691 int event)
692{
693 _enter("{OBJ%x},%d", object->debug_id, event);
694
695 if (!fscache_enqueue_dependents(object, FSCACHE_OBJECT_EV_KILL))
696 return NO_TRANSIT; /* Not finished */
697 return transit_to(WAIT_FOR_CLEARANCE);
698}
699
700/*
David Howellscaaef692013-05-10 19:50:26 +0100701 * Drop an object's attachments
702 */
703static const struct fscache_state *fscache_drop_object(struct fscache_object *object,
704 int event)
David Howells36c955902009-04-03 16:42:38 +0100705{
706 struct fscache_object *parent = object->parent;
David Howells13627292013-05-10 19:50:26 +0100707 struct fscache_cookie *cookie = object->cookie;
David Howells36c955902009-04-03 16:42:38 +0100708 struct fscache_cache *cache = object->cache;
David Howells13627292013-05-10 19:50:26 +0100709 bool awaken = false;
David Howells36c955902009-04-03 16:42:38 +0100710
David Howellscaaef692013-05-10 19:50:26 +0100711 _enter("{OBJ%x,%d},%d", object->debug_id, object->n_children, event);
David Howells36c955902009-04-03 16:42:38 +0100712
David Howells13627292013-05-10 19:50:26 +0100713 ASSERT(cookie != NULL);
714 ASSERT(!hlist_unhashed(&object->cookie_link));
715
David Howells402cb8d2018-04-04 13:41:28 +0100716 if (test_bit(FSCACHE_COOKIE_AUX_UPDATED, &cookie->flags)) {
717 _debug("final update");
718 fscache_update_aux_data(object);
719 }
720
David Howells13627292013-05-10 19:50:26 +0100721 /* Make sure the cookie no longer points here and that the netfs isn't
722 * waiting for us.
723 */
724 spin_lock(&cookie->lock);
725 hlist_del_init(&object->cookie_link);
David Howells94d30ae2013-09-21 00:09:31 +0100726 if (hlist_empty(&cookie->backing_objects) &&
727 test_and_clear_bit(FSCACHE_COOKIE_INVALIDATING, &cookie->flags))
David Howells13627292013-05-10 19:50:26 +0100728 awaken = true;
729 spin_unlock(&cookie->lock);
730
731 if (awaken)
732 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_INVALIDATING);
David Howells6897e3d2009-11-19 18:11:22 +0000733
David Howellscaaef692013-05-10 19:50:26 +0100734 /* Prevent a race with our last child, which has to signal EV_CLEARED
735 * before dropping our spinlock.
736 */
737 spin_lock(&object->lock);
738 spin_unlock(&object->lock);
739
740 /* Discard from the cache's collection of objects */
David Howells36c955902009-04-03 16:42:38 +0100741 spin_lock(&cache->object_list_lock);
742 list_del_init(&object->cache_link);
743 spin_unlock(&cache->object_list_lock);
744
David Howells52bd75f2009-11-19 18:11:08 +0000745 fscache_stat(&fscache_n_cop_drop_object);
David Howells36c955902009-04-03 16:42:38 +0100746 cache->ops->drop_object(object);
David Howells52bd75f2009-11-19 18:11:08 +0000747 fscache_stat_d(&fscache_n_cop_drop_object);
David Howells36c955902009-04-03 16:42:38 +0100748
David Howellscaaef692013-05-10 19:50:26 +0100749 /* The parent object wants to know when all it dependents have gone */
David Howells36c955902009-04-03 16:42:38 +0100750 if (parent) {
751 _debug("release parent OBJ%x {%d}",
752 parent->debug_id, parent->n_children);
753
754 spin_lock(&parent->lock);
755 parent->n_children--;
756 if (parent->n_children == 0)
757 fscache_raise_event(parent, FSCACHE_OBJECT_EV_CLEARED);
758 spin_unlock(&parent->lock);
759 object->parent = NULL;
760 }
761
Tejun Heo8b8edef2010-07-20 22:09:01 +0200762 /* this just shifts the object release to the work processor */
David Howellsa18feb52018-04-04 13:41:27 +0100763 fscache_put_object(object, fscache_obj_put_drop_obj);
David Howellscaaef692013-05-10 19:50:26 +0100764 fscache_stat(&fscache_n_object_dead);
David Howells36c955902009-04-03 16:42:38 +0100765
766 _leave("");
David Howellscaaef692013-05-10 19:50:26 +0100767 return transit_to(OBJECT_DEAD);
David Howells36c955902009-04-03 16:42:38 +0100768}
769
770/*
Tejun Heo8b8edef2010-07-20 22:09:01 +0200771 * get a ref on an object
David Howells36c955902009-04-03 16:42:38 +0100772 */
David Howellsa18feb52018-04-04 13:41:27 +0100773static int fscache_get_object(struct fscache_object *object,
774 enum fscache_obj_ref_trace why)
David Howells36c955902009-04-03 16:42:38 +0100775{
David Howells52bd75f2009-11-19 18:11:08 +0000776 int ret;
David Howells36c955902009-04-03 16:42:38 +0100777
David Howells52bd75f2009-11-19 18:11:08 +0000778 fscache_stat(&fscache_n_cop_grab_object);
David Howellsa18feb52018-04-04 13:41:27 +0100779 ret = object->cache->ops->grab_object(object, why) ? 0 : -EAGAIN;
David Howells52bd75f2009-11-19 18:11:08 +0000780 fscache_stat_d(&fscache_n_cop_grab_object);
781 return ret;
David Howells36c955902009-04-03 16:42:38 +0100782}
783
784/*
David Howellscaaef692013-05-10 19:50:26 +0100785 * Discard a ref on an object
David Howells36c955902009-04-03 16:42:38 +0100786 */
David Howellsa18feb52018-04-04 13:41:27 +0100787static void fscache_put_object(struct fscache_object *object,
788 enum fscache_obj_ref_trace why)
David Howells36c955902009-04-03 16:42:38 +0100789{
David Howells52bd75f2009-11-19 18:11:08 +0000790 fscache_stat(&fscache_n_cop_put_object);
David Howellsa18feb52018-04-04 13:41:27 +0100791 object->cache->ops->put_object(object, why);
David Howells52bd75f2009-11-19 18:11:08 +0000792 fscache_stat_d(&fscache_n_cop_put_object);
David Howells36c955902009-04-03 16:42:38 +0100793}
794
David Howells13627292013-05-10 19:50:26 +0100795/**
796 * fscache_object_destroy - Note that a cache object is about to be destroyed
797 * @object: The object to be destroyed
798 *
799 * Note the imminent destruction and deallocation of a cache object record.
800 */
801void fscache_object_destroy(struct fscache_object *object)
802{
803 fscache_objlist_remove(object);
804
805 /* We can get rid of the cookie now */
David Howellsa18feb52018-04-04 13:41:27 +0100806 fscache_cookie_put(object->cookie, fscache_cookie_put_object);
David Howells13627292013-05-10 19:50:26 +0100807 object->cookie = NULL;
808}
809EXPORT_SYMBOL(fscache_object_destroy);
810
David Howells36c955902009-04-03 16:42:38 +0100811/*
812 * enqueue an object for metadata-type processing
813 */
814void fscache_enqueue_object(struct fscache_object *object)
815{
816 _enter("{OBJ%x}", object->debug_id);
817
David Howellsa18feb52018-04-04 13:41:27 +0100818 if (fscache_get_object(object, fscache_obj_get_queue) >= 0) {
Tejun Heo8b8edef2010-07-20 22:09:01 +0200819 wait_queue_head_t *cong_wq =
820 &get_cpu_var(fscache_object_cong_wait);
821
822 if (queue_work(fscache_object_wq, &object->work)) {
823 if (fscache_object_congested())
824 wake_up(cong_wq);
825 } else
David Howellsa18feb52018-04-04 13:41:27 +0100826 fscache_put_object(object, fscache_obj_put_queue);
Tejun Heo8b8edef2010-07-20 22:09:01 +0200827
828 put_cpu_var(fscache_object_cong_wait);
829 }
David Howells36c955902009-04-03 16:42:38 +0100830}
831
Tejun Heo8b8edef2010-07-20 22:09:01 +0200832/**
833 * fscache_object_sleep_till_congested - Sleep until object wq is congested
David Howellscaaef692013-05-10 19:50:26 +0100834 * @timeoutp: Scheduler sleep timeout
Tejun Heo8b8edef2010-07-20 22:09:01 +0200835 *
836 * Allow an object handler to sleep until the object workqueue is congested.
837 *
838 * The caller must set up a wake up event before calling this and must have set
839 * the appropriate sleep mode (such as TASK_UNINTERRUPTIBLE) and tested its own
840 * condition before calling this function as no test is made here.
841 *
842 * %true is returned if the object wq is congested, %false otherwise.
843 */
844bool fscache_object_sleep_till_congested(signed long *timeoutp)
845{
Christoph Lameter170d8002013-10-15 12:22:29 -0600846 wait_queue_head_t *cong_wq = this_cpu_ptr(&fscache_object_cong_wait);
Tejun Heo8b8edef2010-07-20 22:09:01 +0200847 DEFINE_WAIT(wait);
848
849 if (fscache_object_congested())
850 return true;
851
852 add_wait_queue_exclusive(cong_wq, &wait);
853 if (!fscache_object_congested())
854 *timeoutp = schedule_timeout(*timeoutp);
855 finish_wait(cong_wq, &wait);
856
857 return fscache_object_congested();
858}
859EXPORT_SYMBOL_GPL(fscache_object_sleep_till_congested);
860
David Howells36c955902009-04-03 16:42:38 +0100861/*
David Howellscaaef692013-05-10 19:50:26 +0100862 * Enqueue the dependents of an object for metadata-type processing.
863 *
864 * If we don't manage to finish the list before the scheduler wants to run
865 * again then return false immediately. We return true if the list was
866 * cleared.
David Howells36c955902009-04-03 16:42:38 +0100867 */
David Howellscaaef692013-05-10 19:50:26 +0100868static bool fscache_enqueue_dependents(struct fscache_object *object, int event)
David Howells36c955902009-04-03 16:42:38 +0100869{
870 struct fscache_object *dep;
David Howellscaaef692013-05-10 19:50:26 +0100871 bool ret = true;
David Howells36c955902009-04-03 16:42:38 +0100872
873 _enter("{OBJ%x}", object->debug_id);
874
875 if (list_empty(&object->dependents))
David Howellscaaef692013-05-10 19:50:26 +0100876 return true;
David Howells36c955902009-04-03 16:42:38 +0100877
878 spin_lock(&object->lock);
879
880 while (!list_empty(&object->dependents)) {
881 dep = list_entry(object->dependents.next,
882 struct fscache_object, dep_link);
883 list_del_init(&dep->dep_link);
884
David Howellscaaef692013-05-10 19:50:26 +0100885 fscache_raise_event(dep, event);
David Howellsa18feb52018-04-04 13:41:27 +0100886 fscache_put_object(dep, fscache_obj_put_enq_dep);
David Howells36c955902009-04-03 16:42:38 +0100887
David Howellscaaef692013-05-10 19:50:26 +0100888 if (!list_empty(&object->dependents) && need_resched()) {
889 ret = false;
890 break;
891 }
David Howells36c955902009-04-03 16:42:38 +0100892 }
893
894 spin_unlock(&object->lock);
David Howellscaaef692013-05-10 19:50:26 +0100895 return ret;
David Howells36c955902009-04-03 16:42:38 +0100896}
897
898/*
899 * remove an object from whatever queue it's waiting on
David Howells36c955902009-04-03 16:42:38 +0100900 */
David Howellscaaef692013-05-10 19:50:26 +0100901static void fscache_dequeue_object(struct fscache_object *object)
David Howells36c955902009-04-03 16:42:38 +0100902{
903 _enter("{OBJ%x}", object->debug_id);
904
905 if (!list_empty(&object->dep_link)) {
906 spin_lock(&object->parent->lock);
907 list_del_init(&object->dep_link);
908 spin_unlock(&object->parent->lock);
909 }
910
911 _leave("");
912}
913
914/**
915 * fscache_check_aux - Ask the netfs whether an object on disk is still valid
916 * @object: The object to ask about
917 * @data: The auxiliary data for the object
918 * @datalen: The size of the auxiliary data
919 *
David Howells13627292013-05-10 19:50:26 +0100920 * This function consults the netfs about the coherency state of an object.
921 * The caller must be holding a ref on cookie->n_active (held by
922 * fscache_look_up_object() on behalf of the cache backend during object lookup
923 * and creation).
David Howells36c955902009-04-03 16:42:38 +0100924 */
925enum fscache_checkaux fscache_check_aux(struct fscache_object *object,
David Howellsee1235a2018-04-04 13:41:28 +0100926 const void *data, uint16_t datalen,
927 loff_t object_size)
David Howells36c955902009-04-03 16:42:38 +0100928{
929 enum fscache_checkaux result;
930
931 if (!object->cookie->def->check_aux) {
932 fscache_stat(&fscache_n_checkaux_none);
933 return FSCACHE_CHECKAUX_OKAY;
934 }
935
936 result = object->cookie->def->check_aux(object->cookie->netfs_data,
David Howellsee1235a2018-04-04 13:41:28 +0100937 data, datalen, object_size);
David Howells36c955902009-04-03 16:42:38 +0100938 switch (result) {
939 /* entry okay as is */
940 case FSCACHE_CHECKAUX_OKAY:
941 fscache_stat(&fscache_n_checkaux_okay);
942 break;
943
944 /* entry requires update */
945 case FSCACHE_CHECKAUX_NEEDS_UPDATE:
946 fscache_stat(&fscache_n_checkaux_update);
947 break;
948
949 /* entry requires deletion */
950 case FSCACHE_CHECKAUX_OBSOLETE:
951 fscache_stat(&fscache_n_checkaux_obsolete);
952 break;
953
954 default:
955 BUG();
956 }
957
958 return result;
959}
960EXPORT_SYMBOL(fscache_check_aux);
David Howellsef778e72012-12-20 21:52:36 +0000961
962/*
963 * Asynchronously invalidate an object.
964 */
David Howellscaaef692013-05-10 19:50:26 +0100965static const struct fscache_state *_fscache_invalidate_object(struct fscache_object *object,
966 int event)
David Howellsef778e72012-12-20 21:52:36 +0000967{
968 struct fscache_operation *op;
969 struct fscache_cookie *cookie = object->cookie;
970
David Howellscaaef692013-05-10 19:50:26 +0100971 _enter("{OBJ%x},%d", object->debug_id, event);
972
David Howells13627292013-05-10 19:50:26 +0100973 /* We're going to need the cookie. If the cookie is not available then
974 * retire the object instead.
975 */
976 if (!fscache_use_cookie(object)) {
Matthew Wilcoxe5a95542018-04-10 16:36:48 -0700977 ASSERT(radix_tree_empty(&object->cookie->stores));
David Howells94d30ae2013-09-21 00:09:31 +0100978 set_bit(FSCACHE_OBJECT_RETIRED, &object->flags);
David Howells13627292013-05-10 19:50:26 +0100979 _leave(" [no cookie]");
980 return transit_to(KILL_OBJECT);
981 }
David Howellsef778e72012-12-20 21:52:36 +0000982
983 /* Reject any new read/write ops and abort any that are pending. */
984 fscache_invalidate_writes(cookie);
985 clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags);
986 fscache_cancel_all_ops(object);
987
988 /* Now we have to wait for in-progress reads and writes */
989 op = kzalloc(sizeof(*op), GFP_KERNEL);
David Howells13627292013-05-10 19:50:26 +0100990 if (!op)
991 goto nomem;
David Howellsef778e72012-12-20 21:52:36 +0000992
David Howells08c2e3d2018-04-04 13:41:27 +0100993 fscache_operation_init(cookie, op, object->cache->ops->invalidate_object,
David Howellsd3b97ca2015-02-24 10:05:29 +0000994 NULL, NULL);
David Howells13627292013-05-10 19:50:26 +0100995 op->flags = FSCACHE_OP_ASYNC |
996 (1 << FSCACHE_OP_EXCLUSIVE) |
997 (1 << FSCACHE_OP_UNUSE_COOKIE);
David Howells08c2e3d2018-04-04 13:41:27 +0100998 trace_fscache_page_op(cookie, NULL, op, fscache_page_op_invalidate);
David Howellsef778e72012-12-20 21:52:36 +0000999
1000 spin_lock(&cookie->lock);
1001 if (fscache_submit_exclusive_op(object, op) < 0)
David Howells8d763492012-12-05 13:34:48 +00001002 goto submit_op_failed;
David Howellsef778e72012-12-20 21:52:36 +00001003 spin_unlock(&cookie->lock);
1004 fscache_put_operation(op);
1005
1006 /* Once we've completed the invalidation, we know there will be no data
1007 * stored in the cache and thus we can reinstate the data-check-skip
1008 * optimisation.
1009 */
1010 set_bit(FSCACHE_COOKIE_NO_DATA_YET, &cookie->flags);
1011
1012 /* We can allow read and write requests to come in once again. They'll
1013 * queue up behind our exclusive invalidation operation.
1014 */
David Howellscaaef692013-05-10 19:50:26 +01001015 if (test_and_clear_bit(FSCACHE_COOKIE_INVALIDATING, &cookie->flags))
1016 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_INVALIDATING);
1017 _leave(" [ok]");
1018 return transit_to(UPDATE_OBJECT);
David Howells8d763492012-12-05 13:34:48 +00001019
David Howells13627292013-05-10 19:50:26 +01001020nomem:
David Howellsf09b4432015-02-24 10:05:28 +00001021 fscache_mark_object_dead(object);
David Howells13627292013-05-10 19:50:26 +01001022 fscache_unuse_cookie(object);
1023 _leave(" [ENOMEM]");
1024 return transit_to(KILL_OBJECT);
1025
David Howells8d763492012-12-05 13:34:48 +00001026submit_op_failed:
David Howellsf09b4432015-02-24 10:05:28 +00001027 fscache_mark_object_dead(object);
David Howells8d763492012-12-05 13:34:48 +00001028 spin_unlock(&cookie->lock);
Milosz Tanski920bce22014-08-13 12:58:21 -04001029 fscache_unuse_cookie(object);
David Howells8d763492012-12-05 13:34:48 +00001030 kfree(op);
David Howells8d763492012-12-05 13:34:48 +00001031 _leave(" [EIO]");
David Howellscaaef692013-05-10 19:50:26 +01001032 return transit_to(KILL_OBJECT);
1033}
1034
1035static const struct fscache_state *fscache_invalidate_object(struct fscache_object *object,
1036 int event)
1037{
1038 const struct fscache_state *s;
1039
1040 fscache_stat(&fscache_n_invalidates_run);
1041 fscache_stat(&fscache_n_cop_invalidate_object);
1042 s = _fscache_invalidate_object(object, event);
1043 fscache_stat_d(&fscache_n_cop_invalidate_object);
1044 return s;
1045}
1046
1047/*
David Howells402cb8d2018-04-04 13:41:28 +01001048 * Update auxiliary data.
1049 */
1050static void fscache_update_aux_data(struct fscache_object *object)
1051{
1052 fscache_stat(&fscache_n_updates_run);
1053 fscache_stat(&fscache_n_cop_update_object);
1054 object->cache->ops->update_object(object);
1055 fscache_stat_d(&fscache_n_cop_update_object);
1056}
1057
1058/*
David Howellscaaef692013-05-10 19:50:26 +01001059 * Asynchronously update an object.
1060 */
1061static const struct fscache_state *fscache_update_object(struct fscache_object *object,
1062 int event)
1063{
1064 _enter("{OBJ%x},%d", object->debug_id, event);
1065
David Howells402cb8d2018-04-04 13:41:28 +01001066 fscache_update_aux_data(object);
David Howellscaaef692013-05-10 19:50:26 +01001067
1068 _leave("");
1069 return transit_to(WAIT_FOR_CMD);
David Howellsef778e72012-12-20 21:52:36 +00001070}
David Howells182d9192015-02-19 23:47:31 +00001071
1072/**
1073 * fscache_object_retrying_stale - Note retrying stale object
1074 * @object: The object that will be retried
1075 *
1076 * Note that an object lookup found an on-disk object that was adjudged to be
1077 * stale and has been deleted. The lookup will be retried.
1078 */
1079void fscache_object_retrying_stale(struct fscache_object *object)
1080{
1081 fscache_stat(&fscache_n_cache_no_space_reject);
1082}
1083EXPORT_SYMBOL(fscache_object_retrying_stale);
1084
1085/**
1086 * fscache_object_mark_killed - Note that an object was killed
1087 * @object: The object that was culled
1088 * @why: The reason the object was killed.
1089 *
1090 * Note that an object was killed. Returns true if the object was
1091 * already marked killed, false if it wasn't.
1092 */
1093void fscache_object_mark_killed(struct fscache_object *object,
1094 enum fscache_why_object_killed why)
1095{
1096 if (test_and_set_bit(FSCACHE_OBJECT_KILLED_BY_CACHE, &object->flags)) {
1097 pr_err("Error: Object already killed by cache [%s]\n",
1098 object->cache->identifier);
1099 return;
1100 }
1101
1102 switch (why) {
1103 case FSCACHE_OBJECT_NO_SPACE:
1104 fscache_stat(&fscache_n_cache_no_space_reject);
1105 break;
1106 case FSCACHE_OBJECT_IS_STALE:
1107 fscache_stat(&fscache_n_cache_stale_objects);
1108 break;
1109 case FSCACHE_OBJECT_WAS_RETIRED:
1110 fscache_stat(&fscache_n_cache_retired_objects);
1111 break;
1112 case FSCACHE_OBJECT_WAS_CULLED:
1113 fscache_stat(&fscache_n_cache_culled_objects);
1114 break;
1115 }
1116}
1117EXPORT_SYMBOL(fscache_object_mark_killed);
David Howellse26bfeb2017-01-31 09:45:28 +00001118
1119/*
1120 * The object is dead. We can get here if an object gets queued by an event
1121 * that would lead to its death (such as EV_KILL) when the dispatcher is
1122 * already running (and so can be requeued) but hasn't yet cleared the event
1123 * mask.
1124 */
1125static const struct fscache_state *fscache_object_dead(struct fscache_object *object,
1126 int event)
1127{
1128 if (!test_and_set_bit(FSCACHE_OBJECT_RUN_AFTER_DEAD,
1129 &object->flags))
1130 return NO_TRANSIT;
1131
1132 WARN(true, "FS-Cache object redispatched after death");
1133 return NO_TRANSIT;
1134}