Commit | Line | Data |
---|---|---|
a4f4fdd7 | 1 | /* |
da751ca9 JA |
2 | * syslet engine |
3 | * | |
4 | * IO engine that does regular pread(2)/pwrite(2) to transfer data, but | |
5 | * with syslets to make the execution async. | |
a4f4fdd7 JA |
6 | * |
7 | */ | |
8 | #include <stdio.h> | |
9 | #include <stdlib.h> | |
10 | #include <unistd.h> | |
11 | #include <errno.h> | |
12 | #include <assert.h> | |
b8846354 | 13 | #include <asm/unistd.h> |
a4f4fdd7 JA |
14 | |
15 | #include "../fio.h" | |
16 | #include "../os.h" | |
17 | ||
18 | #ifdef FIO_HAVE_SYSLET | |
19 | ||
20 | struct syslet_data { | |
21 | struct io_u **events; | |
22 | unsigned int nr_events; | |
23 | ||
bf0dc8fa | 24 | struct async_head_user ahu; |
a4f4fdd7 | 25 | struct syslet_uatom **ring; |
9ff9de69 JA |
26 | |
27 | struct syslet_uatom *head, *tail; | |
a4f4fdd7 JA |
28 | }; |
29 | ||
9ff9de69 JA |
30 | static void fio_syslet_complete_atom(struct thread_data *td, |
31 | struct syslet_uatom *atom) | |
32 | { | |
33 | struct syslet_data *sd = td->io_ops->data; | |
5b38ee84 | 34 | struct syslet_uatom *last; |
9ff9de69 | 35 | struct io_u *io_u; |
9ff9de69 JA |
36 | |
37 | /* | |
5b38ee84 JA |
38 | * complete from the beginning of the sequence up to (and |
39 | * including) this atom | |
9ff9de69 | 40 | */ |
5b38ee84 JA |
41 | last = atom; |
42 | io_u = atom->private; | |
43 | atom = io_u->req.head; | |
9ff9de69 JA |
44 | |
45 | /* | |
46 | * now complete in right order | |
47 | */ | |
5b38ee84 | 48 | do { |
9ff9de69 JA |
49 | long ret; |
50 | ||
9ff9de69 JA |
51 | io_u = atom->private; |
52 | ret = *atom->ret_ptr; | |
e2e67912 | 53 | if (ret >= 0) |
9ff9de69 JA |
54 | io_u->resid = io_u->xfer_buflen - ret; |
55 | else if (ret < 0) | |
56 | io_u->error = ret; | |
57 | ||
58 | assert(sd->nr_events < td->iodepth); | |
59 | sd->events[sd->nr_events++] = io_u; | |
9ff9de69 | 60 | |
5b38ee84 JA |
61 | if (atom == last) |
62 | break; | |
9ff9de69 | 63 | |
5b38ee84 JA |
64 | atom = atom->next; |
65 | } while (1); | |
66 | ||
67 | assert(!last->next); | |
9ff9de69 JA |
68 | } |
69 | ||
a4f4fdd7 JA |
70 | /* |
71 | * Inspect the ring to see if we have completed events | |
72 | */ | |
73 | static void fio_syslet_complete(struct thread_data *td) | |
74 | { | |
75 | struct syslet_data *sd = td->io_ops->data; | |
76 | ||
77 | do { | |
78 | struct syslet_uatom *atom; | |
a4f4fdd7 | 79 | |
bf0dc8fa | 80 | atom = sd->ring[sd->ahu.user_ring_idx]; |
a4f4fdd7 JA |
81 | if (!atom) |
82 | break; | |
83 | ||
bf0dc8fa IM |
84 | sd->ring[sd->ahu.user_ring_idx] = NULL; |
85 | if (++sd->ahu.user_ring_idx == td->iodepth) | |
86 | sd->ahu.user_ring_idx = 0; | |
a4f4fdd7 | 87 | |
9ff9de69 | 88 | fio_syslet_complete_atom(td, atom); |
a4f4fdd7 JA |
89 | } while (1); |
90 | } | |
91 | ||
92 | static int fio_syslet_getevents(struct thread_data *td, int min, | |
93 | int fio_unused max, | |
94 | struct timespec fio_unused *t) | |
95 | { | |
96 | struct syslet_data *sd = td->io_ops->data; | |
a4f4fdd7 JA |
97 | long ret; |
98 | ||
99 | do { | |
100 | fio_syslet_complete(td); | |
101 | ||
102 | /* | |
103 | * do we have enough immediate completions? | |
104 | */ | |
105 | if (sd->nr_events >= (unsigned int) min) | |
106 | break; | |
107 | ||
108 | /* | |
109 | * OK, we need to wait for some events... | |
110 | */ | |
9ff9de69 | 111 | ret = async_wait(1, sd->ahu.user_ring_idx, &sd->ahu); |
a4f4fdd7 | 112 | if (ret < 0) |
e49499f8 | 113 | return -errno; |
a4f4fdd7 JA |
114 | } while (1); |
115 | ||
116 | ret = sd->nr_events; | |
117 | sd->nr_events = 0; | |
118 | return ret; | |
119 | } | |
120 | ||
121 | static struct io_u *fio_syslet_event(struct thread_data *td, int event) | |
122 | { | |
123 | struct syslet_data *sd = td->io_ops->data; | |
124 | ||
125 | return sd->events[event]; | |
126 | } | |
127 | ||
128 | static void init_atom(struct syslet_uatom *atom, int nr, void *arg0, | |
a2e1b08a JA |
129 | void *arg1, void *arg2, void *arg3, void *ret_ptr, |
130 | unsigned long flags, void *priv) | |
a4f4fdd7 JA |
131 | { |
132 | atom->flags = flags; | |
133 | atom->nr = nr; | |
134 | atom->ret_ptr = ret_ptr; | |
a2e1b08a | 135 | atom->next = NULL; |
a4f4fdd7 JA |
136 | atom->arg_ptr[0] = arg0; |
137 | atom->arg_ptr[1] = arg1; | |
138 | atom->arg_ptr[2] = arg2; | |
a2e1b08a JA |
139 | atom->arg_ptr[3] = arg3; |
140 | atom->arg_ptr[4] = atom->arg_ptr[5] = NULL; | |
a4f4fdd7 JA |
141 | atom->private = priv; |
142 | } | |
143 | ||
144 | /* | |
145 | * Use seek atom for sync | |
146 | */ | |
147 | static void fio_syslet_prep_sync(struct io_u *io_u, struct fio_file *f) | |
148 | { | |
a2e1b08a | 149 | init_atom(&io_u->req.atom, __NR_fsync, &f->fd, NULL, NULL, NULL, |
7d44a745 | 150 | &io_u->req.ret, 0, io_u); |
a4f4fdd7 JA |
151 | } |
152 | ||
153 | static void fio_syslet_prep_rw(struct io_u *io_u, struct fio_file *f) | |
154 | { | |
155 | int nr; | |
156 | ||
a4f4fdd7 JA |
157 | /* |
158 | * prepare rw | |
159 | */ | |
160 | if (io_u->ddir == DDIR_READ) | |
a2e1b08a | 161 | nr = __NR_pread64; |
a4f4fdd7 | 162 | else |
a2e1b08a | 163 | nr = __NR_pwrite64; |
a4f4fdd7 | 164 | |
a2e1b08a | 165 | init_atom(&io_u->req.atom, nr, &f->fd, &io_u->xfer_buf, |
7d44a745 | 166 | &io_u->xfer_buflen, &io_u->offset, &io_u->req.ret, 0, io_u); |
a4f4fdd7 JA |
167 | } |
168 | ||
169 | static int fio_syslet_prep(struct thread_data fio_unused *td, struct io_u *io_u) | |
170 | { | |
171 | struct fio_file *f = io_u->file; | |
172 | ||
173 | if (io_u->ddir == DDIR_SYNC) | |
174 | fio_syslet_prep_sync(io_u, f); | |
175 | else | |
176 | fio_syslet_prep_rw(io_u, f); | |
177 | ||
178 | return 0; | |
179 | } | |
180 | ||
bf0dc8fa IM |
181 | static void cachemiss_thread_start(void) |
182 | { | |
183 | while (1) | |
7756b0d0 | 184 | async_thread(NULL, NULL); |
bf0dc8fa IM |
185 | } |
186 | ||
187 | #define THREAD_STACK_SIZE (16384) | |
188 | ||
189 | static unsigned long thread_stack_alloc() | |
190 | { | |
5b38ee84 | 191 | return (unsigned long) malloc(THREAD_STACK_SIZE) + THREAD_STACK_SIZE; |
bf0dc8fa IM |
192 | } |
193 | ||
a0a930ef JA |
194 | static void fio_syslet_queued(struct thread_data *td, struct syslet_data *sd) |
195 | { | |
196 | struct syslet_uatom *atom; | |
197 | struct timeval now; | |
198 | ||
199 | fio_gettime(&now, NULL); | |
200 | ||
201 | atom = sd->head; | |
202 | while (atom) { | |
203 | struct io_u *io_u = atom->private; | |
204 | ||
205 | memcpy(&io_u->issue_time, &now, sizeof(now)); | |
206 | io_u_queued(td, io_u); | |
207 | atom = atom->next; | |
208 | } | |
209 | } | |
210 | ||
9ff9de69 | 211 | static int fio_syslet_commit(struct thread_data *td) |
a4f4fdd7 JA |
212 | { |
213 | struct syslet_data *sd = td->io_ops->data; | |
bf0dc8fa | 214 | struct syslet_uatom *done; |
9ff9de69 JA |
215 | |
216 | if (!sd->head) | |
217 | return 0; | |
a4f4fdd7 | 218 | |
5b38ee84 JA |
219 | assert(!sd->tail->next); |
220 | ||
bf0dc8fa IM |
221 | if (!sd->ahu.new_thread_stack) |
222 | sd->ahu.new_thread_stack = thread_stack_alloc(); | |
223 | ||
a0a930ef JA |
224 | fio_syslet_queued(td, sd); |
225 | ||
7d44a745 JA |
226 | /* |
227 | * On sync completion, the atom is returned. So on NULL return | |
228 | * it's queued asynchronously. | |
229 | */ | |
9ff9de69 | 230 | done = async_exec(sd->head, &sd->ahu); |
bf0dc8fa | 231 | |
9ff9de69 | 232 | sd->head = sd->tail = NULL; |
a4f4fdd7 | 233 | |
9ff9de69 JA |
234 | if (done) |
235 | fio_syslet_complete_atom(td, done); | |
a4f4fdd7 | 236 | |
9ff9de69 JA |
237 | return 0; |
238 | } | |
239 | ||
240 | static int fio_syslet_queue(struct thread_data *td, struct io_u *io_u) | |
241 | { | |
242 | struct syslet_data *sd = td->io_ops->data; | |
bf0dc8fa | 243 | |
9ff9de69 JA |
244 | if (sd->tail) { |
245 | sd->tail->next = &io_u->req.atom; | |
246 | sd->tail = &io_u->req.atom; | |
247 | } else | |
248 | sd->head = sd->tail = &io_u->req.atom; | |
a4f4fdd7 | 249 | |
5b38ee84 | 250 | io_u->req.head = sd->head; |
9ff9de69 | 251 | return FIO_Q_QUEUED; |
a4f4fdd7 JA |
252 | } |
253 | ||
db64e9bc | 254 | static int async_head_init(struct syslet_data *sd, unsigned int depth) |
a4f4fdd7 | 255 | { |
a4f4fdd7 JA |
256 | unsigned long ring_size; |
257 | ||
bf0dc8fa | 258 | memset(&sd->ahu, 0, sizeof(struct async_head_user)); |
2ca50be4 | 259 | |
a4f4fdd7 JA |
260 | ring_size = sizeof(struct syslet_uatom *) * depth; |
261 | sd->ring = malloc(ring_size); | |
262 | memset(sd->ring, 0, ring_size); | |
263 | ||
bf0dc8fa IM |
264 | sd->ahu.user_ring_idx = 0; |
265 | sd->ahu.completion_ring = sd->ring; | |
266 | sd->ahu.ring_size_bytes = ring_size; | |
267 | sd->ahu.head_stack = thread_stack_alloc(); | |
5b38ee84 JA |
268 | sd->ahu.head_eip = (unsigned long) cachemiss_thread_start; |
269 | sd->ahu.new_thread_eip = (unsigned long) cachemiss_thread_start; | |
db64e9bc JA |
270 | |
271 | return 0; | |
a4f4fdd7 JA |
272 | } |
273 | ||
2ca50be4 | 274 | static void async_head_exit(struct syslet_data *sd) |
a4f4fdd7 | 275 | { |
7f059a76 | 276 | free(sd->ring); |
a4f4fdd7 JA |
277 | } |
278 | ||
279 | static void fio_syslet_cleanup(struct thread_data *td) | |
280 | { | |
281 | struct syslet_data *sd = td->io_ops->data; | |
282 | ||
283 | if (sd) { | |
2ca50be4 | 284 | async_head_exit(sd); |
a4f4fdd7 JA |
285 | free(sd->events); |
286 | free(sd); | |
287 | td->io_ops->data = NULL; | |
288 | } | |
289 | } | |
290 | ||
291 | static int fio_syslet_init(struct thread_data *td) | |
292 | { | |
293 | struct syslet_data *sd; | |
294 | ||
db64e9bc | 295 | |
a4f4fdd7 JA |
296 | sd = malloc(sizeof(*sd)); |
297 | memset(sd, 0, sizeof(*sd)); | |
298 | sd->events = malloc(sizeof(struct io_u *) * td->iodepth); | |
299 | memset(sd->events, 0, sizeof(struct io_u *) * td->iodepth); | |
db64e9bc JA |
300 | |
301 | /* | |
302 | * This will handily fail for kernels where syslet isn't available | |
303 | */ | |
304 | if (async_head_init(sd, td->iodepth)) { | |
305 | free(sd->events); | |
306 | free(sd); | |
307 | return 1; | |
308 | } | |
309 | ||
a4f4fdd7 | 310 | td->io_ops->data = sd; |
a4f4fdd7 JA |
311 | return 0; |
312 | } | |
313 | ||
314 | static struct ioengine_ops ioengine = { | |
315 | .name = "syslet-rw", | |
316 | .version = FIO_IOOPS_VERSION, | |
317 | .init = fio_syslet_init, | |
318 | .prep = fio_syslet_prep, | |
319 | .queue = fio_syslet_queue, | |
9ff9de69 | 320 | .commit = fio_syslet_commit, |
a4f4fdd7 JA |
321 | .getevents = fio_syslet_getevents, |
322 | .event = fio_syslet_event, | |
323 | .cleanup = fio_syslet_cleanup, | |
b5af8293 JA |
324 | .open_file = generic_open_file, |
325 | .close_file = generic_close_file, | |
a4f4fdd7 JA |
326 | }; |
327 | ||
328 | #else /* FIO_HAVE_SYSLET */ | |
329 | ||
330 | /* | |
331 | * When we have a proper configure system in place, we simply wont build | |
332 | * and install this io engine. For now install a crippled version that | |
333 | * just complains and fails to load. | |
334 | */ | |
335 | static int fio_syslet_init(struct thread_data fio_unused *td) | |
336 | { | |
337 | fprintf(stderr, "fio: syslet not available\n"); | |
338 | return 1; | |
339 | } | |
340 | ||
341 | static struct ioengine_ops ioengine = { | |
342 | .name = "syslet-rw", | |
343 | .version = FIO_IOOPS_VERSION, | |
344 | .init = fio_syslet_init, | |
345 | }; | |
346 | ||
347 | #endif /* FIO_HAVE_SYSLET */ | |
348 | ||
349 | static void fio_init fio_syslet_register(void) | |
350 | { | |
351 | register_ioengine(&ioengine); | |
352 | } | |
353 | ||
354 | static void fio_exit fio_syslet_unregister(void) | |
355 | { | |
356 | unregister_ioengine(&ioengine); | |
357 | } |