t/zbd: check log file for failed assertions
[fio.git] / ioengines.c
... / ...
CommitLineData
1/*
2 * The io parts of the fio tool, includes workers for sync and mmap'ed
3 * io, as well as both posix and linux libaio support.
4 *
5 * sync io is implemented on top of aio.
6 *
7 * This is not really specific to fio, if the get_io_u/put_io_u and
8 * structures was pulled into this as well it would be a perfectly
9 * generic io engine that could be used for other projects.
10 *
11 */
12#include <stdlib.h>
13#include <unistd.h>
14#include <string.h>
15#include <dlfcn.h>
16#include <fcntl.h>
17#include <assert.h>
18
19#include "fio.h"
20#include "diskutil.h"
21#include "zbd.h"
22
23static FLIST_HEAD(engine_list);
24
25static bool check_engine_ops(struct ioengine_ops *ops)
26{
27 if (ops->version != FIO_IOOPS_VERSION) {
28 log_err("bad ioops version %d (want %d)\n", ops->version,
29 FIO_IOOPS_VERSION);
30 return true;
31 }
32
33 if (!ops->queue) {
34 log_err("%s: no queue handler\n", ops->name);
35 return true;
36 }
37
38 /*
39 * sync engines only need a ->queue()
40 */
41 if (ops->flags & FIO_SYNCIO)
42 return false;
43
44 if (!ops->event || !ops->getevents) {
45 log_err("%s: no event/getevents handler\n", ops->name);
46 return true;
47 }
48
49 return false;
50}
51
52void unregister_ioengine(struct ioengine_ops *ops)
53{
54 dprint(FD_IO, "ioengine %s unregistered\n", ops->name);
55 flist_del_init(&ops->list);
56}
57
58void register_ioengine(struct ioengine_ops *ops)
59{
60 dprint(FD_IO, "ioengine %s registered\n", ops->name);
61 flist_add_tail(&ops->list, &engine_list);
62}
63
64static struct ioengine_ops *find_ioengine(const char *name)
65{
66 struct ioengine_ops *ops;
67 struct flist_head *entry;
68
69 flist_for_each(entry, &engine_list) {
70 ops = flist_entry(entry, struct ioengine_ops, list);
71 if (!strcmp(name, ops->name))
72 return ops;
73 }
74
75 return NULL;
76}
77
78#ifdef CONFIG_DYNAMIC_ENGINES
79static void *dlopen_external(struct thread_data *td, const char *engine)
80{
81 char engine_path[PATH_MAX];
82 void *dlhandle;
83
84 sprintf(engine_path, "%s/lib%s.so", FIO_EXT_ENG_DIR, engine);
85
86 dlhandle = dlopen(engine_path, RTLD_LAZY);
87 if (!dlhandle)
88 log_info("Engine %s not found; Either name is invalid, was not built, or fio-engine-%s package is missing.\n",
89 engine, engine);
90
91 return dlhandle;
92}
93#else
94#define dlopen_external(td, engine) (NULL)
95#endif
96
97static struct ioengine_ops *dlopen_ioengine(struct thread_data *td,
98 const char *engine_lib)
99{
100 struct ioengine_ops *ops;
101 void *dlhandle;
102
103 dprint(FD_IO, "dload engine %s\n", engine_lib);
104
105 dlerror();
106 dlhandle = dlopen(engine_lib, RTLD_LAZY);
107 if (!dlhandle) {
108 dlhandle = dlopen_external(td, engine_lib);
109 if (!dlhandle) {
110 td_vmsg(td, -1, dlerror(), "dlopen");
111 return NULL;
112 }
113 }
114
115 /*
116 * Unlike the included modules, external engines should have a
117 * non-static ioengine structure that we can reference.
118 */
119 ops = dlsym(dlhandle, engine_lib);
120 if (!ops)
121 ops = dlsym(dlhandle, "ioengine");
122
123 /*
124 * For some external engines (like C++ ones) it is not that trivial
125 * to provide a non-static ionengine structure that we can reference.
126 * Instead we call a method which allocates the required ioengine
127 * structure.
128 */
129 if (!ops) {
130 get_ioengine_t get_ioengine = dlsym(dlhandle, "get_ioengine");
131
132 if (get_ioengine)
133 get_ioengine(&ops);
134 }
135
136 if (!ops) {
137 td_vmsg(td, -1, dlerror(), "dlsym");
138 dlclose(dlhandle);
139 return NULL;
140 }
141
142 td->io_ops_dlhandle = dlhandle;
143 return ops;
144}
145
146static struct ioengine_ops *__load_ioengine(const char *engine)
147{
148 /*
149 * linux libaio has alias names, so convert to what we want
150 */
151 if (!strncmp(engine, "linuxaio", 8)) {
152 dprint(FD_IO, "converting ioengine name: %s -> libaio\n",
153 engine);
154 engine = "libaio";
155 }
156
157 dprint(FD_IO, "load ioengine %s\n", engine);
158 return find_ioengine(engine);
159}
160
161struct ioengine_ops *load_ioengine(struct thread_data *td)
162{
163 struct ioengine_ops *ops = NULL;
164 const char *name;
165
166 /*
167 * Use ->ioengine_so_path if an external ioengine path is specified.
168 * In this case, ->ioengine is "external" which also means the prefix
169 * for external ioengines "external:" is properly used.
170 */
171 name = td->o.ioengine_so_path ?: td->o.ioengine;
172
173 /*
174 * Try to load ->ioengine first, and if failed try to dlopen(3) either
175 * ->ioengine or ->ioengine_so_path. This is redundant for an external
176 * ioengine with prefix, and also leaves the possibility of unexpected
177 * behavior (e.g. if the "external" ioengine exists), but we do this
178 * so as not to break job files not using the prefix.
179 */
180 ops = __load_ioengine(td->o.ioengine);
181 if (!ops)
182 ops = dlopen_ioengine(td, name);
183
184 /*
185 * If ops is NULL, we failed to load ->ioengine, and also failed to
186 * dlopen(3) either ->ioengine or ->ioengine_so_path as a path.
187 */
188 if (!ops) {
189 log_err("fio: engine %s not loadable\n", name);
190 return NULL;
191 }
192
193 /*
194 * Check that the required methods are there.
195 */
196 if (check_engine_ops(ops))
197 return NULL;
198
199 return ops;
200}
201
202/*
203 * For cleaning up an ioengine which never made it to init().
204 */
205void free_ioengine(struct thread_data *td)
206{
207 dprint(FD_IO, "free ioengine %s\n", td->io_ops->name);
208
209 if (td->eo && td->io_ops->options) {
210 options_free(td->io_ops->options, td->eo);
211 free(td->eo);
212 td->eo = NULL;
213 }
214
215 if (td->io_ops_dlhandle) {
216 dlclose(td->io_ops_dlhandle);
217 td->io_ops_dlhandle = NULL;
218 }
219
220 td->io_ops = NULL;
221}
222
223void close_ioengine(struct thread_data *td)
224{
225 dprint(FD_IO, "close ioengine %s\n", td->io_ops->name);
226
227 if (td->io_ops->cleanup) {
228 td->io_ops->cleanup(td);
229 td->io_ops_data = NULL;
230 }
231
232 free_ioengine(td);
233}
234
235int td_io_prep(struct thread_data *td, struct io_u *io_u)
236{
237 dprint_io_u(io_u, "prep");
238 fio_ro_check(td, io_u);
239
240 lock_file(td, io_u->file, io_u->ddir);
241
242 if (td->io_ops->prep) {
243 int ret = td->io_ops->prep(td, io_u);
244
245 dprint(FD_IO, "prep: io_u %p: ret=%d\n", io_u, ret);
246
247 if (ret)
248 unlock_file(td, io_u->file);
249 return ret;
250 }
251
252 return 0;
253}
254
255int td_io_getevents(struct thread_data *td, unsigned int min, unsigned int max,
256 const struct timespec *t)
257{
258 int r = 0;
259
260 /*
261 * For ioengine=rdma one side operation RDMA_WRITE or RDMA_READ,
262 * server side gets a message from the client
263 * side that the task is finished, and
264 * td->done is set to 1 after td_io_commit(). In this case,
265 * there is no need to reap complete event in server side.
266 */
267 if (td->done)
268 return 0;
269
270 if (min > 0 && td->io_ops->commit) {
271 r = td->io_ops->commit(td);
272 if (r < 0)
273 goto out;
274 }
275 if (max > td->cur_depth)
276 max = td->cur_depth;
277 if (min > max)
278 max = min;
279
280 r = 0;
281 if (max && td->io_ops->getevents)
282 r = td->io_ops->getevents(td, min, max, t);
283out:
284 if (r >= 0) {
285 /*
286 * Reflect that our submitted requests were retrieved with
287 * whatever OS async calls are in the underlying engine.
288 */
289 td->io_u_in_flight -= r;
290 io_u_mark_complete(td, r);
291 } else
292 td_verror(td, r, "get_events");
293
294 dprint(FD_IO, "getevents: %d\n", r);
295 return r;
296}
297
298enum fio_q_status td_io_queue(struct thread_data *td, struct io_u *io_u)
299{
300 const enum fio_ddir ddir = acct_ddir(io_u);
301 unsigned long long buflen = io_u->xfer_buflen;
302 enum fio_q_status ret;
303
304 dprint_io_u(io_u, "queue");
305 fio_ro_check(td, io_u);
306
307 assert((io_u->flags & IO_U_F_FLIGHT) == 0);
308 io_u_set(td, io_u, IO_U_F_FLIGHT);
309
310 /*
311 * If overlap checking was enabled in offload mode we
312 * can release this lock that was acquired when we
313 * started the overlap check because the IO_U_F_FLIGHT
314 * flag is now set
315 */
316 if (td_offload_overlap(td)) {
317 int res = pthread_mutex_unlock(&overlap_check);
318 assert(res == 0);
319 }
320
321 assert(fio_file_open(io_u->file));
322
323 /*
324 * If using a write iolog, store this entry.
325 */
326 log_io_u(td, io_u);
327
328 io_u->error = 0;
329 io_u->resid = 0;
330
331 if (td_ioengine_flagged(td, FIO_SYNCIO) ||
332 (td_ioengine_flagged(td, FIO_ASYNCIO_SYNC_TRIM) &&
333 io_u->ddir == DDIR_TRIM)) {
334 if (fio_fill_issue_time(td))
335 fio_gettime(&io_u->issue_time, NULL);
336
337 /*
338 * only used for iolog
339 */
340 if (td->o.read_iolog_file)
341 memcpy(&td->last_issue, &io_u->issue_time,
342 sizeof(io_u->issue_time));
343 }
344
345
346 if (ddir_rw(ddir)) {
347 if (!(io_u->flags & IO_U_F_VER_LIST)) {
348 td->io_issues[ddir]++;
349 td->io_issue_bytes[ddir] += buflen;
350 }
351 td->rate_io_issue_bytes[ddir] += buflen;
352 }
353
354 ret = td->io_ops->queue(td, io_u);
355 zbd_queue_io_u(io_u, ret);
356
357 unlock_file(td, io_u->file);
358
359 if (ret == FIO_Q_BUSY && ddir_rw(ddir)) {
360 td->io_issues[ddir]--;
361 td->io_issue_bytes[ddir] -= buflen;
362 td->rate_io_issue_bytes[ddir] -= buflen;
363 io_u_clear(td, io_u, IO_U_F_FLIGHT);
364 }
365
366 /*
367 * If an error was seen and the io engine didn't propagate it
368 * back to 'td', do so.
369 */
370 if (io_u->error && !td->error)
371 td_verror(td, io_u->error, "td_io_queue");
372
373 /*
374 * Add warning for O_DIRECT so that users have an easier time
375 * spotting potentially bad alignment. If this triggers for the first
376 * IO, then it's likely an alignment problem or because the host fs
377 * does not support O_DIRECT
378 */
379 if (io_u->error == EINVAL && td->io_issues[io_u->ddir & 1] == 1 &&
380 td->o.odirect) {
381
382 log_info("fio: first direct IO errored. File system may not "
383 "support direct IO, or iomem_align= is bad, or "
384 "invalid block size. Try setting direct=0.\n");
385 }
386
387 if (zbd_unaligned_write(io_u->error) &&
388 td->io_issues[io_u->ddir & 1] == 1 &&
389 td->o.zone_mode != ZONE_MODE_ZBD) {
390 log_info("fio: first I/O failed. If %s is a zoned block device, consider --zonemode=zbd\n",
391 io_u->file->file_name);
392 }
393
394 if (!td->io_ops->commit) {
395 io_u_mark_submit(td, 1);
396 io_u_mark_complete(td, 1);
397 zbd_put_io_u(io_u);
398 }
399
400 if (ret == FIO_Q_COMPLETED) {
401 if (ddir_rw(io_u->ddir) ||
402 (ddir_sync(io_u->ddir) && td->runstate != TD_FSYNCING)) {
403 io_u_mark_depth(td, 1);
404 td->ts.total_io_u[io_u->ddir]++;
405 }
406 } else if (ret == FIO_Q_QUEUED) {
407 td->io_u_queued++;
408
409 if (ddir_rw(io_u->ddir) ||
410 (ddir_sync(io_u->ddir) && td->runstate != TD_FSYNCING))
411 td->ts.total_io_u[io_u->ddir]++;
412
413 if (td->io_u_queued >= td->o.iodepth_batch)
414 td_io_commit(td);
415 }
416
417 if (!td_ioengine_flagged(td, FIO_SYNCIO) &&
418 (!td_ioengine_flagged(td, FIO_ASYNCIO_SYNC_TRIM) ||
419 io_u->ddir != DDIR_TRIM)) {
420 if (fio_fill_issue_time(td))
421 fio_gettime(&io_u->issue_time, NULL);
422
423 /*
424 * only used for iolog
425 */
426 if (td->o.read_iolog_file)
427 memcpy(&td->last_issue, &io_u->issue_time,
428 sizeof(io_u->issue_time));
429 }
430
431 return ret;
432}
433
434int td_io_init(struct thread_data *td)
435{
436 int ret = 0;
437
438 if (td->io_ops->init) {
439 ret = td->io_ops->init(td);
440 if (ret)
441 log_err("fio: io engine %s init failed.%s\n",
442 td->io_ops->name,
443 td->o.iodepth > 1 ?
444 " Perhaps try reducing io depth?" : "");
445 else
446 td->io_ops_init = 1;
447 if (!td->error)
448 td->error = ret;
449 }
450
451 return ret;
452}
453
454void td_io_commit(struct thread_data *td)
455{
456 int ret;
457
458 dprint(FD_IO, "calling ->commit(), depth %d\n", td->cur_depth);
459
460 if (!td->cur_depth || !td->io_u_queued)
461 return;
462
463 io_u_mark_depth(td, td->io_u_queued);
464
465 if (td->io_ops->commit) {
466 ret = td->io_ops->commit(td);
467 if (ret)
468 td_verror(td, -ret, "io commit");
469 }
470
471 /*
472 * Reflect that events were submitted as async IO requests.
473 */
474 td->io_u_in_flight += td->io_u_queued;
475 td->io_u_queued = 0;
476}
477
478int td_io_open_file(struct thread_data *td, struct fio_file *f)
479{
480 if (fio_file_closing(f)) {
481 /*
482 * Open translates to undo closing.
483 */
484 fio_file_clear_closing(f);
485 get_file(f);
486 return 0;
487 }
488 assert(!fio_file_open(f));
489 assert(f->fd == -1);
490 assert(td->io_ops->open_file);
491
492 if (td->io_ops->open_file(td, f)) {
493 if (td->error == EINVAL && td->o.odirect)
494 log_err("fio: destination does not support O_DIRECT\n");
495 if (td->error == EMFILE) {
496 log_err("fio: try reducing/setting openfiles (failed"
497 " at %u of %u)\n", td->nr_open_files,
498 td->o.nr_files);
499 }
500
501 assert(f->fd == -1);
502 assert(!fio_file_open(f));
503 return 1;
504 }
505
506 fio_file_reset(td, f);
507 fio_file_set_open(f);
508 fio_file_clear_closing(f);
509 disk_util_inc(f->du);
510
511 td->nr_open_files++;
512 get_file(f);
513
514 if (f->filetype == FIO_TYPE_PIPE) {
515 if (td_random(td)) {
516 log_err("fio: can't seek on pipes (no random io)\n");
517 goto err;
518 }
519 }
520
521 if (td_ioengine_flagged(td, FIO_DISKLESSIO))
522 goto done;
523
524 if (td->o.invalidate_cache && file_invalidate_cache(td, f))
525 goto err;
526
527 if (td->o.fadvise_hint != F_ADV_NONE &&
528 (f->filetype == FIO_TYPE_BLOCK || f->filetype == FIO_TYPE_FILE)) {
529 int flags;
530
531 if (td->o.fadvise_hint == F_ADV_TYPE) {
532 if (td_random(td))
533 flags = POSIX_FADV_RANDOM;
534 else
535 flags = POSIX_FADV_SEQUENTIAL;
536 } else if (td->o.fadvise_hint == F_ADV_RANDOM)
537 flags = POSIX_FADV_RANDOM;
538 else if (td->o.fadvise_hint == F_ADV_SEQUENTIAL)
539 flags = POSIX_FADV_SEQUENTIAL;
540 else {
541 log_err("fio: unknown fadvise type %d\n",
542 td->o.fadvise_hint);
543 flags = POSIX_FADV_NORMAL;
544 }
545
546 if (posix_fadvise(f->fd, f->file_offset, f->io_size, flags) < 0) {
547 if (!fio_did_warn(FIO_WARN_FADVISE))
548 log_err("fio: fadvise hint failed\n");
549 }
550 }
551#ifdef FIO_HAVE_WRITE_HINT
552 if (fio_option_is_set(&td->o, write_hint) &&
553 (f->filetype == FIO_TYPE_BLOCK || f->filetype == FIO_TYPE_FILE)) {
554 uint64_t hint = td->o.write_hint;
555 int cmd;
556
557 /*
558 * For direct IO, we just need/want to set the hint on
559 * the file descriptor. For buffered IO, we need to set
560 * it on the inode.
561 */
562 if (td->o.odirect)
563 cmd = F_SET_FILE_RW_HINT;
564 else
565 cmd = F_SET_RW_HINT;
566
567 if (fcntl(f->fd, cmd, &hint) < 0) {
568 td_verror(td, errno, "fcntl write hint");
569 goto err;
570 }
571 }
572#endif
573
574 if (td->o.odirect && !OS_O_DIRECT && fio_set_directio(td, f))
575 goto err;
576
577done:
578 log_file(td, f, FIO_LOG_OPEN_FILE);
579 return 0;
580err:
581 disk_util_dec(f->du);
582 if (td->io_ops->close_file)
583 td->io_ops->close_file(td, f);
584 return 1;
585}
586
587int td_io_close_file(struct thread_data *td, struct fio_file *f)
588{
589 if (!fio_file_closing(f))
590 log_file(td, f, FIO_LOG_CLOSE_FILE);
591
592 /*
593 * mark as closing, do real close when last io on it has completed
594 */
595 fio_file_set_closing(f);
596
597 return put_file(td, f);
598}
599
600int td_io_unlink_file(struct thread_data *td, struct fio_file *f)
601{
602 if (td->io_ops->unlink_file)
603 return td->io_ops->unlink_file(td, f);
604 else {
605 int ret;
606
607 ret = unlink(f->file_name);
608 if (ret < 0)
609 return errno;
610
611 return 0;
612 }
613}
614
615int td_io_get_file_size(struct thread_data *td, struct fio_file *f)
616{
617 if (!td->io_ops->get_file_size)
618 return 0;
619
620 return td->io_ops->get_file_size(td, f);
621}
622
623int fio_show_ioengine_help(const char *engine)
624{
625 struct flist_head *entry;
626 struct thread_data td;
627 struct ioengine_ops *io_ops;
628 char *sep;
629 int ret = 1;
630
631 if (!engine || !*engine) {
632 log_info("Available IO engines:\n");
633 flist_for_each(entry, &engine_list) {
634 io_ops = flist_entry(entry, struct ioengine_ops, list);
635 log_info("\t%s\n", io_ops->name);
636 }
637 return 0;
638 }
639 sep = strchr(engine, ',');
640 if (sep) {
641 *sep = 0;
642 sep++;
643 }
644
645 memset(&td, 0, sizeof(struct thread_data));
646 td.o.ioengine = (char *)engine;
647 io_ops = load_ioengine(&td);
648
649 if (!io_ops) {
650 log_info("IO engine %s not found\n", engine);
651 return 1;
652 }
653
654 if (io_ops->options)
655 ret = show_cmd_help(io_ops->options, sep);
656 else
657 log_info("IO engine %s has no options\n", io_ops->name);
658
659 free_ioengine(&td);
660 return ret;
661}