X-Git-Url: https://git.kernel.dk/?p=fio.git;a=blobdiff_plain;f=io_u.c;h=645cba40b4f012096bafff178b697ecff0ec46aa;hp=64399797b9b97a03d3bb4375a75a09d43126a46b;hb=3d7c391dbdc56f261447b2d8ab65e74a42a20c41;hpb=36167d82e5f49dee91c6d2cd426068edee90e36f diff --git a/io_u.c b/io_u.c index 64399797..645cba40 100644 --- a/io_u.c +++ b/io_u.c @@ -7,6 +7,15 @@ #include "fio.h" #include "os.h" +struct io_completion_data { + int nr; /* input */ + endio_handler *handler; /* input */ + + int error; /* output */ + unsigned long bytes_done[2]; /* output */ + struct timeval time; /* output */ +}; + /* * The ->file_map[] contains a map of blocks we have or have not done io * to yet. Used to make sure we cover the entire range in a fair fashion. @@ -190,6 +199,16 @@ void put_io_u(struct thread_data *td, struct io_u *io_u) td->cur_depth--; } +void requeue_io_u(struct thread_data *td, struct io_u **io_u) +{ + struct io_u *__io_u = *io_u; + + list_del(&__io_u->list); + list_add_tail(&__io_u->list, &td->io_u_requeues); + td->cur_depth--; + *io_u = NULL; +} + static int fill_io_u(struct thread_data *td, struct fio_file *f, struct io_u *io_u) { @@ -202,8 +221,8 @@ static int fill_io_u(struct thread_data *td, struct fio_file *f, /* * see if it's time to sync */ - if (td->fsync_blocks && !(td->io_blocks[DDIR_WRITE] % td->fsync_blocks) - && should_fsync(td)) { + if (td->fsync_blocks && !(td->io_issues[DDIR_WRITE] % td->fsync_blocks) + && td->io_issues[DDIR_WRITE] && should_fsync(td)) { io_u->ddir = DDIR_SYNC; io_u->file = f; return 0; @@ -297,16 +316,43 @@ static void io_u_mark_latency(struct thread_data *td, unsigned long msec) td->io_u_lat[index]++; } +static struct fio_file *get_next_file(struct thread_data *td) +{ + unsigned int old_next_file = td->next_file; + struct fio_file *f; + + do { + f = &td->files[td->next_file]; + + td->next_file++; + if (td->next_file >= td->nr_files) + td->next_file = 0; + + if (f->fd != -1) + break; + + f = NULL; + } while (td->next_file != old_next_file); + + return f; +} + struct io_u *__get_io_u(struct thread_data *td) { struct io_u *io_u = NULL; - if (!queue_full(td)) { + if (!list_empty(&td->io_u_requeues)) + io_u = list_entry(td->io_u_requeues.next, struct io_u, list); + else if (!queue_full(td)) { io_u = list_entry(td->io_u_freelist.next, struct io_u, list); io_u->buflen = 0; - io_u->error = 0; io_u->resid = 0; + io_u->file = NULL; + } + + if (io_u) { + io_u->error = 0; list_del(&io_u->list); list_add(&io_u->list, &td->io_u_busylist); td->cur_depth++; @@ -320,14 +366,29 @@ struct io_u *__get_io_u(struct thread_data *td) * Return an io_u to be processed. Gets a buflen and offset, sets direction, * etc. The returned io_u is fully ready to be prepped and submitted. */ -struct io_u *get_io_u(struct thread_data *td, struct fio_file *f) +struct io_u *get_io_u(struct thread_data *td) { + struct fio_file *f; struct io_u *io_u; io_u = __get_io_u(td); if (!io_u) return NULL; + /* + * from a requeue, io_u already setup + */ + if (io_u->file) + return io_u; + + f = get_next_file(td); + if (!f) { + put_io_u(td, io_u); + return NULL; + } + + io_u->file = f; + if (td->zone_bytes >= td->zone_size) { td->zone_bytes = 0; f->last_pos += td->zone_skip; @@ -374,8 +435,8 @@ struct io_u *get_io_u(struct thread_data *td, struct fio_file *f) return io_u; } -void io_completed(struct thread_data *td, struct io_u *io_u, - struct io_completion_data *icd) +static void io_completed(struct thread_data *td, struct io_u *io_u, + struct io_completion_data *icd) { unsigned long msec; @@ -389,6 +450,7 @@ void io_completed(struct thread_data *td, struct io_u *io_u, if (!io_u->error) { unsigned int bytes = io_u->buflen - io_u->resid; const enum fio_ddir idx = io_u->ddir; + int ret; td->io_blocks[idx]++; td->io_bytes[idx] += bytes; @@ -407,25 +469,34 @@ void io_completed(struct thread_data *td, struct io_u *io_u, log_io_piece(td, io_u); icd->bytes_done[idx] += bytes; + + if (icd->handler) { + ret = icd->handler(io_u); + if (ret && !icd->error) + icd->error = ret; + } } else icd->error = io_u->error; } -void init_icd(struct io_completion_data *icd) +static void init_icd(struct io_completion_data *icd, endio_handler *handler, + int nr) { fio_gettime(&icd->time, NULL); + icd->handler = handler; + icd->nr = nr; + icd->error = 0; icd->bytes_done[0] = icd->bytes_done[1] = 0; } -void ios_completed(struct thread_data *td, struct io_completion_data *icd) +static void ios_completed(struct thread_data *td, + struct io_completion_data *icd) { struct io_u *io_u; int i; - init_icd(icd); - for (i = 0; i < icd->nr; i++) { io_u = td->io_ops->event(td, i); @@ -433,3 +504,51 @@ void ios_completed(struct thread_data *td, struct io_completion_data *icd) put_io_u(td, io_u); } } + +long io_u_sync_complete(struct thread_data *td, struct io_u *io_u, + endio_handler *handler) +{ + struct io_completion_data icd; + + init_icd(&icd, handler, 1); + io_completed(td, io_u, &icd); + put_io_u(td, io_u); + + if (!icd.error) + return icd.bytes_done[0] + icd.bytes_done[1]; + + return -1; +} + +long io_u_queued_complete(struct thread_data *td, int min_events, + endio_handler *handler) + +{ + struct timespec ts = { .tv_sec = 0, .tv_nsec = 0, }; + struct timespec *tsp = NULL; + struct io_completion_data icd; + int ret; + + if (min_events > 0) { + tsp = &ts; + ret = td_io_commit(td); + if (ret < 0) { + td_verror(td, -ret); + return ret; + } + } + + ret = td_io_getevents(td, min_events, td->cur_depth, tsp); + if (ret < 0) { + td_verror(td, -ret); + return ret; + } else if (!ret) + return ret; + + init_icd(&icd, handler, ret); + ios_completed(td, &icd); + if (!icd.error) + return icd.bytes_done[0] + icd.bytes_done[1]; + + return -1; +}