X-Git-Url: https://git.kernel.dk/?p=fio.git;a=blobdiff_plain;f=io_u.c;h=645cba40b4f012096bafff178b697ecff0ec46aa;hp=c56ccc6c614b32737d829aa7b170513affce6bce;hb=3d7c391dbdc56f261447b2d8ab65e74a42a20c41;hpb=e9fbdd74f6ac24d93824177a8858c028305c9900;ds=sidebyside diff --git a/io_u.c b/io_u.c index c56ccc6c..645cba40 100644 --- a/io_u.c +++ b/io_u.c @@ -7,6 +7,15 @@ #include "fio.h" #include "os.h" +struct io_completion_data { + int nr; /* input */ + endio_handler *handler; /* input */ + + int error; /* output */ + unsigned long bytes_done[2]; /* output */ + struct timeval time; /* output */ +}; + /* * The ->file_map[] contains a map of blocks we have or have not done io * to yet. Used to make sure we cover the entire range in a fair fashion. @@ -190,6 +199,16 @@ void put_io_u(struct thread_data *td, struct io_u *io_u) td->cur_depth--; } +void requeue_io_u(struct thread_data *td, struct io_u **io_u) +{ + struct io_u *__io_u = *io_u; + + list_del(&__io_u->list); + list_add_tail(&__io_u->list, &td->io_u_requeues); + td->cur_depth--; + *io_u = NULL; +} + static int fill_io_u(struct thread_data *td, struct fio_file *f, struct io_u *io_u) { @@ -202,8 +221,8 @@ static int fill_io_u(struct thread_data *td, struct fio_file *f, /* * see if it's time to sync */ - if (td->fsync_blocks && !(td->io_blocks[DDIR_WRITE] % td->fsync_blocks) - && should_fsync(td)) { + if (td->fsync_blocks && !(td->io_issues[DDIR_WRITE] % td->fsync_blocks) + && td->io_issues[DDIR_WRITE] && should_fsync(td)) { io_u->ddir = DDIR_SYNC; io_u->file = f; return 0; @@ -263,16 +282,77 @@ static void io_u_mark_depth(struct thread_data *td) td->total_io_u++; } +static void io_u_mark_latency(struct thread_data *td, unsigned long msec) +{ + int index = 0; + + switch (msec) { + default: + index++; + case 1024 ... 2047: + index++; + case 512 ... 1023: + index++; + case 256 ... 511: + index++; + case 128 ... 255: + index++; + case 64 ... 127: + index++; + case 32 ... 63: + index++; + case 16 ... 31: + index++; + case 8 ... 15: + index++; + case 4 ... 7: + index++; + case 2 ... 3: + index++; + case 0 ... 1: + break; + } + + td->io_u_lat[index]++; +} + +static struct fio_file *get_next_file(struct thread_data *td) +{ + unsigned int old_next_file = td->next_file; + struct fio_file *f; + + do { + f = &td->files[td->next_file]; + + td->next_file++; + if (td->next_file >= td->nr_files) + td->next_file = 0; + + if (f->fd != -1) + break; + + f = NULL; + } while (td->next_file != old_next_file); + + return f; +} + struct io_u *__get_io_u(struct thread_data *td) { struct io_u *io_u = NULL; - if (!queue_full(td)) { + if (!list_empty(&td->io_u_requeues)) + io_u = list_entry(td->io_u_requeues.next, struct io_u, list); + else if (!queue_full(td)) { io_u = list_entry(td->io_u_freelist.next, struct io_u, list); io_u->buflen = 0; - io_u->error = 0; io_u->resid = 0; + io_u->file = NULL; + } + + if (io_u) { + io_u->error = 0; list_del(&io_u->list); list_add(&io_u->list, &td->io_u_busylist); td->cur_depth++; @@ -286,14 +366,29 @@ struct io_u *__get_io_u(struct thread_data *td) * Return an io_u to be processed. Gets a buflen and offset, sets direction, * etc. The returned io_u is fully ready to be prepped and submitted. */ -struct io_u *get_io_u(struct thread_data *td, struct fio_file *f) +struct io_u *get_io_u(struct thread_data *td) { + struct fio_file *f; struct io_u *io_u; io_u = __get_io_u(td); if (!io_u) return NULL; + /* + * from a requeue, io_u already setup + */ + if (io_u->file) + return io_u; + + f = get_next_file(td); + if (!f) { + put_io_u(td, io_u); + return NULL; + } + + io_u->file = f; + if (td->zone_bytes >= td->zone_size) { td->zone_bytes = 0; f->last_pos += td->zone_skip; @@ -319,29 +414,29 @@ struct io_u *get_io_u(struct thread_data *td, struct fio_file *f) return NULL; } - f->last_pos += io_u->buflen; + f->last_pos = io_u->offset + io_u->buflen; if (td->verify != VERIFY_NONE) populate_verify_io_u(td, io_u); } - if (td_io_prep(td, io_u)) { - put_io_u(td, io_u); - return NULL; - } - /* * Set io data pointers. */ io_u->xfer_buf = io_u->buf; io_u->xfer_buflen = io_u->buflen; + if (td_io_prep(td, io_u)) { + put_io_u(td, io_u); + return NULL; + } + fio_gettime(&io_u->start_time, NULL); return io_u; } -void io_completed(struct thread_data *td, struct io_u *io_u, - struct io_completion_data *icd) +static void io_completed(struct thread_data *td, struct io_u *io_u, + struct io_completion_data *icd) { unsigned long msec; @@ -355,6 +450,7 @@ void io_completed(struct thread_data *td, struct io_u *io_u, if (!io_u->error) { unsigned int bytes = io_u->buflen - io_u->resid; const enum fio_ddir idx = io_u->ddir; + int ret; td->io_blocks[idx]++; td->io_bytes[idx] += bytes; @@ -367,24 +463,39 @@ void io_completed(struct thread_data *td, struct io_u *io_u, add_clat_sample(td, idx, msec); add_bw_sample(td, idx, &icd->time); + io_u_mark_latency(td, msec); if ((td_rw(td) || td_write(td)) && idx == DDIR_WRITE) log_io_piece(td, io_u); icd->bytes_done[idx] += bytes; + + if (icd->handler) { + ret = icd->handler(io_u); + if (ret && !icd->error) + icd->error = ret; + } } else icd->error = io_u->error; } -void ios_completed(struct thread_data *td, struct io_completion_data *icd) +static void init_icd(struct io_completion_data *icd, endio_handler *handler, + int nr) { - struct io_u *io_u; - int i; - fio_gettime(&icd->time, NULL); + icd->handler = handler; + icd->nr = nr; + icd->error = 0; icd->bytes_done[0] = icd->bytes_done[1] = 0; +} + +static void ios_completed(struct thread_data *td, + struct io_completion_data *icd) +{ + struct io_u *io_u; + int i; for (i = 0; i < icd->nr; i++) { io_u = td->io_ops->event(td, i); @@ -393,3 +504,51 @@ void ios_completed(struct thread_data *td, struct io_completion_data *icd) put_io_u(td, io_u); } } + +long io_u_sync_complete(struct thread_data *td, struct io_u *io_u, + endio_handler *handler) +{ + struct io_completion_data icd; + + init_icd(&icd, handler, 1); + io_completed(td, io_u, &icd); + put_io_u(td, io_u); + + if (!icd.error) + return icd.bytes_done[0] + icd.bytes_done[1]; + + return -1; +} + +long io_u_queued_complete(struct thread_data *td, int min_events, + endio_handler *handler) + +{ + struct timespec ts = { .tv_sec = 0, .tv_nsec = 0, }; + struct timespec *tsp = NULL; + struct io_completion_data icd; + int ret; + + if (min_events > 0) { + tsp = &ts; + ret = td_io_commit(td); + if (ret < 0) { + td_verror(td, -ret); + return ret; + } + } + + ret = td_io_getevents(td, min_events, td->cur_depth, tsp); + if (ret < 0) { + td_verror(td, -ret); + return ret; + } else if (!ret) + return ret; + + init_icd(&icd, handler, ret); + ios_completed(td, &icd); + if (!icd.error) + return icd.bytes_done[0] + icd.bytes_done[1]; + + return -1; +}