X-Git-Url: https://git.kernel.dk/?p=fio.git;a=blobdiff_plain;f=engines%2Fsyslet-rw.c;h=2a3095d405c7f0a80ef773877e24acbcdb0009f9;hp=380d19d44c4e3be8ba1af37853322d9c8b87a25a;hb=7756b0d046ef74c32df134c8b716237cc5e16bad;hpb=db64e9bc34d621647a255c918a5b7416ca0b6076;ds=sidebyside diff --git a/engines/syslet-rw.c b/engines/syslet-rw.c index 380d19d4..2a3095d4 100644 --- a/engines/syslet-rw.c +++ b/engines/syslet-rw.c @@ -17,30 +17,45 @@ struct syslet_data { struct io_u **events; unsigned int nr_events; + struct async_head_user ahu; struct syslet_uatom **ring; - unsigned int ring_index; + + struct syslet_uatom *head, *tail; + struct syslet_uatom **event_map; + unsigned int event_map_idx; }; -/* - * Inspect the ring to see if we have completed events - */ -static void fio_syslet_complete(struct thread_data *td) +static void fio_syslet_complete_atom(struct thread_data *td, + struct syslet_uatom *atom) { struct syslet_data *sd = td->io_ops->data; + struct io_u *io_u; + int i, end; - do { - struct syslet_uatom *atom; - struct io_u *io_u; - long ret; + if (!sd->event_map_idx) + return; - atom = sd->ring[sd->ring_index]; - if (!atom) + /* + * Find the start of the string of atoms for this sequence + */ + for (end = sd->event_map_idx - 1; end >= 0; end--) + if (atom == sd->event_map[end]) break; - sd->ring[sd->ring_index] = NULL; - if (++sd->ring_index == td->iodepth) - sd->ring_index = 0; + if (end < 0 || atom != sd->event_map[end]) { + printf("didn't find atom\n"); + return; + } + //printf("end=%d, total %d\n", end, sd->event_map_idx); + + /* + * now complete in right order + */ + for (i = 0; i <= end; i++) { + long ret; + + atom = sd->event_map[i]; io_u = atom->private; ret = *atom->ret_ptr; if (ret > 0) @@ -48,7 +63,42 @@ static void fio_syslet_complete(struct thread_data *td) else if (ret < 0) io_u->error = ret; + assert(sd->nr_events < td->iodepth); sd->events[sd->nr_events++] = io_u; + } + + /* + * Move later completions to the front, if we didn't complete all + */ + if (end == (int) sd->event_map_idx - 1) + sd->event_map_idx = 0; + else { + int nr = sd->event_map_idx - end - 1; + + memmove(sd->event_map, &sd->event_map[end + 1], nr * sizeof(struct syslet_uatom *)); + sd->event_map_idx = nr; + } +} + +/* + * Inspect the ring to see if we have completed events + */ +static void fio_syslet_complete(struct thread_data *td) +{ + struct syslet_data *sd = td->io_ops->data; + + do { + struct syslet_uatom *atom; + + atom = sd->ring[sd->ahu.user_ring_idx]; + if (!atom) + break; + + sd->ring[sd->ahu.user_ring_idx] = NULL; + if (++sd->ahu.user_ring_idx == td->iodepth) + sd->ahu.user_ring_idx = 0; + + fio_syslet_complete_atom(td, atom); } while (1); } @@ -57,7 +107,6 @@ static int fio_syslet_getevents(struct thread_data *td, int min, struct timespec fio_unused *t) { struct syslet_data *sd = td->io_ops->data; - int get_events; long ret; do { @@ -72,10 +121,9 @@ static int fio_syslet_getevents(struct thread_data *td, int min, /* * OK, we need to wait for some events... */ - get_events = min - sd->nr_events; - ret = async_wait(get_events); + ret = async_wait(1, sd->ahu.user_ring_idx, &sd->ahu); if (ret < 0) - return errno; + return -errno; } while (1); ret = sd->nr_events; @@ -91,17 +139,18 @@ static struct io_u *fio_syslet_event(struct thread_data *td, int event) } static void init_atom(struct syslet_uatom *atom, int nr, void *arg0, - void *arg1, void *arg2, void *ret_ptr, - unsigned long flags, void *priv,struct syslet_uatom *next) + void *arg1, void *arg2, void *arg3, void *ret_ptr, + unsigned long flags, void *priv) { atom->flags = flags; atom->nr = nr; atom->ret_ptr = ret_ptr; - atom->next = next; + atom->next = NULL; atom->arg_ptr[0] = arg0; atom->arg_ptr[1] = arg1; atom->arg_ptr[2] = arg2; - atom->arg_ptr[3] = atom->arg_ptr[4] = atom->arg_ptr[5] = NULL; + atom->arg_ptr[3] = arg3; + atom->arg_ptr[4] = atom->arg_ptr[5] = NULL; atom->private = priv; } @@ -110,36 +159,24 @@ static void init_atom(struct syslet_uatom *atom, int nr, void *arg0, */ static void fio_syslet_prep_sync(struct io_u *io_u, struct fio_file *f) { - init_atom(&io_u->seek_atom.atom, __NR_fsync, &f->fd, NULL, NULL, - &io_u->seek_atom.ret, SYSLET_STOP_ON_NEGATIVE, io_u, NULL); + init_atom(&io_u->req.atom, __NR_fsync, &f->fd, NULL, NULL, NULL, + &io_u->req.ret, 0, io_u); } static void fio_syslet_prep_rw(struct io_u *io_u, struct fio_file *f) { int nr; - /* - * prepare seek - */ - io_u->seek_atom.cmd = SEEK_SET; - init_atom(&io_u->seek_atom.atom, __NR_lseek, &f->fd, &io_u->offset, - &io_u->seek_atom.cmd, &io_u->seek_atom.ret, - SYSLET_STOP_ON_NEGATIVE | SYSLET_NO_COMPLETE | - SYSLET_SKIP_TO_NEXT_ON_STOP, - NULL, &io_u->rw_atom.atom); - /* * prepare rw */ if (io_u->ddir == DDIR_READ) - nr = __NR_read; + nr = __NR_pread64; else - nr = __NR_write; + nr = __NR_pwrite64; - init_atom(&io_u->rw_atom.atom, nr, &f->fd, &io_u->xfer_buf, - &io_u->xfer_buflen, &io_u->rw_atom.ret, - SYSLET_STOP_ON_NEGATIVE | SYSLET_SKIP_TO_NEXT_ON_STOP, - io_u, NULL); + init_atom(&io_u->req.atom, nr, &f->fd, &io_u->xfer_buf, + &io_u->xfer_buflen, &io_u->offset, &io_u->req.ret, 0, io_u); } static int fio_syslet_prep(struct thread_data fio_unused *td, struct io_u *io_u) @@ -154,71 +191,81 @@ static int fio_syslet_prep(struct thread_data fio_unused *td, struct io_u *io_u) return 0; } -static int fio_syslet_queue(struct thread_data *td, struct io_u *io_u) +static void cachemiss_thread_start(void) +{ + while (1) + async_thread(NULL, NULL); +} + +#define THREAD_STACK_SIZE (16384) + +static unsigned long thread_stack_alloc() +{ + return (unsigned long)malloc(THREAD_STACK_SIZE) + THREAD_STACK_SIZE; +} + +static int fio_syslet_commit(struct thread_data *td) { struct syslet_data *sd = td->io_ops->data; struct syslet_uatom *done; - long ret; - done = async_exec(&io_u->seek_atom.atom); - if (!done) + if (!sd->head) return 0; + if (!sd->ahu.new_thread_stack) + sd->ahu.new_thread_stack = thread_stack_alloc(); + /* - * completed sync + * On sync completion, the atom is returned. So on NULL return + * it's queued asynchronously. */ - ret = io_u->rw_atom.ret; - if (ret != (long) io_u->xfer_buflen) { - if (ret > 0) { - io_u->resid = io_u->xfer_buflen - ret; - io_u->error = 0; - return ret; - } else - io_u->error = errno; - } + done = async_exec(sd->head, &sd->ahu); - if (!io_u->error) - sd->events[sd->nr_events++] = io_u; - else - td_verror(td, io_u->error); + sd->head = sd->tail = NULL; + + if (done) + fio_syslet_complete_atom(td, done); + + return 0; +} + +static int fio_syslet_queue(struct thread_data *td, struct io_u *io_u) +{ + struct syslet_data *sd = td->io_ops->data; - return io_u->error; + if (sd->tail) { + sd->tail->next = &io_u->req.atom; + sd->tail = &io_u->req.atom; + } else + sd->head = sd->tail = &io_u->req.atom; + + sd->event_map[sd->event_map_idx++] = sd->tail; + return FIO_Q_QUEUED; } static int async_head_init(struct syslet_data *sd, unsigned int depth) { - struct async_head_user ahu; unsigned long ring_size; + memset(&sd->ahu, 0, sizeof(struct async_head_user)); + ring_size = sizeof(struct syslet_uatom *) * depth; sd->ring = malloc(ring_size); memset(sd->ring, 0, ring_size); - memset(&ahu, 0, sizeof(ahu)); - ahu.completion_ring = sd->ring; - ahu.ring_size_bytes = ring_size; - ahu.max_nr_threads = -1; - - if (async_register(&ahu, sizeof(ahu)) < 0) { - perror("async_register"); - fprintf(stderr, "fio: syslet likely not supported\n"); - free(sd->ring); - return 1; - } + sd->ahu.user_ring_idx = 0; + sd->ahu.completion_ring = sd->ring; + sd->ahu.ring_size_bytes = ring_size; + sd->ahu.head_stack = thread_stack_alloc(); + sd->ahu.head_eip = (unsigned long)cachemiss_thread_start; + sd->ahu.new_thread_eip = (unsigned long)cachemiss_thread_start; return 0; } -static void async_head_exit(struct syslet_data *sd, unsigned int depth) +static void async_head_exit(struct syslet_data *sd) { - struct async_head_user ahu; - - memset(&ahu, 0, sizeof(ahu)); - ahu.completion_ring = sd->ring; - ahu.ring_size_bytes = sizeof(struct syslet_uatom *) * depth; - - if (async_unregister(&ahu, sizeof(ahu)) < 0) - perror("async_register"); + free(sd->ring); } static void fio_syslet_cleanup(struct thread_data *td) @@ -226,8 +273,9 @@ static void fio_syslet_cleanup(struct thread_data *td) struct syslet_data *sd = td->io_ops->data; if (sd) { - async_head_exit(sd, td->iodepth); + async_head_exit(sd); free(sd->events); + free(sd->event_map); free(sd); td->io_ops->data = NULL; } @@ -242,6 +290,8 @@ static int fio_syslet_init(struct thread_data *td) memset(sd, 0, sizeof(*sd)); sd->events = malloc(sizeof(struct io_u *) * td->iodepth); memset(sd->events, 0, sizeof(struct io_u *) * td->iodepth); + sd->event_map = malloc(sizeof(struct syslet_uatom *) * td->iodepth); + memset(sd->event_map, 0, sizeof(struct syslet_uatom *) * td->iodepth); /* * This will handily fail for kernels where syslet isn't available @@ -262,6 +312,7 @@ static struct ioengine_ops ioengine = { .init = fio_syslet_init, .prep = fio_syslet_prep, .queue = fio_syslet_queue, + .commit = fio_syslet_commit, .getevents = fio_syslet_getevents, .event = fio_syslet_event, .cleanup = fio_syslet_cleanup,