struct io_u **events;
unsigned int nr_events;
+ struct async_head_user *ahu;
struct syslet_uatom **ring;
unsigned int ring_index;
};
}
static void init_atom(struct syslet_uatom *atom, int nr, void *arg0,
- void *arg1, void *arg2, void *ret_ptr,
- unsigned long flags, void *priv,struct syslet_uatom *next)
+ void *arg1, void *arg2, void *arg3, void *ret_ptr,
+ unsigned long flags, void *priv)
{
atom->flags = flags;
atom->nr = nr;
atom->ret_ptr = ret_ptr;
- atom->next = next;
+ atom->next = NULL;
atom->arg_ptr[0] = arg0;
atom->arg_ptr[1] = arg1;
atom->arg_ptr[2] = arg2;
- atom->arg_ptr[3] = atom->arg_ptr[4] = atom->arg_ptr[5] = NULL;
+ atom->arg_ptr[3] = arg3;
+ atom->arg_ptr[4] = atom->arg_ptr[5] = NULL;
atom->private = priv;
}
*/
static void fio_syslet_prep_sync(struct io_u *io_u, struct fio_file *f)
{
- init_atom(&io_u->seek_atom.atom, __NR_fsync, &f->fd, NULL, NULL,
- &io_u->seek_atom.ret, SYSLET_STOP_ON_NEGATIVE, io_u, NULL);
+ init_atom(&io_u->req.atom, __NR_fsync, &f->fd, NULL, NULL, NULL,
+ &io_u->req.ret, 0, io_u);
}
static void fio_syslet_prep_rw(struct io_u *io_u, struct fio_file *f)
{
int nr;
- /*
- * prepare seek
- */
- io_u->seek_atom.cmd = SEEK_SET;
- init_atom(&io_u->seek_atom.atom, __NR_lseek, &f->fd, &io_u->offset,
- &io_u->seek_atom.cmd, &io_u->seek_atom.ret,
- SYSLET_STOP_ON_NEGATIVE | SYSLET_NO_COMPLETE |
- SYSLET_SKIP_TO_NEXT_ON_STOP,
- NULL, &io_u->rw_atom.atom);
-
/*
* prepare rw
*/
if (io_u->ddir == DDIR_READ)
- nr = __NR_read;
+ nr = __NR_pread64;
else
- nr = __NR_write;
+ nr = __NR_pwrite64;
- init_atom(&io_u->rw_atom.atom, nr, &f->fd, &io_u->xfer_buf,
- &io_u->xfer_buflen, &io_u->rw_atom.ret,
- SYSLET_STOP_ON_NEGATIVE | SYSLET_SKIP_TO_NEXT_ON_STOP,
- io_u, NULL);
+ init_atom(&io_u->req.atom, nr, &f->fd, &io_u->xfer_buf,
+ &io_u->xfer_buflen, &io_u->offset, &io_u->req.ret, 0, io_u);
}
static int fio_syslet_prep(struct thread_data fio_unused *td, struct io_u *io_u)
static int fio_syslet_queue(struct thread_data *td, struct io_u *io_u)
{
struct syslet_data *sd = td->io_ops->data;
- struct syslet_uatom *done;
long ret;
- done = async_exec(&io_u->seek_atom.atom);
- if (!done)
- return 0;
+ /*
+ * On sync completion, the atom is returned. So on NULL return
+ * it's queued asynchronously.
+ */
+ if (!async_exec(&io_u->req.atom))
+ return FIO_Q_QUEUED;
/*
* completed sync
*/
- ret = io_u->rw_atom.ret;
+ ret = io_u->req.ret;
if (ret != (long) io_u->xfer_buflen) {
if (ret > 0) {
io_u->resid = io_u->xfer_buflen - ret;
io_u->error = 0;
- return ret;
+ return FIO_Q_COMPLETED;
} else
io_u->error = errno;
}
else
td_verror(td, io_u->error);
- return io_u->error;
+ return FIO_Q_COMPLETED;
}
-static void async_head_init(struct syslet_data *sd, unsigned int depth)
+static int async_head_init(struct syslet_data *sd, unsigned int depth)
{
- struct async_head_user ahu;
unsigned long ring_size;
+ sd->ahu = malloc(sizeof(struct async_head_user));
+ memset(sd->ahu, 0, sizeof(struct async_head_user));
+
ring_size = sizeof(struct syslet_uatom *) * depth;
sd->ring = malloc(ring_size);
memset(sd->ring, 0, ring_size);
- memset(&ahu, 0, sizeof(ahu));
- ahu.completion_ring = sd->ring;
- ahu.ring_size_bytes = ring_size;
- ahu.max_nr_threads = -1;
+ sd->ahu->completion_ring = sd->ring;
+ sd->ahu->ring_size_bytes = ring_size;
+ sd->ahu->max_nr_threads = -1;
- if (async_register(&ahu, sizeof(ahu)) < 0)
+ if (async_register(sd->ahu, sizeof(*sd->ahu)) < 0) {
perror("async_register");
+ fprintf(stderr, "fio: syslet likely not supported\n");
+ free(sd->ring);
+ free(sd->ahu);
+ return 1;
+ }
+
+ return 0;
}
-static void async_head_exit(struct syslet_data *sd, unsigned int depth)
+static void async_head_exit(struct syslet_data *sd)
{
- struct async_head_user ahu;
-
- memset(&ahu, 0, sizeof(ahu));
- ahu.completion_ring = sd->ring;
- ahu.ring_size_bytes = sizeof(struct syslet_uatom *) * depth;
-
- if (async_unregister(&ahu, sizeof(ahu)) < 0)
+ if (async_unregister(sd->ahu, sizeof(*sd->ahu)) < 0)
perror("async_register");
+
+ free(sd->ahu);
+ free(sd->ring);
}
static void fio_syslet_cleanup(struct thread_data *td)
struct syslet_data *sd = td->io_ops->data;
if (sd) {
- async_head_exit(sd, td->iodepth);
+ async_head_exit(sd);
free(sd->events);
free(sd);
td->io_ops->data = NULL;
{
struct syslet_data *sd;
+
sd = malloc(sizeof(*sd));
memset(sd, 0, sizeof(*sd));
sd->events = malloc(sizeof(struct io_u *) * td->iodepth);
memset(sd->events, 0, sizeof(struct io_u *) * td->iodepth);
+
+ /*
+ * This will handily fail for kernels where syslet isn't available
+ */
+ if (async_head_init(sd, td->iodepth)) {
+ free(sd->events);
+ free(sd);
+ return 1;
+ }
+
td->io_ops->data = sd;
- async_head_init(sd, td->iodepth);
return 0;
}