summary |
shortlog |
log |
commit | commitdiff |
tree
raw |
patch |
inline | side by side (from parent 1:
e7823a9)
I changed the kernel bits (they aren't merged yet), so update
the fio splice engine to reflect that.
Signed-off-by: Jens Axboe <jens.axboe@oracle.com>
#include <errno.h>
#include <assert.h>
#include <sys/poll.h>
#include <errno.h>
#include <assert.h>
#include <sys/poll.h>
return io_u->xfer_buflen;
}
return io_u->xfer_buflen;
}
-static void splice_unmap_io_u(struct thread_data *td, struct io_u *io_u)
-{
- struct spliceio_data *sd = td->io_ops->data;
- struct iovec iov = {
- .iov_base = io_u->xfer_buf,
- .iov_len = io_u->xfer_buflen,
- };
-
- vmsplice(sd->pipe[0], &iov, 1, SPLICE_F_UNMAP);
-}
-
/*
* We can now vmsplice into userspace, so do the transfer by splicing into
* a pipe and vmsplicing that into userspace.
/*
* We can now vmsplice into userspace, so do the transfer by splicing into
* a pipe and vmsplicing that into userspace.
struct spliceio_data *sd = td->io_ops->data;
struct fio_file *f = io_u->file;
struct iovec iov;
struct spliceio_data *sd = td->io_ops->data;
struct fio_file *f = io_u->file;
struct iovec iov;
+ int ret = 0 , buflen, mmap_len;
- buflen = io_u->xfer_buflen;
- p = io_u->xfer_buf;
- io_u->xfer_buf = NULL;
+ mmap_len = buflen = io_u->xfer_buflen;
+
+ map = mmap(io_u->xfer_buf, buflen, PROT_READ, MAP_PRIVATE|OS_MAP_ANON, 0, 0);
+ if (map == MAP_FAILED) {
+ td_verror(td, errno, "mmap io_u");
+ return -1;
+ }
+
+ p = map;
while (buflen) {
int this_len = buflen;
while (buflen) {
int this_len = buflen;
- if (this_len > SPLICE_DEF_SIZE)
+ if (this_len > SPLICE_DEF_SIZE) {
this_len = SPLICE_DEF_SIZE;
this_len = SPLICE_DEF_SIZE;
+ flags = SPLICE_F_MORE;
+ }
- ret = splice(f->fd, &offset, sd->pipe[1], NULL, this_len, SPLICE_F_MORE);
+ ret = splice(f->fd, &offset, sd->pipe[1], NULL, this_len,flags);
if (ret < 0) {
if (errno == ENODATA || errno == EAGAIN)
continue;
if (ret < 0) {
if (errno == ENODATA || errno == EAGAIN)
continue;
+ td_verror(td, errno, "splice-from-fd");
+ break;
while (iov.iov_len) {
ret = vmsplice(sd->pipe[0], &iov, 1, SPLICE_F_MOVE);
while (iov.iov_len) {
ret = vmsplice(sd->pipe[0], &iov, 1, SPLICE_F_MOVE);
- if (ret < 0)
- return -errno;
- else if (!ret)
- return -ENODATA;
+ if (ret < 0) {
+ td_verror(td, errno, "vmsplice");
+ break;
+ } else if (!ret) {
+ td_verror(td, ENODATA, "vmsplice");
+ ret = -1;
+ break;
+ }
- if (!io_u->xfer_buf)
- io_u->xfer_buf = iov.iov_base;
iov.iov_len -= ret;
iov.iov_base += ret;
}
iov.iov_len -= ret;
iov.iov_base += ret;
}
+ if (ret < 0)
+ break;
+ }
+
+ if (munmap(map, mmap_len) < 0) {
+ td_verror(td, errno, "munnap io_u");
+ return -1;
+ if (ret < 0)
+ return ret;
- io_u->unmap = splice_unmap_io_u;
return io_u->xfer_buflen;
}
return io_u->xfer_buflen;
}
*/
sd->vmsplice_to_user = 1;
*/
sd->vmsplice_to_user = 1;
+ /*
+ * And if vmsplice_to_user works, we definitely need aligned
+ * buffers. Just set ->odirect to force that.
+ */
+ if (td_read(td))
+ td->o.odirect = 1;
+
td->io_ops->data = sd;
return 0;
}
td->io_ops->data = sd;
return 0;
}
* Callback for io completion
*/
int (*end_io)(struct thread_data *, struct io_u *);
* Callback for io completion
*/
int (*end_io)(struct thread_data *, struct io_u *);
-
- /*
- * Called when io_u is about to be freed
- */
- void (*unmap)(struct thread_data *, struct io_u *);
unsigned long usr_time;
unsigned long sys_time;
unsigned long ctx;
unsigned long usr_time;
unsigned long sys_time;
unsigned long ctx;
+ unsigned long minf, majf;
/*
* IO depth and latency stats
/*
* IO depth and latency stats
if (ret && !icd->error)
icd->error = ret;
}
if (ret && !icd->error)
icd->error = ret;
}
- if (io_u->unmap)
- io_u->unmap(td, io_u);
} else {
icd->error = io_u->error;
io_u_log_error(td, io_u);
} else {
icd->error = io_u->error;
io_u_log_error(td, io_u);
-#ifndef SPLICE_F_UNMAP
-#define SPLICE_F_UNMAP (0x10)
-#endif
-
#define SPLICE_DEF_SIZE (64*1024)
#ifdef FIO_HAVE_SYSLET
#define SPLICE_DEF_SIZE (64*1024)
#ifdef FIO_HAVE_SYSLET