+static int fio_mmapio_queue(struct thread_data *td, struct io_u *io_u)
+{
+ struct fio_file *f = io_u->file;
+
+ fio_ro_check(td, io_u);
+
+ if (io_u->ddir == DDIR_READ)
+ memcpy(io_u->xfer_buf, io_u->mmap_data, io_u->xfer_buflen);
+ else if (io_u->ddir == DDIR_WRITE)
+ memcpy(io_u->mmap_data, io_u->xfer_buf, io_u->xfer_buflen);
+ else if (io_u->ddir == DDIR_SYNC) {
+ if (msync(f->mmap_ptr, f->mmap_sz, MS_SYNC)) {
+ io_u->error = errno;
+ td_verror(td, io_u->error, "msync");
+ }
+ }
+
+ /*
+ * not really direct, but should drop the pages from the cache
+ */
+ if (td->o.odirect && io_u->ddir != DDIR_SYNC) {
+ if (msync(io_u->mmap_data, io_u->xfer_buflen, MS_SYNC) < 0) {
+ io_u->error = errno;
+ td_verror(td, io_u->error, "msync");
+ }
+ if (madvise(io_u->mmap_data, io_u->xfer_buflen, MADV_DONTNEED) < 0) {
+ io_u->error = errno;
+ td_verror(td, io_u->error, "madvise");
+ }
+ }
+
+ return FIO_Q_COMPLETED;
+}
+
+static int fio_mmapio_init(struct thread_data *td)
+{
+ unsigned long shift, mask;
+
+ mmap_map_size = MMAP_TOTAL_SZ / td->o.nr_files;
+ mask = mmap_map_size;
+ shift = 0;
+ do {
+ mask >>= 1;
+ if (!mask)
+ break;
+ shift++;
+ } while (1);
+
+ mmap_map_mask = 1UL << shift;
+ return 0;
+}
+