X-Git-Url: https://git.kernel.dk/?p=fio.git;a=blobdiff_plain;f=engines%2Fmmap.c;h=55ba1ab36c9a163bec27a788ee05497ef94c6cad;hp=27d5d25515f2016526cbca985e8bf98bc356e326;hb=01fe773df4bc4a35450ce3ef50c8075b3bf55cd0;hpb=b5af82930ccfd7dda6a1b11794efb452eb76d8dc diff --git a/engines/mmap.c b/engines/mmap.c index 27d5d255..55ba1ab3 100644 --- a/engines/mmap.c +++ b/engines/mmap.c @@ -1,135 +1,330 @@ /* - * regular read/write sync io engine + * mmap engine + * + * IO engine that reads/writes from files by doing memcpy to/from + * a memory mapped region of the file. * */ #include #include -#include #include -#include #include #include "../fio.h" -#include "../os.h" +#include "../optgroup.h" +#include "../verify.h" -static int fio_mmapio_queue(struct thread_data *td, struct io_u *io_u) -{ - struct fio_file *f = io_u->file; - unsigned long long real_off = io_u->offset - f->file_offset; +/* + * Limits us to 1GiB of mapped files in total + */ +#define MMAP_TOTAL_SZ (1 * 1024 * 1024 * 1024UL) - if (io_u->ddir == DDIR_READ) - memcpy(io_u->xfer_buf, f->mmap + real_off, io_u->xfer_buflen); - else if (io_u->ddir == DDIR_WRITE) - memcpy(f->mmap + real_off, io_u->xfer_buf, io_u->xfer_buflen); - else if (io_u->ddir == DDIR_SYNC) { - if (msync(f->mmap, f->file_size, MS_SYNC)) - io_u->error = errno; - } +static unsigned long mmap_map_size; - /* - * not really direct, but should drop the pages from the cache - */ - if (td->odirect && io_u->ddir != DDIR_SYNC) { - if (msync(f->mmap + real_off, io_u->xfer_buflen, MS_SYNC) < 0) - io_u->error = errno; - if (madvise(f->mmap + real_off, io_u->xfer_buflen, MADV_DONTNEED) < 0) - io_u->error = errno; - } +struct fio_mmap_data { + void *mmap_ptr; + size_t mmap_sz; + off_t mmap_off; +}; - if (io_u->error) - td_verror(td, io_u->error, "sync"); +#ifdef CONFIG_HAVE_THP +struct mmap_options { + void *pad; + unsigned int thp; +}; - return FIO_Q_COMPLETED; -} +static struct fio_option options[] = { + { + .name = "thp", + .lname = "Transparent Huge Pages", + .type = FIO_OPT_INT, + .off1 = offsetof(struct mmap_options, thp), + .help = "Memory Advise Huge Page", + .category = FIO_OPT_C_ENGINE, + .group = FIO_OPT_G_MMAP, + }, + { + .name = NULL, + }, +}; +#endif + +static bool fio_madvise_file(struct thread_data *td, struct fio_file *f, + size_t length) -static int fio_mmapio_init(struct thread_data *td) { - struct fio_file *f; - int i; + struct fio_mmap_data *fmd = FILE_ENG_DATA(f); +#ifdef CONFIG_HAVE_THP + struct mmap_options *o = td->eo; - if (!td_write(td)) - return 0; + /* Ignore errors on this optional advisory */ + if (o->thp) + madvise(fmd->mmap_ptr, length, MADV_HUGEPAGE); +#endif - /* - * We need to truncate the files to the right size, if - * we are writing to it. - */ - for_each_file(td, f, i) { - if (ftruncate(f->fd, f->file_size) < 0) { - td_verror(td, errno, "ftruncate"); - return 1; + if (!td->o.fadvise_hint) + return true; + + if (!td_random(td)) { + if (posix_madvise(fmd->mmap_ptr, length, POSIX_MADV_SEQUENTIAL) < 0) { + td_verror(td, errno, "madvise"); + return false; + } + } else { + if (posix_madvise(fmd->mmap_ptr, length, POSIX_MADV_RANDOM) < 0) { + td_verror(td, errno, "madvise"); + return false; } } - return 0; + return true; } -static int fio_mmapio_open(struct thread_data *td, struct fio_file *f) +#ifdef CONFIG_HAVE_THP +static int fio_mmap_get_shared(struct thread_data *td) { - int ret, flags; + struct mmap_options *o = td->eo; - ret = generic_open_file(td, f); - if (ret) - return ret; + if (o->thp) + return MAP_PRIVATE; + return MAP_SHARED; +} +#else +static int fio_mmap_get_shared(struct thread_data *td) +{ + return MAP_SHARED; +} +#endif - if (td_rw(td)) +static int fio_mmap_file(struct thread_data *td, struct fio_file *f, + size_t length, off_t off) +{ + struct fio_mmap_data *fmd = FILE_ENG_DATA(f); + int flags = 0, shared = fio_mmap_get_shared(td); + + if (td_rw(td) && !td->o.verify_only) flags = PROT_READ | PROT_WRITE; - else if (td_write(td)) { + else if (td_write(td) && !td->o.verify_only) { flags = PROT_WRITE; - if (td->verify != VERIFY_NONE) + if (td->o.verify != VERIFY_NONE) flags |= PROT_READ; } else flags = PROT_READ; - f->mmap = mmap(NULL, f->file_size, flags, MAP_SHARED, f->fd, f->file_offset); - if (f->mmap == MAP_FAILED) { - f->mmap = NULL; + fmd->mmap_ptr = mmap(NULL, length, flags, shared, f->fd, off); + if (fmd->mmap_ptr == MAP_FAILED) { + fmd->mmap_ptr = NULL; td_verror(td, errno, "mmap"); goto err; } - if (file_invalidate_cache(td, f)) + if (!fio_madvise_file(td, f, length)) goto err; - if (!td_random(td)) { - if (madvise(f->mmap, f->file_size, MADV_SEQUENTIAL) < 0) { - td_verror(td, errno, "madvise"); - goto err; + if (posix_madvise(fmd->mmap_ptr, length, POSIX_MADV_DONTNEED) < 0) { + td_verror(td, errno, "madvise"); + goto err; + } + +#ifdef FIO_MADV_FREE + if (f->filetype == FIO_TYPE_BLOCK) + (void) posix_madvise(fmd->mmap_ptr, fmd->mmap_sz, FIO_MADV_FREE); +#endif + +err: + if (td->error && fmd->mmap_ptr) + munmap(fmd->mmap_ptr, length); + + return td->error; +} + +/* + * Just mmap an appropriate portion, we cannot mmap the full extent + */ +static int fio_mmapio_prep_limited(struct thread_data *td, struct io_u *io_u) +{ + struct fio_file *f = io_u->file; + struct fio_mmap_data *fmd = FILE_ENG_DATA(f); + + if (io_u->buflen > mmap_map_size) { + log_err("fio: bs too big for mmap engine\n"); + return EIO; + } + + fmd->mmap_sz = mmap_map_size; + if (fmd->mmap_sz > f->io_size) + fmd->mmap_sz = f->io_size; + + fmd->mmap_off = io_u->offset; + + return fio_mmap_file(td, f, fmd->mmap_sz, fmd->mmap_off); +} + +/* + * Attempt to mmap the entire file + */ +static int fio_mmapio_prep_full(struct thread_data *td, struct io_u *io_u) +{ + struct fio_file *f = io_u->file; + struct fio_mmap_data *fmd = FILE_ENG_DATA(f); + int ret; + + if (fio_file_partial_mmap(f)) + return EINVAL; + if (io_u->offset != (size_t) io_u->offset || + f->io_size != (size_t) f->io_size) { + fio_file_set_partial_mmap(f); + return EINVAL; + } + + fmd->mmap_sz = f->io_size; + fmd->mmap_off = 0; + + ret = fio_mmap_file(td, f, fmd->mmap_sz, fmd->mmap_off); + if (ret) + fio_file_set_partial_mmap(f); + + return ret; +} + +static int fio_mmapio_prep(struct thread_data *td, struct io_u *io_u) +{ + struct fio_file *f = io_u->file; + struct fio_mmap_data *fmd = FILE_ENG_DATA(f); + int ret; + + /* + * It fits within existing mapping, use it + */ + if (io_u->offset >= fmd->mmap_off && + io_u->offset + io_u->buflen <= fmd->mmap_off + fmd->mmap_sz) + goto done; + + /* + * unmap any existing mapping + */ + if (fmd->mmap_ptr) { + if (munmap(fmd->mmap_ptr, fmd->mmap_sz) < 0) + return errno; + fmd->mmap_ptr = NULL; + } + + if (fio_mmapio_prep_full(td, io_u)) { + td_clear_error(td); + ret = fio_mmapio_prep_limited(td, io_u); + if (ret) + return ret; + } + +done: + io_u->mmap_data = fmd->mmap_ptr + io_u->offset - fmd->mmap_off - + f->file_offset; + return 0; +} + +static enum fio_q_status fio_mmapio_queue(struct thread_data *td, + struct io_u *io_u) +{ + struct fio_file *f = io_u->file; + struct fio_mmap_data *fmd = FILE_ENG_DATA(f); + + fio_ro_check(td, io_u); + + if (io_u->ddir == DDIR_READ) + memcpy(io_u->xfer_buf, io_u->mmap_data, io_u->xfer_buflen); + else if (io_u->ddir == DDIR_WRITE) + memcpy(io_u->mmap_data, io_u->xfer_buf, io_u->xfer_buflen); + else if (ddir_sync(io_u->ddir)) { + if (msync(fmd->mmap_ptr, fmd->mmap_sz, MS_SYNC)) { + io_u->error = errno; + td_verror(td, io_u->error, "msync"); } - } else { - if (madvise(f->mmap, f->file_size, MADV_RANDOM) < 0) { - td_verror(td, errno, "madvise"); - goto err; + } else if (io_u->ddir == DDIR_TRIM) { + int ret = do_io_u_trim(td, io_u); + + if (!ret) + td_verror(td, io_u->error, "trim"); + } + + + /* + * not really direct, but should drop the pages from the cache + */ + if (td->o.odirect && ddir_rw(io_u->ddir)) { + if (msync(io_u->mmap_data, io_u->xfer_buflen, MS_SYNC) < 0) { + io_u->error = errno; + td_verror(td, io_u->error, "msync"); + } + if (posix_madvise(io_u->mmap_data, io_u->xfer_buflen, POSIX_MADV_DONTNEED) < 0) { + io_u->error = errno; + td_verror(td, io_u->error, "madvise"); } } - return 0; + return FIO_Q_COMPLETED; +} -err: - if (f->mmap) - munmap(f->mmap, f->file_size); - generic_close_file(td, f); - return 1; +static int fio_mmapio_init(struct thread_data *td) +{ + struct thread_options *o = &td->o; + + if ((o->rw_min_bs & page_mask) && + (o->odirect || o->fsync_blocks || o->fdatasync_blocks)) { + log_err("fio: mmap options dictate a minimum block size of " + "%llu bytes\n", (unsigned long long) page_size); + return 1; + } + + mmap_map_size = MMAP_TOTAL_SZ / o->nr_files; + return 0; } -static void fio_mmapio_close(struct thread_data fio_unused *td, - struct fio_file *f) +static int fio_mmapio_open_file(struct thread_data *td, struct fio_file *f) { - if (f->mmap) { - munmap(f->mmap, f->file_size); - f->mmap = NULL; + struct fio_mmap_data *fmd; + int ret; + + ret = generic_open_file(td, f); + if (ret) + return ret; + + fmd = calloc(1, sizeof(*fmd)); + if (!fmd) { + int fio_unused __ret; + __ret = generic_close_file(td, f); + return 1; } + + FILE_SET_ENG_DATA(f, fmd); + return 0; +} + +static int fio_mmapio_close_file(struct thread_data *td, struct fio_file *f) +{ + struct fio_mmap_data *fmd = FILE_ENG_DATA(f); + + FILE_SET_ENG_DATA(f, NULL); + free(fmd); + fio_file_clear_partial_mmap(f); + + return generic_close_file(td, f); } static struct ioengine_ops ioengine = { .name = "mmap", .version = FIO_IOOPS_VERSION, - .queue = fio_mmapio_queue, .init = fio_mmapio_init, - .open_file = fio_mmapio_open, - .close_file = fio_mmapio_close, - .flags = FIO_SYNCIO, + .prep = fio_mmapio_prep, + .queue = fio_mmapio_queue, + .open_file = fio_mmapio_open_file, + .close_file = fio_mmapio_close_file, + .get_file_size = generic_get_file_size, + .flags = FIO_SYNCIO | FIO_NOEXTEND, +#ifdef CONFIG_HAVE_THP + .options = options, + .option_struct_size = sizeof(struct mmap_options), +#endif }; static void fio_init fio_mmapio_register(void)