X-Git-Url: https://git.kernel.dk/?a=blobdiff_plain;f=t%2Fio_uring.c;h=538cc7d4e44c99b0dfbde4502780334be7d89b6a;hb=543196617ce45b6a04fb039a3b9c6d06c9b58309;hp=39332c8e6cdc09df0e7ae769bec86059ee6c8a12;hpb=0e47f11b8342ea702fc7cc27929eb5a79e4597a3;p=fio.git diff --git a/t/io_uring.c b/t/io_uring.c index 39332c8e..538cc7d4 100644 --- a/t/io_uring.c +++ b/t/io_uring.c @@ -22,9 +22,7 @@ #include "../arch/arch.h" #include "../lib/types.h" -#include "../os/io_uring.h" - -#define barrier() __asm__ __volatile__("": : :"memory") +#include "../os/linux/io_uring.h" #define min(a, b) ((a < b) ? (a) : (b)) @@ -33,6 +31,7 @@ struct io_sq_ring { unsigned *tail; unsigned *ring_mask; unsigned *ring_entries; + unsigned *flags; unsigned *array; }; @@ -44,11 +43,9 @@ struct io_cq_ring { struct io_uring_cqe *cqes; }; -#define DEPTH 32 - -#define BATCH_SUBMIT 8 -#define BATCH_COMPLETE 8 - +#define DEPTH 128 +#define BATCH_SUBMIT 32 +#define BATCH_COMPLETE 32 #define BS 4096 #define MAX_FDS 16 @@ -57,64 +54,122 @@ static unsigned sq_ring_mask, cq_ring_mask; struct file { unsigned long max_blocks; - int fd; + unsigned pending_ios; + int real_fd; + int fixed_fd; }; struct submitter { pthread_t thread; int ring_fd; - struct drand48_data rand; + int index; struct io_sq_ring sq_ring; struct io_uring_sqe *sqes; - struct iovec iovecs[DEPTH]; struct io_cq_ring cq_ring; int inflight; unsigned long reaps; unsigned long done; unsigned long calls; - unsigned long cachehit, cachemiss; volatile int finish; + + __s32 *fds; + struct file files[MAX_FDS]; unsigned nr_files; unsigned cur_file; + struct iovec iovecs[]; }; -static struct submitter submitters[1]; +static struct submitter *submitter; static volatile int finish; +static int depth = DEPTH; +static int batch_submit = BATCH_SUBMIT; +static int batch_complete = BATCH_COMPLETE; +static int bs = BS; static int polled = 1; /* use IO polling */ -static int fixedbufs = 0; /* use fixed user buffers */ +static int fixedbufs = 1; /* use fixed user buffers */ +static int register_files = 1; /* use fixed files */ static int buffered = 0; /* use buffered IO, not O_DIRECT */ static int sq_thread_poll = 0; /* use kernel submission/poller thread */ static int sq_thread_cpu = -1; /* pin above thread to this CPU */ +static int do_nop = 0; /* no-op SQ ring commands */ +static int nthreads = 1; + +static int vectored = 1; static int io_uring_register_buffers(struct submitter *s) { - struct io_uring_register_buffers reg = { - .iovecs = s->iovecs, - .nr_iovecs = DEPTH - }; + if (do_nop) + return 0; - return syscall(__NR_sys_io_uring_register, s->ring_fd, - IORING_REGISTER_BUFFERS, ®); + return syscall(__NR_io_uring_register, s->ring_fd, + IORING_REGISTER_BUFFERS, s->iovecs, depth); +} + +static int io_uring_register_files(struct submitter *s) +{ + int i; + + if (do_nop) + return 0; + + s->fds = calloc(s->nr_files, sizeof(__s32)); + for (i = 0; i < s->nr_files; i++) { + s->fds[i] = s->files[i].real_fd; + s->files[i].fixed_fd = i; + } + + return syscall(__NR_io_uring_register, s->ring_fd, + IORING_REGISTER_FILES, s->fds, s->nr_files); } static int io_uring_setup(unsigned entries, struct io_uring_params *p) { - return syscall(__NR_sys_io_uring_setup, entries, p); + return syscall(__NR_io_uring_setup, entries, p); +} + +static void io_uring_probe(int fd) +{ + struct io_uring_probe *p; + int ret; + + p = malloc(sizeof(*p) + 256 * sizeof(struct io_uring_probe_op)); + if (!p) + return; + + memset(p, 0, sizeof(*p) + 256 * sizeof(struct io_uring_probe_op)); + ret = syscall(__NR_io_uring_register, fd, IORING_REGISTER_PROBE, p, 256); + if (ret < 0) + goto out; + + if (IORING_OP_READ > p->ops_len) + goto out; + + if ((p->ops[IORING_OP_READ].flags & IO_URING_OP_SUPPORTED)) + vectored = 0; +out: + free(p); } static int io_uring_enter(struct submitter *s, unsigned int to_submit, unsigned int min_complete, unsigned int flags) { - return syscall(__NR_sys_io_uring_enter, s->ring_fd, to_submit, - min_complete, flags); + return syscall(__NR_io_uring_enter, s->ring_fd, to_submit, min_complete, + flags, NULL, 0); } +#ifndef CONFIG_HAVE_GETTID static int gettid(void) { return syscall(__NR_gettid); } +#endif + +static unsigned file_depth(struct submitter *s) +{ + return (depth + s->nr_files - 1) / s->nr_files; +} static void init_io(struct submitter *s, unsigned index) { @@ -123,29 +178,53 @@ static void init_io(struct submitter *s, unsigned index) struct file *f; long r; - f = &s->files[s->cur_file]; - s->cur_file++; - if (s->cur_file == s->nr_files) - s->cur_file = 0; + if (do_nop) { + sqe->opcode = IORING_OP_NOP; + return; + } - lrand48_r(&s->rand, &r); - offset = (r % (f->max_blocks - 1)) * BS; + if (s->nr_files == 1) { + f = &s->files[0]; + } else { + f = &s->files[s->cur_file]; + if (f->pending_ios >= file_depth(s)) { + s->cur_file++; + if (s->cur_file == s->nr_files) + s->cur_file = 0; + f = &s->files[s->cur_file]; + } + } + f->pending_ios++; - sqe->flags = 0; - sqe->opcode = IORING_OP_READV; + r = lrand48(); + offset = (r % (f->max_blocks - 1)) * bs; + + if (register_files) { + sqe->flags = IOSQE_FIXED_FILE; + sqe->fd = f->fixed_fd; + } else { + sqe->flags = 0; + sqe->fd = f->real_fd; + } if (fixedbufs) { - sqe->addr = s->iovecs[index].iov_base; - sqe->len = BS; + sqe->opcode = IORING_OP_READ_FIXED; + sqe->addr = (unsigned long) s->iovecs[index].iov_base; + sqe->len = bs; sqe->buf_index = index; - sqe->flags |= IOSQE_FIXED_BUFFER; + } else if (!vectored) { + sqe->opcode = IORING_OP_READ; + sqe->addr = (unsigned long) s->iovecs[index].iov_base; + sqe->len = bs; + sqe->buf_index = 0; } else { - sqe->addr = &s->iovecs[index]; + sqe->opcode = IORING_OP_READV; + sqe->addr = (unsigned long) &s->iovecs[index]; sqe->len = 1; sqe->buf_index = 0; } sqe->ioprio = 0; - sqe->fd = f->fd; sqe->off = offset; + sqe->user_data = (unsigned long) f; } static int prep_more_ios(struct submitter *s, int max_ios) @@ -156,8 +235,7 @@ static int prep_more_ios(struct submitter *s, int max_ios) next_tail = tail = *ring->tail; do { next_tail++; - barrier(); - if (next_tail == *ring->head) + if (next_tail == atomic_load_acquire(ring->head)) break; index = tail & sq_ring_mask; @@ -167,12 +245,8 @@ static int prep_more_ios(struct submitter *s, int max_ios) tail = next_tail; } while (prepped < max_ios); - if (*ring->tail != tail) { - /* order tail store with writes to sqes above */ - barrier(); - *ring->tail = tail; - barrier(); - } + if (prepped) + atomic_store_release(ring->tail, tail); return prepped; } @@ -180,18 +254,18 @@ static int get_file_size(struct file *f) { struct stat st; - if (fstat(f->fd, &st) < 0) + if (fstat(f->real_fd, &st) < 0) return -1; if (S_ISBLK(st.st_mode)) { unsigned long long bytes; - if (ioctl(f->fd, BLKGETSIZE64, &bytes) != 0) + if (ioctl(f->real_fd, BLKGETSIZE64, &bytes) != 0) return -1; - f->max_blocks = bytes / BS; + f->max_blocks = bytes / bs; return 0; } else if (S_ISREG(st.st_mode)) { - f->max_blocks = st.st_size / BS; + f->max_blocks = st.st_size / bs; return 0; } @@ -206,61 +280,96 @@ static int reap_events(struct submitter *s) head = *ring->head; do { - barrier(); - if (head == *ring->tail) + struct file *f; + + read_barrier(); + if (head == atomic_load_acquire(ring->tail)) break; cqe = &ring->cqes[head & cq_ring_mask]; - if (cqe->res != BS) { - printf("io: unexpected ret=%d\n", cqe->res); - return -1; + if (!do_nop) { + f = (struct file *) (uintptr_t) cqe->user_data; + f->pending_ios--; + if (cqe->res != bs) { + printf("io: unexpected ret=%d\n", cqe->res); + if (polled && cqe->res == -EOPNOTSUPP) + printf("Your filesystem/driver/kernel doesn't support polled IO\n"); + return -1; + } } - if (cqe->flags & IOCQE_FLAG_CACHEHIT) - s->cachehit++; - else - s->cachemiss++; reaped++; head++; } while (1); - s->inflight -= reaped; - *ring->head = head; - barrier(); + if (reaped) { + s->inflight -= reaped; + atomic_store_release(ring->head, head); + } return reaped; } static void *submitter_fn(void *data) { struct submitter *s = data; + struct io_sq_ring *ring = &s->sq_ring; int ret, prepped; printf("submitter=%d\n", gettid()); - srand48_r(pthread_self(), &s->rand); + srand48(pthread_self()); prepped = 0; do { int to_wait, to_submit, this_reap, to_prep; + unsigned ring_flags = 0; - if (!prepped && s->inflight < DEPTH) { - to_prep = min(DEPTH - s->inflight, BATCH_SUBMIT); + if (!prepped && s->inflight < depth) { + to_prep = min(depth - s->inflight, batch_submit); prepped = prep_more_ios(s, to_prep); } s->inflight += prepped; submit_more: to_submit = prepped; submit: - if (s->inflight + BATCH_SUBMIT < DEPTH) + if (to_submit && (s->inflight + to_submit <= depth)) to_wait = 0; else - to_wait = min(s->inflight + to_submit, BATCH_COMPLETE); - - ret = io_uring_enter(s, to_submit, to_wait, - IORING_ENTER_GETEVENTS); - s->calls++; + to_wait = min(s->inflight + to_submit, batch_complete); + + /* + * Only need to call io_uring_enter if we're not using SQ thread + * poll, or if IORING_SQ_NEED_WAKEUP is set. + */ + if (sq_thread_poll) + ring_flags = atomic_load_acquire(ring->flags); + if (!sq_thread_poll || ring_flags & IORING_SQ_NEED_WAKEUP) { + unsigned flags = 0; + + if (to_wait) + flags = IORING_ENTER_GETEVENTS; + if (ring_flags & IORING_SQ_NEED_WAKEUP) + flags |= IORING_ENTER_SQ_WAKEUP; + ret = io_uring_enter(s, to_submit, to_wait, flags); + s->calls++; + } else { + /* for SQPOLL, we submitted it all effectively */ + ret = to_submit; + } - this_reap = reap_events(s); - if (this_reap == -1) - break; + /* + * For non SQ thread poll, we already got the events we needed + * through the io_uring_enter() above. For SQ thread poll, we + * need to loop here until we find enough events. + */ + this_reap = 0; + do { + int r; + r = reap_events(s); + if (r == -1) { + s->finish = 1; + break; + } else if (r > 0) + this_reap += r; + } while (sq_thread_poll && this_reap < to_wait); s->reaps += this_reap; if (ret >= 0) { @@ -297,10 +406,25 @@ submit: return NULL; } +static struct submitter *get_submitter(int offset) +{ + void *ret; + + ret = submitter; + if (offset) + ret += offset * (sizeof(*submitter) + depth * sizeof(struct iovec)); + return ret; +} + static void sig_int(int sig) { + int j; + printf("Exiting on signal %d\n", sig); - submitters[0].finish = 1; + for (j = 0; j < nthreads; j++) { + struct submitter *s = get_submitter(j); + s->finish = 1; + } finish = 1; } @@ -324,25 +448,37 @@ static int setup_ring(struct submitter *s) memset(&p, 0, sizeof(p)); - if (polled) + if (polled && !do_nop) p.flags |= IORING_SETUP_IOPOLL; if (sq_thread_poll) { p.flags |= IORING_SETUP_SQPOLL; - if (sq_thread_cpu != -1) + if (sq_thread_cpu != -1) { p.flags |= IORING_SETUP_SQ_AFF; + p.sq_thread_cpu = sq_thread_cpu; + } } - fd = io_uring_setup(DEPTH, &p); + fd = io_uring_setup(depth, &p); if (fd < 0) { perror("io_uring_setup"); return 1; } s->ring_fd = fd; + io_uring_probe(fd); + if (fixedbufs) { ret = io_uring_register_buffers(s); if (ret < 0) { - perror("io_uring_register"); + perror("io_uring_register_buffers"); + return 1; + } + } + + if (register_files) { + ret = io_uring_register_files(s); + if (ret < 0) { + perror("io_uring_register_files"); return 1; } } @@ -355,6 +491,7 @@ static int setup_ring(struct submitter *s) sring->tail = ptr + p.sq_off.tail; sring->ring_mask = ptr + p.sq_off.ring_mask; sring->ring_entries = ptr + p.sq_off.ring_entries; + sring->flags = ptr + p.sq_off.flags; sring->array = ptr + p.sq_off.array; sq_ring_mask = *sring->ring_mask; @@ -376,33 +513,117 @@ static int setup_ring(struct submitter *s) return 0; } +static void file_depths(char *buf) +{ + char *p; + int i, j; + + buf[0] = '\0'; + p = buf; + for (j = 0; j < nthreads; j++) { + struct submitter *s = get_submitter(j); + + for (i = 0; i < s->nr_files; i++) { + struct file *f = &s->files[i]; + + if (i + 1 == s->nr_files) + p += sprintf(p, "%d", f->pending_ios); + else + p += sprintf(p, "%d, ", f->pending_ios); + } + } +} + +static void usage(char *argv) +{ + printf("%s [options] -- [filenames]\n" + " -d : IO Depth, default %d\n" + " -s : Batch submit, default %d\n" + " -c : Batch complete, default %d\n" + " -b : Block size, default %d\n" + " -p : Polled IO, default %d\n", + argv, DEPTH, BATCH_SUBMIT, BATCH_COMPLETE, BS, polled); + exit(0); +} + int main(int argc, char *argv[]) { - struct submitter *s = &submitters[0]; - unsigned long done, calls, reap, cache_hit, cache_miss; - int err, i, flags, fd; - struct rlimit rlim; + struct submitter *s; + unsigned long done, calls, reap; + int err, i, j, flags, fd, opt; + char *fdepths; void *ret; - if (argc < 2) { - printf("%s: filename\n", argv[0]); + if (!do_nop && argc < 2) { + printf("%s: filename [options]\n", argv[0]); return 1; } - flags = O_RDONLY; + while ((opt = getopt(argc, argv, "d:s:c:b:p:B:F:n:h?")) != -1) { + switch (opt) { + case 'd': + depth = atoi(optarg); + break; + case 's': + batch_submit = atoi(optarg); + break; + case 'c': + batch_complete = atoi(optarg); + break; + case 'b': + bs = atoi(optarg); + break; + case 'p': + polled = !!atoi(optarg); + break; + case 'B': + fixedbufs = !!atoi(optarg); + break; + case 'F': + register_files = !!atoi(optarg); + break; + case 'n': + nthreads = atoi(optarg); + break; + case 'h': + case '?': + default: + usage(argv[0]); + break; + } + } + + submitter = calloc(nthreads, sizeof(*submitter) + + depth * sizeof(struct iovec)); + for (j = 0; j < nthreads; j++) { + s = get_submitter(j); + s->index = j; + s->done = s->calls = s->reaps = 0; + } + + flags = O_RDONLY | O_NOATIME; if (!buffered) flags |= O_DIRECT; - i = 1; - while (i < argc) { - struct file *f = &s->files[s->nr_files]; + j = 0; + i = optind; + printf("i %d, argc %d\n", i, argc); + while (!do_nop && i < argc) { + struct file *f; + s = get_submitter(j); + if (s->nr_files == MAX_FDS) { + printf("Max number of files (%d) reached\n", MAX_FDS); + break; + } fd = open(argv[i], flags); if (fd < 0) { perror("open"); return 1; } - f->fd = fd; + + f = &s->files[s->nr_files]; + f->real_fd = fd; if (get_file_size(f)) { printf("failed getting size of device/file\n"); return 1; @@ -413,80 +634,92 @@ int main(int argc, char *argv[]) } f->max_blocks--; - printf("Added file %s\n", argv[i]); + printf("Added file %s (submitter %d)\n", argv[i], s->index); s->nr_files++; i++; + if (++j >= nthreads) + j = 0; } - rlim.rlim_cur = RLIM_INFINITY; - rlim.rlim_max = RLIM_INFINITY; - if (setrlimit(RLIMIT_MEMLOCK, &rlim) < 0) { - perror("setrlimit"); - return 1; + if (fixedbufs) { + struct rlimit rlim; + + rlim.rlim_cur = RLIM_INFINITY; + rlim.rlim_max = RLIM_INFINITY; + if (setrlimit(RLIMIT_MEMLOCK, &rlim) < 0) { + perror("setrlimit"); + return 1; + } } arm_sig_int(); - for (i = 0; i < DEPTH; i++) { - void *buf; + for (j = 0; j < nthreads; j++) { + s = get_submitter(j); + for (i = 0; i < depth; i++) { + void *buf; - if (posix_memalign(&buf, BS, BS)) { - printf("failed alloc\n"); - return 1; + if (posix_memalign(&buf, bs, bs)) { + printf("failed alloc\n"); + return 1; + } + s->iovecs[i].iov_base = buf; + s->iovecs[i].iov_len = bs; } - s->iovecs[i].iov_base = buf; - s->iovecs[i].iov_len = BS; } - err = setup_ring(s); - if (err) { - printf("ring setup failed: %s, %d\n", strerror(errno), err); - return 1; + for (j = 0; j < nthreads; j++) { + s = get_submitter(j); + + err = setup_ring(s); + if (err) { + printf("ring setup failed: %s, %d\n", strerror(errno), err); + return 1; + } } - printf("polled=%d, fixedbufs=%d, buffered=%d", polled, fixedbufs, buffered); - printf(" QD=%d, sq_ring=%d, cq_ring=%d\n", DEPTH, *s->sq_ring.ring_entries, *s->cq_ring.ring_entries); + s = get_submitter(0); + printf("polled=%d, fixedbufs=%d, register_files=%d, buffered=%d", polled, fixedbufs, register_files, buffered); + printf(" QD=%d, sq_ring=%d, cq_ring=%d\n", depth, *s->sq_ring.ring_entries, *s->cq_ring.ring_entries); - pthread_create(&s->thread, NULL, submitter_fn, s); + for (j = 0; j < nthreads; j++) { + s = get_submitter(j); + pthread_create(&s->thread, NULL, submitter_fn, s); + } - cache_hit = cache_miss = reap = calls = done = 0; + fdepths = malloc(8 * s->nr_files * nthreads); + reap = calls = done = 0; do { unsigned long this_done = 0; unsigned long this_reap = 0; unsigned long this_call = 0; - unsigned long this_cache_hit = 0; - unsigned long this_cache_miss = 0; unsigned long rpc = 0, ipc = 0; - double hit = 0.0; sleep(1); - this_done += s->done; - this_call += s->calls; - this_reap += s->reaps; - this_cache_hit += s->cachehit; - this_cache_miss += s->cachemiss; - if (this_cache_hit && this_cache_miss) { - unsigned long hits, total; - - hits = this_cache_hit - cache_hit; - total = hits + this_cache_miss - cache_miss; - hit = (double) hits / (double) total; - hit *= 100.0; + for (j = 0; j < nthreads; j++) { + this_done += s->done; + this_call += s->calls; + this_reap += s->reaps; } if (this_call - calls) { rpc = (this_done - done) / (this_call - calls); ipc = (this_reap - reap) / (this_call - calls); - } - printf("IOPS=%lu, IOS/call=%lu/%lu, inflight=%u (head=%u tail=%u), Cachehit=%0.2f%%\n", + } else + rpc = ipc = -1; + file_depths(fdepths); + printf("IOPS=%lu, IOS/call=%ld/%ld, inflight=%u (%s)\n", this_done - done, rpc, ipc, s->inflight, - *s->cq_ring.head, *s->cq_ring.tail, hit); + fdepths); done = this_done; calls = this_call; reap = this_reap; - cache_hit = s->cachehit; - cache_miss = s->cachemiss; } while (!finish); - pthread_join(s->thread, &ret); - close(s->ring_fd); + for (j = 0; j < nthreads; j++) { + s = get_submitter(j); + pthread_join(s->thread, &ret); + close(s->ring_fd); + } + free(fdepths); + free(submitter); return 0; }