| 1 | /* |
| 2 | * glusterfs engine |
| 3 | * |
| 4 | * IO engine using Glusterfs's gfapi async interface |
| 5 | * |
| 6 | */ |
| 7 | #include "gfapi.h" |
| 8 | #define NOT_YET 1 |
| 9 | struct fio_gf_iou { |
| 10 | struct io_u *io_u; |
| 11 | int io_complete; |
| 12 | }; |
| 13 | static ulong cb_count = 0, issued = 0; |
| 14 | |
| 15 | static struct io_u *fio_gf_event(struct thread_data *td, int event) |
| 16 | { |
| 17 | struct gf_data *gf_data = td->io_ops->data; |
| 18 | dprint(FD_IO, "%s\n", __FUNCTION__); |
| 19 | return gf_data->aio_events[event]; |
| 20 | } |
| 21 | |
| 22 | static int fio_gf_getevents(struct thread_data *td, unsigned int min, |
| 23 | unsigned int max, const struct timespec *t) |
| 24 | { |
| 25 | struct gf_data *g = td->io_ops->data; |
| 26 | unsigned int events = 0; |
| 27 | struct io_u *io_u; |
| 28 | int i = 0; |
| 29 | struct fio_gf_iou *io = NULL; |
| 30 | |
| 31 | dprint(FD_IO, "%s\n", __FUNCTION__); |
| 32 | do { |
| 33 | io_u_qiter(&td->io_u_all, io_u, i) { |
| 34 | if (!(io_u->flags & IO_U_F_FLIGHT)) |
| 35 | continue; |
| 36 | |
| 37 | io = (struct fio_gf_iou *)io_u->engine_data; |
| 38 | |
| 39 | if (io && io->io_complete) { |
| 40 | io->io_complete = 0; |
| 41 | g->aio_events[events] = io_u; |
| 42 | events++; |
| 43 | |
| 44 | if (events >= max) |
| 45 | break; |
| 46 | } |
| 47 | |
| 48 | } |
| 49 | if (events < min) |
| 50 | usleep(100); |
| 51 | else |
| 52 | break; |
| 53 | |
| 54 | } while (1); |
| 55 | |
| 56 | return events; |
| 57 | } |
| 58 | |
| 59 | static void fio_gf_io_u_free(struct thread_data *td, struct io_u *io_u) |
| 60 | { |
| 61 | struct fio_gf_iou *io = io_u->engine_data; |
| 62 | |
| 63 | if (io) { |
| 64 | if (io->io_complete) { |
| 65 | log_err("incomplete IO found.\n"); |
| 66 | } |
| 67 | io_u->engine_data = NULL; |
| 68 | free(io); |
| 69 | } |
| 70 | fprintf(stderr, "issued %lu finished %lu\n", issued, cb_count); |
| 71 | } |
| 72 | |
| 73 | static int fio_gf_io_u_init(struct thread_data *td, struct io_u *io_u) |
| 74 | { |
| 75 | struct fio_gf_iou *io = NULL; |
| 76 | |
| 77 | dprint(FD_FILE, "%s\n", __FUNCTION__); |
| 78 | |
| 79 | if (!io_u->engine_data) { |
| 80 | io = malloc(sizeof(struct fio_gf_iou)); |
| 81 | if (!io) { |
| 82 | td_verror(td, errno, "malloc"); |
| 83 | return 1; |
| 84 | } |
| 85 | io->io_complete = 0; |
| 86 | io->io_u = io_u; |
| 87 | io_u->engine_data = io; |
| 88 | } |
| 89 | return 0; |
| 90 | } |
| 91 | |
| 92 | static void gf_async_cb(glfs_fd_t * fd, ssize_t ret, void *data) |
| 93 | { |
| 94 | struct io_u *io_u = (struct io_u *)data; |
| 95 | struct fio_gf_iou *iou = (struct fio_gf_iou *)io_u->engine_data; |
| 96 | |
| 97 | dprint(FD_IO, "%s ret %lu\n", __FUNCTION__, ret); |
| 98 | iou->io_complete = 1; |
| 99 | cb_count++; |
| 100 | } |
| 101 | |
| 102 | static int fio_gf_async_queue(struct thread_data fio_unused * td, |
| 103 | struct io_u *io_u) |
| 104 | { |
| 105 | struct gf_data *g = td->io_ops->data; |
| 106 | int r = 0; |
| 107 | |
| 108 | dprint(FD_IO, "%s op %s\n", __FUNCTION__, |
| 109 | io_u->ddir == DDIR_READ ? "read" : io_u->ddir == |
| 110 | DDIR_WRITE ? "write" : io_u->ddir == |
| 111 | DDIR_SYNC ? "sync" : "unknown"); |
| 112 | |
| 113 | fio_ro_check(td, io_u); |
| 114 | |
| 115 | if (io_u->ddir == DDIR_READ) |
| 116 | r = glfs_pread_async(g->fd, io_u->xfer_buf, io_u->xfer_buflen, |
| 117 | io_u->offset, 0, gf_async_cb, |
| 118 | (void *)io_u); |
| 119 | else if (io_u->ddir == DDIR_WRITE) |
| 120 | r = glfs_pwrite_async(g->fd, io_u->xfer_buf, io_u->xfer_buflen, |
| 121 | io_u->offset, 0, gf_async_cb, |
| 122 | (void *)io_u); |
| 123 | else if (io_u->ddir == DDIR_SYNC) { |
| 124 | r = glfs_fsync_async(g->fd, gf_async_cb, (void *)io_u); |
| 125 | } else { |
| 126 | log_err("unsupported operation.\n"); |
| 127 | io_u->error = -EINVAL; |
| 128 | goto failed; |
| 129 | } |
| 130 | if (r) { |
| 131 | log_err("glfs failed.\n"); |
| 132 | io_u->error = r; |
| 133 | goto failed; |
| 134 | } |
| 135 | issued++; |
| 136 | return FIO_Q_QUEUED; |
| 137 | |
| 138 | failed: |
| 139 | io_u->error = r; |
| 140 | td_verror(td, io_u->error, "xfer"); |
| 141 | return FIO_Q_COMPLETED; |
| 142 | } |
| 143 | |
| 144 | int fio_gf_async_setup(struct thread_data *td) |
| 145 | { |
| 146 | int r = 0; |
| 147 | struct gf_data *g = NULL; |
| 148 | #if defined(NOT_YET) |
| 149 | fprintf(stderr, "the async interface is still very experimental...\n"); |
| 150 | #endif |
| 151 | r = fio_gf_setup(td); |
| 152 | if (r) { |
| 153 | return r; |
| 154 | } |
| 155 | td->o.use_thread = 1; |
| 156 | g = td->io_ops->data; |
| 157 | g->aio_events = malloc(td->o.iodepth * sizeof(struct io_u *)); |
| 158 | if (!g->aio_events) { |
| 159 | r = -ENOMEM; |
| 160 | fio_gf_cleanup(td); |
| 161 | return r; |
| 162 | } |
| 163 | |
| 164 | memset(g->aio_events, 0, td->o.iodepth * sizeof(struct io_u *)); |
| 165 | |
| 166 | return r; |
| 167 | |
| 168 | } |
| 169 | |
| 170 | static int fio_gf_async_prep(struct thread_data *td, struct io_u *io_u) |
| 171 | { |
| 172 | dprint(FD_FILE, "%s\n", __FUNCTION__); |
| 173 | |
| 174 | if (!ddir_rw(io_u->ddir)) |
| 175 | return 0; |
| 176 | |
| 177 | return 0; |
| 178 | } |
| 179 | |
| 180 | static struct ioengine_ops ioengine = { |
| 181 | .name = "gfapi_async", |
| 182 | .version = FIO_IOOPS_VERSION, |
| 183 | .init = fio_gf_async_setup, |
| 184 | .cleanup = fio_gf_cleanup, |
| 185 | .prep = fio_gf_async_prep, |
| 186 | .queue = fio_gf_async_queue, |
| 187 | .open_file = fio_gf_open_file, |
| 188 | .close_file = fio_gf_close_file, |
| 189 | .unlink_file = fio_gf_unlink_file, |
| 190 | .get_file_size = fio_gf_get_file_size, |
| 191 | .getevents = fio_gf_getevents, |
| 192 | .event = fio_gf_event, |
| 193 | .io_u_init = fio_gf_io_u_init, |
| 194 | .io_u_free = fio_gf_io_u_free, |
| 195 | .options = gfapi_options, |
| 196 | .option_struct_size = sizeof(struct gf_options), |
| 197 | .flags = FIO_DISKLESSIO, |
| 198 | }; |
| 199 | |
| 200 | static void fio_init fio_gf_register(void) |
| 201 | { |
| 202 | register_ioengine(&ioengine); |
| 203 | } |
| 204 | |
| 205 | static void fio_exit fio_gf_unregister(void) |
| 206 | { |
| 207 | unregister_ioengine(&ioengine); |
| 208 | } |