Merge branch 'master' of https://github.com/jan--f/fio
[fio.git] / rate-submit.c
1 /*
2  * Rated submission helpers
3  *
4  * Copyright (C) 2015 Jens Axboe <axboe@kernel.dk>
5  *
6  */
7 #include "fio.h"
8 #include "ioengine.h"
9 #include "lib/getrusage.h"
10 #include "rate-submit.h"
11
12 static int io_workqueue_fn(struct submit_worker *sw,
13                            struct workqueue_work *work)
14 {
15         struct io_u *io_u = container_of(work, struct io_u, work);
16         const enum fio_ddir ddir = io_u->ddir;
17         struct thread_data *td = sw->priv;
18         int ret;
19
20         dprint(FD_RATE, "io_u %p queued by %u\n", io_u, gettid());
21
22         io_u_set(td, io_u, IO_U_F_NO_FILE_PUT);
23
24         td->cur_depth++;
25
26         do {
27                 ret = td_io_queue(td, io_u);
28                 if (ret != FIO_Q_BUSY)
29                         break;
30                 ret = io_u_queued_complete(td, 1);
31                 if (ret > 0)
32                         td->cur_depth -= ret;
33                 io_u_clear(td, io_u, IO_U_F_FLIGHT);
34         } while (1);
35
36         dprint(FD_RATE, "io_u %p ret %d by %u\n", io_u, ret, gettid());
37
38         io_queue_event(td, io_u, &ret, ddir, NULL, 0, NULL);
39
40         if (ret == FIO_Q_COMPLETED)
41                 td->cur_depth--;
42         else if (ret == FIO_Q_QUEUED) {
43                 unsigned int min_evts;
44
45                 if (td->o.iodepth == 1)
46                         min_evts = 1;
47                 else
48                         min_evts = 0;
49
50                 ret = io_u_queued_complete(td, min_evts);
51                 if (ret > 0)
52                         td->cur_depth -= ret;
53         } else if (ret == FIO_Q_BUSY) {
54                 ret = io_u_queued_complete(td, td->cur_depth);
55                 if (ret > 0)
56                         td->cur_depth -= ret;
57         }
58
59         return 0;
60 }
61
62 static bool io_workqueue_pre_sleep_flush_fn(struct submit_worker *sw)
63 {
64         struct thread_data *td = sw->priv;
65
66         if (td->io_u_queued || td->cur_depth || td->io_u_in_flight)
67                 return true;
68
69         return false;
70 }
71
72 static void io_workqueue_pre_sleep_fn(struct submit_worker *sw)
73 {
74         struct thread_data *td = sw->priv;
75         int ret;
76
77         ret = io_u_quiesce(td);
78         if (ret > 0)
79                 td->cur_depth -= ret;
80 }
81
82 static int io_workqueue_alloc_fn(struct submit_worker *sw)
83 {
84         struct thread_data *td;
85
86         td = calloc(1, sizeof(*td));
87         sw->priv = td;
88         return 0;
89 }
90
91 static void io_workqueue_free_fn(struct submit_worker *sw)
92 {
93         free(sw->priv);
94         sw->priv = NULL;
95 }
96
97 static int io_workqueue_init_worker_fn(struct submit_worker *sw)
98 {
99         struct thread_data *parent = sw->wq->td;
100         struct thread_data *td = sw->priv;
101         int fio_unused ret;
102
103         memcpy(&td->o, &parent->o, sizeof(td->o));
104         memcpy(&td->ts, &parent->ts, sizeof(td->ts));
105         td->o.uid = td->o.gid = -1U;
106         dup_files(td, parent);
107         td->eo = parent->eo;
108         fio_options_mem_dupe(td);
109
110         if (ioengine_load(td))
111                 goto err;
112
113         td->pid = gettid();
114
115         INIT_FLIST_HEAD(&td->io_log_list);
116         INIT_FLIST_HEAD(&td->io_hist_list);
117         INIT_FLIST_HEAD(&td->verify_list);
118         INIT_FLIST_HEAD(&td->trim_list);
119         INIT_FLIST_HEAD(&td->next_rand_list);
120         td->io_hist_tree = RB_ROOT;
121
122         td->o.iodepth = 1;
123         if (td_io_init(td))
124                 goto err_io_init;
125
126         set_epoch_time(td, td->o.log_unix_epoch);
127         fio_getrusage(&td->ru_start);
128         clear_io_state(td, 1);
129
130         td_set_runstate(td, TD_RUNNING);
131         td->flags |= TD_F_CHILD;
132         td->parent = parent;
133         return 0;
134
135 err_io_init:
136         close_ioengine(td);
137 err:
138         return 1;
139
140 }
141
142 static void io_workqueue_exit_worker_fn(struct submit_worker *sw,
143                                         unsigned int *sum_cnt)
144 {
145         struct thread_data *td = sw->priv;
146
147         (*sum_cnt)++;
148         sum_thread_stats(&sw->wq->td->ts, &td->ts, *sum_cnt == 1);
149
150         fio_options_free(td);
151         close_and_free_files(td);
152         if (td->io_ops)
153                 close_ioengine(td);
154         td_set_runstate(td, TD_EXITED);
155 }
156
157 #ifdef CONFIG_SFAA
158 static void sum_val(uint64_t *dst, uint64_t *src)
159 {
160         if (*src) {
161                 __sync_fetch_and_add(dst, *src);
162                 *src = 0;
163         }
164 }
165 #else
166 static void sum_val(uint64_t *dst, uint64_t *src)
167 {
168         if (*src) {
169                 *dst += *src;
170                 *src = 0;
171         }
172 }
173 #endif
174
175 static void pthread_double_unlock(pthread_mutex_t *lock1,
176                                   pthread_mutex_t *lock2)
177 {
178 #ifndef CONFIG_SFAA
179         pthread_mutex_unlock(lock1);
180         pthread_mutex_unlock(lock2);
181 #endif
182 }
183
184 static void pthread_double_lock(pthread_mutex_t *lock1, pthread_mutex_t *lock2)
185 {
186 #ifndef CONFIG_SFAA
187         if (lock1 < lock2) {
188                 pthread_mutex_lock(lock1);
189                 pthread_mutex_lock(lock2);
190         } else {
191                 pthread_mutex_lock(lock2);
192                 pthread_mutex_lock(lock1);
193         }
194 #endif
195 }
196
197 static void sum_ddir(struct thread_data *dst, struct thread_data *src,
198                      enum fio_ddir ddir)
199 {
200         pthread_double_lock(&dst->io_wq.stat_lock, &src->io_wq.stat_lock);
201
202         sum_val(&dst->io_bytes[ddir], &src->io_bytes[ddir]);
203         sum_val(&dst->io_blocks[ddir], &src->io_blocks[ddir]);
204         sum_val(&dst->this_io_blocks[ddir], &src->this_io_blocks[ddir]);
205         sum_val(&dst->this_io_bytes[ddir], &src->this_io_bytes[ddir]);
206         sum_val(&dst->bytes_done[ddir], &src->bytes_done[ddir]);
207
208         pthread_double_unlock(&dst->io_wq.stat_lock, &src->io_wq.stat_lock);
209 }
210
211 static void io_workqueue_update_acct_fn(struct submit_worker *sw)
212 {
213         struct thread_data *src = sw->priv;
214         struct thread_data *dst = sw->wq->td;
215
216         if (td_read(src))
217                 sum_ddir(dst, src, DDIR_READ);
218         if (td_write(src))
219                 sum_ddir(dst, src, DDIR_WRITE);
220         if (td_trim(src))
221                 sum_ddir(dst, src, DDIR_TRIM);
222
223 }
224
225 static struct workqueue_ops rated_wq_ops = {
226         .fn                     = io_workqueue_fn,
227         .pre_sleep_flush_fn     = io_workqueue_pre_sleep_flush_fn,
228         .pre_sleep_fn           = io_workqueue_pre_sleep_fn,
229         .update_acct_fn         = io_workqueue_update_acct_fn,
230         .alloc_worker_fn        = io_workqueue_alloc_fn,
231         .free_worker_fn         = io_workqueue_free_fn,
232         .init_worker_fn         = io_workqueue_init_worker_fn,
233         .exit_worker_fn         = io_workqueue_exit_worker_fn,
234 };
235
236 int rate_submit_init(struct thread_data *td, struct sk_out *sk_out)
237 {
238         if (td->o.io_submit_mode != IO_MODE_OFFLOAD)
239                 return 0;
240
241         return workqueue_init(td, &td->io_wq, &rated_wq_ops, td->o.iodepth, sk_out);
242 }
243
244 void rate_submit_exit(struct thread_data *td)
245 {
246         if (td->o.io_submit_mode != IO_MODE_OFFLOAD)
247                 return;
248
249         workqueue_exit(&td->io_wq);
250 }