#include <unistd.h>
#include "fio.h"
-#include "ioengine.h"
#include "flist.h"
#include "workqueue.h"
#include "lib/getrusage.h"
-struct submit_worker {
- pthread_t thread;
- pthread_mutex_t lock;
- pthread_cond_t cond;
- struct flist_head work_list;
- unsigned int flags;
- unsigned int index;
- uint64_t seq;
- struct workqueue *wq;
- struct thread_data td;
-};
-
enum {
SW_F_IDLE = 1 << 0,
SW_F_RUNNING = 1 << 1,
return sw;
}
-static int all_sw_idle(struct workqueue *wq)
+static bool all_sw_idle(struct workqueue *wq)
{
int i;
struct submit_worker *sw = &wq->workers[i];
if (!(sw->flags & SW_F_IDLE))
- return 0;
+ return false;
}
- return 1;
+ return true;
}
/*
}
/*
- * Must be serialized by caller.
+ * Must be serialized by caller. Returns true for queued, false for busy.
*/
-int workqueue_enqueue(struct workqueue *wq, struct io_u *io_u)
+bool workqueue_enqueue(struct workqueue *wq, struct workqueue_work *work)
{
struct submit_worker *sw;
sw = get_submit_worker(wq);
if (sw) {
- const enum fio_ddir ddir = acct_ddir(io_u);
- struct thread_data *parent = wq->td;
-
- if (ddir_rw(ddir)) {
- parent->io_issues[ddir]++;
- parent->io_issue_bytes[ddir] += io_u->xfer_buflen;
- parent->rate_io_issue_bytes[ddir] += io_u->xfer_buflen;
- }
-
pthread_mutex_lock(&sw->lock);
- flist_add_tail(&io_u->verify_list, &sw->work_list);
+ flist_add_tail(&work->list, &sw->work_list);
sw->seq = ++wq->work_seq;
sw->flags &= ~SW_F_IDLE;
pthread_mutex_unlock(&sw->lock);
pthread_cond_signal(&sw->cond);
- return FIO_Q_QUEUED;
+ return true;
}
- return FIO_Q_BUSY;
+ return false;
}
static void handle_list(struct submit_worker *sw, struct flist_head *list)
{
struct workqueue *wq = sw->wq;
- struct io_u *io_u;
+ struct workqueue_work *work;
while (!flist_empty(list)) {
- io_u = flist_first_entry(list, struct io_u, verify_list);
- flist_del_init(&io_u->verify_list);
- wq->fn(&sw->td, io_u);
+ work = flist_first_entry(list, struct workqueue_work, list);
+ flist_del_init(&work->list);
+ wq->ops.fn(sw, work);
}
}
static int init_submit_worker(struct submit_worker *sw)
{
struct thread_data *parent = sw->wq->td;
- struct thread_data *td = &sw->td;
+ struct thread_data *td = sw->private;
int fio_unused ret;
memcpy(&td->o, &parent->o, sizeof(td->o));
fio_gettime(&td->epoch, NULL);
fio_getrusage(&td->ru_start);
- clear_io_state(td);
+ clear_io_state(td, 1);
td_set_runstate(td, TD_RUNNING);
td->flags |= TD_F_CHILD;
}
#endif
-static void sum_ddir(struct thread_data *dst, struct thread_data *src,
- enum fio_ddir ddir)
+static void pthread_double_unlock(pthread_mutex_t *lock1,
+ pthread_mutex_t *lock2)
{
#ifndef CONFIG_SFAA
- pthread_mutex_lock(&dst->io_wq.stat_lock);
- pthread_mutex_lock(&src->io_wq.stat_lock);
+ pthread_mutex_unlock(lock1);
+ pthread_mutex_unlock(lock2);
#endif
+}
+
+static void pthread_double_lock(pthread_mutex_t *lock1, pthread_mutex_t *lock2)
+{
+#ifndef CONFIG_SFAA
+ if (lock1 < lock2) {
+ pthread_mutex_lock(lock1);
+ pthread_mutex_lock(lock2);
+ } else {
+ pthread_mutex_lock(lock2);
+ pthread_mutex_lock(lock1);
+ }
+#endif
+}
+
+static void sum_ddir(struct thread_data *dst, struct thread_data *src,
+ enum fio_ddir ddir)
+{
+ pthread_double_lock(&dst->io_wq.stat_lock, &src->io_wq.stat_lock);
sum_val(&dst->io_bytes[ddir], &src->io_bytes[ddir]);
sum_val(&dst->io_blocks[ddir], &src->io_blocks[ddir]);
sum_val(&dst->this_io_bytes[ddir], &src->this_io_bytes[ddir]);
sum_val(&dst->bytes_done[ddir], &src->bytes_done[ddir]);
-#ifndef CONFIG_SFAA
- pthread_mutex_unlock(&src->io_wq.stat_lock);
- pthread_mutex_unlock(&dst->io_wq.stat_lock);
-#endif
+ pthread_double_unlock(&dst->io_wq.stat_lock, &src->io_wq.stat_lock);
}
static void update_accounting(struct submit_worker *sw)
{
- struct thread_data *src = &sw->td;
+ struct thread_data *src = sw->private;
struct thread_data *dst = sw->wq->td;
if (td_read(src))
{
struct submit_worker *sw = data;
struct workqueue *wq = sw->wq;
- struct thread_data *td = &sw->td;
unsigned int eflags = 0, ret;
FLIST_HEAD(local_list);
break;
}
- if (td->io_u_queued || td->cur_depth ||
- td->io_u_in_flight) {
+ if (workqueue_pre_sleep_check(sw)) {
pthread_mutex_unlock(&sw->lock);
- io_u_quiesce(td);
+ workqueue_pre_sleep(sw);
pthread_mutex_lock(&sw->lock);
}
static void free_worker(struct submit_worker *sw)
{
- struct thread_data *td = &sw->td;
+ struct thread_data *td = sw->private;
+ struct workqueue *wq = sw->wq;
fio_options_free(td);
close_and_free_files(td);
pthread_cond_destroy(&sw->cond);
pthread_mutex_destroy(&sw->lock);
+
+ if (wq->ops.free_worker_fn)
+ wq->ops.free_worker_fn(sw);
}
static void shutdown_worker(struct submit_worker *sw, unsigned int *sum_cnt)
{
struct thread_data *parent = sw->wq->td;
+ struct thread_data *td = sw->private;
pthread_join(sw->thread, NULL);
(*sum_cnt)++;
- sum_thread_stats(&parent->ts, &sw->td.ts, *sum_cnt);
+ sum_thread_stats(&parent->ts, &td->ts, *sum_cnt == 1);
free_worker(sw);
}
sw = &wq->workers[i];
if (sw->flags & SW_F_ACCOUNTED)
continue;
+ pthread_mutex_lock(&sw->lock);
sw->flags |= SW_F_ACCOUNTED;
+ pthread_mutex_unlock(&sw->lock);
shutdown_worker(sw, &sum_cnt);
shutdown++;
}
sw->wq = wq;
sw->index = index;
+ if (wq->ops.alloc_worker_fn) {
+ ret = wq->ops.alloc_worker_fn(sw);
+ if (ret)
+ return ret;
+ }
+
ret = pthread_create(&sw->thread, NULL, worker_thread, sw);
if (!ret) {
pthread_mutex_lock(&sw->lock);
}
int workqueue_init(struct thread_data *td, struct workqueue *wq,
- workqueue_fn *fn, unsigned max_pending)
+ struct workqueue_ops *ops, unsigned max_pending)
{
unsigned int running;
int i, error;
wq->max_workers = max_pending;
wq->td = td;
- wq->fn = fn;
+ wq->ops = *ops;
wq->work_seq = 0;
wq->next_free_worker = 0;
pthread_cond_init(&wq->flush_cond, NULL);