X-Git-Url: https://git.kernel.dk/?p=fio.git;a=blobdiff_plain;f=backend.c;h=b92877efd1ca8a167555cd753387df6eaf39fbc5;hp=2ec478c1ac206ebf4163a2e854fb77002dba2366;hb=74454ce40f1a5e1e682da0a8acb824a7f6910270;hpb=d01612f3ae2515eb035d0c4ce954d8cb167a0a61 diff --git a/backend.c b/backend.c index 2ec478c1..b92877ef 100644 --- a/backend.c +++ b/backend.c @@ -52,6 +52,7 @@ #include "server.h" #include "lib/getrusage.h" #include "idletime.h" +#include "err.h" static pthread_t disk_util_thread; static struct fio_mutex *disk_thread_mutex; @@ -345,7 +346,7 @@ static inline int runtime_exceeded(struct thread_data *td, struct timeval *t) return 0; if (!td->o.timeout) return 0; - if (mtime_since(&td->epoch, t) >= td->o.timeout * 1000) + if (mtime_since(&td->epoch, t) >= td->o.timeout ) return 1; return 0; @@ -478,6 +479,12 @@ static void do_verify(struct thread_data *td, uint64_t verify_bytes) break; while ((io_u = get_io_u(td)) != NULL) { + if (IS_ERR(io_u)) { + io_u = NULL; + ret = FIO_Q_BUSY; + goto reap; + } + /* * We are only interested in the places where * we wrote or trimmed IOs. Turn those into @@ -574,6 +581,7 @@ sync_done: * completed io_u's first. Note that we can get BUSY even * without IO queued, if the system is resource starved. */ +reap: full = queue_full(td) || (ret == FIO_Q_BUSY && td->cur_depth); if (full || !td->o.iodepth_batch_complete) { min_events = min(td->o.iodepth_batch_complete, @@ -642,13 +650,25 @@ static uint64_t do_io(struct thread_data *td) uint64_t bytes_done[DDIR_RWDIR_CNT] = { 0, 0, 0 }; unsigned int i; int ret = 0; - uint64_t bytes_issued = 0; + uint64_t total_bytes, bytes_issued = 0; if (in_ramp_time(td)) td_set_runstate(td, TD_RAMP); else td_set_runstate(td, TD_RUNNING); + lat_target_init(td); + + /* + * If verify_backlog is enabled, we'll run the verify in this + * handler as well. For that case, we may need up to twice the + * amount of bytes. + */ + total_bytes = td->o.size; + if (td->o.verify != VERIFY_NONE && + (td_write(td) && td->o.verify_backlog)) + total_bytes += td->o.size; + while ((td->o.read_iolog_file && !flist_empty(&td->io_log_list)) || (!flist_empty(&td->trim_list)) || !io_bytes_exceeded(td) || td->o.time_based) { @@ -676,12 +696,22 @@ static uint64_t do_io(struct thread_data *td) if (flow_threshold_exceeded(td)) continue; - if (bytes_issued >= (uint64_t) td->o.size) + if (bytes_issued >= total_bytes) break; io_u = get_io_u(td); - if (!io_u) + if (IS_ERR_OR_NULL(io_u)) { + int err = PTR_ERR(io_u); + + io_u = NULL; + if (err == -EBUSY) { + ret = FIO_Q_BUSY; + goto reap; + } + if (td->o.latency_target) + goto reap; break; + } ddir = io_u->ddir; @@ -692,6 +722,13 @@ static uint64_t do_io(struct thread_data *td) */ if (td->o.verify != VERIFY_NONE && io_u->ddir == DDIR_READ && ((io_u->flags & IO_U_F_VER_LIST) || !td_rw(td))) { + + if (!td->o.verify_pattern_bytes) { + io_u->rand_seed = __rand(&td->__verify_state); + if (sizeof(int) != sizeof(long *)) + io_u->rand_seed *= __rand(&td->__verify_state); + } + if (td->o.verify_async) io_u->end_io = verify_io_u_async; else @@ -702,6 +739,17 @@ static uint64_t do_io(struct thread_data *td) else td_set_runstate(td, TD_RUNNING); + /* + * Always log IO before it's issued, so we know the specific + * order of it. The logged unit will track when the IO has + * completed. + */ + if (td_write(td) && io_u->ddir == DDIR_WRITE && + td->o.do_verify && + td->o.verify != VERIFY_NONE && + !td->o.experimental_verify) + log_io_piece(td, io_u); + ret = td_io_queue(td, io_u); switch (ret) { case FIO_Q_COMPLETED: @@ -776,6 +824,7 @@ sync_done: * can get BUSY even without IO queued, if the system is * resource starved. */ +reap: full = queue_full(td) || (ret == FIO_Q_BUSY && td->cur_depth); if (full || !td->o.iodepth_batch_complete) { min_evts = min(td->o.iodepth_batch_complete, @@ -812,6 +861,8 @@ sync_done: break; } } + if (!in_ramp_time(td) && td->o.latency_target) + lat_target_check(td); if (td->o.thinktime) { unsigned long long b; @@ -981,7 +1032,7 @@ static int init_io_u(struct thread_data *td) * Fill the buffer with the pattern if we are * going to be doing writes. */ - fill_pattern(td, io_u->buf, max_bs, io_u, 0, 0); + fill_verify_pattern(td, io_u->buf, max_bs, io_u, 0, 0); } } @@ -1088,6 +1139,9 @@ static int keep_running(struct thread_data *td) if (diff < td_max_bs(td)) return 0; + if (fio_files_done(td)) + return 0; + return 1; } @@ -1111,6 +1165,44 @@ static int exec_string(struct thread_options *o, const char *string, const char return ret; } +/* + * Dry run to compute correct state of numberio for verification. + */ +static uint64_t do_dry_run(struct thread_data *td) +{ + uint64_t bytes_done[DDIR_RWDIR_CNT] = { 0, 0, 0 }; + + td_set_runstate(td, TD_RUNNING); + + while ((td->o.read_iolog_file && !flist_empty(&td->io_log_list)) || + (!flist_empty(&td->trim_list)) || !io_bytes_exceeded(td)) { + struct io_u *io_u; + int ret; + + if (td->terminate || td->done) + break; + + io_u = get_io_u(td); + if (!io_u) + break; + + io_u->flags |= IO_U_F_FLIGHT; + io_u->error = 0; + io_u->resid = 0; + if (ddir_rw(acct_ddir(io_u))) + td->io_issues[acct_ddir(io_u)]++; + if (ddir_rw(io_u->ddir)) { + io_u_mark_depth(td, 1); + td->ts.total_io_u[io_u->ddir]++; + } + + ret = io_u_sync_complete(td, io_u, bytes_done); + (void) ret; + } + + return bytes_done[DDIR_WRITE] + bytes_done[DDIR_TRIM]; +} + /* * Entry point for the thread based jobs. The process based jobs end up * here as well, after a little setup. @@ -1161,13 +1253,6 @@ static void *thread_main(void *data) fio_mutex_down(td->mutex); dprint(FD_MUTEX, "done waiting on td->mutex\n"); - /* - * the ->mutex mutex is now no longer used, close it to avoid - * eating a file descriptor - */ - fio_mutex_remove(td->mutex); - td->mutex = NULL; - /* * A new gid requires privilege, so we need to do this before setting * the uid. @@ -1324,7 +1409,10 @@ static void *thread_main(void *data) prune_io_piece_log(td); - verify_bytes = do_io(td); + if (td->o.verify_only && (td_write(td) || td_rw(td))) + verify_bytes = do_dry_run(td); + else + verify_bytes = do_io(td); clear_state = 1; @@ -1444,6 +1532,9 @@ err: fio_mutex_remove(td->rusage_sem); td->rusage_sem = NULL; + fio_mutex_remove(td->mutex); + td->mutex = NULL; + td_set_runstate(td, TD_EXITED); return (void *) (uintptr_t) td->error; } @@ -1692,7 +1783,7 @@ static void run_threads(void) if (td->o.start_delay) { spent = mtime_since_genesis(); - if (td->o.start_delay * 1000 > spent) + if (td->o.start_delay > spent) continue; }