X-Git-Url: https://git.kernel.dk/?a=blobdiff_plain;f=io_u.c;h=9a1cd547d75f720bf321b51e8db818bd2764995b;hb=2459bd33b3dbb7a34f28c612d595311a6bc7593d;hp=155d0a3288acb0703e74c824f34ad9ce515d3900;hpb=db83b0abd16bbd6b8f589a993e6f70d9812be6e3;p=fio.git diff --git a/io_u.c b/io_u.c index 155d0a32..9a1cd547 100644 --- a/io_u.c +++ b/io_u.c @@ -795,7 +795,7 @@ void put_io_u(struct thread_data *td, struct io_u *io_u) { const bool needs_lock = td_async_processing(td); - zbd_put_io_u(io_u); + zbd_put_io_u(td, io_u); if (td->parent) td = td->parent; @@ -1326,8 +1326,10 @@ static struct fio_file *__get_next_file(struct thread_data *td) if (f && fio_file_open(f) && !fio_file_closing(f)) { if (td->o.file_service_type == FIO_FSERVICE_SEQ) goto out; - if (td->file_service_left--) - goto out; + if (td->file_service_left) { + td->file_service_left--; + goto out; + } } if (td->o.file_service_type == FIO_FSERVICE_RR || @@ -1369,7 +1371,7 @@ static long set_io_u_file(struct thread_data *td, struct io_u *io_u) if (!fill_io_u(td, io_u)) break; - zbd_put_io_u(io_u); + zbd_put_io_u(td, io_u); put_file_log(td, f); td_io_close_file(td, f); @@ -1387,11 +1389,16 @@ static long set_io_u_file(struct thread_data *td, struct io_u *io_u) return 0; } -static void lat_fatal(struct thread_data *td, struct io_completion_data *icd, +static void lat_fatal(struct thread_data *td, struct io_u *io_u, struct io_completion_data *icd, unsigned long long tnsec, unsigned long long max_nsec) { - if (!td->error) - log_err("fio: latency of %llu nsec exceeds specified max (%llu nsec)\n", tnsec, max_nsec); + if (!td->error) { + log_err("fio: latency of %llu nsec exceeds specified max (%llu nsec): %s %s %llu %llu\n", + tnsec, max_nsec, + io_u->file->file_name, + io_ddir_name(io_u->ddir), + io_u->offset, io_u->buflen); + } td_verror(td, ETIMEDOUT, "max latency exceeded"); icd->error = ETIMEDOUT; } @@ -1886,11 +1893,13 @@ static void account_io_completion(struct thread_data *td, struct io_u *io_u, icd->error = ops->io_u_lat(td, tnsec); } - if (td->o.max_latency && tnsec > td->o.max_latency) - lat_fatal(td, icd, tnsec, td->o.max_latency); - if (td->o.latency_target && tnsec > td->o.latency_target) { - if (lat_target_failed(td)) - lat_fatal(td, icd, tnsec, td->o.latency_target); + if (ddir_rw(idx)) { + if (td->o.max_latency[idx] && tnsec > td->o.max_latency[idx]) + lat_fatal(td, io_u, icd, tnsec, td->o.max_latency[idx]); + if (td->o.latency_target && tnsec > td->o.latency_target) { + if (lat_target_failed(td)) + lat_fatal(td, io_u, icd, tnsec, td->o.latency_target); + } } } @@ -2163,6 +2172,7 @@ void io_u_queued(struct thread_data *td, struct io_u *io_u) static struct frand_state *get_buf_state(struct thread_data *td) { unsigned int v; + unsigned long long i; if (!td->o.dedupe_percentage) return &td->buf_state; @@ -2174,7 +2184,24 @@ static struct frand_state *get_buf_state(struct thread_data *td) v = rand_between(&td->dedupe_state, 1, 100); if (v <= td->o.dedupe_percentage) - return &td->buf_state_prev; + switch (td->o.dedupe_mode) { + case DEDUPE_MODE_REPEAT: + /* + * The caller advances the returned frand_state. + * A copy of prev should be returned instead since + * a subsequent intention to generate a deduped buffer + * might result in generating a unique one + */ + frand_copy(&td->buf_state_ret, &td->buf_state_prev); + return &td->buf_state_ret; + case DEDUPE_MODE_WORKING_SET: + i = rand_between(&td->dedupe_working_set_index_state, 0, td->num_unique_pages - 1); + frand_copy(&td->buf_state_ret, &td->dedupe_working_set_states[i]); + return &td->buf_state_ret; + default: + log_err("unexpected dedupe mode %u\n", td->o.dedupe_mode); + assert(0); + } return &td->buf_state; }