X-Git-Url: https://git.kernel.dk/?p=fio.git;a=blobdiff_plain;f=eta.c;h=647a1bdd8eed7433c2413fde356844720b45e722;hp=970a67dfd0ac8d6672758b99d528f32526f26e4e;hb=cf408d71681396f9b2ac5b277e3849176378c78e;hpb=6fa22eb8d7aec95851b37b64a2c38a17b1da48ee diff --git a/eta.c b/eta.c index 970a67df..647a1bdd 100644 --- a/eta.c +++ b/eta.c @@ -177,12 +177,27 @@ static unsigned long thread_eta(struct thread_data *td) bytes_total = td->fill_device_size; } - if (td->o.zone_size && td->o.zone_skip && bytes_total) { + /* + * If io_size is set, bytes_total is an exact value that does not need + * adjustment. + */ + if (td->o.zone_size && td->o.zone_skip && bytes_total && + !fio_option_is_set(&td->o, io_size)) { unsigned int nr_zones; uint64_t zone_bytes; - zone_bytes = bytes_total + td->o.zone_size + td->o.zone_skip; - nr_zones = (zone_bytes - 1) / (td->o.zone_size + td->o.zone_skip); + /* + * Calculate the upper bound of the number of zones that will + * be processed, including skipped bytes between zones. If this + * is larger than total_io_size (e.g. when --io_size or --size + * specify a small value), use the lower bound to avoid + * adjustments to a negative value that would result in a very + * large bytes_total and an incorrect eta. + */ + zone_bytes = td->o.zone_size + td->o.zone_skip; + nr_zones = (bytes_total + zone_bytes - 1) / zone_bytes; + if (bytes_total < nr_zones * td->o.zone_skip) + nr_zones = bytes_total / zone_bytes; bytes_total -= nr_zones * td->o.zone_skip; } @@ -377,6 +392,9 @@ bool calc_thread_status(struct jobs_eta *je, int force) static unsigned long long disp_io_iops[DDIR_RWDIR_CNT]; static struct timespec rate_prev_time, disp_prev_time; + void *je_rate = (void *) je->rate; + void *je_iops = (void *) je->iops; + if (!force) { if (!(output_format & FIO_OUTPUT_NORMAL) && f_out == stdout) @@ -492,7 +510,7 @@ bool calc_thread_status(struct jobs_eta *je, int force) if (write_bw_log && rate_time > bw_avg_time && !in_ramp_time(td)) { calc_rate(unified_rw_rep, rate_time, io_bytes, rate_io_bytes, - je->rate); + je_rate); memcpy(&rate_prev_time, &now, sizeof(now)); add_agg_sample(sample_val(je->rate[DDIR_READ]), DDIR_READ, 0); add_agg_sample(sample_val(je->rate[DDIR_WRITE]), DDIR_WRITE, 0); @@ -504,8 +522,8 @@ bool calc_thread_status(struct jobs_eta *je, int force) if (!force && !eta_time_within_slack(disp_time)) return false; - calc_rate(unified_rw_rep, disp_time, io_bytes, disp_io_bytes, je->rate); - calc_iops(unified_rw_rep, disp_time, io_iops, disp_io_iops, je->iops); + calc_rate(unified_rw_rep, disp_time, io_bytes, disp_io_bytes, je_rate); + calc_iops(unified_rw_rep, disp_time, io_iops, disp_io_iops, je_iops); memcpy(&disp_prev_time, &now, sizeof(now));