Fix bandwidth logging for mixed read/write workloads.
[fio.git] / stat.c
diff --git a/stat.c b/stat.c
index 2400cd70f04c49a5d2ba74e3d6fa590c5ae2e65c..68744281a92de519cbb1f350fb459dfcca517cfd 100644 (file)
--- a/stat.c
+++ b/stat.c
@@ -440,12 +440,14 @@ static void show_ddir_status(struct group_run_stats *rs, struct thread_stat *ts,
                                        ts->percentile_list);
        }
        if (calc_lat(&ts->bw_stat[ddir], &min, &max, &mean, &dev)) {
-               double p_of_agg;
+               double p_of_agg = 100.0;
                const char *bw_str = "KB";
 
-               p_of_agg = mean * 100 / (double) rs->agg[ddir];
-               if (p_of_agg > 100.0)
-                       p_of_agg = 100.0;
+               if (rs->agg[ddir]) {
+                       p_of_agg = mean * 100 / (double) rs->agg[ddir];
+                       if (p_of_agg > 100.0)
+                               p_of_agg = 100.0;
+               }
 
                if (mean > 999999.9) {
                        min /= 1000.0;
@@ -653,9 +655,14 @@ static void show_ddir_status_terse(struct thread_stat *ts,
                free(ovals);
 
        if (calc_lat(&ts->bw_stat[ddir], &min, &max, &mean, &dev)) {
-               double p_of_agg;
+               double p_of_agg = 100.0;
+
+               if (rs->agg[ddir]) {
+                       p_of_agg = mean * 100 / (double) rs->agg[ddir];
+                       if (p_of_agg > 100.0)
+                               p_of_agg = 100.0;
+               }
 
-               p_of_agg = mean * 100 / (double) rs->agg[ddir];
                log_info(";%lu;%lu;%f%%;%f;%f", min, max, p_of_agg, mean, dev);
        } else
                log_info(";%lu;%lu;%f%%;%f;%f", 0UL, 0UL, 0.0, 0.0, 0.0);
@@ -1128,6 +1135,9 @@ static void __add_log_sample(struct io_log *iolog, unsigned long val,
 {
        const int nr_samples = iolog->nr_samples;
 
+       if (!iolog->nr_samples)
+               iolog->avg_last = t;
+
        if (iolog->nr_samples == iolog->max_samples) {
                int new_size = sizeof(struct io_sample) * iolog->max_samples*2;
 
@@ -1142,14 +1152,66 @@ static void __add_log_sample(struct io_log *iolog, unsigned long val,
        iolog->nr_samples++;
 }
 
+static inline void reset_io_stat(struct io_stat *ios)
+{
+       ios->max_val = ios->min_val = ios->samples = 0;
+       ios->mean.u.f = ios->S.u.f = 0;
+}
+
 static void add_log_sample(struct thread_data *td, struct io_log *iolog,
                           unsigned long val, enum fio_ddir ddir,
                           unsigned int bs)
 {
+       unsigned long elapsed, this_window;
+
        if (!ddir_rw(ddir))
                return;
 
-       __add_log_sample(iolog, val, ddir, bs, mtime_since_now(&td->epoch));
+       elapsed = mtime_since_now(&td->epoch);
+
+       /*
+        * If no time averaging, just add the log sample.
+        */
+       if (!iolog->avg_msec) {
+               __add_log_sample(iolog, val, ddir, bs, elapsed);
+               return;
+       }
+
+       /*
+        * Add the sample. If the time period has passed, then
+        * add that entry to the log and clear.
+        */
+       add_stat_sample(&iolog->avg_window[ddir], val);
+
+       /*
+        * If period hasn't passed, adding the above sample is all we
+        * need to do.
+        */
+       this_window = elapsed - iolog->avg_last;
+       if (this_window < iolog->avg_msec)
+               return;
+
+       /*
+        * Note an entry in the log. Use the mean from the logged samples,
+        * making sure to properly round up. Only write a log entry if we
+        * had actual samples done.
+        */
+       if (iolog->avg_window[DDIR_READ].samples) {
+               unsigned long mr;
+
+               mr = iolog->avg_window[DDIR_READ].mean.u.f + 0.50;
+               __add_log_sample(iolog, mr, DDIR_READ, 0, elapsed);
+       }
+       if (iolog->avg_window[DDIR_WRITE].samples) {
+               unsigned long mw;
+
+               mw = iolog->avg_window[DDIR_WRITE].mean.u.f + 0.50;
+               __add_log_sample(iolog, mw, DDIR_WRITE, 0, elapsed);
+       }
+
+       reset_io_stat(&iolog->avg_window[DDIR_READ]);
+       reset_io_stat(&iolog->avg_window[DDIR_WRITE]);
+       iolog->avg_last = elapsed;
 }
 
 void add_agg_sample(unsigned long val, enum fio_ddir ddir, unsigned int bs)
@@ -1229,16 +1291,27 @@ void add_bw_sample(struct thread_data *td, enum fio_ddir ddir, unsigned int bs,
        spent = mtime_since(&td->bw_sample_time, t);
        if (spent < td->o.bw_avg_time)
                return;
+       
+       /* 
+        * Compute both read and write rates for the interval.
+        */
+       for (ddir = DDIR_READ; ddir <= DDIR_WRITE; ddir++) {
+               uint64_t delta;
 
-       rate = (td->this_io_bytes[ddir] - td->stat_io_bytes[ddir]) *
-                       1000 / spent / 1024;
-       add_stat_sample(&ts->bw_stat[ddir], rate);
+               delta = td->this_io_bytes[ddir] - td->stat_io_bytes[ddir];
+               if (!delta)
+                       continue; /* No entries for interval */
 
-       if (td->bw_log)
-               add_log_sample(td, td->bw_log, rate, ddir, bs);
+               rate = delta * 1000 / spent / 1024;
+               add_stat_sample(&ts->bw_stat[ddir], rate);
+
+               if (td->bw_log)
+                       add_log_sample(td, td->bw_log, rate, ddir, bs);
+
+               td->stat_io_bytes[ddir] = td->this_io_bytes[ddir];
+       }
 
        fio_gettime(&td->bw_sample_time, NULL);
-       td->stat_io_bytes[ddir] = td->this_io_bytes[ddir];
 }
 
 void add_iops_sample(struct thread_data *td, enum fio_ddir ddir,
@@ -1258,10 +1331,8 @@ void add_iops_sample(struct thread_data *td, enum fio_ddir ddir,
 
        add_stat_sample(&ts->iops_stat[ddir], iops);
 
-       if (td->iops_log) {
-               assert(iops);
+       if (td->iops_log)
                add_log_sample(td, td->iops_log, iops, ddir, 0);
-       }
 
        fio_gettime(&td->iops_sample_time, NULL);
        td->stat_io_blocks[ddir] = td->this_io_blocks[ddir];