X-Git-Url: https://git.kernel.dk/?p=fio.git;a=blobdiff_plain;f=stat.c;h=83ffd0690afca59cc6d1ff266bd717ff70d9395e;hp=7bcae310f8d41bb388abc1af6b773b6125ea0492;hb=07511a632ef219eb578d8c7f12825a275c640bcf;hpb=802ad4a83e92a30b5fdccf117d59fbb69068c054 diff --git a/stat.c b/stat.c index 7bcae310..83ffd069 100644 --- a/stat.c +++ b/stat.c @@ -121,13 +121,17 @@ static void show_clat_percentiles(unsigned int *io_u_plat, unsigned long nr, fio_fp64_t *plist) { unsigned long sum = 0; - unsigned int len, i, j = 0; - int is_last = 0; + unsigned int len, i, j = 0, minv = -1U, maxv = 0; + unsigned int *ovals = NULL, oval_len = 0; + int is_last, scale_down; len = 0; while (len < FIO_IO_U_LIST_MAX_LEN && plist[len].u.f != 0.0) len++; + if (!len) + return; + /* * Sort the percentile list. Note that it may already be sorted if * we are using the default values, but since it's a short list this @@ -136,31 +140,75 @@ static void show_clat_percentiles(unsigned int *io_u_plat, unsigned long nr, if (len > 1) qsort((void*)plist, len, sizeof(plist[0]), double_cmp); - log_info(" clat percentiles (usec) :"); - + /* + * Calculate bucket values, note down max and min values + */ + is_last = 0; for (i = 0; i < FIO_IO_U_PLAT_NR && !is_last; i++) { sum += io_u_plat[i]; while (sum >= (plist[j].u.f / 100.0 * nr)) { assert(plist[j].u.f <= 100.0); - /* for formatting */ - if (j != 0 && (j % 4) == 0) - log_info(" "); - - /* end of the list */ - is_last = (j == len - 1); + if (j == oval_len) { + oval_len += 100; + ovals = realloc(ovals, oval_len * sizeof(unsigned int)); + } - log_info(" %2.2fth=%u%c", plist[j], plat_idx_to_val(i), - (is_last? '\n' : ',')); + ovals[j] = plat_idx_to_val(i); + if (ovals[j] < minv) + minv = ovals[j]; + if (ovals[j] > maxv) + maxv = ovals[j]; + is_last = (j == len - 1); if (is_last) break; - if (j % 4 == 3) /* for formatting */ - log_info("\n"); j++; } } + + /* + * We default to usecs, but if the value range is such that we + * should scale down to msecs, do that. + */ + if (minv > 2000 && maxv > 99999) { + scale_down = 1; + log_info(" clat percentiles (msec):\n |"); + } else { + scale_down = 0; + log_info(" clat percentiles (usec):\n |"); + } + + for (j = 0; j < len; j++) { + char fbuf[8]; + + /* for formatting */ + if (j != 0 && (j % 4) == 0) + log_info(" |"); + + /* end of the list */ + is_last = (j == len - 1); + + if (plist[j].u.f < 10.0) + sprintf(fbuf, " %2.2f", plist[j].u.f); + else + sprintf(fbuf, "%2.2f", plist[j].u.f); + + if (scale_down) + ovals[j] = (ovals[j] + 999) / 1000; + + log_info(" %sth=[%5u]%c", fbuf, ovals[j], is_last ? '\n' : ','); + + if (is_last) + break; + + if (j % 4 == 3) /* for formatting */ + log_info("\n"); + } + + if (ovals) + free(ovals); } static int calc_lat(struct io_stat *is, unsigned long *min, unsigned long *max, @@ -191,7 +239,7 @@ void show_group_stats(struct group_run_stats *rs) const char *ddir_str[] = { " READ", " WRITE" }; int i; - log_info("\nRun status group %d (all jobs):\n", rs->groupid); + log_info("Run status group %d (all jobs):\n", rs->groupid); for (i = 0; i <= DDIR_WRITE; i++) { const int i2p = is_power_of_2(rs->kb_base); @@ -372,7 +420,7 @@ static void show_ddir_status(struct group_run_stats *rs, struct thread_stat *ts, double p_of_agg; p_of_agg = mean * 100 / (double) rs->agg[ddir]; - log_info(" bw (KB/s) : min=%5lu, max=%5lu, per=%3.2f%%," + log_info(" bw (KB/s) : min=%5lu, max=%5lu, per=%3.2f%%," " avg=%5.02f, stdev=%5.02f\n", min, max, p_of_agg, mean, dev); } @@ -642,23 +690,109 @@ static void sum_stat(struct io_stat *dst, struct io_stat *src, int nr) dst->S.u.f = S; } +void sum_group_stats(struct group_run_stats *dst, struct group_run_stats *src) +{ + int i; + + for (i = 0; i < 2; i++) { + if (dst->max_run[i] < src->max_run[i]) + dst->max_run[i] = src->max_run[i]; + if (dst->min_run[i] && dst->min_run[i] > src->min_run[i]) + dst->min_run[i] = src->min_run[i]; + if (dst->max_bw[i] < src->max_bw[i]) + dst->max_bw[i] = src->max_bw[i]; + if (dst->min_bw[i] && dst->min_bw[i] > src->min_bw[i]) + dst->min_bw[i] = src->min_bw[i]; + + dst->io_kb[i] += src->io_kb[i]; + dst->agg[i] += src->agg[i]; + } + +} + +void sum_thread_stats(struct thread_stat *dst, struct thread_stat *src, int nr) +{ + int l, k; + + for (l = 0; l <= DDIR_WRITE; l++) { + sum_stat(&dst->clat_stat[l], &src->clat_stat[l], nr); + sum_stat(&dst->slat_stat[l], &src->slat_stat[l], nr); + sum_stat(&dst->lat_stat[l], &src->lat_stat[l], nr); + sum_stat(&dst->bw_stat[l], &src->bw_stat[l], nr); + + dst->io_bytes[l] += src->io_bytes[l]; + + if (dst->runtime[l] < src->runtime[l]) + dst->runtime[l] = src->runtime[l]; + } + + dst->usr_time += src->usr_time; + dst->sys_time += src->sys_time; + dst->ctx += src->ctx; + dst->majf += src->majf; + dst->minf += src->minf; + + for (k = 0; k < FIO_IO_U_MAP_NR; k++) + dst->io_u_map[k] += src->io_u_map[k]; + for (k = 0; k < FIO_IO_U_MAP_NR; k++) + dst->io_u_submit[k] += src->io_u_submit[k]; + for (k = 0; k < FIO_IO_U_MAP_NR; k++) + dst->io_u_complete[k] += src->io_u_complete[k]; + for (k = 0; k < FIO_IO_U_LAT_U_NR; k++) + dst->io_u_lat_u[k] += src->io_u_lat_u[k]; + for (k = 0; k < FIO_IO_U_LAT_M_NR; k++) + dst->io_u_lat_m[k] += src->io_u_lat_m[k]; + + for (k = 0; k <= 2; k++) { + dst->total_io_u[k] += src->total_io_u[k]; + dst->short_io_u[k] += src->short_io_u[k]; + } + + for (k = 0; k <= DDIR_WRITE; k++) { + int m; + for (m = 0; m < FIO_IO_U_PLAT_NR; m++) + dst->io_u_plat[k][m] += src->io_u_plat[k][m]; + } + + dst->total_run_time += src->total_run_time; + dst->total_submit += src->total_submit; + dst->total_complete += src->total_complete; +} + +void init_group_run_stat(struct group_run_stats *gs) +{ + memset(gs, 0, sizeof(*gs)); + gs->min_bw[0] = gs->min_run[0] = ~0UL; + gs->min_bw[1] = gs->min_run[1] = ~0UL; +} + +void init_thread_stat(struct thread_stat *ts) +{ + int j; + + memset(ts, 0, sizeof(*ts)); + + for (j = 0; j <= DDIR_WRITE; j++) { + ts->lat_stat[j].min_val = -1UL; + ts->clat_stat[j].min_val = -1UL; + ts->slat_stat[j].min_val = -1UL; + ts->bw_stat[j].min_val = -1UL; + } + ts->groupid = -1; +} + void show_run_stats(void) { struct group_run_stats *runstats, *rs; struct thread_data *td; struct thread_stat *threadstats, *ts; - int i, j, k, l, nr_ts, last_ts, idx; + int i, j, nr_ts, last_ts, idx; int kb_base_warned = 0; runstats = malloc(sizeof(struct group_run_stats) * (groupid + 1)); - for (i = 0; i < groupid + 1; i++) { - rs = &runstats[i]; - - memset(rs, 0, sizeof(*rs)); - rs->min_bw[0] = rs->min_run[0] = ~0UL; - rs->min_bw[1] = rs->min_run[1] = ~0UL; - } + for (i = 0; i < groupid + 1; i++) + init_group_run_stat(&runstats[i]); /* * find out how many threads stats we need. if group reporting isn't @@ -680,18 +814,8 @@ void show_run_stats(void) threadstats = malloc(nr_ts * sizeof(struct thread_stat)); - for (i = 0; i < nr_ts; i++) { - ts = &threadstats[i]; - - memset(ts, 0, sizeof(*ts)); - for (j = 0; j <= DDIR_WRITE; j++) { - ts->lat_stat[j].min_val = -1UL; - ts->clat_stat[j].min_val = -1UL; - ts->slat_stat[j].min_val = -1UL; - ts->bw_stat[j].min_val = -1UL; - } - ts->groupid = -1; - } + for (i = 0; i < nr_ts; i++) + init_thread_stat(&threadstats[i]); j = 0; last_ts = -1; @@ -755,50 +879,7 @@ void show_run_stats(void) } } - for (l = 0; l <= DDIR_WRITE; l++) { - sum_stat(&ts->clat_stat[l], &td->ts.clat_stat[l], idx); - sum_stat(&ts->slat_stat[l], &td->ts.slat_stat[l], idx); - sum_stat(&ts->lat_stat[l], &td->ts.lat_stat[l], idx); - sum_stat(&ts->bw_stat[l], &td->ts.bw_stat[l], idx); - - ts->io_bytes[l] += td->ts.io_bytes[l]; - - if (ts->runtime[l] < td->ts.runtime[l]) - ts->runtime[l] = td->ts.runtime[l]; - } - - ts->usr_time += td->ts.usr_time; - ts->sys_time += td->ts.sys_time; - ts->ctx += td->ts.ctx; - ts->majf += td->ts.majf; - ts->minf += td->ts.minf; - - for (k = 0; k < FIO_IO_U_MAP_NR; k++) - ts->io_u_map[k] += td->ts.io_u_map[k]; - for (k = 0; k < FIO_IO_U_MAP_NR; k++) - ts->io_u_submit[k] += td->ts.io_u_submit[k]; - for (k = 0; k < FIO_IO_U_MAP_NR; k++) - ts->io_u_complete[k] += td->ts.io_u_complete[k]; - for (k = 0; k < FIO_IO_U_LAT_U_NR; k++) - ts->io_u_lat_u[k] += td->ts.io_u_lat_u[k]; - for (k = 0; k < FIO_IO_U_LAT_M_NR; k++) - ts->io_u_lat_m[k] += td->ts.io_u_lat_m[k]; - - - for (k = 0; k <= 2; k++) { - ts->total_io_u[k] += td->ts.total_io_u[k]; - ts->short_io_u[k] += td->ts.short_io_u[k]; - } - - for (k = 0; k <= DDIR_WRITE; k++) { - int m; - for (m = 0; m < FIO_IO_U_PLAT_NR; m++) - ts->io_u_plat[k][m] += td->ts.io_u_plat[k][m]; - } - - ts->total_run_time += td->ts.total_run_time; - ts->total_submit += td->ts.total_submit; - ts->total_complete += td->ts.total_complete; + sum_thread_stats(ts, &td->ts, idx); } for (i = 0; i < nr_ts; i++) { @@ -874,7 +955,12 @@ void show_run_stats(void) show_group_stats(rs); } - show_disk_util(); + if (is_backend) + fio_server_send_du(); + else + show_disk_util(); + + free_disk_util(); } free(runstats); @@ -1004,7 +1090,7 @@ void add_bw_sample(struct thread_data *td, enum fio_ddir ddir, unsigned int bs, if (!ddir_rw(ddir)) return; - spent = mtime_since(&td->stat_sample_time[ddir], t); + spent = mtime_since(&td->bw_sample_time, t); if (spent < td->o.bw_avg_time) return; @@ -1015,6 +1101,32 @@ void add_bw_sample(struct thread_data *td, enum fio_ddir ddir, unsigned int bs, if (td->bw_log) add_log_sample(td, td->bw_log, rate, ddir, bs); - fio_gettime(&td->stat_sample_time[ddir], NULL); + fio_gettime(&td->bw_sample_time, NULL); td->stat_io_bytes[ddir] = td->this_io_bytes[ddir]; } + +void add_iops_sample(struct thread_data *td, enum fio_ddir ddir, + struct timeval *t) +{ + struct thread_stat *ts = &td->ts; + unsigned long spent, iops; + + if (!ddir_rw(ddir)) + return; + + spent = mtime_since(&td->iops_sample_time, t); + if (spent < td->o.iops_avg_time) + return; + + iops = ((td->this_io_blocks[ddir] - td->stat_io_blocks[ddir]) * 1000) / spent; + + add_stat_sample(&ts->iops_stat[ddir], iops); + + if (td->iops_log) { + assert(iops); + add_log_sample(td, td->iops_log, iops, ddir, 0); + } + + fio_gettime(&td->iops_sample_time, NULL); + td->stat_io_blocks[ddir] = td->this_io_blocks[ddir]; +}