static void convert_ts(struct thread_stat *dst, struct thread_stat *src)
{
- int i, j;
+ int i, j, k;
dst->error = le32_to_cpu(src->error);
dst->thread_number = le32_to_cpu(src->thread_number);
dst->majf = le64_to_cpu(src->majf);
dst->clat_percentiles = le32_to_cpu(src->clat_percentiles);
dst->lat_percentiles = le32_to_cpu(src->lat_percentiles);
+ dst->slat_percentiles = le32_to_cpu(src->slat_percentiles);
dst->percentile_precision = le64_to_cpu(src->percentile_precision);
for (i = 0; i < FIO_IO_U_LIST_MAX_LEN; i++) {
for (i = 0; i < FIO_IO_U_LAT_M_NR; i++)
dst->io_u_lat_m[i] = le64_to_cpu(src->io_u_lat_m[i]);
- for (i = 0; i < DDIR_RWDIR_CNT; i++)
- for (j = 0; j < FIO_IO_U_PLAT_NR; j++)
- dst->io_u_plat[i][j] = le64_to_cpu(src->io_u_plat[i][j]);
+ for (i = 0; i < FIO_LAT_CNT; i++)
+ for (j = 0; j < DDIR_RWDIR_CNT; j++)
+ for (k = 0; k < FIO_IO_U_PLAT_NR; k++)
+ dst->io_u_plat[i][j][k] = le64_to_cpu(src->io_u_plat[i][j][k]);
for (j = 0; j < FIO_IO_U_PLAT_NR; j++)
dst->io_u_sync_plat[j] = le64_to_cpu(src->io_u_sync_plat[j]);
dst->nr_block_infos = le64_to_cpu(src->nr_block_infos);
for (i = 0; i < dst->nr_block_infos; i++)
dst->block_infos[i] = le32_to_cpu(src->block_infos[i]);
+ for (i = 0; i < DDIR_RWDIR_CNT; i++) {
+ for (j = 0; j < FIO_IO_U_PLAT_NR; j++) {
+ dst->io_u_plat_high_prio[i][j] = le64_to_cpu(src->io_u_plat_high_prio[i][j]);
+ dst->io_u_plat_low_prio[i][j] = le64_to_cpu(src->io_u_plat_low_prio[i][j]);
+ }
+ convert_io_stat(&dst->clat_high_prio_stat[i], &src->clat_high_prio_stat[i]);
+ convert_io_stat(&dst->clat_low_prio_stat[i], &src->clat_low_prio_stat[i]);
+ }
dst->ss_dur = le64_to_cpu(src->ss_dur);
dst->ss_state = le32_to_cpu(src->ss_state);
struct flist_head *opt_list = &client->opt_lists[pdu->groupid];
flist_add_tail(&p->list, opt_list);
+ } else {
+ free(p->value);
+ free(p->name);
+ free(p);
}
}
s->time = le64_to_cpu(s->time);
s->data.val = le64_to_cpu(s->data.val);
- s->__ddir = le32_to_cpu(s->__ddir);
+ s->__ddir = __le32_to_cpu(s->__ddir);
s->bs = le64_to_cpu(s->bs);
if (ret->log_offset) {