Add support for latency probing over an interval of load
[fio.git] / cconv.c
diff --git a/cconv.c b/cconv.c
index 534bfb07214078f4a0d12f84691579edb8259a57..4040be28f497cae6dcc57f96287fb29c987f0c89 100644 (file)
--- a/cconv.c
+++ b/cconv.c
@@ -37,6 +37,7 @@ static void free_thread_options_to_cpu(struct thread_options *o)
        free(o->mmapfile);
        free(o->read_iolog_file);
        free(o->write_iolog_file);
+       free(o->merge_blktrace_file);
        free(o->bw_log_file);
        free(o->lat_log_file);
        free(o->iops_log_file);
@@ -73,6 +74,7 @@ void convert_thread_options_to_cpu(struct thread_options *o,
        string_to_cpu(&o->mmapfile, top->mmapfile);
        string_to_cpu(&o->read_iolog_file, top->read_iolog_file);
        string_to_cpu(&o->write_iolog_file, top->write_iolog_file);
+       string_to_cpu(&o->merge_blktrace_file, top->merge_blktrace_file);
        string_to_cpu(&o->bw_log_file, top->bw_log_file);
        string_to_cpu(&o->lat_log_file, top->lat_log_file);
        string_to_cpu(&o->iops_log_file, top->iops_log_file);
@@ -98,6 +100,12 @@ void convert_thread_options_to_cpu(struct thread_options *o,
        o->iodepth_batch_complete_min = le32_to_cpu(top->iodepth_batch_complete_min);
        o->iodepth_batch_complete_max = le32_to_cpu(top->iodepth_batch_complete_max);
        o->serialize_overlap = le32_to_cpu(top->serialize_overlap);
+       o->iodepth_mode = le32_to_cpu(top->iodepth_mode);
+       o->lat_step_low = le32_to_cpu(top->lat_step_low);
+       o->lat_step_high = le32_to_cpu(top->lat_step_high);
+       o->lat_step_inc = le32_to_cpu(top->lat_step_inc);
+       o->lat_step_ramp = le32_to_cpu(top->lat_step_ramp);
+       o->lat_step_run = le32_to_cpu(top->lat_step_run);
        o->size = le64_to_cpu(top->size);
        o->io_size = le64_to_cpu(top->io_size);
        o->size_percent = le32_to_cpu(top->size_percent);
@@ -223,6 +231,7 @@ void convert_thread_options_to_cpu(struct thread_options *o,
        o->zone_range = le64_to_cpu(top->zone_range);
        o->zone_size = le64_to_cpu(top->zone_size);
        o->zone_skip = le64_to_cpu(top->zone_skip);
+       o->zone_mode = le32_to_cpu(top->zone_mode);
        o->lockmem = le64_to_cpu(top->lockmem);
        o->offset_increment = le64_to_cpu(top->offset_increment);
        o->number_ios = le64_to_cpu(top->number_ios);
@@ -303,6 +312,12 @@ void convert_thread_options_to_cpu(struct thread_options *o,
 
        for (i = 0; i < FIO_IO_U_LIST_MAX_LEN; i++)
                o->percentile_list[i].u.f = fio_uint64_to_double(le64_to_cpu(top->percentile_list[i].u.i));
+
+       for (i = 0; i < FIO_IO_U_LIST_MAX_LEN; i++)
+               o->merge_blktrace_scalars[i].u.f = fio_uint64_to_double(le64_to_cpu(top->merge_blktrace_scalars[i].u.i));
+
+       for (i = 0; i < FIO_IO_U_LIST_MAX_LEN; i++)
+               o->merge_blktrace_iters[i].u.f = fio_uint64_to_double(le64_to_cpu(top->merge_blktrace_iters[i].u.i));
 #if 0
        uint8_t cpumask[FIO_TOP_STR_MAX];
        uint8_t verify_cpumask[FIO_TOP_STR_MAX];
@@ -329,6 +344,7 @@ void convert_thread_options_to_net(struct thread_options_pack *top,
        string_to_net(top->mmapfile, o->mmapfile);
        string_to_net(top->read_iolog_file, o->read_iolog_file);
        string_to_net(top->write_iolog_file, o->write_iolog_file);
+       string_to_net(top->merge_blktrace_file, o->merge_blktrace_file);
        string_to_net(top->bw_log_file, o->bw_log_file);
        string_to_net(top->lat_log_file, o->lat_log_file);
        string_to_net(top->iops_log_file, o->iops_log_file);
@@ -353,6 +369,12 @@ void convert_thread_options_to_net(struct thread_options_pack *top,
        top->iodepth_batch_complete_min = cpu_to_le32(o->iodepth_batch_complete_min);
        top->iodepth_batch_complete_max = cpu_to_le32(o->iodepth_batch_complete_max);
        top->serialize_overlap = cpu_to_le32(o->serialize_overlap);
+       top->iodepth_mode = cpu_to_le32(o->iodepth_mode);
+       top->lat_step_low = cpu_to_le32(o->lat_step_low);
+       top->lat_step_high = cpu_to_le32(o->lat_step_high);
+       top->lat_step_inc = cpu_to_le32(o->lat_step_inc);
+       top->lat_step_ramp = cpu_to_le32(o->lat_step_ramp);
+       top->lat_step_run = cpu_to_le32(o->lat_step_run);
        top->size_percent = cpu_to_le32(o->size_percent);
        top->fill_device = cpu_to_le32(o->fill_device);
        top->file_append = cpu_to_le32(o->file_append);
@@ -548,6 +570,7 @@ void convert_thread_options_to_net(struct thread_options_pack *top,
        top->zone_range = __cpu_to_le64(o->zone_range);
        top->zone_size = __cpu_to_le64(o->zone_size);
        top->zone_skip = __cpu_to_le64(o->zone_skip);
+       top->zone_mode = __cpu_to_le32(o->zone_mode);
        top->lockmem = __cpu_to_le64(o->lockmem);
        top->ddir_seq_add = __cpu_to_le64(o->ddir_seq_add);
        top->file_size_low = __cpu_to_le64(o->file_size_low);
@@ -563,6 +586,12 @@ void convert_thread_options_to_net(struct thread_options_pack *top,
 
        for (i = 0; i < FIO_IO_U_LIST_MAX_LEN; i++)
                top->percentile_list[i].u.i = __cpu_to_le64(fio_double_to_uint64(o->percentile_list[i].u.f));
+
+       for (i = 0; i < FIO_IO_U_LIST_MAX_LEN; i++)
+               top->merge_blktrace_scalars[i].u.i = __cpu_to_le64(fio_double_to_uint64(o->merge_blktrace_scalars[i].u.f));
+
+       for (i = 0; i < FIO_IO_U_LIST_MAX_LEN; i++)
+               top->merge_blktrace_iters[i].u.i = __cpu_to_le64(fio_double_to_uint64(o->merge_blktrace_iters[i].u.f));
 #if 0
        uint8_t cpumask[FIO_TOP_STR_MAX];
        uint8_t verify_cpumask[FIO_TOP_STR_MAX];