qsort(bssplit, td->o.bssplit_nr[ddir], sizeof(struct bssplit), bs_cmp);
td->o.bssplit[ddir] = bssplit;
return 0;
-
}
static int str_bssplit_cb(void *data, const char *input)
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(description),
.help = "Text job description",
+ .category = FIO_OPT_G_DESC,
},
{
.name = "name",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(name),
.help = "Name of this job",
+ .category = FIO_OPT_G_DESC,
},
{
.name = "directory",
.off1 = td_var_offset(directory),
.cb = str_directory_cb,
.help = "Directory to store files in",
+ .category = FIO_OPT_G_FILE,
},
{
.name = "filename",
.cb = str_filename_cb,
.prio = -1, /* must come after "directory" */
.help = "File(s) to use for the workload",
+ .category = FIO_OPT_G_FILE,
},
{
.name = "kb_base",
.prio = 1,
.def = "1024",
.help = "How many bytes per KB for reporting (1000 or 1024)",
+ .category = FIO_OPT_G_MISC,
},
{
.name = "lockfile",
.help = "Lock file when doing IO to it",
.parent = "filename",
.def = "none",
+ .category = FIO_OPT_G_FILE,
.posval = {
{ .ival = "none",
.oval = FILE_LOCK_NONE,
.off1 = td_var_offset(opendir),
.cb = str_opendir_cb,
.help = "Recursively add files from this directory and down",
+ .category = FIO_OPT_G_FILE,
},
{
.name = "rw",
.help = "IO direction",
.def = "read",
.verify = rw_verify,
+ .category = FIO_OPT_G_IO_DDIR,
.posval = {
{ .ival = "read",
.oval = TD_DDIR_READ,
.off1 = td_var_offset(rw_seq),
.help = "IO offset generator modifier",
.def = "sequential",
+ .category = FIO_OPT_G_IO_DDIR,
.posval = {
{ .ival = "sequential",
.oval = RW_SEQ_SEQ,
.off1 = td_var_offset(ioengine),
.help = "IO engine to use",
.def = FIO_PREFERRED_ENGINE,
+ .category = FIO_OPT_G_IO,
.posval = {
{ .ival = "sync",
.help = "Use read/write",
.help = "Number of IO buffers to keep in flight",
.minval = 1,
.def = "1",
+ .category = FIO_OPT_G_IO,
},
{
.name = "iodepth_batch",
.parent = "iodepth",
.minval = 1,
.def = "1",
+ .category = FIO_OPT_G_IO,
},
{
.name = "iodepth_batch_complete",
.parent = "iodepth",
.minval = 0,
.def = "1",
+ .category = FIO_OPT_G_IO,
},
{
.name = "iodepth_low",
.off1 = td_var_offset(iodepth_low),
.help = "Low water mark for queuing depth",
.parent = "iodepth",
+ .category = FIO_OPT_G_IO,
},
{
.name = "size",
.type = FIO_OPT_STR_VAL,
.cb = str_size_cb,
.help = "Total size of device or files",
+ .category = FIO_OPT_G_IO,
},
{
.name = "fill_device",
.off1 = td_var_offset(fill_device),
.help = "Write until an ENOSPC error occurs",
.def = "0",
+ .category = FIO_OPT_G_IO,
},
{
.name = "filesize",
.off2 = td_var_offset(file_size_high),
.minval = 1,
.help = "Size of individual files",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_FILE,
},
{
.name = "offset",
.off1 = td_var_offset(start_offset),
.help = "Start IO from this offset",
.def = "0",
+ .category = FIO_OPT_G_IO,
},
{
.name = "bs",
.help = "Block size unit",
.def = "4k",
.parent = "rw",
+ .category = FIO_OPT_G_IO,
},
{
.name = "ba",
.minval = 1,
.help = "IO block offset alignment",
.parent = "rw",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_IO_BUF,
},
{
.name = "bsrange",
.minval = 1,
.help = "Set block size range (in more detail than bs)",
.parent = "rw",
+ .category = FIO_OPT_G_IO,
},
{
.name = "bssplit",
.cb = str_bssplit_cb,
.help = "Set a specific mix of block sizes",
.parent = "rw",
+ .category = FIO_OPT_G_IO,
},
{
.name = "bs_unaligned",
.off1 = td_var_offset(bs_unaligned),
.help = "Don't sector align IO buffer sizes",
.parent = "rw",
+ .category = FIO_OPT_G_IO,
},
{
.name = "randrepeat",
.help = "Use repeatable random IO pattern",
.def = "1",
.parent = "rw",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_RAND,
},
{
.name = "use_os_rand",
.help = "Set to use OS random generator",
.def = "0",
.parent = "rw",
+ .category = FIO_OPT_G_RAND,
},
{
.name = "norandommap",
.off1 = td_var_offset(norandommap),
.help = "Accept potential duplicate random blocks",
.parent = "rw",
+ .category = FIO_OPT_G_RAND,
},
{
.name = "softrandommap",
.help = "Set norandommap if randommap allocation fails",
.parent = "norandommap",
.def = "0",
+ .category = FIO_OPT_G_RAND,
},
{
.name = "nrfiles",
.off1 = td_var_offset(nr_files),
.help = "Split job workload between this number of files",
.def = "1",
+ .category = FIO_OPT_G_FILE,
},
{
.name = "openfiles",
.type = FIO_OPT_INT,
.off1 = td_var_offset(open_files),
.help = "Number of files to keep open at the same time",
+ .category = FIO_OPT_G_FILE,
},
{
.name = "file_service_type",
.off1 = td_var_offset(file_service_type),
.help = "How to select which file to service next",
.def = "roundrobin",
+ .category = FIO_OPT_G_FILE,
.posval = {
{ .ival = "random",
.oval = FIO_FSERVICE_RANDOM,
.off1 = td_var_offset(fallocate_mode),
.help = "Whether pre-allocation is performed when laying out files",
.def = "posix",
+ .category = FIO_OPT_G_FILE,
.posval = {
{ .ival = "none",
.oval = FIO_FALLOCATE_NONE,
.off1 = td_var_offset(fadvise_hint),
.help = "Use fadvise() to advise the kernel on IO pattern",
.def = "1",
+ .category = FIO_OPT_G_FILE,
},
{
.name = "fsync",
.off1 = td_var_offset(fsync_blocks),
.help = "Issue fsync for writes every given number of blocks",
.def = "0",
+ .category = FIO_OPT_G_FILE,
},
{
.name = "fdatasync",
.off1 = td_var_offset(fdatasync_blocks),
.help = "Issue fdatasync for writes every given number of blocks",
.def = "0",
+ .category = FIO_OPT_G_FILE,
},
{
.name = "write_barrier",
.off1 = td_var_offset(barrier_blocks),
.help = "Make every Nth write a barrier write",
.def = "0",
+ .category = FIO_OPT_G_IO,
},
#ifdef FIO_HAVE_SYNC_FILE_RANGE
{
.cb = str_sfr_cb,
.off1 = td_var_offset(sync_file_range),
.help = "Use sync_file_range()",
+ .category = FIO_OPT_G_FILE,
},
#endif
{
.off1 = td_var_offset(odirect),
.help = "Use O_DIRECT IO (negates buffered)",
.def = "0",
+ .category = FIO_OPT_G_IO,
},
{
.name = "buffered",
.neg = 1,
.help = "Use buffered IO (negates direct)",
.def = "1",
+ .category = FIO_OPT_G_IO,
},
{
.name = "overwrite",
.off1 = td_var_offset(overwrite),
.help = "When writing, set whether to overwrite current data",
.def = "0",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_FILE,
},
{
.name = "loops",
.off1 = td_var_offset(loops),
.help = "Number of times to run the job",
.def = "1",
+ .category = FIO_OPT_G_MISC,
},
{
.name = "numjobs",
.off1 = td_var_offset(numjobs),
.help = "Duplicate this job this many times",
.def = "1",
+ .category = FIO_OPT_G_MISC,
},
{
.name = "startdelay",
.off1 = td_var_offset(start_delay),
.help = "Only start job when this period has passed",
.def = "0",
+ .category = FIO_OPT_G_MISC,
},
{
.name = "runtime",
.off1 = td_var_offset(timeout),
.help = "Stop workload when this amount of time has passed",
.def = "0",
+ .category = FIO_OPT_G_MISC,
},
{
.name = "time_based",
.type = FIO_OPT_STR_SET,
.off1 = td_var_offset(time_based),
.help = "Keep running until runtime/timeout is met",
+ .category = FIO_OPT_G_MISC,
},
{
.name = "ramp_time",
.type = FIO_OPT_STR_VAL_TIME,
.off1 = td_var_offset(ramp_time),
.help = "Ramp up time before measuring performance",
+ .category = FIO_OPT_G_MISC,
},
{
.name = "clocksource",
.cb = fio_clock_source_cb,
.off1 = td_var_offset(clocksource),
.help = "What type of timing source to use",
+ .category = FIO_OPT_G_OS,
.posval = {
{ .ival = "gettimeofday",
.oval = CS_GTOD,
.off1 = td_var_offset(mem_type),
.help = "Backing type for IO buffers",
.def = "malloc",
+ .category = FIO_OPT_G_IO_BUF | FIO_OPT_G_MEM,
.posval = {
{ .ival = "malloc",
.oval = MEM_MALLOC,
.help = "IO memory buffer offset alignment",
.def = "0",
.parent = "iomem",
+ .category = FIO_OPT_G_IO_BUF | FIO_OPT_G_MEM,
},
{
.name = "verify",
.help = "Verify data written",
.cb = str_verify_cb,
.def = "0",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_VERIFY,
.posval = {
{ .ival = "0",
.oval = VERIFY_NONE,
.help = "Run verification stage after write",
.def = "1",
.parent = "verify",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_VERIFY,
},
{
.name = "verifysort",
.help = "Sort written verify blocks for read back",
.def = "1",
.parent = "verify",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_VERIFY,
},
{
.name = "verify_interval",
.minval = 2 * sizeof(struct verify_header),
.help = "Store verify buffer header every N bytes",
.parent = "verify",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_VERIFY,
},
{
.name = "verify_offset",
.def = "0",
.cb = str_verify_offset_cb,
.parent = "verify",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_VERIFY,
},
{
.name = "verify_pattern",
.cb = str_verify_pattern_cb,
.help = "Fill pattern for IO buffers",
.parent = "verify",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_VERIFY,
},
{
.name = "verify_fatal",
.def = "0",
.help = "Exit on a single verify failure, don't continue",
.parent = "verify",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_VERIFY | FIO_OPT_G_ERR,
},
{
.name = "verify_dump",
.def = "0",
.help = "Dump contents of good and bad blocks on failure",
.parent = "verify",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_VERIFY | FIO_OPT_G_ERR,
},
{
.name = "verify_async",
.def = "0",
.help = "Number of async verifier threads to use",
.parent = "verify",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_VERIFY,
},
{
.name = "verify_backlog",
.off1 = td_var_offset(verify_backlog),
.help = "Verify after this number of blocks are written",
.parent = "verify",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_VERIFY,
},
{
.name = "verify_backlog_batch",
.off1 = td_var_offset(verify_batch),
.help = "Verify this number of IO blocks",
.parent = "verify",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_VERIFY,
},
#ifdef FIO_HAVE_CPU_AFFINITY
{
.cb = str_verify_cpus_allowed_cb,
.help = "Set CPUs allowed for async verify threads",
.parent = "verify_async",
+ .category = FIO_OPT_G_OS | FIO_OPT_G_CPU | FIO_OPT_G_VERIFY,
},
#endif
#ifdef FIO_HAVE_TRIM
.help = "Number of verify blocks to discard/trim",
.parent = "verify",
.def = "0",
+ .category = FIO_OPT_G_IO,
},
{
.name = "trim_verify_zero",
.off1 = td_var_offset(trim_zero),
.parent = "trim_percentage",
.def = "1",
+ .category = FIO_OPT_G_IO,
},
{
.name = "trim_backlog",
.off1 = td_var_offset(trim_backlog),
.help = "Trim after this number of blocks are written",
.parent = "trim_percentage",
+ .category = FIO_OPT_G_IO,
},
{
.name = "trim_backlog_batch",
.off1 = td_var_offset(trim_batch),
.help = "Trim this number of IO blocks",
.parent = "trim_percentage",
+ .category = FIO_OPT_G_IO,
},
#endif
{
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(write_iolog_file),
.help = "Store IO pattern to file",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_LOG,
},
{
.name = "read_iolog",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(read_iolog_file),
.help = "Playback IO pattern from file",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_LOG,
},
{
.name = "replay_no_stall",
.def = "0",
.parent = "read_iolog",
.help = "Playback IO pattern file as fast as possible without stalls",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_LOG,
},
{
.name = "replay_redirect",
.off1 = td_var_offset(replay_redirect),
.parent = "read_iolog",
.help = "Replay all I/O onto this device, regardless of trace device",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_LOG,
},
{
.name = "exec_prerun",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(exec_prerun),
.help = "Execute this file prior to running job",
+ .category = FIO_OPT_G_MISC | FIO_OPT_G_OS,
},
{
.name = "exec_postrun",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(exec_postrun),
.help = "Execute this file after running job",
+ .category = FIO_OPT_G_MISC | FIO_OPT_G_OS,
},
#ifdef FIO_HAVE_IOSCHED_SWITCH
{
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(ioscheduler),
.help = "Use this IO scheduler on the backing device",
+ .category = FIO_OPT_G_OS | FIO_OPT_G_IO,
},
#endif
{
.off1 = td_var_offset(zone_size),
.help = "Amount of data to read per zone",
.def = "0",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_ZONE,
},
{
.name = "zonerange",
.off1 = td_var_offset(zone_range),
.help = "Give size of an IO zone",
.def = "0",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_ZONE,
},
{
.name = "zoneskip",
.off1 = td_var_offset(zone_skip),
.help = "Space between IO zones",
.def = "0",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_ZONE,
},
{
.name = "lockmem",
.cb = str_lockmem_cb,
.help = "Lock down this amount of memory",
.def = "0",
+ .category = FIO_OPT_G_OS | FIO_OPT_G_MEM,
},
{
.name = "rwmixread",
.maxval = 100,
.help = "Percentage of mixed workload that is reads",
.def = "50",
+ .category = FIO_OPT_G_IO,
},
{
.name = "rwmixwrite",
.maxval = 100,
.help = "Percentage of mixed workload that is writes",
.def = "50",
+ .category = FIO_OPT_G_IO,
},
{
.name = "rwmixcycle",
.type = FIO_OPT_DEPRECATED,
+ .category = FIO_OPT_G_IO,
},
{
.name = "nice",
.minval = -19,
.maxval = 20,
.def = "0",
+ .category = FIO_OPT_G_OS | FIO_OPT_G_CPU,
},
#ifdef FIO_HAVE_IOPRIO
{
.help = "Set job IO priority value",
.minval = 0,
.maxval = 7,
+ .category = FIO_OPT_G_OS | FIO_OPT_G_CPU,
},
{
.name = "prioclass",
.help = "Set job IO priority class",
.minval = 0,
.maxval = 3,
+ .category = FIO_OPT_G_OS | FIO_OPT_G_CPU,
},
#endif
{
.off1 = td_var_offset(thinktime),
.help = "Idle time between IO buffers (usec)",
.def = "0",
+ .category = FIO_OPT_G_MISC,
},
{
.name = "thinktime_spin",
.help = "Start think time by spinning this amount (usec)",
.def = "0",
.parent = "thinktime",
+ .category = FIO_OPT_G_MISC,
},
{
.name = "thinktime_blocks",
.help = "IO buffer period between 'thinktime'",
.def = "1",
.parent = "thinktime",
+ .category = FIO_OPT_G_MISC,
},
{
.name = "rate",
.off1 = td_var_offset(rate[0]),
.off2 = td_var_offset(rate[1]),
.help = "Set bandwidth rate",
+ .category = FIO_OPT_G_IO,
},
{
.name = "ratemin",
.off2 = td_var_offset(ratemin[1]),
.help = "Job must meet this rate or it will be shutdown",
.parent = "rate",
+ .category = FIO_OPT_G_IO,
},
{
.name = "rate_iops",
.off1 = td_var_offset(rate_iops[0]),
.off2 = td_var_offset(rate_iops[1]),
.help = "Limit IO used to this number of IO operations/sec",
+ .category = FIO_OPT_G_IO,
},
{
.name = "rate_iops_min",
.off2 = td_var_offset(rate_iops_min[1]),
.help = "Job must meet this rate or it will be shut down",
.parent = "rate_iops",
+ .category = FIO_OPT_G_IO,
},
{
.name = "ratecycle",
.help = "Window average for rate limits (msec)",
.def = "1000",
.parent = "rate",
+ .category = FIO_OPT_G_IO,
},
{
.name = "invalidate",
.off1 = td_var_offset(invalidate_cache),
.help = "Invalidate buffer/page cache prior to running job",
.def = "1",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_CACHE,
},
{
.name = "sync",
.help = "Use O_SYNC for buffered writes",
.def = "0",
.parent = "buffered",
+ .category = FIO_OPT_G_IO | FIO_OPT_G_FILE,
},
{
.name = "bwavgtime",
" (msec)",
.def = "500",
.parent = "write_bw_log",
+ .category = FIO_OPT_G_LOG | FIO_OPT_G_STAT,
},
{
.name = "iopsavgtime",
.help = "Time window over which to calculate IOPS (msec)",
.def = "500",
.parent = "write_iops_log",
+ .category = FIO_OPT_G_LOG | FIO_OPT_G_STAT,
},
{
.name = "create_serialize",
.off1 = td_var_offset(create_serialize),
.help = "Serialize creating of job files",
.def = "1",
+ .category = FIO_OPT_G_FILE,
},
{
.name = "create_fsync",
.off1 = td_var_offset(create_fsync),
.help = "fsync file after creation",
.def = "1",
+ .category = FIO_OPT_G_FILE,
},
{
.name = "create_on_open",
.off1 = td_var_offset(create_on_open),
.help = "Create files when they are opened for IO",
.def = "0",
+ .category = FIO_OPT_G_FILE,
},
{
.name = "pre_read",
.off1 = td_var_offset(pre_read),
.help = "Pre-read files before starting official testing",
.def = "0",
+ .category = FIO_OPT_G_FILE | FIO_OPT_G_CACHE,
},
{
.name = "cpuload",
.type = FIO_OPT_INT,
.off1 = td_var_offset(cpuload),
.help = "Use this percentage of CPU",
+ .category = FIO_OPT_G_CPU,
},
{
.name = "cpuchunks",
.help = "Length of the CPU burn cycles (usecs)",
.def = "50000",
.parent = "cpuload",
+ .category = FIO_OPT_G_CPU,
},
#ifdef FIO_HAVE_CPU_AFFINITY
{
.type = FIO_OPT_INT,
.cb = str_cpumask_cb,
.help = "CPU affinity mask",
+ .category = FIO_OPT_G_CPU | FIO_OPT_G_OS,
},
{
.name = "cpus_allowed",
.type = FIO_OPT_STR,
.cb = str_cpus_allowed_cb,
.help = "Set CPUs allowed",
+ .category = FIO_OPT_G_CPU | FIO_OPT_G_OS,
},
#endif
{
.off1 = td_var_offset(end_fsync),
.help = "Include fsync at the end of job",
.def = "0",
+ .category = FIO_OPT_G_FILE,
},
{
.name = "fsync_on_close",
.off1 = td_var_offset(fsync_on_close),
.help = "fsync files on close",
.def = "0",
+ .category = FIO_OPT_G_FILE,
},
{
.name = "unlink",
.off1 = td_var_offset(unlink),
.help = "Unlink created files after job has completed",
.def = "0",
+ .category = FIO_OPT_G_FILE,
},
{
.name = "exitall",
.type = FIO_OPT_STR_SET,
.cb = str_exitall_cb,
.help = "Terminate all jobs when one exits",
+ .category = FIO_OPT_G_MISC | FIO_OPT_G_JOB,
},
{
.name = "stonewall",
.type = FIO_OPT_STR_SET,
.off1 = td_var_offset(stonewall),
.help = "Insert a hard barrier between this job and previous",
+ .category = FIO_OPT_G_MISC | FIO_OPT_G_JOB,
},
{
.name = "new_group",
.type = FIO_OPT_STR_SET,
.off1 = td_var_offset(new_group),
.help = "Mark the start of a new group (for reporting)",
+ .category = FIO_OPT_G_MISC | FIO_OPT_G_JOB,
},
{
.name = "thread",
.type = FIO_OPT_STR_SET,
.off1 = td_var_offset(use_thread),
.help = "Use threads instead of forks",
+ .category = FIO_OPT_G_MISC | FIO_OPT_G_OS | FIO_OPT_G_JOB,
},
{
.name = "write_bw_log",
.off1 = td_var_offset(write_bw_log),
.cb = str_write_bw_log_cb,
.help = "Write log of bandwidth during run",
+ .category = FIO_OPT_G_LOG,
},
{
.name = "write_lat_log",
.off1 = td_var_offset(write_lat_log),
.cb = str_write_lat_log_cb,
.help = "Write log of latency during run",
+ .category = FIO_OPT_G_LOG,
},
{
.name = "write_iops_log",
.off1 = td_var_offset(write_iops_log),
.cb = str_write_iops_log_cb,
.help = "Write log of IOPS during run",
+ .category = FIO_OPT_G_LOG,
},
{
.name = "log_avg_msec",
.off1 = td_var_offset(log_avg_msec),
.help = "Average bw/iops/lat logs over this period of time",
.def = "0",
+ .category = FIO_OPT_G_LOG,
},
{
.name = "hugepage-size",
.off1 = td_var_offset(hugepage_size),
.help = "When using hugepages, specify size of each page",
.def = __fio_stringify(FIO_HUGE_PAGE),
+ .category = FIO_OPT_G_OS | FIO_OPT_G_MEM,
},
{
.name = "group_reporting",
.type = FIO_OPT_STR_SET,
.off1 = td_var_offset(group_reporting),
.help = "Do reporting on a per-group basis",
+ .category = FIO_OPT_G_MISC,
},
{
.name = "zero_buffers",
.type = FIO_OPT_STR_SET,
.off1 = td_var_offset(zero_buffers),
.help = "Init IO buffers to all zeroes",
+ .category = FIO_OPT_G_IO_BUF,
},
{
.name = "refill_buffers",
.type = FIO_OPT_STR_SET,
.off1 = td_var_offset(refill_buffers),
.help = "Refill IO buffers on every IO submit",
+ .category = FIO_OPT_G_IO_BUF,
},
{
.name = "scramble_buffers",
.off1 = td_var_offset(scramble_buffers),
.help = "Slightly scramble buffers on every IO submit",
.def = "1",
+ .category = FIO_OPT_G_IO_BUF,
+ },
+ {
+ .name = "buffer_compress_percentage",
+ .type = FIO_OPT_INT,
+ .off1 = td_var_offset(compress_percentage),
+ .maxval = 100,
+ .minval = 1,
+ .help = "How compressible the buffer is (approximately)",
+ },
+ {
+ .name = "buffer_compress_chunk",
+ .type = FIO_OPT_INT,
+ .off1 = td_var_offset(compress_chunk),
+ .parent = "buffer_compress_percentage",
+ .help = "Size of compressible region in buffer",
},
{
.name = "clat_percentiles",
.off1 = td_var_offset(clat_percentiles),
.help = "Enable the reporting of completion latency percentiles",
.def = "1",
+ .category = FIO_OPT_G_STAT,
},
{
.name = "percentile_list",
.maxlen = FIO_IO_U_LIST_MAX_LEN,
.minfp = 0.0,
.maxfp = 100.0,
+ .category = FIO_OPT_G_STAT,
},
#ifdef FIO_HAVE_DISK_UTIL
.off1 = td_var_offset(do_disk_util),
.help = "Log disk utilization statistics",
.def = "1",
+ .category = FIO_OPT_G_OS | FIO_OPT_G_STAT,
},
#endif
{
.help = "Greatly reduce number of gettimeofday() calls",
.cb = str_gtod_reduce_cb,
.def = "0",
+ .category = FIO_OPT_G_OS | FIO_OPT_G_MISC | FIO_OPT_G_STAT,
},
{
.name = "disable_lat",
.help = "Disable latency numbers",
.parent = "gtod_reduce",
.def = "0",
+ .category = FIO_OPT_G_OS | FIO_OPT_G_MISC | FIO_OPT_G_STAT,
},
{
.name = "disable_clat",
.help = "Disable completion latency numbers",
.parent = "gtod_reduce",
.def = "0",
+ .category = FIO_OPT_G_OS | FIO_OPT_G_MISC | FIO_OPT_G_STAT,
},
{
.name = "disable_slat",
.help = "Disable submission latency numbers",
.parent = "gtod_reduce",
.def = "0",
+ .category = FIO_OPT_G_OS | FIO_OPT_G_MISC | FIO_OPT_G_STAT,
},
{
.name = "disable_bw_measurement",
.help = "Disable bandwidth logging",
.parent = "gtod_reduce",
.def = "0",
+ .category = FIO_OPT_G_OS | FIO_OPT_G_MISC | FIO_OPT_G_STAT,
},
{
.name = "gtod_cpu",
.cb = str_gtod_cpu_cb,
.help = "Set up dedicated gettimeofday() thread on this CPU",
.verify = gtod_cpu_verify,
+ .category = FIO_OPT_G_OS | FIO_OPT_G_MISC | FIO_OPT_G_STAT,
},
{
.name = "continue_on_error",
.off1 = td_var_offset(continue_on_error),
.help = "Continue on non-fatal errors during IO",
.def = "none",
+ .category = FIO_OPT_G_MISC | FIO_OPT_G_ERR,
.posval = {
{ .ival = "none",
.oval = ERROR_TYPE_NONE,
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(profile),
.help = "Select a specific builtin performance test",
+ .category = FIO_OPT_G_MISC | FIO_OPT_G_JOB,
},
{
.name = "cgroup",
.type = FIO_OPT_STR_STORE,
.off1 = td_var_offset(cgroup),
.help = "Add job to cgroup of this name",
+ .category = FIO_OPT_G_MISC | FIO_OPT_G_OS,
},
{
.name = "cgroup_weight",
.help = "Use given weight for cgroup",
.minval = 100,
.maxval = 1000,
+ .category = FIO_OPT_G_MISC | FIO_OPT_G_OS,
},
{
.name = "cgroup_nodelete",
.off1 = td_var_offset(cgroup_nodelete),
.help = "Do not delete cgroups after job completion",
.def = "0",
+ .category = FIO_OPT_G_MISC | FIO_OPT_G_OS,
},
{
.name = "uid",
.type = FIO_OPT_INT,
.off1 = td_var_offset(uid),
.help = "Run job with this user ID",
+ .category = FIO_OPT_G_OS | FIO_OPT_G_JOB,
},
{
.name = "gid",
.type = FIO_OPT_INT,
.off1 = td_var_offset(gid),
.help = "Run job with this group ID",
+ .category = FIO_OPT_G_OS | FIO_OPT_G_JOB,
},
{
.name = "flow_id",
.off1 = td_var_offset(flow_id),
.help = "The flow index ID to use",
.def = "0",
+ .category = FIO_OPT_G_IO,
},
{
.name = "flow",
.help = "Weight for flow control of this job",
.parent = "flow_id",
.def = "0",
+ .category = FIO_OPT_G_IO,
},
{
.name = "flow_watermark",
" with non-zero flow.",
.parent = "flow_id",
.def = "1024",
+ .category = FIO_OPT_G_IO,
},
{
.name = "flow_sleep",
" back by the flow control mechanism",
.parent = "flow_id",
.def = "0",
+ .category = FIO_OPT_G_IO,
},
{
.name = NULL,