* Map of job/command line options
*/
static struct fio_option options[] = {
+ {
+ .name = "description",
+ .type = FIO_OPT_STR_STORE,
+ .off1 = td_var_offset(description),
+ .help = "Text job description",
+ },
{
.name = "name",
.type = FIO_OPT_STR_STORE,
.help = "IO engine to use",
.def = "sync",
.posval = { "sync", "libaio", "posixaio", "mmap", "splice",
- "sg", "null", },
+ "sg", "null", "net", "syslet-rw" },
+ },
+ {
+ .name = "iodepth",
+ .type = FIO_OPT_INT,
+ .off1 = td_var_offset(iodepth),
+ .help = "Amount of IO buffers to keep in flight",
+ .def = "1",
+ },
+ {
+ .name = "size",
+ .type = FIO_OPT_STR_VAL,
+ .off1 = td_var_offset(total_file_size),
+ .help = "Size of device or file",
+ },
+ {
+ .name = "bs",
+ .type = FIO_OPT_STR_VAL_INT,
+ .off1 = td_var_offset(bs[DDIR_READ]),
+ .off2 = td_var_offset(bs[DDIR_WRITE]),
+ .help = "Block size unit",
+ .def = "4k",
+ },
+ {
+ .name = "bsrange",
+ .type = FIO_OPT_RANGE,
+ .off1 = td_var_offset(min_bs[DDIR_READ]),
+ .off2 = td_var_offset(max_bs[DDIR_READ]),
+ .off3 = td_var_offset(min_bs[DDIR_WRITE]),
+ .off4 = td_var_offset(max_bs[DDIR_WRITE]),
+ .help = "Set block size range (in more detail than bs)",
+ },
+ {
+ .name = "bs_unaligned",
+ .type = FIO_OPT_STR_SET,
+ .off1 = td_var_offset(bs_unaligned),
+ .help = "Don't sector align IO buffer sizes",
+ },
+ {
+ .name = "offset",
+ .type = FIO_OPT_STR_VAL,
+ .off1 = td_var_offset(start_offset),
+ .help = "Start IO from this offset",
+ .def = "0",
+ },
+ {
+ .name = "randrepeat",
+ .type = FIO_OPT_BOOL,
+ .off1 = td_var_offset(rand_repeatable),
+ .help = "Use repeatable random IO pattern",
+ .def = "1",
+ },
+ {
+ .name = "norandommap",
+ .type = FIO_OPT_STR_SET,
+ .off1 = td_var_offset(norandommap),
+ .help = "Accept potential duplicate random blocks",
+ },
+ {
+ .name = "nrfiles",
+ .type = FIO_OPT_INT,
+ .off1 = td_var_offset(nr_files),
+ .help = "Split job workload between this number of files",
+ .def = "1",
+ },
+ {
+ .name = "fsync",
+ .type = FIO_OPT_INT,
+ .off1 = td_var_offset(fsync_blocks),
+ .help = "Issue fsync for writes every given number of blocks",
+ .def = "0",
+ },
+ {
+ .name = "direct",
+ .type = FIO_OPT_BOOL,
+ .off1 = td_var_offset(odirect),
+ .help = "Use O_DIRECT IO (negates buffered)",
+ .def = "0",
+ },
+ {
+ .name = "buffered",
+ .type = FIO_OPT_BOOL,
+ .off1 = td_var_offset(odirect),
+ .neg = 1,
+ .help = "Use buffered IO (negates direct)",
+ .def = "1",
+ },
+ {
+ .name = "overwrite",
+ .type = FIO_OPT_BOOL,
+ .off1 = td_var_offset(overwrite),
+ .help = "When writing, set whether to overwrite current data",
+ .def = "0",
+ },
+ {
+ .name = "loops",
+ .type = FIO_OPT_INT,
+ .off1 = td_var_offset(loops),
+ .help = "Number of times to run the job",
+ .def = "1",
+ },
+ {
+ .name = "numjobs",
+ .type = FIO_OPT_INT,
+ .off1 = td_var_offset(numjobs),
+ .help = "Duplicate this job this many times",
+ .def = "1",
+ },
+ {
+ .name = "startdelay",
+ .type = FIO_OPT_INT,
+ .off1 = td_var_offset(start_delay),
+ .help = "Only start job when this period has passed",
+ .def = "0",
+ },
+ {
+ .name = "runtime",
+ .alias = "timeout",
+ .type = FIO_OPT_STR_VAL_TIME,
+ .off1 = td_var_offset(timeout),
+ .help = "Stop workload when this amount of time has passed",
+ .def = "0",
},
{
.name = "mem",
.help = "Use this IO scheduler on the backing device",
},
#endif
- {
- .name = "size",
- .type = FIO_OPT_STR_VAL,
- .off1 = td_var_offset(total_file_size),
- .help = "Size of device or file",
- },
- {
- .name = "bs",
- .type = FIO_OPT_STR_VAL_INT,
- .off1 = td_var_offset(bs[DDIR_READ]),
- .off2 = td_var_offset(bs[DDIR_WRITE]),
- .help = "Block size unit",
- .def = "4k",
- },
- {
- .name = "offset",
- .type = FIO_OPT_STR_VAL,
- .off1 = td_var_offset(start_offset),
- .help = "Start IO from this offset",
- .def = "0",
- },
{
.name = "zonesize",
.type = FIO_OPT_STR_VAL,
.help = "Lock down this amount of memory",
.def = "0",
},
- {
- .name = "bsrange",
- .type = FIO_OPT_RANGE,
- .off1 = td_var_offset(min_bs[DDIR_READ]),
- .off2 = td_var_offset(max_bs[DDIR_READ]),
- .off3 = td_var_offset(min_bs[DDIR_WRITE]),
- .off4 = td_var_offset(max_bs[DDIR_WRITE]),
- .help = "Set block size range",
- },
- {
- .name = "randrepeat",
- .type = FIO_OPT_INT,
- .off1 = td_var_offset(rand_repeatable),
- .help = "Use repeatable random IO pattern",
- .def = "1",
- },
- {
- .name = "nrfiles",
- .type = FIO_OPT_INT,
- .off1 = td_var_offset(nr_files),
- .help = "Split job workload between this number of files",
- .def = "1",
- },
- {
- .name = "iodepth",
- .type = FIO_OPT_INT,
- .off1 = td_var_offset(iodepth),
- .help = "Amount of IO buffers to keep in flight",
- .def = "1",
- },
- {
- .name = "fsync",
- .type = FIO_OPT_INT,
- .off1 = td_var_offset(fsync_blocks),
- .help = "Issue fsync for writes every given number of blocks",
- .def = "0",
- },
{
.name = "rwmixcycle",
.type = FIO_OPT_INT,
.name = "thinktime",
.type = FIO_OPT_INT,
.off1 = td_var_offset(thinktime),
- .help = "Idle time between IO buffers",
+ .help = "Idle time between IO buffers (usec)",
+ .def = "0",
+ },
+ {
+ .name = "thinktime_spin",
+ .type = FIO_OPT_INT,
+ .off1 = td_var_offset(thinktime_spin),
+ .help = "Start thinktime by spinning this amount (usec)",
.def = "0",
},
{
.help = "Window average for rate limits (msec)",
.def = "1000",
},
- {
- .name = "startdelay",
- .type = FIO_OPT_INT,
- .off1 = td_var_offset(start_delay),
- .help = "Only start job when this period has passed",
- .def = "0",
- },
- {
- .name = "timeout",
- .type = FIO_OPT_STR_VAL_TIME,
- .off1 = td_var_offset(timeout),
- .help = "Stop workload when this amount of time has passed",
- .def = "0",
- },
{
.name = "invalidate",
- .type = FIO_OPT_INT,
+ .type = FIO_OPT_BOOL,
.off1 = td_var_offset(invalidate_cache),
.help = "Invalidate buffer/page cache prior to running job",
.def = "1",
},
{
.name = "sync",
- .type = FIO_OPT_INT,
+ .type = FIO_OPT_BOOL,
.off1 = td_var_offset(sync_io),
.help = "Use O_SYNC for buffered writes",
.def = "0",
},
{
.name = "create_serialize",
- .type = FIO_OPT_INT,
+ .type = FIO_OPT_BOOL,
.off1 = td_var_offset(create_serialize),
.help = "Serialize creating of job files",
.def = "1",
},
{
.name = "create_fsync",
- .type = FIO_OPT_INT,
+ .type = FIO_OPT_BOOL,
.off1 = td_var_offset(create_fsync),
.help = "Fsync file after creation",
.def = "1",
},
- {
- .name = "loops",
- .type = FIO_OPT_INT,
- .off1 = td_var_offset(loops),
- .help = "Number of times to run the job",
- .def = "1",
- },
- {
- .name = "numjobs",
- .type = FIO_OPT_INT,
- .off1 = td_var_offset(numjobs),
- .help = "Duplicate this job this many times",
- .def = "1",
- },
{
.name = "cpuload",
.type = FIO_OPT_INT,
.off1 = td_var_offset(cpucycle),
.help = "Length of the CPU burn cycles",
},
- {
- .name = "direct",
- .type = FIO_OPT_INT,
- .off1 = td_var_offset(odirect),
- .help = "Use O_DIRECT IO",
- .def = "1",
- },
- {
- .name = "overwrite",
- .type = FIO_OPT_INT,
- .off1 = td_var_offset(overwrite),
- .help = "When writing, set whether to overwrite current data",
- .def = "0",
- },
#ifdef FIO_HAVE_CPU_AFFINITY
{
.name = "cpumask",
#endif
{
.name = "end_fsync",
- .type = FIO_OPT_INT,
+ .type = FIO_OPT_BOOL,
.off1 = td_var_offset(end_fsync),
.help = "Include fsync at the end of job",
.def = "0",
},
{
.name = "unlink",
- .type = FIO_OPT_INT,
+ .type = FIO_OPT_BOOL,
.off1 = td_var_offset(unlink),
.help = "Unlink created files after job has completed",
- .def = "1",
+ .def = "0",
},
{
.name = "exitall",
{
.name = "thread",
.type = FIO_OPT_STR_SET,
- .off1 = td_var_offset(thread),
+ .off1 = td_var_offset(use_thread),
.help = "Use threads instead of forks",
},
{
.off1 = td_var_offset(write_lat_log),
.help = "Write log of latency during run",
},
- {
- .name = "norandommap",
- .type = FIO_OPT_STR_SET,
- .off1 = td_var_offset(norandommap),
- .help = "Accept potential duplicate random blocks",
- },
- {
- .name = "bs_unaligned",
- .type = FIO_OPT_STR_SET,
- .off1 = td_var_offset(bs_unaligned),
- .help = "Don't sector align IO buffer sizes",
- },
{
.name = "hugepage-size",
.type = FIO_OPT_STR_VAL,
FILE *f_err = NULL;
static int write_lat_log = 0;
-static int write_bw_log = 0;
+int write_bw_log = 0;
/*
* Return a free job structure.
if (td == &def_thread)
return;
+ if (td->error)
+ fprintf(f_out, "fio: %s\n", td->verror);
+
memset(&threads[td->thread_number - 1], 0, sizeof(*td));
thread_number--;
}
*/
if (td->filetype == FIO_TYPE_CHAR && td->odirect)
td->odirect = 0;
+
+ /*
+ * thinktime_spin must be less than thinktime
+ */
+ if (td->thinktime_spin > td->thinktime)
+ td->thinktime_spin = td->thinktime;
}
/*
static char *to_kmg(unsigned int val)
{
char *buf = malloc(32);
- char post[] = { 0, 'K', 'M', 'G', 'P', 0 };
+ char post[] = { 0, 'K', 'M', 'G', 'P', 'E', 0 };
char *p = post;
do {
td->io_ops->flags |= FIO_RAWIO;
td->filetype = FIO_TYPE_FILE;
- if (!stat(jobname, &sb)) {
+ if (td->filename && !lstat(td->filename, &sb)) {
if (S_ISBLK(sb.st_mode))
td->filetype = FIO_TYPE_BD;
else if (S_ISCHR(sb.st_mode))
char tmp[PATH_MAX];
int len = 0;
- if (td->directory && td->directory[0] != '\0')
+ if (td->directory && td->directory[0] != '\0') {
+ if (lstat(td->directory, &sb) < 0) {
+ log_err("fio: %s is not a directory\n", td->directory);
+ td_verror(td, errno);
+ return 1;
+ }
+ if (!S_ISDIR(sb.st_mode)) {
+ log_err("fio: %s is not a directory\n", td->directory);
+ return 1;
+ }
len = sprintf(tmp, "%s/", td->directory);
+ }
td->files = malloc(sizeof(struct fio_file) * td->nr_files);
c3 = to_kmg(td->min_bs[DDIR_WRITE]);
c4 = to_kmg(td->max_bs[DDIR_WRITE]);
- fprintf(f_out, "%s: (g=%d): rw=%s, odir=%u, bs=%s-%s/%s-%s, rate=%u, ioengine=%s, iodepth=%u\n", td->name, td->groupid, ddir_str[ddir], td->odirect, c1, c2, c3, c4, td->rate, td->io_ops->name, td->iodepth);
+ fprintf(f_out, "%s: (g=%d): rw=%s, bs=%s-%s/%s-%s, ioengine=%s, iodepth=%u\n", td->name, td->groupid, ddir_str[ddir], c1, c2, c3, c4, td->io_ops->name, td->iodepth);
free(c1);
free(c2);
return 0;
if (td->rand_repeatable)
- seeds[3] = FIO_RANDSEED;
+ seeds[3] = FIO_RANDSEED * td->thread_number;
if (!td->norandommap) {
for_each_file(td, f, i) {
- blocks = (f->file_size + td->rw_min_bs - 1) / td->rw_min_bs;
+ blocks = (f->real_file_size + td->rw_min_bs - 1) / td->rw_min_bs;
num_maps = (blocks + BLOCKS_PER_MAP-1)/ BLOCKS_PER_MAP;
f->file_map = malloc(num_maps * sizeof(long));
f->num_maps = num_maps;
for (i = 0; i < strlen(line); i++) {
if (line[i] == ';')
return 1;
+ if (line[i] == '#')
+ return 1;
if (!isspace(line[i]) && !iscntrl(line[i]))
return 0;
}
f_out = stdout;
f_err = stderr;
+ options_init(options);
+
dupe_job_options();
if (setup_thread_area())