X-Git-Url: https://git.kernel.dk/?p=fio.git;a=blobdiff_plain;f=HOWTO;h=482d69651c0362c3afd5863512dc32f7d73f3733;hp=f8b488faac91285682cb9a9027b9ddaa7ecc804a;hb=a3f001f5f679ae4efc5f34f682b16a2673ec2bac;hpb=774af4a2ced9a268a47527f9f61f6d23ffe8bbe5 diff --git a/HOWTO b/HOWTO index f8b488fa..482d6965 100644 --- a/HOWTO +++ b/HOWTO @@ -767,20 +767,14 @@ ioengine=str Defines how the job issues io to the file. The following defines engine specific options. libhdfs Read and write through Hadoop (HDFS). - The 'filename' option is used to specify host, - port of the hdfs name-node to connect. This - engine interprets offsets a little + This engine interprets offsets a little differently. In HDFS, files once created cannot be modified. So random writes are not possible. To imitate this, libhdfs engine - expects bunch of small files to be created - over HDFS, and engine will randomly pick a - file out of those files based on the offset - generated by fio backend. (see the example - job file to create such files, use rw=write - option). Please note, you might want to set - necessary environment variables to work with - hdfs/libhdfs properly. + creates bunch of small files, and engine will + pick a file out of those files based on the + offset enerated by fio backend. Each jobs uses + it's own connection to HDFS. mtd Read, write and erase an MTD character device (e.g., /dev/mtd0). Discards are treated as @@ -1049,7 +1043,7 @@ rate=int Cap the bandwidth used by this job. The number is in bytes/sec, will only limit writes (to 500KB/sec), the latter will only limit reads. -ratemin=int Tell fio to do whatever it can to maintain at least this +rate_min=int Tell fio to do whatever it can to maintain at least this bandwidth. Failing to meet this requirement, will cause the job to exit. The same format as rate is used for read vs write separation. @@ -1064,11 +1058,14 @@ rate_iops_min=int If fio doesn't meet this rate of IO, it will cause the job to exit. The same format as rate is used for read vs write separation. -rate_poisson=bool Real world random request flow follows Poisson process - (https://en.wikipedia.org/wiki/Poisson_process). With - rate_poisson=1, fio tries to simulate request flow under Poisson - process (instead of even distribution, which is the default). - Default: false. +rate_process=str This option controls how fio manages rated IO + submissions. The default is 'linear', which submits IO in a + linear fashion with fixed delays between IOs that gets + adjusted based on IO completion rates. If this is set to + 'poisson', fio will submit IO based on a more real world + random request flow, known as the Poisson process + (https://en.wikipedia.org/wiki/Poisson_process). The lambda + will be 10^6 / IOPS for the given workload. latency_target=int If set, fio will attempt to find the max performance point that the given workload will run at while maintaining a @@ -1087,7 +1084,7 @@ latency_percentile=float The percentage of IOs that must fall within the max_latency=int If set, fio will exit the job if it exceeds this maximum latency. It will exit with an ETIME error. -ratecycle=int Average bandwidth for 'rate' and 'ratemin' over this number +rate_cycle=int Average bandwidth for 'rate' and 'rate_min' over this number of milliseconds. cpumask=int Set the CPU affinity of this job. The parameter given is a @@ -1224,6 +1221,9 @@ exitall When one job finishes, terminate the rest. The default is to wait for each job to finish, sometimes that is not the desired action. +exitall_on_error When one job finishes in error, terminate the rest. The + default is to wait for each job to finish. + bwavgtime=int Average the calculated bandwidth over the given time. Value is specified in milliseconds. @@ -1577,11 +1577,15 @@ log_compression=int If this is set, fio will compress the IO logs as in the specified log file. This feature depends on the availability of zlib. -log_store_compressed=bool If set, and log_compression is also set, - fio will store the log files in a compressed format. They - can be decompressed with fio, using the --inflate-log - command line parameter. The files will be stored with a - .fz suffix. +log_compression_cpus=str Define the set of CPUs that are allowed to + handle online log compression for the IO jobs. This can + provide better isolation between performance sensitive jobs, + and background compression work. + +log_store_compressed=bool If set, fio will store the log files in a + compressed format. They can be decompressed with fio, using + the --inflate-log command line parameter. The files will be + stored with a .fz suffix. block_error_percentiles=bool If set, record errors in trim block-sized units from writes and trims and output a histogram of @@ -1775,11 +1779,13 @@ that defines them is selected. If the job is a TCP listener or UDP reader, the hostname is not used and must be omitted unless it is a valid UDP multicast address. +[libhdfs] namenode=str The host name or IP address of a HDFS cluster namenode to contact. [netsplice] port=int [net] port=int The TCP or UDP port to bind to or connect to. If this is used with numjobs to spawn multiple instances of the same job type, then this will be the starting port number since fio will use a range of ports. +[libhdfs] port=int the listening port of the HFDS cluster namenode. [netsplice] interface=str [net] interface=str The IP address of the network interface used to send or @@ -1838,6 +1844,9 @@ be the starting port number since fio will use a range of ports. [mtd] skip_bad=bool Skip operations against known bad blocks. +[libhdfs] hdfsdirectory libhdfs will create chunk in this HDFS directory +[libhdfs] chunck_size the size of the chunck to use for each file. + 6.0 Interpreting the output ---------------------------