Rename job files
authorJens Axboe <axboe@kernel.dk>
Thu, 11 Apr 2013 09:54:08 +0000 (11:54 +0200)
committerJens Axboe <axboe@kernel.dk>
Thu, 11 Apr 2013 09:54:08 +0000 (11:54 +0200)
gfio uses a .fio or .ini ending.

Signed-off-by: Jens Axboe <axboe@kernel.dk>
42 files changed:
examples/1mbs_clients [deleted file]
examples/1mbs_clients.fio [new file with mode: 0644]
examples/aio-read [deleted file]
examples/aio-read.fio [new file with mode: 0644]
examples/cpuio [deleted file]
examples/cpuio.fio [new file with mode: 0644]
examples/disk-zone-profile [deleted file]
examples/disk-zone-profile.fio [new file with mode: 0644]
examples/e4defrag [deleted file]
examples/e4defrag.fio [new file with mode: 0644]
examples/e4defrag2 [deleted file]
examples/e4defrag2.fio [new file with mode: 0644]
examples/enospc-pressure [deleted file]
examples/enospc-pressure.fio [new file with mode: 0644]
examples/falloc [deleted file]
examples/falloc.fio [new file with mode: 0644]
examples/flow [deleted file]
examples/flow.fio [new file with mode: 0644]
examples/fsx [deleted file]
examples/fsx.fio [new file with mode: 0644]
examples/fusion-aw-sync.fio [new file with mode: 0644]
examples/fusion-aw-sync.ini [deleted file]
examples/iometer-file-access-server [deleted file]
examples/iometer-file-access-server.fio [new file with mode: 0644]
examples/netio [deleted file]
examples/netio.fio [new file with mode: 0644]
examples/null [deleted file]
examples/null.fio [new file with mode: 0644]
examples/numa [deleted file]
examples/numa.fio [new file with mode: 0644]
examples/rdmaio-client [deleted file]
examples/rdmaio-client.fio [new file with mode: 0644]
examples/rdmaio-server [deleted file]
examples/rdmaio-server.fio [new file with mode: 0644]
examples/ssd-test [deleted file]
examples/ssd-test.fio [new file with mode: 0644]
examples/surface-scan [deleted file]
examples/surface-scan.fio [new file with mode: 0644]
examples/tiobench-example [deleted file]
examples/tiobench-example.fio [new file with mode: 0644]
examples/zipf [deleted file]
examples/zipf.fio [new file with mode: 0644]

diff --git a/examples/1mbs_clients b/examples/1mbs_clients
deleted file mode 100644 (file)
index 505cd87..0000000
+++ /dev/null
@@ -1,104 +0,0 @@
-; Keep adding 1024kb/s reading clients at 4 seconds
-[global]
-size=32m
-rw=read
-directory=tmp
-rate=1250
-ratemin=1024
-
-[file1]
-startdelay=0
-
-[file2]
-startdelay=4
-
-[file3]
-startdelay=8
-
-[file4]
-startdelay=12
-
-[file5]
-startdelay=16
-
-[file6]
-startdelay=20
-
-[file7]
-startdelay=24
-
-[file8]
-startdelay=28
-
-[file9]
-startdelay=32
-
-[file10]
-startdelay=36
-
-[file11]
-startdelay=40
-
-[file12]
-startdelay=44
-
-[file13]
-startdelay=48
-
-[file14]
-startdelay=52
-
-[file15]
-startdelay=56
-
-[file16]
-startdelay=60
-
-[file17]
-startdelay=64
-
-[file18]
-startdelay=68
-
-[file19]
-startdelay=72
-
-[file20]
-startdelay=76
-
-[file21]
-startdelay=80
-
-[file22]
-startdelay=84
-
-[file23]
-startdelay=88
-
-[file24]
-startdelay=92
-
-[file25]
-startdelay=96
-
-[file26]
-startdelay=100
-
-[file27]
-startdelay=104
-
-[file28]
-startdelay=108
-
-[file29]
-startdelay=112
-
-[file30]
-startdelay=116
-
-[file31]
-startdelay=120
-
-[file32]
-startdelay=124
-
diff --git a/examples/1mbs_clients.fio b/examples/1mbs_clients.fio
new file mode 100644 (file)
index 0000000..505cd87
--- /dev/null
@@ -0,0 +1,104 @@
+; Keep adding 1024kb/s reading clients at 4 seconds
+[global]
+size=32m
+rw=read
+directory=tmp
+rate=1250
+ratemin=1024
+
+[file1]
+startdelay=0
+
+[file2]
+startdelay=4
+
+[file3]
+startdelay=8
+
+[file4]
+startdelay=12
+
+[file5]
+startdelay=16
+
+[file6]
+startdelay=20
+
+[file7]
+startdelay=24
+
+[file8]
+startdelay=28
+
+[file9]
+startdelay=32
+
+[file10]
+startdelay=36
+
+[file11]
+startdelay=40
+
+[file12]
+startdelay=44
+
+[file13]
+startdelay=48
+
+[file14]
+startdelay=52
+
+[file15]
+startdelay=56
+
+[file16]
+startdelay=60
+
+[file17]
+startdelay=64
+
+[file18]
+startdelay=68
+
+[file19]
+startdelay=72
+
+[file20]
+startdelay=76
+
+[file21]
+startdelay=80
+
+[file22]
+startdelay=84
+
+[file23]
+startdelay=88
+
+[file24]
+startdelay=92
+
+[file25]
+startdelay=96
+
+[file26]
+startdelay=100
+
+[file27]
+startdelay=104
+
+[file28]
+startdelay=108
+
+[file29]
+startdelay=112
+
+[file30]
+startdelay=116
+
+[file31]
+startdelay=120
+
+[file32]
+startdelay=124
+
diff --git a/examples/aio-read b/examples/aio-read
deleted file mode 100644 (file)
index 173a4e1..0000000
+++ /dev/null
@@ -1,20 +0,0 @@
-; Read 4 files with aio at different depths
-[global]
-ioengine=libaio
-buffered=0
-rw=randread
-bs=128k
-size=512m
-directory=/data1
-
-[file1]
-iodepth=4
-
-[file2]
-iodepth=32
-
-[file3]
-iodepth=8
-
-[file4]
-iodepth=16
diff --git a/examples/aio-read.fio b/examples/aio-read.fio
new file mode 100644 (file)
index 0000000..173a4e1
--- /dev/null
@@ -0,0 +1,20 @@
+; Read 4 files with aio at different depths
+[global]
+ioengine=libaio
+buffered=0
+rw=randread
+bs=128k
+size=512m
+directory=/data1
+
+[file1]
+iodepth=4
+
+[file2]
+iodepth=32
+
+[file3]
+iodepth=8
+
+[file4]
+iodepth=16
diff --git a/examples/cpuio b/examples/cpuio
deleted file mode 100644 (file)
index 577e072..0000000
+++ /dev/null
@@ -1,8 +0,0 @@
-[global]
-ioengine=cpuio
-time_based
-runtime=10
-
-[burn50percent]
-cpuload=50
-
diff --git a/examples/cpuio.fio b/examples/cpuio.fio
new file mode 100644 (file)
index 0000000..577e072
--- /dev/null
@@ -0,0 +1,8 @@
+[global]
+ioengine=cpuio
+time_based
+runtime=10
+
+[burn50percent]
+cpuload=50
+
diff --git a/examples/disk-zone-profile b/examples/disk-zone-profile
deleted file mode 100644 (file)
index 96e5669..0000000
+++ /dev/null
@@ -1,14 +0,0 @@
-; Read disk in zones of 128m/2g, generating a plot of that afterwards
-; should give a nice picture of the zoning of this drive
-
-[global]
-bs=64k
-direct=1
-rw=read
-ioengine=libaio
-iodepth=2
-zonesize=256m
-zoneskip=2g
-write_bw_log
-
-[/dev/sdb]
diff --git a/examples/disk-zone-profile.fio b/examples/disk-zone-profile.fio
new file mode 100644 (file)
index 0000000..96e5669
--- /dev/null
@@ -0,0 +1,14 @@
+; Read disk in zones of 128m/2g, generating a plot of that afterwards
+; should give a nice picture of the zoning of this drive
+
+[global]
+bs=64k
+direct=1
+rw=read
+ioengine=libaio
+iodepth=2
+zonesize=256m
+zoneskip=2g
+write_bw_log
+
+[/dev/sdb]
diff --git a/examples/e4defrag b/examples/e4defrag
deleted file mode 100644 (file)
index cb94e85..0000000
+++ /dev/null
@@ -1,41 +0,0 @@
-[global]
-ioengine=e4defrag
-directory=/scratch
-nrfiles=1
-filesize=100M
-size=100M
-bs=32k
-#group_reporting
-
-[isolated-e4defrag]
-# It is important to disable buffered io
-buffered=0
-donorname=file.def
-filename=file1
-inplace=0
-rw=write
-
-# Run e4defrag and aio-dio workers in parallel
-[e4defrag]
-stonewall
-time_based=30
-runtime=30
-ioengine=e4defrag
-buffered=0
-donorname=file.def
-filename=file1
-inplace=0
-rw=write
-
-[random-aio-32k]
-ioengine=libaio
-runtime=30
-verify=md5
-direct=1
-bs=64k
-iodepth=128
-filename=file1
-rw=randrw
-numjobs=4
-
-
diff --git a/examples/e4defrag.fio b/examples/e4defrag.fio
new file mode 100644 (file)
index 0000000..cb94e85
--- /dev/null
@@ -0,0 +1,41 @@
+[global]
+ioengine=e4defrag
+directory=/scratch
+nrfiles=1
+filesize=100M
+size=100M
+bs=32k
+#group_reporting
+
+[isolated-e4defrag]
+# It is important to disable buffered io
+buffered=0
+donorname=file.def
+filename=file1
+inplace=0
+rw=write
+
+# Run e4defrag and aio-dio workers in parallel
+[e4defrag]
+stonewall
+time_based=30
+runtime=30
+ioengine=e4defrag
+buffered=0
+donorname=file.def
+filename=file1
+inplace=0
+rw=write
+
+[random-aio-32k]
+ioengine=libaio
+runtime=30
+verify=md5
+direct=1
+bs=64k
+iodepth=128
+filename=file1
+rw=randrw
+numjobs=4
+
+
diff --git a/examples/e4defrag2 b/examples/e4defrag2
deleted file mode 100644 (file)
index 818618a..0000000
+++ /dev/null
@@ -1,88 +0,0 @@
-#################################################
-# Hardcode defragmentation patterns
-# Please be carefull, it can trigger kernel panic
-#################################################
-[global]
-ioengine=e4defrag
-group_reporting
-directory=/scratch
-nrfiles=1
-filesize=100M
-size=100M
-donorname=file.def
-bs=32k
-
-###########
-# Run several defragmentation threads for different files, but
-# use shared donor file
-[parallel-e4defrag]
-buffered=0
-inplace=0
-rw=write
-numjobs=4
-
-########
-# Run two defragmentation threads, each thread use another's file
-# as donor file
-
-[e4defrag-1]
-stonewall
-inplace=0
-rw=write
-donorname=e4defrag-2
-
-[e4defrag-2]
-inplace=0
-rw=write
-donorname=e4defrag-1
-
-###########
-# Run random defragment activity 
-[e4defrag-fuzzer-4k]
-stonewall
-inplace=1
-bs=4k
-rw=randwrite
-filename=file
-donorname=file.def
-
-########
-# Run random e4defrag and various aio workers in parallel
-[e4defrag-fuzzer-4k]
-stonewall
-continue_on_error=all
-inplace=1
-bs=4k
-donorname=file3.def
-filename=file3
-time_based=30
-rw=randwrite
-
-[buffered-aio-32k]
-continue_on_error=none
-verify=md5
-buffered=1
-ioengine=libaio
-iodepth=128
-bs=32k
-filename=file3
-rw=randrw
-runtime=30
-time_based=30
-numjobs=4
-
-[direct-aio-32k]
-continue_on_error=none
-verify=md5
-buffered=0
-direct=1
-ioengine=libaio
-iodepth=128
-bs=32k
-filename=file3
-rw=randrw
-runtime=30
-time_based=30
-numjobs=4
-
-
diff --git a/examples/e4defrag2.fio b/examples/e4defrag2.fio
new file mode 100644 (file)
index 0000000..818618a
--- /dev/null
@@ -0,0 +1,88 @@
+#################################################
+# Hardcode defragmentation patterns
+# Please be carefull, it can trigger kernel panic
+#################################################
+[global]
+ioengine=e4defrag
+group_reporting
+directory=/scratch
+nrfiles=1
+filesize=100M
+size=100M
+donorname=file.def
+bs=32k
+
+###########
+# Run several defragmentation threads for different files, but
+# use shared donor file
+[parallel-e4defrag]
+buffered=0
+inplace=0
+rw=write
+numjobs=4
+
+########
+# Run two defragmentation threads, each thread use another's file
+# as donor file
+
+[e4defrag-1]
+stonewall
+inplace=0
+rw=write
+donorname=e4defrag-2
+
+[e4defrag-2]
+inplace=0
+rw=write
+donorname=e4defrag-1
+
+###########
+# Run random defragment activity 
+[e4defrag-fuzzer-4k]
+stonewall
+inplace=1
+bs=4k
+rw=randwrite
+filename=file
+donorname=file.def
+
+########
+# Run random e4defrag and various aio workers in parallel
+[e4defrag-fuzzer-4k]
+stonewall
+continue_on_error=all
+inplace=1
+bs=4k
+donorname=file3.def
+filename=file3
+time_based=30
+rw=randwrite
+
+[buffered-aio-32k]
+continue_on_error=none
+verify=md5
+buffered=1
+ioengine=libaio
+iodepth=128
+bs=32k
+filename=file3
+rw=randrw
+runtime=30
+time_based=30
+numjobs=4
+
+[direct-aio-32k]
+continue_on_error=none
+verify=md5
+buffered=0
+direct=1
+ioengine=libaio
+iodepth=128
+bs=32k
+filename=file3
+rw=randrw
+runtime=30
+time_based=30
+numjobs=4
+
+
diff --git a/examples/enospc-pressure b/examples/enospc-pressure
deleted file mode 100644 (file)
index ca9d8f7..0000000
+++ /dev/null
@@ -1,51 +0,0 @@
-#
-# Test for race-condition DIO-write vs punch_hole
-# If race exist dio may rewrite punched block after
-# it was allocated to another file, we will catch that
-# by verifying blocks content
-#
-[global]
-ioengine=libaio 
-directory=/scratch
-# File size is reasonably huge to provoke ENOSPC
-filesize=128G
-size=999G
-iodepth=128
-
-# Expect write failure due to ENOSPC, skip error dump
-continue_on_error=write
-ignore_error=,ENOSPC
-error_dump=0
-fallocate=none
-exitall
-
-# Two threads (dio and punch_hole) operate on single file:'raicer',
-# We do not care about data content here
-[dio-raicer]
-bs=128k 
-direct=1
-buffered=0 
-rw=randwrite
-runtime=100
-filename=raicer
-time_based
-
-[punch_hole-raicer]
-bs=4k
-rw=randtrim
-filename=raicer
-
-# Verifier thread continiously write to newly allcated blocks
-# and veryfy written content
-[aio-dio-verifier]
-create_on_open=1
-verify=crc32c-intel
-verify_fatal=1
-verify_dump=1
-verify_backlog=1024
-verify_async=4
-direct=1
-# block size should be equals to fs block size to prevent short writes
-bs=4k
-rw=randrw
-filename=aio-dio-verifier
diff --git a/examples/enospc-pressure.fio b/examples/enospc-pressure.fio
new file mode 100644 (file)
index 0000000..ca9d8f7
--- /dev/null
@@ -0,0 +1,51 @@
+#
+# Test for race-condition DIO-write vs punch_hole
+# If race exist dio may rewrite punched block after
+# it was allocated to another file, we will catch that
+# by verifying blocks content
+#
+[global]
+ioengine=libaio 
+directory=/scratch
+# File size is reasonably huge to provoke ENOSPC
+filesize=128G
+size=999G
+iodepth=128
+
+# Expect write failure due to ENOSPC, skip error dump
+continue_on_error=write
+ignore_error=,ENOSPC
+error_dump=0
+fallocate=none
+exitall
+
+# Two threads (dio and punch_hole) operate on single file:'raicer',
+# We do not care about data content here
+[dio-raicer]
+bs=128k 
+direct=1
+buffered=0 
+rw=randwrite
+runtime=100
+filename=raicer
+time_based
+
+[punch_hole-raicer]
+bs=4k
+rw=randtrim
+filename=raicer
+
+# Verifier thread continiously write to newly allcated blocks
+# and veryfy written content
+[aio-dio-verifier]
+create_on_open=1
+verify=crc32c-intel
+verify_fatal=1
+verify_dump=1
+verify_backlog=1024
+verify_async=4
+direct=1
+# block size should be equals to fs block size to prevent short writes
+bs=4k
+rw=randrw
+filename=aio-dio-verifier
diff --git a/examples/falloc b/examples/falloc
deleted file mode 100644 (file)
index fa30731..0000000
+++ /dev/null
@@ -1,54 +0,0 @@
-[global]
-ioengine=falloc
-iodepth=1
-direct=0
-buffered=0
-directory=/scratch
-nrfiles=1
-size=100M
-filesize=100M
-group_reporting
-
-
-# Run falloc and punch_hole threads in parallel
-# After activity file will be highly fragmented
-[falloc-fuzzer]
-stonewall
-runtime=10
-time_based=10
-bssplit=4k/10:64k/50:32k/40
-rw=randwrite
-numjobs=1
-filename=fragmented_file
-
-[punch hole-fuzzer]
-bs=4k
-runtime=10
-time_based=10
-rw=randtrim
-numjobs=2
-filename=fragmented_file
-
-## Mesure IO performance on fragmented file
-[sequential aio-dio write]
-stonewall
-ioengine=libaio
-numjobs=1
-iodepth=128
-buffered=0
-direct=1
-rw=write
-bs=64k
-filename=fragmented_file
-
-[sequential buffered read]
-stonewall
-ioengine=sync
-numjobs=1
-iodepth=1
-buffered=1
-direct=0
-rw=read
-bs=64k
-filename=fragmented_file
-
diff --git a/examples/falloc.fio b/examples/falloc.fio
new file mode 100644 (file)
index 0000000..fa30731
--- /dev/null
@@ -0,0 +1,54 @@
+[global]
+ioengine=falloc
+iodepth=1
+direct=0
+buffered=0
+directory=/scratch
+nrfiles=1
+size=100M
+filesize=100M
+group_reporting
+
+
+# Run falloc and punch_hole threads in parallel
+# After activity file will be highly fragmented
+[falloc-fuzzer]
+stonewall
+runtime=10
+time_based=10
+bssplit=4k/10:64k/50:32k/40
+rw=randwrite
+numjobs=1
+filename=fragmented_file
+
+[punch hole-fuzzer]
+bs=4k
+runtime=10
+time_based=10
+rw=randtrim
+numjobs=2
+filename=fragmented_file
+
+## Mesure IO performance on fragmented file
+[sequential aio-dio write]
+stonewall
+ioengine=libaio
+numjobs=1
+iodepth=128
+buffered=0
+direct=1
+rw=write
+bs=64k
+filename=fragmented_file
+
+[sequential buffered read]
+stonewall
+ioengine=sync
+numjobs=1
+iodepth=1
+buffered=1
+direct=0
+rw=read
+bs=64k
+filename=fragmented_file
+
diff --git a/examples/flow b/examples/flow
deleted file mode 100644 (file)
index 4b078cf..0000000
+++ /dev/null
@@ -1,25 +0,0 @@
-# Example usage of flows. The below will have roughly a 1:8 difference
-# between job2 and job1.
-[global]
-norandommap
-thread
-time_based
-runtime=30
-direct=1
-ioengine=libaio
-iodepth=256
-size=100g
-bs=8k
-filename=/tmp/testfile
-flow_watermark=100
-flow_sleep=1000
-
-[job2]
-numjobs=1
-rw=write
-flow=-8
-
-[job1]
-numjobs=1
-rw=randread
-flow=1
diff --git a/examples/flow.fio b/examples/flow.fio
new file mode 100644 (file)
index 0000000..4b078cf
--- /dev/null
@@ -0,0 +1,25 @@
+# Example usage of flows. The below will have roughly a 1:8 difference
+# between job2 and job1.
+[global]
+norandommap
+thread
+time_based
+runtime=30
+direct=1
+ioengine=libaio
+iodepth=256
+size=100g
+bs=8k
+filename=/tmp/testfile
+flow_watermark=100
+flow_sleep=1000
+
+[job2]
+numjobs=1
+rw=write
+flow=-8
+
+[job1]
+numjobs=1
+rw=randread
+flow=1
diff --git a/examples/fsx b/examples/fsx
deleted file mode 100644 (file)
index 6b48c6f..0000000
+++ /dev/null
@@ -1,12 +0,0 @@
-; This job file works pretty works similarly to running fsx-linux
-; with -r 4096 -w 4096 -Z -N 500000
-[file]
-ioengine=libaio
-iodepth=1
-rw=randrw
-size=256k
-bs=4k
-norandommap
-direct=1
-loops=500000
-rwmixcycle=40
diff --git a/examples/fsx.fio b/examples/fsx.fio
new file mode 100644 (file)
index 0000000..6b48c6f
--- /dev/null
@@ -0,0 +1,12 @@
+; This job file works pretty works similarly to running fsx-linux
+; with -r 4096 -w 4096 -Z -N 500000
+[file]
+ioengine=libaio
+iodepth=1
+rw=randrw
+size=256k
+bs=4k
+norandommap
+direct=1
+loops=500000
+rwmixcycle=40
diff --git a/examples/fusion-aw-sync.fio b/examples/fusion-aw-sync.fio
new file mode 100644 (file)
index 0000000..c4639f0
--- /dev/null
@@ -0,0 +1,15 @@
+# Example Job File that randomly writes 8k worth of data atomically for
+# 60 seconds.
+[rw_aw_file_sync]
+rw=randwrite
+ioengine=fusion-aw-sync
+blocksize=8k
+blockalign=8k
+
+filename=/mnt/fs/file
+randrepeat=1
+fallocate=none
+direct=1
+invalidate=0
+runtime=60
+time_based
diff --git a/examples/fusion-aw-sync.ini b/examples/fusion-aw-sync.ini
deleted file mode 100644 (file)
index c4639f0..0000000
+++ /dev/null
@@ -1,15 +0,0 @@
-# Example Job File that randomly writes 8k worth of data atomically for
-# 60 seconds.
-[rw_aw_file_sync]
-rw=randwrite
-ioengine=fusion-aw-sync
-blocksize=8k
-blockalign=8k
-
-filename=/mnt/fs/file
-randrepeat=1
-fallocate=none
-direct=1
-invalidate=0
-runtime=60
-time_based
diff --git a/examples/iometer-file-access-server b/examples/iometer-file-access-server
deleted file mode 100644 (file)
index 291bace..0000000
+++ /dev/null
@@ -1,18 +0,0 @@
-# This job file tries to mimic the Intel IOMeter File Server Access Pattern
-[global]
-description=Emulation of Intel IOmeter File Server Access Pattern
-
-[iometer]
-bssplit=512/10:1k/5:2k/5:4k/60:8k/2:16k/4:32k/4:64k/10
-rw=randrw
-rwmixread=80
-direct=1
-size=4g
-ioengine=libaio
-# IOMeter defines the server loads as the following:
-# iodepth=1    Linear
-# iodepth=4    Very Light
-# iodepth=8    Light
-# iodepth=64   Moderate
-# iodepth=256  Heavy
-iodepth=64
diff --git a/examples/iometer-file-access-server.fio b/examples/iometer-file-access-server.fio
new file mode 100644 (file)
index 0000000..291bace
--- /dev/null
@@ -0,0 +1,18 @@
+# This job file tries to mimic the Intel IOMeter File Server Access Pattern
+[global]
+description=Emulation of Intel IOmeter File Server Access Pattern
+
+[iometer]
+bssplit=512/10:1k/5:2k/5:4k/60:8k/2:16k/4:32k/4:64k/10
+rw=randrw
+rwmixread=80
+direct=1
+size=4g
+ioengine=libaio
+# IOMeter defines the server loads as the following:
+# iodepth=1    Linear
+# iodepth=4    Very Light
+# iodepth=8    Light
+# iodepth=64   Moderate
+# iodepth=256  Heavy
+iodepth=64
diff --git a/examples/netio b/examples/netio
deleted file mode 100644 (file)
index 3b1a7cd..0000000
+++ /dev/null
@@ -1,20 +0,0 @@
-# Example network job, just defines two clients that send/recv data
-[global]
-ioengine=net
-#Use hostname=/tmp.fio.sock for local unix domain sockets
-port=8888
-#Use =udp for UDP, =unix for local unix domain socket
-protocol=tcp
-bs=4k
-size=10g
-#set the below option to enable end-to-end data integrity tests
-#verify=md5
-
-[receiver]
-listen
-rw=read
-
-[sender]
-hostname=localhost
-startdelay=1
-rw=write
diff --git a/examples/netio.fio b/examples/netio.fio
new file mode 100644 (file)
index 0000000..3b1a7cd
--- /dev/null
@@ -0,0 +1,20 @@
+# Example network job, just defines two clients that send/recv data
+[global]
+ioengine=net
+#Use hostname=/tmp.fio.sock for local unix domain sockets
+port=8888
+#Use =udp for UDP, =unix for local unix domain socket
+protocol=tcp
+bs=4k
+size=10g
+#set the below option to enable end-to-end data integrity tests
+#verify=md5
+
+[receiver]
+listen
+rw=read
+
+[sender]
+hostname=localhost
+startdelay=1
+rw=write
diff --git a/examples/null b/examples/null
deleted file mode 100644 (file)
index 9d2f3e0..0000000
+++ /dev/null
@@ -1,10 +0,0 @@
-[global]
-bs=4k
-gtod_reduce=1
-
-[null]
-ioengine=null
-size=100g
-rw=randread
-norandommap
-time_based=0
diff --git a/examples/null.fio b/examples/null.fio
new file mode 100644 (file)
index 0000000..9d2f3e0
--- /dev/null
@@ -0,0 +1,10 @@
+[global]
+bs=4k
+gtod_reduce=1
+
+[null]
+ioengine=null
+size=100g
+rw=randread
+norandommap
+time_based=0
diff --git a/examples/numa b/examples/numa
deleted file mode 100644 (file)
index b81964f..0000000
+++ /dev/null
@@ -1,21 +0,0 @@
-; setup numa policy for each thread
-; 'numactl --show' to determine the maximum numa nodes
-[global]
-ioengine=libaio
-buffered=0
-rw=randread
-bs=512K
-iodepth=16
-size=512m
-filename=/dev/sdb1
-
-; Fix memory blocks (512K * 16) in numa node 0
-[job1]
-numa_cpu_nodes=0
-numa_mem_policy=bind:0
-
-; Interleave memory blocks (512K * 16) in numa node 0 and 1
-[job2]
-numa_cpu_nodes=0-1
-numa_mem_policy=interleave:0-1
-
diff --git a/examples/numa.fio b/examples/numa.fio
new file mode 100644 (file)
index 0000000..b81964f
--- /dev/null
@@ -0,0 +1,21 @@
+; setup numa policy for each thread
+; 'numactl --show' to determine the maximum numa nodes
+[global]
+ioengine=libaio
+buffered=0
+rw=randread
+bs=512K
+iodepth=16
+size=512m
+filename=/dev/sdb1
+
+; Fix memory blocks (512K * 16) in numa node 0
+[job1]
+numa_cpu_nodes=0
+numa_mem_policy=bind:0
+
+; Interleave memory blocks (512K * 16) in numa node 0 and 1
+[job2]
+numa_cpu_nodes=0-1
+numa_mem_policy=interleave:0-1
+
diff --git a/examples/rdmaio-client b/examples/rdmaio-client
deleted file mode 100644 (file)
index 7c660c9..0000000
+++ /dev/null
@@ -1,11 +0,0 @@
-# Example rdma client job
-[global]
-ioengine=rdma
-filename=[ip_addr]/[port]/[RDMA_WRITE/RDMA_READ/SEND]
-bs=1m
-size=100g
-
-[sender]
-rw=write
-iodepth=1
-iodepth_batch_complete=1
\ No newline at end of file
diff --git a/examples/rdmaio-client.fio b/examples/rdmaio-client.fio
new file mode 100644 (file)
index 0000000..7c660c9
--- /dev/null
@@ -0,0 +1,11 @@
+# Example rdma client job
+[global]
+ioengine=rdma
+filename=[ip_addr]/[port]/[RDMA_WRITE/RDMA_READ/SEND]
+bs=1m
+size=100g
+
+[sender]
+rw=write
+iodepth=1
+iodepth_batch_complete=1
\ No newline at end of file
diff --git a/examples/rdmaio-server b/examples/rdmaio-server
deleted file mode 100644 (file)
index 9348859..0000000
+++ /dev/null
@@ -1,10 +0,0 @@
-# Example rdma server job
-[global]
-ioengine=rdma
-filename=[ip_addr]/[port]
-bs=1m
-size=100g
-
-[receiver]
-rw=read
-iodepth=16
\ No newline at end of file
diff --git a/examples/rdmaio-server.fio b/examples/rdmaio-server.fio
new file mode 100644 (file)
index 0000000..9348859
--- /dev/null
@@ -0,0 +1,10 @@
+# Example rdma server job
+[global]
+ioengine=rdma
+filename=[ip_addr]/[port]
+bs=1m
+size=100g
+
+[receiver]
+rw=read
+iodepth=16
\ No newline at end of file
diff --git a/examples/ssd-test b/examples/ssd-test
deleted file mode 100644 (file)
index c84cf50..0000000
+++ /dev/null
@@ -1,37 +0,0 @@
-# Do some important numbers on SSD drives, to gauge what kind of
-# performance you might get out of them.
-#
-# Sequential read and write speeds are tested, these are expected to be
-# high. Random reads should also be fast, random writes are where crap
-# drives are usually separated from the good drives.
-#
-# This uses a queue depth of 4. New SATA SSD's will support up to 32
-# in flight commands, so it may also be interesting to increase the queue
-# depth and compare. Note that most real-life usage will not see that
-# large of a queue depth, so 4 is more representative of normal use.
-#
-[global]
-bs=4k
-ioengine=libaio
-iodepth=4
-size=1g
-direct=1
-runtime=60
-directory=/mount-point-of-ssd
-filename=ssd.test.file
-
-[seq-read]
-rw=read
-stonewall
-
-[rand-read]
-rw=randread
-stonewall
-
-[seq-write]
-rw=write
-stonewall
-
-[rand-write]
-rw=randwrite
-stonewall
diff --git a/examples/ssd-test.fio b/examples/ssd-test.fio
new file mode 100644 (file)
index 0000000..c84cf50
--- /dev/null
@@ -0,0 +1,37 @@
+# Do some important numbers on SSD drives, to gauge what kind of
+# performance you might get out of them.
+#
+# Sequential read and write speeds are tested, these are expected to be
+# high. Random reads should also be fast, random writes are where crap
+# drives are usually separated from the good drives.
+#
+# This uses a queue depth of 4. New SATA SSD's will support up to 32
+# in flight commands, so it may also be interesting to increase the queue
+# depth and compare. Note that most real-life usage will not see that
+# large of a queue depth, so 4 is more representative of normal use.
+#
+[global]
+bs=4k
+ioengine=libaio
+iodepth=4
+size=1g
+direct=1
+runtime=60
+directory=/mount-point-of-ssd
+filename=ssd.test.file
+
+[seq-read]
+rw=read
+stonewall
+
+[rand-read]
+rw=randread
+stonewall
+
+[seq-write]
+rw=write
+stonewall
+
+[rand-write]
+rw=randwrite
+stonewall
diff --git a/examples/surface-scan b/examples/surface-scan
deleted file mode 100644 (file)
index dc3373a..0000000
+++ /dev/null
@@ -1,23 +0,0 @@
-; writes 512 byte verification blocks until the disk is full,
-; then verifies written data
-[global]
-thread=1
-bs=64k
-direct=1
-ioengine=sync
-verify=meta
-verify_pattern=0xaa555aa5
-verify_interval=512
-
-[write-phase]
-filename=datafile.tmp  ; or use a full disk, for example /dev/sda
-rw=write
-fill_device=1
-do_verify=0
-
-[verify-phase]
-stonewall
-create_serialize=0
-filename=datafile.tmp
-rw=read
-do_verify=1
diff --git a/examples/surface-scan.fio b/examples/surface-scan.fio
new file mode 100644 (file)
index 0000000..dc3373a
--- /dev/null
@@ -0,0 +1,23 @@
+; writes 512 byte verification blocks until the disk is full,
+; then verifies written data
+[global]
+thread=1
+bs=64k
+direct=1
+ioengine=sync
+verify=meta
+verify_pattern=0xaa555aa5
+verify_interval=512
+
+[write-phase]
+filename=datafile.tmp  ; or use a full disk, for example /dev/sda
+rw=write
+fill_device=1
+do_verify=0
+
+[verify-phase]
+stonewall
+create_serialize=0
+filename=datafile.tmp
+rw=read
+do_verify=1
diff --git a/examples/tiobench-example b/examples/tiobench-example
deleted file mode 100644 (file)
index 5a4493e..0000000
+++ /dev/null
@@ -1,24 +0,0 @@
-; tiobench like setup, add more fX files between the stonewalls to
-; create more threads
-
-[global]
-direct=1
-size=512m
-bsrange=4k-4k
-timeout=60
-numjobs=4      ; 4 simultaneous threads for each job
-
-[f1]
-rw=write
-
-[f2]
-stonewall
-rw=randwrite
-
-[f3]
-stonewall
-rw=read
-
-[f4]
-stonewall
-rw=randread
diff --git a/examples/tiobench-example.fio b/examples/tiobench-example.fio
new file mode 100644 (file)
index 0000000..5a4493e
--- /dev/null
@@ -0,0 +1,24 @@
+; tiobench like setup, add more fX files between the stonewalls to
+; create more threads
+
+[global]
+direct=1
+size=512m
+bsrange=4k-4k
+timeout=60
+numjobs=4      ; 4 simultaneous threads for each job
+
+[f1]
+rw=write
+
+[f2]
+stonewall
+rw=randwrite
+
+[f3]
+stonewall
+rw=read
+
+[f4]
+stonewall
+rw=randread
diff --git a/examples/zipf b/examples/zipf
deleted file mode 100644 (file)
index fcfa38d..0000000
+++ /dev/null
@@ -1,10 +0,0 @@
-# Example job file for using a zipf distribution instead
-# of a purely random workload where each block is read
-# or written once.
-[job]
-ioengine=null
-rw=randread
-norandommap
-size=1280m
-bs=4k
-random_distribution=zipf:0.5
diff --git a/examples/zipf.fio b/examples/zipf.fio
new file mode 100644 (file)
index 0000000..fcfa38d
--- /dev/null
@@ -0,0 +1,10 @@
+# Example job file for using a zipf distribution instead
+# of a purely random workload where each block is read
+# or written once.
+[job]
+ioengine=null
+rw=randread
+norandommap
+size=1280m
+bs=4k
+random_distribution=zipf:0.5