projects
/
fio.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
t/btrace2fio: don't dec inflight for requeue
[fio.git]
/
io_u.c
diff --git
a/io_u.c
b/io_u.c
index eac871bfe9d91890e70e9110bd575e9a725b1a4c..e8894d5dca62ff07a9e39174191ae877490793b5 100644
(file)
--- a/
io_u.c
+++ b/
io_u.c
@@
-90,21
+90,11
@@
static int __get_next_rand_offset(struct thread_data *td, struct fio_file *f,
return 1;
if (td->o.random_generator == FIO_RAND_GEN_TAUSWORTHE) {
return 1;
if (td->o.random_generator == FIO_RAND_GEN_TAUSWORTHE) {
- uint64_t rmax;
-
- rmax = td->o.use_os_rand ? OS_RAND_MAX : FRAND_MAX;
-
- if (td->o.use_os_rand) {
- rmax = OS_RAND_MAX;
- r = os_random_long(&td->random_state);
- } else {
- rmax = FRAND_MAX;
- r = __rand(&td->__random_state);
- }
+ r = __rand(&td->random_state);
dprint(FD_RANDOM, "off rand %llu\n", (unsigned long long) r);
dprint(FD_RANDOM, "off rand %llu\n", (unsigned long long) r);
- *b = lastb * (r / ((uint64_t)
rmax
+ 1.0));
+ *b = lastb * (r / ((uint64_t)
FRAND_MAX
+ 1.0));
} else {
uint64_t off = 0;
} else {
uint64_t off = 0;
@@
-200,13
+190,8
@@
static int should_do_random(struct thread_data *td, enum fio_ddir ddir)
if (td->o.perc_rand[ddir] == 100)
return 1;
if (td->o.perc_rand[ddir] == 100)
return 1;
- if (td->o.use_os_rand) {
- r = os_random_long(&td->seq_rand_state[ddir]);
- v = 1 + (int) (100.0 * (r / (OS_RAND_MAX + 1.0)));
- } else {
- r = __rand(&td->__seq_rand_state[ddir]);
- v = 1 + (int) (100.0 * (r / (FRAND_MAX + 1.0)));
- }
+ r = __rand(&td->seq_rand_state[ddir]);
+ v = 1 + (int) (100.0 * (r / (FRAND_MAX + 1.0)));
return v <= td->o.perc_rand[ddir];
}
return v <= td->o.perc_rand[ddir];
}
@@
-221,7
+206,6
@@
static int get_next_rand_offset(struct thread_data *td, struct fio_file *f,
return get_off_from_method(td, f, ddir, b);
if (!flist_empty(&td->next_rand_list)) {
return get_off_from_method(td, f, ddir, b);
if (!flist_empty(&td->next_rand_list)) {
- struct rand_off *r;
fetch:
r = flist_first_entry(&td->next_rand_list, struct rand_off, list);
flist_del(&r->list);
fetch:
r = flist_first_entry(&td->next_rand_list, struct rand_off, list);
flist_del(&r->list);
@@
-271,20
+255,32
@@
static int get_next_rand_block(struct thread_data *td, struct fio_file *f,
static int get_next_seq_offset(struct thread_data *td, struct fio_file *f,
enum fio_ddir ddir, uint64_t *offset)
{
static int get_next_seq_offset(struct thread_data *td, struct fio_file *f,
enum fio_ddir ddir, uint64_t *offset)
{
+ struct thread_options *o = &td->o;
+
assert(ddir_rw(ddir));
assert(ddir_rw(ddir));
- if (f->last_pos >= f->io_size + get_start_offset(td, f) && td->o.time_based)
+ if (f->last_pos >= f->io_size + get_start_offset(td, f) &&
+ o->time_based)
f->last_pos = f->last_pos - f->io_size;
if (f->last_pos < f->real_file_size) {
uint64_t pos;
f->last_pos = f->last_pos - f->io_size;
if (f->last_pos < f->real_file_size) {
uint64_t pos;
- if (f->last_pos == f->file_offset &&
td->o.
ddir_seq_add < 0)
+ if (f->last_pos == f->file_offset &&
o->
ddir_seq_add < 0)
f->last_pos = f->real_file_size;
pos = f->last_pos - f->file_offset;
f->last_pos = f->real_file_size;
pos = f->last_pos - f->file_offset;
- if (pos)
- pos += td->o.ddir_seq_add;
+ if (pos && o->ddir_seq_add) {
+ pos += o->ddir_seq_add;
+
+ /*
+ * If we reach beyond the end of the file
+ * with holed IO, wrap around to the
+ * beginning again.
+ */
+ if (pos >= f->real_file_size)
+ pos = f->file_offset;
+ }
*offset = pos;
return 0;
*offset = pos;
return 0;
@@
-424,7
+420,7
@@
static unsigned int __get_next_buflen(struct thread_data *td, struct io_u *io_u,
int ddir = io_u->ddir;
unsigned int buflen = 0;
unsigned int minbs, maxbs;
int ddir = io_u->ddir;
unsigned int buflen = 0;
unsigned int minbs, maxbs;
- unsigned long r
, rand_max
;
+ unsigned long r;
assert(ddir_rw(ddir));
assert(ddir_rw(ddir));
@@
-443,20
+439,12
@@
static unsigned int __get_next_buflen(struct thread_data *td, struct io_u *io_u,
if (!io_u_fits(td, io_u, minbs))
return 0;
if (!io_u_fits(td, io_u, minbs))
return 0;
- if (td->o.use_os_rand)
- rand_max = OS_RAND_MAX;
- else
- rand_max = FRAND_MAX;
-
do {
do {
- if (td->o.use_os_rand)
- r = os_random_long(&td->bsrange_state);
- else
- r = __rand(&td->__bsrange_state);
+ r = __rand(&td->bsrange_state);
if (!td->o.bssplit_nr[ddir]) {
buflen = 1 + (unsigned int) ((double) maxbs *
if (!td->o.bssplit_nr[ddir]) {
buflen = 1 + (unsigned int) ((double) maxbs *
- (r / (
rand_max
+ 1.0)));
+ (r / (
FRAND_MAX
+ 1.0)));
if (buflen < minbs)
buflen = minbs;
} else {
if (buflen < minbs)
buflen = minbs;
} else {
@@
-468,7
+456,7
@@
static unsigned int __get_next_buflen(struct thread_data *td, struct io_u *io_u,
buflen = bsp->bs;
perc += bsp->perc;
buflen = bsp->bs;
perc += bsp->perc;
- if ((r <= ((
rand_max
/ 100L) * perc)) &&
+ if ((r <= ((
FRAND_MAX
/ 100L) * perc)) &&
io_u_fits(td, io_u, buflen))
break;
}
io_u_fits(td, io_u, buflen))
break;
}
@@
-517,13
+505,8
@@
static inline enum fio_ddir get_rand_ddir(struct thread_data *td)
unsigned int v;
unsigned long r;
unsigned int v;
unsigned long r;
- if (td->o.use_os_rand) {
- r = os_random_long(&td->rwmix_state);
- v = 1 + (int) (100.0 * (r / (OS_RAND_MAX + 1.0)));
- } else {
- r = __rand(&td->__rwmix_state);
- v = 1 + (int) (100.0 * (r / (FRAND_MAX + 1.0)));
- }
+ r = __rand(&td->rwmix_state);
+ v = 1 + (int) (100.0 * (r / (FRAND_MAX + 1.0)));
if (v <= td->o.rwmix[DDIR_READ])
return DDIR_READ;
if (v <= td->o.rwmix[DDIR_READ])
return DDIR_READ;
@@
-597,8
+580,8
@@
static enum fio_ddir rate_ddir(struct thread_data *td, enum fio_ddir ddir)
if (td_rw(td) && __should_check_rate(td, odir))
td->rate_pending_usleep[odir] -= usec;
if (td_rw(td) && __should_check_rate(td, odir))
td->rate_pending_usleep[odir] -= usec;
- if (ddir
_trim(ddir)
)
- return
ddir
;
+ if (ddir
== DDIR_TRIM
)
+ return
DDIR_TRIM
;
return ddir;
}
return ddir;
}
@@
-748,9
+731,17
@@
static int fill_io_u(struct thread_data *td, struct io_u *io_u)
* See if it's time to switch to a new zone
*/
if (td->zone_bytes >= td->o.zone_size && td->o.zone_skip) {
* See if it's time to switch to a new zone
*/
if (td->zone_bytes >= td->o.zone_size && td->o.zone_skip) {
+ struct fio_file *f = io_u->file;
+
td->zone_bytes = 0;
td->zone_bytes = 0;
- io_u->file->file_offset += td->o.zone_range + td->o.zone_skip;
- io_u->file->last_pos = io_u->file->file_offset;
+ f->file_offset += td->o.zone_range + td->o.zone_skip;
+
+ /*
+ * Wrap from the beginning, if we exceed the file size
+ */
+ if (f->file_offset >= f->real_file_size)
+ f->file_offset = f->real_file_size - f->file_offset;
+ f->last_pos = f->file_offset;
td->io_skip_bytes += td->o.zone_skip;
}
td->io_skip_bytes += td->o.zone_skip;
}
@@
-967,15
+958,9
@@
static struct fio_file *get_next_file_rand(struct thread_data *td,
int opened = 0;
unsigned long r;
int opened = 0;
unsigned long r;
- if (td->o.use_os_rand) {
- r = os_random_long(&td->next_file_state);
- fno = (unsigned int) ((double) td->o.nr_files
- * (r / (OS_RAND_MAX + 1.0)));
- } else {
- r = __rand(&td->__next_file_state);
- fno = (unsigned int) ((double) td->o.nr_files
+ r = __rand(&td->next_file_state);
+ fno = (unsigned int) ((double) td->o.nr_files
* (r / (FRAND_MAX + 1.0)));
* (r / (FRAND_MAX + 1.0)));
- }
f = td->files[fno];
if (fio_file_done(f))
f = td->files[fno];
if (fio_file_done(f))
@@
-1281,7
+1266,7
@@
void lat_target_check(struct thread_data *td)
* If latency target is enabled, we might be ramping up or down and not
* using the full queue depth available.
*/
* If latency target is enabled, we might be ramping up or down and not
* using the full queue depth available.
*/
-int queue_full(struct thread_data *td)
+int queue_full(
const
struct thread_data *td)
{
const int qempty = io_u_qempty(&td->io_u_freelist);
{
const int qempty = io_u_qempty(&td->io_u_freelist);
@@
-1487,7
+1472,8
@@
struct io_u *get_io_u(struct thread_data *td)
if (io_u->ddir == DDIR_WRITE) {
if (td->flags & TD_F_REFILL_BUFFERS) {
io_u_fill_buffer(td, io_u,
if (io_u->ddir == DDIR_WRITE) {
if (td->flags & TD_F_REFILL_BUFFERS) {
io_u_fill_buffer(td, io_u,
- io_u->xfer_buflen, io_u->xfer_buflen);
+ td->o.min_bs[DDIR_WRITE],
+ io_u->xfer_buflen);
} else if ((td->flags & TD_F_SCRAMBLE_BUFFERS) &&
!(td->flags & TD_F_COMPRESS))
do_scramble = 1;
} else if ((td->flags & TD_F_SCRAMBLE_BUFFERS) &&
!(td->flags & TD_F_COMPRESS))
do_scramble = 1;
@@
-1864,22
+1850,29
@@
void fill_io_buffer(struct thread_data *td, void *buf, unsigned int min_write,
else if (!td->o.zero_buffers) {
unsigned int perc = td->o.compress_percentage;
struct frand_state *rs;
else if (!td->o.zero_buffers) {
unsigned int perc = td->o.compress_percentage;
struct frand_state *rs;
+ unsigned int left = max_bs;
- rs = get_buf_state(td);
+ do {
+ rs = get_buf_state(td);
- if (perc) {
- unsigned int seg = min_write;
+ min_write = min(min_write, left);
- seg = min(min_write, td->o.compress_chunk);
- if (!seg)
- seg = min_write;
+ if (perc) {
+ unsigned int seg = min_write;
- fill_random_buf_percentage(rs, buf, perc, seg,max_bs);
- save_buf_state(td, rs);
- } else {
- fill_random_buf(rs, buf, max_bs);
+ seg = min(min_write, td->o.compress_chunk);
+ if (!seg)
+ seg = min_write;
+
+ fill_random_buf_percentage(rs, buf, perc, seg,
+ min_write);
+ } else
+ fill_random_buf(rs, buf, min_write);
+
+ buf += min_write;
+ left -= min_write;
save_buf_state(td, rs);
save_buf_state(td, rs);
- }
+ }
while (left);
} else
memset(buf, 0, max_bs);
}
} else
memset(buf, 0, max_bs);
}