if (get_next_rand_offset(td, f, ddir, &b))
return 1;
} else {
- if (f->last_pos >= f->real_file_size)
- return 1;
-
- b = f->last_pos / td->o.min_bs[ddir];
+ if (f->last_pos >= f->real_file_size) {
+ if (!td_random(td) || get_next_rand_offset(td, f, ddir, &b))
+ return 1;
+ } else
+ b = (f->last_pos - f->file_offset) / td->o.min_bs[ddir];
}
io_u->offset = (b * td->o.min_bs[ddir]) + f->file_offset;
buflen = td->o.min_bs[ddir];
else {
r = os_random_long(&td->bsrange_state);
- buflen = (unsigned int) (1 + (double) (td->o.max_bs[ddir] - 1) * r / (RAND_MAX + 1.0));
+ if (!td->o.bssplit_nr)
+ buflen = (unsigned int) (1 + (double) (td->o.max_bs[ddir] - 1) * r / (RAND_MAX + 1.0));
+ else {
+ long perc = 0;
+ unsigned int i;
+
+ for (i = 0; i < td->o.bssplit_nr; i++) {
+ struct bssplit *bsp = &td->o.bssplit[i];
+
+ buflen = bsp->bs;
+ perc += bsp->perc;
+ if (r <= ((LONG_MAX / 100L) * perc))
+ break;
+ }
+ }
if (!td->o.bs_unaligned)
buflen = (buflen + td->o.min_bs[ddir] - 1) & ~(td->o.min_bs[ddir] - 1);
}
static int fill_io_u(struct thread_data *td, struct io_u *io_u)
{
+ if (td->io_ops->flags & FIO_NOIO)
+ goto out;
+
/*
* see if it's time to sync
*/
io_u_mark_latency(td, usec);
if (td_write(td) && idx == DDIR_WRITE &&
+ td->o.do_verify &&
td->o.verify != VERIFY_NONE)
log_io_piece(td, io_u);