*b = 0;
i = 0;
- while ((*b) * td->rw_min_bs < f->file_size) {
+ while ((*b) * td->rw_min_bs < f->real_file_size) {
if (f->file_map[i] != -1UL) {
*b += ffz(f->file_map[i]);
return 0;
b = f->last_pos / td->min_bs[ddir];
*offset = (b * td->min_bs[ddir]) + f->file_offset;
- if (*offset > f->file_size)
+ if (*offset > f->real_file_size)
return 1;
return 0;
buflen = td->min_bs[ddir];
else {
r = os_random_long(&td->bsrange_state);
- buflen = (1 + (double) (td->max_bs[ddir] - 1) * r / (RAND_MAX + 1.0));
+ buflen = (unsigned int) (1 + (double) (td->max_bs[ddir] - 1) * r / (RAND_MAX + 1.0));
if (!td->bs_unaligned)
buflen = (buflen + td->min_bs[ddir] - 1) & ~(td->min_bs[ddir] - 1);
}
* mixed read/write workload, check the rwmix cycle and switch if
* necessary.
*/
-static int get_rw_ddir(struct thread_data *td)
+static enum fio_ddir get_rw_ddir(struct thread_data *td)
{
if (td_rw(td)) {
struct timeval now;
return NULL;
}
- if (io_u->buflen + io_u->offset > f->file_size) {
+ if (io_u->buflen + io_u->offset > f->real_file_size) {
if (td->io_ops->flags & FIO_RAWIO) {
put_io_u(td, io_u);
return NULL;
}
- io_u->buflen = f->file_size - io_u->offset;
+ io_u->buflen = f->real_file_size - io_u->offset;
}
if (io_u->ddir != DDIR_SYNC) {
if (!io_u->error) {
unsigned int bytes = io_u->buflen - io_u->resid;
- const int idx = io_u->ddir;
+ const enum fio_ddir idx = io_u->ddir;
td->io_blocks[idx]++;
td->io_bytes[idx] += bytes;