summary |
shortlog |
log |
commit | commitdiff |
tree
raw |
patch |
inline | side by side (from parent 1:
691394d)
Really should have been split up, but...
- Check this_io_bytes at the bottom of do_io() so that async engines
have a chance to queue pending IO before deeming this job done.
- dprint() should use log_info(), may not be stdout we want.
- last block bug in get_next_free_block(), if max blocks wasn't a
multiple of the bitmap size.
- Add more debug points.
Signed-off-by: Jens Axboe <jens.axboe@oracle.com>
- /*
- * stop job if we failed doing any IO
- */
- if ((td->this_io_bytes[0] + td->this_io_bytes[1]) == 0)
- td->done = 1;
-
if (td->o.fill_device && td->error == ENOSPC) {
td->error = 0;
td->terminate = 1;
if (td->o.fill_device && td->error == ENOSPC) {
td->error = 0;
td->terminate = 1;
}
} else
cleanup_pending_aio(td);
}
} else
cleanup_pending_aio(td);
+
+ /*
+ * stop job if we failed doing any IO
+ */
+ if ((td->this_io_bytes[0] + td->this_io_bytes[1]) == 0)
+ td->done = 1;
}
static void cleanup_io_u(struct thread_data *td)
}
static void cleanup_io_u(struct thread_data *td)
+ /*
+ * reset file done count if we are to start over
+ */
+ if (td->o.time_based || td->o.loops)
td->nr_done_files = 0;
for_each_file(td, f, i)
td->nr_done_files = 0;
for_each_file(td, f, i)
do { \
if (((type) & fio_debug) == 0) \
break; \
do { \
if (((type) & fio_debug) == 0) \
break; \
+ log_info(str, ##args); \
+static inline void dprint_io_u(struct io_u *io_u, const char *p)
+{
+ struct fio_file *f = io_u->file;
+
+ dprint(FD_IO, "%s: io_u %p: off=%llu/len=%lu/ddir=%d", p, io_u,
+ io_u->offset, io_u->buflen, io_u->ddir);
+ if (f)
+ dprint(FD_IO, "/%s", f->file_name);
+ dprint(FD_IO, "\n");
+}
+
io_u->buflen = blocks * min_bs;
}
io_u->buflen = blocks * min_bs;
}
+static inline unsigned long long last_block(struct thread_data *td,
+ struct fio_file *f)
+{
+ unsigned long long max_blocks;
+
+ max_blocks = f->io_size / td->o.rw_min_bs;
+ if (!max_blocks)
+ return 0;
+
+ return max_blocks - 1;
+}
+
/*
* Return the next free block in the map.
*/
/*
* Return the next free block in the map.
*/
while ((*b) * td->o.rw_min_bs < f->real_file_size) {
if (f->file_map[i] != -1UL) {
*b += fio_ffz(f->file_map[i]);
while ((*b) * td->o.rw_min_bs < f->real_file_size) {
if (f->file_map[i] != -1UL) {
*b += fio_ffz(f->file_map[i]);
+ if (*b > last_block(td, f))
+ break;
f->last_free_lookup = i;
return 0;
}
f->last_free_lookup = i;
return 0;
}
+ dprint(FD_IO, "failed finding a free block\n");
return 1;
}
static int get_next_rand_offset(struct thread_data *td, struct fio_file *f,
return 1;
}
static int get_next_rand_offset(struct thread_data *td, struct fio_file *f,
- int ddir, unsigned long long *b)
- unsigned long long max_blocks = f->io_size / td->o.min_bs[ddir];
unsigned long long r, rb;
int loops = 5;
do {
r = os_random_long(&td->random_state);
unsigned long long r, rb;
int loops = 5;
do {
r = os_random_long(&td->random_state);
- if (!max_blocks)
- *b = 0;
- else
- *b = ((max_blocks - 1) * r / (unsigned long long) (RAND_MAX+1.0));
+ *b = last_block(td, f);
+
/*
* if we are not maintaining a random map, we are done.
*/
/*
* if we are not maintaining a random map, we are done.
*/
static int get_next_offset(struct thread_data *td, struct io_u *io_u)
{
struct fio_file *f = io_u->file;
static int get_next_offset(struct thread_data *td, struct io_u *io_u)
{
struct fio_file *f = io_u->file;
- const int ddir = io_u->ddir;
unsigned long long b;
if (td_random(td) && (td->o.ddir_nr && !--td->ddir_nr)) {
td->ddir_nr = td->o.ddir_nr;
unsigned long long b;
if (td_random(td) && (td->o.ddir_nr && !--td->ddir_nr)) {
td->ddir_nr = td->o.ddir_nr;
- if (get_next_rand_offset(td, f, ddir, &b))
+ if (get_next_rand_offset(td, f, &b))
return 1;
} else {
if (f->last_pos >= f->real_file_size) {
return 1;
} else {
if (f->last_pos >= f->real_file_size) {
- if (!td_random(td) || get_next_rand_offset(td, f, ddir, &b))
+ if (!td_random(td) || get_next_rand_offset(td, f, &b))
- b = (f->last_pos - f->file_offset) / td->o.min_bs[ddir];
+ b = (f->last_pos - f->file_offset) / td->o.rw_min_bs;
- io_u->offset = (b * td->o.min_bs[ddir]) + f->file_offset;
- if (io_u->offset >= f->real_file_size)
+ io_u->offset = (b * td->o.rw_min_bs) + f->file_offset;
+ if (io_u->offset >= f->real_file_size) {
+ dprint(FD_IO, "get_next_offset: offset %llu >= size %llu\n",
+ io_u->offset, f->real_file_size);
* No log, let the seq/rand engine retrieve the next buflen and
* position.
*/
* No log, let the seq/rand engine retrieve the next buflen and
* position.
*/
- if (get_next_offset(td, io_u))
+ if (get_next_offset(td, io_u)) {
+ dprint(FD_IO, "io_u %p, failed getting offset\n", io_u);
io_u->buflen = get_next_buflen(td, io_u);
io_u->buflen = get_next_buflen(td, io_u);
+ if (!io_u->buflen) {
+ dprint(FD_IO, "io_u %p, failed getting buflen\n", io_u);
- if (io_u->offset + io_u->buflen > io_u->file->real_file_size)
+ if (io_u->offset + io_u->buflen > io_u->file->real_file_size) {
+ dprint(FD_IO, "io_u %p, offset too large\n", io_u);
/*
* mark entry before potentially trimming io_u
/*
* mark entry before potentially trimming io_u
* If using a write iolog, store this entry.
*/
out:
* If using a write iolog, store this entry.
*/
out:
+ dprint_io_u(io_u, "fill_io_u");
td->zone_bytes += io_u->buflen;
log_io_u(td, io_u);
return 0;
td->zone_bytes += io_u->buflen;
log_io_u(td, io_u);
return 0;
if (f->flags & FIO_FILE_DONE)
continue;
if (f->flags & FIO_FILE_DONE)
continue;
- if ((!goodf || (f->flags & goodf)) && !(f->flags & badf))
+ if ((!goodf || (f->flags & goodf)) && !(f->flags & badf)) {
+ dprint(FD_FILE, "get_next_file_rand: %p\n", f);
f = NULL;
} while (td->next_file != old_next_file);
f = NULL;
} while (td->next_file != old_next_file);
+ dprint(FD_FILE, "get_next_file_rr: %p\n", f);
assert(td->o.nr_files <= td->files_index);
assert(td->o.nr_files <= td->files_index);
- if (!td->nr_open_files || td->nr_done_files >= td->o.nr_files)
+ if (!td->nr_open_files || td->nr_done_files >= td->o.nr_files) {
+ dprint(FD_FILE, "get_next_file: nr_open=%d, nr_done=%d, nr_files=%d\n", td->nr_open_files, td->nr_done_files, td->o.nr_files);
f = td->file_service_file;
if (f && (f->flags & FIO_FILE_OPEN) && td->file_service_left--)
f = td->file_service_file;
if (f && (f->flags & FIO_FILE_OPEN) && td->file_service_left--)
if (td->o.file_service_type == FIO_FSERVICE_RR)
f = get_next_file_rr(td, FIO_FILE_OPEN, FIO_FILE_CLOSING);
if (td->o.file_service_type == FIO_FSERVICE_RR)
f = get_next_file_rr(td, FIO_FILE_OPEN, FIO_FILE_CLOSING);
td->file_service_file = f;
td->file_service_left = td->file_service_nr - 1;
td->file_service_file = f;
td->file_service_left = td->file_service_nr - 1;
+out:
+ dprint(FD_FILE, "get_next_file: %p\n", f);
struct io_u *io_u;
io_u = __get_io_u(td);
struct io_u *io_u;
io_u = __get_io_u(td);
+ if (!io_u) {
+ dprint(FD_IO, "__get_io_u failed\n");
/*
* from a requeue, io_u already setup
/*
* from a requeue, io_u already setup
if (td->o.read_iolog_file) {
if (read_iolog_get(td, io_u))
goto err_put;
if (td->o.read_iolog_file) {
if (read_iolog_get(td, io_u))
goto err_put;
- } else if (set_io_u_file(td, io_u))
+ } else if (set_io_u_file(td, io_u)) {
+ dprint(FD_IO, "io_u %p, setting file failed\n", io_u);
f = io_u->file;
assert(f->flags & FIO_FILE_OPEN);
if (io_u->ddir != DDIR_SYNC) {
f = io_u->file;
assert(f->flags & FIO_FILE_OPEN);
if (io_u->ddir != DDIR_SYNC) {
+ if (!io_u->buflen) {
+ dprint(FD_IO, "get_io_u: zero buflen on %p\n", io_u);
f->last_pos = io_u->offset + io_u->buflen;
f->last_pos = io_u->offset + io_u->buflen;
+ dprint(FD_IO, "get_io_u failed\n");
put_io_u(td, io_u);
return NULL;
}
put_io_u(td, io_u);
return NULL;
}
+ dprint_io_u(io_u, "io complete");
+
assert(io_u->flags & IO_U_F_FLIGHT);
io_u->flags &= ~IO_U_F_FLIGHT;
assert(io_u->flags & IO_U_F_FLIGHT);
io_u->flags &= ~IO_U_F_FLIGHT;
-static void dprint_io_u(struct io_u *io_u, const char *p)
-{
- struct fio_file *f = io_u->file;
-
- dprint(FD_IO, "%s: off=%llu/len=%lu/ddir=%d", p, io_u->offset,
- io_u->buflen, io_u->ddir);
- if (f)
- dprint(FD_IO, "/%s", f->file_name);
- dprint(FD_IO, "\n");
-}
-
int td_io_prep(struct thread_data *td, struct io_u *io_u)
{
dprint_io_u(io_u, "prep");
fio_ro_check(td, io_u);
int td_io_prep(struct thread_data *td, struct io_u *io_u)
{
dprint_io_u(io_u, "prep");
fio_ro_check(td, io_u);
- if (td->io_ops->prep)
- return td->io_ops->prep(td, io_u);
+ if (td->io_ops->prep) {
+ int ret = td->io_ops->prep(td, io_u);
+
+ dprint(FD_IO, "->prep(%p)=%d\n", io_u, ret);
+ return ret;
+ }