{
struct posixaio_data *pd = td->io_ops->data;
os_aiocb_t *suspend_list[SUSPEND_ENTRIES];
- struct flist_head *entry;
struct timespec start;
int have_timeout = 0;
- int suspend_entries = 0;
+ int suspend_entries;
+ struct io_u *io_u;
unsigned int r;
+ int i;
if (t && !fill_timespec(&start))
have_timeout = 1;
memset(&start, 0, sizeof(start));
r = 0;
- memset(suspend_list, 0, sizeof(*suspend_list));
restart:
- flist_for_each(entry, &td->io_u_busylist) {
- struct io_u *io_u = flist_entry(entry, struct io_u, list);
+ memset(suspend_list, 0, sizeof(*suspend_list));
+ suspend_entries = 0;
+ io_u_qiter(&td->io_u_all, io_u, i) {
int err;
- if (io_u->seen)
+ if (io_u->seen || !(io_u->flags & IO_U_F_FLIGHT))
continue;
err = aio_error(&io_u->aiocb);
return FIO_Q_COMPLETED;
#endif
}
-
+
if (ret) {
+ int aio_err = aio_error(aiocb);
+
/*
* At least OSX has a very low limit on the number of pending
* IOs, so if it returns EAGAIN, we are out of resources
* to queue more. Just return FIO_Q_BUSY to naturally
* drop off at this depth.
*/
- if (errno == EAGAIN)
+ if (aio_err == EAGAIN)
return FIO_Q_BUSY;
- io_u->error = errno;
+ io_u->error = aio_err;
td_verror(td, io_u->error, "xfer");
return FIO_Q_COMPLETED;
}