unsigned short index;
u32 sequence;
bool has_user;
- bool needs_lock;
+ bool in_async;
bool needs_fixed_file;
};
ret2 = -EAGAIN;
/* Catch -EAGAIN return for forced non-blocking submission */
if (!force_nonblock || ret2 != -EAGAIN) {
- kiocb_done(kiocb, ret2, nxt, s->needs_lock);
+ kiocb_done(kiocb, ret2, nxt, s->in_async);
} else {
- /*
- * If ->needs_lock is true, we're already in async
- * context.
- */
- if (!s->needs_lock)
+ if (!s->in_async)
io_async_list_note(READ, req, iov_count);
ret = -EAGAIN;
}
ret = -EAGAIN;
if (force_nonblock && !(kiocb->ki_flags & IOCB_DIRECT)) {
- /* If ->needs_lock is true, we're already in async context. */
- if (!s->needs_lock)
+ if (!s->in_async)
io_async_list_note(WRITE, req, iov_count);
goto out_free;
}
else
ret2 = loop_rw_iter(WRITE, file, kiocb, &iter);
if (!force_nonblock || ret2 != -EAGAIN) {
- kiocb_done(kiocb, ret2, nxt, s->needs_lock);
+ kiocb_done(kiocb, ret2, nxt, s->in_async);
} else {
- /*
- * If ->needs_lock is true, we're already in async
- * context.
- */
- if (!s->needs_lock)
+ if (!s->in_async)
io_async_list_note(WRITE, req, iov_count);
ret = -EAGAIN;
}
return -EAGAIN;
/* workqueue context doesn't hold uring_lock, grab it now */
- if (s->needs_lock)
+ if (s->in_async)
mutex_lock(&ctx->uring_lock);
io_iopoll_req_issued(req);
- if (s->needs_lock)
+ if (s->in_async)
mutex_unlock(&ctx->uring_lock);
}
if (!ret) {
s->has_user = cur_mm != NULL;
- s->needs_lock = true;
+ s->in_async = true;
do {
ret = __io_submit_sqe(ctx, req, s, &nxt, false);
/*
-EFAULT);
} else {
s.has_user = has_user;
- s.needs_lock = true;
+ s.in_async = true;
s.needs_fixed_file = true;
io_submit_sqe(ctx, &s, statep, &link);
submitted++;
out:
s.has_user = true;
- s.needs_lock = false;
+ s.in_async = false;
s.needs_fixed_file = false;
submit++;
io_submit_sqe(ctx, &s, statep, &link);