}
seq_printf(m, "UserBufs:\t%u\n", ctx->buf_table.nr);
for (i = 0; has_lock && i < ctx->buf_table.nr; i++) {
- struct io_mapped_ubuf *buf = ctx->buf_table.nodes[i]->buf;
+ struct io_mapped_ubuf *buf = NULL;
- seq_printf(m, "%5u: 0x%llx/%u\n", i, buf->ubuf, buf->len);
+ if (ctx->buf_table.nodes[i])
+ buf = ctx->buf_table.nodes[i]->buf;
+ if (buf)
+ seq_printf(m, "%5u: 0x%llx/%u\n", i, buf->ubuf, buf->len);
+ else
+ seq_printf(m, "%5u: <none>\n", i);
}
if (has_lock && !xa_empty(&ctx->personalities)) {
unsigned long index;
static void io_preinit_req(struct io_kiocb *req, struct io_ring_ctx *ctx)
{
req->ctx = ctx;
- req->rsrc_nodes[IORING_RSRC_FILE] = rsrc_empty_node;
- req->rsrc_nodes[IORING_RSRC_BUFFER] = rsrc_empty_node;
+ req->rsrc_nodes[IORING_RSRC_FILE] = NULL;
+ req->rsrc_nodes[IORING_RSRC_BUFFER] = NULL;
req->link = NULL;
req->async_data = NULL;
/* not necessary, but safer to zero */
notif->file = NULL;
notif->task = current;
io_get_task_refs(1);
- notif->rsrc_nodes[IORING_RSRC_FILE] = rsrc_empty_node;
- notif->rsrc_nodes[IORING_RSRC_BUFFER] = rsrc_empty_node;
+ notif->rsrc_nodes[IORING_RSRC_FILE] = NULL;
+ notif->rsrc_nodes[IORING_RSRC_BUFFER] = NULL;
nd = io_notif_to_data(notif);
nd->zc_report = false;
#define IORING_MAX_FIXED_FILES (1U << 20)
#define IORING_MAX_REG_BUFFERS (1U << 14)
-static const struct io_mapped_ubuf dummy_ubuf = {
- /* set invalid range, so io_import_fixed() fails meeting it */
- .ubuf = -1UL,
- .len = UINT_MAX,
-};
-
-const struct io_rsrc_node empty_node = {
- .type = IORING_RSRC_BUFFER,
- .buf = (struct io_mapped_ubuf *) &dummy_ubuf,
-};
-
int __io_account_mem(struct user_struct *user, unsigned long nr_pages)
{
unsigned long page_limit, cur_pages, new_pages;
{
unsigned int i;
- if (node->buf != &dummy_ubuf) {
+ if (node->buf) {
struct io_mapped_ubuf *imu = node->buf;
if (!refcount_dec_and_test(&imu->refs))
err = io_buffer_validate(iov);
if (err)
break;
- if (!iov->iov_base && tag) {
- err = -EINVAL;
- break;
- }
node = io_sqe_buffer_register(ctx, iov, &last_hpage);
if (IS_ERR(node)) {
err = PTR_ERR(node);
break;
}
+ if (tag) {
+ if (!node) {
+ err = -EINVAL;
+ break;
+ }
+ node->tag = tag;
+ }
i = array_index_nospec(up->offset + done, ctx->buf_table.nr);
io_reset_rsrc_node(&ctx->buf_table, i);
ctx->buf_table.nodes[i] = node;
- if (tag)
- node->tag = tag;
if (ctx->compat)
user_data += sizeof(struct compat_iovec);
else
/* check previously registered pages */
for (i = 0; i < ctx->buf_table.nr; i++) {
struct io_rsrc_node *node = ctx->buf_table.nodes[i];
- struct io_mapped_ubuf *imu = node->buf;
+ struct io_mapped_ubuf *imu;
+ if (!node)
+ continue;
+ imu = node->buf;
for (j = 0; j < imu->nr_bvecs; j++) {
if (!PageCompound(imu->bvec[j].bv_page))
continue;
bool coalesced;
if (!iov->iov_base)
- return rsrc_empty_node;
+ return NULL;
node = io_rsrc_node_alloc(ctx, IORING_RSRC_BUFFER);
if (!node)
ret = -EFAULT;
break;
}
- if (tag && !iov->iov_base) {
- ret = -EINVAL;
- break;
- }
}
node = io_sqe_buffer_register(ctx, iov, &last_hpage);
ret = PTR_ERR(node);
break;
}
- if (tag)
+ if (tag) {
+ if (!node) {
+ ret = -EINVAL;
+ break;
+ }
node->tag = tag;
+ }
data.nodes[i] = node;
}
struct io_rsrc_node *dst_node, *src_node;
src_node = io_rsrc_node_lookup(&src_ctx->buf_table, i);
- if (src_node == rsrc_empty_node) {
- dst_node = rsrc_empty_node;
+ if (!src_node) {
+ dst_node = NULL;
} else {
dst_node = io_rsrc_node_alloc(ctx, IORING_RSRC_BUFFER);
if (!dst_node) {
int io_register_rsrc(struct io_ring_ctx *ctx, void __user *arg,
unsigned int size, unsigned int type);
-extern const struct io_rsrc_node empty_node;
-#define rsrc_empty_node (struct io_rsrc_node *) &empty_node
-
static inline struct io_rsrc_node *io_rsrc_node_lookup(struct io_rsrc_data *data,
int index)
{
static inline void io_put_rsrc_node(struct io_rsrc_node *node)
{
- if (node != rsrc_empty_node && !--node->refs)
+ if (node && !--node->refs)
io_free_rsrc_node(node);
}
static inline void io_req_put_rsrc_nodes(struct io_kiocb *req)
{
- if (req->rsrc_nodes[IORING_RSRC_FILE] != rsrc_empty_node) {
- io_put_rsrc_node(req->rsrc_nodes[IORING_RSRC_FILE]);
- req->rsrc_nodes[IORING_RSRC_FILE] = rsrc_empty_node;
- }
- if (req->rsrc_nodes[IORING_RSRC_BUFFER] != rsrc_empty_node) {
- io_put_rsrc_node(req->rsrc_nodes[IORING_RSRC_BUFFER]);
- req->rsrc_nodes[IORING_RSRC_BUFFER] = rsrc_empty_node;
- }
+ io_put_rsrc_node(req->rsrc_nodes[IORING_RSRC_FILE]);
+ io_put_rsrc_node(req->rsrc_nodes[IORING_RSRC_BUFFER]);
+ req->rsrc_nodes[IORING_RSRC_FILE] = NULL;
+ req->rsrc_nodes[IORING_RSRC_BUFFER] = NULL;
}
static inline void io_req_assign_rsrc_node(struct io_kiocb *req,
struct io_rsrc_node *node)
{
- if (node != rsrc_empty_node) {
- node->refs++;
- req->rsrc_nodes[node->type] = node;
- }
+ node->refs++;
+ req->rsrc_nodes[node->type] = node;
}
int io_files_update(struct io_kiocb *req, unsigned int issue_flags);
if (unlikely(sp->flags & ~valid_flags))
return -EINVAL;
sp->splice_fd_in = READ_ONCE(sqe->splice_fd_in);
- sp->rsrc_node = rsrc_empty_node;
+ sp->rsrc_node = NULL;
req->flags |= REQ_F_FORCE_ASYNC;
return 0;
}