Skip to content

Commit

Permalink
io_uring: stash ctx task reference for SQPOLL
Browse files Browse the repository at this point in the history
commit 2aede0e upstream.

We can grab a reference to the task instead of stashing away the task
files_struct. This is doable without creating a circular reference
between the ring fd and the task itself.

Reviewed-by: Pavel Begunkov <asml.silence@gmail.com>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
  • Loading branch information
axboe authored and gregkh committed Nov 1, 2020
1 parent cf38360 commit 89197cd
Showing 1 changed file with 34 additions and 13 deletions.
47 changes: 34 additions & 13 deletions fs/io_uring.c
Expand Up @@ -265,7 +265,16 @@ struct io_ring_ctx {
/* IO offload */
struct io_wq *io_wq;
struct task_struct *sqo_thread; /* if using sq thread polling */
struct mm_struct *sqo_mm;

/*
* For SQPOLL usage - we hold a reference to the parent task, so we
* have access to the ->files
*/
struct task_struct *sqo_task;

/* Only used for accounting purposes */
struct mm_struct *mm_account;

wait_queue_head_t sqo_wait;

/*
Expand Down Expand Up @@ -969,9 +978,10 @@ static int __io_sq_thread_acquire_mm(struct io_ring_ctx *ctx)
{
if (!current->mm) {
if (unlikely(!(ctx->flags & IORING_SETUP_SQPOLL) ||
!mmget_not_zero(ctx->sqo_mm)))
!ctx->sqo_task->mm ||
!mmget_not_zero(ctx->sqo_task->mm)))
return -EFAULT;
kthread_use_mm(ctx->sqo_mm);
kthread_use_mm(ctx->sqo_task->mm);
}

return 0;
Expand Down Expand Up @@ -7591,11 +7601,11 @@ static void io_unaccount_mem(struct io_ring_ctx *ctx, unsigned long nr_pages,
if (ctx->limit_mem)
__io_unaccount_mem(ctx->user, nr_pages);

if (ctx->sqo_mm) {
if (ctx->mm_account) {
if (acct == ACCT_LOCKED)
ctx->sqo_mm->locked_vm -= nr_pages;
ctx->mm_account->locked_vm -= nr_pages;
else if (acct == ACCT_PINNED)
atomic64_sub(nr_pages, &ctx->sqo_mm->pinned_vm);
atomic64_sub(nr_pages, &ctx->mm_account->pinned_vm);
}
}

Expand All @@ -7610,11 +7620,11 @@ static int io_account_mem(struct io_ring_ctx *ctx, unsigned long nr_pages,
return ret;
}

if (ctx->sqo_mm) {
if (ctx->mm_account) {
if (acct == ACCT_LOCKED)
ctx->sqo_mm->locked_vm += nr_pages;
ctx->mm_account->locked_vm += nr_pages;
else if (acct == ACCT_PINNED)
atomic64_add(nr_pages, &ctx->sqo_mm->pinned_vm);
atomic64_add(nr_pages, &ctx->mm_account->pinned_vm);
}

return 0;
Expand Down Expand Up @@ -7918,9 +7928,12 @@ static void io_ring_ctx_free(struct io_ring_ctx *ctx)
{
io_finish_async(ctx);
io_sqe_buffer_unregister(ctx);
if (ctx->sqo_mm) {
mmdrop(ctx->sqo_mm);
ctx->sqo_mm = NULL;

if (ctx->sqo_task) {
put_task_struct(ctx->sqo_task);
ctx->sqo_task = NULL;
mmdrop(ctx->mm_account);
ctx->mm_account = NULL;
}

io_sqe_files_unregister(ctx);
Expand Down Expand Up @@ -8665,8 +8678,16 @@ static int io_uring_create(unsigned entries, struct io_uring_params *p,
ctx->user = user;
ctx->creds = get_current_cred();

ctx->sqo_task = get_task_struct(current);

/*
* This is just grabbed for accounting purposes. When a process exits,
* the mm is exited and dropped before the files, hence we need to hang
* on to this mm purely for the purposes of being able to unaccount
* memory (locked/pinned vm). It's not used for anything else.
*/
mmgrab(current->mm);
ctx->sqo_mm = current->mm;
ctx->mm_account = current->mm;

/*
* Account memory _before_ installing the file descriptor. Once
Expand Down

0 comments on commit 89197cd

Please sign in to comment.