static void io_rsrc_put_work(struct work_struct *work);
 
 static void io_req_task_queue(struct io_kiocb *req);
-static void io_submit_flush_completions(struct io_comp_state *cs,
-                                       struct io_ring_ctx *ctx);
+static void io_submit_flush_completions(struct io_ring_ctx *ctx);
 static bool io_poll_remove_waitqs(struct io_kiocb *req);
 static int io_req_prep_async(struct io_kiocb *req);
 
                return;
        if (ctx->submit_state.comp.nr) {
                mutex_lock(&ctx->uring_lock);
-               io_submit_flush_completions(&ctx->submit_state.comp, ctx);
+               io_submit_flush_completions(ctx);
                mutex_unlock(&ctx->uring_lock);
        }
        percpu_ref_put(&ctx->refs);
                list_add(&req->compl.list, &state->comp.free_list);
 }
 
-static void io_submit_flush_completions(struct io_comp_state *cs,
-                                       struct io_ring_ctx *ctx)
+static void io_submit_flush_completions(struct io_ring_ctx *ctx)
 {
+       struct io_comp_state *cs = &ctx->submit_state.comp;
        int i, nr = cs->nr;
        struct io_kiocb *req;
        struct req_batch rb;
 
                        cs->reqs[cs->nr++] = req;
                        if (cs->nr == ARRAY_SIZE(cs->reqs))
-                               io_submit_flush_completions(cs, ctx);
+                               io_submit_flush_completions(ctx);
                } else {
                        io_put_req(req);
                }
        if (state->link.head)
                io_queue_sqe(state->link.head);
        if (state->comp.nr)
-               io_submit_flush_completions(&state->comp, ctx);
+               io_submit_flush_completions(ctx);
        if (state->plug_started)
                blk_finish_plug(&state->plug);
        io_state_file_put(state);