};
 
 enum io_uring_cmd_flags {
-       IO_URING_F_NONBLOCK             = 1,
-       IO_URING_F_COMPLETE_DEFER       = 2,
+       IO_URING_F_COMPLETE_DEFER       = 1,
+       /* int's last bit, sign checks are usually faster than a bit test */
+       IO_URING_F_NONBLOCK             = INT_MIN,
 };
 
 struct io_mapped_ubuf {
 }
 
 static struct io_buffer *io_buffer_select(struct io_kiocb *req, size_t *len,
-                                         int bgid, bool needs_lock)
+                                         int bgid, unsigned int issue_flags)
 {
        struct io_buffer *kbuf = req->kbuf;
        struct io_buffer *head;
+       bool needs_lock = !(issue_flags & IO_URING_F_NONBLOCK);
 
        if (req->flags & REQ_F_BUFFER_SELECTED)
                return kbuf;
 }
 
 static void __user *io_rw_buffer_select(struct io_kiocb *req, size_t *len,
-                                       bool needs_lock)
+                                       unsigned int issue_flags)
 {
        struct io_buffer *kbuf;
        u16 bgid;
 
        bgid = req->buf_index;
-       kbuf = io_buffer_select(req, len, bgid, needs_lock);
+       kbuf = io_buffer_select(req, len, bgid, issue_flags);
        if (IS_ERR(kbuf))
                return kbuf;
        return u64_to_user_ptr(kbuf->addr);
 
 #ifdef CONFIG_COMPAT
 static ssize_t io_compat_import(struct io_kiocb *req, struct iovec *iov,
-                               bool needs_lock)
+                               unsigned int issue_flags)
 {
        struct compat_iovec __user *uiov;
        compat_ssize_t clen;
                return -EINVAL;
 
        len = clen;
-       buf = io_rw_buffer_select(req, &len, needs_lock);
+       buf = io_rw_buffer_select(req, &len, issue_flags);
        if (IS_ERR(buf))
                return PTR_ERR(buf);
        iov[0].iov_base = buf;
 #endif
 
 static ssize_t __io_iov_buffer_select(struct io_kiocb *req, struct iovec *iov,
-                                     bool needs_lock)
+                                     unsigned int issue_flags)
 {
        struct iovec __user *uiov = u64_to_user_ptr(req->rw.addr);
        void __user *buf;
        len = iov[0].iov_len;
        if (len < 0)
                return -EINVAL;
-       buf = io_rw_buffer_select(req, &len, needs_lock);
+       buf = io_rw_buffer_select(req, &len, issue_flags);
        if (IS_ERR(buf))
                return PTR_ERR(buf);
        iov[0].iov_base = buf;
 }
 
 static ssize_t io_iov_buffer_select(struct io_kiocb *req, struct iovec *iov,
-                                   bool needs_lock)
+                                   unsigned int issue_flags)
 {
        if (req->flags & REQ_F_BUFFER_SELECTED) {
                struct io_buffer *kbuf = req->kbuf;
 
 #ifdef CONFIG_COMPAT
        if (req->ctx->compat)
-               return io_compat_import(req, iov, needs_lock);
+               return io_compat_import(req, iov, issue_flags);
 #endif
 
-       return __io_iov_buffer_select(req, iov, needs_lock);
+       return __io_iov_buffer_select(req, iov, issue_flags);
 }
 
 static int io_import_iovec(int rw, struct io_kiocb *req, struct iovec **iovec,
-                          struct iov_iter *iter, bool needs_lock)
+                          struct iov_iter *iter, unsigned int issue_flags)
 {
        void __user *buf = u64_to_user_ptr(req->rw.addr);
        size_t sqe_len = req->rw.len;
 
        if (opcode == IORING_OP_READ || opcode == IORING_OP_WRITE) {
                if (req->flags & REQ_F_BUFFER_SELECT) {
-                       buf = io_rw_buffer_select(req, &sqe_len, needs_lock);
+                       buf = io_rw_buffer_select(req, &sqe_len, issue_flags);
                        if (IS_ERR(buf))
                                return PTR_ERR(buf);
                        req->rw.len = sqe_len;
        }
 
        if (req->flags & REQ_F_BUFFER_SELECT) {
-               ret = io_iov_buffer_select(req, *iovec, needs_lock);
+               ret = io_iov_buffer_select(req, *iovec, issue_flags);
                if (!ret)
                        iov_iter_init(iter, rw, *iovec, 1, (*iovec)->iov_len);
                *iovec = NULL;
        struct iovec *iov = iorw->s.fast_iov;
        int ret;
 
-       ret = io_import_iovec(rw, req, &iov, &iorw->s.iter, false);
+       /* submission path, ->uring_lock should already be taken */
+       ret = io_import_iovec(rw, req, &iov, &iorw->s.iter, IO_URING_F_NONBLOCK);
        if (unlikely(ret < 0))
                return ret;
 
        } else {
                s = &__s;
                iovec = s->fast_iov;
-               ret = io_import_iovec(READ, req, &iovec, &s->iter, !force_nonblock);
+               ret = io_import_iovec(READ, req, &iovec, &s->iter, issue_flags);
                if (ret < 0)
                        return ret;
 
        } else {
                s = &__s;
                iovec = s->fast_iov;
-               ret = io_import_iovec(WRITE, req, &iovec, &s->iter, !force_nonblock);
+               ret = io_import_iovec(WRITE, req, &iovec, &s->iter, issue_flags);
                if (ret < 0)
                        return ret;
                iov_iter_save_state(&s->iter, &s->iter_state);
 }
 
 static struct io_buffer *io_recv_buffer_select(struct io_kiocb *req,
-                                              bool needs_lock)
+                                              unsigned int issue_flags)
 {
        struct io_sr_msg *sr = &req->sr_msg;
 
-       return io_buffer_select(req, &sr->len, sr->bgid, needs_lock);
+       return io_buffer_select(req, &sr->len, sr->bgid, issue_flags);
 }
 
 static inline unsigned int io_put_recv_kbuf(struct io_kiocb *req)
        }
 
        if (req->flags & REQ_F_BUFFER_SELECT) {
-               kbuf = io_recv_buffer_select(req, !force_nonblock);
+               kbuf = io_recv_buffer_select(req, issue_flags);
                if (IS_ERR(kbuf))
                        return PTR_ERR(kbuf);
                kmsg->fast_iov[0].iov_base = u64_to_user_ptr(kbuf->addr);
                return -ENOTSOCK;
 
        if (req->flags & REQ_F_BUFFER_SELECT) {
-               kbuf = io_recv_buffer_select(req, !force_nonblock);
+               kbuf = io_recv_buffer_select(req, issue_flags);
                if (IS_ERR(kbuf))
                        return PTR_ERR(kbuf);
                buf = u64_to_user_ptr(kbuf->addr);