/linux/io_uring/ |
H A D | net.h | 24 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags); 28 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags); 30 int io_send(struct io_kiocb *req, unsigned int issue_flags); 33 int io_recvmsg(struct io_kiocb *req, unsigned int issue_flags); 34 int io_recv(struct io_kiocb *req, unsigned int issue_flags); 39 int io_accept(struct io_kiocb *req, unsigned int issue_flags); 42 int io_socket(struct io_kiocb *req, unsigned int issue_flags); 45 int io_connect(struct io_kiocb *req, unsigned int issue_flags); 47 int io_send_zc(struct io_kiocb *req, unsigned int issue_flags); 53 int io_bind(struct io_kiocb *req, unsigned int issue_flags); [all …]
|
H A D | uring_cmd.c | 40 if (issue_flags & IO_URING_F_UNLOCKED) in io_req_uring_cleanup() 81 unsigned int issue_flags) in io_uring_cmd_del_cancelable() argument 90 io_ring_submit_lock(ctx, issue_flags); in io_uring_cmd_del_cancelable() 105 unsigned int issue_flags) in io_uring_cmd_mark_cancelable() argument 151 unsigned issue_flags) in io_uring_cmd_done() argument 241 issue_flags |= IO_URING_F_SQE128; in io_uring_cmd() 243 issue_flags |= IO_URING_F_CQE32; in io_uring_cmd() 245 issue_flags |= IO_URING_F_COMPAT; in io_uring_cmd() 249 issue_flags |= IO_URING_F_IOPOLL; in io_uring_cmd() 289 unsigned int issue_flags) in io_uring_cmd_getsockopt() argument [all …]
|
H A D | kbuf.h | 67 unsigned int issue_flags); 69 unsigned int issue_flags); 74 int io_remove_buffers(struct io_kiocb *req, unsigned int issue_flags); 77 int io_provide_buffers(struct io_kiocb *req, unsigned int issue_flags); 121 return io_kbuf_recycle_legacy(req, issue_flags); in io_kbuf_recycle() 195 int nbufs, unsigned issue_flags) in __io_put_kbufs() argument 207 __io_put_kbuf(req, len, issue_flags); in __io_put_kbufs() 213 unsigned issue_flags) in io_put_kbuf() argument 215 return __io_put_kbufs(req, len, 1, issue_flags); in io_put_kbuf() 219 int nbufs, unsigned issue_flags) in io_put_kbufs() argument [all …]
|
H A D | openclose.c | 114 int io_openat2(struct io_kiocb *req, unsigned int issue_flags) in io_openat2() argument 128 if (issue_flags & IO_URING_F_NONBLOCK) { in io_openat2() 153 (!resolve_nonblock && (issue_flags & IO_URING_F_NONBLOCK))) in io_openat2() 158 if ((issue_flags & IO_URING_F_NONBLOCK) && !nonblock_set) in io_openat2() 164 ret = io_fixed_fd_install(req, issue_flags, file, in io_openat2() 175 int io_openat(struct io_kiocb *req, unsigned int issue_flags) in io_openat() argument 177 return io_openat2(req, issue_flags); in io_openat() 193 io_ring_submit_lock(ctx, issue_flags); in __io_close_fixed() 195 io_ring_submit_unlock(ctx, issue_flags); in __io_close_fixed() 224 int io_close(struct io_kiocb *req, unsigned int issue_flags) in io_close() argument [all …]
|
H A D | msg_ring.c | 42 unsigned int issue_flags) in io_double_lock_ctx() argument 49 if (!(issue_flags & IO_URING_F_UNLOCKED)) { in io_double_lock_ctx() 159 if (unlikely(io_double_lock_ctx(target_ctx, issue_flags))) in io_msg_ring_data() 176 io_ring_submit_lock(ctx, issue_flags); in io_msg_grab_file() 183 io_ring_submit_unlock(ctx, issue_flags); in io_msg_grab_file() 194 if (unlikely(io_double_lock_ctx(target_ctx, issue_flags))) in io_msg_install_complete() 262 src_file = io_msg_grab_file(req, issue_flags); in io_msg_send_fd() 271 return io_msg_install_complete(req, issue_flags); in io_msg_send_fd() 294 int io_msg_ring(struct io_kiocb *req, unsigned int issue_flags) in io_msg_ring() argument 305 ret = io_msg_ring_data(req, issue_flags); in io_msg_ring() [all …]
|
H A D | openclose.h | 3 int __io_close_fixed(struct io_ring_ctx *ctx, unsigned int issue_flags, 7 int io_openat(struct io_kiocb *req, unsigned int issue_flags); 11 int io_openat2(struct io_kiocb *req, unsigned int issue_flags); 14 int io_close(struct io_kiocb *req, unsigned int issue_flags); 17 int io_install_fixed_fd(struct io_kiocb *req, unsigned int issue_flags);
|
H A D | net.c | 446 unsigned int issue_flags) in io_req_msg_cleanup() argument 449 io_netmsg_recycle(req, issue_flags); in io_req_msg_cleanup() 493 unsigned issue_flags) in io_send_finish() argument 543 if (issue_flags & IO_URING_F_NONBLOCK) in io_sendmsg() 566 io_req_msg_cleanup(req, issue_flags); in io_sendmsg() 672 io_req_msg_cleanup(req, issue_flags); in io_send() 846 issue_flags); in io_recv_finish() 886 io_req_msg_cleanup(req, issue_flags); in io_recv_finish() 1008 io_kbuf_recycle(req, issue_flags); in io_recvmsg() 1055 io_kbuf_recycle(req, issue_flags); in io_recvmsg() [all …]
|
H A D | fs.h | 4 int io_renameat(struct io_kiocb *req, unsigned int issue_flags); 8 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags); 12 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags); 16 int io_symlinkat(struct io_kiocb *req, unsigned int issue_flags); 19 int io_linkat(struct io_kiocb *req, unsigned int issue_flags);
|
H A D | futex.h | 7 int io_futex_wait(struct io_kiocb *req, unsigned int issue_flags); 8 int io_futexv_wait(struct io_kiocb *req, unsigned int issue_flags); 9 int io_futex_wake(struct io_kiocb *req, unsigned int issue_flags); 13 unsigned int issue_flags); 21 unsigned int issue_flags) in io_futex_cancel() argument
|
H A D | xattr.c | 108 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags) in io_fgetxattr() argument 113 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fgetxattr() 123 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags) in io_getxattr() argument 130 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_getxattr() 206 static int __io_setxattr(struct io_kiocb *req, unsigned int issue_flags, in __io_setxattr() argument 221 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags) in io_fsetxattr() argument 225 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fsetxattr() 227 ret = __io_setxattr(req, issue_flags, &req->file->f_path); in io_fsetxattr() 232 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags) in io_setxattr() argument 239 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_setxattr() [all …]
|
H A D | xattr.h | 6 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags); 9 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags); 12 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags); 15 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags);
|
H A D | futex.c | 117 unsigned int issue_flags) in io_futex_cancel() argument 126 io_ring_submit_lock(ctx, issue_flags); in io_futex_cancel() 136 io_ring_submit_unlock(ctx, issue_flags); in io_futex_cancel() 272 io_ring_submit_lock(ctx, issue_flags); in io_futexv_wait() 280 io_ring_submit_unlock(ctx, issue_flags); in io_futexv_wait() 316 io_ring_submit_unlock(ctx, issue_flags); in io_futexv_wait() 320 int io_futex_wait(struct io_kiocb *req, unsigned int issue_flags) in io_futex_wait() argument 333 io_ring_submit_lock(ctx, issue_flags); in io_futex_wait() 350 io_ring_submit_unlock(ctx, issue_flags); in io_futex_wait() 357 io_ring_submit_unlock(ctx, issue_flags); in io_futex_wait() [all …]
|
H A D | cancel.c | 104 unsigned issue_flags) in io_try_cancel() argument 119 ret = io_poll_cancel(ctx, cd, issue_flags); in io_try_cancel() 123 ret = io_waitid_cancel(ctx, cd, issue_flags); in io_try_cancel() 127 ret = io_futex_cancel(ctx, cd, issue_flags); in io_try_cancel() 167 unsigned int issue_flags) in __io_async_cancel() argument 175 ret = io_try_cancel(tctx, cd, issue_flags); in __io_async_cancel() 184 io_ring_submit_lock(ctx, issue_flags); in __io_async_cancel() 194 io_ring_submit_unlock(ctx, issue_flags); in __io_async_cancel() 198 int io_async_cancel(struct io_kiocb *req, unsigned int issue_flags) in io_async_cancel() argument 216 issue_flags); in io_async_cancel() [all …]
|
H A D | sync.c | 40 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags) in io_sync_file_range() argument 46 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_sync_file_range() 70 int io_fsync(struct io_kiocb *req, unsigned int issue_flags) in io_fsync() argument 77 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fsync() 99 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags) in io_fallocate() argument 105 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fallocate()
|
H A D | poll.h | 34 int io_poll_add(struct io_kiocb *req, unsigned int issue_flags); 37 int io_poll_remove(struct io_kiocb *req, unsigned int issue_flags); 41 unsigned issue_flags); 42 int io_arm_poll_handler(struct io_kiocb *req, unsigned issue_flags);
|
H A D | rw.c | 91 unsigned int issue_flags) in __io_import_iovec() argument 137 unsigned int issue_flags) in io_import_iovec() argument 207 io_rw_recycle(req, issue_flags); in io_req_rw_cleanup() 590 unsigned int issue_flags) in kiocb_done() argument 606 io_req_rw_cleanup(req, issue_flags); in kiocb_done() 943 ret = __io_read(req, issue_flags); in io_read() 962 ret = __io_read(req, issue_flags); in io_read_mshot() 973 if (io_kbuf_recycle(req, issue_flags)) in io_read_mshot() 979 io_kbuf_recycle(req, issue_flags); in io_read_mshot() 1011 io_req_rw_cleanup(req, issue_flags); in io_read_mshot() [all …]
|
H A D | fs.c | 81 int io_renameat(struct io_kiocb *req, unsigned int issue_flags) in io_renameat() argument 86 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_renameat() 130 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags) in io_unlinkat() argument 135 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_unlinkat() 177 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags) in io_mkdirat() argument 182 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_mkdirat() 227 int io_symlinkat(struct io_kiocb *req, unsigned int issue_flags) in io_symlinkat() argument 232 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_symlinkat() 272 int io_linkat(struct io_kiocb *req, unsigned int issue_flags) in io_linkat() argument 277 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_linkat()
|
H A D | splice.c | 48 int io_tee(struct io_kiocb *req, unsigned int issue_flags) in io_tee() argument 56 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_tee() 59 in = io_file_get_fixed(req, sp->splice_fd_in, issue_flags); in io_tee() 88 int io_splice(struct io_kiocb *req, unsigned int issue_flags) in io_splice() argument 97 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_splice() 100 in = io_file_get_fixed(req, sp->splice_fd_in, issue_flags); in io_splice()
|
H A D | kbuf.c | 61 io_ring_submit_lock(ctx, issue_flags); in io_kbuf_recycle_legacy() 69 io_ring_submit_unlock(ctx, issue_flags); in io_kbuf_recycle_legacy() 87 if (issue_flags & IO_URING_F_UNLOCKED) { in __io_put_kbuf() 137 unsigned int issue_flags) in io_ring_buffer_select() argument 175 unsigned int issue_flags) in io_buffer_select() argument 288 unsigned int issue_flags) in io_buffers_select() argument 294 io_ring_submit_lock(ctx, issue_flags); in io_buffers_select() 316 io_ring_submit_unlock(ctx, issue_flags); in io_buffers_select() 447 io_ring_submit_lock(ctx, issue_flags); in io_remove_buffers() 457 io_ring_submit_unlock(ctx, issue_flags); in io_remove_buffers() [all …]
|
H A D | sync.h | 4 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags); 7 int io_fsync(struct io_kiocb *req, unsigned int issue_flags); 9 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags);
|
H A D | advise.c | 51 int io_madvise(struct io_kiocb *req, unsigned int issue_flags) in io_madvise() argument 57 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_madvise() 96 int io_fadvise(struct io_kiocb *req, unsigned int issue_flags) in io_fadvise() argument 101 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK && io_fadvise_force_async(fa)); in io_fadvise()
|
H A D | poll.c | 583 unsigned issue_flags) in __io_arm_poll_handler() argument 605 ipt->owning = issue_flags & IO_URING_F_UNLOCKED; in __io_arm_poll_handler() 609 if (issue_flags & IO_URING_F_UNLOCKED) in __io_arm_poll_handler() 688 unsigned issue_flags) in io_req_alloc_apoll() argument 748 apoll = io_req_alloc_apoll(req, issue_flags); in io_arm_poll_handler() 755 io_kbuf_recycle(req, issue_flags); in io_arm_poll_handler() 889 unsigned issue_flags) in io_poll_cancel() argument 897 io_ring_submit_lock(ctx, issue_flags); in io_poll_cancel() 899 io_ring_submit_unlock(ctx, issue_flags); in io_poll_cancel() 999 io_ring_submit_lock(ctx, issue_flags); in io_poll_remove() [all …]
|
H A D | rw.h | 21 int io_read(struct io_kiocb *req, unsigned int issue_flags); 22 int io_write(struct io_kiocb *req, unsigned int issue_flags); 27 int io_read_mshot(struct io_kiocb *req, unsigned int issue_flags);
|
/linux/include/linux/io_uring/ |
H A D | cmd.h | 47 unsigned issue_flags); 58 unsigned int issue_flags); 70 ssize_t ret2, unsigned issue_flags) in io_uring_cmd_done() argument 79 unsigned int issue_flags) in io_uring_cmd_mark_cancelable() argument
|
/linux/drivers/nvme/host/ |
H A D | ioctl.c | 409 unsigned issue_flags) in nvme_uring_task_cb() argument 495 if (issue_flags & IO_URING_F_NONBLOCK) { in nvme_uring_cmd_io() 499 if (issue_flags & IO_URING_F_IOPOLL) in nvme_uring_cmd_io() 622 static int nvme_uring_cmd_checks(unsigned int issue_flags) in nvme_uring_cmd_checks() argument 633 unsigned int issue_flags) in nvme_ns_uring_cmd() argument 640 ret = nvme_uring_cmd_checks(issue_flags); in nvme_ns_uring_cmd() 663 return nvme_ns_uring_cmd(ns, ioucmd, issue_flags); in nvme_ns_chr_uring_cmd() 754 unsigned int issue_flags) in nvme_ns_head_chr_uring_cmd() argument 763 ret = nvme_ns_uring_cmd(ns, ioucmd, issue_flags); in nvme_ns_head_chr_uring_cmd() 775 if (issue_flags & IO_URING_F_IOPOLL) in nvme_dev_uring_cmd() [all …]
|