diff options
author | Jens Axboe <axboe@kernel.dk> | 2024-03-13 18:56:14 +0300 |
---|---|---|
committer | Jens Axboe <axboe@kernel.dk> | 2024-04-15 17:10:26 +0300 |
commit | 3ab1db3c6039e02a9deb9d5091d28d559917a645 (patch) | |
tree | 5b9e578aedf9848f2b0a8c8830fe2a4218aa56dc /io_uring/io_uring.h | |
parent | 62346c6cb28b043f2a6e95337d9081ec0b37b5f5 (diff) | |
download | linux-3ab1db3c6039e02a9deb9d5091d28d559917a645.tar.xz |
io_uring: get rid of remap_pfn_range() for mapping rings/sqes
Rather than use remap_pfn_range() for this and manually free later,
switch to using vm_insert_pages() and have it Just Work.
If possible, allocate a single compound page that covers the range that
is needed. If that works, then we can just use page_address() on that
page. If we fail to get a compound page, allocate single pages and use
vmap() to map them into the kernel virtual address space.
This just covers the rings/sqes, the other remaining user of the mmap
remap_pfn_range() user will be converted separately. Once that is done,
we can kill the old alloc/free code.
Signed-off-by: Jens Axboe <axboe@kernel.dk>
Diffstat (limited to 'io_uring/io_uring.h')
-rw-r--r-- | io_uring/io_uring.h | 2 |
1 files changed, 2 insertions, 0 deletions
diff --git a/io_uring/io_uring.h b/io_uring/io_uring.h index dbd9a2b870eb..75230d914007 100644 --- a/io_uring/io_uring.h +++ b/io_uring/io_uring.h @@ -70,6 +70,8 @@ bool io_req_post_cqe(struct io_kiocb *req, s32 res, u32 cflags); void __io_commit_cqring_flush(struct io_ring_ctx *ctx); struct page **io_pin_pages(unsigned long ubuf, unsigned long len, int *npages); +int io_uring_mmap_pages(struct io_ring_ctx *ctx, struct vm_area_struct *vma, + struct page **pages, int npages); struct file *io_file_get_normal(struct io_kiocb *req, int fd); struct file *io_file_get_fixed(struct io_kiocb *req, int fd, |