Am 15.07.24 um 11:11 schrieb Daniel Vetter:
On Thu, Jul 11, 2024 at 11:00:02AM +0200, Christian König wrote:
Am 11.07.24 um 09:42 schrieb Huan Yang:
Some user may need load file into dma-buf, current
way is:
    1. allocate a dma-buf, get dma-buf fd
    2. mmap dma-buf fd into vaddr
    3. read(file_fd, vaddr, fsz)
This is too heavy if fsz reached to GB.
You need to describe a bit more why that is to heavy. I can only assume you
need to save memory bandwidth and avoid the extra copy with the CPU.

This patch implement a feature called DMA_HEAP_IOCTL_ALLOC_READ_FILE.
User need to offer a file_fd which you want to load into dma-buf, then,
it promise if you got a dma-buf fd, it will contains the file content.
Interesting idea, that has at least more potential than trying to enable
direct I/O on mmap()ed DMA-bufs.

The approach with the new IOCTL might not work because it is a very
specialized use case.

But IIRC there was a copy_file_range callback in the file_operations
structure you could use for that. I'm just not sure when and how that's used
with the copy_file_range() system call.
I'm not sure any of those help, because internally they're all still based
on struct page (or maybe in the future on folios). And that's the thing
dma-buf can't give you, at least without peaking behind the curtain.

I think an entirely different option would be malloc+udmabuf. That
essentially handles the impendence-mismatch between direct I/O and dma-buf
on the dma-buf side. The downside is that it'll make the permanently
pinned memory accounting and tracking issues even more apparent, but I
guess eventually we do need to sort that one out.

Oh, very good idea!

Just one minor correction: it's not malloc+udmabuf, but rather create_memfd()+udmabuf.

And you need to complete your direct I/O before creating the udmabuf since that reference will prevent direct I/O from working.

Regards,
Christian.




And since all the patches here are only for the pages system heap I'm
guess udmabuf should work out for the use-case here? Worth a shot at
least.
-Sima

Regards,
Christian.

Notice, file_fd depends on user how to open this file. So, both buffer
I/O and Direct I/O is supported.

Signed-off-by: Huan Yang <l...@vivo.com>
---
   drivers/dma-buf/dma-heap.c    | 525 +++++++++++++++++++++++++++++++++-
   include/linux/dma-heap.h      |  57 +++-
   include/uapi/linux/dma-heap.h |  32 +++
   3 files changed, 611 insertions(+), 3 deletions(-)

diff --git a/drivers/dma-buf/dma-heap.c b/drivers/dma-buf/dma-heap.c
index 2298ca5e112e..abe17281adb8 100644
--- a/drivers/dma-buf/dma-heap.c
+++ b/drivers/dma-buf/dma-heap.c
@@ -15,9 +15,11 @@
   #include <linux/list.h>
   #include <linux/slab.h>
   #include <linux/nospec.h>
+#include <linux/highmem.h>
   #include <linux/uaccess.h>
   #include <linux/syscalls.h>
   #include <linux/dma-heap.h>
+#include <linux/vmalloc.h>
   #include <uapi/linux/dma-heap.h>
   #define DEVNAME "dma_heap"
@@ -43,12 +45,462 @@ struct dma_heap {
        struct cdev heap_cdev;
   };
+/**
+ * struct dma_heap_file - wrap the file, read task for dma_heap allocate use.
+ * @file:              file to read from.
+ *
+ * @cred:              kthread use, user cred copy to use for the read.
+ *
+ * @max_batch:         maximum batch size to read, if collect match batch,
+ *                     trigger read, default 128MB, must below file size.
+ *
+ * @fsz:               file size.
+ *
+ * @direct:            use direct IO?
+ */
+struct dma_heap_file {
+       struct file *file;
+       struct cred *cred;
+       size_t max_batch;
+       size_t fsz;
+       bool direct;
+};
+
+/**
+ * struct dma_heap_file_work - represents a dma_heap file read real work.
+ * @vaddr:             contigous virtual address alloc by vmap, file read need.
+ *
+ * @start_size:                file read start offset, same to 
@dma_heap_file_task->roffset.
+ *
+ * @need_size:         file read need size, same to @dma_heap_file_task->rsize.
+ *
+ * @heap_file:         file wrapper.
+ *
+ * @list:              child node of @dma_heap_file_control->works.
+ *
+ * @refp:              same @dma_heap_file_task->ref, if end of read, put ref.
+ *
+ * @failp:             if any work io failed, set it true, pointp 
@dma_heap_file_task->fail.
+ */
+struct dma_heap_file_work {
+       void *vaddr;
+       ssize_t start_size;
+       ssize_t need_size;
+       struct dma_heap_file *heap_file;
+       struct list_head list;
+       atomic_t *refp;
+       bool *failp;
+};
+
+/**
+ * struct dma_heap_file_task - represents a dma_heap file read process
+ * @ref:               current file work counter, if zero, allocate and read
+ *                     done.
+ *
+ * @roffset:           last read offset, current prepared work' begin file
+ *                     start offset.
+ *
+ * @rsize:             current allocated page size use to read, if reach 
rbatch,
+ *                     trigger commit.
+ *
+ * @rbatch:            current prepared work's batch, below @dma_heap_file's
+ *                     batch.
+ *
+ * @heap_file:         current dma_heap_file
+ *
+ * @parray:            used for vmap, size is @dma_heap_file's batch's number
+ *                     pages.(this is maximum). Due to single thread file read,
+ *                     one page array reuse each work prepare is OK.
+ *                     Each index in parray is PAGE_SIZE.(vmap need)
+ *
+ * @pindex:            current allocated page filled in @parray's index.
+ *
+ * @fail:              any work failed when file read?
+ *
+ * dma_heap_file_task is the production of file read, will prepare each work
+ * during allocate dma_buf pages, if match current batch, then trigger commit
+ * and prepare next work. After all batch queued, user going on prepare dma_buf
+ * and so on, but before return dma_buf fd, need to wait file read end and
+ * check read result.
+ */
+struct dma_heap_file_task {
+       atomic_t ref;
+       size_t roffset;
+       size_t rsize;
+       size_t rbatch;
+       struct dma_heap_file *heap_file;
+       struct page **parray;
+       unsigned int pindex;
+       bool fail;
+};
+
+/**
+ * struct dma_heap_file_control - global control of dma_heap file read.
+ * @works:             @dma_heap_file_work's list head.
+ *
+ * @lock:              only lock for @works.
+ *
+ * @threadwq:          wait queue for @work_thread, if commit work, 
@work_thread
+ *                     wakeup and read this work's file contains.
+ *
+ * @workwq:            used for main thread wait for file read end, if 
allocation
+ *                     end before file read. @dma_heap_file_task ref effect 
this.
+ *
+ * @work_thread:       file read kthread. the dma_heap_file_task work's 
consumer.
+ *
+ * @heap_fwork_cachep: @dma_heap_file_work's cachep, it's alloc/free 
frequently.
+ *
+ * @nr_work:           global number of how many work committed.
+ */
+struct dma_heap_file_control {
+       struct list_head works;
+       spinlock_t lock;
+       wait_queue_head_t threadwq;
+       wait_queue_head_t workwq;
+       struct task_struct *work_thread;
+       struct kmem_cache *heap_fwork_cachep;
+       atomic_t nr_work;
+};
+
+static struct dma_heap_file_control *heap_fctl;
   static LIST_HEAD(heap_list);
   static DEFINE_MUTEX(heap_list_lock);
   static dev_t dma_heap_devt;
   static struct class *dma_heap_class;
   static DEFINE_XARRAY_ALLOC(dma_heap_minors);
+/**
+ * map_pages_to_vaddr - map each scatter page into contiguous virtual address.
+ * @heap_ftask:                prepared and need to commit's work.
+ *
+ * Cached pages need to trigger file read, this function map each scatter page
+ * into contiguous virtual address, so that file read can easy use.
+ * Now that we get vaddr page, cached pages can return to original user, so we
+ * will not effect dma-buf export even if file read not end.
+ */
+static void *map_pages_to_vaddr(struct dma_heap_file_task *heap_ftask)
+{
+       return vmap(heap_ftask->parray, heap_ftask->pindex, VM_MAP,
+                   PAGE_KERNEL);
+}
+
+bool dma_heap_prepare_file_read(struct dma_heap_file_task *heap_ftask,
+                               struct page *page)
+{
+       struct page **array = heap_ftask->parray;
+       int index = heap_ftask->pindex;
+       int num = compound_nr(page), i;
+       unsigned long sz = page_size(page);
+
+       heap_ftask->rsize += sz;
+       for (i = 0; i < num; ++i)
+               array[index++] = &page[i];
+       heap_ftask->pindex = index;
+
+       return heap_ftask->rsize >= heap_ftask->rbatch;
+}
+
+static struct dma_heap_file_work *
+init_file_work(struct dma_heap_file_task *heap_ftask)
+{
+       struct dma_heap_file_work *heap_fwork;
+       struct dma_heap_file *heap_file = heap_ftask->heap_file;
+
+       if (READ_ONCE(heap_ftask->fail))
+               return NULL;
+
+       heap_fwork = kmem_cache_alloc(heap_fctl->heap_fwork_cachep, GFP_KERNEL);
+       if (unlikely(!heap_fwork))
+               return NULL;
+
+       heap_fwork->vaddr = map_pages_to_vaddr(heap_ftask);
+       if (unlikely(!heap_fwork->vaddr)) {
+               kmem_cache_free(heap_fctl->heap_fwork_cachep, heap_fwork);
+               return NULL;
+       }
+
+       heap_fwork->heap_file = heap_file;
+       heap_fwork->start_size = heap_ftask->roffset;
+       heap_fwork->need_size = heap_ftask->rsize;
+       heap_fwork->refp = &heap_ftask->ref;
+       heap_fwork->failp = &heap_ftask->fail;
+       atomic_inc(&heap_ftask->ref);
+       return heap_fwork;
+}
+
+static void destroy_file_work(struct dma_heap_file_work *heap_fwork)
+{
+       vunmap(heap_fwork->vaddr);
+       atomic_dec(heap_fwork->refp);
+       wake_up(&heap_fctl->workwq);
+
+       kmem_cache_free(heap_fctl->heap_fwork_cachep, heap_fwork);
+}
+
+int dma_heap_submit_file_read(struct dma_heap_file_task *heap_ftask)
+{
+       struct dma_heap_file_work *heap_fwork = init_file_work(heap_ftask);
+       struct page *last = NULL;
+       struct dma_heap_file *heap_file = heap_ftask->heap_file;
+       size_t start = heap_ftask->roffset;
+       struct file *file = heap_file->file;
+       size_t fsz = heap_file->fsz;
+
+       if (unlikely(!heap_fwork))
+               return -ENOMEM;
+
+       /**
+        * If file size is not page aligned, direct io can't process the tail.
+        * So, if reach to tail, remain the last page use buffer read.
+        */
+       if (heap_file->direct && start + heap_ftask->rsize > fsz) {
+               heap_fwork->need_size -= PAGE_SIZE;
+               last = heap_ftask->parray[heap_ftask->pindex - 1];
+       }
+
+       spin_lock(&heap_fctl->lock);
+       list_add_tail(&heap_fwork->list, &heap_fctl->works);
+       spin_unlock(&heap_fctl->lock);
+       atomic_inc(&heap_fctl->nr_work);
+
+       wake_up(&heap_fctl->threadwq);
+
+       if (last) {
+               char *buf, *pathp;
+               ssize_t err;
+               void *buffer;
+
+               buf = kmalloc(PATH_MAX, GFP_KERNEL);
+               if (unlikely(!buf))
+                       return -ENOMEM;
+
+               start = PAGE_ALIGN_DOWN(fsz);
+
+               pathp = file_path(file, buf, PATH_MAX);
+               if (IS_ERR(pathp)) {
+                       kfree(buf);
+                       return PTR_ERR(pathp);
+               }
+
+               buffer = kmap_local_page(last); // use page's kaddr.
+               err = kernel_read_file_from_path(pathp, start, &buffer,
+                                                fsz - start, &fsz,
+                                                READING_POLICY);
+               kunmap_local(buffer);
+               kfree(buf);
+               if (err < 0) {
+                       pr_err("failed to use buffer kernel_read_file %s, err=%ld, 
[%ld, %ld], f_sz=%ld\n",
+                              pathp, err, start, fsz, fsz);
+
+                       return err;
+               }
+       }
+
+       heap_ftask->roffset += heap_ftask->rsize;
+       heap_ftask->rsize = 0;
+       heap_ftask->pindex = 0;
+       heap_ftask->rbatch = min_t(size_t,
+                                  PAGE_ALIGN(fsz) - heap_ftask->roffset,
+                                  heap_ftask->rbatch);
+       return 0;
+}
+
+bool dma_heap_wait_for_file_read(struct dma_heap_file_task *heap_ftask)
+{
+       wait_event_freezable(heap_fctl->workwq,
+                            atomic_read(&heap_ftask->ref) == 0);
+       return heap_ftask->fail;
+}
+
+bool dma_heap_destroy_file_read(struct dma_heap_file_task *heap_ftask)
+{
+       bool fail;
+
+       dma_heap_wait_for_file_read(heap_ftask);
+       fail = heap_ftask->fail;
+       kvfree(heap_ftask->parray);
+       kfree(heap_ftask);
+       return fail;
+}
+
+struct dma_heap_file_task *
+dma_heap_declare_file_read(struct dma_heap_file *heap_file)
+{
+       struct dma_heap_file_task *heap_ftask =
+               kzalloc(sizeof(*heap_ftask), GFP_KERNEL);
+       if (unlikely(!heap_ftask))
+               return NULL;
+
+       /**
+        * Batch is the maximum size which we prepare work will meet.
+        * So, direct alloc this number's page array is OK.
+        */
+       heap_ftask->parray = kvmalloc_array(heap_file->max_batch >> PAGE_SHIFT,
+                                           sizeof(struct page *), GFP_KERNEL);
+       if (unlikely(!heap_ftask->parray))
+               goto put;
+
+       heap_ftask->heap_file = heap_file;
+       heap_ftask->rbatch = heap_file->max_batch;
+       return heap_ftask;
+put:
+       kfree(heap_ftask);
+       return NULL;
+}
+
+static void __work_this_io(struct dma_heap_file_work *heap_fwork)
+{
+       struct dma_heap_file *heap_file = heap_fwork->heap_file;
+       struct file *file = heap_file->file;
+       ssize_t start = heap_fwork->start_size;
+       ssize_t size = heap_fwork->need_size;
+       void *buffer = heap_fwork->vaddr;
+       const struct cred *old_cred;
+       ssize_t err;
+
+       // use real task's cred to read this file.
+       old_cred = override_creds(heap_file->cred);
+       err = kernel_read_file(file, start, &buffer, size, &heap_file->fsz,
+                              READING_POLICY);
+       if (err < 0) {
+               pr_err("use kernel_read_file, err=%ld, [%ld, %ld], f_sz=%ld\n",
+                      err, start, (start + size), heap_file->fsz);
+               WRITE_ONCE(*heap_fwork->failp, true);
+       }
+       // recovery to my cred.
+       revert_creds(old_cred);
+}
+
+static int dma_heap_file_control_thread(void *data)
+{
+       struct dma_heap_file_control *heap_fctl =
+               (struct dma_heap_file_control *)data;
+       struct dma_heap_file_work *worker, *tmp;
+       int nr_work;
+
+       LIST_HEAD(pages);
+       LIST_HEAD(workers);
+
+       while (true) {
+               wait_event_freezable(heap_fctl->threadwq,
+                                    atomic_read(&heap_fctl->nr_work) > 0);
+recheck:
+               spin_lock(&heap_fctl->lock);
+               list_splice_init(&heap_fctl->works, &workers);
+               spin_unlock(&heap_fctl->lock);
+
+               if (unlikely(kthread_should_stop())) {
+                       list_for_each_entry_safe(worker, tmp, &workers, list) {
+                               list_del(&worker->list);
+                               destroy_file_work(worker);
+                       }
+                       break;
+               }
+
+               nr_work = 0;
+               list_for_each_entry_safe(worker, tmp, &workers, list) {
+                       ++nr_work;
+                       list_del(&worker->list);
+                       __work_this_io(worker);
+
+                       destroy_file_work(worker);
+               }
+               atomic_sub(nr_work, &heap_fctl->nr_work);
+
+               if (atomic_read(&heap_fctl->nr_work) > 0)
+                       goto recheck;
+       }
+       return 0;
+}
+
+size_t dma_heap_file_size(struct dma_heap_file *heap_file)
+{
+       return heap_file->fsz;
+}
+
+static int prepare_dma_heap_file(struct dma_heap_file *heap_file, int file_fd,
+                                size_t batch)
+{
+       struct file *file;
+       size_t fsz;
+       int ret;
+
+       file = fget(file_fd);
+       if (!file)
+               return -EINVAL;
+
+       fsz = i_size_read(file_inode(file));
+       if (fsz < batch) {
+               ret = -EINVAL;
+               goto err;
+       }
+
+       /**
+        * Selinux block our read, but actually we are reading the stand-in
+        * for this file.
+        * So save current's cred and when going to read, override mine, and
+        * end of read, revert.
+        */
+       heap_file->cred = prepare_kernel_cred(current);
+       if (unlikely(!heap_file->cred)) {
+               ret = -ENOMEM;
+               goto err;
+       }
+
+       heap_file->file = file;
+       heap_file->max_batch = batch;
+       heap_file->fsz = fsz;
+
+       heap_file->direct = file->f_flags & O_DIRECT;
+
+#define DMA_HEAP_SUGGEST_DIRECT_IO_SIZE (1UL << 30)
+       if (!heap_file->direct && fsz >= DMA_HEAP_SUGGEST_DIRECT_IO_SIZE)
+               pr_warn("alloc read file better to use O_DIRECT to read larget 
file\n");
+
+       return 0;
+
+err:
+       fput(file);
+       return ret;
+}
+
+static void destroy_dma_heap_file(struct dma_heap_file *heap_file)
+{
+       fput(heap_file->file);
+       put_cred(heap_file->cred);
+}
+
+static int dma_heap_buffer_alloc_read_file(struct dma_heap *heap, int file_fd,
+                                          size_t batch, unsigned int fd_flags,
+                                          unsigned int heap_flags)
+{
+       struct dma_buf *dmabuf;
+       int fd;
+       struct dma_heap_file heap_file;
+
+       fd = prepare_dma_heap_file(&heap_file, file_fd, batch);
+       if (fd)
+               goto error_file;
+
+       dmabuf = heap->ops->allocate_read_file(heap, &heap_file, fd_flags,
+                                              heap_flags);
+       if (IS_ERR(dmabuf)) {
+               fd = PTR_ERR(dmabuf);
+               goto error;
+       }
+
+       fd = dma_buf_fd(dmabuf, fd_flags);
+       if (fd < 0) {
+               dma_buf_put(dmabuf);
+               /* just return, as put will call release and that will free */
+       }
+
+error:
+       destroy_dma_heap_file(&heap_file);
+error_file:
+       return fd;
+}
+
   static int dma_heap_buffer_alloc(struct dma_heap *heap, size_t len,
                                 u32 fd_flags,
                                 u64 heap_flags)
@@ -93,6 +545,38 @@ static int dma_heap_open(struct inode *inode, struct file 
*file)
        return 0;
   }
+static long dma_heap_ioctl_allocate_read_file(struct file *file, void *data)
+{
+       struct dma_heap_allocation_file_data *heap_allocation_file = data;
+       struct dma_heap *heap = file->private_data;
+       int fd;
+
+       if (heap_allocation_file->fd || !heap_allocation_file->file_fd)
+               return -EINVAL;
+
+       if (heap_allocation_file->fd_flags & ~DMA_HEAP_VALID_FD_FLAGS)
+               return -EINVAL;
+
+       if (heap_allocation_file->heap_flags & ~DMA_HEAP_VALID_HEAP_FLAGS)
+               return -EINVAL;
+
+       if (!heap->ops->allocate_read_file)
+               return -EINVAL;
+
+       fd = dma_heap_buffer_alloc_read_file(
+               heap, heap_allocation_file->file_fd,
+               heap_allocation_file->batch ?
+                       PAGE_ALIGN(heap_allocation_file->batch) :
+                       DEFAULT_ADI_BATCH,
+               heap_allocation_file->fd_flags,
+               heap_allocation_file->heap_flags);
+       if (fd < 0)
+               return fd;
+
+       heap_allocation_file->fd = fd;
+       return 0;
+}
+
   static long dma_heap_ioctl_allocate(struct file *file, void *data)
   {
        struct dma_heap_allocation_data *heap_allocation = data;
@@ -121,6 +605,7 @@ static long dma_heap_ioctl_allocate(struct file *file, void 
*data)
   static unsigned int dma_heap_ioctl_cmds[] = {
        DMA_HEAP_IOCTL_ALLOC,
+       DMA_HEAP_IOCTL_ALLOC_AND_READ,
   };
   static long dma_heap_ioctl(struct file *file, unsigned int ucmd,
@@ -170,6 +655,9 @@ static long dma_heap_ioctl(struct file *file, unsigned int 
ucmd,
        case DMA_HEAP_IOCTL_ALLOC:
                ret = dma_heap_ioctl_allocate(file, kdata);
                break;
+       case DMA_HEAP_IOCTL_ALLOC_AND_READ:
+               ret = dma_heap_ioctl_allocate_read_file(file, kdata);
+               break;
        default:
                ret = -ENOTTY;
                goto err;
@@ -316,11 +804,44 @@ static int dma_heap_init(void)
        dma_heap_class = class_create(DEVNAME);
        if (IS_ERR(dma_heap_class)) {
-               unregister_chrdev_region(dma_heap_devt, NUM_HEAP_MINORS);
-               return PTR_ERR(dma_heap_class);
+               ret = PTR_ERR(dma_heap_class);
+               goto fail_class;
        }
        dma_heap_class->devnode = dma_heap_devnode;
+       heap_fctl = kzalloc(sizeof(*heap_fctl), GFP_KERNEL);
+       if (unlikely(!heap_fctl)) {
+               ret =  -ENOMEM;
+               goto fail_alloc;
+       }
+
+       INIT_LIST_HEAD(&heap_fctl->works);
+       init_waitqueue_head(&heap_fctl->threadwq);
+       init_waitqueue_head(&heap_fctl->workwq);
+
+       heap_fctl->work_thread = kthread_run(dma_heap_file_control_thread,
+                                            heap_fctl, "heap_fwork_t");
+       if (IS_ERR(heap_fctl->work_thread)) {
+               ret = -ENOMEM;
+               goto fail_thread;
+       }
+
+       heap_fctl->heap_fwork_cachep = KMEM_CACHE(dma_heap_file_work, 0);
+       if (unlikely(!heap_fctl->heap_fwork_cachep)) {
+               ret = -ENOMEM;
+               goto fail_cache;
+       }
+
        return 0;
+
+fail_cache:
+       kthread_stop(heap_fctl->work_thread);
+fail_thread:
+       kfree(heap_fctl);
+fail_alloc:
+       class_destroy(dma_heap_class);
+fail_class:
+       unregister_chrdev_region(dma_heap_devt, NUM_HEAP_MINORS);
+       return ret;
   }
   subsys_initcall(dma_heap_init);
diff --git a/include/linux/dma-heap.h b/include/linux/dma-heap.h
index 064bad725061..9c25383f816c 100644
--- a/include/linux/dma-heap.h
+++ b/include/linux/dma-heap.h
@@ -12,12 +12,17 @@
   #include <linux/cdev.h>
   #include <linux/types.h>
+#define DEFAULT_ADI_BATCH (128 << 20)
+
   struct dma_heap;
+struct dma_heap_file_task;
+struct dma_heap_file;
   /**
    * struct dma_heap_ops - ops to operate on a given heap
    * @allocate:                allocate dmabuf and return struct dma_buf ptr
- *
+ * @allocate_read_file: allocate dmabuf and read file, then return struct
+ * dma_buf ptr.
    * allocate returns dmabuf on success, ERR_PTR(-errno) on error.
    */
   struct dma_heap_ops {
@@ -25,6 +30,11 @@ struct dma_heap_ops {
                                    unsigned long len,
                                    u32 fd_flags,
                                    u64 heap_flags);
+
+       struct dma_buf *(*allocate_read_file)(struct dma_heap *heap,
+                                             struct dma_heap_file *heap_file,
+                                             u32 fd_flags,
+                                             u64 heap_flags);
   };
   /**
@@ -65,4 +75,49 @@ const char *dma_heap_get_name(struct dma_heap *heap);
    */
   struct dma_heap *dma_heap_add(const struct dma_heap_export_info *exp_info);
+/**
+ * dma_heap_destroy_file_read - waits for a file read to complete then destroy 
it
+ * Returns: true if the file read failed, false otherwise
+ */
+bool dma_heap_destroy_file_read(struct dma_heap_file_task *heap_ftask);
+
+/**
+ * dma_heap_wait_for_file_read - waits for a file read to complete
+ * Returns: true if the file read failed, false otherwise
+ */
+bool dma_heap_wait_for_file_read(struct dma_heap_file_task *heap_ftask);
+
+/**
+ * dma_heap_alloc_file_read - Declare a task to read file when allocate pages.
+ * @heap_file:         target file to read
+ *
+ * Return NULL if failed, otherwise return a struct pointer.
+ */
+struct dma_heap_file_task *
+dma_heap_declare_file_read(struct dma_heap_file *heap_file);
+
+/**
+ * dma_heap_prepare_file_read - cache each allocated page until we meet this 
batch.
+ * @heap_ftask:                prepared and need to commit's work.
+ * @page:              current allocated page. don't care which order.
+ *
+ * Returns true if reach to batch, false so go on prepare.
+ */
+bool dma_heap_prepare_file_read(struct dma_heap_file_task *heap_ftask,
+                               struct page *page);
+
+/**
+ * dma_heap_commit_file_read -  prepare collect enough memory, going to 
trigger IO
+ * @heap_ftask:                        info that current IO needs
+ *
+ * This commit will also check if reach to tail read.
+ * For direct I/O submissions, it is necessary to pay attention to file reads
+ * that are not page-aligned. For the unaligned portion of the read, buffer IO
+ * needs to be triggered.
+ * Returns:
+ *   0 if all right, -errno if something wrong
+ */
+int dma_heap_submit_file_read(struct dma_heap_file_task *heap_ftask);
+size_t dma_heap_file_size(struct dma_heap_file *heap_file);
+
   #endif /* _DMA_HEAPS_H */
diff --git a/include/uapi/linux/dma-heap.h b/include/uapi/linux/dma-heap.h
index a4cf716a49fa..8c20e8b74eed 100644
--- a/include/uapi/linux/dma-heap.h
+++ b/include/uapi/linux/dma-heap.h
@@ -39,6 +39,27 @@ struct dma_heap_allocation_data {
        __u64 heap_flags;
   };
+/**
+ * struct dma_heap_allocation_file_data - metadata passed from userspace for
+ *                                      allocations and read file
+ * @fd:                        will be populated with a fd which provides the
+ *                     handle to the allocated dma-buf
+ * @file_fd:           file descriptor to read from(suggested to use O_DIRECT 
open file)
+ * @batch:             how many memory alloced then file read(bytes), default 
128MB
+ *                     will auto aligned to PAGE_SIZE
+ * @fd_flags:          file descriptor flags used when allocating
+ * @heap_flags:                flags passed to heap
+ *
+ * Provided by userspace as an argument to the ioctl
+ */
+struct dma_heap_allocation_file_data {
+       __u32 fd;
+       __u32 file_fd;
+       __u32 batch;
+       __u32 fd_flags;
+       __u64 heap_flags;
+};
+
   #define DMA_HEAP_IOC_MAGIC           'H'
   /**
@@ -50,4 +71,15 @@ struct dma_heap_allocation_data {
   #define DMA_HEAP_IOCTL_ALLOC _IOWR(DMA_HEAP_IOC_MAGIC, 0x0,\
                                      struct dma_heap_allocation_data)
+/**
+ * DOC: DMA_HEAP_IOCTL_ALLOC_AND_READ - allocate memory from pool and both
+ *                                     read file when allocate memory.
+ *
+ * Takes a dma_heap_allocation_file_data struct and returns it with the fd 
field
+ * populated with the dmabuf handle of the allocation. When return, the dma-buf
+ * content is read from file.
+ */
+#define DMA_HEAP_IOCTL_ALLOC_AND_READ \
+       _IOWR(DMA_HEAP_IOC_MAGIC, 0x1, struct dma_heap_allocation_file_data)
+
   #endif /* _UAPI_LINUX_DMABUF_POOL_H */

Reply via email to