On Tue 11-12-18 12:38:00, Josef Bacik wrote:
> All of the arguments to these functions come from the vmf, and the
> following patches are going to add more arguments.  Cut down on the
> amount of arguments passed by simply passing in the vmf to these two
> helpers.
> 
> Signed-off-by: Josef Bacik <[email protected]>

The patch looks good. You can add:

Reviewed-by: Jan Kara <[email protected]>

                                                                Honza

> ---
>  mm/filemap.c | 28 ++++++++++++++--------------
>  1 file changed, 14 insertions(+), 14 deletions(-)
> 
> diff --git a/mm/filemap.c b/mm/filemap.c
> index 03bce38d8f2b..8fc45f24b201 100644
> --- a/mm/filemap.c
> +++ b/mm/filemap.c
> @@ -2309,20 +2309,20 @@ EXPORT_SYMBOL(generic_file_read_iter);
>   * Synchronous readahead happens when we don't even find
>   * a page in the page cache at all.
>   */
> -static void do_sync_mmap_readahead(struct vm_area_struct *vma,
> -                                struct file_ra_state *ra,
> -                                struct file *file,
> -                                pgoff_t offset)
> +static void do_sync_mmap_readahead(struct vm_fault *vmf)
>  {
> +     struct file *file = vmf->vma->vm_file;
> +     struct file_ra_state *ra = &file->f_ra;
>       struct address_space *mapping = file->f_mapping;
> +     pgoff_t offset = vmf->pgoff;
>  
>       /* If we don't want any read-ahead, don't bother */
> -     if (vma->vm_flags & VM_RAND_READ)
> +     if (vmf->vma->vm_flags & VM_RAND_READ)
>               return;
>       if (!ra->ra_pages)
>               return;
>  
> -     if (vma->vm_flags & VM_SEQ_READ) {
> +     if (vmf->vma->vm_flags & VM_SEQ_READ) {
>               page_cache_sync_readahead(mapping, ra, file, offset,
>                                         ra->ra_pages);
>               return;
> @@ -2352,16 +2352,16 @@ static void do_sync_mmap_readahead(struct 
> vm_area_struct *vma,
>   * Asynchronous readahead happens when we find the page and PG_readahead,
>   * so we want to possibly extend the readahead further..
>   */
> -static void do_async_mmap_readahead(struct vm_area_struct *vma,
> -                                 struct file_ra_state *ra,
> -                                 struct file *file,
> -                                 struct page *page,
> -                                 pgoff_t offset)
> +static void do_async_mmap_readahead(struct vm_fault *vmf,
> +                                 struct page *page)
>  {
> +     struct file *file = vmf->vma->vm_file;
> +     struct file_ra_state *ra = &file->f_ra;
>       struct address_space *mapping = file->f_mapping;
> +     pgoff_t offset = vmf->pgoff;
>  
>       /* If we don't want any read-ahead, don't bother */
> -     if (vma->vm_flags & VM_RAND_READ)
> +     if (vmf->vma->vm_flags & VM_RAND_READ)
>               return;
>       if (ra->mmap_miss > 0)
>               ra->mmap_miss--;
> @@ -2418,10 +2418,10 @@ vm_fault_t filemap_fault(struct vm_fault *vmf)
>                * We found the page, so try async readahead before
>                * waiting for the lock.
>                */
> -             do_async_mmap_readahead(vmf->vma, ra, file, page, offset);
> +             do_async_mmap_readahead(vmf, page);
>       } else if (!page) {
>               /* No page in the page cache at all */
> -             do_sync_mmap_readahead(vmf->vma, ra, file, offset);
> +             do_sync_mmap_readahead(vmf);
>               count_vm_event(PGMAJFAULT);
>               count_memcg_event_mm(vmf->vma->vm_mm, PGMAJFAULT);
>               ret = VM_FAULT_MAJOR;
> -- 
> 2.14.3
> 
-- 
Jan Kara <[email protected]>
SUSE Labs, CR

Reply via email to