On Mon, Feb 09, 2026 at 04:35:47PM +0100, David Hildenbrand (Arm) wrote:
> On 2/3/26 21:56, Peter Xu wrote:
>
> > +static vm_fault_t fault_process_userfaultfd(struct vm_fault *vmf)
> > +{
> > +   struct vm_area_struct *vma = vmf->vma;
> > +   struct inode *inode = file_inode(vma->vm_file);
> > +   /*
> > +    * NOTE: we could double check this hook present when
> > +    * UFFDIO_REGISTER on MISSING or MINOR for a file driver.
> > +    */
> > +   struct folio *folio =
> > +       vma->vm_ops->uffd_ops->get_folio_noalloc(inode, vmf->pgoff);
> > +
> > +   if (!IS_ERR_OR_NULL(folio)) {
> > +           /*
> > +            * TODO: provide a flag for get_folio_noalloc() to avoid
> > +            * locking (or even the extra reference?)
> > +            */
> > +           folio_unlock(folio);
> > +           folio_put(folio);
> > +           if (userfaultfd_minor(vma))
> > +                   return handle_userfault(vmf, VM_UFFD_MINOR);
> > +   } else {
> > +           return handle_userfault(vmf, VM_UFFD_MISSING);
> > +   }
> > +
> > +   return 0;
> > +}
> > +
> >   /*
> >    * The mmap_lock must have been held on entry, and may have been
> >    * released depending on flags and vma->vm_ops->fault() return value.
> > @@ -5370,16 +5397,20 @@ static vm_fault_t __do_fault(struct vm_fault *vmf)
> >                     return VM_FAULT_OOM;
> >     }
> > +   /*
> > +    * If this is an userfaultfd trap, process it in advance before
> > +    * triggering the genuine fault handler.
> > +    */
> > +   if (userfaultfd_missing(vma) || userfaultfd_minor(vma)) {
> > +           ret = fault_process_userfaultfd(vmf);
> > +           if (ret)
> > +                   return ret;
> > +   }

I agree this is neater than handling VM_FAULT_UFFD.
I'd just move the checks for userfaultfd_minor() and userfaultfd_missing()
inside fault_process_userfaultfd().

> > +
> >     ret = vma->vm_ops->fault(vmf);
> >     if (unlikely(ret & (VM_FAULT_ERROR | VM_FAULT_NOPAGE | VM_FAULT_RETRY |
> > -                       VM_FAULT_DONE_COW | VM_FAULT_UFFD_MINOR |
> > -                       VM_FAULT_UFFD_MISSING))) {
> > -           if (ret & VM_FAULT_UFFD_MINOR)
> > -                   return handle_userfault(vmf, VM_UFFD_MINOR);
> > -           if (ret & VM_FAULT_UFFD_MISSING)
> > -                   return handle_userfault(vmf, VM_UFFD_MISSING);
> > +                       VM_FAULT_DONE_COW)))
> >             return ret;
> > -   }
> >     folio = page_folio(vmf->page);
> >     if (unlikely(PageHWPoison(vmf->page))) {
> > diff --git a/mm/shmem.c b/mm/shmem.c
> > index eafd7986fc2ec..5286f28b3e443 100644
> > --- a/mm/shmem.c
> > +++ b/mm/shmem.c
> > @@ -2484,13 +2484,6 @@ static int shmem_get_folio_gfp(struct inode *inode, 
> > pgoff_t index,
> >     fault_mm = vma ? vma->vm_mm : NULL;
> >     folio = filemap_get_entry(inode->i_mapping, index);
> > -   if (folio && vma && userfaultfd_minor(vma)) {
> > -           if (!xa_is_value(folio))
> > -                   folio_put(folio);
> > -           *fault_type = VM_FAULT_UFFD_MINOR;
> > -           return 0;
> > -   }
> > -
> >     if (xa_is_value(folio)) {
> >             error = shmem_swapin_folio(inode, index, &folio,
> >                                        sgp, gfp, vma, fault_type);
> > @@ -2535,11 +2528,6 @@ static int shmem_get_folio_gfp(struct inode *inode, 
> > pgoff_t index,
> >      * Fast cache lookup and swap lookup did not find it: allocate.
> >      */
> > -   if (vma && userfaultfd_missing(vma)) {
> > -           *fault_type = VM_FAULT_UFFD_MISSING;
> > -           return 0;
> > -   }
> > -
> >     /* Find hugepage orders that are allowed for anonymous shmem and tmpfs. 
> > */
> >     orders = shmem_allowable_huge_orders(inode, vma, index, write_end, 
> > false);
> >     if (orders > 0) {
> > diff --git a/virt/kvm/guest_memfd.c b/virt/kvm/guest_memfd.c
> > index 14cca057fc0ec..bd0de685f42f8 100644
> > --- a/virt/kvm/guest_memfd.c
> > +++ b/virt/kvm/guest_memfd.c
> > @@ -421,26 +421,6 @@ static vm_fault_t kvm_gmem_fault_user_mapping(struct 
> > vm_fault *vmf)
> >     folio = __filemap_get_folio(inode->i_mapping, vmf->pgoff,
> >                                 FGP_LOCK | FGP_ACCESSED, 0);
> > -   if (userfaultfd_armed(vmf->vma)) {
> > -           /*
> > -            * If userfaultfd is registered in minor mode and a folio
> > -            * exists, return VM_FAULT_UFFD_MINOR to trigger the
> > -            * userfaultfd handler.
> > -            */
> > -           if (userfaultfd_minor(vmf->vma) && !IS_ERR_OR_NULL(folio)) {
> > -                   ret = VM_FAULT_UFFD_MINOR;
> > -                   goto out_folio;
> > -           }
> > -
> > -           /*
> > -            * Check if userfaultfd is registered in missing mode. If so,
> > -            * check if a folio exists in the page cache. If not, return
> > -            * VM_FAULT_UFFD_MISSING to trigger the userfaultfd handler.
> > -            */
> > -           if (userfaultfd_missing(vmf->vma) && IS_ERR_OR_NULL(folio))
> > -                   return VM_FAULT_UFFD_MISSING;
> > -   }
> > -
> >     /* folio not in the pagecache, try to allocate */
> >     if (IS_ERR(folio))
> >             folio = __kvm_gmem_folio_alloc(inode, vmf->pgoff);
> 
> That looks better in general. We should likely find a better/more consistent
> name for fault_process_userfaultfd().

__do_userfault()? :)
 
> -- 
> Cheers,
> 
> David

-- 
Sincerely yours,
Mike.

Reply via email to