On 03/16/2018 08:47 PM, John Hubbard wrote:
> On 03/16/2018 07:36 PM, John Hubbard wrote:
>> On 03/16/2018 12:14 PM, jgli...@redhat.com wrote:
>>> From: Ralph Campbell <rcampb...@nvidia.com>
>>>
>>
>> <snip>
>>
>>> +static void hmm_release(struct mmu_notifier *mn, struct mm_struct *mm)
>>> +{
>>> +   struct hmm *hmm = mm->hmm;
>>> +   struct hmm_mirror *mirror;
>>> +   struct hmm_mirror *mirror_next;
>>> +
>>> +   down_write(&hmm->mirrors_sem);
>>> +   list_for_each_entry_safe(mirror, mirror_next, &hmm->mirrors, list) {
>>> +           list_del_init(&mirror->list);
>>> +           if (mirror->ops->release)
>>> +                   mirror->ops->release(mirror);
>>> +   }
>>> +   up_write(&hmm->mirrors_sem);
>>> +}
>>> +
>>
>> OK, as for actual code review:
>>
>> This part of the locking looks good. However, I think it can race against
>> hmm_mirror_register(), because hmm_mirror_register() will just add a new 
>> mirror regardless.
>>
>> So:
>>
>> thread 1                                      thread 2
>> --------------                                -----------------
>> hmm_release                                   hmm_mirror_register 
>>     down_write(&hmm->mirrors_sem);                <blocked: waiting for sem>
>>         // deletes all list items
>>     up_write
>>                                                   unblocked: adds new mirror
>>                                               
>>

Mark Hairgrove just pointed out some more fun facts:

1. Because hmm_mirror_register() needs to be called with an mm that has a 
non-zero
refcount, you generally cannot get an hmm_release callback, so the above race 
should
not happen.

2. We looked around, and the code is missing a call to 
mmu_notifier_unregister().
That means that it is going to leak memory and not let the mm get released 
either.

Maybe having each mirror have its own mmu notifier callback is a possible way
to solve this.

thanks,
-- 
John Hubbard
NVIDIA

Reply via email to