On 02/12/18 00:58, Rasmus Villemoes wrote:
> On 2018-02-12 07:27, frowand.l...@gmail.com wrote:
>> From: Frank Rowand <frank.row...@sony.com>
>> Create a cache of the nodes that contain a phandle property. Use this
>> cache to find the node for a given phandle value instead of scanning
>> the devicetree to find the node. If the phandle value is not found
>> in the cache, of_find_node_by_phandle() will fall back to the tree
>> scan algorithm.
>> The cache is initialized in of_core_init().
>> The cache is freed via a late_initcall_sync() if modules are not
> Maybe a few words about the memory consumption of this solution versus
> the other proposed ones.
The patch comment is about this patch, not the other proposals.
Please do not take that as a snippy response. There were several
emails in the previous thread that discussed memory. In that
thread I responded as to how I would address the concerns. If
anyone wants to raise concerns about memory usage as a result of
this version of the patch they should do so in this current thread.
> Other nits below.
>> +static void of_populate_phandle_cache(void)
>> + unsigned long flags;
>> + phandle max_phandle;
>> + u32 nodes = 0;
>> + struct device_node *np;
>> + if (phandle_cache)
>> + return;
> What's the point of that check?
Sanity check to make sure a memory leak of a previous cache
does not occur. I'll change it to free the cache if it
There is only one instance of of_populate_cache() being called,
so this is a theoretical issue. I intend to add another caller
in the devicetree overlay code in the future, but do not want
to do that now, to avoid a conflict with the overlay patch
series that has been in parallel development, and for which
I submitted v2 shortly after this set of patches.
> And shouldn't it be done inside the
> spinlock if at all?
Not an issue yet, but I'll keep my eye on the possibility of races
when I add a call to of_populate_cache() from the overlay code.
>> + max_phandle = live_tree_max_phandle();
>> + raw_spin_lock_irqsave(&devtree_lock, flags);
>> + for_each_of_allnodes(np)
>> + nodes++;
> Why not save a walk over all nodes and a spin_lock/unlock pair by
> combining the node count with the max_phandle computation? But you've
> just moved the existing live_tree_max_phandle, so probably better as a
> followup patch.
I'll consider adding the node counting into live_tree_max_phandle() later.
The other user of live_tree_max_phandle() is being modified in my overlay
patch series (see mention above). I don't want to create a conflict between
the two series.
>> + /* sanity cap for malformed tree */
>> + if (max_phandle > nodes)
>> + max_phandle = nodes;
>> + phandle_cache = kzalloc((max_phandle + 1) * sizeof(*phandle_cache),
>> + GFP_ATOMIC);
> Maybe kcalloc. Sure, you've capped max_phandle so there's no real risk
> of overflow.
OK, will do.
>> + for_each_of_allnodes(np)
>> + if (np->phandle != OF_PHANDLE_ILLEGAL &&
>> + np->phandle <= max_phandle &&
>> + np->phandle)
> I'd reverse the order of these conditions so that for all the nodes with
> no phandle we only do the np->phandle check. Also, extra whitespace
> before &&.
>> + phandle_cache[np->phandle] = np;
>> + max_phandle_cache = max_phandle;
>> + raw_spin_unlock_irqrestore(&devtree_lock, flags);