On 03/05/2018 12:41 PM, Aaron Lu wrote:
> On Fri, Mar 02, 2018 at 06:55:25PM +0100, Vlastimil Babka wrote:
>> On 03/01/2018 03:00 PM, Michal Hocko wrote:
>>>
>>> I am really surprised that this has such a big impact.
>>
>> It's even stranger to me. Struct page is 64 bytes these days, exactly a
>> a cache line. Unless that changed, Intel CPUs prefetched a "buddy" cache
>> line (that forms an aligned 128 bytes block with the one we touch).
>> Which is exactly a order-0 buddy struct page! Maybe that implicit
>> prefetching stopped at L2 and explicit goes all the way to L1, can't
> 
> The Intel Architecture Optimization Manual section 7.3.2 says:
> 
> prefetchT0 - fetch data into all cache levels
> Intel Xeon Processors based on Nehalem, Westmere, Sandy Bridge and newer
> microarchitectures: 1st, 2nd and 3rd level cache.
> 
> prefetchT2 - fetch data into 2nd and 3rd level caches (identical to
> prefetchT1)
> Intel Xeon Processors based on Nehalem, Westmere, Sandy Bridge and newer
> microarchitectures: 2nd and 3rd level cache.
> 
> prefetchNTA - fetch data into non-temporal cache close to the processor,
> minimizing cache pollution
> Intel Xeon Processors based on Nehalem, Westmere, Sandy Bridge and newer
> microarchitectures: must fetch into 3rd level cache with fast replacement.
> 
> I tried 'prefetcht0' and 'prefetcht2' instead of the default
> 'prefetchNTA' on a 2 sockets Intel Skylake, the two ended up with about
> the same performance number as prefetchNTA. I had expected prefetchT0 to
> deliver a better score if it was indeed due to L1D since prefetchT2 will
> not place data into L1 while prefetchT0 will, but looks like it is not
> the case here.
> 
> It feels more like the buddy cacheline isn't in any level of the caches
> without prefetch for some reason.

So the adjacent line prefetch might be disabled? Could you check bios or
the MSR mentioned in
https://software.intel.com/en-us/articles/disclosure-of-hw-prefetcher-control-on-some-intel-processors

>> remember. Would that make such a difference? It would be nice to do some
>> perf tests with cache counters to see what is really going on...
> 
> Compare prefetchT2 to no-prefetch, I saw these metrics change:
> 
> no-prefetch          change  prefetchT2       metrics
>             \                          \
>          stddev                     stddev
> ------------------------------------------------------------------------
>       0.18            +0.0        0.18        perf-stat.branch-miss-rate%     
>                                   
>  8.268e+09            +3.8%  8.585e+09        perf-stat.branch-misses         
>                       
>  2.333e+10            +4.7%  2.443e+10        perf-stat.cache-misses          
>                                   
>  2.402e+11            +5.0%  2.522e+11        perf-stat.cache-references      
>                               
>       3.52            -1.1%       3.48        perf-stat.cpi                   
>                                   
>       0.02            -0.0        0.01 ±3%    perf-stat.dTLB-load-miss-rate%  
>                          
>  8.677e+08            -7.3%  8.048e+08 ±3%    perf-stat.dTLB-load-misses      
>                                   
>       1.18            +0.0        1.19        perf-stat.dTLB-store-miss-rate% 
>             
>  2.359e+10            +6.0%  2.502e+10        perf-stat.dTLB-store-misses     
>                                   
>  1.979e+12            +5.0%  2.078e+12        perf-stat.dTLB-stores           
>           
>  6.126e+09           +10.1%  6.745e+09 ±3%    perf-stat.iTLB-load-misses      
>                                   
>       3464            -8.4%       3172 ±3%    
> perf-stat.instructions-per-iTLB-miss            
>       0.28            +1.1%       0.29        perf-stat.ipc                   
>                                   
>  2.929e+09            +5.1%  3.077e+09        perf-stat.minor-faults          
>                
>  9.244e+09            +4.7%  9.681e+09        perf-stat.node-loads            
>                                   
>  2.491e+08            +5.8%  2.634e+08        perf-stat.node-store-misses     
>           
>  6.472e+09            +6.1%  6.869e+09        perf-stat.node-stores           
>                                   
>  2.929e+09            +5.1%  3.077e+09        perf-stat.page-faults           
>                
>    2182469            -4.2%    2090977        perf-stat.path-length
> 
> Not sure if this is useful though...

Looks like most stats increased in absolute values as the work done
increased and this is a time-limited benchmark? Although number of
instructions (calculated from itlb misses and insns-per-itlb-miss) shows
less than 1% increase, so dunno. And the improvement comes from reduced
dTLB-load-misses? That makes no sense for order-0 buddy struct pages
which always share a page. And the memmap mapping should use huge pages.
BTW what is path-length?

Reply via email to