On 12/14/2014 4:59 AM, Chris Wilson wrote:
One of the things wbinvd is considered evil for is that it blocks the
CPU for an indeterminate amount of time - upsetting latency critcial
aspects of the OS. For example, the x86/mm has similar code to use
wbinvd for large clflushes that caused a bit of controversy with RT:
http://linux-kernel.2935.n7.nabble.com/PATCH-x86-Use-clflush-instead-of-wbinvd-whenever-possible-when-changing-mapping-td493751.html
and also the use of wbinvd in the nvidia driver has also been noted as
evil by RT folks.
However as the wbinvd still exists, it can't be all that bad...
Yeah there are definitely tradeoffs here. In this particular case,
we're trying to flush out a ~140M object on every frame, which just
seems silly.
There's definitely room for optimization in Mesa too; avoiding a mapping
that marks a large bo as dirty would be good, but if we improve the
kernel in this area, even sloppy apps and existing binaries will speed up.
Maybe we could apply this only on !llc systems or something? I wonder
how much wbinvd performance varies across microarchitectures; maybe
Thomas's issue isn't really relevant anymore (at least one can hope).
When digging into this, I found that an optimization to remove the IPI
for wbinvd was clobbered during a merge; maybe that should be
resurrected too. Surely a single, global wbinvd is sufficient; we don't
need to do n_cpus^2 wbinvd + the associated invalidation bus signals here...
Alternately, we could insert some delays into this path just to make it
extra clear to userspace that they really shouldn't be hitting this in
the common case (and provide some additional interfaces to let them
avoid it by allowing flushing and dirty management in userspace).
Jesse
_______________________________________________
Intel-gfx mailing list
[email protected]
http://lists.freedesktop.org/mailman/listinfo/intel-gfx