https://bugs.documentfoundation.org/show_bug.cgi?id=152587
V Stuart Foote <[email protected]> changed: What |Removed |Added ---------------------------------------------------------------------------- CC| |[email protected], | |[email protected], | |[email protected] --- Comment #6 from V Stuart Foote <[email protected]> --- (In reply to Alex Thurgood from comment #5) > @Stuart : should I rerun the tests on macOS M1 with the software rendering > option ticked as well / instead ? Probably, but remember these are testing the rendering results to a virtual vcl canavs for simple sample graphics--comparing bitmap against "expected" result. The "pass", "quirky", "fail" or "not tested" would either match or not match between the two rendering modes, and may or may not be the same for GPU drivers for the the M1 silicon vs. Intel GPUs or discrete GPUs. As I don't drive macOS and at present only have Intel based macMini to work with I can not compare the results for Skia and non-Skia render paths on M1 silicon. But I would think dev comment quikee and Luboš, or even Akshit since his GSOC 2021 contribution [1][2], on the intended use of the tests-- both triggered during launch, and alternatively run ad-hoc via the "Run Graphics Tests" dialog-- as to what extent they affect GPU configuration. Of the 109 tests, what significance to assign to the "not run", "quirky" and "fail" status, and more importantly how to assess a difference between the render engines. And also, would there be some methodology (similar to the BreakPad modeled crash report uploads) to centrally upload the graphics test results and GPU/system details for devs to analyze. =-ref-= [1] https://blog.documentfoundation.org/blog/2021/10/19/libreoffice-and-google-summer-of-code-2021-the-results/ [2] https://theproglevblog.blogspot.com/2021/08/google-summer-of-code-2021-project.html -- You are receiving this mail because: You are the assignee for the bug.
