When comparing two test results the current behavior sets a status of skip for tests that were not run in one set of results. This is confusing and silly, so instead give them a status of N/A which gives a clear understanding of what skipped, and what simply didn't run either by exclusion or because it didn't exist yet.
V3: Change N/A to "Not Run" Signed-off-by: Dylan Baker <baker.dyla...@gmail.com> --- framework/summary.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/framework/summary.py b/framework/summary.py index fc08d2b..3fc03e7 100644 --- a/framework/summary.py +++ b/framework/summary.py @@ -282,7 +282,7 @@ class Result: self.options = result['options'] self.glxinfo = result['glxinfo'] self.lspci = result['lspci'] - self.time = result.get('time_elapsed', 'skip') + self.time = result.get('time_elapsed', 'Not Run') self.name = result['name'] self.tests = result['tests'] @@ -418,7 +418,7 @@ class HTMLIndex(list): self._testResult(each.name, key, each.tests[key]['result']) except KeyError: self.append({'type': 'other', - 'text': '<td class="skip">skip</td>'}) + 'text': '<td class="skip">Not Run</td>'}) self._endRow() def _newRow(self): -- 1.8.1.4 _______________________________________________ Piglit mailing list Piglit@lists.freedesktop.org http://lists.freedesktop.org/mailman/listinfo/piglit