On Thu, 23 Aug 2018 at 05:54, Oleksandr Terentiev <otere...@cisco.com> wrote:
> Thank you Anibal for the fast response > > On 22.08.18 19:50, Anibal Limon wrote: > > > > On Wed, 22 Aug 2018 at 11:39, Oleksandr Terentiev <otere...@cisco.com> > wrote: > >> Hi, >> >> I launched util-linux ptest using automated/linux/ptest/ptest.yaml from >> https://git.linaro.org/qa/test-definitions.git and received the >> following results: >> https://pastebin.com/nj9PYQzE >> >> As you can see some tests failed. However, case util-linux marked as >> passed. It looks like ptest.py only analyze return code of ptest-runner >> -d <ptest_dir> <ptest_name> command. And since ptest-runner finishes >> correctly exit code is 0. Therefore all tests are always marked as >> passed, and users never know when some of the tests fail. >> >> Maybe it worth to analyze each test? >> > > Talking about each ptest the result comes from the ptest script in the OE > recipe [1], for convention if the OE ptest returns 0 means pass, so > needs to be fixed in the OE ptest [2]. > > I’ve read https://wiki.yoctoproject.org/wiki/Ptest carefully a few times > more. There are prescriptions about output format. But I didn’t find any > mention about return code processing or a reference to the convention you > mentioned in the answer. > > I looked through some OE run-ptest scripts. I suspect they don’t verify if > some of their tests failed, and exit with 0 even if all their tests failed. > > > http://git.openembedded.org/openembedded-core/tree/meta/recipes-core/util-linux/util-linux/run-ptest > > http://git.openembedded.org/openembedded-core/tree/meta/recipes-support/attr/acl/run-ptest > > http://git.openembedded.org/openembedded-core/tree/meta/recipes-support/attr/files/run-ptest > > http://git.openembedded.org/openembedded-core/tree/meta/recipes-core/dbus/dbus/run-ptest > > http://git.openembedded.org/openembedded-core/tree/meta/recipes-devtools/e2fsprogs/e2fsprogs/run-ptest > > http://git.openembedded.org/openembedded-core/tree/meta/recipes-extended/gawk/gawk/run-ptest > > Right, looks that the OEQA test case was update since i worked on it [1], so now it takes into account the pass/fail of every ptest. So the ptest.py needs to implement the same behavior. Regards, Anibal [1] http://git.openembedded.org/openembedded-core/tree/meta/lib/oeqa/runtime/cases/ptest.py#n80 > > Regarding the LAVA ptest.py script, I mark the run as succeed if there is > no critical error in the ptest-runner and we have a QA-reports tool to > analyse pass/fails > in detail for every ptest executed [3]. > > I heard about QA-reports tool but I’ve never used it before, so maybe I > missed something. > From > https://qa-reports.linaro.org/qcomlt/openembedded-rpb-sumo/build/37/testrun/1890442/suite/linux-ptest/tests/ > I see all ptests passed. Still, in log > > https://qa-reports.linaro.org/qcomlt/openembedded-rpb-sumo/build/37/testrun/1890442/log > I found 54 failed tests and wasn’t able to find a report which indicates > those failures. > > Is there such a report? It would be really useful to know that some tests > failed. > > Thanks > > > > [1] > http://git.openembedded.org/openembedded-core/tree/meta/recipes-core/util-linux/util-linux/run-ptest > [2] https://wiki.yoctoproject.org/wiki/Ptest > [3] > https://qa-reports.linaro.org/qcomlt/openembedded-rpb-sumo/build/37/testrun/1890442/ > > Regards, > Anibal > >> >> Best regards, >> Alex >> >> >
_______________________________________________ linaro-validation mailing list linaro-validation@lists.linaro.org https://lists.linaro.org/mailman/listinfo/linaro-validation