On 9/19/19 6:56 PM, Cleber Rosa wrote: > On Wed, Sep 18, 2019 at 09:14:58PM -0400, Cleber Rosa wrote: >> On Wed, Sep 18, 2019 at 05:16:54PM +1000, David Gibson wrote: >>> Hi, >>> >>> I'm finding make check-acceptance is currently useless for me as a >>> pre-pull test, because a bunch of the tests are not at all reliable. >>> There are a bunch which I'm still investigating, but for now I'm >>> looking at the MIPS Malta SSH tests. >>> >>> There seem to be at least two problems here. First, the test includes >>> a download of a pretty big guest disk image. This can easily exhaust >>> the 2m30 timeout on its own. >>> >> >> You're correct that successes and failures on those tests depend >> largely on bandwith. On a shared environment I used for tests >> the download of those images take roughly 400 seconds, resulting >> in failures. On my own machine, around 60, and the tests pass. >> >> There's a conceptual and conflicting problem in that the environment >> for tests to run should be prepared beforehand. The conflicting >> solutions can be: >> >> * extensive bootstrapping of the test execution environment, such >> as the installation of guests from ISOs or installation trees, or >> the download of "default" images wether the tests will use it or >> not (this is what Avocado-VT does/requires) >> >> * keeping test assets in the tree (Avocado allows this if you have >> a your_test.py.data/ directory), but it's not practical for large >> files or files that can't or shouldn't be redistributed >> >>> Even without the timeout, it makes the test really slow, even on >>> repeated runs. Is there some way we can make the image download part >>> of "building" the tests rather than actually running the testsuite, so >>> that a) the test themselves go faster and b) we don't include the >>> download in the test timeout - obviously the download speed is hugely >>> dependent on factors that aren't really related to what we're testing >>> here. >>> >> >> On Avocado version 72.0 we attempted to minimize the isse by >> implementing a "vmimage" command. So, if you expect to use Fedora 30 >> aarch64 images, you could run before your tests: >> >> $ avocado vmimage get --distro fedora --distro-version 30 --arch aarch64 >> >> And to list the images on your cache: >> >> $ avocado vmimage list >> >> Unfortunately, this test doesn't use the vmimage API. Actually that >> is fine because not all test assets map nicely to the vmimage goal, >> and should keep using the more generic (and lower level) fetch_asset(). >> >> We're now working on various "asset fetcher" improvements that should >> allow us to check/cache all assets before a test is executed. Also, >> we're adding a mode in which the "fetch_asset()" API will default to >> cancel (aka SKIP) a test if the asset could not be downloaded. >> >> If you're interested in the card we're using to track that new feature: >> >> >> https://trello.com/c/T3SC1sZs/1521-implement-fetch-assets-command-line-parameter >> >> Another possibility that we've prototyped, and we'll be working on >> further, is to make a specific part of the "test" code execution >> (really a pre-test phase) to be executed without a timeout and even be >> tried a number of times before bailing out and skipping the test. >> >>> In the meantime, I tried hacking it by just increasing the timeout to >>> 10m. That got several of the tests working for me, but one still >>> failed. Specifically 'LinuxSSH.test_mips_malta32eb_kernel3_2_0' still >>> timed out for me, but now after booting the guest, rather than during >>> the image download. Looking at the avocado log file I'm seeing a >>> bunch of soft lockup messages from the guest console, AFAICT. So it >>> looks like we have a real bug here, which I suspect has been >>> overlooked precisely because the download problems mean this test >>> isn't reliable. >>> >> >> I've schedulled a 100 executions of `make check-acceptance` builds, with >> the linux_ssh_mips_malta.py tests having a 1500 seconds timeout. The >> very first execution already brought interesting results: >> >> ... >> (15/39) >> /home/cleber/src/qemu/tests/acceptance/linux_ssh_mips_malta.py:LinuxSSH.test_mips_malta32eb_kernel3_2_0: >> PASS (198.38 s) >> (16/39) >> /home/cleber/src/qemu/tests/acceptance/linux_ssh_mips_malta.py:LinuxSSH.test_mips_malta64el_kernel3_2_0: >> FAIL: Failure message found in console: Oops (22.83 s) >> >> I'll let you know about my full results. This should also serve as a >> starting point to a discussion about the reliability of other tests, >> as you mentioned before. > > Out of the 100 executions on a ppc64le host, the results that contain > failures and errors: > > 15-/home/cleber/src/qemu/tests/acceptance/linux_ssh_mips_malta.py:LinuxSSH.test_mips_malta32eb_kernel3_2_0 > - PASS: 92 > - INTERRUPTED: 4 > - FAIL: 4 > 16-/home/cleber/src/qemu/tests/acceptance/linux_ssh_mips_malta.py:LinuxSSH.test_mips_malta64el_kernel3_2_0 > - PASS: 95 > - FAIL: 5 > > FAIL means that self.fail() was called, which means 'Oops' was found > in the console. INTERRUPTED here means that the test timeout kicked > in, and I can back David's statements about soft lockups. > > Let me know if anyone wants access to the full logs/results.
Can you check if the FAIL case are this bug please? https://bugs.launchpad.net/qemu/+bug/1833661 Thanks, Phil.