I have now enqueued another run of OST for the next VDSM patch in line (that was merged around the time of the first failure) https://gerrit.ovirt. org/#/c/75299/2 to see if it was the culprit. Not sure why we allow only one (*globally for all developers) *concurrent run of the manual OST job... I've created a jira issue for this https://ovirt-jira.atlassian.net/browse/OST-61
Martin On Mon, Apr 10, 2017 at 3:29 PM, Martin Betak <[email protected]> wrote: > Actually it appears that this failure is not engine related (tried > reverting my patches from saturday but still received similar errors with > lago in the [ add_secondary_storage_domains] test). > But when running the current engine master with VDSM cutoff before the > time of failure, it passed http://jenkins.ovirt.or > g/job/ovirt-system-tests_manual/229/ > > This was the VDSM commit that *worked* > *Commit 5c8aff6177bdaa81ee11c20d417c9bb10e651fb8 by Francesco Romani > <http://jenkins.ovirt.org/user/[email protected]/>* > > On Mon, Apr 10, 2017 at 11:08 AM, Michal Skrivanek < > [email protected]> wrote: > >> >> > On 9 Apr 2017, at 09:42, Barak Korren <[email protected]> wrote: >> > >> > On 9 April 2017 at 10:39, Yaniv Kaul <[email protected]> wrote: >> >> 1. The error: >> >> 2017-04-08 14:12:11,376-04 ERROR >> >> [org.ovirt.vdsm.jsonrpc.client.JsonRpcClient] (ResponseWorker) [] Not >> able >> >> to update response for "fc0cef3f-d8fa-4074-b315-e36cc2f63fa1" >> >> >> >> Is still seen, not sure why. >> >> >> >> 2. MOM is not available (unrelated, but still a bug) >> >> >> >> 3. New error I have not seen in the past on host1 >> >> (http://jenkins.ovirt.org/job/test-repo_ovirt_experimental_m >> aster/6257/artifact/exported-artifacts/basic-suit-master-el7 >> /test_logs/basic-suite-master/post-002_bootstrap.py/lago-bas >> ic-suite-master-host1/_var_log/vdsm/vdsm.log >> >> ) : >> >> >> >> 2017-04-08 14:12:22,208-0400 WARN (jsonrpc/6) [storage.LVM] lvm pvs >> failed: >> >> 5 [] [' Failed to find physical volume >> >> "/dev/mapper/360014050eacbb3c8b21428fb6683f074".'] (lvm:325) >> >> >> >> 4. New warning: >> >> 2017-04-08 14:13:17,327-0400 WARN (check/loop) [storage.asyncutils] >> Call >> >> <bound method DirectioChecker._check of <DirectioChecker >> >> /dev/0eb0f05c-0507-4f02-ba13-b8d865538d7a/metadata running >> >> next_check=4295183.36 at 0x2fb9190>> delayed by 1.43 seconds >> >> (asyncutils:138) >> >> >> >> >> >> Do we know when it began? >> > >> > Yes. I linked to the patch that seems to have started this. >> >> This should be easy to revert >> Working on it >> Then we’ll see what exactly is the culprit >> >> > >> > >> > >> > -- >> > Barak Korren >> > [email protected] >> > RHCE, RHCi, RHV-DevOps Team >> > https://ifireball.wordpress.com/ >> > _______________________________________________ >> > Devel mailing list >> > [email protected] >> > http://lists.ovirt.org/mailman/listinfo/devel >> > >> > >> >> _______________________________________________ >> Devel mailing list >> [email protected] >> http://lists.ovirt.org/mailman/listinfo/devel >> > >
_______________________________________________ Devel mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/devel
