yes. I suspect this is the problem. In the host I can see in the yum log the times the vdsm packages are installed: [dron@dron post-002_bootstrap.py]$ less lago-basic-suite-4-3-host-1/_var_log/yum.log |grep vdsm Mar 25 07:48:49 Installed: vdsm-api-4.30.11-23.git276b602.el7.noarch Mar 25 07:48:52 Installed: vdsm-yajsonrpc-4.30.11-23.git276b602.el7.noarch Mar 25 07:49:00 Installed: vdsm-common-4.30.11-23.git276b602.el7.noarch Mar 25 07:49:00 Installed: vdsm-hook-vhostmd-4.30.11-23.git276b602.el7.noarch Mar 25 07:49:38 Installed: vdsm-network-4.30.11-23.git276b602.el7.x86_64 Mar 25 07:49:38 Installed: vdsm-python-4.30.11-23.git276b602.el7.noarch Mar 25 07:49:39 Installed: vdsm-client-4.30.11-23.git276b602.el7.noarch Mar 25 07:49:39 Installed: vdsm-jsonrpc-4.30.11-23.git276b602.el7.noarch Mar 25 07:49:39 Installed: vdsm-http-4.30.11-23.git276b602.el7.noarch Mar 25 07:49:42 Installed: vdsm-hook-openstacknet-4.30.11-23.git276b602.el7.noarch Mar 25 07:49:57 Installed: vdsm-hook-vmfex-dev-4.30.11-23.git276b602.el7.noarch Mar 25 07:49:58 Installed: vdsm-4.30.11-23.git276b602.el7.x86_64 Mar 25 07:49:58 Installed: vdsm-hook-ethtool-options-4.30.11-23.git276b602.el7.noarch Mar 25 07:49:58 Installed: vdsm-hook-fcoe-4.30.11-23.git276b602.el7.noarch
and in the engine you can see that the engine already failed installation: 2019-03-25 07:48:33,920-04 INFO [org.ovirt.engine.core.bll.hostdeploy.AddVdsCommand] (default task-1) [6baa2a84-b322-42fd-855d-b25272a19a23] Running command: AddVdsCommand internal: false. Entities affected : ID: 319fcdf5-5941-46b1-a75c-d335bf500f82 Type: ClusterAction group CREATE_HOST with role type ADMIN On Mon, Mar 25, 2019 at 1:35 PM Dafna Ron <[email protected]> wrote: > I am also adding didi > should this not be logged in host-deploy? > > > On Mon, Mar 25, 2019 at 1:33 PM Marcin Sobczyk <[email protected]> > wrote: > >> Indeed in my failed job, I can see, that 'vdsm-python' is installed >> *after* engine's attempt to run 'vdsm-tool': >> >> lago-basic-suite-4-3-host-1/_var_log/yum.log >> 403:Mar 22 17:54:12 Installed: vdsm-python-4.30.11-23.git276b602.el7.noarch >> >> lago-basic-suite-4-3-engine/_var_log/ovirt-engine/engine.log >> 1580:2019-03-22 17:53:03,793-04 DEBUG >> [org.ovirt.engine.core.uutils.ssh.SSHClient] (default task-1) >> [d86d9b56-2564-44a9-a8d4-01a4d7f5a529] Executing: '/usr/bin/vdsm-tool >> vdsm-id' >> 1581:2019-03-22 17:53:04,043-04 WARN >> [org.ovirt.engine.core.bll.hostdeploy.AddVdsCommand] (default task-1) >> [d86d9b56-2564-44a9-a8d4-01a4d7f5a529] Failed to initiate vdsm-id request on >> host: Command returned failure code 127 during SSH session >> 'root@lago-basic-suite-4-3-host-1' >> >> >> On 3/25/19 2:22 PM, Dafna Ron wrote: >> >> and its only failing on one host as well. >> >> >> On Mon, Mar 25, 2019 at 1:17 PM Marcin Sobczyk <[email protected]> >> wrote: >> >>> Hmmm, that would suggest, that 'vdsm-tool' command is not found on the >>> host. There was no work done around 'vdsm-tool's 'vdsm-id' recently. >>> On 3/25/19 2:08 PM, Dafna Ron wrote: >>> >>> Actually, I can see in engine that it is trying to start vdsm >>> installation on the host: >>> >>> 2019-03-25 07:48:33,123-04 DEBUG >>> [org.ovirt.engine.core.aaa.filters.SsoRestApiAuthFilter] (default task-1) >>> [] Entered SsoRestApiAuthFilter >>> 2019-03-25 07:48:33,123-04 DEBUG >>> [org.ovirt.engine.core.aaa.filters.SsoRestApiAuthFilter] (default task-1) >>> [] SsoRestApiAuthFilter authenticating with sso >>> 2019-03-25 07:48:33,123-04 DEBUG >>> [org.ovirt.engine.core.aaa.filters.SsoRestApiAuthFilter] (default task-1) >>> [] SsoRestApiAuthFilter authenticating using BEARER header >>> 2019-03-25 07:48:33,123-04 DEBUG >>> [org.ovirt.engine.core.aaa.filters.SsoRestApiAuthFilter] (default task-1) >>> [] SsoRestApiAuthFilter successfully authenticated using BEARER header >>> 2019-03-25 07:48:33,123-04 DEBUG >>> [org.ovirt.engine.core.aaa.filters.SsoRestApiNegotiationFilter] (default >>> task-1) [] Entered SsoRestApiNegotiationFilter >>> 2019-03-25 07:48:33,124-04 DEBUG >>> [org.ovirt.engine.core.aaa.filters.SsoRestApiNegotiationFilter] (default >>> task-1) [] SsoRestApiNegotiationFilter Not performing Negotiate Auth >>> 2019-03-25 07:48:33,143-04 DEBUG >>> [org.ovirt.engine.core.dal.dbbroker.PostgresDbEngineDialect$PostgresSimpleJdbcCall] >>> (default task-1) [6baa2a84-b322-42fd-855d-b25272a19a23] Compiled stored >>> procedure. Call string is [{call getclusterforuserbyclustername(?, ?, ?)}] >>> 2019-03-25 07:48:33,143-04 DEBUG >>> [org.ovirt.engine.core.dal.dbbroker.PostgresDbEngineDialect$PostgresSimpleJdbcCall] >>> (default task-1) [6baa2a84-b322-42fd-855d-b25272a19a23] SqlCall for >>> procedure [GetClusterForUserByClusterName] compiled >>> 2019-03-25 07:48:33,145-04 DEBUG >>> [org.ovirt.engine.core.common.di.interceptor.DebugLoggingInterceptor] >>> (default task-1) [6baa2a84-b322-42fd-855d-b25272a19a23] method: runQuery, >>> params: [GetClusterByName, NameQueryParameters:{refresh='false', >>> filtered='false'}], timeElapsed: 7ms >>> 2019-03-25 07:48:33,161-04 DEBUG >>> [org.ovirt.engine.core.dal.dbbroker.PostgresDbEngineDialect$PostgresSimpleJdbcCall] >>> (default task-1) [] Compiled stored procedure. Call string is [{call >>> getvmstaticbyname(?)}] >>> 2019-03-25 07:48:33,161-04 DEBUG >>> [org.ovirt.engine.core.dal.dbbroker.PostgresDbEngineDialect$PostgresSimpleJdbcCall] >>> (default task-1) [] SqlCall for procedure [GetVmStaticByName] compiled >>> 2019-03-25 07:48:33,206-04 DEBUG >>> [org.ovirt.engine.core.dal.dbbroker.PostgresDbEngineDialect$PostgresSimpleJdbcCall] >>> (default task-1) [6baa2a84-b322-42fd-855d-b25272a19a23] Compiled stored >>> procedure. Call string is [{call getvdsbyname(?)}] >>> 2019-03-25 07:48:33,206-04 DEBUG >>> [org.ovirt.engine.core.dal.dbbroker.PostgresDbEngineDialect$PostgresSimpleJdbcCall] >>> (default task-1) [6baa2a84-b322-42fd-855d-b25272a19a23] SqlCall for >>> procedure [GetVdsByName] compiled >>> 2019-03-25 07:48:33,223-04 DEBUG >>> [org.ovirt.engine.core.common.di.interceptor.DebugLoggingInterceptor] >>> (default task-1) [6baa2a84-b322-42fd-855d-b25272a19a23] method: getByName, >>> params: [lago-basic-suite-4-3-host-1], timeElapsed: 20ms >>> 2019-03-25 07:48:33,225-04 DEBUG >>> [org.ovirt.engine.core.dal.dbbroker.PostgresDbEngineDialect$PostgresSimpleJdbcCall] >>> (default task-1) [6baa2a84-b322-42fd-855d-b25272a19a23] Compiled stored >>> procedure. Call string is [{call getvdsbyhostname(?)}] >>> 2019-03-25 07:48:33,225-04 DEBUG >>> [org.ovirt.engine.core.dal.dbbroker.PostgresDbEngineDialect$PostgresSimpleJdbcCall] >>> (default task-1) [6baa2a84-b322-42fd-855d-b25272a19a23] SqlCall for >>> procedure [GetVdsByHostName] compiled >>> 2019-03-25 07:48:33,252-04 DEBUG >>> [org.ovirt.engine.core.common.di.interceptor.DebugLoggingInterceptor] >>> (default task-1) [6baa2a84-b322-42fd-855d-b25272a19a23] method: >>> getAllForHostname, params: [lago-basic-suite-4-3-host-1], timeElapsed: 28ms >>> 2019-03-25 07:48:33,254-04 DEBUG >>> [org.ovirt.engine.core.dal.dbbroker.PostgresDbEngineDialect$PostgresSimpleJdbcCall] >>> (default task-1) [6baa2a84-b322-42fd-855d-b25272a19a23] Compiled stored >>> procedure. Call string is [{call getstorage_poolsbyclusterid(?)}] >>> 2019-03-25 07:48:33,254-04 DEBUG >>> [org.ovirt.engine.core.dal.dbbroker.PostgresDbEngineDialect$PostgresSimpleJdbcCall] >>> (default task-1) [6baa2a84-b322-42fd-855d-b25272a19a23] SqlCall for >>> procedure [Getstorage_poolsByClusterId] compiled >>> 2019-03-25 07:48:33,262-04 DEBUG >>> [org.ovirt.engine.core.uutils.ssh.SSHClient] (default task-1) >>> [6baa2a84-b322-42fd-855d-b25272a19a23] Connecting >>> 'root@lago-basic-suite-4-3-host-1' >>> 2019-03-25 07:48:33,461-04 DEBUG >>> [org.ovirt.engine.core.uutils.ssh.SSHClient] (default task-1) >>> [6baa2a84-b322-42fd-855d-b25272a19a23] Connected: >>> 'root@lago-basic-suite-4-3-host-1' >>> 2019-03-25 07:48:33,467-04 DEBUG >>> [org.ovirt.engine.core.uutils.ssh.SSHClient] (default task-1) >>> [6baa2a84-b322-42fd-855d-b25272a19a23] Authenticating: >>> 'root@lago-basic-suite-4-3-host-1' >>> 2019-03-25 07:48:33,599-04 DEBUG >>> [org.ovirt.engine.core.uutils.ssh.SSHClient] (default task-1) >>> [6baa2a84-b322-42fd-855d-b25272a19a23] Authenticated: >>> 'root@lago-basic-suite-4-3-host-1' >>> 2019-03-25 07:48:33,600-04 DEBUG >>> [org.ovirt.engine.core.uutils.ssh.SSHClient] (default task-1) >>> [6baa2a84-b322-42fd-855d-b25272a19a23] Executing: '/usr/bin/vdsm-tool >>> vdsm-id' >>> 2019-03-25 07:48:33,902-04 WARN >>> [org.ovirt.engine.core.bll.hostdeploy.AddVdsCommand] (default task-1) >>> [6baa2a84-b322-42fd-855d-b25272a19a23] Failed to initiate vdsm-id request >>> on host: Command returned failure code 127 during SSH session >>> 'root@lago-basic-suite-4-3-host-1' >>> 2019-03-25 07:48:33,920-04 INFO >>> [org.ovirt.engine.core.bll.hostdeploy.AddVdsCommand] (default task-1) >>> [6baa2a84-b322-42fd-855d-b25272a19a23] Running command: AddVdsCommand >>> internal: false. Entities affected : ID: >>> 319fcdf5-5941-46b1-a75c-d335bf500f82 Type: ClusterAction group CREATE_HOST >>> with role type ADMIN >>> >>> >>> >>> >>> On Mon, Mar 25, 2019 at 12:53 PM Dafna Ron <[email protected]> wrote: >>> >>>> we have another failing patch with the same test: >>>> http://jenkins.ovirt.org/job/ovirt-4.3_change-queue-tester/360/ >>>> >>>> its obviously not related to the patch but there is something that is >>>> causing these failures randomly. >>>> From what I can see in the current failed job, you are correct and >>>> host-1 and engine does not even try to deploy host-1. >>>> >>>> i can see host-1 is getting error 127 in lago for ntpd and I can see >>>> network manager errors in the host messages log >>>> >>>> In the host messages log I can see several messages that I suspect may >>>> cause issues in communication between engine and host: >>>> >>>> Mar 25 07:50:09 lago-basic-suite-4-3-host-1 sasldblistusers2: >>>> _sasldb_getkeyhandle has failed >>>> Mar 25 07:50:10 lago-basic-suite-4-3-host-1 saslpasswd2: error deleting >>>> entry from sasldb: BDB0073 DB_NOTFOUND: No matching key/data pair found >>>> Mar 25 07:50:10 lago-basic-suite-4-3-host-1 saslpasswd2: error deleting >>>> entry from sasldb: BDB0073 DB_NOTFOUND: No matching key/data pair found >>>> Mar 25 07:50:10 lago-basic-suite-4-3-host-1 saslpasswd2: error deleting >>>> entry from sasldb: BDB0073 DB_NOTFOUND: No matching key/data pair found >>>> >>>> >>>> ces/12) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bonding: >>>> bondscan-UMJa2S is being created... >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bonding: >>>> bondscan-UMJa2S is being deleted... >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 NetworkManager[2658]: >>>> <info> [1553514613.7774] manager: (bondscan-UMJa2S): new Bond device >>>> (/org/freedesktop/NetworkManager/Devices/13) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: (unregistered >>>> net_device) (unregistering): Released all slaves >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bonding: >>>> bondscan-liwvMR is being created... >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bondscan-liwvMR: >>>> option fail_over_mac: invalid value (3) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bondscan-liwvMR: >>>> option ad_select: invalid value (3) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bondscan-liwvMR: >>>> option lacp_rate: mode dependency failed, not supported in mode >>>> balance-rr(0) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bondscan-liwvMR: >>>> option arp_validate: invalid value (7) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bondscan-liwvMR: >>>> option xmit_hash_policy: invalid value (5) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 NetworkManager[2658]: >>>> <info> [1553514613.8002] manager: (bondscan-liwvMR): new Bond device >>>> (/org/freedesktop/NetworkManager/Devices/14) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bondscan-liwvMR: >>>> option primary_reselect: invalid value (3) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bondscan-liwvMR: >>>> option arp_all_targets: invalid value (2) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bonding: >>>> bondscan-liwvMR is being deleted... >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: (unregistered >>>> net_device) (unregistering): Released all slaves >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bonding: >>>> bondscan-liwvMR is being created... >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bondscan-liwvMR: >>>> option fail_over_mac: invalid value (3) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bondscan-liwvMR: >>>> option ad_select: invalid value (3) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bondscan-liwvMR: >>>> option lacp_rate: mode dependency failed, not supported in mode >>>> active-backup(1) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bondscan-liwvMR: >>>> option arp_validate: invalid value (7) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 NetworkManager[2658]: >>>> <info> [1553514613.8429] manager: (bondscan-liwvMR): new Bond device >>>> (/org/freedesktop/NetworkManager/Devices/15) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bondscan-liwvMR: >>>> option xmit_hash_policy: invalid value (5) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bondscan-liwvMR: >>>> option primary_reselect: invalid value (3) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bondscan-liwvMR: >>>> option arp_all_targets: invalid value (2) >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: bonding: >>>> bondscan-liwvMR is being deleted... >>>> Mar 25 07:50:13 lago-basic-suite-4-3-host-1 kernel: (unregistered >>>> net_device) (unregistering): Released all slaves >>>> >>>> >>>> On Mon, Mar 25, 2019 at 11:24 AM Marcin Sobczyk <[email protected]> >>>> wrote: >>>> >>>>> For the failed job, the engine didn't even try to deploy on host-1: >>>>> >>>>> >>>>> https://jenkins.ovirt.org/job/ovirt-4.3_change-queue-tester/339/artifact/basic-suite.el7.x86_64/test_logs/basic-suite-4.3/post-002_bootstrap.py/lago-basic-suite-4-3-engine/_var_log/ovirt-engine/host-deploy/ >>>>> >>>>> Martin, do you know what could be the reason for that? >>>>> >>>>> I can see in the logs for both successful and unsuccessful >>>>> basic-suite-4.3 runs, that there is no 'ntpdate' on host-1: >>>>> >>>>> 2019-03-25 10:14:46,350::ssh.py::ssh::58::lago.ssh::DEBUG::Running >>>>> d0c49b54 on lago-basic-suite-4-3-host-1: ntpdate -4 >>>>> lago-basic-suite-4-3-engine >>>>> 2019-03-25 10:14:46,383::ssh.py::ssh::81::lago.ssh::DEBUG::Command >>>>> d0c49b54 on lago-basic-suite-4-3-host-1 returned with 127 >>>>> 2019-03-25 10:14:46,384::ssh.py::ssh::96::lago.ssh::DEBUG::Command >>>>> d0c49b54 on lago-basic-suite-4-3-host-1 errors: >>>>> bash: ntpdate: command not found >>>>> >>>>> On host-0 everything is ok: >>>>> >>>>> 2019-03-25 10:14:46,917::ssh.py::ssh::58::lago.ssh::DEBUG::Running >>>>> d11b2a64 on lago-basic-suite-4-3-host-0: ntpdate -4 >>>>> lago-basic-suite-4-3-engine >>>>> 2019-03-25 10:14:53,088::ssh.py::ssh::81::lago.ssh::DEBUG::Command >>>>> d11b2a64 on lago-basic-suite-4-3-host-0 returned with 0 >>>>> 2019-03-25 10:14:53,088::ssh.py::ssh::89::lago.ssh::DEBUG::Command >>>>> d11b2a64 on lago-basic-suite-4-3-host-0 output: >>>>> 25 Mar 06:14:53 ntpdate[6646]: adjust time server 192.168.202.2 offset >>>>> 0.017150 sec >>>>> >>>>> On 3/25/19 10:13 AM, Eyal Edri wrote: >>>>> >>>>> Still fails, now on a different component. ( ovirt-web-ui-extentions ) >>>>> >>>>> https://jenkins.ovirt.org/job/ovirt-4.3_change-queue-tester/339/ >>>>> >>>>> On Fri, Mar 22, 2019 at 3:59 PM Dan Kenigsberg <[email protected]> >>>>> wrote: >>>>> >>>>>> >>>>>> >>>>>> On Fri, Mar 22, 2019 at 3:21 PM Marcin Sobczyk <[email protected]> >>>>>> wrote: >>>>>> >>>>>>> Dafna, >>>>>>> >>>>>>> in 'verify_add_hosts' we specifically wait for single host to be up >>>>>>> with a timeout: >>>>>>> >>>>>>> 144 up_hosts = hosts_service.list(search='datacenter={} AND >>>>>>> status=up'.format(DC_NAME)) >>>>>>> 145 if len(up_hosts): >>>>>>> 146 return True >>>>>>> >>>>>>> The log files say, that it took ~50 secs for one of the hosts to be >>>>>>> up (seems reasonable) and no timeout is being reported. >>>>>>> Just after running 'verify_add_hosts', we run >>>>>>> 'add_master_storage_domain', which calls '_hosts_in_dc' function. >>>>>>> That function does the exact same check, but it fails: >>>>>>> >>>>>>> 113 hosts = hosts_service.list(search='datacenter={} AND >>>>>>> status=up'.format(dc_name)) >>>>>>> 114 if hosts: >>>>>>> 115 if random_host: >>>>>>> 116 return random.choice(hosts) >>>>>>> >>>>>>> I don't think it is relevant to our current failure; but I consider >>>>>> random_host=True as a bad practice. As if we do not have enough moving >>>>>> parts, we are adding intentional randomness. Reproducibility is far more >>>>>> important than coverage - particularly for a shared system test like OST. >>>>>> >>>>>>> 117 else: >>>>>>> 118 return sorted(hosts, key=lambda host: host.name) >>>>>>> 119 raise RuntimeError('Could not find hosts that are up in DC %s' >>>>>>> % dc_name) >>>>>>> >>>>>>> >>>>>>> I'm also not able to reproduce this issue locally on my server. The >>>>>>> investigation continues... >>>>>>> >>>>>> >>>>>> I think that it would be fair to take the filtering by host state out >>>>>> of Engine and into the test, where we can easily log the current status >>>>>> of >>>>>> each host. Then we'd have better understanding on the next failure. >>>>>> >>>>>> On 3/22/19 1:17 PM, Marcin Sobczyk wrote: >>>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> sure, I'm on it - it's weird though, I did ran 4.3 basic suite for >>>>>>> this patch manually and everything was ok. >>>>>>> On 3/22/19 1:05 PM, Dafna Ron wrote: >>>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> We are failing branch 4.3 for test: >>>>>>> 002_bootstrap.add_master_storage_domain >>>>>>> >>>>>>> It seems that in one of the hosts, the vdsm is not starting >>>>>>> there is nothing in vdsm.log or in supervdsm.log >>>>>>> >>>>>>> CQ identified this patch as the suspected root cause: >>>>>>> >>>>>>> https://gerrit.ovirt.org/#/c/98748/ - vdsm: client: Add support for >>>>>>> flow id >>>>>>> >>>>>>> Milan, Marcin, can you please have a look? >>>>>>> >>>>>>> full logs: >>>>>>> >>>>>>> >>>>>>> http://jenkins.ovirt.org/job/ovirt-4.3_change-queue-tester/326/artifact/basic-suite.el7.x86_64/test_logs/basic-suite-4.3/post-002_bootstrap.py/ >>>>>>> >>>>>>> the only error I can see is about host not being up (makes sense as >>>>>>> vdsm is not running) >>>>>>> >>>>>>> Stacktrace >>>>>>> >>>>>>> File "/usr/lib64/python2.7/unittest/case.py", line 369, in run >>>>>>> testMethod() >>>>>>> File "/usr/lib/python2.7/site-packages/nose/case.py", line 197, in >>>>>>> runTest >>>>>>> self.test(*self.arg) >>>>>>> File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line >>>>>>> 142, in wrapped_test >>>>>>> test() >>>>>>> File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line >>>>>>> 60, in wrapper >>>>>>> return func(get_test_prefix(), *args, **kwargs) >>>>>>> File >>>>>>> "/home/jenkins/workspace/ovirt-4.3_change-queue-tester/ovirt-system-tests/basic-suite-4.3/test-scenarios/002_bootstrap.py", >>>>>>> line 417, in add_master_storage_domain >>>>>>> add_iscsi_storage_domain(prefix) >>>>>>> File >>>>>>> "/home/jenkins/workspace/ovirt-4.3_change-queue-tester/ovirt-system-tests/basic-suite-4.3/test-scenarios/002_bootstrap.py", >>>>>>> line 561, in add_iscsi_storage_domain >>>>>>> host=_random_host_from_dc(api, DC_NAME), >>>>>>> File >>>>>>> "/home/jenkins/workspace/ovirt-4.3_change-queue-tester/ovirt-system-tests/basic-suite-4.3/test-scenarios/002_bootstrap.py", >>>>>>> line 122, in _random_host_from_dc >>>>>>> return _hosts_in_dc(api, dc_name, True) >>>>>>> File >>>>>>> "/home/jenkins/workspace/ovirt-4.3_change-queue-tester/ovirt-system-tests/basic-suite-4.3/test-scenarios/002_bootstrap.py", >>>>>>> line 119, in _hosts_in_dc >>>>>>> raise RuntimeError('Could not find hosts that are up in DC %s' % >>>>>>> dc_name) >>>>>>> 'Could not find hosts that are up in DC test-dc\n-------------------- >>>>>>> >> begin captured logging << --------------------\nlago.ssh: DEBUG: >>>>>>> start task:937bdea7-a2a3-47ad-9383-36647ea37ddf:Get ssh client for >>>>>>> lago-basic-suite-4-3-engine:\nlago.ssh: DEBUG: end >>>>>>> task:937bdea7-a2a3-47ad-9383-36647ea37ddf:Get ssh client for >>>>>>> lago-basic-suite-4-3-engine:\nlago.ssh: DEBUG: Running c07b5ee2 on >>>>>>> lago-basic-suite-4-3-engine: cat /root/multipath.txt\nlago.ssh: DEBUG: >>>>>>> Command c07b5ee2 on lago-basic-suite-4-3-engine returned with >>>>>>> 0\nlago.ssh: DEBUG: Command c07b5ee2 on lago-basic-suite-4-3-engine >>>>>>> output:\n >>>>>>> 3600140516f88cafa71243648ea218995\n360014053e28f60001764fed9978ec4b3\n360014059edc777770114a6484891dcf1\n36001405d93d8585a50d43a4ad0bd8d19\n36001405e31361631de14bcf87d43e55a\n\n----------- >>>>>>> >>>>>>> _______________________________________________ >>>>>>> Devel mailing list -- [email protected] >>>>>>> To unsubscribe send an email to [email protected] >>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>> oVirt Code of Conduct: >>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>> List Archives: >>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/J4NCHXTK5ZYLXWW36DZKAUL5DN7WBNW4/ >>>>>>> >>>>>> _______________________________________________ >>>>>> Devel mailing list -- [email protected] >>>>>> To unsubscribe send an email to [email protected] >>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>> oVirt Code of Conduct: >>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>> List Archives: >>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/ULS4OKU2YZFDQT5EDFYKLW5GFA52YZ7U/ >>>>>> >>>>> >>>>> >>>>> -- >>>>> >>>>> Eyal edri >>>>> >>>>> >>>>> MANAGER >>>>> >>>>> RHV/CNV DevOps >>>>> >>>>> EMEA VIRTUALIZATION R&D >>>>> >>>>> >>>>> Red Hat EMEA <https://www.redhat.com/> >>>>> <https://red.ht/sig> TRIED. TESTED. TRUSTED. >>>>> <https://redhat.com/trusted> >>>>> phone: +972-9-7692018 >>>>> irc: eedri (on #tlv #rhev-dev #rhev-integ) >>>>> >>>>> _______________________________________________ >>>>> Devel mailing list -- [email protected] >>>>> To unsubscribe send an email to [email protected] >>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>> oVirt Code of Conduct: >>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>> List Archives: >>>>> https://lists.ovirt.org/archives/list/[email protected]/message/XOP6ZCUIDUVC2XNVBS2X7OAHGOXJZROL/ >>>>> >>>>
_______________________________________________ Devel mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/XXZGAZZOIJTZQ2LZVIGWAFUYKE6TXN36/
