[ovirt-users] Re: major network changes
That's strange...Do you have any kind of proxy there? Best Regards,Strahil Nikolov В четвъртък, 25 юли 2019 г., 17:31:26 ч. Гринуич+3, carl langlois написа: Thanks Strahil, When i run the command it get stuck waiting for the response. About to connect() to ovengine port 443 (#0) * Trying 10.16.248.74... * Connected to ovengine (10.16.248.74) port 443 (#0) * Initializing NSS with certpath: sql:/etc/pki/nssdb * skipping SSL peer certificate verification * SSL connection using TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 > GET /ovirt-engine/services/health HTTP/1.1 > User-Agent: curl/7.29.0 > Host: ovengine > Accept: */* > in the error_log i have always [proxy_ajp:debug] [pid 2164] mod_proxy_ajp.c(276): [client 10.16.248.65:58768] AH00872: APR_BUCKET_IS_EOS [Thu Jul 25 10:26:48.452910 2019] [proxy_ajp:debug] [pid 2164] mod_proxy_ajp.c(282): [client 10.16.248.65:58768] AH00873: data to read (max 8186 at 4) [Thu Jul 25 10:26:48.452916 2019] [proxy_ajp:debug] [pid 2164] mod_proxy_ajp.c(296): [client 10.16.248.65:58768] AH00875: got 0 bytes of data [Thu Jul 25 10:26:49.444631 2019] [proxy:debug] [pid 2771] proxy_util.c(1843): AH00925: initializing worker ajp://127.0.0.1:8702 shared [Thu Jul 25 10:26:49.444686 2019] [proxy:debug] [pid 2771] proxy_util.c(1885): AH00927: initializing worker ajp://127.0.0.1:8702 local [Thu Jul 25 10:26:49.444716 2019] [proxy:debug] [pid 2771] proxy_util.c(1936): AH00931: initialized single connection worker in child 2771 for (127.0.0.1) [Thu Jul 25 10:26:49.444745 2019] [proxy:debug] [pid 2771] proxy_util.c(1843): AH00925: initializing worker proxy:reverse shared [Thu Jul 25 10:26:49.444878 2019] [proxy:debug] [pid 2771] proxy_util.c(1885): AH00927: initializing worker proxy:reverse local [Thu Jul 25 10:26:49.444938 2019] [proxy:debug] [pid 2771] proxy_util.c(1936): AH00931: initialized single connection worker in child 2771 for (*) it keep initializing after the connection after an attempt to connect. Any ideas?Thanks & Regards Carl On Wed, Jul 24, 2019 at 11:00 PM Strahil wrote: The CA can be downloaded via the web , or you can tell curl to just ignore the engine's cert via the '-k' flag. It will show you if the health page is working. Best Regards, Strahil Nikolov On Jul 24, 2019 19:39, carl langlois wrote: Strahil, not sure what to put for the --cacert. Yes Derek your are right at one point the port 8702 stop listening. tcp6 0 0 127.0.0.1:8702 :::* LISTEN 1607/ovirt-engine After some time the line above disappear. I am trying to figure why this port is being close after some time when the engine is running on the host on the 248.x network. On the 236.x network this port is kept alive all the time.If you have any hint on why this port is closing do not hesitate because i am starting to be out of ideas. :-) Thanks & Regards Carl On Wed, Jul 24, 2019 at 11:11 AM Strahil Nikolov wrote: A healthy engine should report:[root@ovirt1 ~]# curl --cacert CA https://engine.localdomain/ovirt-engine/services/health;echoDB Up!Welcome to Health Status! Of course you can use the '-k' switch to verify the situation. Best Regards,Strahil Nikolov В сряда, 24 юли 2019 г., 17:43:59 ч. Гринуич+3, Derek Atkins написа: Hi, carl langlois writes: > If i try to access http://ovengine/ovirt-engine/services/health > i always get "Service Unavailable" in the browser and each time i it reload in > the browser i get in the error_log > > [proxy_ajp:error] [pid 1868] [client 10.8.1.76:63512] AH00896: failed to make > connection to backend: 127.0.0.1 > [Tue Jul 23 14:04:10.074023 2019] [proxy:error] [pid 1416] (111)Connection > refused: AH00957: AJP: attempt to connect to 127.0.0.1:8702 (127.0.0.1) failed Sounds like a service isn't running on port 8702. > Thanks & Regards > > Carl -derek -- Derek Atkins 617-623-3745 de...@ihtfp.com www.ihtfp.com Computer and Internet Security Consultant ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/3FKCB4JKOGO5HYUBFNZLRTXXRUY2BMJR/
[ovirt-users] Re: major network changes
Hi, Thanks you for all your input We have pinpoint the cause of our "liveliness check: failure. In our engine we have a external provider that connect to our domain controller with LDAP request. When we removed this provider the hosted-engine host was able to check for liveliness on the engine. Not sure why the provider is preventing the engine to go live. Any thoughts? Thanks & Regards Carl On Wed, Jul 24, 2019 at 11:00 PM Strahil wrote: > The CA can be downloaded via the web , or you can tell curl to just > ignore the engine's cert via the '-k' flag. > It will show you if the health page is working. > > Best Regards, > Strahil Nikolov > On Jul 24, 2019 19:39, carl langlois wrote: > > Strahil, not sure what to put for the --cacert. > > Yes Derek your are right at one point the port 8702 stop listening. > > tcp6 0 0 127.0.0.1:8702 :::* > LISTEN 1607/ovirt-engine > > After some time the line above disappear. I am trying to figure why this > port is being close after some time when the engine is running on the host > on the 248.x network. On the 236.x network this port is kept alive all the > time. > If you have any hint on why this port is closing do not hesitate because i > am starting to be out of ideas. :-) > > > Thanks & Regards > > Carl > > > > > > > On Wed, Jul 24, 2019 at 11:11 AM Strahil Nikolov > wrote: > > A healthy engine should report: > [root@ovirt1 ~]# curl --cacert CA > https://engine.localdomain/ovirt-engine/services/health;echo > DB Up!Welcome to Health Status! > > Of course you can use the '-k' switch to verify the situation. > > Best Regards, > Strahil Nikolov > > В сряда, 24 юли 2019 г., 17:43:59 ч. Гринуич+3, Derek Atkins < > de...@ihtfp.com> написа: > > > Hi, > > carl langlois writes: > > > If i try to access http://ovengine/ovirt-engine/services/health > > i always get "Service Unavailable" in the browser and each time i it > reload in > > the browser i get in the error_log > > > > [proxy_ajp:error] [pid 1868] [client 10.8.1.76:63512] AH00896: failed > to make > > connection to backend: 127.0.0.1 > > [Tue Jul 23 14:04:10.074023 2019] [proxy:error] [pid 1416] > (111)Connection > > refused: AH00957: AJP: attempt to connect to 127.0.0.1:8702 (127.0.0.1) > failed > > Sounds like a service isn't running on port 8702. > > > Thanks & Regards > > > > Carl > > -derek > > -- > Derek Atkins617-623-3745 > > de...@ihtfp.com > www.ihtfp.com > Computer and Internet Security Consultant > > ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/4Z2RKPLZKS5YXNPIUEGP5ASTCKVDJNS6/
[ovirt-users] Re: major network changes
Hi, carl langlois writes: > Strahil, not sure what to put for the --cacert. > > Yes Derek your are right at one point the port 8702 stop listening. > > tcp6 0 0 127.0.0.1:8702 :::* LISTEN > 1607/ovirt-engine Can you try running 'lsof' to figure out what application has that port open? Then you can figure out why it's dying. > After some time the line above disappear. I am trying to figure why this port > is being close after some time when the engine is running on the host on the > 248.x network. On the 236.x network this port is kept alive all the time. > If you have any hint on why this port is closing do not hesitate because i am > starting to be out of ideas. :-) > > Thanks & Regards > > Carl -derek -- Derek Atkins 617-623-3745 de...@ihtfp.com www.ihtfp.com Computer and Internet Security Consultant ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/PH5NE5FKZXSQKTDCBVJLAQHYTJ2VZWH5/
[ovirt-users] Re: major network changes
Thanks Strahil, When i run the command it get stuck waiting for the response. About to connect() to ovengine port 443 (#0) * Trying 10.16.248.74... * Connected to ovengine (10.16.248.74) port 443 (#0) * Initializing NSS with certpath: sql:/etc/pki/nssdb * skipping SSL peer certificate verification * SSL connection using TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 > GET /ovirt-engine/services/health HTTP/1.1 > User-Agent: curl/7.29.0 > Host: ovengine > Accept: */* > in the error_log i have always [proxy_ajp:debug] [pid 2164] mod_proxy_ajp.c(276): [client 10.16.248.65:58768] AH00872: APR_BUCKET_IS_EOS [Thu Jul 25 10:26:48.452910 2019] [proxy_ajp:debug] [pid 2164] mod_proxy_ajp.c(282): [client 10.16.248.65:58768] AH00873: data to read (max 8186 at 4) [Thu Jul 25 10:26:48.452916 2019] [proxy_ajp:debug] [pid 2164] mod_proxy_ajp.c(296): [client 10.16.248.65:58768] AH00875: got 0 bytes of data [Thu Jul 25 10:26:49.444631 2019] [proxy:debug] [pid 2771] proxy_util.c(1843): AH00925: initializing worker ajp://127.0.0.1:8702 shared [Thu Jul 25 10:26:49.444686 2019] [proxy:debug] [pid 2771] proxy_util.c(1885): AH00927: initializing worker ajp://127.0.0.1:8702 local [Thu Jul 25 10:26:49.444716 2019] [proxy:debug] [pid 2771] proxy_util.c(1936): AH00931: initialized single connection worker in child 2771 for (127.0.0.1) [Thu Jul 25 10:26:49.444745 2019] [proxy:debug] [pid 2771] proxy_util.c(1843): AH00925: initializing worker proxy:reverse shared [Thu Jul 25 10:26:49.444878 2019] [proxy:debug] [pid 2771] proxy_util.c(1885): AH00927: initializing worker proxy:reverse local [Thu Jul 25 10:26:49.444938 2019] [proxy:debug] [pid 2771] proxy_util.c(1936): AH00931: initialized single connection worker in child 2771 for (*) it keep initializing after the connection after an attempt to connect. Any ideas? Thanks & Regards Carl On Wed, Jul 24, 2019 at 11:00 PM Strahil wrote: > The CA can be downloaded via the web , or you can tell curl to just > ignore the engine's cert via the '-k' flag. > It will show you if the health page is working. > > Best Regards, > Strahil Nikolov > On Jul 24, 2019 19:39, carl langlois wrote: > > Strahil, not sure what to put for the --cacert. > > Yes Derek your are right at one point the port 8702 stop listening. > > tcp6 0 0 127.0.0.1:8702 :::* > LISTEN 1607/ovirt-engine > > After some time the line above disappear. I am trying to figure why this > port is being close after some time when the engine is running on the host > on the 248.x network. On the 236.x network this port is kept alive all the > time. > If you have any hint on why this port is closing do not hesitate because i > am starting to be out of ideas. :-) > > > Thanks & Regards > > Carl > > > > > > > On Wed, Jul 24, 2019 at 11:11 AM Strahil Nikolov > wrote: > > A healthy engine should report: > [root@ovirt1 ~]# curl --cacert CA > https://engine.localdomain/ovirt-engine/services/health;echo > DB Up!Welcome to Health Status! > > Of course you can use the '-k' switch to verify the situation. > > Best Regards, > Strahil Nikolov > > В сряда, 24 юли 2019 г., 17:43:59 ч. Гринуич+3, Derek Atkins < > de...@ihtfp.com> написа: > > > Hi, > > carl langlois writes: > > > If i try to access http://ovengine/ovirt-engine/services/health > > i always get "Service Unavailable" in the browser and each time i it > reload in > > the browser i get in the error_log > > > > [proxy_ajp:error] [pid 1868] [client 10.8.1.76:63512] AH00896: failed > to make > > connection to backend: 127.0.0.1 > > [Tue Jul 23 14:04:10.074023 2019] [proxy:error] [pid 1416] > (111)Connection > > refused: AH00957: AJP: attempt to connect to 127.0.0.1:8702 (127.0.0.1) > failed > > Sounds like a service isn't running on port 8702. > > > Thanks & Regards > > > > Carl > > -derek > > -- > Derek Atkins617-623-3745 > > de...@ihtfp.com > www.ihtfp.com > Computer and Internet Security Consultant > > ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/YE5TRRCXLPETYFRZ6TZPSLZPHAIDXPF3/
[ovirt-users] Re: major network changes
The CA can be downloaded via the web , or you can tell curl to just ignore the engine's cert via the '-k' flag. It will show you if the health page is working. Best Regards, Strahil NikolovOn Jul 24, 2019 19:39, carl langlois wrote: > > Strahil, not sure what to put for the --cacert. > > Yes Derek your are right at one point the port 8702 stop listening. > > tcp6 0 0 127.0.0.1:8702 :::* LISTEN > 1607/ovirt-engine > > After some time the line above disappear. I am trying to figure why this port > is being close after some time when the engine is running on the host on the > 248.x network. On the 236.x network this port is kept alive all the time. > If you have any hint on why this port is closing do not hesitate because i am > starting to be out of ideas. :-) > > > Thanks & Regards > > Carl > > > > > > > On Wed, Jul 24, 2019 at 11:11 AM Strahil Nikolov > wrote: >> >> A healthy engine should report: >> [root@ovirt1 ~]# curl --cacert CA >> https://engine.localdomain/ovirt-engine/services/health;echo >> DB Up!Welcome to Health Status! >> >> Of course you can use the '-k' switch to verify the situation. >> >> Best Regards, >> Strahil Nikolov >> >> В сряда, 24 юли 2019 г., 17:43:59 ч. Гринуич+3, Derek Atkins >> написа: >> >> >> Hi, >> >> carl langlois writes: >> >> > If i try to access http://ovengine/ovirt-engine/services/health >> > i always get "Service Unavailable" in the browser and each time i it >> > reload in >> > the browser i get in the error_log >> > >> > [proxy_ajp:error] [pid 1868] [client 10.8.1.76:63512] AH00896: failed to >> > make >> > connection to backend: 127.0.0.1 >> > [Tue Jul 23 14:04:10.074023 2019] [proxy:error] [pid 1416] (111)Connection >> > refused: AH00957: AJP: attempt to connect to 127.0.0.1:8702 (127.0.0.1) >> > failed >> >> Sounds like a service isn't running on port 8702. >> >> > Thanks & Regards >> > >> > Carl >> >> -derek >> >> -- >> Derek Atkins 617-623-3745 >> >> de...@ihtfp.com >> www.ihtfp.com >> Computer and Internet Security Consultant___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/UXWT4R7WYOIZXXIAWZC2S4XNN7V5HFYU/
[ovirt-users] Re: major network changes
Strahil, not sure what to put for the --cacert. Yes Derek your are right at one point the port 8702 stop listening. tcp6 0 0 127.0.0.1:8702 :::*LISTEN 1607/ovirt-engine After some time the line above disappear. I am trying to figure why this port is being close after some time when the engine is running on the host on the 248.x network. On the 236.x network this port is kept alive all the time. If you have any hint on why this port is closing do not hesitate because i am starting to be out of ideas. :-) Thanks & Regards Carl On Wed, Jul 24, 2019 at 11:11 AM Strahil Nikolov wrote: > A healthy engine should report: > [root@ovirt1 ~]# curl --cacert CA > https://engine.localdomain/ovirt-engine/services/health;echo > DB Up!Welcome to Health Status! > > Of course you can use the '-k' switch to verify the situation. > > Best Regards, > Strahil Nikolov > > В сряда, 24 юли 2019 г., 17:43:59 ч. Гринуич+3, Derek Atkins < > de...@ihtfp.com> написа: > > > Hi, > > carl langlois writes: > > > If i try to access http://ovengine/ovirt-engine/services/health > > i always get "Service Unavailable" in the browser and each time i it > reload in > > the browser i get in the error_log > > > > [proxy_ajp:error] [pid 1868] [client 10.8.1.76:63512] AH00896: failed > to make > > connection to backend: 127.0.0.1 > > [Tue Jul 23 14:04:10.074023 2019] [proxy:error] [pid 1416] > (111)Connection > > refused: AH00957: AJP: attempt to connect to 127.0.0.1:8702 (127.0.0.1) > failed > > Sounds like a service isn't running on port 8702. > > > Thanks & Regards > > > > Carl > > -derek > > -- > Derek Atkins617-623-3745 > > de...@ihtfp.com > www.ihtfp.com > Computer and Internet Security Consultant > ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/X2TLXH46EU5OJNQ3ZYZIZJG73HR6KXY7/
[ovirt-users] Re: major network changes
A healthy engine should report:[root@ovirt1 ~]# curl --cacert CA https://engine.localdomain/ovirt-engine/services/health;echoDB Up!Welcome to Health Status! Of course you can use the '-k' switch to verify the situation. Best Regards,Strahil Nikolov В сряда, 24 юли 2019 г., 17:43:59 ч. Гринуич+3, Derek Atkins написа: Hi, carl langlois writes: > If i try to access http://ovengine/ovirt-engine/services/health > i always get "Service Unavailable" in the browser and each time i it reload in > the browser i get in the error_log > > [proxy_ajp:error] [pid 1868] [client 10.8.1.76:63512] AH00896: failed to make > connection to backend: 127.0.0.1 > [Tue Jul 23 14:04:10.074023 2019] [proxy:error] [pid 1416] (111)Connection > refused: AH00957: AJP: attempt to connect to 127.0.0.1:8702 (127.0.0.1) failed Sounds like a service isn't running on port 8702. > Thanks & Regards > > Carl -derek -- Derek Atkins 617-623-3745 de...@ihtfp.com www.ihtfp.com Computer and Internet Security Consultant ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/RLIGSXA5DMFKYWOAZQOYH6EOB7DMQVDL/
[ovirt-users] Re: major network changes
Hi, carl langlois writes: > If i try to access http://ovengine/ovirt-engine/services/health > i always get "Service Unavailable" in the browser and each time i it reload in > the browser i get in the error_log > > [proxy_ajp:error] [pid 1868] [client 10.8.1.76:63512] AH00896: failed to make > connection to backend: 127.0.0.1 > [Tue Jul 23 14:04:10.074023 2019] [proxy:error] [pid 1416] (111)Connection > refused: AH00957: AJP: attempt to connect to 127.0.0.1:8702 (127.0.0.1) failed Sounds like a service isn't running on port 8702. > Thanks & Regards > > Carl -derek -- Derek Atkins 617-623-3745 de...@ihtfp.com www.ihtfp.com Computer and Internet Security Consultant ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/QMW4OB7AIVE2YYU2OYIGZPVW5F4VTLLK/
[ovirt-users] Re: major network changes
If i try to access http://ovengine/ovirt-engine/services/health i always get "Service Unavailable" in the browser and each time i it reload in the browser i get in the error_log [proxy_ajp:error] [pid 1868] [client 10.8.1.76:63512] AH00896: failed to make connection to backend: 127.0.0.1 [Tue Jul 23 14:04:10.074023 2019] [proxy:error] [pid 1416] (111)Connection refused: AH00957: AJP: attempt to connect to 127.0.0.1:8702 (127.0.0.1) failed Thanks & Regards Carl On Tue, Jul 23, 2019 at 12:59 PM carl langlois wrote: > Hi > At one point we did have issue with DNS resolution(mainly the reverse > lookup). But that was fix. Yes we can ping both network and vice-versa. > > Not sure how to multi-home the engine. Will do some research on that. > > I did find something in the error_log on the engine. > > In the /etc/httpd/logs/error_log i always get this messages. > > [Tue Jul 23 11:21:52.430555 2019] [proxy:error] [pid 3189] AH00959: > ap_proxy_connect_backend disabling worker for (127.0.0.1) for 5s > [Tue Jul 23 11:21:52.430562 2019] [proxy_ajp:error] [pid 3189] [client > 10.16.248.65:35154] AH00896: failed to make connection to backend: 127.0.0.1 > > The 10.16.248.65 is the new address of the host that was move to the new > network. > > > Thanks & Regards > Carl > > > > > On Tue, Jul 23, 2019 at 11:52 AM Strahil wrote: > >> According to another post in the mailing list, the Engine Hosts (that has >> ovirt-ha-agent/ovirt-ha-broker running) is checking http:// >> {fqdn}/ovirt-engine/services/health >> >> As the IP is changed, I think you need to check the URL before and after >> thr mifgration. >> >> Best Regards, >> Strahil NikolovOn Jul 23, 2019 16:41, Derek Atkins >> wrote: >> > >> > Hi, >> > >> > If I understand it correctly, the HE Hosts try to ping (or SSH, or >> > otherwise reach) the Engine host. If it reaches it, then it passes the >> > liveness check. If it cannot reach it, then it fails. So to me this >> error >> > means that there is some configuration, somewhere, that is trying to >> reach >> > the engine on the old address (which fails when the engine has the new >> > address). >> > >> > I do not know where in the *host* configuration this data lives, so I >> > cannot suggest where you need to change it. >> > >> > Can 10.16.248.x reach 10.8.236.x and vice-versa? >> > >> > Maybe multi-home the engine on both networks for now until you figure >> it out? >> > >> > -derek >> > >> > On Tue, July 23, 2019 9:13 am, carl langlois wrote: >> > > Hi, >> > > >> > > We have managed to stabilize the DNS udpate in out network. Now the >> > > current >> > > situation is. >> > > I have 3 hosts that can run the engine (hosted-engine). >> > > They were all in the 10.8.236.x. Now i have moved one of them in the >> > > 10.16.248.x. >> > > >> > > If i boot the engine on one of the host that is in the 10.8.236.x the >> > > engine is going up with status "good". I can access the engine UI. I >> can >> > > see all my hosts even the one in the 10.16.248.x network. >> > > >> > > But if i boot the engine on the hosted-engine host that was switch to >> the >> > > 10.16.248.x the engine is booting. I can ssh to it but the status is >> > > always >> > > " fail for liveliness check". >> > > The main difference is that when i boot on the host that is in the >> > > 10.16.248.x network the engine gets a address in the 248.x network. >> > > >> > > On the engine i have this in the >> > > /var/log/ovirt-engine-dwh/ovirt-engine-dwhd.log >> > > 019-07-23 >> > > >> 09:05:30|MFzehi|YYTDiS|jTq2w8|OVIRT_ENGINE_DWH|SampleTimeKeepingJob|Default|5|tWarn|tWarn_1|Can >> >> > > not sample data, oVirt Engine is not updating the statistics. Please >> check >> > > your oVirt Engine status.|9704 >> > > the engine.log seems okey. >> > > >> > > So i need to understand what this " liveliness check" do(or try to >> do) so >> > > i >> > > can investigate why the engine status is not becoming good. >> > > >> > > The initial deployment was done in the 10.8.236.x network. Maybe is >> as >> > > something to do with that. >> > > >> > > Thanks & Regards >> > > >> > > Carl >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > On Thu, Jul 18, 2019 at 8:53 AM Miguel Duarte de Mora Barroso < >> > > mdbarr...@redhat.com> wrote: >> > > >> > >> On Thu, Jul 18, 2019 at 2:50 PM Miguel Duarte de Mora Barroso >> > >> wrote: >> > >> > >> > >> > On Thu, Jul 18, 2019 at 1:57 PM carl langlois < >> crl.langl...@gmail.com> >> > >> wrote: >> > >> > > >> > >> > > Hi Miguel, >> > >> > > >> > >> > > I have managed to change the config for the ovn-controler. >> > >> > > with those commands >> > >> > > ovs-vsctl set Open_vSwitch . external-ids:ovn-remote=ssl: >> > >> 10.16.248.74:6642 >> > >> > > ovs-vsctl set Open_vSwitch . >> external-ids:ovn-encap-ip=10.16.248.65 >> > >> > > and restating the services >> > >> > >> > >> > Yes, that's what the script is supposed to do, check [0]. >> > >> > >>
[ovirt-users] Re: major network changes
Hi At one point we did have issue with DNS resolution(mainly the reverse lookup). But that was fix. Yes we can ping both network and vice-versa. Not sure how to multi-home the engine. Will do some research on that. I did find something in the error_log on the engine. In the /etc/httpd/logs/error_log i always get this messages. [Tue Jul 23 11:21:52.430555 2019] [proxy:error] [pid 3189] AH00959: ap_proxy_connect_backend disabling worker for (127.0.0.1) for 5s [Tue Jul 23 11:21:52.430562 2019] [proxy_ajp:error] [pid 3189] [client 10.16.248.65:35154] AH00896: failed to make connection to backend: 127.0.0.1 The 10.16.248.65 is the new address of the host that was move to the new network. Thanks & Regards Carl On Tue, Jul 23, 2019 at 11:52 AM Strahil wrote: > According to another post in the mailing list, the Engine Hosts (that has > ovirt-ha-agent/ovirt-ha-broker running) is checking http:// > {fqdn}/ovirt-engine/services/health > > As the IP is changed, I think you need to check the URL before and after > thr mifgration. > > Best Regards, > Strahil NikolovOn Jul 23, 2019 16:41, Derek Atkins > wrote: > > > > Hi, > > > > If I understand it correctly, the HE Hosts try to ping (or SSH, or > > otherwise reach) the Engine host. If it reaches it, then it passes the > > liveness check. If it cannot reach it, then it fails. So to me this > error > > means that there is some configuration, somewhere, that is trying to > reach > > the engine on the old address (which fails when the engine has the new > > address). > > > > I do not know where in the *host* configuration this data lives, so I > > cannot suggest where you need to change it. > > > > Can 10.16.248.x reach 10.8.236.x and vice-versa? > > > > Maybe multi-home the engine on both networks for now until you figure it > out? > > > > -derek > > > > On Tue, July 23, 2019 9:13 am, carl langlois wrote: > > > Hi, > > > > > > We have managed to stabilize the DNS udpate in out network. Now the > > > current > > > situation is. > > > I have 3 hosts that can run the engine (hosted-engine). > > > They were all in the 10.8.236.x. Now i have moved one of them in the > > > 10.16.248.x. > > > > > > If i boot the engine on one of the host that is in the 10.8.236.x the > > > engine is going up with status "good". I can access the engine UI. I > can > > > see all my hosts even the one in the 10.16.248.x network. > > > > > > But if i boot the engine on the hosted-engine host that was switch to > the > > > 10.16.248.x the engine is booting. I can ssh to it but the status is > > > always > > > " fail for liveliness check". > > > The main difference is that when i boot on the host that is in the > > > 10.16.248.x network the engine gets a address in the 248.x network. > > > > > > On the engine i have this in the > > > /var/log/ovirt-engine-dwh/ovirt-engine-dwhd.log > > > 019-07-23 > > > > 09:05:30|MFzehi|YYTDiS|jTq2w8|OVIRT_ENGINE_DWH|SampleTimeKeepingJob|Default|5|tWarn|tWarn_1|Can > > > > not sample data, oVirt Engine is not updating the statistics. Please > check > > > your oVirt Engine status.|9704 > > > the engine.log seems okey. > > > > > > So i need to understand what this " liveliness check" do(or try to do) > so > > > i > > > can investigate why the engine status is not becoming good. > > > > > > The initial deployment was done in the 10.8.236.x network. Maybe is as > > > something to do with that. > > > > > > Thanks & Regards > > > > > > Carl > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > On Thu, Jul 18, 2019 at 8:53 AM Miguel Duarte de Mora Barroso < > > > mdbarr...@redhat.com> wrote: > > > > > >> On Thu, Jul 18, 2019 at 2:50 PM Miguel Duarte de Mora Barroso > > >> wrote: > > >> > > > >> > On Thu, Jul 18, 2019 at 1:57 PM carl langlois < > crl.langl...@gmail.com> > > >> wrote: > > >> > > > > >> > > Hi Miguel, > > >> > > > > >> > > I have managed to change the config for the ovn-controler. > > >> > > with those commands > > >> > > ovs-vsctl set Open_vSwitch . external-ids:ovn-remote=ssl: > > >> 10.16.248.74:6642 > > >> > > ovs-vsctl set Open_vSwitch . > external-ids:ovn-encap-ip=10.16.248.65 > > >> > > and restating the services > > >> > > > >> > Yes, that's what the script is supposed to do, check [0]. > > >> > > > >> > Not sure why running vdsm-tool didn't work for you. > > >> > > > >> > > > > >> > > But even with this i still have the "fail for liveliness check" > when > > >> starting the ovirt engine. But one thing i notice with our new > network > > >> is > > >> that the reverse DNS does not work(IP -> hostname). The forward is > > >> working > > >> fine. I am trying to see with our IT why it is not working. > > >> > > > >> > Do you guys use OVN? If not, you could disable the provider, > install > > >> > the hosted-engine VM, then, if needed, re-add / re-activate it . > > >> > > >> I'm assuming it fails for the same reason you've stated initially - > > >> i.e. ovn-controller is involve
[ovirt-users] Re: major network changes
According to another post in the mailing list, the Engine Hosts (that has ovirt-ha-agent/ovirt-ha-broker running) is checking http://{fqdn}/ovirt-engine/services/health As the IP is changed, I think you need to check the URL before and after thr mifgration. Best Regards, Strahil NikolovOn Jul 23, 2019 16:41, Derek Atkins wrote: > > Hi, > > If I understand it correctly, the HE Hosts try to ping (or SSH, or > otherwise reach) the Engine host. If it reaches it, then it passes the > liveness check. If it cannot reach it, then it fails. So to me this error > means that there is some configuration, somewhere, that is trying to reach > the engine on the old address (which fails when the engine has the new > address). > > I do not know where in the *host* configuration this data lives, so I > cannot suggest where you need to change it. > > Can 10.16.248.x reach 10.8.236.x and vice-versa? > > Maybe multi-home the engine on both networks for now until you figure it out? > > -derek > > On Tue, July 23, 2019 9:13 am, carl langlois wrote: > > Hi, > > > > We have managed to stabilize the DNS udpate in out network. Now the > > current > > situation is. > > I have 3 hosts that can run the engine (hosted-engine). > > They were all in the 10.8.236.x. Now i have moved one of them in the > > 10.16.248.x. > > > > If i boot the engine on one of the host that is in the 10.8.236.x the > > engine is going up with status "good". I can access the engine UI. I can > > see all my hosts even the one in the 10.16.248.x network. > > > > But if i boot the engine on the hosted-engine host that was switch to the > > 10.16.248.x the engine is booting. I can ssh to it but the status is > > always > > " fail for liveliness check". > > The main difference is that when i boot on the host that is in the > > 10.16.248.x network the engine gets a address in the 248.x network. > > > > On the engine i have this in the > > /var/log/ovirt-engine-dwh/ovirt-engine-dwhd.log > > 019-07-23 > > 09:05:30|MFzehi|YYTDiS|jTq2w8|OVIRT_ENGINE_DWH|SampleTimeKeepingJob|Default|5|tWarn|tWarn_1|Can > > > > not sample data, oVirt Engine is not updating the statistics. Please check > > your oVirt Engine status.|9704 > > the engine.log seems okey. > > > > So i need to understand what this " liveliness check" do(or try to do) so > > i > > can investigate why the engine status is not becoming good. > > > > The initial deployment was done in the 10.8.236.x network. Maybe is as > > something to do with that. > > > > Thanks & Regards > > > > Carl > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > On Thu, Jul 18, 2019 at 8:53 AM Miguel Duarte de Mora Barroso < > > mdbarr...@redhat.com> wrote: > > > >> On Thu, Jul 18, 2019 at 2:50 PM Miguel Duarte de Mora Barroso > >> wrote: > >> > > >> > On Thu, Jul 18, 2019 at 1:57 PM carl langlois > >> wrote: > >> > > > >> > > Hi Miguel, > >> > > > >> > > I have managed to change the config for the ovn-controler. > >> > > with those commands > >> > > ovs-vsctl set Open_vSwitch . external-ids:ovn-remote=ssl: > >> 10.16.248.74:6642 > >> > > ovs-vsctl set Open_vSwitch . external-ids:ovn-encap-ip=10.16.248.65 > >> > > and restating the services > >> > > >> > Yes, that's what the script is supposed to do, check [0]. > >> > > >> > Not sure why running vdsm-tool didn't work for you. > >> > > >> > > > >> > > But even with this i still have the "fail for liveliness check" when > >> starting the ovirt engine. But one thing i notice with our new network > >> is > >> that the reverse DNS does not work(IP -> hostname). The forward is > >> working > >> fine. I am trying to see with our IT why it is not working. > >> > > >> > Do you guys use OVN? If not, you could disable the provider, install > >> > the hosted-engine VM, then, if needed, re-add / re-activate it . > >> > >> I'm assuming it fails for the same reason you've stated initially - > >> i.e. ovn-controller is involved; if it is not, disregard this msg :) > >> > > >> > [0] - > >> https://github.com/oVirt/ovirt-provider-ovn/blob/master/driver/scripts/setup_ovn_controller.sh#L24 > >> > >> > > >> > > > >> > > Regards. > >> > > Carl > >> ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/UA6TWA5UHXT6NSLB7ZTAYWGMANQL3BED/
[ovirt-users] Re: major network changes
Hi Carl, I think there is another thread here related to the migration to another network. As far as I know, the check liveliness try's to access the ovirt's health page. Does the new engine's ip has A/PTR record setup? Also, check the engine logs, once the HostedEngine VM is up and running. Best Regards, Strahil NikolovOn Jul 23, 2019 16:13, carl langlois wrote: > > Hi, > > We have managed to stabilize the DNS udpate in out network. Now the current > situation is. > I have 3 hosts that can run the engine (hosted-engine). > They were all in the 10.8.236.x. Now i have moved one of them in the > 10.16.248.x. > > If i boot the engine on one of the host that is in the 10.8.236.x the engine > is going up with status "good". I can access the engine UI. I can see all my > hosts even the one in the 10.16.248.x network. > > But if i boot the engine on the hosted-engine host that was switch to the > 10.16.248.x the engine is booting. I can ssh to it but the status is always " > fail for liveliness check". > The main difference is that when i boot on the host that is in the > 10.16.248.x network the engine gets a address in the 248.x network. > > On the engine i have this in the > /var/log/ovirt-engine-dwh/ovirt-engine-dwhd.log > 019-07-23 > 09:05:30|MFzehi|YYTDiS|jTq2w8|OVIRT_ENGINE_DWH|SampleTimeKeepingJob|Default|5|tWarn|tWarn_1|Can > not sample data, oVirt Engine is not updating the statistics. Please check > your oVirt Engine status.|9704 > the engine.log seems okey. > > So i need to understand what this " liveliness check" do(or try to do) so i > can investigate why the engine status is not becoming good. > > The initial deployment was done in the 10.8.236.x network. Maybe is as > something to do with that. > > Thanks & Regards > > Carl > > > > > > > > > > > > > > > > > > > On Thu, Jul 18, 2019 at 8:53 AM Miguel Duarte de Mora Barroso > wrote: >> >> On Thu, Jul 18, 2019 at 2:50 PM Miguel Duarte de Mora Barroso >> wrote: >> > >> > On Thu, Jul 18, 2019 at 1:57 PM carl langlois >> > wrote: >> > > >> > > Hi Miguel, >> > > >> > > I have managed to change the config for the ovn-controler. >> > > with those commands >> > > ovs-vsctl set Open_vSwitch . >> > >external-ids:ovn-remote=ssl:10.16.248.74:6642 >> > > ovs-vsctl set Open_vSwitch . external-ids:ovn-encap-ip=10.16.248.65 >> > > and restating the services >> > >> > Yes, that's what the script is supposed to do, check [0]. >> > >> > Not sure why running vdsm-tool didn't work for you. >> > >> > > >> > > But even with this i still have the "fail for liveliness check" when >> > > starting the ovirt engine. But one thing i notice with our new network >> > > is that the reverse DNS does not work(IP -> hostname). The forward is >> > > working fine. I am trying to see with our IT why it is not working. >> > >> > Do you guys use OVN? If not, you could disable the provider, install >> > the hosted-engine VM, then, if needed, re-add / re-activate it . >> >> I'm assuming it fails for the same reason you've stated initially - >> i.e. ovn-controller is involved; if it is not, disregard this msg :) >> > >> > [0] - >> > https://github.com/oVirt/ovirt-provider-ovn/blob/master/driver/scripts/setup_ovn_controller.sh#L24 >> > >> > > >> > > Regards. >> > > Carl >> > > >> > > On Thu, Jul 18, 2019 at 4:03 AM Miguel Duarte de Mora Barroso >> > > wrote: >> > >> >> > >> On Wed, Jul 17, 2019 at 7:07 PM carl langlois >> > >> wrote: >> > >> > >> > >> > Hi >> > >> > Here is the output of the command >> > >> > >> > >> > [root@ovhost1 ~]# vdsm-tool --vvverbose ovn-config 10.16.248.74 >> > >> > ovirtmgmt >> > >> > MainThread::DEBUG::2019-07-17 13:02:52,___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/RP3F6XF3KZ4FOSZR4WXQVV4CH4WAIM5Z/
[ovirt-users] Re: major network changes
Hi, If I understand it correctly, the HE Hosts try to ping (or SSH, or otherwise reach) the Engine host. If it reaches it, then it passes the liveness check. If it cannot reach it, then it fails. So to me this error means that there is some configuration, somewhere, that is trying to reach the engine on the old address (which fails when the engine has the new address). I do not know where in the *host* configuration this data lives, so I cannot suggest where you need to change it. Can 10.16.248.x reach 10.8.236.x and vice-versa? Maybe multi-home the engine on both networks for now until you figure it out? -derek On Tue, July 23, 2019 9:13 am, carl langlois wrote: > Hi, > > We have managed to stabilize the DNS udpate in out network. Now the > current > situation is. > I have 3 hosts that can run the engine (hosted-engine). > They were all in the 10.8.236.x. Now i have moved one of them in the > 10.16.248.x. > > If i boot the engine on one of the host that is in the 10.8.236.x the > engine is going up with status "good". I can access the engine UI. I can > see all my hosts even the one in the 10.16.248.x network. > > But if i boot the engine on the hosted-engine host that was switch to the > 10.16.248.x the engine is booting. I can ssh to it but the status is > always > " fail for liveliness check". > The main difference is that when i boot on the host that is in the > 10.16.248.x network the engine gets a address in the 248.x network. > > On the engine i have this in the > /var/log/ovirt-engine-dwh/ovirt-engine-dwhd.log > 019-07-23 > 09:05:30|MFzehi|YYTDiS|jTq2w8|OVIRT_ENGINE_DWH|SampleTimeKeepingJob|Default|5|tWarn|tWarn_1|Can > not sample data, oVirt Engine is not updating the statistics. Please check > your oVirt Engine status.|9704 > the engine.log seems okey. > > So i need to understand what this " liveliness check" do(or try to do) so > i > can investigate why the engine status is not becoming good. > > The initial deployment was done in the 10.8.236.x network. Maybe is as > something to do with that. > > Thanks & Regards > > Carl > > > > > > > > > > > > > > > > > > > On Thu, Jul 18, 2019 at 8:53 AM Miguel Duarte de Mora Barroso < > mdbarr...@redhat.com> wrote: > >> On Thu, Jul 18, 2019 at 2:50 PM Miguel Duarte de Mora Barroso >> wrote: >> > >> > On Thu, Jul 18, 2019 at 1:57 PM carl langlois >> wrote: >> > > >> > > Hi Miguel, >> > > >> > > I have managed to change the config for the ovn-controler. >> > > with those commands >> > > ovs-vsctl set Open_vSwitch . external-ids:ovn-remote=ssl: >> 10.16.248.74:6642 >> > > ovs-vsctl set Open_vSwitch . external-ids:ovn-encap-ip=10.16.248.65 >> > > and restating the services >> > >> > Yes, that's what the script is supposed to do, check [0]. >> > >> > Not sure why running vdsm-tool didn't work for you. >> > >> > > >> > > But even with this i still have the "fail for liveliness check" when >> starting the ovirt engine. But one thing i notice with our new network >> is >> that the reverse DNS does not work(IP -> hostname). The forward is >> working >> fine. I am trying to see with our IT why it is not working. >> > >> > Do you guys use OVN? If not, you could disable the provider, install >> > the hosted-engine VM, then, if needed, re-add / re-activate it . >> >> I'm assuming it fails for the same reason you've stated initially - >> i.e. ovn-controller is involved; if it is not, disregard this msg :) >> > >> > [0] - >> https://github.com/oVirt/ovirt-provider-ovn/blob/master/driver/scripts/setup_ovn_controller.sh#L24 >> > >> > > >> > > Regards. >> > > Carl >> > > >> > > On Thu, Jul 18, 2019 at 4:03 AM Miguel Duarte de Mora Barroso < >> mdbarr...@redhat.com> wrote: >> > >> >> > >> On Wed, Jul 17, 2019 at 7:07 PM carl langlois >> >> wrote: >> > >> > >> > >> > Hi >> > >> > Here is the output of the command >> > >> > >> > >> > [root@ovhost1 ~]# vdsm-tool --vvverbose ovn-config 10.16.248.74 >> ovirtmgmt >> > >> > MainThread::DEBUG::2019-07-17 >> 13:02:52,581::cmdutils::150::root::(exec_cmd) lshw -json -disable usb >> -disable pcmcia -disable isapnp -disable ide -disable scsi -disable dmi >> -disable memory -disable cpuinfo (cwd None) >> > >> > MainThread::DEBUG::2019-07-17 >> 13:02:52,738::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; >> = 0 >> > >> > MainThread::DEBUG::2019-07-17 >> 13:02:52,741::routes::109::root::(get_gateway) The gateway 10.16.248.1 >> is >> duplicated for the device ovirtmgmt >> > >> > MainThread::DEBUG::2019-07-17 >> 13:02:52,742::routes::109::root::(get_gateway) The gateway 10.16.248.1 >> is >> duplicated for the device ovirtmgmt >> > >> > MainThread::DEBUG::2019-07-17 >> 13:02:52,742::cmdutils::150::root::(exec_cmd) /sbin/tc qdisc show (cwd >> None) >> > >> > MainThread::DEBUG::2019-07-17 >> 13:02:52,744::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; >> = 0 >> > >> > MainThread::DEBUG::2019-07-17 >> 13:02:52,745::cmdutils::150::root::(exec_cmd) /sbin/tc class show dev >> enp2s0f1 classid 0:1388 (cwd None) >> > >> > MainThrea
[ovirt-users] Re: major network changes
Hi, We have managed to stabilize the DNS udpate in out network. Now the current situation is. I have 3 hosts that can run the engine (hosted-engine). They were all in the 10.8.236.x. Now i have moved one of them in the 10.16.248.x. If i boot the engine on one of the host that is in the 10.8.236.x the engine is going up with status "good". I can access the engine UI. I can see all my hosts even the one in the 10.16.248.x network. But if i boot the engine on the hosted-engine host that was switch to the 10.16.248.x the engine is booting. I can ssh to it but the status is always " fail for liveliness check". The main difference is that when i boot on the host that is in the 10.16.248.x network the engine gets a address in the 248.x network. On the engine i have this in the /var/log/ovirt-engine-dwh/ovirt-engine-dwhd.log 019-07-23 09:05:30|MFzehi|YYTDiS|jTq2w8|OVIRT_ENGINE_DWH|SampleTimeKeepingJob|Default|5|tWarn|tWarn_1|Can not sample data, oVirt Engine is not updating the statistics. Please check your oVirt Engine status.|9704 the engine.log seems okey. So i need to understand what this " liveliness check" do(or try to do) so i can investigate why the engine status is not becoming good. The initial deployment was done in the 10.8.236.x network. Maybe is as something to do with that. Thanks & Regards Carl On Thu, Jul 18, 2019 at 8:53 AM Miguel Duarte de Mora Barroso < mdbarr...@redhat.com> wrote: > On Thu, Jul 18, 2019 at 2:50 PM Miguel Duarte de Mora Barroso > wrote: > > > > On Thu, Jul 18, 2019 at 1:57 PM carl langlois > wrote: > > > > > > Hi Miguel, > > > > > > I have managed to change the config for the ovn-controler. > > > with those commands > > > ovs-vsctl set Open_vSwitch . external-ids:ovn-remote=ssl: > 10.16.248.74:6642 > > > ovs-vsctl set Open_vSwitch . external-ids:ovn-encap-ip=10.16.248.65 > > > and restating the services > > > > Yes, that's what the script is supposed to do, check [0]. > > > > Not sure why running vdsm-tool didn't work for you. > > > > > > > > But even with this i still have the "fail for liveliness check" when > starting the ovirt engine. But one thing i notice with our new network is > that the reverse DNS does not work(IP -> hostname). The forward is working > fine. I am trying to see with our IT why it is not working. > > > > Do you guys use OVN? If not, you could disable the provider, install > > the hosted-engine VM, then, if needed, re-add / re-activate it . > > I'm assuming it fails for the same reason you've stated initially - > i.e. ovn-controller is involved; if it is not, disregard this msg :) > > > > [0] - > https://github.com/oVirt/ovirt-provider-ovn/blob/master/driver/scripts/setup_ovn_controller.sh#L24 > > > > > > > > Regards. > > > Carl > > > > > > On Thu, Jul 18, 2019 at 4:03 AM Miguel Duarte de Mora Barroso < > mdbarr...@redhat.com> wrote: > > >> > > >> On Wed, Jul 17, 2019 at 7:07 PM carl langlois > wrote: > > >> > > > >> > Hi > > >> > Here is the output of the command > > >> > > > >> > [root@ovhost1 ~]# vdsm-tool --vvverbose ovn-config 10.16.248.74 > ovirtmgmt > > >> > MainThread::DEBUG::2019-07-17 > 13:02:52,581::cmdutils::150::root::(exec_cmd) lshw -json -disable usb > -disable pcmcia -disable isapnp -disable ide -disable scsi -disable dmi > -disable memory -disable cpuinfo (cwd None) > > >> > MainThread::DEBUG::2019-07-17 > 13:02:52,738::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 > > >> > MainThread::DEBUG::2019-07-17 > 13:02:52,741::routes::109::root::(get_gateway) The gateway 10.16.248.1 is > duplicated for the device ovirtmgmt > > >> > MainThread::DEBUG::2019-07-17 > 13:02:52,742::routes::109::root::(get_gateway) The gateway 10.16.248.1 is > duplicated for the device ovirtmgmt > > >> > MainThread::DEBUG::2019-07-17 > 13:02:52,742::cmdutils::150::root::(exec_cmd) /sbin/tc qdisc show (cwd None) > > >> > MainThread::DEBUG::2019-07-17 > 13:02:52,744::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 > > >> > MainThread::DEBUG::2019-07-17 > 13:02:52,745::cmdutils::150::root::(exec_cmd) /sbin/tc class show dev > enp2s0f1 classid 0:1388 (cwd None) > > >> > MainThread::DEBUG::2019-07-17 > 13:02:52,747::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 > > >> > MainThread::DEBUG::2019-07-17 > 13:02:52,766::cmdutils::150::root::(exec_cmd) > /usr/share/openvswitch/scripts/ovs-ctl status (cwd None) > > >> > MainThread::DEBUG::2019-07-17 > 13:02:52,777::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 > > >> > MainThread::DEBUG::2019-07-17 > 13:02:52,778::vsctl::67::root::(commit) Executing commands: > /usr/bin/ovs-vsctl --timeout=5 --oneline --format=json -- list Bridge -- > list Port -- list Interface > > >> > MainThread::DEBUG::2019-07-17 > 13:02:52,778::cmdutils::150::root::(exec_cmd) /usr/bin/ovs-vsctl > --timeout=5 --oneline --format=json -- list Bridge -- list Port -- list > Interface (cwd None) > > >> > MainThread::DEBUG::2019-07-17 > 13:02:52,799::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 > >
[ovirt-users] Re: major network changes
On Thu, Jul 18, 2019 at 2:50 PM Miguel Duarte de Mora Barroso wrote: > > On Thu, Jul 18, 2019 at 1:57 PM carl langlois wrote: > > > > Hi Miguel, > > > > I have managed to change the config for the ovn-controler. > > with those commands > > ovs-vsctl set Open_vSwitch . external-ids:ovn-remote=ssl:10.16.248.74:6642 > > ovs-vsctl set Open_vSwitch . external-ids:ovn-encap-ip=10.16.248.65 > > and restating the services > > Yes, that's what the script is supposed to do, check [0]. > > Not sure why running vdsm-tool didn't work for you. > > > > > But even with this i still have the "fail for liveliness check" when > > starting the ovirt engine. But one thing i notice with our new network is > > that the reverse DNS does not work(IP -> hostname). The forward is working > > fine. I am trying to see with our IT why it is not working. > > Do you guys use OVN? If not, you could disable the provider, install > the hosted-engine VM, then, if needed, re-add / re-activate it . I'm assuming it fails for the same reason you've stated initially - i.e. ovn-controller is involved; if it is not, disregard this msg :) > > [0] - > https://github.com/oVirt/ovirt-provider-ovn/blob/master/driver/scripts/setup_ovn_controller.sh#L24 > > > > > Regards. > > Carl > > > > On Thu, Jul 18, 2019 at 4:03 AM Miguel Duarte de Mora Barroso > > wrote: > >> > >> On Wed, Jul 17, 2019 at 7:07 PM carl langlois > >> wrote: > >> > > >> > Hi > >> > Here is the output of the command > >> > > >> > [root@ovhost1 ~]# vdsm-tool --vvverbose ovn-config 10.16.248.74 ovirtmgmt > >> > MainThread::DEBUG::2019-07-17 > >> > 13:02:52,581::cmdutils::150::root::(exec_cmd) lshw -json -disable usb > >> > -disable pcmcia -disable isapnp -disable ide -disable scsi -disable dmi > >> > -disable memory -disable cpuinfo (cwd None) > >> > MainThread::DEBUG::2019-07-17 > >> > 13:02:52,738::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; > >> > = 0 > >> > MainThread::DEBUG::2019-07-17 > >> > 13:02:52,741::routes::109::root::(get_gateway) The gateway 10.16.248.1 > >> > is duplicated for the device ovirtmgmt > >> > MainThread::DEBUG::2019-07-17 > >> > 13:02:52,742::routes::109::root::(get_gateway) The gateway 10.16.248.1 > >> > is duplicated for the device ovirtmgmt > >> > MainThread::DEBUG::2019-07-17 > >> > 13:02:52,742::cmdutils::150::root::(exec_cmd) /sbin/tc qdisc show (cwd > >> > None) > >> > MainThread::DEBUG::2019-07-17 > >> > 13:02:52,744::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; > >> > = 0 > >> > MainThread::DEBUG::2019-07-17 > >> > 13:02:52,745::cmdutils::150::root::(exec_cmd) /sbin/tc class show dev > >> > enp2s0f1 classid 0:1388 (cwd None) > >> > MainThread::DEBUG::2019-07-17 > >> > 13:02:52,747::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; > >> > = 0 > >> > MainThread::DEBUG::2019-07-17 > >> > 13:02:52,766::cmdutils::150::root::(exec_cmd) > >> > /usr/share/openvswitch/scripts/ovs-ctl status (cwd None) > >> > MainThread::DEBUG::2019-07-17 > >> > 13:02:52,777::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; > >> > = 0 > >> > MainThread::DEBUG::2019-07-17 13:02:52,778::vsctl::67::root::(commit) > >> > Executing commands: /usr/bin/ovs-vsctl --timeout=5 --oneline > >> > --format=json -- list Bridge -- list Port -- list Interface > >> > MainThread::DEBUG::2019-07-17 > >> > 13:02:52,778::cmdutils::150::root::(exec_cmd) /usr/bin/ovs-vsctl > >> > --timeout=5 --oneline --format=json -- list Bridge -- list Port -- list > >> > Interface (cwd None) > >> > MainThread::DEBUG::2019-07-17 > >> > 13:02:52,799::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; > >> > = 0 > >> > netlink/events::DEBUG::2019-07-17 > >> > 13:02:52,802::concurrent::192::root::(run) START thread > >> > (func= >> > method Monitor._scan of >> > 0x7f99fb618c90>>, args=(), kwargs={}) > >> > netlink/events::DEBUG::2019-07-17 > >> > 13:02:54,805::concurrent::195::root::(run) FINISH thread > >> > > >> > Using default PKI files > >> > > >> > I do not see any indication of the config?? > >> > >> And afterwards when you execute "ovs-vsctl list Open_vSwitch" does it > >> reflect the updated value ? > >> > >> This command would have to be performed in the node where hosted > >> engine will be hosted - not sure if it's possible to determine before > >> hand which one it will be. If not, you should run it in all the nodes > >> in the cluster, to be sure. > >> > >> > > >> > Regards > >> > Carl > >> > > >> > On Wed, Jul 17, 2019 at 11:40 AM carl langlois > >> > wrote: > >> >> > >> >> Hi > >> >> > >> >> I have open a bug https://bugzilla.redhat.com/show_bug.cgi?id=1730776 > >> >> > >> >> I have try this command "vdsm-tool ovn-config 10.16.248.74 ovirtmgmt" > >> >> on one of the host but nothing changed. After a restart of the > >> >> ovn-controler i still get > >> >> > >> >> 2019-07-17T15:38:52.572Z|00033|reconnect|INFO|ssl:10.8.236.244:6642: > >> >> waiting 8 seconds before reconnect > >> >> 2019-07-17T15:39:00.578Z|00034|reconnect|INFO|ssl:10.8.236.244:6642
[ovirt-users] Re: major network changes
On Thu, Jul 18, 2019 at 1:57 PM carl langlois wrote: > > Hi Miguel, > > I have managed to change the config for the ovn-controler. > with those commands > ovs-vsctl set Open_vSwitch . external-ids:ovn-remote=ssl:10.16.248.74:6642 > ovs-vsctl set Open_vSwitch . external-ids:ovn-encap-ip=10.16.248.65 > and restating the services Yes, that's what the script is supposed to do, check [0]. Not sure why running vdsm-tool didn't work for you. > > But even with this i still have the "fail for liveliness check" when starting > the ovirt engine. But one thing i notice with our new network is that the > reverse DNS does not work(IP -> hostname). The forward is working fine. I am > trying to see with our IT why it is not working. Do you guys use OVN? If not, you could disable the provider, install the hosted-engine VM, then, if needed, re-add / re-activate it . [0] - https://github.com/oVirt/ovirt-provider-ovn/blob/master/driver/scripts/setup_ovn_controller.sh#L24 > > Regards. > Carl > > On Thu, Jul 18, 2019 at 4:03 AM Miguel Duarte de Mora Barroso > wrote: >> >> On Wed, Jul 17, 2019 at 7:07 PM carl langlois wrote: >> > >> > Hi >> > Here is the output of the command >> > >> > [root@ovhost1 ~]# vdsm-tool --vvverbose ovn-config 10.16.248.74 ovirtmgmt >> > MainThread::DEBUG::2019-07-17 >> > 13:02:52,581::cmdutils::150::root::(exec_cmd) lshw -json -disable usb >> > -disable pcmcia -disable isapnp -disable ide -disable scsi -disable dmi >> > -disable memory -disable cpuinfo (cwd None) >> > MainThread::DEBUG::2019-07-17 >> > 13:02:52,738::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 >> > MainThread::DEBUG::2019-07-17 >> > 13:02:52,741::routes::109::root::(get_gateway) The gateway 10.16.248.1 is >> > duplicated for the device ovirtmgmt >> > MainThread::DEBUG::2019-07-17 >> > 13:02:52,742::routes::109::root::(get_gateway) The gateway 10.16.248.1 is >> > duplicated for the device ovirtmgmt >> > MainThread::DEBUG::2019-07-17 >> > 13:02:52,742::cmdutils::150::root::(exec_cmd) /sbin/tc qdisc show (cwd >> > None) >> > MainThread::DEBUG::2019-07-17 >> > 13:02:52,744::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 >> > MainThread::DEBUG::2019-07-17 >> > 13:02:52,745::cmdutils::150::root::(exec_cmd) /sbin/tc class show dev >> > enp2s0f1 classid 0:1388 (cwd None) >> > MainThread::DEBUG::2019-07-17 >> > 13:02:52,747::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 >> > MainThread::DEBUG::2019-07-17 >> > 13:02:52,766::cmdutils::150::root::(exec_cmd) >> > /usr/share/openvswitch/scripts/ovs-ctl status (cwd None) >> > MainThread::DEBUG::2019-07-17 >> > 13:02:52,777::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 >> > MainThread::DEBUG::2019-07-17 13:02:52,778::vsctl::67::root::(commit) >> > Executing commands: /usr/bin/ovs-vsctl --timeout=5 --oneline --format=json >> > -- list Bridge -- list Port -- list Interface >> > MainThread::DEBUG::2019-07-17 >> > 13:02:52,778::cmdutils::150::root::(exec_cmd) /usr/bin/ovs-vsctl >> > --timeout=5 --oneline --format=json -- list Bridge -- list Port -- list >> > Interface (cwd None) >> > MainThread::DEBUG::2019-07-17 >> > 13:02:52,799::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 >> > netlink/events::DEBUG::2019-07-17 >> > 13:02:52,802::concurrent::192::root::(run) START thread >> > (func=> > method Monitor._scan of > > 0x7f99fb618c90>>, args=(), kwargs={}) >> > netlink/events::DEBUG::2019-07-17 >> > 13:02:54,805::concurrent::195::root::(run) FINISH thread >> > >> > Using default PKI files >> > >> > I do not see any indication of the config?? >> >> And afterwards when you execute "ovs-vsctl list Open_vSwitch" does it >> reflect the updated value ? >> >> This command would have to be performed in the node where hosted >> engine will be hosted - not sure if it's possible to determine before >> hand which one it will be. If not, you should run it in all the nodes >> in the cluster, to be sure. >> >> > >> > Regards >> > Carl >> > >> > On Wed, Jul 17, 2019 at 11:40 AM carl langlois >> > wrote: >> >> >> >> Hi >> >> >> >> I have open a bug https://bugzilla.redhat.com/show_bug.cgi?id=1730776 >> >> >> >> I have try this command "vdsm-tool ovn-config 10.16.248.74 ovirtmgmt" on >> >> one of the host but nothing changed. After a restart of the ovn-controler >> >> i still get >> >> >> >> 2019-07-17T15:38:52.572Z|00033|reconnect|INFO|ssl:10.8.236.244:6642: >> >> waiting 8 seconds before reconnect >> >> 2019-07-17T15:39:00.578Z|00034|reconnect|INFO|ssl:10.8.236.244:6642: >> >> connecting... >> >> 2019-07-17T15:39:05.720Z|00035|fatal_signal|WARN|terminating with signal >> >> 15 (Terminated) >> >> 2019-07-17T15:39:05.863Z|1|vlog|INFO|opened log file >> >> /var/log/openvswitch/ovn-controller.log >> >> 2019-07-17T15:39:05.864Z|2|reconnect|INFO|unix:/var/run/openvswitch/db.sock: >> >> connecting... >> >> 2019-07-17T15:39:05.864Z|3|reconnect|INFO|unix:/var/run/openvswitch/db.sock: >> >> connected >> >> 2019-07-17T15:39:05.865Z
[ovirt-users] Re: major network changes
Hi Miguel, I have managed to change the config for the ovn-controler. with those commands ovs-vsctl set Open_vSwitch . external-ids:ovn-remote=ssl:10.16.248.74:6642 ovs-vsctl set Open_vSwitch . external-ids:ovn-encap-ip=10.16.248.65 and restating the services But even with this i still have the "fail for liveliness check" when starting the ovirt engine. But one thing i notice with our new network is that the reverse DNS does not work(IP -> hostname). The forward is working fine. I am trying to see with our IT why it is not working. Regards. Carl On Thu, Jul 18, 2019 at 4:03 AM Miguel Duarte de Mora Barroso < mdbarr...@redhat.com> wrote: > On Wed, Jul 17, 2019 at 7:07 PM carl langlois > wrote: > > > > Hi > > Here is the output of the command > > > > [root@ovhost1 ~]# vdsm-tool --vvverbose ovn-config 10.16.248.74 > ovirtmgmt > > MainThread::DEBUG::2019-07-17 > 13:02:52,581::cmdutils::150::root::(exec_cmd) lshw -json -disable usb > -disable pcmcia -disable isapnp -disable ide -disable scsi -disable dmi > -disable memory -disable cpuinfo (cwd None) > > MainThread::DEBUG::2019-07-17 > 13:02:52,738::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 > > MainThread::DEBUG::2019-07-17 > 13:02:52,741::routes::109::root::(get_gateway) The gateway 10.16.248.1 is > duplicated for the device ovirtmgmt > > MainThread::DEBUG::2019-07-17 > 13:02:52,742::routes::109::root::(get_gateway) The gateway 10.16.248.1 is > duplicated for the device ovirtmgmt > > MainThread::DEBUG::2019-07-17 > 13:02:52,742::cmdutils::150::root::(exec_cmd) /sbin/tc qdisc show (cwd None) > > MainThread::DEBUG::2019-07-17 > 13:02:52,744::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 > > MainThread::DEBUG::2019-07-17 > 13:02:52,745::cmdutils::150::root::(exec_cmd) /sbin/tc class show dev > enp2s0f1 classid 0:1388 (cwd None) > > MainThread::DEBUG::2019-07-17 > 13:02:52,747::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 > > MainThread::DEBUG::2019-07-17 > 13:02:52,766::cmdutils::150::root::(exec_cmd) > /usr/share/openvswitch/scripts/ovs-ctl status (cwd None) > > MainThread::DEBUG::2019-07-17 > 13:02:52,777::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 > > MainThread::DEBUG::2019-07-17 13:02:52,778::vsctl::67::root::(commit) > Executing commands: /usr/bin/ovs-vsctl --timeout=5 --oneline --format=json > -- list Bridge -- list Port -- list Interface > > MainThread::DEBUG::2019-07-17 > 13:02:52,778::cmdutils::150::root::(exec_cmd) /usr/bin/ovs-vsctl > --timeout=5 --oneline --format=json -- list Bridge -- list Port -- list > Interface (cwd None) > > MainThread::DEBUG::2019-07-17 > 13:02:52,799::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 > > netlink/events::DEBUG::2019-07-17 > 13:02:52,802::concurrent::192::root::(run) START thread > (func= method Monitor._scan of 0x7f99fb618c90>>, args=(), kwargs={}) > > netlink/events::DEBUG::2019-07-17 > 13:02:54,805::concurrent::195::root::(run) FINISH thread > > > Using default PKI files > > > > I do not see any indication of the config?? > > And afterwards when you execute "ovs-vsctl list Open_vSwitch" does it > reflect the updated value ? > > This command would have to be performed in the node where hosted > engine will be hosted - not sure if it's possible to determine before > hand which one it will be. If not, you should run it in all the nodes > in the cluster, to be sure. > > > > > Regards > > Carl > > > > On Wed, Jul 17, 2019 at 11:40 AM carl langlois > wrote: > >> > >> Hi > >> > >> I have open a bug https://bugzilla.redhat.com/show_bug.cgi?id=1730776 > >> > >> I have try this command "vdsm-tool ovn-config 10.16.248.74 ovirtmgmt" > on one of the host but nothing changed. After a restart of the > ovn-controler i still get > >> > >> 2019-07-17T15:38:52.572Z|00033|reconnect|INFO|ssl:10.8.236.244:6642: > waiting 8 seconds before reconnect > >> 2019-07-17T15:39:00.578Z|00034|reconnect|INFO|ssl:10.8.236.244:6642: > connecting... > >> 2019-07-17T15:39:05.720Z|00035|fatal_signal|WARN|terminating with > signal 15 (Terminated) > >> 2019-07-17T15:39:05.863Z|1|vlog|INFO|opened log file > /var/log/openvswitch/ovn-controller.log > >> > 2019-07-17T15:39:05.864Z|2|reconnect|INFO|unix:/var/run/openvswitch/db.sock: > connecting... > >> > 2019-07-17T15:39:05.864Z|3|reconnect|INFO|unix:/var/run/openvswitch/db.sock: > connected > >> 2019-07-17T15:39:05.865Z|4|reconnect|INFO|ssl:10.8.236.244:6642: > connecting... > >> 2019-07-17T15:39:06.865Z|5|reconnect|INFO|ssl:10.8.236.244:6642: > connection attempt timed out > >> 2019-07-17T15:39:06.865Z|6|reconnect|INFO|ssl:10.8.236.244:6642: > waiting 1 seconds before reconnect > >> 2019-07-17T15:39:07.867Z|7|reconnect|INFO|ssl:10.8.236.244:6642: > connecting... > >> 2019-07-17T15:39:08.867Z|8|reconnect|INFO|ssl:10.8.236.244:6642: > connection attempt timed out > >> 2019-07-17T15:39:08.868Z|9|reconnect|INFO|ssl:10.8.236.244:6642: > waiting 2 seconds before reconnect > >> 2019-07-17T15:39:10.870Z|00010|reconnect|IN
[ovirt-users] Re: major network changes
On Wed, Jul 17, 2019 at 7:07 PM carl langlois wrote: > > Hi > Here is the output of the command > > [root@ovhost1 ~]# vdsm-tool --vvverbose ovn-config 10.16.248.74 ovirtmgmt > MainThread::DEBUG::2019-07-17 13:02:52,581::cmdutils::150::root::(exec_cmd) > lshw -json -disable usb -disable pcmcia -disable isapnp -disable ide -disable > scsi -disable dmi -disable memory -disable cpuinfo (cwd None) > MainThread::DEBUG::2019-07-17 13:02:52,738::cmdutils::158::root::(exec_cmd) > SUCCESS: = ''; = 0 > MainThread::DEBUG::2019-07-17 13:02:52,741::routes::109::root::(get_gateway) > The gateway 10.16.248.1 is duplicated for the device ovirtmgmt > MainThread::DEBUG::2019-07-17 13:02:52,742::routes::109::root::(get_gateway) > The gateway 10.16.248.1 is duplicated for the device ovirtmgmt > MainThread::DEBUG::2019-07-17 13:02:52,742::cmdutils::150::root::(exec_cmd) > /sbin/tc qdisc show (cwd None) > MainThread::DEBUG::2019-07-17 13:02:52,744::cmdutils::158::root::(exec_cmd) > SUCCESS: = ''; = 0 > MainThread::DEBUG::2019-07-17 13:02:52,745::cmdutils::150::root::(exec_cmd) > /sbin/tc class show dev enp2s0f1 classid 0:1388 (cwd None) > MainThread::DEBUG::2019-07-17 13:02:52,747::cmdutils::158::root::(exec_cmd) > SUCCESS: = ''; = 0 > MainThread::DEBUG::2019-07-17 13:02:52,766::cmdutils::150::root::(exec_cmd) > /usr/share/openvswitch/scripts/ovs-ctl status (cwd None) > MainThread::DEBUG::2019-07-17 13:02:52,777::cmdutils::158::root::(exec_cmd) > SUCCESS: = ''; = 0 > MainThread::DEBUG::2019-07-17 13:02:52,778::vsctl::67::root::(commit) > Executing commands: /usr/bin/ovs-vsctl --timeout=5 --oneline --format=json -- > list Bridge -- list Port -- list Interface > MainThread::DEBUG::2019-07-17 13:02:52,778::cmdutils::150::root::(exec_cmd) > /usr/bin/ovs-vsctl --timeout=5 --oneline --format=json -- list Bridge -- list > Port -- list Interface (cwd None) > MainThread::DEBUG::2019-07-17 13:02:52,799::cmdutils::158::root::(exec_cmd) > SUCCESS: = ''; = 0 > netlink/events::DEBUG::2019-07-17 13:02:52,802::concurrent::192::root::(run) > START thread > (func= object at 0x7f99fb618c90>>, args=(), kwargs={}) > netlink/events::DEBUG::2019-07-17 13:02:54,805::concurrent::195::root::(run) > FINISH thread > Using default PKI files > > I do not see any indication of the config?? And afterwards when you execute "ovs-vsctl list Open_vSwitch" does it reflect the updated value ? This command would have to be performed in the node where hosted engine will be hosted - not sure if it's possible to determine before hand which one it will be. If not, you should run it in all the nodes in the cluster, to be sure. > > Regards > Carl > > On Wed, Jul 17, 2019 at 11:40 AM carl langlois wrote: >> >> Hi >> >> I have open a bug https://bugzilla.redhat.com/show_bug.cgi?id=1730776 >> >> I have try this command "vdsm-tool ovn-config 10.16.248.74 ovirtmgmt" on one >> of the host but nothing changed. After a restart of the ovn-controler i >> still get >> >> 2019-07-17T15:38:52.572Z|00033|reconnect|INFO|ssl:10.8.236.244:6642: waiting >> 8 seconds before reconnect >> 2019-07-17T15:39:00.578Z|00034|reconnect|INFO|ssl:10.8.236.244:6642: >> connecting... >> 2019-07-17T15:39:05.720Z|00035|fatal_signal|WARN|terminating with signal 15 >> (Terminated) >> 2019-07-17T15:39:05.863Z|1|vlog|INFO|opened log file >> /var/log/openvswitch/ovn-controller.log >> 2019-07-17T15:39:05.864Z|2|reconnect|INFO|unix:/var/run/openvswitch/db.sock: >> connecting... >> 2019-07-17T15:39:05.864Z|3|reconnect|INFO|unix:/var/run/openvswitch/db.sock: >> connected >> 2019-07-17T15:39:05.865Z|4|reconnect|INFO|ssl:10.8.236.244:6642: >> connecting... >> 2019-07-17T15:39:06.865Z|5|reconnect|INFO|ssl:10.8.236.244:6642: >> connection attempt timed out >> 2019-07-17T15:39:06.865Z|6|reconnect|INFO|ssl:10.8.236.244:6642: waiting >> 1 seconds before reconnect >> 2019-07-17T15:39:07.867Z|7|reconnect|INFO|ssl:10.8.236.244:6642: >> connecting... >> 2019-07-17T15:39:08.867Z|8|reconnect|INFO|ssl:10.8.236.244:6642: >> connection attempt timed out >> 2019-07-17T15:39:08.868Z|9|reconnect|INFO|ssl:10.8.236.244:6642: waiting >> 2 seconds before reconnect >> 2019-07-17T15:39:10.870Z|00010|reconnect|INFO|ssl:10.8.236.244:6642: >> connecting... >> 2019-07-17T15:39:12.872Z|00011|reconnect|INFO|ssl:10.8.236.244:6642: >> connection attempt timed out >> 2019-07-17T15:39:12.872Z|00012|reconnect|INFO|ssl:10.8.236.244:6642: waiting >> 4 seconds before reconnect >> >> >> >> On Wed, Jul 17, 2019 at 10:56 AM Miguel Duarte de Mora Barroso >> wrote: >>> >>> On Wed, Jul 17, 2019 at 3:01 PM carl langlois >>> wrote: >>> > >>> > Hi Miguel >>> > >>> > if i do ovs-vsctl list Open_vSwitch i get >>> > >>> > uuid : ce94c4b1-7eb2-42e3-8bfd-96e1dec40dea >>> > bridges : [9b0738ee-594d-4a87-8967-049a8b1a5774] >>> > cur_cfg : 1 >>> > datapath_types : [netdev, system] >>> > db_version : "7.14.0" >>> > external_ids
[ovirt-users] Re: major network changes
Hi Here is the output of the command [root@ovhost1 ~]# vdsm-tool --vvverbose ovn-config 10.16.248.74 ovirtmgmt MainThread::DEBUG::2019-07-17 13:02:52,581::cmdutils::150::root::(exec_cmd) lshw -json -disable usb -disable pcmcia -disable isapnp -disable ide -disable scsi -disable dmi -disable memory -disable cpuinfo (cwd None) MainThread::DEBUG::2019-07-17 13:02:52,738::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 MainThread::DEBUG::2019-07-17 13:02:52,741::routes::109::root::(get_gateway) The gateway 10.16.248.1 is duplicated for the device ovirtmgmt MainThread::DEBUG::2019-07-17 13:02:52,742::routes::109::root::(get_gateway) The gateway 10.16.248.1 is duplicated for the device ovirtmgmt MainThread::DEBUG::2019-07-17 13:02:52,742::cmdutils::150::root::(exec_cmd) /sbin/tc qdisc show (cwd None) MainThread::DEBUG::2019-07-17 13:02:52,744::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 MainThread::DEBUG::2019-07-17 13:02:52,745::cmdutils::150::root::(exec_cmd) /sbin/tc class show dev enp2s0f1 classid 0:1388 (cwd None) MainThread::DEBUG::2019-07-17 13:02:52,747::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 MainThread::DEBUG::2019-07-17 13:02:52,766::cmdutils::150::root::(exec_cmd) /usr/share/openvswitch/scripts/ovs-ctl status (cwd None) MainThread::DEBUG::2019-07-17 13:02:52,777::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 MainThread::DEBUG::2019-07-17 13:02:52,778::vsctl::67::root::(commit) Executing commands: /usr/bin/ovs-vsctl --timeout=5 --oneline --format=json -- list Bridge -- list Port -- list Interface MainThread::DEBUG::2019-07-17 13:02:52,778::cmdutils::150::root::(exec_cmd) /usr/bin/ovs-vsctl --timeout=5 --oneline --format=json -- list Bridge -- list Port -- list Interface (cwd None) MainThread::DEBUG::2019-07-17 13:02:52,799::cmdutils::158::root::(exec_cmd) SUCCESS: = ''; = 0 netlink/events::DEBUG::2019-07-17 13:02:52,802::concurrent::192::root::(run) START thread (func=>, args=(), kwargs={}) netlink/events::DEBUG::2019-07-17 13:02:54,805::concurrent::195::root::(run) FINISH thread Using default PKI files I do not see any indication of the config?? Regards Carl On Wed, Jul 17, 2019 at 11:40 AM carl langlois wrote: > Hi > > I have open a bug https://bugzilla.redhat.com/show_bug.cgi?id=1730776 > > I have try this command "vdsm-tool ovn-config 10.16.248.74 ovirtmgmt" on > one of the host but nothing changed. After a restart of the ovn-controler i > still get > > 2019-07-17T15:38:52.572Z|00033|reconnect|INFO|ssl:10.8.236.244:6642: > waiting 8 seconds before reconnect > 2019-07-17T15:39:00.578Z|00034|reconnect|INFO|ssl:10.8.236.244:6642: > connecting... > 2019-07-17T15:39:05.720Z|00035|fatal_signal|WARN|terminating with signal > 15 (Terminated) > 2019-07-17T15:39:05.863Z|1|vlog|INFO|opened log file > /var/log/openvswitch/ovn-controller.log > 2019-07-17T15:39:05.864Z|2|reconnect|INFO|unix:/var/run/openvswitch/db.sock: > connecting... > 2019-07-17T15:39:05.864Z|3|reconnect|INFO|unix:/var/run/openvswitch/db.sock: > connected > 2019-07-17T15:39:05.865Z|4|reconnect|INFO|ssl:10.8.236.244:6642: > connecting... > 2019-07-17T15:39:06.865Z|5|reconnect|INFO|ssl:10.8.236.244:6642: > connection attempt timed out > 2019-07-17T15:39:06.865Z|6|reconnect|INFO|ssl:10.8.236.244:6642: > waiting 1 seconds before reconnect > 2019-07-17T15:39:07.867Z|7|reconnect|INFO|ssl:10.8.236.244:6642: > connecting... > 2019-07-17T15:39:08.867Z|8|reconnect|INFO|ssl:10.8.236.244:6642: > connection attempt timed out > 2019-07-17T15:39:08.868Z|9|reconnect|INFO|ssl:10.8.236.244:6642: > waiting 2 seconds before reconnect > 2019-07-17T15:39:10.870Z|00010|reconnect|INFO|ssl:10.8.236.244:6642: > connecting... > 2019-07-17T15:39:12.872Z|00011|reconnect|INFO|ssl:10.8.236.244:6642: > connection attempt timed out > 2019-07-17T15:39:12.872Z|00012|reconnect|INFO|ssl:10.8.236.244:6642: > waiting 4 seconds before reconnect > > > > On Wed, Jul 17, 2019 at 10:56 AM Miguel Duarte de Mora Barroso < > mdbarr...@redhat.com> wrote: > >> On Wed, Jul 17, 2019 at 3:01 PM carl langlois >> wrote: >> > >> > Hi Miguel >> > >> > if i do ovs-vsctl list Open_vSwitch i get >> > >> > uuid : ce94c4b1-7eb2-42e3-8bfd-96e1dec40dea >> > bridges : [9b0738ee-594d-4a87-8967-049a8b1a5774] >> > cur_cfg : 1 >> > datapath_types : [netdev, system] >> > db_version : "7.14.0" >> > external_ids: {hostname="ovhost2", ovn-bridge-mappings="", >> ovn-encap-ip="10.8.236.150", ovn-encap-type=geneve, ovn-remote="ssl: >> 10.8.236.244:6642", system-id="7c39d07b-1d54-417b-bf56-7a0f1a07f832"} >> > iface_types : [geneve, gre, internal, lisp, patch, stt, system, >> tap, vxlan] >> > manager_options : [] >> > next_cfg: 1 >> > other_config: {} >> > ovs_version : "2.7.3" >> > ssl : [] >> > statistics : {} >> > system_type : centos >> > system_version : "7" >> > >> > I can see two addresses that
[ovirt-users] Re: major network changes
Hi I have open a bug https://bugzilla.redhat.com/show_bug.cgi?id=1730776 I have try this command "vdsm-tool ovn-config 10.16.248.74 ovirtmgmt" on one of the host but nothing changed. After a restart of the ovn-controler i still get 2019-07-17T15:38:52.572Z|00033|reconnect|INFO|ssl:10.8.236.244:6642: waiting 8 seconds before reconnect 2019-07-17T15:39:00.578Z|00034|reconnect|INFO|ssl:10.8.236.244:6642: connecting... 2019-07-17T15:39:05.720Z|00035|fatal_signal|WARN|terminating with signal 15 (Terminated) 2019-07-17T15:39:05.863Z|1|vlog|INFO|opened log file /var/log/openvswitch/ovn-controller.log 2019-07-17T15:39:05.864Z|2|reconnect|INFO|unix:/var/run/openvswitch/db.sock: connecting... 2019-07-17T15:39:05.864Z|3|reconnect|INFO|unix:/var/run/openvswitch/db.sock: connected 2019-07-17T15:39:05.865Z|4|reconnect|INFO|ssl:10.8.236.244:6642: connecting... 2019-07-17T15:39:06.865Z|5|reconnect|INFO|ssl:10.8.236.244:6642: connection attempt timed out 2019-07-17T15:39:06.865Z|6|reconnect|INFO|ssl:10.8.236.244:6642: waiting 1 seconds before reconnect 2019-07-17T15:39:07.867Z|7|reconnect|INFO|ssl:10.8.236.244:6642: connecting... 2019-07-17T15:39:08.867Z|8|reconnect|INFO|ssl:10.8.236.244:6642: connection attempt timed out 2019-07-17T15:39:08.868Z|9|reconnect|INFO|ssl:10.8.236.244:6642: waiting 2 seconds before reconnect 2019-07-17T15:39:10.870Z|00010|reconnect|INFO|ssl:10.8.236.244:6642: connecting... 2019-07-17T15:39:12.872Z|00011|reconnect|INFO|ssl:10.8.236.244:6642: connection attempt timed out 2019-07-17T15:39:12.872Z|00012|reconnect|INFO|ssl:10.8.236.244:6642: waiting 4 seconds before reconnect On Wed, Jul 17, 2019 at 10:56 AM Miguel Duarte de Mora Barroso < mdbarr...@redhat.com> wrote: > On Wed, Jul 17, 2019 at 3:01 PM carl langlois > wrote: > > > > Hi Miguel > > > > if i do ovs-vsctl list Open_vSwitch i get > > > > uuid : ce94c4b1-7eb2-42e3-8bfd-96e1dec40dea > > bridges : [9b0738ee-594d-4a87-8967-049a8b1a5774] > > cur_cfg : 1 > > datapath_types : [netdev, system] > > db_version : "7.14.0" > > external_ids: {hostname="ovhost2", ovn-bridge-mappings="", > ovn-encap-ip="10.8.236.150", ovn-encap-type=geneve, ovn-remote="ssl: > 10.8.236.244:6642", system-id="7c39d07b-1d54-417b-bf56-7a0f1a07f832"} > > iface_types : [geneve, gre, internal, lisp, patch, stt, system, > tap, vxlan] > > manager_options : [] > > next_cfg: 1 > > other_config: {} > > ovs_version : "2.7.3" > > ssl : [] > > statistics : {} > > system_type : centos > > system_version : "7" > > > > I can see two addresses that are on the old network.. > > Yes, those are it. > > Use the tool I mentioned to update that to the correct addresses on > the network, and re-try. > > vdsm-tool ovn-config > > > Regards > > Carl > > > > > > On Wed, Jul 17, 2019 at 8:21 AM carl langlois > wrote: > >> > >> Hi Miguel, > >> > >> I will surely open a bugs, any specific ovirt componenent to select > when openeing the bug? > > ovirt-engine > > >> > >> When you say that the hosted-engine should have trigger a the update. > Do you mean is was suppose to trigger the update and did not work or it is > something missing? > > I sincerely do not know. @Dominik Holler, could you shed some light into > this ? > > >> Could i have missed a step when switching the network? > >> > >> Also if i try to do ovs-vsctl list . The list command require a Table > name. Not sure what table to use? > >> > >> Regards > >> Carl > >> > >> > >> > >> On Wed, Jul 17, 2019 at 4:21 AM Miguel Duarte de Mora Barroso < > mdbarr...@redhat.com> wrote: > >>> > >>> On Tue, Jul 16, 2019 at 8:48 PM carl langlois > wrote: > >>> > > >>> > Hi > >>> > > >>> > We are in a process of changing our network connection. Our current > network is using 10.8.256.x and we will change to 10.16.248.x. We have a HA > ovirt cluster (around 10 nodes) currently configure on the 10.8.256.x. So > my question is is it possible to relocate the ovirt cluster to the > 10.16.248.x. We have tried to move everything to the new network without > success. All the node seem to boot up properly, our gluster storage also > work properly. > >>> > When we try to start the hosted-engine it goes up but fail the > liveliness check. We have notice in the > /var/log/openvswitch/ovn-controller.log that he is triying to connect to > the hold ip address of the hosted-engine vm. > >>> > 019-07-16T18:41:29.483Z|01992|reconnect|INFO|ssl:10.8.236.244:6642: > waiting 8 seconds before reconnect > >>> > 2019-07-16T18:41:37.489Z|01993|reconnect|INFO|ssl:10.8.236.244:6642: > connecting... > >>> > 2019-07-16T18:41:45.497Z|01994|reconnect|INFO|ssl:10.8.236.244:6642: > connection attempt timed out > >>> > > >>> > So my question is were is the 10.8.236.244 come from. > >>> > >>> Looks like the ovn controllers were not updated during the network > change. > >>> > >>> The wrong IP is configured with
[ovirt-users] Re: major network changes
On Wed, Jul 17, 2019 at 3:01 PM carl langlois wrote: > > Hi Miguel > > if i do ovs-vsctl list Open_vSwitch i get > > uuid : ce94c4b1-7eb2-42e3-8bfd-96e1dec40dea > bridges : [9b0738ee-594d-4a87-8967-049a8b1a5774] > cur_cfg : 1 > datapath_types : [netdev, system] > db_version : "7.14.0" > external_ids: {hostname="ovhost2", ovn-bridge-mappings="", > ovn-encap-ip="10.8.236.150", ovn-encap-type=geneve, > ovn-remote="ssl:10.8.236.244:6642", > system-id="7c39d07b-1d54-417b-bf56-7a0f1a07f832"} > iface_types : [geneve, gre, internal, lisp, patch, stt, system, tap, > vxlan] > manager_options : [] > next_cfg: 1 > other_config: {} > ovs_version : "2.7.3" > ssl : [] > statistics : {} > system_type : centos > system_version : "7" > > I can see two addresses that are on the old network.. Yes, those are it. Use the tool I mentioned to update that to the correct addresses on the network, and re-try. vdsm-tool ovn-config > Regards > Carl > > > On Wed, Jul 17, 2019 at 8:21 AM carl langlois wrote: >> >> Hi Miguel, >> >> I will surely open a bugs, any specific ovirt componenent to select when >> openeing the bug? ovirt-engine >> >> When you say that the hosted-engine should have trigger a the update. Do you >> mean is was suppose to trigger the update and did not work or it is >> something missing? I sincerely do not know. @Dominik Holler, could you shed some light into this ? >> Could i have missed a step when switching the network? >> >> Also if i try to do ovs-vsctl list . The list command require a Table name. >> Not sure what table to use? >> >> Regards >> Carl >> >> >> >> On Wed, Jul 17, 2019 at 4:21 AM Miguel Duarte de Mora Barroso >> wrote: >>> >>> On Tue, Jul 16, 2019 at 8:48 PM carl langlois >>> wrote: >>> > >>> > Hi >>> > >>> > We are in a process of changing our network connection. Our current >>> > network is using 10.8.256.x and we will change to 10.16.248.x. We have a >>> > HA ovirt cluster (around 10 nodes) currently configure on the 10.8.256.x. >>> > So my question is is it possible to relocate the ovirt cluster to the >>> > 10.16.248.x. We have tried to move everything to the new network without >>> > success. All the node seem to boot up properly, our gluster storage also >>> > work properly. >>> > When we try to start the hosted-engine it goes up but fail the liveliness >>> > check. We have notice in the /var/log/openvswitch/ovn-controller.log that >>> > he is triying to connect to the hold ip address of the hosted-engine vm. >>> > 019-07-16T18:41:29.483Z|01992|reconnect|INFO|ssl:10.8.236.244:6642: >>> > waiting 8 seconds before reconnect >>> > 2019-07-16T18:41:37.489Z|01993|reconnect|INFO|ssl:10.8.236.244:6642: >>> > connecting... >>> > 2019-07-16T18:41:45.497Z|01994|reconnect|INFO|ssl:10.8.236.244:6642: >>> > connection attempt timed out >>> > >>> > So my question is were is the 10.8.236.244 come from. >>> >>> Looks like the ovn controllers were not updated during the network change. >>> >>> The wrong IP is configured within openvswitch, you can see it in the >>> (offending) nodes through "ovs-vsctl list . ". It'll be a key in the >>> 'external_ids' column called 'ovn-remote' . >>> >>> This is not the solution, but a work-around; you could try to >>> configure the ovn controllers via: >>> vdsm-tool ovn-config >>> >>> Despite the provided work-around, I really think the hosted engine >>> should have triggered the ansible role that in turn triggers this >>> reconfiguration. >>> >>> Would you open a bug with this information ? >>> >>> >>> > >>> > The routing table for one of our host look like this >>> > >>> > estination Gateway Genmask Flags Metric RefUse >>> > Iface >>> > default gateway 0.0.0.0 UG0 00 >>> > ovirtmgmt >>> > 10.16.248.0 0.0.0.0 255.255.255.0 U 0 00 >>> > ovirtmgmt >>> > link-local 0.0.0.0 255.255.0.0 U 1002 00 >>> > eno1 >>> > link-local 0.0.0.0 255.255.0.0 U 1003 00 >>> > eno2 >>> > link-local 0.0.0.0 255.255.0.0 U 1025 00 >>> > ovirtmgmt >>> > >>> > Any help would be really appreciated. >>> > >>> > Regards >>> > Carl >>> > >>> > >>> > >>> > >>> > ___ >>> > Users mailing list -- users@ovirt.org >>> > To unsubscribe send an email to users-le...@ovirt.org >>> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>> > oVirt Code of Conduct: >>> > https://www.ovirt.org/community/about/community-guidelines/ >>> > List Archives: >>> > https://lists.ovirt.org/archives/list/users@ovirt.org/message/DBQUWEPPDK2JDFU4HOGNURK7AB3FDINC/ ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy
[ovirt-users] Re: major network changes
Hi Miguel if i do ovs-vsctl list Open_vSwitch i get uuid : ce94c4b1-7eb2-42e3-8bfd-96e1dec40dea bridges : [9b0738ee-594d-4a87-8967-049a8b1a5774] cur_cfg : 1 datapath_types : [netdev, system] db_version : "7.14.0" external_ids: {hostname="ovhost2", ovn-bridge-mappings="", ovn-encap-ip="10.8.236.150", ovn-encap-type=geneve, ovn-remote="ssl: 10.8.236.244:6642", system-id="7c39d07b-1d54-417b-bf56-7a0f1a07f832"} iface_types : [geneve, gre, internal, lisp, patch, stt, system, tap, vxlan] manager_options : [] next_cfg: 1 other_config: {} ovs_version : "2.7.3" ssl : [] statistics : {} system_type : centos system_version : "7" I can see two addresses that are on the old network.. Regards Carl On Wed, Jul 17, 2019 at 8:21 AM carl langlois wrote: > Hi Miguel, > > I will surely open a bugs, any specific ovirt componenent to select when > openeing the bug? > > When you say that the hosted-engine should have trigger a the update. Do > you mean is was suppose to trigger the update and did not work or it is > something missing? > Could i have missed a step when switching the network? > > Also if i try to do ovs-vsctl list . The list command require a Table > name. Not sure what table to use? > > Regards > Carl > > > > On Wed, Jul 17, 2019 at 4:21 AM Miguel Duarte de Mora Barroso < > mdbarr...@redhat.com> wrote: > >> On Tue, Jul 16, 2019 at 8:48 PM carl langlois >> wrote: >> > >> > Hi >> > >> > We are in a process of changing our network connection. Our current >> network is using 10.8.256.x and we will change to 10.16.248.x. We have a HA >> ovirt cluster (around 10 nodes) currently configure on the 10.8.256.x. So >> my question is is it possible to relocate the ovirt cluster to the >> 10.16.248.x. We have tried to move everything to the new network without >> success. All the node seem to boot up properly, our gluster storage also >> work properly. >> > When we try to start the hosted-engine it goes up but fail the >> liveliness check. We have notice in the >> /var/log/openvswitch/ovn-controller.log that he is triying to connect to >> the hold ip address of the hosted-engine vm. >> > 019-07-16T18:41:29.483Z|01992|reconnect|INFO|ssl:10.8.236.244:6642: >> waiting 8 seconds before reconnect >> > 2019-07-16T18:41:37.489Z|01993|reconnect|INFO|ssl:10.8.236.244:6642: >> connecting... >> > 2019-07-16T18:41:45.497Z|01994|reconnect|INFO|ssl:10.8.236.244:6642: >> connection attempt timed out >> > >> > So my question is were is the 10.8.236.244 come from. >> >> Looks like the ovn controllers were not updated during the network change. >> >> The wrong IP is configured within openvswitch, you can see it in the >> (offending) nodes through "ovs-vsctl list . ". It'll be a key in the >> 'external_ids' column called 'ovn-remote' . >> >> This is not the solution, but a work-around; you could try to >> configure the ovn controllers via: >> vdsm-tool ovn-config >> >> Despite the provided work-around, I really think the hosted engine >> should have triggered the ansible role that in turn triggers this >> reconfiguration. >> >> Would you open a bug with this information ? >> >> >> > >> > The routing table for one of our host look like this >> > >> > estination Gateway Genmask Flags Metric RefUse >> Iface >> > default gateway 0.0.0.0 UG0 00 >> ovirtmgmt >> > 10.16.248.0 0.0.0.0 255.255.255.0 U 0 00 >> ovirtmgmt >> > link-local 0.0.0.0 255.255.0.0 U 1002 00 >> eno1 >> > link-local 0.0.0.0 255.255.0.0 U 1003 00 >> eno2 >> > link-local 0.0.0.0 255.255.0.0 U 1025 00 >> ovirtmgmt >> > >> > Any help would be really appreciated. >> > >> > Regards >> > Carl >> > >> > >> > >> > >> > ___ >> > Users mailing list -- users@ovirt.org >> > To unsubscribe send an email to users-le...@ovirt.org >> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >> > oVirt Code of Conduct: >> https://www.ovirt.org/community/about/community-guidelines/ >> > List Archives: >> https://lists.ovirt.org/archives/list/users@ovirt.org/message/DBQUWEPPDK2JDFU4HOGNURK7AB3FDINC/ >> > ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/2G5ONK6TBKOZSB4YGH2AKF5KWFEKM32U/
[ovirt-users] Re: major network changes
Hi Miguel, I will surely open a bugs, any specific ovirt componenent to select when openeing the bug? When you say that the hosted-engine should have trigger a the update. Do you mean is was suppose to trigger the update and did not work or it is something missing? Could i have missed a step when switching the network? Also if i try to do ovs-vsctl list . The list command require a Table name. Not sure what table to use? Regards Carl On Wed, Jul 17, 2019 at 4:21 AM Miguel Duarte de Mora Barroso < mdbarr...@redhat.com> wrote: > On Tue, Jul 16, 2019 at 8:48 PM carl langlois > wrote: > > > > Hi > > > > We are in a process of changing our network connection. Our current > network is using 10.8.256.x and we will change to 10.16.248.x. We have a HA > ovirt cluster (around 10 nodes) currently configure on the 10.8.256.x. So > my question is is it possible to relocate the ovirt cluster to the > 10.16.248.x. We have tried to move everything to the new network without > success. All the node seem to boot up properly, our gluster storage also > work properly. > > When we try to start the hosted-engine it goes up but fail the > liveliness check. We have notice in the > /var/log/openvswitch/ovn-controller.log that he is triying to connect to > the hold ip address of the hosted-engine vm. > > 019-07-16T18:41:29.483Z|01992|reconnect|INFO|ssl:10.8.236.244:6642: > waiting 8 seconds before reconnect > > 2019-07-16T18:41:37.489Z|01993|reconnect|INFO|ssl:10.8.236.244:6642: > connecting... > > 2019-07-16T18:41:45.497Z|01994|reconnect|INFO|ssl:10.8.236.244:6642: > connection attempt timed out > > > > So my question is were is the 10.8.236.244 come from. > > Looks like the ovn controllers were not updated during the network change. > > The wrong IP is configured within openvswitch, you can see it in the > (offending) nodes through "ovs-vsctl list . ". It'll be a key in the > 'external_ids' column called 'ovn-remote' . > > This is not the solution, but a work-around; you could try to > configure the ovn controllers via: > vdsm-tool ovn-config > > Despite the provided work-around, I really think the hosted engine > should have triggered the ansible role that in turn triggers this > reconfiguration. > > Would you open a bug with this information ? > > > > > > The routing table for one of our host look like this > > > > estination Gateway Genmask Flags Metric RefUse > Iface > > default gateway 0.0.0.0 UG0 00 > ovirtmgmt > > 10.16.248.0 0.0.0.0 255.255.255.0 U 0 00 > ovirtmgmt > > link-local 0.0.0.0 255.255.0.0 U 1002 00 > eno1 > > link-local 0.0.0.0 255.255.0.0 U 1003 00 > eno2 > > link-local 0.0.0.0 255.255.0.0 U 1025 00 > ovirtmgmt > > > > Any help would be really appreciated. > > > > Regards > > Carl > > > > > > > > > > ___ > > Users mailing list -- users@ovirt.org > > To unsubscribe send an email to users-le...@ovirt.org > > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > > List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/DBQUWEPPDK2JDFU4HOGNURK7AB3FDINC/ > ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/6NG5DUQXMBPEQKOR2HKP5VSAU5OPR3US/
[ovirt-users] Re: major network changes
On Tue, Jul 16, 2019 at 8:48 PM carl langlois wrote: > > Hi > > We are in a process of changing our network connection. Our current network > is using 10.8.256.x and we will change to 10.16.248.x. We have a HA ovirt > cluster (around 10 nodes) currently configure on the 10.8.256.x. So my > question is is it possible to relocate the ovirt cluster to the 10.16.248.x. > We have tried to move everything to the new network without success. All the > node seem to boot up properly, our gluster storage also work properly. > When we try to start the hosted-engine it goes up but fail the liveliness > check. We have notice in the /var/log/openvswitch/ovn-controller.log that he > is triying to connect to the hold ip address of the hosted-engine vm. > 019-07-16T18:41:29.483Z|01992|reconnect|INFO|ssl:10.8.236.244:6642: waiting 8 > seconds before reconnect > 2019-07-16T18:41:37.489Z|01993|reconnect|INFO|ssl:10.8.236.244:6642: > connecting... > 2019-07-16T18:41:45.497Z|01994|reconnect|INFO|ssl:10.8.236.244:6642: > connection attempt timed out > > So my question is were is the 10.8.236.244 come from. Looks like the ovn controllers were not updated during the network change. The wrong IP is configured within openvswitch, you can see it in the (offending) nodes through "ovs-vsctl list . ". It'll be a key in the 'external_ids' column called 'ovn-remote' . This is not the solution, but a work-around; you could try to configure the ovn controllers via: vdsm-tool ovn-config Despite the provided work-around, I really think the hosted engine should have triggered the ansible role that in turn triggers this reconfiguration. Would you open a bug with this information ? > > The routing table for one of our host look like this > > estination Gateway Genmask Flags Metric RefUse Iface > default gateway 0.0.0.0 UG0 00 > ovirtmgmt > 10.16.248.0 0.0.0.0 255.255.255.0 U 0 00 > ovirtmgmt > link-local 0.0.0.0 255.255.0.0 U 1002 00 eno1 > link-local 0.0.0.0 255.255.0.0 U 1003 00 eno2 > link-local 0.0.0.0 255.255.0.0 U 1025 00 > ovirtmgmt > > Any help would be really appreciated. > > Regards > Carl > > > > > ___ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/DBQUWEPPDK2JDFU4HOGNURK7AB3FDINC/ ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/YGT2TEE7PBVNLZGKN7TSF77K7V7A3H2R/