[ovirt-users] Re: Please Help, Ovirt Node Hosted Engine Deployment Problems 4.3.2

2020-01-14 Thread William Smyth
I'm running into the exact same issue as described in this thread. The
HostedEngine VM (HE) loses connectivity during the deployment while on the
NAT'd network and the deployment fails. My HE VM is never migrated to
shared storage because that part of the process hasn't been reached, nor
have I been asked for the shared storage details, which are simple NFS
shares. I've tried the deployment from the CLI and through Cockpit, both
times using the Ansible deployment, and both times reaching the same
conclusion.

I did manage to connect to the Centos 7 host that is running the HE guest
with virt-manager and moved the NIC to the 'ovirtmgmt' vswitch using
macvtap, which allowed me to access HE from the outside world. One of the
big issues though is that the HE isn't on shared storage and I'm not sure
how to move it there. I'm running 2 Centos 7 hosts with different CPU
makes, AMD and Intel. So i have two separate single-host clusters but I'm
trying to at least have the HA setup for the HE, which I think is still
possible. I know that I won't be able to do a live migration, but that's
not a big deal. I'm afraid my failed deployment of HE is preventing this
and I'm at an impasse. I don't want to give up on oVirt but it's becoming
difficult to get my lab environment operational and I'm wasting a lot of
time troubleshooting.

Any help or suggestions would be greatly appreciated.


William Smyth
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/DJY22FPYIQIXU62U4F7GQ7ZH2RDI7IBW/


[ovirt-users] Re: Please Help, Ovirt Node Hosted Engine Deployment Problems 4.3.2

2019-05-06 Thread Simone Tiraboschi
On Fri, May 3, 2019 at 8:14 PM Todd Barton 
wrote:

> Simone,
>
> It appears 192.168.122.13 stops routing correctly during the final stage
> of deployment.  After a failure of final stage, I can restart the
> hosted-engine VM from the cockpit and I can ping 192.168.122.13 from the
> Host again.  If I retry the final stage of deployment again, 192.168.122.13
> stop routing correctly from Host during that process.  Below are two ping
> commands...the 1st one is after deploy failure (screen shot previous email)
> and the second one is after force-restarting the hosted-engine VM.
>
> [root@ovirt-dr-standalone ~]# ping 192.168.122.13
> PING 192.168.122.13 (192.168.122.13) 56(84) bytes of data.
> From 192.168.122.1 icmp_seq=21 Destination Host Unreachable
> From 192.168.122.1 icmp_seq=22 Destination Host Unreachable
> From 192.168.122.1 icmp_seq=23 Destination Host Unreachable
> From 192.168.122.1 icmp_seq=24 Destination Host Unreachable
> From 192.168.122.1 icmp_seq=25 Destination Host Unreachable
> From 192.168.122.1 icmp_seq=26 Destination Host Unreachable
> From 192.168.122.1 icmp_seq=27 Destination Host Unreachable
> ^C
> --- 192.168.122.13 ping statistics ---
> 40 packets transmitted, 0 received, +7 errors, 100% packet loss, time
> 39041ms pipe 4
>
> [root@ovirt-dr-standalone ~]# ping 192.168.122.13
> PING 192.168.122.13 (192.168.122.13) 56(84) bytes of data.
> 64 bytes from 192.168.122.13: icmp_seq=2 ttl=64 time=0.560 ms
> 64 bytes from 192.168.122.13: icmp_seq=3 ttl=64 time=0.592 ms
> 64 bytes from 192.168.122.13: icmp_seq=4 ttl=64 time=0.345 ms
> 64 bytes from 192.168.122.13: icmp_seq=5 ttl=64 time=0.265 ms
> 64 bytes from 192.168.122.13: icmp_seq=6 ttl=64 time=0.374 ms
> 64 bytes from 192.168.122.13: icmp_seq=7 ttl=64 time=0.390 ms
> 64 bytes from 192.168.122.13: icmp_seq=8 ttl=64 time=0.635 ms
> 64 bytes from 192.168.122.13: icmp_seq=9 ttl=64 time=0.466 ms
> 64 bytes from 192.168.122.13: icmp_seq=10 ttl=64 time=0.376 ms
> 64 bytes from 192.168.122.13: icmp_seq=11 ttl=64 time=0.435 ms
> 64 bytes from 192.168.122.13: icmp_seq=12 ttl=64 time=0.567 ms
> 64 bytes from 192.168.122.13: icmp_seq=13 ttl=64 time=0.442 ms
> 64 bytes from 192.168.122.13: icmp_seq=14 ttl=64 time=0.402 ms
> ^C
> --- 192.168.122.13 ping statistics ---
> 14 packets transmitted, 13 received, 7% packet loss, time 13000ms rtt
> min/avg/max/mdev = 0.265/0.449/0.635/0.108 ms
>
>
> This appears to roughly be the same issue as preparing the vm...the
> network setup goes "off the reservation" when the deployment is making
> changes.  Maybe this is something caused by the virtualization setup, but
> I've read about others doing ovirt  hosts in VMs (like
> https://www.ovirt.org/blog/2018/02/up-and-running-with-ovirt-4-2-and-gluster-storage.html
> ).
>

I tried reproducing this with oVirt node nested over KVM and everything
worked as expected for me.
Honestly I'm not hyper-v expert but I'd suggest to try changing something
(not sure exactly what) about the network definition on hyper-v side.



>
> Any suggestions?  I'm getting to the point where I may need to throw in
> the towel on this setup, but it would be greatly advantageous to have a VM
> lab so I can test changes/upgrades.  I would love to find a way to make
> this work.
>
> *Todd Barton*
>
>
>
>  On Fri, 03 May 2019 12:04:47 -0400 *Simone Tiraboschi
> >* wrote 
>
>
>
> On Fri, May 3, 2019 at 5:27 PM Todd Barton <
> tcbar...@ipvoicedatasystems.com> wrote:
>
>
> Simone/Dominik,
>
> Double reply below and more info with latest attempt.
>
> ---
>
> Simone...answers to your questions, using CAPS to make my responses easier
> to see/read.
>
> "If I correctly understood you environment you have:
> - A pfsense software firewall/router on 10.1.1.1
> - Your host on 10.1.1.61
> - You are accessing cockpit from a browser running on a machine on
> 10.1.1.101 on the same subnet"
>
> YES
>
> "And the issue is that once the engine created the management bridge, your
> client machine on 10.1.1.101 wasn't able anymore to reach your host on
> 10.1.1.61. Am I right?"
>
> YES
>
> "In this case the default gateway or other routes should't be an issue
> since your client is inside the same subnet."
>
> CORRECT, IMO
>
> "Do you think we are loosing some piece of your network configuration
> creating the management bridge such as a custom MTU or a VLAN id or
> something like that?"
>
> NO CUSTOM SETUP HERE...RUNNING A PLAIN/BASIC NETWORK
>
> "Do you think pfsense can start blocking/dropping the traffic for any
> reason?"
>
> NO, ONLY USING PFSENSE TO PROVIDE DHCP AND DNS IN TEST/LAB ENVIRONMENT.
>
> "Todd, another hint: how did you exposed cockpit over firewalld?
> Maybe something in firewalld zone configuration?"
>
> I'M USING THE OVIRT NODE MINIMAL/UTILITY INSTALL AND I DIDN'T DO ANY
> CUSTOMIZATION OUTSIDE OF THE INSTALL/SETUP.  IF ITS FIREWALLD, THEN ITS NOT
> SOMETHING THE NODE SETUP DID PROPERLY.  I CAN SEND CONFIG INFO IF YOU WOULD
> LIKE SEE IT, BUT THERE ARE MORE DEVELOPMENTS/INFO

[ovirt-users] Re: Please Help, Ovirt Node Hosted Engine Deployment Problems 4.3.2

2019-05-03 Thread Todd Barton
Simone,



It appears 192.168.122.13 stops routing correctly during the final stage of 
deployment.  After a failure of final stage, I can restart the hosted-engine VM 
from the cockpit and I can ping 192.168.122.13 from the Host again.  If I retry 
the final stage of deployment again, 192.168.122.13 stop routing correctly from 
Host during that process.  Below are two ping commands...the 1st one is after 
deploy failure (screen shot previous email) and the second one is after 
force-restarting the hosted-engine VM.  



[root@ovirt-dr-standalone ~]# ping 192.168.122.13 


PING 192.168.122.13 (192.168.122.13) 56(84) bytes of data.

>From 192.168.122.1 icmp_seq=21 Destination Host Unreachable

>From 192.168.122.1 icmp_seq=22 Destination Host Unreachable

>From 192.168.122.1 icmp_seq=23 Destination Host Unreachable

>From 192.168.122.1 icmp_seq=24 Destination Host Unreachable

>From 192.168.122.1 icmp_seq=25 Destination Host Unreachable

>From 192.168.122.1 icmp_seq=26 Destination Host Unreachable

>From 192.168.122.1 icmp_seq=27 Destination Host Unreachable

^C

--- 192.168.122.13 ping statistics ---

40 packets transmitted, 0 received, +7 errors, 100% packet loss, time 39041ms 
pipe 4



[root@ovirt-dr-standalone ~]# ping 192.168.122.13

PING 192.168.122.13 (192.168.122.13) 56(84) bytes of data.

64 bytes from 192.168.122.13: icmp_seq=2 ttl=64 time=0.560 ms 

64 bytes from 192.168.122.13: icmp_seq=3 ttl=64 time=0.592 ms

64 bytes from 192.168.122.13: icmp_seq=4 ttl=64 time=0.345 ms

64 bytes from 192.168.122.13: icmp_seq=5 ttl=64 time=0.265 ms

64 bytes from 192.168.122.13: icmp_seq=6 ttl=64 time=0.374 ms

64 bytes from 192.168.122.13: icmp_seq=7 ttl=64 time=0.390 ms

64 bytes from 192.168.122.13: icmp_seq=8 ttl=64 time=0.635 ms

64 bytes from 192.168.122.13: icmp_seq=9 ttl=64 time=0.466 ms

64 bytes from 192.168.122.13: icmp_seq=10 ttl=64 time=0.376 ms

64 bytes from 192.168.122.13: icmp_seq=11 ttl=64 time=0.435 ms

64 bytes from 192.168.122.13: icmp_seq=12 ttl=64 time=0.567 ms

64 bytes from 192.168.122.13: icmp_seq=13 ttl=64 time=0.442 ms

64 bytes from 192.168.122.13: icmp_seq=14 ttl=64 time=0.402 ms

^C

--- 192.168.122.13 ping statistics ---

14 packets transmitted, 13 received, 7% packet loss, time 13000ms rtt 
min/avg/max/mdev = 0.265/0.449/0.635/0.108 ms





This appears to roughly be the same issue as preparing the vm...the network 
setup goes "off the reservation" when the deployment is making changes.  Maybe 
this is something caused by the virtualization setup, but I've read about 
others doing ovirt  hosts in VMs (like 
https://www.ovirt.org/blog/2018/02/up-and-running-with-ovirt-4-2-and-gluster-storage.html).



Any suggestions?  I'm getting to the point where I may need to throw in the 
towel on this setup, but it would be greatly advantageous to have a VM lab so I 
can test changes/upgrades.  I would love to find a way to make this work.

 

Todd Barton








 On Fri, 03 May 2019 12:04:47 -0400 Simone Tiraboschi  
wrote 







On Fri, May 3, 2019 at 5:27 PM Todd Barton 
 wrote:



Simone/Dominik,



Double reply below and more info with latest attempt.



---



Simone...answers to your questions, using CAPS to make my responses easier to 
see/read.



"If I correctly understood you environment you have:

- A pfsense software firewall/router on 10.1.1.1

- Your host on 10.1.1.61

- You are accessing cockpit from a browser running on a machine on 10.1.1.101 
on the same subnet"



YES



"And the issue is that once the engine created the management bridge, your 
client machine on 10.1.1.101 wasn't able anymore to reach your host on 
10.1.1.61. Am I right?"



YES



"In this case the default gateway or other routes should't be an issue since 
your client is inside the same subnet."



CORRECT, IMO



"Do you think we are loosing some piece of your network configuration creating 
the management bridge such as a custom MTU or a VLAN id or something like that?"

 

NO CUSTOM SETUP HERE...RUNNING A PLAIN/BASIC NETWORK



"Do you think pfsense can start blocking/dropping the traffic for any reason?"



NO, ONLY USING PFSENSE TO PROVIDE DHCP AND DNS IN TEST/LAB ENVIRONMENT.



"Todd, another hint: how did you exposed cockpit over firewalld?

Maybe something in firewalld zone configuration?"



I'M USING THE OVIRT NODE MINIMAL/UTILITY INSTALL AND I DIDN'T DO ANY 
CUSTOMIZATION OUTSIDE OF THE INSTALL/SETUP.  IF ITS FIREWALLD, THEN ITS NOT 
SOMETHING THE NODE SETUP DID PROPERLY.  I CAN SEND CONFIG INFO IF YOU WOULD 
LIKE SEE IT, BUT THERE ARE MORE DEVELOPMENTS/INFO FURTHER DOWN IN THE EMAIL 
FROM ADDITIONAL EFFORTS.




---



Dominik,



It is in a virtualized setup as I'm testing install/setup of this version of 
ovirt/node mainly as a lab/testing setup, but I was hoping to use this 
environment as a temporary data center to keep critical VMs running if 
necessary while performing the reinstall/rebuild of my physical syste

[ovirt-users] Re: Please Help, Ovirt Node Hosted Engine Deployment Problems 4.3.2

2019-05-03 Thread Simone Tiraboschi
On Fri, May 3, 2019 at 5:27 PM Todd Barton 
wrote:

> Simone/Dominik,
>
> Double reply below and more info with latest attempt.
>
> ---
>
> Simone...answers to your questions, using CAPS to make my responses easier
> to see/read.
>
> "If I correctly understood you environment you have:
> - A pfsense software firewall/router on 10.1.1.1
> - Your host on 10.1.1.61
> - You are accessing cockpit from a browser running on a machine on
> 10.1.1.101 on the same subnet"
>
> YES
>
> "And the issue is that once the engine created the management bridge, your
> client machine on 10.1.1.101 wasn't able anymore to reach your host on
> 10.1.1.61. Am I right?"
>
> YES
>
> "In this case the default gateway or other routes should't be an issue
> since your client is inside the same subnet."
>
> CORRECT, IMO
>
> "Do you think we are loosing some piece of your network configuration
> creating the management bridge such as a custom MTU or a VLAN id or
> something like that?"
>
> NO CUSTOM SETUP HERE...RUNNING A PLAIN/BASIC NETWORK
>
> "Do you think pfsense can start blocking/dropping the traffic for any
> reason?"
>
> NO, ONLY USING PFSENSE TO PROVIDE DHCP AND DNS IN TEST/LAB ENVIRONMENT.
>
> "Todd, another hint: how did you exposed cockpit over firewalld?
> Maybe something in firewalld zone configuration?"
>
> I'M USING THE OVIRT NODE MINIMAL/UTILITY INSTALL AND I DIDN'T DO ANY
> CUSTOMIZATION OUTSIDE OF THE INSTALL/SETUP.  IF ITS FIREWALLD, THEN ITS NOT
> SOMETHING THE NODE SETUP DID PROPERLY.  I CAN SEND CONFIG INFO IF YOU WOULD
> LIKE SEE IT, BUT THERE ARE MORE DEVELOPMENTS/INFO FURTHER DOWN IN THE EMAIL
> FROM ADDITIONAL EFFORTS.
>
> ---
>
> Dominik,
>
> It is in a virtualized setup as I'm testing install/setup of this version
> of ovirt/node mainly as a lab/testing setup, but I was hoping to use this
> environment as a temporary data center to keep critical VMs running if
> necessary while performing the reinstall/rebuild of my physical system.
> I'm running this in hyper-v with the virtual switch/network setup as a
> private network with no restrictions from what I can see.  The VMs are
> setup to allow mac spoofing and the mac addresses are all unique.
>
> The virtualization could the be the culprit, but I have no idea how/why it
> would be causing a problem during this specific point of the install.
>
> See new info below and I'm curious about your thoughts on this issue.
>
> ---
>
>  Additional info from "Redeploy" option 
>
> To see what would happen, I attempted the "Redeploy" option in the cockpit
> after the reboot described in previous email.  Upon Redeploy using the same
> settings, it made it through the "Prepare VM" stage.  I didn't manually
> perform a hosted-engine cleanup command, but it looked like deploy script
> cleaned up everything before attempting the redeployment.  I've repeated
> this behavior twice, so for some reason the redeployment works after the
> 1st failure.
>
> Continuing on to specify the "storage" settings to finalize deployment, it
> failed at "Obtain SSO token using username/password credentials" because it
> couldn't connect to the engine.  The HostedEngineLocal is running on the
> Host with 192.168.122.13 as the IP (the temp IP).  Trying to ping that
> address from the Host gets a "Destination Host Unreachable" from
> 192.168.122.1.  Logging into the HostedEngineLocal console from the
> cockpit, I can't ping 192.168.122.13 getting the same unreachable message
> from within the hosted-engine,
>

This looks really strange to me.
I'd suggest to focus on this.


> but I can ping the Host address 10.1.1.61 from within the hosted-engine.
>
> I'm assuming the hosted-engine should still be on this temp/private IP
> until after completion of the HE deployment.
>

Until 85% of the process.
Then it will shutdown the local bootstrap VM (running on default natted
libvirt network) and it will transfer the content of it's disk over the
disk created by the engine on the shared storage.
Latest step is simply activating ovirt-ha-agent to let it start the engine
VM from the shared storage as expected.
The final engine VM will be attached to the management bridge created by
the deployment process.


> It seems like if I could make 192.168.122.13 routable from the Host, I
> could get this to work.  Any advice on how to fix this?...I don't
> understand why I can't ping 192.168.122.13 from anywhere.
>

The default libvirt network has its specific bridge; the host has
address 192.168.122.1
there and it will be able to communicate to the VM running on
192.168.122.13.
Other machines are not required to communicate with the engine during the
deployment so we are not routing 192.168.122.1/24 neither masquerading it
for NAT traversal.


>
> I've attached logs and ip info commands from Host as well as screen shots
> from cockpit including storage/final deployment error and hosted-engine
> basic networking info.
>
> Thanks,
>
> Todd B.
>
>
>
>
>
>
>
>
>
>  On Thu, 0

[ovirt-users] Re: Please Help, Ovirt Node Hosted Engine Deployment Problems 4.3.2

2019-05-02 Thread Dominik Holler
On Thu, 2 May 2019 09:57:08 +0200
Simone Tiraboschi  wrote:

> On Thu, May 2, 2019 at 5:22 AM Todd Barton 
> wrote:
> 
> > Didi,
> >
> > I was able to carve out some time to attempt the original basic setup
> > again this evening.  The result was similar to my original post.  During HE
> > deployment, in the process of waiting for the host to come up (cockpit
> > message), the networking is disrupted while building the bridged network
> > and the host becomes unreachable.
> >
> > In this state, I can't ping the host from external machine and the
> > ping/nslookup is non-functional from within the host.  Nslookup returns
> > "connection time out; no servers could be reached".  The networking appears
> > to be completely down although various command make it appear operational.
> >
> > Upon rebooting the Host (the host locked up on reboot attempt and needed
> > to be reset), the message appears "libvirt-guests is configured not to
> > start any guests on boot".  After the reboot, the cockpit becomes
> > responsive again and loging-in displays the "This system is already
> > registered ovirt-dr-he-standalone.ipvoicedatasystems.lan!" with a
> > "Redeploy" button.  Looking at the networking setup in cockpit, it appears
> > the "ovritmgmt" network is setup, but the hosted engine did not complete
> > deployment and startup.  The /etc/host file still contains the temporary IP
> > address used in deployment and a HostedEngineLocal is listed under virtual
> > machines, but it is not running.
> >
> > Please advise with any help/input on why this is happening.  *Your help
> > is much appreciated.*
> >
> >
> > Here are the settings and diagnostic info/logs.
> >
> > This is a single-host hyper-converged setup for lab testing.
> >
> > - Host behind pfsense firewall with gateway IP address 10.1.1.1/24.  The
> > Host machine and the machine accessing the cockpit from IP address
> > 10.1.1.101 are the only devices on the subnet (other than the router).  It
> > really can't get any simpler.
> >
> > - Host setup with single nic eth0
> > - Hostname is setup as fully FQDN on Host
> > - Static IP setup on Host with gateway and DNS server set to 10.1.1.1
> > - FQDNs confirmed resolvable on subnet via dns server at 10.1.1.1 in
> > pfsense
> >   Host = ovirt-dr-standalone.ipvoicedatasystems.lan , IP = 10.1.1.61
> >   Hosted Engine VM = ovirt-dr-he-standalone.ipvoicedatasystems.lan ,
> > IP = 10.1.1.60
> >
> > - Gluster portion of cockpit setup installed as expected without problems
> >
> >
> Everything defined here looks OK to me.
> 
> 
> > - Hosted-Engine cockpit deployment executed with settings in attached
> > screen shots.
> > - Hosted engine setup and vdsm logs are attached in zip before the reboot.
> > - Other network info captured in text files included in zip.
> > - Screen shot of post reboot network setup in cockpit.
> >
> >
> According to VDSM logs
> 
> setupNetworks got executed here:
> 
> 2019-05-01 20:22:14,656-0400 INFO  (jsonrpc/0) [api.network] START
> setupNetworks(networks={u'ovirtmgmt': {u'ipv6autoconf': True, u'nic':
> u'eth0', u'ipaddr': u'10.1.1.61', u'switch': u'legacy', u'mtu': 1500,
> u'netmask': u'255.255.255.0', u'dhcpv6': False, u'STP': u'no', u'bridged':
> u'true', u'gateway': u'10.1.1.1', u'defaultRoute': True}}, bondings={},
> options={u'connectivityCheck': u'true', u'connectivityTimeout': 120,
> u'commitOnSuccess': False}) from=:::192.168.122.13,47544,
> flow_id=2e7d10f2 (api:48)
> 
> and it successfully completed at:
> 2019-05-01 20:22:22,904-0400 INFO  (jsonrpc/6) [jsonrpc.JsonRpcServer] RPC
> call Host.confirmConnectivity succeeded in 0.00 seconds (__init__:312)
> 2019-05-01 20:22:22,916-0400 INFO  (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC
> call Host.confirmConnectivity succeeded in 0.00 seconds (__init__:312)
> 2019-05-01 20:22:22,917-0400 INFO  (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC
> call Host.confirmConnectivity succeeded in 0.00 seconds (__init__:312)
> 2019-05-01 20:22:23,469-0400 INFO  (jsonrpc/2) [jsonrpc.JsonRpcServer] RPC
> call Host.confirmConnectivity succeeded in 0.00 seconds (__init__:312)
> 2019-05-01 20:22:23,583-0400 INFO  (jsonrpc/0) [api.network] FINISH
> setupNetworks return={'status': {'message': 'Done', 'code': 0}}
> from=:::192.168.122.13,47544, flow_id=2e7d10f2 (api:54)
> 2019-05-01 20:22:23,583-0400 INFO  (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC
> call Host.setupNetworks succeeded in 8.93 seconds (__init__:312)
> 2019-05-01 20:22:24,033-0400 INFO  (jsonrpc/3) [jsonrpc.JsonRpcServer] RPC
> call Host.confirmConnectivity succeeded in 0.01 seconds (__init__:312)
> 
> Host.confirmConnectivity means that, after setupNetworks, the bootstrap
> engine VM was still able to reach the engine as expected.
> 
> 
> and indeed after that we see:
> 
> 2019-05-01 20:22:24,051-0400 INFO  (jsonrpc/4) [api.host] START
> getCapabilities() from=:::192.168.122.13,47544, flow_id=2e7d10f2
> (api:48)
> ...
> 2019-05-01 20:22:25,557-0400 INFO  (jsonrpc/4) [api.host] FINISH
> getCap

[ovirt-users] Re: Please Help, Ovirt Node Hosted Engine Deployment Problems 4.3.2

2019-05-02 Thread Simone Tiraboschi
On Thu, May 2, 2019 at 9:57 AM Simone Tiraboschi 
wrote:

>
>
> On Thu, May 2, 2019 at 5:22 AM Todd Barton <
> tcbar...@ipvoicedatasystems.com> wrote:
>
>> Didi,
>>
>> I was able to carve out some time to attempt the original basic setup
>> again this evening.  The result was similar to my original post.  During HE
>> deployment, in the process of waiting for the host to come up (cockpit
>> message), the networking is disrupted while building the bridged network
>> and the host becomes unreachable.
>>
>> In this state, I can't ping the host from external machine and the
>> ping/nslookup is non-functional from within the host.  Nslookup returns
>> "connection time out; no servers could be reached".  The networking appears
>> to be completely down although various command make it appear operational.
>>
>> Upon rebooting the Host (the host locked up on reboot attempt and needed
>> to be reset), the message appears "libvirt-guests is configured not to
>> start any guests on boot".  After the reboot, the cockpit becomes
>> responsive again and loging-in displays the "This system is already
>> registered ovirt-dr-he-standalone.ipvoicedatasystems.lan!" with a
>> "Redeploy" button.  Looking at the networking setup in cockpit, it appears
>> the "ovritmgmt" network is setup, but the hosted engine did not complete
>> deployment and startup.  The /etc/host file still contains the temporary IP
>> address used in deployment and a HostedEngineLocal is listed under virtual
>> machines, but it is not running.
>>
>> Please advise with any help/input on why this is happening.  *Your help
>> is much appreciated.*
>>
>>
>> Here are the settings and diagnostic info/logs.
>>
>> This is a single-host hyper-converged setup for lab testing.
>>
>> - Host behind pfsense firewall with gateway IP address 10.1.1.1/24.  The
>> Host machine and the machine accessing the cockpit from IP address
>> 10.1.1.101 are the only devices on the subnet (other than the router).  It
>> really can't get any simpler.
>>
>> - Host setup with single nic eth0
>> - Hostname is setup as fully FQDN on Host
>> - Static IP setup on Host with gateway and DNS server set to 10.1.1.1
>> - FQDNs confirmed resolvable on subnet via dns server at 10.1.1.1 in
>> pfsense
>>   Host = ovirt-dr-standalone.ipvoicedatasystems.lan , IP = 10.1.1.61
>>   Hosted Engine VM = ovirt-dr-he-standalone.ipvoicedatasystems.lan ,
>> IP = 10.1.1.60
>>
>> - Gluster portion of cockpit setup installed as expected without problems
>>
>>
> Everything defined here looks OK to me.
>
>
>> - Hosted-Engine cockpit deployment executed with settings in attached
>> screen shots.
>> - Hosted engine setup and vdsm logs are attached in zip before the reboot.
>> - Other network info captured in text files included in zip.
>> - Screen shot of post reboot network setup in cockpit.
>>
>>
> According to VDSM logs
>
> setupNetworks got executed here:
>
> 2019-05-01 20:22:14,656-0400 INFO  (jsonrpc/0) [api.network] START
> setupNetworks(networks={u'ovirtmgmt': {u'ipv6autoconf': True, u'nic':
> u'eth0', u'ipaddr': u'10.1.1.61', u'switch': u'legacy', u'mtu': 1500,
> u'netmask': u'255.255.255.0', u'dhcpv6': False, u'STP': u'no', u'bridged':
> u'true', u'gateway': u'10.1.1.1', u'defaultRoute': True}}, bondings={},
> options={u'connectivityCheck': u'true', u'connectivityTimeout': 120,
> u'commitOnSuccess': False}) from=:::192.168.122.13,47544,
> flow_id=2e7d10f2 (api:48)
>
> and it successfully completed at:
> 2019-05-01 20:22:22,904-0400 INFO  (jsonrpc/6) [jsonrpc.JsonRpcServer] RPC
> call Host.confirmConnectivity succeeded in 0.00 seconds (__init__:312)
> 2019-05-01 20:22:22,916-0400 INFO  (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC
> call Host.confirmConnectivity succeeded in 0.00 seconds (__init__:312)
> 2019-05-01 20:22:22,917-0400 INFO  (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC
> call Host.confirmConnectivity succeeded in 0.00 seconds (__init__:312)
> 2019-05-01 20:22:23,469-0400 INFO  (jsonrpc/2) [jsonrpc.JsonRpcServer] RPC
> call Host.confirmConnectivity succeeded in 0.00 seconds (__init__:312)
> 2019-05-01 20:22:23,583-0400 INFO  (jsonrpc/0) [api.network] FINISH
> setupNetworks return={'status': {'message': 'Done', 'code': 0}}
> from=:::192.168.122.13,47544, flow_id=2e7d10f2 (api:54)
> 2019-05-01 20:22:23,583-0400 INFO  (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC
> call Host.setupNetworks succeeded in 8.93 seconds (__init__:312)
> 2019-05-01 20:22:24,033-0400 INFO  (jsonrpc/3) [jsonrpc.JsonRpcServer] RPC
> call Host.confirmConnectivity succeeded in 0.01 seconds (__init__:312)
>
> Host.confirmConnectivity means that, after setupNetworks, the bootstrap
> engine VM was still able to reach the engine as expected.
>
>
> and indeed after that we see:
>
> 2019-05-01 20:22:24,051-0400 INFO  (jsonrpc/4) [api.host] START
> getCapabilities() from=:::192.168.122.13,47544, flow_id=2e7d10f2
> (api:48)
> ...
> 2019-05-01 20:22:25,557-0400 INFO  (jsonrpc/4) [api.host] FINISH
> getCapabilities return={'status': {'m

[ovirt-users] Re: Please Help, Ovirt Node Hosted Engine Deployment Problems 4.3.2

2019-05-02 Thread Simone Tiraboschi
On Thu, May 2, 2019 at 5:22 AM Todd Barton 
wrote:

> Didi,
>
> I was able to carve out some time to attempt the original basic setup
> again this evening.  The result was similar to my original post.  During HE
> deployment, in the process of waiting for the host to come up (cockpit
> message), the networking is disrupted while building the bridged network
> and the host becomes unreachable.
>
> In this state, I can't ping the host from external machine and the
> ping/nslookup is non-functional from within the host.  Nslookup returns
> "connection time out; no servers could be reached".  The networking appears
> to be completely down although various command make it appear operational.
>
> Upon rebooting the Host (the host locked up on reboot attempt and needed
> to be reset), the message appears "libvirt-guests is configured not to
> start any guests on boot".  After the reboot, the cockpit becomes
> responsive again and loging-in displays the "This system is already
> registered ovirt-dr-he-standalone.ipvoicedatasystems.lan!" with a
> "Redeploy" button.  Looking at the networking setup in cockpit, it appears
> the "ovritmgmt" network is setup, but the hosted engine did not complete
> deployment and startup.  The /etc/host file still contains the temporary IP
> address used in deployment and a HostedEngineLocal is listed under virtual
> machines, but it is not running.
>
> Please advise with any help/input on why this is happening.  *Your help
> is much appreciated.*
>
>
> Here are the settings and diagnostic info/logs.
>
> This is a single-host hyper-converged setup for lab testing.
>
> - Host behind pfsense firewall with gateway IP address 10.1.1.1/24.  The
> Host machine and the machine accessing the cockpit from IP address
> 10.1.1.101 are the only devices on the subnet (other than the router).  It
> really can't get any simpler.
>
> - Host setup with single nic eth0
> - Hostname is setup as fully FQDN on Host
> - Static IP setup on Host with gateway and DNS server set to 10.1.1.1
> - FQDNs confirmed resolvable on subnet via dns server at 10.1.1.1 in
> pfsense
>   Host = ovirt-dr-standalone.ipvoicedatasystems.lan , IP = 10.1.1.61
>   Hosted Engine VM = ovirt-dr-he-standalone.ipvoicedatasystems.lan ,
> IP = 10.1.1.60
>
> - Gluster portion of cockpit setup installed as expected without problems
>
>
Everything defined here looks OK to me.


> - Hosted-Engine cockpit deployment executed with settings in attached
> screen shots.
> - Hosted engine setup and vdsm logs are attached in zip before the reboot.
> - Other network info captured in text files included in zip.
> - Screen shot of post reboot network setup in cockpit.
>
>
According to VDSM logs

setupNetworks got executed here:

2019-05-01 20:22:14,656-0400 INFO  (jsonrpc/0) [api.network] START
setupNetworks(networks={u'ovirtmgmt': {u'ipv6autoconf': True, u'nic':
u'eth0', u'ipaddr': u'10.1.1.61', u'switch': u'legacy', u'mtu': 1500,
u'netmask': u'255.255.255.0', u'dhcpv6': False, u'STP': u'no', u'bridged':
u'true', u'gateway': u'10.1.1.1', u'defaultRoute': True}}, bondings={},
options={u'connectivityCheck': u'true', u'connectivityTimeout': 120,
u'commitOnSuccess': False}) from=:::192.168.122.13,47544,
flow_id=2e7d10f2 (api:48)

and it successfully completed at:
2019-05-01 20:22:22,904-0400 INFO  (jsonrpc/6) [jsonrpc.JsonRpcServer] RPC
call Host.confirmConnectivity succeeded in 0.00 seconds (__init__:312)
2019-05-01 20:22:22,916-0400 INFO  (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC
call Host.confirmConnectivity succeeded in 0.00 seconds (__init__:312)
2019-05-01 20:22:22,917-0400 INFO  (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC
call Host.confirmConnectivity succeeded in 0.00 seconds (__init__:312)
2019-05-01 20:22:23,469-0400 INFO  (jsonrpc/2) [jsonrpc.JsonRpcServer] RPC
call Host.confirmConnectivity succeeded in 0.00 seconds (__init__:312)
2019-05-01 20:22:23,583-0400 INFO  (jsonrpc/0) [api.network] FINISH
setupNetworks return={'status': {'message': 'Done', 'code': 0}}
from=:::192.168.122.13,47544, flow_id=2e7d10f2 (api:54)
2019-05-01 20:22:23,583-0400 INFO  (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC
call Host.setupNetworks succeeded in 8.93 seconds (__init__:312)
2019-05-01 20:22:24,033-0400 INFO  (jsonrpc/3) [jsonrpc.JsonRpcServer] RPC
call Host.confirmConnectivity succeeded in 0.01 seconds (__init__:312)

Host.confirmConnectivity means that, after setupNetworks, the bootstrap
engine VM was still able to reach the engine as expected.


and indeed after that we see:

2019-05-01 20:22:24,051-0400 INFO  (jsonrpc/4) [api.host] START
getCapabilities() from=:::192.168.122.13,47544, flow_id=2e7d10f2
(api:48)
...
2019-05-01 20:22:25,557-0400 INFO  (jsonrpc/4) [api.host] FINISH
getCapabilities return={'status': {'message': 'Done', 'code': 0}, 'info':
{u'HBAInventory': {u'iSCSI': [{u'InitiatorName':
u'iqn.1994-05.com.redhat:79982989d81e'}], u'FC': []}, u'packages2':
{u'kernel': {u'release': u'957.10.1.el7.x86_64', u'version': u'3.10.0'},
u'glusterfs

[ovirt-users] Re: Please Help, Ovirt Node Hosted Engine Deployment Problems 4.3.2

2019-05-01 Thread Todd Barton
Thanks again...I've done all the detail work back in the 3.x days and I thought 
(and was hoping) the node/cockpit setup would make this easier to get 
everything lined up for the HE deploy, but it is not working as expected.  I've 
followed best practices/recommendations, but realize there are not absolute 
specifics in these recommendations...there are a lot of either/or 
statements...which is why I was asking for recommendations.  I've reviewed many 
articles including the "up and running" one like 
https://ovirt.org/blog/2018/02/up-and-running-with-ovirt-4-2-and-gluster-storage.html,
  In everything I've looked at there isn't anything new or different vs what 
I've already done or attempted.



I was very methodical in my initial attempts as I did in my initial install of 
v3.3 years ago which took many attempts and methodical configuration to get it 
up and setup the way I wanted.  What I'm trying to understand is why its not 
coming up in a lab setting with what I would consider to be a pretty remedial 
setup.



I'll get back to a basic setup and run through the process again today or 
tomorrow and post logs of the failure.



Regards,



Todd Barton






 On Wed, 01 May 2019 01:50:49 -0400 Yedidyah Bar David  
wrote 



On Tue, Apr 30, 2019 at 4:09 PM Todd Barton 

 wrote: 

> 

> Thanks a bunch for the reply Didi and Simone.  I will admit this last setup 
> was a bit of wild attempt to see if i could get it working somehow so maybe 
> it wasn't the best example to submit...and yeah, should have been /24 
> subnets.  Initially I tried the single nic setup, but the outcome seemed to 
> be the same scenario. 

> 

> Honestly I've run through this setup so many times in the last week its all a 
> blur.  I started messing multiple nics in latest attempts to see if this was 
> something specific I should do in a cockpit setup as one of the articles I 
> read suggested multiple interfaces to separate traffic. 

> 

> My "production" 4.0 environment (currently a failed upgrade with a down host 
> that I can't seem to get back online) is 3 host gluster on 4 bonded 1Gbps 
> links.  With the exception of the upgrade issue/failure, it has been 
> rock-solid with good performance and I've only restarted hosts on upgrades in 
> 4+ years.  There are a few networking changes i would like to make in a 
> rebuild, but I wanted to test various options before implementing.  Getting a 
> single nic environment was the initial goal to get started. 

> 

> I'm doing this testing in a virtualized setup with pfsense as the 
> firewall/router and I can setup hosts/nics however I want.  I will start over 
> again with more straightforward setup and get more data on failure.  
> Considering I can setup the environment how i want, what would be your 
> recommended config for a single nic(or single bond) setup using cockpit?  
> Static IPs with host file resolution, DHCP with mac specific IPs, etc. 



Much of such decisions is a matter of personal preferences, 

acquaintance with the relevant technologies and tooling you have 

around them, local needs/policies/mandates, existing infrastructure, 

etc. 



If you search the net, e.g. for "ovirt best practices" or "RHV best 

practices", you can find various articles etc. that can provide some 

good guidelines/ideas. 



I suggest to read around a bit, then spend some good time on planning, 

then carefully and systematically implement your design, verifying 

each step right after doing it. When you run into problems, tell us 

:-). Ideally, IMO, you should not give up on your design due to such 

problems and try workarounds, inferior (in your eyes) solutions, etc., 

unless you manage to find existing open bugs that describe your 

problem and you decide you can't want until they are solved. Instead, 

try to fix problems, perhaps with the list members' help. 



I realize spending a week on what is in your perception a simple, 

straightforward task, does not leave you in the best mood for such a 

methodical next attempt. Perhaps first take a break and do something 

else :-), then start from a clean and fresh hardware/software 

environment and mind. 



Good luck and best regards, 



> 

> Thank you, 

> 

> Todd Barton 

> 

> 

> 

> 

>  On Tue, 30 Apr 2019 05:20:04 -0400 Simone Tiraboschi 
>  wrote  

> 

> 

> 

> On Tue, Apr 30, 2019 at 9:50 AM Yedidyah Bar David  
> wrote: 

> 

> On Tue, Apr 30, 2019 at 5:09 AM Todd Barton 

>  wrote: 

> > 

> > I've having to rebuild an environment that started back in the early 3.x 
> > days.  A lot has changed and I'm attempting to use the Ovirt Node based 
> > setup to build a new environment, but I can't get through the hosted engine 
> > deployment process via the cockpit (I've done command line as well).  I've 
> > tried static DHCP address and static IPs as well as confir

[ovirt-users] Re: Please Help, Ovirt Node Hosted Engine Deployment Problems 4.3.2

2019-04-30 Thread Yedidyah Bar David
On Tue, Apr 30, 2019 at 4:09 PM Todd Barton
 wrote:
>
> Thanks a bunch for the reply Didi and Simone.  I will admit this last setup 
> was a bit of wild attempt to see if i could get it working somehow so maybe 
> it wasn't the best example to submit...and yeah, should have been /24 
> subnets.  Initially I tried the single nic setup, but the outcome seemed to 
> be the same scenario.
>
> Honestly I've run through this setup so many times in the last week its all a 
> blur.  I started messing multiple nics in latest attempts to see if this was 
> something specific I should do in a cockpit setup as one of the articles I 
> read suggested multiple interfaces to separate traffic.
>
> My "production" 4.0 environment (currently a failed upgrade with a down host 
> that I can't seem to get back online) is 3 host gluster on 4 bonded 1Gbps 
> links.  With the exception of the upgrade issue/failure, it has been 
> rock-solid with good performance and I've only restarted hosts on upgrades in 
> 4+ years.  There are a few networking changes i would like to make in a 
> rebuild, but I wanted to test various options before implementing.  Getting a 
> single nic environment was the initial goal to get started.
>
> I'm doing this testing in a virtualized setup with pfsense as the 
> firewall/router and I can setup hosts/nics however I want.  I will start over 
> again with more straightforward setup and get more data on failure.  
> Considering I can setup the environment how i want, what would be your 
> recommended config for a single nic(or single bond) setup using cockpit?  
> Static IPs with host file resolution, DHCP with mac specific IPs, etc.

Much of such decisions is a matter of personal preferences,
acquaintance with the relevant technologies and tooling you have
around them, local needs/policies/mandates, existing infrastructure,
etc.

If you search the net, e.g. for "ovirt best practices" or "RHV best
practices", you can find various articles etc. that can provide some
good guidelines/ideas.

I suggest to read around a bit, then spend some good time on planning,
then carefully and systematically implement your design, verifying
each step right after doing it. When you run into problems, tell us
:-). Ideally, IMO, you should not give up on your design due to such
problems and try workarounds, inferior (in your eyes) solutions, etc.,
unless you manage to find existing open bugs that describe your
problem and you decide you can't want until they are solved. Instead,
try to fix problems, perhaps with the list members' help.

I realize spending a week on what is in your perception a simple,
straightforward task, does not leave you in the best mood for such a
methodical next attempt. Perhaps first take a break and do something
else :-), then start from a clean and fresh hardware/software
environment and mind.

Good luck and best regards,

>
> Thank you,
>
> Todd Barton
>
>
>
>
>  On Tue, 30 Apr 2019 05:20:04 -0400 Simone Tiraboschi 
>  wrote 
>
>
>
> On Tue, Apr 30, 2019 at 9:50 AM Yedidyah Bar David  wrote:
>
> On Tue, Apr 30, 2019 at 5:09 AM Todd Barton
>  wrote:
> >
> > I've having to rebuild an environment that started back in the early 3.x 
> > days.  A lot has changed and I'm attempting to use the Ovirt Node based 
> > setup to build a new environment, but I can't get through the hosted engine 
> > deployment process via the cockpit (I've done command line as well).  I've 
> > tried static DHCP address and static IPs as well as confirmed I have 
> > resolvable host-names.  This is a test environment so I can work through 
> > any issues in deployment.
> >
> > When the cockpit is displaying the waiting for host to come up task, the 
> > cockpit gets disconnected.  It appears to a happen when the bridge network 
> > is setup.  At that point, the deployment is messed up and I can't return to 
> > the cockpit.  I've tried this with one or two nic/interfaces and tried 
> > every permutation of static and dynamic ip addresses.  I've spent a week 
> > trying different setups and I've got to be doing something stupid.
> >
> > Attached is a screen capture of the resulting IP info after my latest try 
> > failing.  I used two nics, one for the gluster and bridge network and the 
> > other for the ovirt cockpit access.  I can't access cockpit on either ip 
> > address after the failure.
> >
> > I've attempted this setup as both a single host hyper-converged setup and a 
> > three host hyper-converged environment...same issue in both.
> >
> > Can someone please help me or give me some thoughts on what is wrong?
>
> There are two parts here: 1. Fix it so that you can continue (and so
> that if it happens to you on production, you know what to do) 2. Fix
> the code so that it does not happen again. They are not necessarily
> identical (or even very similar).
>
> At the point in time of taking the screen capture:
>
> 1. Did the ovirtmgmt bridge get the IP address of the intended nic? Which one?
>
> 2. Did you chec

[ovirt-users] Re: Please Help, Ovirt Node Hosted Engine Deployment Problems 4.3.2

2019-04-30 Thread Todd Barton
Thanks a bunch for the reply Didi and Simone.  I will admit this last setup was 
a bit of wild attempt to see if i could get it working somehow so maybe it 
wasn't the best example to submit...and yeah, should have been /24 subnets.  
Initially I tried the single nic setup, but the outcome seemed to be the same 
scenario.



Honestly I've run through this setup so many times in the last week its all a 
blur.  I started messing multiple nics in latest attempts to see if this was 
something specific I should do in a cockpit setup as one of the articles I read 
suggested multiple interfaces to separate traffic.



My "production" 4.0 environment  (currently a failed upgrade with a down host 
that I can't seem to get back online) is 3 host gluster on 4 bonded 1Gbps 
links.  With the exception of the upgrade issue/failure, it has been rock-solid 
with good performance and I've only restarted hosts on upgrades in 4+ years.  
There are a few networking changes i would like to make in a rebuild, but I 
wanted to test various options before implementing.  Getting a single nic 
environment was the initial goal to get started.



I'm doing this testing in a virtualized setup with pfsense as the 
firewall/router and I can setup hosts/nics however I want.  I will start over 
again with more straightforward setup and get more data on failure.  
Considering I can setup the environment how i want, what would be your 
recommended config for a single nic(or single bond) setup using cockpit?  
Static IPs with host file resolution, DHCP with mac specific IPs, etc.



Thank you,


Todd Barton










 On Tue, 30 Apr 2019 05:20:04 -0400 Simone Tiraboschi  
wrote 







On Tue, Apr 30, 2019 at 9:50 AM Yedidyah Bar David  
wrote:

On Tue, Apr 30, 2019 at 5:09 AM Todd Barton

  wrote:

 >

 > I've having to rebuild an environment that started back in the early 3.x 
 > days.  A lot has changed and I'm attempting to use the Ovirt Node based 
 > setup to build a new environment, but I can't get through the hosted engine 
 > deployment process via the cockpit (I've done command line as well).  I've 
 > tried static DHCP address and static IPs as well as confirmed I have 
 > resolvable host-names.  This is a test environment so I can work through any 
 > issues in deployment.

 >

 > When the cockpit is displaying the waiting for host to come up task, the 
 > cockpit gets disconnected.  It appears to a happen when the bridge network 
 > is setup.  At that point, the deployment is messed up and I can't return to 
 > the cockpit.  I've tried this with one or two nic/interfaces and tried every 
 > permutation of static and dynamic ip addresses.  I've spent a week trying 
 > different setups and I've got to be doing something stupid.

 >

 > Attached is a screen capture of the resulting IP info after my latest try 
 > failing.  I used two nics, one for the gluster and bridge network and the 
 > other for the ovirt cockpit access.  I can't access cockpit on either ip 
 > address after the failure.

 >

 > I've attempted this setup as both a single host hyper-converged setup and a 
 > three host hyper-converged environment...same issue in both.

 >

 > Can someone please help me or give me some thoughts on what is wrong?

 

 There are two parts here: 1. Fix it so that you can continue (and so

 that if it happens to you on production, you know what to do) 2. Fix

 the code so that it does not happen again. They are not necessarily

 identical (or even very similar).

 

 At the point in time of taking the screen capture:

 

 1. Did the ovirtmgmt bridge get the IP address of the intended nic? Which one?

 

 2. Did you check routing? Default gateway, or perhaps you had/have

 specific other routes?

 

 3. What nics are in the bridge? Can you check/share output of 'brctl show'?

 

 4. Probably not related, just noting: You have there (currently on

 eth0 and on ovirtmgmt, perhaps you tried other combinations):

 http://10.1.2.61/16 and http://10.1.1.61/16 . It seems like you wanted two 
different

 subnets, but are actually using a single one. Perhaps you intended to

 use http://10.1.2.61/24 and http://10.1.1.61/24.


 

Good catch: the issue comes exactly form here!

Please see:

https://bugzilla.redhat.com/1694626



The issue happens when the user has two interfaces configured on the same IP 
subnet, the default gateway is configured to be reached from one of the two 
interfaces and the user chooses to create the management bridge on the other 
one.

When the engine, adding the host, creates the management bridge it also tries 
to configure the default gateway on the bridge and for some reason this disrupt 
the external connectivity on the host and the the user is going to loose it.



If you intend to use one interface for gluster and the other for the management 
network I'd strongly suggest to use two distinct subnets having the default 
gateway on the subnet

[ovirt-users] Re: Please Help, Ovirt Node Hosted Engine Deployment Problems 4.3.2

2019-04-30 Thread Strahil

> Each data has to be written on the host itself and on the two remote ones so 
> you are going to have 1000 mbps / 2 (external replicas ) / 8 (bit/bytes) = a 
> max of 62.5 MB/s sustained throughput shared between all the VMs and this 
> ignoring all the overheads.
> In practice it will be much less ending in a barely usable environment.


I am currently in this type of setup (replica 2 arbiter 1 on 1 gbit/s  both 
storage and oVirt connectivity)  and the maximum write (sequential) speed I can 
reach is aprox 89 MB/s.
I guess with replica 3 the maximum will be as Simone stated.
My maximum read speed is aprox 500 MB/s  despite that local NVMe is capable of 
1.3 GB/s and I guess  FUSE  is not using  the  local brick no matter  I have 
set it up with cluster.use-local  (or whatever  it was called).

If possible , get a 10 gbit/s connectivity or multiple  gbit interfaces  ( as  
I'm planing to do in the nearest future).

Best Regards,
Strahil Nikolov___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/GGV5VUKY372HU5O4HCC6CIDQSTDIEXZP/


[ovirt-users] Re: Please Help, Ovirt Node Hosted Engine Deployment Problems 4.3.2

2019-04-30 Thread Simone Tiraboschi
On Tue, Apr 30, 2019 at 9:50 AM Yedidyah Bar David  wrote:

> On Tue, Apr 30, 2019 at 5:09 AM Todd Barton
>  wrote:
> >
> > I've having to rebuild an environment that started back in the early 3.x
> days.  A lot has changed and I'm attempting to use the Ovirt Node based
> setup to build a new environment, but I can't get through the hosted engine
> deployment process via the cockpit (I've done command line as well).  I've
> tried static DHCP address and static IPs as well as confirmed I have
> resolvable host-names.  This is a test environment so I can work through
> any issues in deployment.
> >
> > When the cockpit is displaying the waiting for host to come up task, the
> cockpit gets disconnected.  It appears to a happen when the bridge network
> is setup.  At that point, the deployment is messed up and I can't return to
> the cockpit.  I've tried this with one or two nic/interfaces and tried
> every permutation of static and dynamic ip addresses.  I've spent a week
> trying different setups and I've got to be doing something stupid.
> >
> > Attached is a screen capture of the resulting IP info after my latest
> try failing.  I used two nics, one for the gluster and bridge network and
> the other for the ovirt cockpit access.  I can't access cockpit on either
> ip address after the failure.
> >
> > I've attempted this setup as both a single host hyper-converged setup
> and a three host hyper-converged environment...same issue in both.
> >
> > Can someone please help me or give me some thoughts on what is wrong?
>
> There are two parts here: 1. Fix it so that you can continue (and so
> that if it happens to you on production, you know what to do) 2. Fix
> the code so that it does not happen again. They are not necessarily
> identical (or even very similar).
>
> At the point in time of taking the screen capture:
>
> 1. Did the ovirtmgmt bridge get the IP address of the intended nic? Which
> one?
>
> 2. Did you check routing? Default gateway, or perhaps you had/have
> specific other routes?
>
> 3. What nics are in the bridge? Can you check/share output of 'brctl show'?
>
> 4. Probably not related, just noting: You have there (currently on
> eth0 and on ovirtmgmt, perhaps you tried other combinations):
> 10.1.2.61/16 and 10.1.1.61/16 . It seems like you wanted two different
> subnets, but are actually using a single one. Perhaps you intended to
> use 10.1.2.61/24 and 10.1.1.61/24.
>

Good catch: the issue comes exactly form here!
Please see:
https://bugzilla.redhat.com/1694626

The issue happens when the user has two interfaces configured on the same
IP subnet, the default gateway is configured to be reached from one of the
two interfaces and the user chooses to create the management bridge on the
other one.
When the engine, adding the host, creates the management bridge it also
tries to configure the default gateway on the bridge and for some reason
this disrupt the external connectivity on the host and the the user is
going to loose it.

If you intend to use one interface for gluster and the other for the
management network I'd strongly suggest to use two distinct subnets having
the default gateway on the subnet you are going to use for the management
network.

If you want to use two interfaces for reliability reasons I'd strongly
suggest to create a bond of the two instead.

Please also notice that deploying a three host hyper-converged environment
over a single 1 gbps interface will be really penalizing in terms of
storage performances.
Each data has to be written on the host itself and on the two remote ones
so you are going to have 1000 mbps / 2 (external replicas ) / 8 (bit/bytes)
= a max of 62.5 MB/s sustained throughput shared between all the VMs and
this ignoring all the overheads.
In practice it will be much less ending in a barely usable environment.

I'd strongly suggest to move to a 10 gbps environment if possible, or to
bond a few 1 gbps nics for gluster.


5. Can you ping from/to these two addresses from/to some other machine
> on the network? Your laptop? The storage?
>
> 6. If possible, please check/share relevant logs, including (from the
> host) /var/log/vdsm/* and /var/log/ovirt-hosted-engine-setup/*.
>
> Thanks and best regards,
> --
> Didi
> ___
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/JIYWEUXPA25BK3K23MPBISRGZN76AWV3/
>
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users

[ovirt-users] Re: Please Help, Ovirt Node Hosted Engine Deployment Problems 4.3.2

2019-04-30 Thread Yedidyah Bar David
On Tue, Apr 30, 2019 at 5:09 AM Todd Barton
 wrote:
>
> I've having to rebuild an environment that started back in the early 3.x 
> days.  A lot has changed and I'm attempting to use the Ovirt Node based setup 
> to build a new environment, but I can't get through the hosted engine 
> deployment process via the cockpit (I've done command line as well).  I've 
> tried static DHCP address and static IPs as well as confirmed I have 
> resolvable host-names.  This is a test environment so I can work through any 
> issues in deployment.
>
> When the cockpit is displaying the waiting for host to come up task, the 
> cockpit gets disconnected.  It appears to a happen when the bridge network is 
> setup.  At that point, the deployment is messed up and I can't return to the 
> cockpit.  I've tried this with one or two nic/interfaces and tried every 
> permutation of static and dynamic ip addresses.  I've spent a week trying 
> different setups and I've got to be doing something stupid.
>
> Attached is a screen capture of the resulting IP info after my latest try 
> failing.  I used two nics, one for the gluster and bridge network and the 
> other for the ovirt cockpit access.  I can't access cockpit on either ip 
> address after the failure.
>
> I've attempted this setup as both a single host hyper-converged setup and a 
> three host hyper-converged environment...same issue in both.
>
> Can someone please help me or give me some thoughts on what is wrong?

There are two parts here: 1. Fix it so that you can continue (and so
that if it happens to you on production, you know what to do) 2. Fix
the code so that it does not happen again. They are not necessarily
identical (or even very similar).

At the point in time of taking the screen capture:

1. Did the ovirtmgmt bridge get the IP address of the intended nic? Which one?

2. Did you check routing? Default gateway, or perhaps you had/have
specific other routes?

3. What nics are in the bridge? Can you check/share output of 'brctl show'?

4. Probably not related, just noting: You have there (currently on
eth0 and on ovirtmgmt, perhaps you tried other combinations):
10.1.2.61/16 and 10.1.1.61/16 . It seems like you wanted two different
subnets, but are actually using a single one. Perhaps you intended to
use 10.1.2.61/24 and 10.1.1.61/24.

5. Can you ping from/to these two addresses from/to some other machine
on the network? Your laptop? The storage?

6. If possible, please check/share relevant logs, including (from the
host) /var/log/vdsm/* and /var/log/ovirt-hosted-engine-setup/*.

Thanks and best regards,
-- 
Didi
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/JIYWEUXPA25BK3K23MPBISRGZN76AWV3/