[ovirt-users] Re: hosted-engine --vm-status returns "vm not running on this host" on all hosts

2021-02-13 Thread Yedidyah Bar David
On Fri, Feb 12, 2021 at 12:03 AM  wrote:
>
> I have tried to upgrade from 4.3 to 4.4 in  a self hosted environment and I 
> now have multiple problems. One of which is that neither of my two hosts 
> think they are running the hosted engine. What could cause this?

Please provide more details. What did you do, what happened. Relevant logs. etc.

Best regards,
-- 
Didi
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/GKFZFAKP3QCIHMHLHXCBQBUZMWCVITG6/


[ovirt-users] Re: Problems after upgrade from 4.4.3 to 4.4.4

2021-02-13 Thread tfe...@swissonline.ch
Hi

On host1, all services are running, including “sanlock”, “supervdsmd” and 
“vdsmd”.
On host2, the services "ovsdb-server” and "ovirt-imageio” are not running, and 
as a result, “vdsmd” is also not running.
So the cluster is down.

In the log of host2, I can find "ovsdb-server.service: Failed at step EXEC 
spawning /usr/share/openvswitch/scripts/ovs-ctl: Exec format error”.

host1 has status “Unassigned” and host2 has status “NonResponsive”.
So the SPM role cannot be assigned to another host.

Any ideas how to fix this?

Kind regards
Toni 


> On 13 Feb 2021, at 18:31, Strahil Nikolov  wrote:
> 
> Have you tried to set another host as SPM ?
> Also, you can mark the host was rebooted ( I assume you got no power 
> management configured) from Hosts -> 3 dots in upper right in UI.
> 
> Also check if sanlock, supervdsmd and vdsmd services are healthyand running.
> 
> Best Regards,
> Strahil Nikolov
> 
> On Fri, Feb 12, 2021 at 1:46, tfe...@swissonline.ch
>  wrote:
> Hi
> 
> I have problems after upgrading my 2-node cluster from 4.4.3 to 4.4.4.
> 
> Initially, I performed the upgrade of the oVirt hosts using the oVirt GUI (I 
> wasn't planning any changes).
> 
> It appears that the upgrade broke the system.
> On host1, the ovirt-engine was configured to run on the oVirt host itself 
> (not self-hosted engine).
> 
> After the upgrade, the oVirt GUI didn't load in the Browser anymore.
> I tried to fix the issue by migrating to self-hosted engine, which did not 
> work, so I ran engine restore and engine-setup in order to get back to the 
> initial state.
> I am now able to login to the oVirt GUI again, but I am having the following 
> problems:
> host1 is in status "Unassigned", and it has the SPM role. It cannot be set to 
> maintenance mode, nor re-installed from GUI, but I am able to reboot the host 
> from oVirt.
> All Storage Domains are inactive. (all NFS)
> In the /var/log/messages log, I can see the following message appearing 
> frequently: "vdsm[5935]: ERROR ssl handshake: socket error, address: 
> :::192.168.100.61"
> The cluster is down and no VM's can be run. I don't know how to fix either of 
> the issues.
> 
> Does anyone have an idea?
> I am appending a tar file containing log files to this email.
> 
> http://gofile.me/5fp92/d7iGEqh3H 
> Many thanks
> 
> Toni
> 
> 
> 
> ___
> Users mailing list -- users@ovirt.org 
> To unsubscribe send an email to users-le...@ovirt.org 
> 
> Privacy Statement: https://www.ovirt.org/privacy-policy.html 
> 
> oVirt Code of Conduct: 
> https://www.ovirt.org/community/about/community-guidelines/ 
> 
> List Archives: 
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/3QM4ATVPNXWJVNT2BCY7IFX63JYROZSD/
>  
> 

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/5KH6ZAVU6XAI5H6P5BOFUNGYF34FJHVJ/


[ovirt-users] Cockpit and Gluster

2021-02-13 Thread C Williams
Hello,

I have a hardware RAID 10 array and want to use Cockpit to deploy a hosted
engine with replica 3 Gluster.

Should I pick the JBOD option for this ?

Will my disk alignments be correct for my PVs, VGs, LVs and xfs formatting?

I don't see RAID 10. I see where it was removed in 4.2.8 -- I think.

Thank You All For Your Help !
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/LA66XNZ776CFAU6XTOUILQ2OOSCRR3PP/


[ovirt-users] Re: Problems after upgrade from 4.4.3 to 4.4.4

2021-02-13 Thread Strahil Nikolov via Users
Have you tried to set another host as SPM ?Also, you can mark the host was 
rebooted ( I assume you got no power management configured) from Hosts -> 3 
dots in upper right in UI.
Also check if sanlock, supervdsmd and vdsmd services are healthyand running.
Best Regards,Strahil Nikolov
 
 
  On Fri, Feb 12, 2021 at 1:46, tfe...@swissonline.ch 
wrote:   
Hi

I have problems after upgrading my 2-node cluster from 4.4.3 to 4.4.4.

Initially, I performed the upgrade of the oVirt hosts using the oVirt GUI (I 
wasn't planning any changes).

It appears that the upgrade broke the system.
 

On host1, the ovirt-engine was configured to run on the oVirt host itself (not 
self-hosted engine).

After the upgrade, the oVirt GUI didn't load in the Browser anymore.
 

I tried to fix the issue by migrating to self-hosted engine, which did not 
work, so I ran engine restore and engine-setup in order to get back to the 
initial state.
 

I am now able to login to the oVirt GUI again, but I am having the following 
problems:
 

   - host1 is in status "Unassigned", and it has the SPM role. It cannot be set 
to maintenance mode, nor re-installed from GUI, but I am able to reboot the 
host from oVirt.
   - All Storage Domains are inactive. (all NFS)
   - In the /var/log/messages log, I can see the following message appearing 
frequently: "vdsm[5935]: ERROR ssl handshake: socket error, address: 
:::192.168.100.61"   
 

The cluster is down and no VM's can be run. I don't know how to fix either of 
the issues.

Does anyone have an idea?
 

I am appending a tar file containing log files to this email.
http://gofile.me/5fp92/d7iGEqh3H

Many thanks

Toni
 




___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3QM4ATVPNXWJVNT2BCY7IFX63JYROZSD/
  
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/AFL2JH5EPMYCXXZ6WHZRDGQUNOJFSNM2/


[ovirt-users] Re: Single Node Hyperconverged - Failing Engine Deployment - Network setup?

2021-02-13 Thread jhamiltonactually
Hi
Sorry about the delay! Here is what I did - which for some reason got further 
than before, but still failed at the last hurdle! Maybe just doing this 
exercise of commenting my workflow helped me to fix errors I'd not seen before 
- I don't know! Have any packages been updated I don't know about?

https://docs.google.com/document/d/1SslZJYokA-Ho8SN8Vr5IzTnItuIDA6ACjilvLgCLcQ0/edit?usp=sharing

Here are the logs from the failed gluster deploy, before I commented out the 
filter in lvm.conf 
https://docs.google.com/document/d/1uvICx4NBMjC68hhPN98f-LAykyHUPvX5ipr-dUKk32s/edit?usp=sharing

I will update this thread as may be of use to some. Is there anywhere I should 
post it that other might find easier? Any advice on this final hurdle would be 
very gratefully received!

I have to go do some family stuff, but will update later if I find a solution!
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/T5PG64I23UALXBCE2HAW7OKO4TWYTFH5/


[ovirt-users] Re: strange behavior of fresh ovirt setup..

2021-02-13 Thread Edward Berger
I recall some earlier ovirt users list postings with similar issues with
the NAS.
One thought they solved it by changing the ovirt storage domain config NFS
version to a lower version instead of whatever it was defaulting to.
The other was blaming the NAS software ZFS underlying the NFS causing some
issues.

CentOS mdraid disks that are xfs formatted and exported NFS to oVirt seems
to work fine in the instances I'm aware of.

On Thu, Feb 11, 2021 at 2:51 AM  wrote:

> Justas Stankūnas 
>
> Wed, Feb 10, 11:29 PM (10 hours ago)
>
> to infra-owner
> Hi,
>
> so i have just installed oVirt on 4 nodes (oVirt node 4.4.4) with a hosted
> engine. For storage I have TrueNAS dedicated machine for NFS.
> hosts are all identical Lenovo thinkcentre m92 with (Intel(R) Core(TM)
> i5-3470T CPU @ 2.90GHz and 16 GB of RAM).
> everything is connected with a physical managed switch over 1Gbps each.
>
> everything is setup without any additional configuration. There is only a
> single vm (hosted engine). but hosts are constant errors and warnings being
> presented and I am lost even from where to start looking.
>
>
>
> Time|   Message
> Feb 10, 2021, 11:07:38 PM| |VM HostedEngine is not responding.
> Feb 10, 2021, 10:46:48 PM| |VM HostedEngine is not responding.
> Feb 10, 2021, 10:46:35 PM| |Failed to connect Host host1.local to Storage
> Pool Home
> Feb 10, 2021, 10:46:35 PM| |Host host1.local cannot access the Storage
> Domain(s) ovirt_manager_nfs attached to the Data Center Home. Setting Host
> state to Non-Operational.
> Feb 10, 2021, 10:46:35 PM| |Host host1.local reports about one of the
> Active Storage Domains as Problematic.
> Feb 10, 2021, 10:46:17 PM| |VDSM host1.local command Get Host Statistics
> failed: Connection timeout for host 'host1.local', last response arrived
> 17522 ms ago.
> Feb 10, 2021, 10:36:27 PM| |VM HostedEngine is not responding.
> Feb 10, 2021, 10:25:54 PM| |VM HostedEngine is not responding.
> Feb 10, 2021, 10:05:03 PM| |VM HostedEngine is not responding.
> Feb 10, 2021, 10:04:54 PM| |Failed to connect Host host1.local to Storage
> Pool Home
> Feb 10, 2021, 10:04:54 PM| |Host host1.local cannot access the Storage
> Domain(s) ovirt_manager_nfs attached to the Data Center Home. Setting Host
> state to Non-Operational.
> Feb 10, 2021, 10:04:54 PM| |Host host1.local reports about one of the
> Active Storage Domains as Problematic.
> Feb 10, 2021, 10:04:33 PM| |Invalid status on Data Center Home. Setting
> Data Center status to Non Responsive (On host host3.local, Error: Network
> error during communication with the Host.).
> Feb 10, 2021, 10:04:33 PM| |VDSM host3.local command SpmStatusVDS failed:
> Connection timeout for host 'host3.local', last response arrived 2184 ms
> ago.
> Feb 10, 2021, 9:54:41 PM| |VM HostedEngine is not responding.
> Feb 10, 2021, 9:33:47 PM| |VM HostedEngine is not responding.
> Feb 10, 2021, 9:25:52 PM| |VDSM host2.local command Get Host Statistics
> failed: Message timeout which can be caused by communication issues
> Feb 10, 2021, 9:23:22 PM| |VM HostedEngine is not responding.
> Feb 10, 2021, 9:12:51 PM| |Failed to connect Host host1.local to Storage
> Pool Home
> Feb 10, 2021, 9:12:51 PM| |Host host1.local cannot access the Storage
> Domain(s) nfs_vms attached to the Data Center Home. Setting Host state to
> Non-Operational.
> Feb 10, 2021, 9:12:51 PM| |Host host1.local reports about one of the
> Active Storage Domains as Problematic.
> Feb 10, 2021, 9:12:44 PM| |Invalid status on Data Center Home. Setting
> Data Center status to Non Responsive (On host host3.local, Error: Network
> error during communication with the Host.).
> Feb 10, 2021, 9:12:44 PM| |VDSM host1.local command Get Host Statistics
> failed: Connection timeout for host 'host1.local', last response arrived
> 16869 ms ago.
> Feb 10, 2021, 9:12:44 PM| |VDSM host3.local command SpmStatusVDS failed:
> Connection timeout for host 'host3.local', last response arrived 16931 ms
> ago.
> Feb 10, 2021, 9:02:51 PM| |VM HostedEngine is not responding.
> Feb 10, 2021, 8:52:25 PM| |VM HostedEngine is not responding.
> Feb 10, 2021, 8:31:34 PM| |VM HostedEngine is not responding.
> Feb 10, 2021, 8:21:06 PM| |VM HostedEngine is not responding.
> Feb 10, 2021, 8:20:36 PM| |Invalid status on Data Center Home. Setting
> Data Center status to Non Responsive (On host host3.local, Error: Network
> error during communication with the Host.).
> Feb 10, 2021, 8:20:36 PM| |VDSM host3.local command Get Host Statistics
> failed: Connection timeout for host 'host3.local', last response arrived
> 23346 ms ago.
> Feb 10, 2021, 8:10:11 PM| |Failed to connect Host host1.local to Storage
> Pool Home
> Feb 10, 2021, 8:10:11 PM| |Host host1.local cannot access the Storage
> Domain(s) ovirt_manager_nfs, nfs_vms, iso attached to the Data Center Home.
> Setting Host state to Non-Operational.
> Feb 10, 2021, 8:10:10 PM| |Host host1.local reports about one of the
> Active Storage Domains as Problematic.
> Feb 10, 2021, 8:10:01 PM|