[ovirt-users] Upgrading librbd1 and librados2 for oVirt 4.2.x

2018-08-30 Thread Andreas Elvers
Hello, I am in need of upgrading librbd1 and librados2 for a oVirt 4.2.x cluster. The cluster was installed via node ng. Taking the repository for Ceph Mimic or Luminous will end with a dependency problem because liburcu-cds.so.1 is already installed as a more recent version. A hint on how to

[ovirt-users] Re: 4.2 / 4.3 : Moving the hosted-engine to another storage

2019-04-02 Thread Andreas Elvers
> No need for that; but you will required to redeploy them from the new > engine to update their configuration. so I keep the old engine running while deploying the new engine on a different storage? Curious. I don't understand what "redeploy them [the old engine hosts] from the new engine to

[ovirt-users] Global maintenance and fencing of hosts

2019-04-12 Thread Andreas Elvers
I am wondering whether global maintenance inhibits fencing of non-responsive hosts. Is this so? Background: I plan on migrating the engine from one cluster to another. I understand this means to backup/restore the engine. While migrating the engine it is shut down and all VMs will continue

[ovirt-users] Re: Migrate self-hosted engine between cluster

2019-04-12 Thread Andreas Elvers
I am in the process of migrating the engine to a new cluster. I hope I will accomplish it this weekend. Fingers crossed. What you need to know: The migration is really a backup and restore process. 1. You create a backup of the engine. 2. Place the cluster into global maintenance and shutdown

[ovirt-users] Re: Migrate self-hosted engine between cluster

2019-04-12 Thread Andreas Elvers
This one is probably saving my weekend. Thanks a lot for your great work. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct:

[ovirt-users] Re: botched 3.6 -> 4.0/1/2 upgrade, how to recover

2019-05-13 Thread Andreas Elvers
If you can: restart your setup from scratch with at least 3 hosts. Really. If you don't face harsh consequences for doing so. It isn't just worth the hassle and pain. Use at least 3 hosts because: A node can corrupt due to many reasons. Natural causes, self inflicted ones. A corrupt node can

[ovirt-users] Re: deprecating export domain?

2019-05-17 Thread Andreas Elvers
You're so right. Thanks. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct:

[ovirt-users] Re: ovirt 4.3.3.7 cannot create a gluster storage domain

2019-05-17 Thread Andreas Elvers
Yeah. But I think this ist just an artefact of the current version. All images are in sync. dom_md/ids is an obsolete file anyway as the docs say. see vdsm/block-storage-domains https://www.ovirt.org/develop/developer-guide/vdsm/block-storage-domains.html

[ovirt-users] Re: Can't bring upgraded to 4.3 host back to cluster

2019-06-11 Thread Andreas Elvers
> probably one of the problems that I > haven't switched the cluster from iptables to firewalld. But this is just > my guess. > When switching to from 4.2.8 to 4.3.3 I did not change one host from iptables to firewalld as well. I was still able to change it later even if the documentation

[ovirt-users] Re: Does anyone have a positive experience with physical host to oVirt conversion?

2019-06-19 Thread Andreas Elvers
For converting my Debian boxes I do this. It is nothing fancy. Things you would do with plain physical systems as well. I have a VM for preparing VM images where I attach newly created disks, rsyncing all data from old host to the new image. Then I shut down the preparation VM, remove the new

[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-09 Thread Andreas Elvers
Yeah. I digged that bug report up as well. Our oVirt setup comprises of 6 oVirt hosts which form 2 datacenters Datacenter: Luise (has a ceph storage domain) node01 to node03 are a hyper converged oVirt glusterfs. This is cluster Luise01. My original is to move the engine onto this cluster.

[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-09 Thread Andreas Elvers
Short correction: All nodes oVirt and ceph are connected with 3 stacked Cisco 3750e providing 10 and 1Gbit ports. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement:

[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-09 Thread Andreas Elvers
Deployment on NFS again failed with TASK [ovirt.hosted_engine_setup : Wait for OVF_STORE disk content]. I'm lost. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement:

[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-09 Thread Andreas Elvers
I re-checked the versions with [root@node05 ~]# rpm -qa | grep engine ovirt-ansible-engine-setup-1.1.9-1.el7.noarch ovirt-engine-appliance-4.3-20190416.1.el7.x86_64 ovirt-hosted-engine-ha-2.3.1-1.el7.noarch ovirt-hosted-engine-setup-2.3.7-1.el7.noarch

[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-09 Thread Andreas Elvers
I updated all nodes to 4.3.3.1, but I did not remove the 4.2 repo from /etc/yum.repos.d. I did now remove the old repos, ovirt-engine-appliance is now shown as 4.3. I tried deployment to the hyper converged cluster. First try failed again with "The target Data Center does not contain the

[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-09 Thread Andreas Elvers
I'm current on 4.3.3.1 The software versions I am using: ovirt-hosted-engine-setup: 2.3.7 ovirt-ansible-hosted-engine-setup: 1.0.17 ovirt-engine-appliance: 4.2 I think I found the bug. :-/ ___ Users mailing list -- users@ovirt.org To unsubscribe send

[ovirt-users] Re: HostedEngine migration from ovirt1:/engine to gluster1:/engine (new ip)

2019-05-13 Thread Andreas Elvers
Shouldn't this be done by restoring the engine? Initial engine host and storage parameters are collected while doing the restore. It might be a bit far stretched, but at least be an automated repeatable experience. Are there really procedures where you manipulate the DHW directly? I never saw

[ovirt-users] Gluster volume heals and after 5 seconds has /dom_md/ids dirty again

2019-05-13 Thread Andreas Elvers
I restored my engine to a gluster volume named :/engine on a three node hyperconverged oVirt 4.3.3.1 cluster. Before restoring I was checking the status of the volumes. They were clean. No heal entries. All peers connected. gluster volume status looked good. Then I restored. This went well. The

[ovirt-users] Re: Gluster volume heals and after 5 seconds has /dom_md/ids dirty again

2019-05-13 Thread Andreas Elvers
Please note that I am running a hyper-converged NodeNG setup. I understand that upgrading single components is not really possible with a ovirt Node NG. And could probably break the datacenter upgrade path. Could you point out some reference for your suggestions? Docs, Bug reports or the

[ovirt-users] Re: Gluster volume heals and after 5 seconds has /dom_md/ids dirty again

2019-05-13 Thread Andreas Elvers
Yes. After a reboot you could have a sync issue for up to a few hours. But this issue persists now for 24 days. Additionally I see errors in the glustershd.log of the two hosts that are having heal info for that volume. The first node shows as OK and has no errors in its glustershd.log. The

[ovirt-users] Re: Gluster volume heals and after 5 seconds has /dom_md/ids dirty again

2019-05-13 Thread Andreas Elvers
> What version of gluster are you running at the moment? I'm running glusterfs-5.5-1.el7.x86_64 the one that comes with oVirt Node 4.3.3.1 ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy

[ovirt-users] Re: Gluster volume heals and after 5 seconds has /dom_md/ids dirty again

2019-05-14 Thread Andreas Elvers
There are plans for Node NG 4.3.5 to switch to Gluster V6. So far I still have unsynced entries with gluster 5.5. I tried to reset every brick, but that resolved to having the same heal info for that volume. I'm currently re-installing a node that has heal info on the gluster volume. Let's

[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-22 Thread Andreas Elvers
Sorry. The thread became separated. Yeah. I finished the migration. Simone proposed I had a corrupted backup. And yes. I was backtracking my backups to that point where I switched all datacenters to 4.3 compatibility and before doing any engine migration. Lucky me, I had that backup available.

[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-22 Thread Andreas Elvers
No need to worry. I'm so much impressed with the quality of oVirt. It has so many different setup scenarios, mine even a tech preview style one ;-) And good idea to file a bug for usability. To wrap things up: Me typing Luise1 instead of Luise01 was my error to begin with. It triggered

[ovirt-users] Re: ovirt 4.3.3.7 cannot create a gluster storage domain

2019-05-20 Thread Andreas Elvers
> Without this file [dom_md/ids], you will not have any > kind of storage. Ok. Sounds I'm kind of in trouble with that file being un-healable by gluster? ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org

[ovirt-users] Re: deprecating export domain?

2019-05-20 Thread Andreas Elvers
Thanks for the list of options. Will try it. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct:

[ovirt-users] OvfUpdateIntervalInMinutes restored to original value too early?

2019-05-10 Thread Andreas Elvers
I'm having a problem with restore. It fails while "Wait for OVF_STORE disk content". Looking at the ansible output and at the bug report https://bugzilla.redhat.com/show_bug.cgi?id=1644748 isn't the original value of 60 minutes for OvfUpdateIntervalInMinutes restored too early? [ INFO ] TASK

[ovirt-users] Re: ovirt 4.3.3.7 cannot create a gluster storage domain

2019-05-16 Thread Andreas Elvers
Why did you move to gluster v6? For the kicks? :-) The devs are currently evaluating for themselves whether they can switch to V6 for the upcoming releases. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to

[ovirt-users] Re: Gluster volume heals and after 5 seconds has /dom_md/ids dirty again

2019-05-15 Thread Andreas Elvers
ovirt node NG is a very good choice. Install experience, upgrade experience is very nice. 4.3 just got out of beta at Redhat. So the woes we experience is due to be a early adopter. ovirt node NG provides a stable selection of components and versions. ovirt node NG is tested. Spinning up your

[ovirt-users] Re: deprecating export domain?

2019-05-15 Thread Andreas Elvers
Maybe I overlooked the information, but in recent RHVE 4.3 docs the information how to use the export storage domain have been removed and there is no alternative to do so, but to detach a data domain and attach it somewhere else. But how can I move my VMs one by one to a new storage domain on

[ovirt-users] Re: oVirt upgrade version from 4.2 to 4.3

2019-05-16 Thread Andreas Elvers
AND HAVE BACKUPS OF THE ENGINE. But I suppose you already have automated backups of the engine. Right? After raising the datacenter compatibility to 4.3 the backups of 4.2 oVirt are incompatible with 4.3! ___ Users mailing list -- users@ovirt.org To

[ovirt-users] Re: oVirt upgrade version from 4.2 to 4.3

2019-05-16 Thread Andreas Elvers
You can take a look at the RHEV 4.3 documentation. https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.3/html/upgrade_guide/index There is another good read at: https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.3/html/administration_guide/index You

[ovirt-users] Re: OvfUpdateIntervalInMinutes restored to original value too early?

2019-05-12 Thread Andreas Elvers
You were right Simone. I used an older backup that I knew was just before I started with migrating the engine. I was able to restore that backup. I have now the engine on my gluster backed Cluster. Thanks ! ___ Users mailing list -- users@ovirt.org To

[ovirt-users] Re: Every now and then Backup question

2019-05-20 Thread Andreas Elvers
Bareos (a fork of Bacula) is nice. And they promote http://relax-and-recover.org/ a desaster recovery strategy. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement:

[ovirt-users] Re: Hosted engine restore went very wrong

2019-05-07 Thread Andreas Elvers
First I rebooted the hosted I tried to deploy to. The node status is now "OK" again. I will try to deploy on that node again. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement:

[ovirt-users] Re: Hosted engine restore went very wrong

2019-05-07 Thread Andreas Elvers
Deployment looked good, but the race condition (which I faced 5 times out of seven deployments) is back again. [ INFO ] TASK [ovirt.hosted_engine_setup : Add HE disks] [ INFO ] changed: [localhost] [ INFO ] TASK [ovirt.hosted_engine_setup : Register disk details] [ INFO ] ok: [localhost] [

[ovirt-users] Hosted engine restore went very wrong

2019-05-07 Thread Andreas Elvers
I really feel like an idiot. I tried to move our hosted engine from our Default datacenter to our Ceph Datacenter. I ran intro problems which were correctly addressed. see: https://lists.ovirt.org/archives/list/users@ovirt.org/thread/ZFCLFWRN6XR6KMHMC63O7J37D5GNPVKZ/ It was in case a race

[ovirt-users] Re: Hosted engine restore went very wrong

2019-05-07 Thread Andreas Elvers
The second deployment still triggers "The target Data Center does not contain the Virtual Disk". Trying again. This time deleting the old data in the gluster volume. Changing number of engine CPUs from 4 to 2. Hope this gives the race condition some different gravity.

[ovirt-users] Re: Hosted engine restore went very wrong

2019-05-07 Thread Andreas Elvers
Still no success. The ansible script errors out again at "The target Data Center does not contain the Virtual Disk". trying a fourth time. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement:

[ovirt-users] Re: Hosted engine restore went very wrong

2019-05-07 Thread Andreas Elvers
Meanwhile I was able to remove the old wrecked engine with the LocalHostedEngine on Port 6900. It was stable enough to re-install that node with engine undeployment selected. So I'm now able to bring back the engine to the old Default Datacenter and NFS.

[ovirt-users] Re: Stale hosted engine node information harmful ?

2019-05-02 Thread Andreas Elvers
on > all nodes > ovirt-ha-agent.service & ovirt-ha-broker.service . Once everywhere is > stopped, start > the 2 services on all nodes and wait 4-5min. > Last verify the status from each host, before removing global maintenance. > > Best Regards, > Strahil NikolovOn May

[ovirt-users] Re: Problem with restoring engine

2019-05-06 Thread Andreas Elvers
Yes! I can confirm that a second deployment succeeded on the second run. Great :-) I was also able to delete the Local Engine with virsh. I'd like to point out for reference, that this is only possible if a sasldb user is created like this: # saslpasswd2 -a libvirt username Password:

[ovirt-users] Re: Stale hosted engine node information harmful ?

2019-05-02 Thread Andreas Elvers
Yeah. That did the trick. What I did was: Enable global maintenance. On every affected host: # systemctl stop ovirt-ha-agent # hosted-engine --clean-metadata --host-id=1 --force-clean (repeat for every id) Finally start the stopped ha-agents. Check the status: # hosted-engine --vm-status Disable

[ovirt-users] Stale hosted engine node information harmful ?

2019-05-02 Thread Andreas Elvers
I have 5 nodes (node01 to node05). Originally all those nodes were part of our default datacenter/cluster with a NFS storage domain for vmdisk, engine and iso-images. All five nodes were engine HA nodes. Later node01, node02 and node03 were re-installed to have engine HA removed. Then those

[ovirt-users] Problem with restoring engine

2019-05-05 Thread Andreas Elvers
Hello today I tried to migrate the hosted engine from our Default Datacenter (NFS) to our Ceph Datacenter. The deployment worked with the automatic "hosted-engine --deploy --restore-from-file=backup/file_name" command. Perfect. Only thing is: I messed up with the cluster name. The name should

[ovirt-users] Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-08 Thread Andreas Elvers
Hello, when trying to deploy the engine using "hosted-engine --deploy --restore-from-file=myenginebackup" the ansible playbook errors out at [ INFO ] TASK [ovirt.hosted_engine_setup : Trigger hosted engine OVF update and enable the serial console] [ INFO ] changed: [localhost] [ INFO ]

[ovirt-users] Re: Storage domain format version info

2019-07-09 Thread Andreas Elvers
I'd like to add a question on storage domains. When will storage domains become obsolete? Like kind of a road map. I'm asking because we are operating a OpenStack Cinder V1 Domain to access Ceph RBD images. I know this is only a former tech preview, but this preview is doing solid work. So I'd

[ovirt-users] Re: Upgrading from 4.2.8 to 4.3.3 broke Node NG GlusterFS

2019-04-24 Thread Andreas Elvers
After rebooting the node that was not able to mount the gluster volume things improved eventually. SPM went away and restarted for the Datacenter and suddenly node03 was able to mount the gluster volume. In between I was down to 1/3 active Bricks which results in read only glusterfs. I was

[ovirt-users] Upgrading from 4.2.8 to 4.3.3 broke Node NG GlusterFS

2019-04-24 Thread Andreas Elvers
Hi, I am currently upgrading my oVirt setup from 4.2.8 to 4.3.3.1. The setup consists of: Datacenter/Cluster Default: [fully upgraded to 4.3.3.1] 2 nodes (node04,node05)- NFS storage domain with self hosted engine Datacenter Luise: Cluster1: 3 nodes (node01,node02,node03) - Node NG with

[ovirt-users] Re: Upgrading from 4.2.8 to 4.3.3 broke Node NG GlusterFS

2019-04-24 Thread Andreas Elvers
Restarting improved things a little bit. Still bricks on node03 are shown as down, but "gluster volume status" is looking better. Saiph:~ andreas$ ssh node01 gluster volume status vmstore Status of volume: vmstore Gluster process TCP Port RDMA Port Online Pid

[ovirt-users] Re: Upgrading from 4.2.8 to 4.3.3 broke Node NG GlusterFS

2019-04-24 Thread Andreas Elvers
"systemctl restart glusterd" on node03 did not help. Still getting: node03#: ls -l /rhev/data-center/mnt/glusterSD/node01.infra.solutions.work:_vmstore ls: cannot access /rhev/data-center/mnt/glusterSD/node01.infra.solutions.work:_vmstore: Transport endpoint is not connected Engine still

[ovirt-users] Re: Upgrading from 4.2.8 to 4.3.3 broke Node NG GlusterFS

2019-04-24 Thread Andreas Elvers
The file handle is stale so find will display: "find: '/rhev/data-center/mnt/glusterSD/node01.infra.solutions.work:_vmstore': Transport endpoint is not connected" "stat /rhev/data-center/mnt/glusterSD/node01.infra.solutions.work:_vmstore" will output stat: cannot stat

[ovirt-users] Upgrading librbd1 and librados2 for oVirt 4.2.x

2018-08-24 Thread andreas . elvers+ovirtforum
Hallo Group ! I am in need of upgrading librbd1 and librados2 for a oVirt 4.2.x cluster. The cluster was installed via node ng. Taking the repository for Ceph Mimic or Luminous will end in a dependency problem because liburcu-cds.so.1 is already installed as a more recent version provided by

[ovirt-users] Re: Upgrading librbd1 and librados2 for oVirt 4.2.x

2018-08-24 Thread andreas . elvers+ovirtforum
I found a way. [root@node02 yum.repos.d]# cat luminious.repo [ovirt-4.2-centos-ceph-luminous] enabled=1 name = CentOS-7 - ceph luminous baseurl = http://mirror.centos.org/centos/7/storage/$basearch/ceph-luminous/ gpgcheck = 1 enabled = 1 gpgkey =

[ovirt-users] Re: 4.2 / 4.3 : Moving the hosted-engine to another storage

2019-04-02 Thread andreas . elvers+ovirtforum
Thanks for your answer. > Yes, now you can do it via backup and restore: > take a backup of the engine with engine-backup and restore it on a new > hosted-engine VM on a new storage domain with: > hosted-engine --deploy --restore-from-file=mybackup.tar.gz That is great news. Just to clear things

[ovirt-users] 4.2 / 4.3 : Moving the hosted-engine to another storage

2019-04-01 Thread andreas . elvers+ovirtforum
Hi friends of oVirt, Roughly 3 years ago a user asked about the options he had to move the hosted engine to some other storage. The answer by Simone Tiraboschi was that it would largely not be possible, because of references in the database to the node the engine was hosted. This information