[ovirt-users] Upgrading librbd1 and librados2 for oVirt 4.2.x

2018-08-30 Thread Andreas Elvers
Hello,

I am in need of upgrading librbd1 and librados2 for a oVirt 4.2.x cluster.
The cluster was installed via node ng.
 
Taking the  repository for Ceph Mimic or Luminous will end with a dependency
problem because liburcu-cds.so.1 is already installed as a more recent version.

A hint on how to get a more recent versions of the above libraries would be very
much appreciated.

Thanks,
- Andreas


___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZA74LCDEUKUIWFQVI6HNWYDVAQOPGUS4/


[ovirt-users] Re: 4.2 / 4.3 : Moving the hosted-engine to another storage

2019-04-02 Thread Andreas Elvers

> No need for that; but you will required to redeploy them from the new
> engine to update their configuration.

so I keep the old engine running while deploying the new engine on a different
storage? Curious. 

I don't understand what "redeploy them [the old engine hosts] from 
the new engine to update their configuration" means. 

In fact the current engine incarnation is running on an old cluster, that
has no access to the new storage (the cluster is to go away). 
The engine is also managing the new cluster to which we want to move the 
engine. 
The engine is the only piece that keeps us from shutting down the old cluster. 
That's the motivation for restoring the engine on the new cluster.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WVOFX4ZXDF4PKR5JHW4BKR45ZQRC56CS/


[ovirt-users] Global maintenance and fencing of hosts

2019-04-12 Thread Andreas Elvers
I am wondering whether global maintenance inhibits fencing of non-responsive 
hosts. Is this so? 

Background: I plan on migrating the engine from one cluster to another. I 
understand this means to backup/restore the engine. While migrating the engine 
it is shut down and all VMs will continue running. This is good. When starting 
the engine in the new location, I really don't want the engine to fence any 
host on its own, because of reasons I can not yet know.

So is global maintenance enough to suppress fencing, or do I have to deactivate 
fencing on all hosts? 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3KRJRAWBFC4QAXVYXUGXVA3324USBHBN/


[ovirt-users] Re: Migrate self-hosted engine between cluster

2019-04-12 Thread Andreas Elvers
I am in the process of migrating the engine to a new cluster. I hope I will 
accomplish it this weekend. Fingers crossed.

What you need to know:

The migration is really a backup and restore process. 

1. You create a backup of the engine.
2. Place the cluster into global maintenance and shutdown the engine. 
3. Then you create a new engine on the other cluster and before engine-setup 
you restore the engine backup.
4. Start the new engine.

From here I'm not yet sure what should happen in what order:

- re-install the nodes with the old engine to remove the Engine HA Setup from 
that host.
- re-install nodes on the new cluster to receive the Engine HA Setup.

There is documentation about backup and restore of the engine here: 
https://ovirt.org/documentation/self-hosted/chap-Backing_up_and_Restoring_an_EL-Based_Self-Hosted_Environment.html
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/H4IDZJHKCLOVCDEN4ZRJI4KSPNPLWP5H/


[ovirt-users] Re: Migrate self-hosted engine between cluster

2019-04-12 Thread Andreas Elvers
This one is probably saving my weekend. Thanks a lot for your great work.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/2IJDVBDF5ZOUL2MJMXBSSBVF2ZBZKNIG/


[ovirt-users] Re: botched 3.6 -> 4.0/1/2 upgrade, how to recover

2019-05-13 Thread Andreas Elvers
If you can: restart your setup from scratch with at least 3 hosts. Really. If 
you don't face harsh consequences for doing so. It isn't just worth the hassle 
and pain. Use at least 3 hosts because: A node can corrupt due to many reasons. 
Natural causes, self inflicted ones. A corrupt node can only recover through 
re-installation. This is triggered by the engine with VMs safe on the other 
nodes. On a single node system how can this be possible with only a corrupted 
node at hand? If you really can only stick to a one host setup you could try 
Proxmox. They have a community edition. 

For multi host setups choose oVirt. It is just great.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/NISRIS3ZGGBSSQTNOE4PF53HBROI7RQ5/


[ovirt-users] Re: deprecating export domain?

2019-05-17 Thread Andreas Elvers
You're so right. Thanks.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/KIT4FV6ATTRGEEDTU75PDU556NRGX3TO/


[ovirt-users] Re: ovirt 4.3.3.7 cannot create a gluster storage domain

2019-05-17 Thread Andreas Elvers
Yeah. But I think this ist just an artefact of the current version. All images 
are in sync.
 dom_md/ids is an obsolete file anyway as the docs say.

see vdsm/block-storage-domains
https://www.ovirt.org/develop/developer-guide/vdsm/block-storage-domains.html
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MVQXFPVRJCHGL4YJ3KAJQLXXBTGJ6LNO/


[ovirt-users] Re: Can't bring upgraded to 4.3 host back to cluster

2019-06-11 Thread Andreas Elvers
> probably one of the problems that I
> haven't switched the cluster from iptables to firewalld. But this is just
> my guess.
> 

When switching to from 4.2.8 to 4.3.3 I did not change one host from iptables 
to firewalld as well. I was still able to change it later even if the 
documentation somewhere said iptables support is to be removed in 4.3.  
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QMJGVX4LFSMGXKRN3ZS7UOAYMFBWC6XS/


[ovirt-users] Re: Does anyone have a positive experience with physical host to oVirt conversion?

2019-06-19 Thread Andreas Elvers
For converting my Debian boxes I do this. It is nothing fancy. Things you would 
do with plain physical systems as well.

I have a VM for preparing VM images where I attach newly created disks, 
rsyncing all data from old host to the new image. 

Then I shut down the preparation VM, remove the new disk from it.

Create a new VM and attach the disk to that VM.

I boot a rescue CD, fix /etc/fstab and networking, Re-Install grub to the disk.

Done.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/4TKO7YKDOZEM67XTVAIZ24IK7LTCB5AH/


[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-09 Thread Andreas Elvers
Yeah. I digged that bug report up as well.

Our oVirt setup comprises of

6 oVirt hosts which form 2 datacenters 

Datacenter: Luise (has a ceph storage domain)
node01 to node03 are a hyper converged oVirt glusterfs. This is cluster 
Luise01. My original is to move the engine onto this cluster. 

Datacenter: Default Cluster Default
node04 and node05. It uses an nfs storage: see below. The engine used to live 
here. 

Datacenter Luise Cluster: Luise02
node06 is currently the only node. This cluster will gain all nodes from the 
Default Datacenter as soon as it is decommissioned.

All nodes are by Supermicro 4 cores 48GB Intel Xeon E5506 @ 2.13 GHz
They use two bonded 1Gbit Links for connectivity. I was sceptical with 
performance, but we are running multiple around 10-20 TYPO3 CMSes plus lesser 
IO hungry things in around 3 to 6 VMs per Node. They all use Ceph as storage 
domain.

All nodes I tried to deploy to are without virtual load.

Storage:


The glusterfs is provided by the hyper converged setup provided by oVirt Node 
NG. It is quite small because is just our gateway to boot the cinder VM that 
will guide our ceph nodes to their rbd destiny. Every Node has a 250GB SSD. 
This gives us 90GB for the :/engine gluster volume. There is also a 100GB 
volume for the supporting VMs that provide the bootstrap to cinder. There is 
little IO going to these volumes.

The ceph is a 10Gbit connected cluster of 3 Nodes with SSD OSDs. The oVirt 
connect via a Cisco 

The NFS server is serving a Raid-10 btrfs formatted filesystem added with a 
bcache.

My migration history so far:

Tried to Migrate to Luise Datacenter I did a typo with entering Luise1 instead 
the correct cluster Luise01. Ironically that migration succeeded. 

Tried to Migrate back to Default. First one erred out with "Disk Image is not 
available in the cluster". Second one succeeded after following your advice to 
try it a second time. I removed the accidentally created Luise1 cluster.

Tried to Migrate to Luise again. This time I forgot to completely shut down the 
hosted engine on Default Datacenter. I shut it down and tried again.

All subsequent retires to deploy to the hyper converged Luise01 cluster failed. 
I used the Local Engine to re-install node05 which was hosting the engine I 
forget to shut down. I was not able to access the web interface for some 
reason, so I supposed it was damaged in a way. 

After 4 tries of deployment to our hyper converged cluster that all erred with 
the "Disk Image is not available" I'm not trying to go back to the Default 
Datacenter. Twice it erred out at "Wait for OVF_STORE disk content"
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/DCIZCN5F6J4DBQHNDM3U4O7HX5KTDVUH/


[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-09 Thread Andreas Elvers
Short correction: All nodes oVirt and ceph are connected with 3 stacked Cisco 
3750e providing 10 and 1Gbit ports.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/JXEKEGGLYIIPVPGJGV6HQ476AVVWNZID/


[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-09 Thread Andreas Elvers
Deployment on NFS again failed with TASK [ovirt.hosted_engine_setup : Wait for 
OVF_STORE disk content].

I'm lost.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/DFG5ECI2DADG7GVREBFAOLL7BQSNDGWK/


[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-09 Thread Andreas Elvers
I re-checked the versions with

[root@node05 ~]# rpm -qa | grep engine
ovirt-ansible-engine-setup-1.1.9-1.el7.noarch
ovirt-engine-appliance-4.3-20190416.1.el7.x86_64
ovirt-hosted-engine-ha-2.3.1-1.el7.noarch
ovirt-hosted-engine-setup-2.3.7-1.el7.noarch
ovirt-ansible-hosted-engine-setup-1.0.17-1.el7.noarch
python-ovirt-engine-sdk4-4.3.1-2.el7.x86_64

so probably I was already on the latest versions. I did not do any updated. 
First I was doing a "rpm info" to get the version.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WOXKWWN6QVF4D6HORBL3ZD7YZQ4XDCZD/


[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-09 Thread Andreas Elvers
I updated all nodes to 4.3.3.1, but I did not remove the 4.2 repo from 
/etc/yum.repos.d. 

I did now remove the old repos, ovirt-engine-appliance is now shown as 4.3.

I tried deployment to the hyper converged cluster. First try failed again with 
"The target Data Center does not contain the Virtual Disk.". I'm retrying. 
Should this fail again, I will try a deployment to the old NFS backed Default 
Datacenter.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/J626ORT5IYZL3HCJRLGLV5N3D6Q2UWZF/


[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-09 Thread Andreas Elvers
I'm current on 4.3.3.1

The software versions I am using:

ovirt-hosted-engine-setup: 2.3.7
ovirt-ansible-hosted-engine-setup: 1.0.17
ovirt-engine-appliance: 4.2

I think I found the bug. :-/ 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/J4Y5HE77S33TCWPME4CGCT7OLG3MGDSC/


[ovirt-users] Re: HostedEngine migration from ovirt1:/engine to gluster1:/engine (new ip)

2019-05-13 Thread Andreas Elvers
Shouldn't this be done by restoring the engine? Initial engine host and storage 
parameters are collected while doing the restore. It might be a bit far 
stretched, but at least be an automated repeatable experience.

Are there really procedures where you manipulate the DHW directly? I never saw 
a reference in the documentation. 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/5JYINPWJINWQPIN2LQO2ZTEUGHY3YKQ2/


[ovirt-users] Gluster volume heals and after 5 seconds has /dom_md/ids dirty again

2019-05-13 Thread Andreas Elvers
I restored my engine to a gluster volume named :/engine on a three node 
hyperconverged oVirt 4.3.3.1 cluster. Before restoring I was checking the 
status of the volumes. They were clean. No heal entries. All peers connected. 
gluster volume status looked good. Then I restored. This went well. The engine 
is up. But the engine gluster volume shows entries on node02 and node03. The 
engine was installed to node01. I have to deploy the engine to the other two 
hosts to reach full HA, but I bet maintenance is not possible until the volume 
is healed. 

I tried "gluster volume heal engine" also with added "full". The heal entries 
will disappear for a few seconds and then /dom_md/ids will pop up again. The 
__DIRECT_IO_TEST__ will join later. The split-brain info has no entries. Is 
this some kind of hidden split brain? Maybe there is data on node01 brick which 
got not synced to the other two nodes? I can only speculate. Gluster docs say: 
this should heal. But it doesn't.  I have two other volumes. Those are fine. 
One of them containing 3 VMs that are running. I also tried to shut down the 
engine, so no-one was using the volume. Then heal. Same effect. Those two files 
will always show up. But none other. Heal can always be started successfully 
from any of the participating nodes.

Reset the volume bricks one by one and cross fingers? 

[root@node03 ~]#  gluster volume heal engine info
Brick node01.infra.solutions.work:/gluster_bricks/engine/engine
Status: Connected
Number of entries: 0

Brick node02.infra.solutions.work:/gluster_bricks/engine/engine
/9f4d5ae9-e01d-4b73-8b6d-e349279e9782/dom_md/ids
/__DIRECT_IO_TEST__
Status: Connected
Number of entries: 2

Brick node03.infra.solutions.work:/gluster_bricks/engine/engine
/9f4d5ae9-e01d-4b73-8b6d-e349279e9782/dom_md/ids
/__DIRECT_IO_TEST__
Status: Connected
Number of entries: 2
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/L3YCRPRAGPUMBZIBFOPT6L4B7H4M6HLS/


[ovirt-users] Re: Gluster volume heals and after 5 seconds has /dom_md/ids dirty again

2019-05-13 Thread Andreas Elvers
Please note that I am running a hyper-converged NodeNG setup. I understand that 
upgrading single components is not really possible with a ovirt Node NG. And 
could probably break the datacenter upgrade path.

Could you point out some reference for your suggestions? Docs, Bug reports or 
the sorts?

> I encountered serious issues with 5.3-5.5 (crashing bricks, multiple brick 
> processes for
> the same brick causing disconnects and excessive heals). I had better luck 
> with 5.6,
> although it’s not clear to me if the duplicate brick process issue is still 
> present in
> that version. I finally jumped to 6 which has been more stable for me. I’d 
> recommend
> upgrading at least to 5.6 if not going right to 6.1.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/BFOB6QEST6EAGEZGSCM4GO7BFRUYCKEI/


[ovirt-users] Re: Gluster volume heals and after 5 seconds has /dom_md/ids dirty again

2019-05-13 Thread Andreas Elvers
Yes. After a reboot you could have a sync issue for up to a few hours. But this 
issue persists now for 24 days. Additionally I see errors in the glustershd.log 
of the two hosts that are having heal info for that volume. The first node 
shows as OK and has no errors in its glustershd.log.

The errors are like this:

[2019-05-13 15:18:40.808945] W [MSGID: 114031] 
[client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-engine-client-1: remote 
operation failed. Path:  
(95ba9fb2-b0ae-436c-9c31-2779cf202235) [No such file or directory]
[2019-05-13 15:18:40.809113] W [MSGID: 114031] 
[client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-engine-client-2: remote 
operation failed. Path:  
(95ba9fb2-b0ae-436c-9c31-2779cf202235) [No such file or directory]
[root@node02 ~]#

Looks like the first node is sane and the other two are the masters but are not 
so sane. :-/ 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/4JROBW3YYGM65YNATOF7EHSMMA3H6NFL/


[ovirt-users] Re: Gluster volume heals and after 5 seconds has /dom_md/ids dirty again

2019-05-13 Thread Andreas Elvers
> What version of gluster are you running at the moment?

I'm running glusterfs-5.5-1.el7.x86_64 the one that comes with oVirt Node 
4.3.3.1
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/OPMOUE5FQRYF4Q6ZMPBDFM6EYAXRNM44/


[ovirt-users] Re: Gluster volume heals and after 5 seconds has /dom_md/ids dirty again

2019-05-14 Thread Andreas Elvers
There are plans for Node NG 4.3.5 to switch to Gluster V6. 

So far I still have unsynced entries with gluster 5.5. I tried to reset every 
brick, but that resolved to having the same heal info for that volume.

I'm currently re-installing a node that has heal info on the gluster volume. 
Let's see what happens after that.

See: 
https://lists.ovirt.org/archives/list/de...@ovirt.org/thread/IVTYCFKCQ5P2WZ43CM245PTX5T54PLQX/
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/VUBOALBOM2HWS5ZA2G3MP52E4FKRPBYQ/


[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-22 Thread Andreas Elvers
Sorry. The thread became separated. Yeah. I finished the migration. Simone 
proposed I had a corrupted backup. And yes. I was backtracking my backups to 
that point where I switched all datacenters to 4.3 compatibility and before 
doing any engine migration. Lucky me, I had that backup available.

I think my backups became corrupted, because I did a backup from the engine 
after I moved it to a cluster that was originally not defined. I had a typo in 
the cluster name. And used that backup and tried to get the engine back to my 
nfs storage. That worked. I did another backup at that place again and from 
that point on the backup would always break the deployment for whatever reason. 

The logs are available at 

https://lists.ovirt.org/archives/list/users@ovirt.org/thread/HXJZWN7L6WXSDLQB3SMIAUEJGI6QFBGC/
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/TNHQO4VPZES7NXUR5KI3LACBQ4NK2FM4/


[ovirt-users] Re: Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-22 Thread Andreas Elvers
No need to worry. I'm so much impressed with the quality of oVirt. It has so 
many different setup scenarios, mine even a tech preview style one ;-) 

And good idea to file a bug for usability. To wrap things up:

Me typing Luise1 instead of Luise01 was my error to begin with. It triggered 
everything. Luise01 is our hyper converged oVirt node cluster. It is our ceph 
bootstrap. Yeah... How could you forget the zero..

node01, node02 and node03 are forming Luise01.

I ordered deployment of the engine to node01 on Luise1 and hosted-engine 
happily did so. Creating a new cluster Luise1 on node01 masking the glusterfs 
bricks on node01 for the remaining Luise01 cluster nodes.

This most probably also triggered this issue: 
https://lists.ovirt.org/archives/list/users@ovirt.org/thread/L3YCRPRAGPUMBZIBFOPT6L4B7H4M6HLS/

It has already been addressed by Nir Soffer. See 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/OSZGHT65OUUSGKTLYKJVZPVRFTBPI3EO/
 

I will create a bug report for that. A node that has already joined a cluster 
should not be a viable destination for an engine deployment to a new to create 
cluster. 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/YIDVCKVYKRGQUDK4R22Z6GNWRHHZWNMF/


[ovirt-users] Re: ovirt 4.3.3.7 cannot create a gluster storage domain

2019-05-20 Thread Andreas Elvers

> Without this file [dom_md/ids], you will not have any
> kind of storage.

Ok. Sounds I'm kind of in trouble with that file being un-healable by gluster?
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/FIR74ALG7WJFIPNIAHZH4PRBY7UI2QRO/


[ovirt-users] Re: deprecating export domain?

2019-05-20 Thread Andreas Elvers
Thanks for the list of options. Will try it.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/KFHXK5QYWZRN3G2MNLTSKWN4EK42LVQH/


[ovirt-users] OvfUpdateIntervalInMinutes restored to original value too early?

2019-05-10 Thread Andreas Elvers
I'm having a problem with restore. It fails while "Wait for OVF_STORE disk 
content". Looking at the ansible output and at the bug report 
https://bugzilla.redhat.com/show_bug.cgi?id=1644748 isn't the original value of 
60 minutes for OvfUpdateIntervalInMinutes restored too early?

[ INFO  ] TASK [ovirt.hosted_engine_setup : Fetch host SPM_ID]
[ INFO  ] changed: [localhost -> engine.infra.solutions.work]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Parse host SPM_ID]
[ INFO  ] ok: [localhost -> engine.infra.solutions.work]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Restore original 
DisableFenceAtStartupInSec]
[ INFO  ] changed: [localhost -> engine.infra.solutions.work]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Remove DisableFenceAtStartupInSec 
temporary file]
[ INFO  ] changed: [localhost -> engine.infra.solutions.work]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Restore original 
OvfUpdateIntervalInMinutes]
[ INFO  ] changed: [localhost -> engine.infra.solutions.work]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Remove OvfUpdateIntervalInMinutes 
temporary file]
[ INFO  ] changed: [localhost -> engine.infra.solutions.work]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Removing temporary value]
[ INFO  ] changed: [localhost -> engine.infra.solutions.work]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Restoring original value]
[ INFO  ] changed: [localhost -> engine.infra.solutions.work]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Remove temporary directory for 
ansible as postgres user]
[ INFO  ] changed: [localhost -> engine.infra.solutions.work]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Fail if he_force_ip4 and 
he_force_ip6 are set at the same time]
[ INFO  ] skipping: [localhost]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Prepare getent key]
[ INFO  ] skipping: [localhost]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Trigger hosted engine OVF update 
and enable the serial console]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Wait until OVF update finishes]
[ INFO  ] ok: [localhost]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Parse OVF_STORE disk list]
[ INFO  ] ok: [localhost]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Check OVF_STORE volume status]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Wait for OVF_STORE disk content]
[ ERROR ] {u'_ansible_parsed': True, u'stderr_lines': [u'20+0 records in', 
u'20+0 records out', u'10240 bytes (10 kB) copied, 0.000164261 s, 62.3 MB/s', 
u'tar: 1e74609b-51e1-45c8-9106-2596ee59ba3a.ovf: Not found in archive', u'tar: 
Exiting with failure status due to previous errors'], u'changed': True, u'end': 
u'2019-05-10 12:57:57.610068', u'_ansible_item_label': {u'image_id': 
u'59d16f28-64ed-4c3f-aaad-7d98e2718f53', u'name': u'OVF_STORE', u'id': 
u'98bd5cfc-809d-4d40-83e0-98e379af5f97'}, u'stdout': u'', u'failed': True, 
u'_ansible_item_result': True, u'msg': u'non-zero return code', u'rc': 2, 
u'start': u'2019-05-10 12:57:56.928294', u'attempts': 12, u'cmd': u"vdsm-client 
Image prepare storagepoolID=597f329c-0296-03af-0369-0139 
storagedomainID=8f115ea6-91ec-4e10-9240-bc8eda47b7ff 
imageID=98bd5cfc-809d-4d40-83e0-98e379af5f97 
volumeID=59d16f28-64ed-4c3f-aaad-7d98e2718f53 | grep path | awk '{ print $2 }' 
| xargs -I{} sudo -u vdsm dd if={} | tar -tvf - 1e74609b-51e1-45c8-9106-259
 6ee59ba3a.ovf", u'item': {u'image_id': 
u'59d16f28-64ed-4c3f-aaad-7d98e2718f53', u'name': u'OVF_STORE', u'id': 
u'98bd5cfc-809d-4d40-83e0-98e379af5f97'}, u'delta': u'0:00:00.681774', 
u'invocation': {u'module_args': {u'warn': False, u'executable': None, 
u'_uses_shell': True, u'_raw_params': u"vdsm-client Image prepare 
storagepoolID=597f329c-0296-03af-0369-0139 
storagedomainID=8f115ea6-91ec-4e10-9240-bc8eda47b7ff 
imageID=98bd5cfc-809d-4d40-83e0-98e379af5f97 
volumeID=59d16f28-64ed-4c3f-aaad-7d98e2718f53 | grep path | awk '{ print $2 }' 
| xargs -I{} sudo -u vdsm dd if={} | tar -tvf - 
1e74609b-51e1-45c8-9106-2596ee59ba3a.ovf", u'removes': None, u'argv': None, 
u'creates': None, u'chdir': None, u'stdin': None}}, u'stdout_lines': [], 
u'stderr': u'20+0 records in\n20+0 records out\n10240 bytes (10 kB) copied, 
0.000164261 s, 62.3 MB/s\ntar: 1e74609b-51e1-45c8-9106-2596ee59ba3a.ovf: Not 
found in archive\ntar: Exiting with failure status due to previous errors', 
u'_ansible_no_log': False}
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/HJ272VCIHH5NIVDHJWP4HRQLV7BYQJHI/


[ovirt-users] Re: ovirt 4.3.3.7 cannot create a gluster storage domain

2019-05-16 Thread Andreas Elvers
Why  did you move to gluster v6? For the kicks? :-) The devs are currently 
evaluating for themselves whether they can switch to V6 for the upcoming 
releases.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/JYZRKA4QBTXYDR3WXFRW7IXLCSGGVSLC/


[ovirt-users] Re: Gluster volume heals and after 5 seconds has /dom_md/ids dirty again

2019-05-15 Thread Andreas Elvers
ovirt node NG is a very good choice. Install experience, upgrade experience is 
very nice. 4.3 just got out of beta at Redhat. So the woes we experience is due 
to be a early adopter.

ovirt node NG provides a stable selection of components and versions. ovirt 
node NG is tested. Spinning up your own oVirt mix seems more error prone to me.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/DQWUV6EMBELX65YKK2CDJULNBPKFHH6N/


[ovirt-users] Re: deprecating export domain?

2019-05-15 Thread Andreas Elvers
Maybe I overlooked the information, but in recent RHVE 4.3 docs the information 
how to use the export storage domain have been removed and there is no 
alternative to do so, but to detach a data domain and attach it somewhere else. 
But how can I move my VMs one by one to a new storage domain on a different 
datacenter without completely detaching the original storage domain? 

I don't want to bring down all of my VMs on the old storage domain for import. 
I want to export and import them one by one. When all VMs are moved to the new 
data center only then I want to decommission
the old data center.

What is the rationale to deprecate the export storage and already remove 
documentation when there seems to be no alternative available?
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/O4HXTTYAFOC53XQR432ZFMN3JL6IJBIC/


[ovirt-users] Re: oVirt upgrade version from 4.2 to 4.3

2019-05-16 Thread Andreas Elvers
AND HAVE BACKUPS OF THE ENGINE. But I suppose you already have automated 
backups of the engine. Right?

After raising the datacenter compatibility to 4.3 the backups of 4.2 oVirt are 
incompatible with 4.3!
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/4OK5TM7E2KNGQONN746WA57ID66IDKKY/


[ovirt-users] Re: oVirt upgrade version from 4.2 to 4.3

2019-05-16 Thread Andreas Elvers
You can take a look at the RHEV 4.3 documentation.

https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.3/html/upgrade_guide/index

There is another good read at:

https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.3/html/administration_guide/index

You should probably wait for 4.3.4 to arrive, since 4.3.3 is still a little bit 
rough on the edges. I can currently only create a new VM via the templates, 
because the normal creation dialog errors, when I change the datacenter. 
Otherwise the upgrade went very well.

If you are running off of oVirt Node NG it is very straightforward.

I did:

1. Check of all nodes are using firewalld. You have to switch all nodes to use 
firewalld, because Iptables is deprecated and removed in 4.3 as it was said in 
documentation. You have to re-install every node for that. So every node has to 
go through maintenance.

2. Upgrade the engine. First to the 4.2.8 then to 4.3.x

My upgrade flow for this is:

Engine Update START

Minor Upgrades first

enable global maintenance mode

login to engine
engine-upgrade-check
yum update "ovirt-*-setup*"
engine-setup
yum update
when success: disable global maintenance mode

Major Upgrade

enable global maintenance mode

login to engine
yum install 
http://resources.ovirt.org/pub/yum-repo/ovirt-release[releasenumber].rpm
engine-upgrade-check
yum update "ovirt-*-setup*"
engine-setup
remove the old ovirt release from /etc/yum.repos.d
yum update
when success: disable global maintenance mode

Engine Upgrade END

3. Upgrade the nodes. 

First to 4.2.8 then as specified in ovirt release info:

yum install 
https://resources.ovirt.org/pub/ovirt-4.3/rpm/el7/noarch/ovirt-node-ng-image-update-4.3.3-1.el7.noarch.rpm

4. Upgrade the Datacenter Compatibility to 4.3
This includes rebooting all VMs at one point. You can read about it in the 
documentation.

Then you're done.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SBTRBHU6WMA3BGHAVX3BZKQF3PAQBC5H/


[ovirt-users] Re: OvfUpdateIntervalInMinutes restored to original value too early?

2019-05-12 Thread Andreas Elvers
You were right Simone. I used an older backup that I knew was just before I 
started with migrating the engine. I was able to restore that backup. I have 
now the engine on my gluster backed Cluster.

Thanks !
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/BNF7YZJLNBFDIKXEOZ4AXUNWJ5EEOUOG/


[ovirt-users] Re: Every now and then Backup question

2019-05-20 Thread Andreas Elvers
Bareos (a fork of Bacula) is nice. And they promote 
http://relax-and-recover.org/ a desaster recovery strategy.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ODNXJDWIZBNCL445SUTMKENDQWUZCCX3/


[ovirt-users] Re: Hosted engine restore went very wrong

2019-05-07 Thread Andreas Elvers
First I rebooted the hosted I tried to deploy to. The node status is now "OK" 
again. I will try to deploy on that node again.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/7QPI7HXV3G3K2XYUM7ZS3JNDC2Y5PCXM/


[ovirt-users] Re: Hosted engine restore went very wrong

2019-05-07 Thread Andreas Elvers
Deployment looked good, but the race condition (which I faced 5 times out of 
seven deployments) is back again.

[ INFO  ] TASK [ovirt.hosted_engine_setup : Add HE disks]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Register disk details]
[ INFO  ] ok: [localhost]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Set default graphics protocols]
[ INFO  ] ok: [localhost]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Check if FIPS is enabled]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Select graphic protocols]
[ INFO  ] skipping: [localhost]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Add VM]
[ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is "[Cannot 
attach Virtual Disk. The target Data Center does not contain the Virtual 
Disk.]". HTTP response code is 409.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault 
reason is \"Operation Failed\". Fault detail is \"[Cannot attach Virtual Disk. 
The target Data Center does not contain the Virtual Disk.]\". HTTP response 
code is 409."}
[ ERROR ] Failed to execute stage 'Closing up': Failed executing 
ansible-playbook
[ INFO  ] Stage: Clean up
[ INFO  ] Cleaning temporary resources
[ INFO  ] TASK [ovirt.hosted_engine_setup : Execute just a specific set of 
steps]
[ INFO  ] ok: [localhost]


Will try again, but yesterday I got this error two times in a row.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/5OTPHSX2B5NAPC7M6JLUMNXVQWYBOLUB/


[ovirt-users] Hosted engine restore went very wrong

2019-05-07 Thread Andreas Elvers
I really feel like an idiot. I tried to move our hosted engine from our Default 
datacenter to our Ceph Datacenter.
I ran intro problems which were correctly addressed. 
see: 
https://lists.ovirt.org/archives/list/users@ovirt.org/thread/ZFCLFWRN6XR6KMHMC63O7J37D5GNPVKZ/

It was in case a race condition. I was able to bring back the engine to our 
Default Cluster. And then I tried to 
do the move again to our Ceph Datacenter.

I got the error "The target Data Center does not contain the Virtual Disk" 
twice yesterday. Because it was late,
I decided to do it in the next morning.

I did a new backup from the engine. Copied it over to the new node of the Ceph 
Datacenter and started the 
hosted-engine --deploy. But I FORGET TO SHUTDOWN the other engine! Oh man.

The deploy script errored out with:

[ ERROR ] fatal: [localhost]: FAILED! => {"censored": "the output has been 
hidden due to the fact that 'no_log: true' was specified for this result", 
"changed": false}

[ ERROR ] fatal: [localhost -> engine.infra.solutions.work]: FAILED! => 
{"changed": false, "msg": "There was a failure deploying the engine on the 
local engine VM. The system may not be provisioned accord

Then I realised something was different this time.
I shutdown and undefined the Local Engine. The node is now in a degraded state. 
Is it possible to start the deployment again on a degraded node?
I started the old engine again, but I'm not able to reach the login page.

Any Idea what to do next?
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3DMBPWWXL7XWNXO7REZTJJPUE54WBMJJ/


[ovirt-users] Re: Hosted engine restore went very wrong

2019-05-07 Thread Andreas Elvers
The second deployment still triggers "The target Data Center does not contain 
the Virtual Disk".

Trying again. This time deleting the old data in the gluster volume. Changing 
number of engine CPUs from 4 to 2. Hope this gives the race condition some 
different gravity.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/PPE6QW3XZ3DNIOBCMTFODS3T4RNPOVZH/


[ovirt-users] Re: Hosted engine restore went very wrong

2019-05-07 Thread Andreas Elvers
Still no success. The ansible script errors out again at "The target Data 
Center does not contain the Virtual Disk". trying a fourth time.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/2IIJMBLONEIWASKYNCRZPVSJRIHTVJN5/


[ovirt-users] Re: Hosted engine restore went very wrong

2019-05-07 Thread Andreas Elvers
Meanwhile I was able to remove the old wrecked engine with the 
LocalHostedEngine on Port 6900. It was stable enough to re-install that node 
with engine undeployment selected. So I'm now able to bring back the engine to 
the old Default Datacenter and NFS.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QXNFLO7J57KIARQ2FHXYHGOCZOP5FJFG/


[ovirt-users] Re: Stale hosted engine node information harmful ?

2019-05-02 Thread Andreas Elvers
I did that. There is no change. Node01,02 and 03 are still present in 
"hosted-engine --vm-status".


> In a Red Hat Solution , it is recommended  to restart ovirt-ha-agent &
> ovirt-ha-broker.
> 
> I usually set the global maintenance and wait 20s-30s . Then I just stop on 
> all nodes
> ovirt-ha-agent.service & ovirt-ha-broker.service . Once everywhere is 
> stopped, start
> the 2 services on all nodes and wait 4-5min.
> Last verify the status from each host, before removing global maintenance.
> 
> Best Regards,
> Strahil NikolovOn May 2, 2019 12:30, Andreas Elvers
> https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/675IE4RGUAROUDCYMLW4EVI5B2M6VC2M/


[ovirt-users] Re: Problem with restoring engine

2019-05-06 Thread Andreas Elvers
Yes! I can confirm that a second deployment succeeded on the second run. Great 
:-) 

I was also able to delete the Local Engine with virsh. I'd like to point out 
for reference, that this is only possible if a sasldb user is created like this:

# saslpasswd2 -a libvirt username
Password: 

After that it was possible to do this with the created credentials

# virsh destroy HostedEngineLocal 
# virsh undefine HostedEngineLocal

Glancing at the ansible output of the deployment process, the playbook seems to 
take care of old local engines anyway. 

Thanks again for your help Simone.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZRAODEEX5IZ7PQLDH7QNYONQJTYM6ATI/


[ovirt-users] Re: Stale hosted engine node information harmful ?

2019-05-02 Thread Andreas Elvers
Yeah. That did the trick. What I did was:

Enable global maintenance.
On every affected host:
# systemctl stop ovirt-ha-agent
# hosted-engine --clean-metadata --host-id=1 --force-clean (repeat for every id)
Finally start the stopped ha-agents.
Check the status: 
# hosted-engine --vm-status
Disable global maintenance..
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZDPNWLP5CYZWGFJSH3WNFBJEVIUPQSPC/


[ovirt-users] Stale hosted engine node information harmful ?

2019-05-02 Thread Andreas Elvers
I have 5 nodes (node01 to node05). Originally all those nodes were part of our 
default datacenter/cluster with a NFS storage domain for vmdisk, engine and 
iso-images. All five nodes were engine HA nodes.
Later node01, node02 and node03 were re-installed to have engine HA removed. 
Then those nodes were removed from the default cluster. Eventually node01,02 
and 03 were completely re-installed to host our new Ceph/Gluster based 
datecenter. The engine is still running on the old default Datacenter. Now I 
wish to move it over to our ceph/gluster datacenter.

when I look at the current output of "hosted-engine --vm-status" I see:

--== Host node01.infra.solutions.work (id: 1) status ==--

conf_on_shared_storage : True
Status up-to-date  : False
Hostname   : node01.infra.solutions.work
Host ID: 1
Engine status  : unknown stale-data
Score  : 0
stopped: True
Local maintenance  : False
crc32  : e437bff4
local_conf_timestamp   : 155627
Host timestamp : 155877
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=155877 (Fri Aug  3 13:09:19 2018)
host-id=1
score=0
vm_conf_refresh_time=155627 (Fri Aug  3 13:05:08 2018)
conf_on_shared_storage=True
maintenance=False
state=AgentStopped
stopped=True


--== Host node02.infra.solutions.work (id: 2) status ==--

conf_on_shared_storage : True
Status up-to-date  : False
Hostname   : node02.infra.solutions.work
Host ID: 2
Engine status  : unknown stale-data
Score  : 0
stopped: True
Local maintenance  : False
crc32  : 11185b04
local_conf_timestamp   : 154757
Host timestamp : 154856
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=154856 (Fri Aug  3 13:22:19 2018)
host-id=2
score=0
vm_conf_refresh_time=154757 (Fri Aug  3 13:20:40 2018)
conf_on_shared_storage=True
maintenance=False
state=AgentStopped
stopped=True


--== Host node03.infra.solutions.work (id: 3) status ==--

conf_on_shared_storage : True
Status up-to-date  : False
Hostname   : node03.infra.solutions.work
Host ID: 3
Engine status  : unknown stale-data
Score  : 0
stopped: False
Local maintenance  : True
crc32  : 9595bed9
local_conf_timestamp   : 14363
Host timestamp : 14362
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=14362 (Thu Aug  2 18:03:25 2018)
host-id=3
score=0
vm_conf_refresh_time=14363 (Thu Aug  2 18:03:25 2018)
conf_on_shared_storage=True
maintenance=True
state=LocalMaintenance
stopped=False


--== Host node04.infra.solutions.work (id: 4) status ==--

conf_on_shared_storage : True
Status up-to-date  : True
Hostname   : node04.infra.solutions.work
Host ID: 4
Engine status  : {"health": "good", "vm": "up", "detail": 
"Up"}
Score  : 3400
stopped: False
Local maintenance  : False
crc32  : 245854b1
local_conf_timestamp   : 317498
Host timestamp : 317498
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=317498 (Thu May  2 09:44:47 2019)
host-id=4
score=3400
vm_conf_refresh_time=317498 (Thu May  2 09:44:47 2019)
conf_on_shared_storage=True
maintenance=False
state=EngineUp
stopped=False


--== Host node05.infra.solutions.work (id: 5) status ==--

conf_on_shared_storage : True
Status up-to-date  : True
Hostname   : node05.infra.solutions.work
Host ID: 5
Engine status  : {"reason": "vm not running on this host", 
"health": "bad", "vm": "down", "detail": "unknown"}
Score  : 3400
stopped: False
Local maintenance  : False
crc32  : 0711afa0
local_conf_timestamp   : 318044
Host timestamp 

[ovirt-users] Problem with restoring engine

2019-05-05 Thread Andreas Elvers
Hello today I tried to migrate the hosted engine from our Default Datacenter 
(NFS) to our Ceph Datacenter. The deployment worked with the automatic  
"hosted-engine --deploy --restore-from-file=backup/file_name" command. Perfect.

Only thing is: I messed up with the cluster name. The name should be Luise01 
but I entered Luise1. Duh...

Now I want to bring the engine back to the Default Datacenter. Easy thing. Just 
repeat the same steps again.

1. Enable global ha maintenenace
2. Stop and disable the engine
3. create the engine backup
4 ... continue with all the steps from chapter 13.1.8 RHEV Docs 4.3 Beta.

Everything looked great. The ansible playbook was running, then asking for the 
storage domain. I entered the NFS path. It got registered, but then the ansible 
playbook  errors out with 

[ INFO  ] TASK [ovirt.hosted_engine_setup : Add VM]
[ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is "[Cannot 
attach Virtual Disk. The target Data Center does not contain the Virtual 
Disk.]". HTTP response code is 409.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault 
reason is \"Operation Failed\". Fault detail is \"[Cannot attach Virtual Disk. 
The target Data Center does not contain the Virtual Disk.]\". HTTP response 
code is 409."}
[ ERROR ] Failed to execute stage 'Closing up': Failed executing 
ansible-playbook
[ INFO  ] Stage: Clean up
[ INFO  ] Cleaning temporary resources

I see that there is a bug report on 
https://bugzilla.redhat.com/show_bug.cgi?id=1649424

Any idea how to get around this error ?

Additionally I now have a HostedEngineLocal (shut off) on that node... How do I 
remove it?
engine-cleanup ?

Have to get some sleep.

best regards. 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZFCLFWRN6XR6KMHMC63O7J37D5GNPVKZ/


[ovirt-users] Engine restore errors out on "Wait for OVF_STORE disk content"

2019-05-08 Thread Andreas Elvers
Hello,

when trying to deploy the engine using "hosted-engine --deploy 
--restore-from-file=myenginebackup" the ansible playbook errors out at 

[ INFO  ] TASK [ovirt.hosted_engine_setup : Trigger hosted engine OVF update 
and enable the serial console]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Wait until OVF update finishes]
[ INFO  ] ok: [localhost]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Parse OVF_STORE disk list]
[ INFO  ] ok: [localhost]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Check OVF_STORE volume status]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [ovirt.hosted_engine_setup : Wait for OVF_STORE disk content]
[ ERROR ] {u'_ansible_parsed': True, u'stderr_lines': [u'20+0 records in', 
u'20+0 records out', u'10240 bytes (10 kB) copied, 0.000141645 s, 72.3 MB/s', 
u'tar: ebb09b0e-2d03-40f0-8fa4-c40b18612a54.ovf: Not found in archive', u'tar: 
Exiting with failure status due to previous errors'], u'changed': True, u'end': 
u'2019-05-08 15:21:47.595195', u'_ansible_item_label': {u'image_id': 
u'65fd6c57-033c-4c95-87c1-b16c26e4bc98', u'name': u'OVF_STORE', u'id': 
u'9ff8b389-5e24-4166-9842-f1d6104b662b'}, u'stdout': u'', u'failed': True, 
u'_ansible_item_result': True, u'msg': u'non-zero return code', u'rc': 2, 
u'start': u'2019-05-08 15:21:46.906877', u'attempts': 12, u'cmd': u"vdsm-client 
Image prepare storagepoolID=597f329c-0296-03af-0369-0139 
storagedomainID=f708ced4-e339-4d02-a07f-78f1a30fc2a8 
imageID=9ff8b389-5e24-4166-9842-f1d6104b662b 
volumeID=65fd6c57-033c-4c95-87c1-b16c26e4bc98 | grep path | awk '{ print $2 }' 
| xargs -I{} sudo -u vdsm dd if={} | tar -tvf - ebb09b0e-2d03-40f0-8fa4-c40
 b18612a54.ovf", u'item': {u'image_id': 
u'65fd6c57-033c-4c95-87c1-b16c26e4bc98', u'name': u'OVF_STORE', u'id': 
u'9ff8b389-5e24-4166-9842-f1d6104b662b'}, u'delta': u'0:00:00.688318', 
u'invocation': {u'module_args': {u'warn': False, u'executable': None, 
u'_uses_shell': True, u'_raw_params': u"vdsm-client Image prepare 
storagepoolID=597f329c-0296-03af-0369-0139 
storagedomainID=f708ced4-e339-4d02-a07f-78f1a30fc2a8 
imageID=9ff8b389-5e24-4166-9842-f1d6104b662b 
volumeID=65fd6c57-033c-4c95-87c1-b16c26e4bc98 | grep path | awk '{ print $2 }' 
| xargs -I{} sudo -u vdsm dd if={} | tar -tvf - 
ebb09b0e-2d03-40f0-8fa4-c40b18612a54.ovf", u'removes': None, u'argv': None, 
u'creates': None, u'chdir': None, u'stdin': None}}, u'stdout_lines': [], 
u'stderr': u'20+0 records in\n20+0 records out\n10240 bytes (10 kB) copied, 
0.000141645 s, 72.3 MB/s\ntar: ebb09b0e-2d03-40f0-8fa4-c40b18612a54.ovf: Not 
found in archive\ntar: Exiting with failure status due to previous errors', 
u'_ansible_no_log': False}
[ ERROR ] {u'_ansible_parsed': True, u'stderr_lines': [u'20+0 records in', 
u'20+0 records out', u'10240 bytes (10 kB) copied, 0.000140541 s, 72.9 MB/s', 
u'tar: ebb09b0e-2d03-40f0-8fa4-c40b18612a54.ovf: Not found in archive', u'tar: 
Exiting with failure status due to previous errors'], u'changed': True, u'end': 
u'2019-05-08 15:24:01.387469', u'_ansible_item_label': {u'image_id': 
u'dacf9ad8-77b9-4205-8ca2-d6877627ad4a', u'name': u'OVF_STORE', u'id': 
u'8691076a-8e45-4429-a18a-5faebef866cc'}, u'stdout': u'', u'failed': True, 
u'_ansible_item_result': True, u'msg': u'non-zero return code', u'rc': 2, 
u'start': u'2019-05-08 15:24:00.660309', u'attempts': 12, u'cmd': u"vdsm-client 
Image prepare storagepoolID=597f329c-0296-03af-0369-0139 
storagedomainID=f708ced4-e339-4d02-a07f-78f1a30fc2a8 
imageID=8691076a-8e45-4429-a18a-5faebef866cc 
volumeID=dacf9ad8-77b9-4205-8ca2-d6877627ad4a | grep path | awk '{ print $2 }' 
| xargs -I{} sudo -u vdsm dd if={} | tar -tvf - ebb09b0e-2d03-40f0-8fa4-c40
 b18612a54.ovf", u'item': {u'image_id': 
u'dacf9ad8-77b9-4205-8ca2-d6877627ad4a', u'name': u'OVF_STORE', u'id': 
u'8691076a-8e45-4429-a18a-5faebef866cc'}, u'delta': u'0:00:00.727160', 
u'invocation': {u'module_args': {u'warn': False, u'executable': None, 
u'_uses_shell': True, u'_raw_params': u"vdsm-client Image prepare 
storagepoolID=597f329c-0296-03af-0369-0139 
storagedomainID=f708ced4-e339-4d02-a07f-78f1a30fc2a8 
imageID=8691076a-8e45-4429-a18a-5faebef866cc 
volumeID=dacf9ad8-77b9-4205-8ca2-d6877627ad4a | grep path | awk '{ print $2 }' 
| xargs -I{} sudo -u vdsm dd if={} | tar -tvf - 
ebb09b0e-2d03-40f0-8fa4-c40b18612a54.ovf", u'removes': None, u'argv': None, 
u'creates': None, u'chdir': None, u'stdin': None}}, u'stdout_lines': [], 
u'stderr': u'20+0 records in\n20+0 records out\n10240 bytes (10 kB) copied, 
0.000140541 s, 72.9 MB/s\ntar: ebb09b0e-2d03-40f0-8fa4-c40b18612a54.ovf: Not 
found in archive\ntar: Exiting with failure status due to previous errors', 
u'_ansible_no_log': False}
[ ERROR ] Failed to execute stage 'Closing up': Failed executing 
ansible-playbook

I tried twice. Same result. Should I retry?

Is it safe to use the local hosted engine for starting stopping vms? I'm kind 
of headless for some days :-)

Best regards.

[ovirt-users] Re: Storage domain format version info

2019-07-09 Thread Andreas Elvers
I'd like to add a question on storage domains.

When will storage domains become obsolete? Like kind of a road map.

I'm asking because we are operating a OpenStack Cinder V1 Domain to access Ceph 
RBD images. I know this is only a former tech preview, but this preview is 
doing solid work. So I'd like to know, when this will go away. 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/OQ3CODAESWZNABAEZADFG2XSVZZV66VF/


[ovirt-users] Re: Upgrading from 4.2.8 to 4.3.3 broke Node NG GlusterFS

2019-04-24 Thread Andreas Elvers
After rebooting the node that was not able to mount the gluster volume things 
improved eventually. SPM went away and restarted for the Datacenter and 
suddenly node03 was able to mount the gluster volume. In between I was down to 
1/3 active Bricks which results in read only glusterfs. I was lucky to have the 
Engine still on NFS. But anyway... 

Thanks for your thoughts.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/5NKHRCJWEZGXSBKRMR447RCX6GWAAZV6/


[ovirt-users] Upgrading from 4.2.8 to 4.3.3 broke Node NG GlusterFS

2019-04-24 Thread Andreas Elvers
Hi,

I am currently upgrading my oVirt setup from 4.2.8 to 4.3.3.1.

The setup consists of:

Datacenter/Cluster Default: [fully upgraded to 4.3.3.1]
   2 nodes (node04,node05)- NFS storage domain with self hosted engine 

Datacenter Luise:
   Cluster1: 3 nodes (node01,node02,node03) - Node NG with GlusterFS - Ceph 
Cinder storage domain
  [Node1 and Node3 are upgraded to 4.3.3.1, Node2 is on 4.2.8]
   Cluster2: 1 node (node06)  - only Ceph Cinder storage domain [fully upgraded 
to 4.3.3.1]


Problems started when upgrading Luise/Cluster1 with GlusterFS:
(I always waited for GlusterFS to be fully synced before proceeding to the next 
step)

- Upgrade node01 to 4.3.3 -> OK
- Upgrade node03 to 4.3.3.1 -> OK
- Upgrade node01 to 4.3.3.1 -> GlusterFS became unstable.


I now get the error message:

VDSM node03.infra.solutions.work command ConnectStoragePoolVDS failed: Cannot 
find master domain: u'spUUID=f3218bf7-6158-4b2b-b272-51cdc3280376, 
msdUUID=02a32017-cbe6-4407-b825-4e558b784157'

And on node03 there is a problem with Gluster:

node03#: ls -l 
/rhev/data-center/mnt/glusterSD/node01.infra.solutions.work:_vmstore
ls: cannot access 
/rhev/data-center/mnt/glusterSD/node01.infra.solutions.work:_vmstore: Transport 
endpoint is not connected

The directory is available on node01 and node02.

The engine is reporting the brick on node03 as down. Node03 and Node06 are 
shown as NonOperational, because they are not able to access the gluster 
storage domain. 

A “gluster peer status” on node1, node2, and node3 shows all peers connected.

“gluster volume heal vmstore info” shows for all nodes:


gluster volume heal vmstore info
Brick node01.infra.solutions.work:/gluster_bricks/vmstore/vmstore
Status: Transport endpoint is not connected
Number of entries: -

Brick node02.infra.solutions.work:/gluster_bricks/vmstore/vmstore



/02a32017-cbe6-4407-b825-4e558b784157/dom_md/ids
/.shard/40948f85-2212-47f9-bd5e-102a8dd632b8.66

/.shard/40948f85-2212-47f9-bd5e-102a8dd632b8.60
/02a32017-cbe6-4407-b825-4e558b784157/images/a3a10398-9698-4b73-84d9-9735448e3534/6161e310-4ad6-42d9-8117-5a89c5b2b4b6


/.shard/40948f85-2212-47f9-bd5e-102a8dd632b8.96


/.shard/d66880de-3fa1-4362-8c43-574a173c5f7d.133


/.shard/40948f85-2212-47f9-bd5e-102a8dd632b8.38
/.shard/40948f85-2212-47f9-bd5e-102a8dd632b8.67
/__DIRECT_IO_TEST__


/02a32017-cbe6-4407-b825-4e558b784157/images/493188b2-c137-4440-99ee-43a753842a7d/9aa2d139-e3bd-406b-8fe0-b189123eaa73

/.shard/40948f85-2212-47f9-bd5e-102a8dd632b8.64
/.shard/d66880de-3fa1-4362-8c43-574a173c5f7d.132



/.shard/40948f85-2212-47f9-bd5e-102a8dd632b8.44
/.shard/40948f85-2212-47f9-bd5e-102a8dd632b8.9
/.shard/40948f85-2212-47f9-bd5e-102a8dd632b8.69

/02a32017-cbe6-4407-b825-4e558b784157/images/12e647fb-20aa-4957-b659-05fa75a9215e/f7e4b2a3-ab84-4eb5-a4e7-7208ddad8156





/.shard/40948f85-2212-47f9-bd5e-102a8dd632b8.35
/.shard/40948f85-2212-47f9-bd5e-102a8dd632b8.32


/.shard/40948f85-2212-47f9-bd5e-102a8dd632b8.39


/.shard/40948f85-2212-47f9-bd5e-102a8dd632b8.34
/.shard/40948f85-2212-47f9-bd5e-102a8dd632b8.68
Status: Connected
Number of entries: 47

Brick node03.infra.solutions.work:/gluster_bricks/vmstore/vmstore
/02a32017-cbe6-4407-b825-4e558b784157/images/12e647fb-20aa-4957-b659-05fa75a9215e/f7e4b2a3-ab84-4eb5-a4e7-7208ddad8156











/.shard/d66880de-3fa1-4362-8c43-574a173c5f7d.133






/02a32017-cbe6-4407-b825-4e558b784157/images/493188b2-c137-4440-99ee-43a753842a7d/9aa2d139-e3bd-406b-8fe0-b189123eaa73






/.shard/40948f85-2212-47f9-bd5e-102a8dd632b8.44







/02a32017-cbe6-4407-b825-4e558b784157/dom_md/ids



/02a32017-cbe6-4407-b825-4e558b784157/images/a3a10398-9698-4b73-84d9-9735448e3534/6161e310-4ad6-42d9-8117-5a89c5b2b4b6



/.shard/d66880de-3fa1-4362-8c43-574a173c5f7d.132



/__DIRECT_IO_TEST__
Status: Connected
Number of entries: 47

On Node03 there are several self healing processes, that seem to be doing 
nothing.

Oh well.. What now?

Best regards,
- Andreas
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/R5GS6AQXTEQRMUQNMEBDC72YG3A5JFF6/


[ovirt-users] Re: Upgrading from 4.2.8 to 4.3.3 broke Node NG GlusterFS

2019-04-24 Thread Andreas Elvers
Restarting improved things a little bit. Still bricks on node03 are shown as 
down, but "gluster volume status" is looking better.

Saiph:~ andreas$ ssh node01 gluster volume status vmstore
Status of volume: vmstore
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick node01.infra.solutions.work:/gluster_
bricks/vmstore/vmstore  49157 0  Y   24543
Brick node02.infra.solutions.work:/gluster_
bricks/vmstore/vmstore  49154 0  Y   23795
Brick node03.infra.solutions.work:/gluster_
bricks/vmstore/vmstore  49157 0  Y   1617
Self-heal Daemon on localhost   N/A   N/AY   32121
Self-heal Daemon on node03.infra.solutions.
workN/A   N/AY   25798
Self-heal Daemon on node02.infra.solutions.
workN/A   N/AY   30879

Task Status of Volume vmstore
--
There are no active volume tasks

Saiph:~ andreas$ ssh node02 gluster volume status vmstore
Status of volume: vmstore
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick node01.infra.solutions.work:/gluster_
bricks/vmstore/vmstore  49157 0  Y   24543
Brick node02.infra.solutions.work:/gluster_
bricks/vmstore/vmstore  49154 0  Y   23795
Brick node03.infra.solutions.work:/gluster_
bricks/vmstore/vmstore  49157 0  Y   1617
Self-heal Daemon on localhost   N/A   N/AY   30879
Self-heal Daemon on node03.infra.solutions.
workN/A   N/AY   25798
Self-heal Daemon on node01.infra.solutions.
workN/A   N/AY   32121

Task Status of Volume vmstore
--
There are no active volume tasks

Saiph:~ andreas$ ssh node03 gluster volume status vmstore
Status of volume: vmstore
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick node01.infra.solutions.work:/gluster_
bricks/vmstore/vmstore  49157 0  Y   24543
Brick node02.infra.solutions.work:/gluster_
bricks/vmstore/vmstore  49154 0  Y   23795
Brick node03.infra.solutions.work:/gluster_
bricks/vmstore/vmstore  49157 0  Y   1617
Self-heal Daemon on localhost   N/A   N/AY   25798
Self-heal Daemon on node01.infra.solutions.
workN/A   N/AY   32121
Self-heal Daemon on node02.infra.solutions.
workN/A   N/AY   30879

Task Status of Volume vmstore
--
There are no active volume tasks
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/AAGJ42GF267NOFEQXNJRUQJD7C5UCOM5/


[ovirt-users] Re: Upgrading from 4.2.8 to 4.3.3 broke Node NG GlusterFS

2019-04-24 Thread Andreas Elvers
"systemctl restart glusterd" on node03 did not help. Still getting:

node03#: ls -l 
/rhev/data-center/mnt/glusterSD/node01.infra.solutions.work:_vmstore 
ls: cannot access 
/rhev/data-center/mnt/glusterSD/node01.infra.solutions.work:_vmstore: Transport 
endpoint is not connected

Engine still shows bricks on node03 as down.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/2JMM4UBZ54TNHCFUFYDX2OOVCKEMXFBX/


[ovirt-users] Re: Upgrading from 4.2.8 to 4.3.3 broke Node NG GlusterFS

2019-04-24 Thread Andreas Elvers
The file handle is stale so find will display:

"find: '/rhev/data-center/mnt/glusterSD/node01.infra.solutions.work:_vmstore': 
Transport endpoint is not connected"

"stat /rhev/data-center/mnt/glusterSD/node01.infra.solutions.work:_vmstore" 
will output
stat: cannot stat 
'/rhev/data-center/mnt/glusterSD/node01.infra.solutions.work:_vmstore': 
Transport endpoint is not connected

All Nodes are peering with the other nodes:
- 
Saiph:~ andreas$ ssh node01 gluster peer status
Number of Peers: 2

Hostname: node02.infra.solutions.work
Uuid: 87fab40a-2395-41ce-857d-0b846e078cdb
State: Peer in Cluster (Connected)

Hostname: node03.infra.solutions.work
Uuid: 49025f81-e7c1-4760-be03-f36e0f403d26
State: Peer in Cluster (Connected)

Saiph:~ andreas$ ssh node02 gluster peer status
Number of Peers: 2

Hostname: node03.infra.solutions.work
Uuid: 49025f81-e7c1-4760-be03-f36e0f403d26
State: Peer in Cluster (Disconnected)

Hostname: node01.infra.solutions.work
Uuid: f25e6bff-e5e2-465f-a33e-9148bef94633
State: Peer in Cluster (Connected)

ssh node03 gluster peer status
Number of Peers: 2

Hostname: node02.infra.solutions.work
Uuid: 87fab40a-2395-41ce-857d-0b846e078cdb
State: Peer in Cluster (Connected)

Hostname: node01.infra.solutions.work
Uuid: f25e6bff-e5e2-465f-a33e-9148bef94633
State: Peer in Cluster (Connected)
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/DI6AWTLIQIPWNK2M7PBABQ4TAPB4J3S3/


[ovirt-users] Upgrading librbd1 and librados2 for oVirt 4.2.x

2018-08-24 Thread andreas . elvers+ovirtforum
Hallo Group !

I am in need of upgrading librbd1 and librados2 for a oVirt 4.2.x cluster.
The cluster was installed via node ng.

Taking the  repository for Ceph Mimic or Luminous will end in a dependency
problem because liburcu-cds.so.1 is already installed as a more recent version
provided by the oVirt repository.

A hint on how to get a more recent versions of the above libraries would be very
much appreciated.

Thanks,
- Andreas
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/YA3WNB6O7B4ZRI3PIG46IZCIMDFXSYZL/


[ovirt-users] Re: Upgrading librbd1 and librados2 for oVirt 4.2.x

2018-08-24 Thread andreas . elvers+ovirtforum
I found a way.

[root@node02 yum.repos.d]# cat luminious.repo

[ovirt-4.2-centos-ceph-luminous]
enabled=1
name = CentOS-7 - ceph luminous
baseurl = http://mirror.centos.org/centos/7/storage/$basearch/ceph-luminous/
gpgcheck = 1
enabled = 1
gpgkey = 
https://raw.githubusercontent.com/CentOS-Storage-SIG/centos-release-storage-common/master/RPM-GPG-KEY-CentOS-SIG-Storage
includepkgs = librados2 librbd1 lttng-ust
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/BW5KVW5UCJDWM4ZSLONZQ547OHYI33Z6/


[ovirt-users] Re: 4.2 / 4.3 : Moving the hosted-engine to another storage

2019-04-02 Thread andreas . elvers+ovirtforum
Thanks for your answer.

> Yes, now you can do it via backup and restore:
> take a backup of the engine with engine-backup and restore it on a new
> hosted-engine VM on a new storage domain with:
> hosted-engine --deploy --restore-from-file=mybackup.tar.gz

That is great news. Just to clear things up completely:

- This backup command is sufficient ? No further switches ?
  "engine-backup --mode=backup --file=[EngineBackupFile] --log=[LogFILE]"

- Do I need to place the node to which I want to deploy engine into maintenance
  before doing the backup? This is referred in the stated documentation as:

  "If a hosted-engine host is carrying a virtual load at the time of backup 
[...] 
   then a host [...] cannot be used to deploy a restored self-hosted engine."

   This is still true? If yes: all other precautions apply from that 
documentation?

- before creating the new engine on the new storage, the old engine need to be 
  un-deployed on all engine-HA hosts. I am unable to find information about this
  issue. Just un-deploy via web ui? 

- I want to deploy on a gluster volume that has been automatically created by 
the
  Node setup. The file /etc/gluster/glusterd.vol does not carry 
  "option rpc-auth-allow-insecure on" which is mentioned in the documentation.
  Do I need to follow the instructions for gluster, or are settings already 
sufficiently
  set by the automatic gluster deployment done by oVirt Node setup? I already
  have some VM images running on that gluster storage anyway.

Thanks for your help.
- Andreas
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/TQNWNUHROMQLO2RSPRCMW3E5J7Y4USJD/


[ovirt-users] 4.2 / 4.3 : Moving the hosted-engine to another storage

2019-04-01 Thread andreas . elvers+ovirtforum
Hi friends of oVirt,

Roughly 3 years ago a user asked about the options he had to move the hosted 
engine to some other storage.

The answer by Simone Tiraboschi was that it would largely not be possible, 
because of references in the database to the node the engine was hosted. This 
information would prevent a successful move of the engine even with 
backup/restore.

The situation seem to have improved, but I'm not sure. So I ask.

We have to move our engine away from our older Cluster with NFS Storage 
backends (engine, volumes, iso-images). 

The engine should be restored on our new cluster that has a gluster volume 
available for the engine. Additionally this 3-node cluster is running Guests 
from a Cinder/Ceph storage Domain.

I want to restore the engine on a different cluster to a different storage 
domain. 

Reading the documentation at 
https://www.ovirt.org/documentation/self-hosted/chap-Backing_up_and_Restoring_an_EL-Based_Self-Hosted_Environment.html
 I am wondering whether oVirt Nodes (formerly Node-NG) are capable of restoring 
an engine at all. Do I need EL-based Nodes? We are currently running on oVirt 
Nodes.

- Andreas
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/Y7EPAGSARSUFGYRABDX7M7BNAVBRAFHS/