Re: [ovirt-users] ovirt-engine ui not showing up/inaccessible

2016-04-15 Thread Alexander Wels
On Friday, April 15, 2016 10:12:44 AM Gene Fontanilla wrote:
> Hi,
> 
> our office had power outage for more than an hour and continues
> fluctuations, so i decided to pull the plug after 2nd fluctuation and wait
> for the power to become stable.
> 
> After the power became stable, I decided to turn the ovirt-engine up 1st,
> after booting up. I tried accessing my engine but the webui is not
> accessible, then i looked at the /var/log/ovirt-engine/engine.log
> 
> and saw this error:
> 
> 2016-04-15 06:34:56,265 ERROR [org.ovirt.engine.core.bll.Backend] (MSC
> service thread 1-28) Error in getting DB connection. The database is
> inaccessible. Original exception is: DataAccessResourceFailureException:
> Error retreiving database metadata; nested exception is
> org.springframework.jdbc.support.MetaDataAccessException: Could not get
> Connection for extracting meta data; nested exception is
> org.springframework.jdbc.CannotGetJdbcConnectionException: Could not get
> JDBC Connection; nested exception is java.sql.SQLException:
> javax.resource.ResourceException: IJ000453: Unable to get managed
> connection for java:/ENGINEDataSource
> 
> Can anyone help me out? I need in that environment up for the final stage
> of my testing.
> 
> regards,
> Gene

Sounds like postgresql isn't starting for some reason. Maybe check the status 
of that and manually start it if it is not up.

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] ovirt-engine ui not showing up/inaccessible

2016-04-15 Thread Gene Fontanilla
Hi,

our office had power outage for more than an hour and continues
fluctuations, so i decided to pull the plug after 2nd fluctuation and wait
for the power to become stable.

After the power became stable, I decided to turn the ovirt-engine up 1st,
after booting up. I tried accessing my engine but the webui is not
accessible, then i looked at the /var/log/ovirt-engine/engine.log

and saw this error:

2016-04-15 06:34:56,265 ERROR [org.ovirt.engine.core.bll.Backend] (MSC
service thread 1-28) Error in getting DB connection. The database is
inaccessible. Original exception is: DataAccessResourceFailureException:
Error retreiving database metadata; nested exception is
org.springframework.jdbc.support.MetaDataAccessException: Could not get
Connection for extracting meta data; nested exception is
org.springframework.jdbc.CannotGetJdbcConnectionException: Could not get
JDBC Connection; nested exception is java.sql.SQLException:
javax.resource.ResourceException: IJ000453: Unable to get managed
connection for java:/ENGINEDataSource

Can anyone help me out? I need in that environment up for the final stage
of my testing.

regards,
Gene
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] nfs storage permission problem

2016-04-15 Thread Bill James

just to close off this issue.
I found problem.
My /etc/exports file had a space in it where it shouldn't have (between 
* and options).
This seemed to acceptable with older versions of centos7.2 but on my 
newer host with latest centos7.2 (3.10.0-327.13.1.el7.x86_64) it only 
mounted read-only. maybe default changed from rw to ro?


I fixed exports file and now both versions of centos work.

/mount_point *(rw,sync,no_subtree_check,all_squash,anonuid=36,anongid=36)



On 4/13/16 4:39 PM, Brett I. Holcomb wrote:

On Wed, 2016-04-13 at 15:52 -0700, Bill James wrote:
[vdsm@ovirt4 test /]$ touch 
/rhev/data-center/mnt/ovirt3-ks.test.j2noc.com:_ovirt-store_nfs/test
touch: cannot touch 
‘/rhev/data-center/mnt/ovirt3-ks.test.j2noc.com:_ovirt-store_nfs/test’: 
Read-only file system


Hmm, read-only.  :-(

ovirt3-ks.test.j2noc.com:/ovirt-store/nfs on 
/rhev/data-center/mnt/ovirt3-ks.test.j2noc.com:_ovirt-store_nfs type 
nfs4 
(*rw*,relatime,vers=4.0,rsize=1048576,wsize=1048576,namlen=255,soft,nosharecache,proto=tcp,port=0,timeo=600,retrans=6,sec=sys,clientaddr=10.100.108.96,local_lock=none,addr=10.100.108.33)


now to figure out why


[root@ovirt4 test ~]# ls -la /rhev/data-center/mnt/
total 8
drwxr-xr-x 4 vdsm kvm  110 Apr 13 15:30 .
drwxr-xr-x 3 vdsm kvm   16 Apr 13 08:06 ..
drwxr-xr-x 3 vdsm kvm 4096 Mar 11 15:19 
netappqa3:_vol_cloud__images_ovirt__QA__export
drwxr-xr-x 3 vdsm kvm 4096 Mar 11 15:17 
netappqa3:_vol_cloud__images_ovirt__QA__ISOs




export and ISO domain mount fine too. (and rw)



ovirt-engine-3.6.4.1-1.el7.centos.noarch


On 04/13/2016 03:21 PM, Brett I. Holcomb wrote:

On Wed, 2016-04-13 at 15:09 -0700, Bill James wrote:

I have a cluster working fine with 2 nodes.
I'm trying to add a third and it is complaining:


StorageServerAccessPermissionError: Permission settings on the
specified
path do not allow access to the storage. Verify permission settings
on
the specified storage path.: 'path =
/rhev/data-center/mnt/ovirt3-ks.test.j2noc.com:_ovirt-store_nfs'


if I try the commands manually as vdsm they work fine and the volume
mounts.


[vdsm@ovirt4 test /]$ mkdir -p
/rhev/data-center/mnt/ovirt3-ks.test.j2noc.com:_ovirt-store_nfs
[vdsm@ovirt4 test /]$ sudo -n /usr/bin/mount -t nfs -o
soft,nosharecache,timeo=600,retrans=6
ovirt3-ks.test.j2noc.com:/ovirt-store/nfs
/rhev/data-center/mnt/ovirt3-ks.test.j2noc.com:_ovirt-store_nfs
[vdsm@ovirt4 test /]$ df -h
/rhev/data-center/mnt/ovirt3-ks.test.j2noc.com:_ovirt-store_nfs
Filesystem Size  Used Avail Use%
Mounted on
ovirt3-ks.test.j2noc.com:/ovirt-store/nfs  1.1T  305G  759G  29%
/rhev/data-center/mnt/ovirt3-ks.test.j2noc.com:_ovirt-store_nfs


After manually mounting the NFS volumes and activating the node it
still
fails.


2016-04-13 14:55:16,559 WARN
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector
]
(DefaultQuartzScheduler_Worker-61) [64ceea1d] Correlation ID:
64ceea1d,
Job ID: a47b74c7-2ae0-43f9-9bdf-e50963a28895, Call Stack: null,
Custom
Event ID: -1, Message: Host ovirt4.test.j2noc.com cannot access the
Storage Domain(s)  attached to the Data Center Default.
Setting
Host state to Non-Operational.


Not sure what "UNKNOWN" storage is, unless its one I deleted earlier
that somehow isn't really removed.

Also tried "reinstall" on node. same issue.

Attached are engine and vdsm logs.


Thanks.

___
Users mailing list
Users@ovirt.org 
http://lists.ovirt.org/mailman/listinfo/users

Try adding anonuid=36,anongid=36 to the mount and make sure 36:36 is
the owner group on the mount point.  I found this,http://www.ovirt.org
/documentation/how-to/troubleshooting/troubleshooting-nfs-storage-
issues/, helpful.

___
Users mailing list
Users@ovirt.org 
http://lists.ovirt.org/mailman/listinfo/users


___
Users mailing list
Users@ovirt.org 
http://lists.ovirt.org/mailman/listinfo/users 



Try adding the anonuid=36,anongid=36 to the NFS mount options.




___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] several questions about serial console

2016-04-15 Thread Nathanaël Blanchet



Le 15/04/2016 17:27, Nathanaël Blanchet a écrit :

Hi all,

About serial console:

  * how to get out of a selectionned vm where we are on the login
prompt (why not backing up to the vm menu) rather than killing the
ssh process or closing the terminal? Usual "^] " doesn't work there.
  * according to

https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Virtualization/3.6/html-single/Virtual_Machine_Management_Guide/index.html

# ssh -t -p  ovirt-vmconsole@/MANAGER_IP/  --vm-name/vm1 /

should allow to connect directly to a vm on its serial port, 
and it is very useful when there are a big number of vm. In reality, 
we get a SSH issue : "unknown option -- -"
An other question is : Why the vm order is not alphabetic? it could 
simplify the search when too many vms are displayed. And a simple "sort 
-k2" command should do the stuff...


  * if we want to add 5 users with UserVmManager role on 150 vms and I
can't use group for this stuff, this means I need to do this with
an ovirt-shell script like :
# for i in $(cat /tmp/ids.ovirt); do for j in $(cat
/tmp/list_all);do ovirt-shell -E "add permission --parent-vm-name
$j --user-id $i --role-name UserVmManager"; done; done
and 5*150 API connections only because I can't add several user
ids on the same "add permission" line ? It's doable, but not not
very convinient and very long if I have many more users to add.
  * Why can't we add permission by user-name and not user-id ?

Thnak you for your help.


--
Nathanaël Blanchet

Supervision réseau
Pôle Infrastrutures Informatiques
227 avenue Professeur-Jean-Louis-Viala
34193 MONTPELLIER CEDEX 5   
Tél. 33 (0)4 67 54 84 55
Fax  33 (0)4 67 54 84 14
blanc...@abes.fr  



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


--
Nathanaël Blanchet

Supervision réseau
Pôle Infrastrutures Informatiques
227 avenue Professeur-Jean-Louis-Viala
34193 MONTPELLIER CEDEX 5   
Tél. 33 (0)4 67 54 84 55
Fax  33 (0)4 67 54 84 14
blanc...@abes.fr

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] several questions about serial console

2016-04-15 Thread Nathanaël Blanchet

Hi all,

About serial console:

 * how to get out of a selectionned vm where we are on the login prompt
   (why not backing up to the vm menu) rather than killing the ssh
   process or closing the terminal? Usual "^] " doesn't work there.
 * according to
   
https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Virtualization/3.6/html-single/Virtual_Machine_Management_Guide/index.html


   # ssh -t -p  ovirt-vmconsole@/MANAGER_IP/  --vm-name/vm1 /

should allow to connect directly to a vm on its serial port, 
and it is very useful when there are a big number of vm. In reality, we 
get a SSH issue : "unknown option -- -"


 * if we want to add 5 users with UserVmManager role on 150 vms and I
   can't use group for this stuff, this means I need to do this with an
   ovirt-shell script like :
   # for i in $(cat /tmp/ids.ovirt); do for j in $(cat
   /tmp/list_all);do ovirt-shell -E "add permission --parent-vm-name $j
   --user-id $i --role-name UserVmManager"; done; done
   and 5*150 API connections only because I can't add several user ids
   on the same "add permission" line ? It's doable, but not not very
   convinient and very long if I have many more users to add.
 * Why can't we add permission by user-name and not user-id ?

Thnak you for your help.


--
Nathanaël Blanchet

Supervision réseau
Pôle Infrastrutures Informatiques
227 avenue Professeur-Jean-Louis-Viala
34193 MONTPELLIER CEDEX 5   
Tél. 33 (0)4 67 54 84 55
Fax  33 (0)4 67 54 84 14
blanc...@abes.fr

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted engine on gluster problem

2016-04-15 Thread Luiz Claudio Prazeres Goncalves
I'm not planning to move to ovirt 4 until it gets stable, so would be great
to backport to 3.6 or ,ideally, gets developed on the next release of 3.6
branch. Considering the urgency (its a single point of failure) x
complexity wouldn't be hard to make the proposed fix.

I'm using today a production environment on top of gluster replica 3 and
this is the only SPF I have.

Thanks
Luiz

Em sex, 15 de abr de 2016 03:05, Sandro Bonazzola 
escreveu:

> On Thu, Apr 14, 2016 at 7:35 PM, Nir Soffer  wrote:
>
>> On Wed, Apr 13, 2016 at 4:34 PM, Luiz Claudio Prazeres Goncalves
>>  wrote:
>> > Nir, here is the problem:
>> > https://bugzilla.redhat.com/show_bug.cgi?id=1298693
>> >
>> > When you do a hosted-engine --deploy and pick "glusterfs" you don't
>> have a
>> > way to define the mount options, therefore, the use of the
>> > "backupvol-server", however when you create a storage domain from the
>> UI you
>> > can, like the attached screen shot.
>> >
>> >
>> > In the hosted-engine --deploy, I would expect a flow which includes not
>> only
>> > the "gluster" entrypoint, but also the gluster mount options which is
>> > missing today. This option would be optional, but would remove the
>> single
>> > point of failure described on the Bug 1298693.
>> >
>> > for example:
>> >
>> > Existing entry point on the "hosted-engine --deploy" flow
>> > gluster1.xyz.com:/engine
>>
>> I agree, this feature must be supported.
>>
>
> It will, and it's currently targeted to 4.0.
>
>
>
>>
>> > Missing option on the "hosted-engine --deploy" flow :
>> > backupvolfile-server=gluster2.xyz.com
>> ,fetch-attempts=3,log-level=WARNING,log-file=/var/log/glusterfs/gluster_engine_domain.log
>> >
>> > Sandro, it seems to me a simple solution which can be easily fixed.
>> >
>> > What do you think?
>> >
>> > Regards
>> > -Luiz
>> >
>> >
>> >
>> > 2016-04-13 4:15 GMT-03:00 Sandro Bonazzola :
>> >>
>> >>
>> >>
>> >> On Tue, Apr 12, 2016 at 6:47 PM, Nir Soffer 
>> wrote:
>> >>>
>> >>> On Tue, Apr 12, 2016 at 3:05 PM, Luiz Claudio Prazeres Goncalves
>> >>>  wrote:
>> >>> > Hi Sandro, I've been using gluster with 3 external hosts for a while
>> >>> > and
>> >>> > things are working pretty well, however this single point of failure
>> >>> > looks
>> >>> > like a simple feature to implement,but critical to anyone who wants
>> to
>> >>> > use
>> >>> > gluster on production  . This is not hyperconvergency which has
>> other
>> >>> > issues/implications. So , why not have this feature out on 3.6
>> branch?
>> >>> > It
>> >>> > looks like just let vdsm use the 'backupvol-server' option when
>> >>> > mounting the
>> >>> > engine domain and make the property tests.
>> >>>
>> >>> Can you explain what is the problem, and what is the suggested
>> solution?
>> >>>
>> >>> Engine and vdsm already support the backupvol-server option - you can
>> >>> define this option in the storage domain options when you create a
>> >>> gluster
>> >>> storage domain. With this option vdsm should be able to connect to
>> >>> gluster
>> >>> storage domain even if a brick is down.
>> >>>
>> >>> If you don't have this option in engine , you probably cannot add it
>> with
>> >>> hosted
>> >>> engine setup, since for editing it you must put the storage domain in
>> >>> maintenance
>> >>> and if you do this the engine vm will be killed :-) This is is one of
>> >>> the issues with
>> >>> engine managing the storage domain it runs on.
>> >>>
>> >>> I think the best way to avoid this issue, is to add a DNS entry
>> >>> providing the addresses
>> >>> of all the gluster bricks, and use this address for the gluster
>> >>> storage domain. This way
>> >>> the glusterfs mount helper can mount the domain even if one of the
>> >>> gluster bricks
>> >>> are down.
>> >>>
>> >>> Again, we will need some magic from the hosted engine developers to
>> >>> modify the
>> >>> address of the hosted engine gluster domain on existing system.
>> >>
>> >>
>> >> Magic won't happen without a bz :-) please open one describing what's
>> >> requested.
>> >>
>> >>
>> >>>
>> >>>
>> >>> Nir
>> >>>
>> >>> >
>> >>> > Could you add this feature to the next release of 3.6 branch?
>> >>> >
>> >>> > Thanks
>> >>> > Luiz
>> >>> >
>> >>> > Em ter, 12 de abr de 2016 05:03, Sandro Bonazzola <
>> sbona...@redhat.com>
>> >>> > escreveu:
>> >>> >>
>> >>> >> On Mon, Apr 11, 2016 at 11:44 PM, Bond, Darryl <
>> db...@nrggos.com.au>
>> >>> >> wrote:
>> >>> >>>
>> >>> >>> My setup is hyperconverged. I have placed my test results in
>> >>> >>> https://bugzilla.redhat.com/show_bug.cgi?id=1298693
>> >>> >>>
>> >>> >>
>> >>> >> Ok, so you're aware about the limitation of the single point of
>> >>> >> failure.
>> >>> >> If you drop the host referenced in hosted engine configuration for
>> the
>> >>> >> initial setup it won't be able to connect to shared storage even if
>> >>> >> the
>> >>> >> other hosts in the cluster are up 

Re: [ovirt-users] serial console and permission

2016-04-15 Thread Michal Skrivanek

> On 14 Apr 2016, at 23:27, Nathanaël Blanchet  wrote:
> 
> Hi all,
> 
> I've successfully set up the serial console feature for all my vms.
> But the only way I found to make it work is to add each user as a 
> UserVmManager role, whereas they have the SuperUser role at the datacenter 
> level.I know there is an open bug on it for this.

Hi,
yes, https://bugzilla.redhat.com/show_bug.cgi?id=1320343

> A second bug is that adding a group with UserVmManager as permission on a vm 
> (instead of a simple user) doesn't allow to get the serial console.

That’s by design. The console is giving you a root-like access hence regular 
user permissions are not enough

Thanks,
michal

> Thank you for your help
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
> 
> 

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hosted-engine stuck "failed liveliness check" "detail": "up"

2016-04-15 Thread Paul Groeneweg | Pazion
Thanks for the help!

I managed to fix it  :-)

I made a device from the hosted engine file with losetup and for the LVM
again.
>From there I was able to fsck my partition. And answered a lot of y to the
list of inconsistences.
I removed the loop back devices.
I turned off maintenance mode and hosted engine started :-)
Few hours later , restore of the pgsql db and scratching my head, very
happy it is running again without any VMS being offline meanwhile :)

I will defintely save the way to start hosted-engine with cd iso, as this
seems less error sensitive and easier to do.



Thanks!


Op vr 15 apr. 2016 om 11:35 schreef Simone Tiraboschi :

> On Fri, Apr 15, 2016 at 11:02 AM, Paul Groeneweg | Pazion
>  wrote:
> > Thanks!
> >
> > I managed to get the console through:
> >
> > hosted-engine --add-console-password
> > /bin/remote-viewer vnc://localhost:5900
> >
> > Turns out, there seems to be some corruption on the partition:
> > http://screencast.com/t/6iR0U3QuI
> >
> > Is there a way to boot from CD, so I can start rescue mode?
>
> Yes,
> you have to set global maintenance mode to avoid VM start attempts on
> other hosts.
>
> You have to poweroff the engine VM with:
>  hosted-engine --vm-shutdown/--vm-poweroff (probably the second one in
> your case)
> Then you have to create a copy of
> /var/run/ovirt-hosted-engine-ha/vm.conf and you have to manually edit
> it in order to attach the ISO image and change the boot order.
> At that point you can start the engine VM with your custom configuration
> with:
>  hosted-engine --vm-start --vm-conf=my_custom_vm.conf
>
> Good luck!
>
> > Op vr 15 apr. 2016 om 10:58 schreef Martin Sivak :
> >>
> >> Hi,
> >>
> >> you can access the console using vnc or use virsh to get access to the
> >> serial console.
> >>
> >> Check the following commands on the host where the VM is currently
> >> running:
> >>
> >> virsh -r list
> >> virsh -r console HostedEngine
> >> virsh -r vncdisplay HostedEngine
> >>
> >> Those should give you enough pointers to connect to the VM.
> >>
> >> Regards
> >>
> >> Martin Sivak
> >>
> >>
> >>
> >> On Fri, Apr 15, 2016 at 10:14 AM, Paul Groeneweg | Pazion
> >>  wrote:
> >> >
> >> > Tonight my server with NFS hosted-engine mount crashed.
> >> > Now all is back online ,except the hosted engine. I can't ping  or ssh
> >> > the
> >> > machine
> >> >
> >> > when I do hosted-engine --vm-status, I get:
> >> >
> >> > ..
> >> > --== Host 2 status ==--
> >> >
> >> > Status up-to-date  : True
> >> > Hostname   : geisha-3.pazion.nl
> >> > Host ID: 2
> >> > Engine status  : {"reason": "failed liveliness
> >> > check",
> >> > "health": "bad", "vm": "up", "detail": "up"}
> >> > Score  : 3400
> >> > stopped: False
> >> > Local maintenance  : False
> >> > crc32  : d71d7c6b
> >> > Host timestamp : 4404
> >> > 
> >> >
> >> > I tried restarting all services/nfs mounts, start hosted engine on
> other
> >> > hosts, but all the same host up, but liveliness failed and unable to
> >> > access
> >> > the network/IP.
> >> >
> >> > I imagine it is stuck at the console requiring a fsck check maybe?
> >> > Is there a way to access the boot display directly?
> >> >
> >> > Any help is highly appreciated!
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> > ___
> >> > Users mailing list
> >> > Users@ovirt.org
> >> > http://lists.ovirt.org/mailman/listinfo/users
> >> >
> >
> >
> > ___
> > Users mailing list
> > Users@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> >
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] vhostmd vdsm-hook

2016-04-15 Thread Simon Barrett
I'm trying to use the vhostmd vdsm host to access ovirt node metrics from 
within a VM. Vhostmd is running and updating the /dev/shm/vhostmd0 on the ovirt 
node.

The part I'm stuck on is: "This disk image is exported read-only to guests. 
Guests can read the disk image to see metrics" from 
http://www.ovirt.org/develop/developer-guide/vdsm/hook/vhostmd/

Does the hook do this by default? I don't see any new read-only device mounted 
in the guest. Is there additional work I need to do to mount this and access 
the data from within the guest?

Many thanks,

Simon
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Fwd: Re: HA agent fails to start

2016-04-15 Thread Simone Tiraboschi
On Fri, Apr 15, 2016 at 8:45 AM, Richard Neuboeck  wrote:
> On 04/14/2016 11:03 PM, Simone Tiraboschi wrote:
>> On Thu, Apr 14, 2016 at 10:38 PM, Simone Tiraboschi  
>> wrote:
>>> On Thu, Apr 14, 2016 at 6:53 PM, Richard Neuboeck  
>>> wrote:
 On 14.04.16 18:46, Simone Tiraboschi wrote:
> On Thu, Apr 14, 2016 at 4:04 PM, Richard Neuboeck  
> wrote:
>> On 04/14/2016 02:14 PM, Simone Tiraboschi wrote:
>>> On Thu, Apr 14, 2016 at 12:51 PM, Richard Neuboeck
>>>  wrote:
 On 04/13/2016 10:00 AM, Simone Tiraboschi wrote:
> On Wed, Apr 13, 2016 at 9:38 AM, Richard Neuboeck 
>  wrote:
>> The answers file shows the setup time of both machines.
>>
>> On both machines hosted-engine.conf got rotated right before I wrote
>> this mail. Is it possible that I managed to interrupt the rotation 
>> with
>> the reboot so the backup was accurate but the update not yet written 
>> to
>> hosted-engine.conf?
>
> AFAIK we don't have any rotation mechanism for that file; something
> else you have in place on that host?

 Those machines are all CentOS 7.2 minimal installs. The only
 adaptation I do is installing vim, removing postfix and installing
 exim, removing firewalld and installing iptables-service. Then I add
 the oVirt repos (3.6 and 3.6-snapshot) and deploy the host.

 But checking lsof shows that 'ovirt-ha-agent --no-daemon' has access
 to the config file (and the one ending with ~):

 # lsof | grep 'hosted-engine.conf~'
 ovirt-ha- 193446   vdsm  351u  REG
 253,01021135070683
 /etc/ovirt-hosted-engine/hosted-engine.conf~
>>>
>>> This is not that much relevant if the file was renamed after
>>> ovirt-ha-agent opened it.
>>> Try this:
>>>
>>> [root@c72he20160405h1 ovirt-hosted-engine-setup]# tail -n1 -f
>>> /etc/ovirt-hosted-engine/hosted-engine.conf &
>>> [1] 28866
>>> [root@c72he20160405h1 ovirt-hosted-engine-setup]# port=
>>>
>>> [root@c72he20160405h1 ovirt-hosted-engine-setup]# lsof | grep 
>>> hosted-engine.conf
>>> tail  28866  root3r  REG
>>> 253,0  10141595898 /etc/ovirt-hosted-engine/hosted-engine.conf
>>> [root@c72he20160405h1 ovirt-hosted-engine-setup]# mv
>>> /etc/ovirt-hosted-engine/hosted-engine.conf
>>> /etc/ovirt-hosted-engine/hosted-engine.conf_123
>>> [root@c72he20160405h1 ovirt-hosted-engine-setup]# lsof | grep 
>>> hosted-engine.conf
>>> tail  28866  root3r  REG
>>> 253,0  10141595898
>>> /etc/ovirt-hosted-engine/hosted-engine.conf_123
>>> [root@c72he20160405h1 ovirt-hosted-engine-setup]#
>>>
>>
>> I've issued the commands you suggested but I don't know how that
>> helps to find the process accessing the config files.
>>
>> After moving the hosted-engine.conf file the HA agent crashed
>> logging the information that the config file is not available.
>>
>> Here is the output from every command:
>>
>> # tail -n1 -f /etc/ovirt-hosted-engine/hosted-engine.conf &
>> [1] 167865
>> [root@cube-two ~]# port=
>> # lsof | grep hosted-engine.conf
>> ovirt-ha- 166609   vdsm5u  REG
>> 253,01021134433491
>> /etc/ovirt-hosted-engine/hosted-engine.conf (deleted)
>> ovirt-ha- 166609   vdsm7u  REG
>> 253,01021134433453
>> /etc/ovirt-hosted-engine/hosted-engine.conf (deleted)
>> ovirt-ha- 166609   vdsm8u  REG
>> 253,01021134433489
>> /etc/ovirt-hosted-engine/hosted-engine.conf (deleted)
>> ovirt-ha- 166609   vdsm9u  REG
>> 253,01021134433493
>> /etc/ovirt-hosted-engine/hosted-engine.conf~
>> ovirt-ha- 166609   vdsm   10u  REG
>> 253,01021134433495
>> /etc/ovirt-hosted-engine/hosted-engine.conf
>> tail  167865   root3r  REG
>> 253,01021134433493
>> /etc/ovirt-hosted-engine/hosted-engine.conf~
>> # mv /etc/ovirt-hosted-engine/hosted-engine.conf
>> /etc/ovirt-hosted-engine/hosted-engine.conf_123
>> # lsof | grep hosted-engine.conf
>> ovirt-ha- 166609   vdsm5u  REG
>> 253,01021134433491
>> /etc/ovirt-hosted-engine/hosted-engine.conf (deleted)
>> ovirt-ha- 166609   vdsm7u  REG
>> 253,01021134433453
>> /etc/ovirt-hosted-engine/hosted-engine.conf (deleted)

Re: [ovirt-users] hosted-engine stuck "failed liveliness check" "detail": "up"

2016-04-15 Thread Simone Tiraboschi
On Fri, Apr 15, 2016 at 11:02 AM, Paul Groeneweg | Pazion
 wrote:
> Thanks!
>
> I managed to get the console through:
>
> hosted-engine --add-console-password
> /bin/remote-viewer vnc://localhost:5900
>
> Turns out, there seems to be some corruption on the partition:
> http://screencast.com/t/6iR0U3QuI
>
> Is there a way to boot from CD, so I can start rescue mode?

Yes,
you have to set global maintenance mode to avoid VM start attempts on
other hosts.

You have to poweroff the engine VM with:
 hosted-engine --vm-shutdown/--vm-poweroff (probably the second one in
your case)
Then you have to create a copy of
/var/run/ovirt-hosted-engine-ha/vm.conf and you have to manually edit
it in order to attach the ISO image and change the boot order.
At that point you can start the engine VM with your custom configuration with:
 hosted-engine --vm-start --vm-conf=my_custom_vm.conf

Good luck!

> Op vr 15 apr. 2016 om 10:58 schreef Martin Sivak :
>>
>> Hi,
>>
>> you can access the console using vnc or use virsh to get access to the
>> serial console.
>>
>> Check the following commands on the host where the VM is currently
>> running:
>>
>> virsh -r list
>> virsh -r console HostedEngine
>> virsh -r vncdisplay HostedEngine
>>
>> Those should give you enough pointers to connect to the VM.
>>
>> Regards
>>
>> Martin Sivak
>>
>>
>>
>> On Fri, Apr 15, 2016 at 10:14 AM, Paul Groeneweg | Pazion
>>  wrote:
>> >
>> > Tonight my server with NFS hosted-engine mount crashed.
>> > Now all is back online ,except the hosted engine. I can't ping  or ssh
>> > the
>> > machine
>> >
>> > when I do hosted-engine --vm-status, I get:
>> >
>> > ..
>> > --== Host 2 status ==--
>> >
>> > Status up-to-date  : True
>> > Hostname   : geisha-3.pazion.nl
>> > Host ID: 2
>> > Engine status  : {"reason": "failed liveliness
>> > check",
>> > "health": "bad", "vm": "up", "detail": "up"}
>> > Score  : 3400
>> > stopped: False
>> > Local maintenance  : False
>> > crc32  : d71d7c6b
>> > Host timestamp : 4404
>> > 
>> >
>> > I tried restarting all services/nfs mounts, start hosted engine on other
>> > hosts, but all the same host up, but liveliness failed and unable to
>> > access
>> > the network/IP.
>> >
>> > I imagine it is stuck at the console requiring a fsck check maybe?
>> > Is there a way to access the boot display directly?
>> >
>> > Any help is highly appreciated!
>> >
>> >
>> >
>> >
>> >
>> >
>> > ___
>> > Users mailing list
>> > Users@ovirt.org
>> > http://lists.ovirt.org/mailman/listinfo/users
>> >
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hosted-engine stuck "failed liveliness check" "detail": "up"

2016-04-15 Thread Paul Groeneweg | Pazion
Thanks!

I managed to get the console through:

hosted-engine --add-console-password
/bin/remote-viewer vnc://localhost:5900

Turns out, there seems to be some corruption on the partition:
http://screencast.com/t/6iR0U3QuI

Is there a way to boot from CD, so I can start rescue mode?


Op vr 15 apr. 2016 om 10:58 schreef Martin Sivak :

> Hi,
>
> you can access the console using vnc or use virsh to get access to the
> serial console.
>
> Check the following commands on the host where the VM is currently running:
>
> virsh -r list
> virsh -r console HostedEngine
> virsh -r vncdisplay HostedEngine
>
> Those should give you enough pointers to connect to the VM.
>
> Regards
>
> Martin Sivak
>
>
>
> On Fri, Apr 15, 2016 at 10:14 AM, Paul Groeneweg | Pazion
>  wrote:
> >
> > Tonight my server with NFS hosted-engine mount crashed.
> > Now all is back online ,except the hosted engine. I can't ping  or ssh
> the
> > machine
> >
> > when I do hosted-engine --vm-status, I get:
> >
> > ..
> > --== Host 2 status ==--
> >
> > Status up-to-date  : True
> > Hostname   : geisha-3.pazion.nl
> > Host ID: 2
> > Engine status  : {"reason": "failed liveliness
> check",
> > "health": "bad", "vm": "up", "detail": "up"}
> > Score  : 3400
> > stopped: False
> > Local maintenance  : False
> > crc32  : d71d7c6b
> > Host timestamp : 4404
> > 
> >
> > I tried restarting all services/nfs mounts, start hosted engine on other
> > hosts, but all the same host up, but liveliness failed and unable to
> access
> > the network/IP.
> >
> > I imagine it is stuck at the console requiring a fsck check maybe?
> > Is there a way to access the boot display directly?
> >
> > Any help is highly appreciated!
> >
> >
> >
> >
> >
> >
> > ___
> > Users mailing list
> > Users@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> >
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hosted-engine stuck "failed liveliness check" "detail": "up"

2016-04-15 Thread Martin Sivak
Hi,

you can access the console using vnc or use virsh to get access to the
serial console.

Check the following commands on the host where the VM is currently running:

virsh -r list
virsh -r console HostedEngine
virsh -r vncdisplay HostedEngine

Those should give you enough pointers to connect to the VM.

Regards

Martin Sivak



On Fri, Apr 15, 2016 at 10:14 AM, Paul Groeneweg | Pazion
 wrote:
>
> Tonight my server with NFS hosted-engine mount crashed.
> Now all is back online ,except the hosted engine. I can't ping  or ssh the
> machine
>
> when I do hosted-engine --vm-status, I get:
>
> ..
> --== Host 2 status ==--
>
> Status up-to-date  : True
> Hostname   : geisha-3.pazion.nl
> Host ID: 2
> Engine status  : {"reason": "failed liveliness check",
> "health": "bad", "vm": "up", "detail": "up"}
> Score  : 3400
> stopped: False
> Local maintenance  : False
> crc32  : d71d7c6b
> Host timestamp : 4404
> 
>
> I tried restarting all services/nfs mounts, start hosted engine on other
> hosts, but all the same host up, but liveliness failed and unable to access
> the network/IP.
>
> I imagine it is stuck at the console requiring a fsck check maybe?
> Is there a way to access the boot display directly?
>
> Any help is highly appreciated!
>
>
>
>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Move hosted storage

2016-04-15 Thread Marcin Michta

  
  
Hi,

I created hosted engine on the NFS storage at server1. I wonder -
can I move (migrate) hosted storage to server2 (also NFS) and how to
make work after all.


  
 
  Thanks in advance! 
  Marcin.

  

  

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] hosted-engine stuck "failed liveliness check" "detail": "up"

2016-04-15 Thread Paul Groeneweg | Pazion
Tonight my server with NFS hosted-engine mount crashed.
Now all is back online ,except the hosted engine. I can't ping  or ssh the
machine

when I do hosted-engine --vm-status, I get:

..
--== Host 2 status ==--

Status up-to-date  : True
Hostname   : geisha-3.pazion.nl
Host ID: 2
Engine status  : {"reason": "failed liveliness check",
"health": "bad", "vm": "up", "detail": "up"}
Score  : 3400
stopped: False
Local maintenance  : False
crc32  : d71d7c6b
Host timestamp : 4404


I tried restarting all services/nfs mounts, start hosted engine on other
hosts, but all the same host up, but liveliness failed and unable to access
the network/IP.

I imagine it is stuck at the console requiring a fsck check maybe?
Is there a way to access the boot display directly?

Any help is highly appreciated!
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Fwd: Re: HA agent fails to start

2016-04-15 Thread Richard Neuboeck
On 04/14/2016 11:03 PM, Simone Tiraboschi wrote:
> On Thu, Apr 14, 2016 at 10:38 PM, Simone Tiraboschi  
> wrote:
>> On Thu, Apr 14, 2016 at 6:53 PM, Richard Neuboeck  
>> wrote:
>>> On 14.04.16 18:46, Simone Tiraboschi wrote:
 On Thu, Apr 14, 2016 at 4:04 PM, Richard Neuboeck  
 wrote:
> On 04/14/2016 02:14 PM, Simone Tiraboschi wrote:
>> On Thu, Apr 14, 2016 at 12:51 PM, Richard Neuboeck
>>  wrote:
>>> On 04/13/2016 10:00 AM, Simone Tiraboschi wrote:
 On Wed, Apr 13, 2016 at 9:38 AM, Richard Neuboeck 
  wrote:
> The answers file shows the setup time of both machines.
>
> On both machines hosted-engine.conf got rotated right before I wrote
> this mail. Is it possible that I managed to interrupt the rotation 
> with
> the reboot so the backup was accurate but the update not yet written 
> to
> hosted-engine.conf?

 AFAIK we don't have any rotation mechanism for that file; something
 else you have in place on that host?
>>>
>>> Those machines are all CentOS 7.2 minimal installs. The only
>>> adaptation I do is installing vim, removing postfix and installing
>>> exim, removing firewalld and installing iptables-service. Then I add
>>> the oVirt repos (3.6 and 3.6-snapshot) and deploy the host.
>>>
>>> But checking lsof shows that 'ovirt-ha-agent --no-daemon' has access
>>> to the config file (and the one ending with ~):
>>>
>>> # lsof | grep 'hosted-engine.conf~'
>>> ovirt-ha- 193446   vdsm  351u  REG
>>> 253,01021135070683
>>> /etc/ovirt-hosted-engine/hosted-engine.conf~
>>
>> This is not that much relevant if the file was renamed after
>> ovirt-ha-agent opened it.
>> Try this:
>>
>> [root@c72he20160405h1 ovirt-hosted-engine-setup]# tail -n1 -f
>> /etc/ovirt-hosted-engine/hosted-engine.conf &
>> [1] 28866
>> [root@c72he20160405h1 ovirt-hosted-engine-setup]# port=
>>
>> [root@c72he20160405h1 ovirt-hosted-engine-setup]# lsof | grep 
>> hosted-engine.conf
>> tail  28866  root3r  REG
>> 253,0  10141595898 /etc/ovirt-hosted-engine/hosted-engine.conf
>> [root@c72he20160405h1 ovirt-hosted-engine-setup]# mv
>> /etc/ovirt-hosted-engine/hosted-engine.conf
>> /etc/ovirt-hosted-engine/hosted-engine.conf_123
>> [root@c72he20160405h1 ovirt-hosted-engine-setup]# lsof | grep 
>> hosted-engine.conf
>> tail  28866  root3r  REG
>> 253,0  10141595898
>> /etc/ovirt-hosted-engine/hosted-engine.conf_123
>> [root@c72he20160405h1 ovirt-hosted-engine-setup]#
>>
>
> I've issued the commands you suggested but I don't know how that
> helps to find the process accessing the config files.
>
> After moving the hosted-engine.conf file the HA agent crashed
> logging the information that the config file is not available.
>
> Here is the output from every command:
>
> # tail -n1 -f /etc/ovirt-hosted-engine/hosted-engine.conf &
> [1] 167865
> [root@cube-two ~]# port=
> # lsof | grep hosted-engine.conf
> ovirt-ha- 166609   vdsm5u  REG
> 253,01021134433491
> /etc/ovirt-hosted-engine/hosted-engine.conf (deleted)
> ovirt-ha- 166609   vdsm7u  REG
> 253,01021134433453
> /etc/ovirt-hosted-engine/hosted-engine.conf (deleted)
> ovirt-ha- 166609   vdsm8u  REG
> 253,01021134433489
> /etc/ovirt-hosted-engine/hosted-engine.conf (deleted)
> ovirt-ha- 166609   vdsm9u  REG
> 253,01021134433493
> /etc/ovirt-hosted-engine/hosted-engine.conf~
> ovirt-ha- 166609   vdsm   10u  REG
> 253,01021134433495
> /etc/ovirt-hosted-engine/hosted-engine.conf
> tail  167865   root3r  REG
> 253,01021134433493
> /etc/ovirt-hosted-engine/hosted-engine.conf~
> # mv /etc/ovirt-hosted-engine/hosted-engine.conf
> /etc/ovirt-hosted-engine/hosted-engine.conf_123
> # lsof | grep hosted-engine.conf
> ovirt-ha- 166609   vdsm5u  REG
> 253,01021134433491
> /etc/ovirt-hosted-engine/hosted-engine.conf (deleted)
> ovirt-ha- 166609   vdsm7u  REG
> 253,01021134433453
> /etc/ovirt-hosted-engine/hosted-engine.conf (deleted)
> ovirt-ha- 166609   vdsm8u  REG
> 253,01021134433489
> /etc/ovirt-hosted-engine/hosted-engine.conf (deleted)
> ovirt-ha- 

Re: [ovirt-users] Hosted engine on gluster problem

2016-04-15 Thread Sandro Bonazzola
On Thu, Apr 14, 2016 at 7:35 PM, Nir Soffer  wrote:

> On Wed, Apr 13, 2016 at 4:34 PM, Luiz Claudio Prazeres Goncalves
>  wrote:
> > Nir, here is the problem:
> > https://bugzilla.redhat.com/show_bug.cgi?id=1298693
> >
> > When you do a hosted-engine --deploy and pick "glusterfs" you don't have
> a
> > way to define the mount options, therefore, the use of the
> > "backupvol-server", however when you create a storage domain from the UI
> you
> > can, like the attached screen shot.
> >
> >
> > In the hosted-engine --deploy, I would expect a flow which includes not
> only
> > the "gluster" entrypoint, but also the gluster mount options which is
> > missing today. This option would be optional, but would remove the single
> > point of failure described on the Bug 1298693.
> >
> > for example:
> >
> > Existing entry point on the "hosted-engine --deploy" flow
> > gluster1.xyz.com:/engine
>
> I agree, this feature must be supported.
>

It will, and it's currently targeted to 4.0.



>
> > Missing option on the "hosted-engine --deploy" flow :
> > backupvolfile-server=gluster2.xyz.com
> ,fetch-attempts=3,log-level=WARNING,log-file=/var/log/glusterfs/gluster_engine_domain.log
> >
> > Sandro, it seems to me a simple solution which can be easily fixed.
> >
> > What do you think?
> >
> > Regards
> > -Luiz
> >
> >
> >
> > 2016-04-13 4:15 GMT-03:00 Sandro Bonazzola :
> >>
> >>
> >>
> >> On Tue, Apr 12, 2016 at 6:47 PM, Nir Soffer  wrote:
> >>>
> >>> On Tue, Apr 12, 2016 at 3:05 PM, Luiz Claudio Prazeres Goncalves
> >>>  wrote:
> >>> > Hi Sandro, I've been using gluster with 3 external hosts for a while
> >>> > and
> >>> > things are working pretty well, however this single point of failure
> >>> > looks
> >>> > like a simple feature to implement,but critical to anyone who wants
> to
> >>> > use
> >>> > gluster on production  . This is not hyperconvergency which has other
> >>> > issues/implications. So , why not have this feature out on 3.6
> branch?
> >>> > It
> >>> > looks like just let vdsm use the 'backupvol-server' option when
> >>> > mounting the
> >>> > engine domain and make the property tests.
> >>>
> >>> Can you explain what is the problem, and what is the suggested
> solution?
> >>>
> >>> Engine and vdsm already support the backupvol-server option - you can
> >>> define this option in the storage domain options when you create a
> >>> gluster
> >>> storage domain. With this option vdsm should be able to connect to
> >>> gluster
> >>> storage domain even if a brick is down.
> >>>
> >>> If you don't have this option in engine , you probably cannot add it
> with
> >>> hosted
> >>> engine setup, since for editing it you must put the storage domain in
> >>> maintenance
> >>> and if you do this the engine vm will be killed :-) This is is one of
> >>> the issues with
> >>> engine managing the storage domain it runs on.
> >>>
> >>> I think the best way to avoid this issue, is to add a DNS entry
> >>> providing the addresses
> >>> of all the gluster bricks, and use this address for the gluster
> >>> storage domain. This way
> >>> the glusterfs mount helper can mount the domain even if one of the
> >>> gluster bricks
> >>> are down.
> >>>
> >>> Again, we will need some magic from the hosted engine developers to
> >>> modify the
> >>> address of the hosted engine gluster domain on existing system.
> >>
> >>
> >> Magic won't happen without a bz :-) please open one describing what's
> >> requested.
> >>
> >>
> >>>
> >>>
> >>> Nir
> >>>
> >>> >
> >>> > Could you add this feature to the next release of 3.6 branch?
> >>> >
> >>> > Thanks
> >>> > Luiz
> >>> >
> >>> > Em ter, 12 de abr de 2016 05:03, Sandro Bonazzola <
> sbona...@redhat.com>
> >>> > escreveu:
> >>> >>
> >>> >> On Mon, Apr 11, 2016 at 11:44 PM, Bond, Darryl  >
> >>> >> wrote:
> >>> >>>
> >>> >>> My setup is hyperconverged. I have placed my test results in
> >>> >>> https://bugzilla.redhat.com/show_bug.cgi?id=1298693
> >>> >>>
> >>> >>
> >>> >> Ok, so you're aware about the limitation of the single point of
> >>> >> failure.
> >>> >> If you drop the host referenced in hosted engine configuration for
> the
> >>> >> initial setup it won't be able to connect to shared storage even if
> >>> >> the
> >>> >> other hosts in the cluster are up since the entry point is down.
> >>> >> Note that hyperconverged deployment is not supported in 3.6.
> >>> >>
> >>> >>
> >>> >>>
> >>> >>>
> >>> >>> Short description of setup:
> >>> >>>
> >>> >>> 3 hosts with 2 disks each set up with gluster replica 3 across the
> 6
> >>> >>> disks volume name hosted-engine.
> >>> >>>
> >>> >>> Hostname hosted-storage configured in /etc//hosts to point to the
> >>> >>> host1.
> >>> >>>
> >>> >>> Installed hosted engine on host1 with the hosted engine storage
> path
> >>> >>> =
> >>> >>> hosted-storage:/hosted-engine
> >>> >>>
> >>> >>> Install first engine on h1 successful.