Re: [ovirt-users] Mailing-Lists upgrade

2017-09-06 Thread Karli Sjöberg
On tor, 2017-09-07 at 13:12 +0900, Marc Dequènes (Duck) wrote:
> Quack,
> 
> On behalf of the oVirt infra team, I'd like to announce the current
> Mailing-Lists system is going to be upgraded to a brand new Mailman 3
> installation on Monday during the slot 11:00-12:00 JST.
> 
> It should not take a full hour to migrate as we already made
> incremental
> synchronization with the current system but better keep some margin.
> The
> system will then take over delivery of the mails but might be a bit
> slow
> at first as it needs to reindex all the archived mails (which might
> take
> a few hours).
> 
> To manage your subscriptions and delivery settings you can do this
> easily on the much nicer web interface (https://lists.ovirt.org).
> There
> is a notion of account so you don't need to login separately for each
> ML.
> 
> You can Sign In using Fedora, GitHub or Google or create a local
> account
> if you prefer. Please keep in mind signing in with a different method
> would create separate accounts (which cannot be merged at the
> moment).
> But you can easily link your account to other authentication methods
> in
> your settings (click on you name in the up-right corner -> Account ->
> Account Connections).

Hey Duck!

Stupid question perhaps but, have I understood it correctly that all of
us are going to have to create new accounts for the new mail-list
system, if we don´t wish to log in using Fedora, GitHub or Google? Or
are the existing local accounts going to be migrated as well?

That's how I interpreted it at least.

/K

> 
> As for the original mail archives, because the previous system did
> not
> have stable URLs, we cannot create mappings to the new pages. We
> decided
> to keep the old archives around on the same URL (/pipermail), so the
> Internet links would still work fine.
> 
> Hope you'd be happy with the new system.
> \_o<
> 
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Mailing-Lists upgrade

2017-09-06 Thread Duck
Quack,

On behalf of the oVirt infra team, I'd like to announce the current
Mailing-Lists system is going to be upgraded to a brand new Mailman 3
installation on Monday during the slot 11:00-12:00 JST.

It should not take a full hour to migrate as we already made incremental
synchronization with the current system but better keep some margin. The
system will then take over delivery of the mails but might be a bit slow
at first as it needs to reindex all the archived mails (which might take
a few hours).

To manage your subscriptions and delivery settings you can do this
easily on the much nicer web interface (https://lists.ovirt.org). There
is a notion of account so you don't need to login separately for each ML.

You can Sign In using Fedora, GitHub or Google or create a local account
if you prefer. Please keep in mind signing in with a different method
would create separate accounts (which cannot be merged at the moment).
But you can easily link your account to other authentication methods in
your settings (click on you name in the up-right corner -> Account ->
Account Connections).

As for the original mail archives, because the previous system did not
have stable URLs, we cannot create mappings to the new pages. We decided
to keep the old archives around on the same URL (/pipermail), so the
Internet links would still work fine.

Hope you'd be happy with the new system.
\_o<



signature.asc
Description: OpenPGP digital signature
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Failed to check for available updates on host *** failed with message...

2017-09-06 Thread Manuel Luis Aznar
Hello there,

I am getting this error in the ovirt hosted engine:

2017-09-06 16:45:56,160+01 ERROR
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(pool-7-thread-1) [4ff48065] EVENT_ID: HOST_AVAILABLE_UPDATES_FAILED(839),
Correlation ID: null, Call Stack: null, Custom Event ID: -1, Message: *Failed
to check for available updates on host host1.bajada.es
 with message 'SSH authentication to
'r...@host1.bajada.es ' failed. Please verify
provided credentials. Make sure key is authorized at host'.*

I am up and running: *oVirt Engine Version: 4.1.1.8-1.el7.centos*.

I suppose that this error is related to SSH public and private keys. Am I
right?¿

Thanks for all in advance
Any help would be appreciated
Manuel Luis Aznar
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt selft-hosted with NFS on top gluster

2017-09-06 Thread Abi Askushi
For a first idea I use:

dd if=/dev/zero of=testfile bs=1GB count=1

When testing on the gluster mount point using above command I hardly get
10MB/s. (On the same time the network traffic hardly reaches 100Mbit).

When testing our of the gluster (for example at /root) I get 600 - 700MB/s.

When I mount the gluster volume with NFS and test on it I get 90 - 100
MB/s, (almost 10x from gluster results) which is the max I can get
considering I have only 1 Gbit network for the storage.

Also, when using glusterfs the general VM performance is very poor and disk
write benchmarks show that is it at least 4 times slower then when the VM
is hosted on the same data store when NFS mounted.

I don't know why I hitting such a significant performance penalty, and
every possible tweak that I was able to find out there did not make any
difference on the performance.

The hardware I am using is pretty decent for the purposes intended:
3 nodes, each node having with 32 MB of RAM, 16 physical CPU cores, 2 TB of
storage in RAID5 (4 disks), of which 1.5 TB are sliced for the data store
of ovirt where VMs are stored.

The gluster configuration is the following:

Volume Name: vms
Type: Replicate
Volume ID: 4513340d-7919-498b-bfe0-d836b5cea40b
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x (2 + 1) = 3
Transport-type: tcp
Bricks:
Brick1: gluster0:/gluster/vms/brick
Brick2: gluster1:/gluster/vms/brick
Brick3: gluster2:/gluster/vms/brick (arbiter)
Options Reconfigured:
nfs.export-volumes: on
nfs.disable: off
performance.readdir-ahead: on
transport.address-family: inet
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: on
performance.low-prio-threads: 32
network.remote-dio: off
cluster.eager-lock: off
cluster.quorum-type: auto
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 1
features.shard: on
user.cifs: off
storage.owner-uid: 36
storage.owner-gid: 36
network.ping-timeout: 30
performance.strict-o-direct: on
cluster.granular-entry-heal: enable
features.shard-block-size: 64MB
performance.client-io-threads: on
client.event-threads: 4
server.event-threads: 4
performance.write-behind-window-size: 4MB
performance.cache-size: 1GB

In case I can provide any other details let me know.
At the moment I already switched to gluster based NFS but I have a similar
setup with 2 nodes  where the data store is mounted through gluster (and
again relatively good hardware) where I might check any tweaks or
improvements on this setup.

Thanx


On Wed, Sep 6, 2017 at 5:32 PM, Yaniv Kaul  wrote:

>
>
> On Wed, Sep 6, 2017 at 3:32 PM, Abi Askushi 
> wrote:
>
>> Hi All,
>>
>> I've playing with ovirt self hosted engine setup and I even use it to
>> production for several VM. The setup I have is 3 server with gluster
>> storage in replica 2+1 (1 arbiter).
>> The data storage domain where VMs are stored is mounted with gluster
>> through ovirt. The performance I get for the VMs is very low and I was
>> thinking to switch and mount the same storage through NFS instead of
>> glusterfs.
>>
>
> I don't see how it'll improve performance.
> I suggest you share the gluster configuration (as well as the storage HW)
> so we can understand why the performance is low.
> Y.
>
>
>>
>> The only think I am hesitant is how can I ensure high availability of the
>> storage when I loose one server? I was thinking to have at /etc/hosts sth
>> like below:
>>
>> 10.100.100.1 nfsmount
>> 10.100.100.2 nfsmount
>> 10.100.100.3 nfsmount
>>
>> then use nfsmount as the server name when adding this domain through
>> ovirt GUI.
>> Are there any other more elegant solutions? What do you do for such cases?
>> Note: gluster has the back-vol-file option which provides a lean way to
>> have redundancy on the mount point and I am using this when mounting with
>> glusterfs.
>>
>> Thanx
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] disk attachment to VM

2017-09-06 Thread Yaniv Kaul
On Tue, Sep 5, 2017 at 4:53 PM, Benny Zlotnik  wrote:

> Hi,
>
> Look at [1], however there are caveats so be sure to pay close attention
> to the warning section.
>

You could also use it via device pass-through.
Y.


>
> [1] - https://github.com/oVirt/vdsm/blob/master/vdsm_hooks/local
> disk/README
>
> On Tue, Sep 5, 2017 at 4:52 PM, Benny Zlotnik  wrote:
>
>> Hi,
>>
>> Look at [1], however there are caveats so be sure to pay close attention
>> to the warning section.
>>
>> [1] - https://github.com/oVirt/vdsm/blob/master/vdsm_hooks/local
>> disk/README
>>
>>
>> On Tue, Sep 5, 2017 at 4:40 PM, Erekle Magradze <
>> erekle.magra...@recogizer.de> wrote:
>>
>>> Hey Guys,
>>> Is there a way to attach an SSD directly to the oVirt VM?
>>> Thanks in advance
>>> Cheers
>>> Erekle
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>>
>>
>>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] cpu, core and thread mappings

2017-09-06 Thread Yaniv Kaul
On Wed, Sep 6, 2017 at 2:47 PM, Gianluca Cecchi 
wrote:

> Hello,
> I was talking with a guy expert in VMware and discussing performance of
> VMs in respect of virtual cpus assigned to them in relation with mapping
> with the real hw of the hypervisor underneath.
>
> One of the topics was numa usage and its overheads in case of a "too" big
> VM, in terms of both number of vcpus and memory amount.
> Eg:
> suppose host has 2 intel based sockets, with 6 cores and HT enabled and
> has 96Gb of ram (distributed 48+48 between the 2 processors)
> suppose I configure a VM with 16 vcpus (2:4:2): would be the mapping
> respected at physical level or only a sort of "hint" for the hypervisor?
> Can I say that it would perform better if I configure it 12 vcpus and
> mapping 1:6:2, because it can stay all inside one cpu?
>

Hard to say without relationship to the workload. You are losing 4 vCPUs -
perhaps those can be used for something (the OS) while the rest of them
could be used by the application, for example?


>
> And what if I define a VM with 52Gb of ram? Can I say that it would
> perform in general better if I try to get it all in one cpu related memory
> slots (eg not more than 48Gb in my example)?
>
> Hard to say without relationship to the workload - will it need all the
memory? Will it be accessing all of it, in random order?

If you've maxed out one node, you just need more memory from the other
node.


> Are there any documents going more deeply in these sort of considerations?
>
>
It is so workload dependent that there will not be a one size fit all.

Also, if one goes and sizes so that the biggest VM is able to all-stay
> inside one cpu-memory, does it make sense to say that it will perform
> better in this scenario a cluster composed by 4 nodes, each one with 1
> socket and 48Gb of memory instead of a cluster of 2 nodes, each one with 2
> sockets and 96Gb of ram?
>

You could have used affinity.

See [1] for some details on Redis. Note that IO (specifically network) is
just as important - and its impact is much more profound.
Y.

[1] https://redis.io/topics/benchmarks


>
> Hope I have clarified my questions/doubts.
>
>
> Thanks in advance for any insight,
> Gianluca
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt selft-hosted with NFS on top gluster

2017-09-06 Thread Yaniv Kaul
On Wed, Sep 6, 2017 at 3:32 PM, Abi Askushi  wrote:

> Hi All,
>
> I've playing with ovirt self hosted engine setup and I even use it to
> production for several VM. The setup I have is 3 server with gluster
> storage in replica 2+1 (1 arbiter).
> The data storage domain where VMs are stored is mounted with gluster
> through ovirt. The performance I get for the VMs is very low and I was
> thinking to switch and mount the same storage through NFS instead of
> glusterfs.
>

I don't see how it'll improve performance.
I suggest you share the gluster configuration (as well as the storage HW)
so we can understand why the performance is low.
Y.


>
> The only think I am hesitant is how can I ensure high availability of the
> storage when I loose one server? I was thinking to have at /etc/hosts sth
> like below:
>
> 10.100.100.1 nfsmount
> 10.100.100.2 nfsmount
> 10.100.100.3 nfsmount
>
> then use nfsmount as the server name when adding this domain through ovirt
> GUI.
> Are there any other more elegant solutions? What do you do for such cases?
> Note: gluster has the back-vol-file option which provides a lean way to
> have redundancy on the mount point and I am using this when mounting with
> glusterfs.
>
> Thanx
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Cannot install ovirt guest agent- repo does not exists

2017-09-06 Thread Yohan JAROSZ
Dear Vinzenz, dear list

It seems that the repo evilissimo is not here anymore: 
http://download.opensuse.org/repositories/home:/
So we can’t install overt agent as stated in the docs: 
https://www.ovirt.org/documentation/how-to/guest-agent/install-the-guest-agent-in-ubuntu/
(http://download.opensuse.org/repositories/home:/evilissimo:/ubuntu:/16.04/xUbuntu_16.04/)

best,
Yo.




Yohan Jarosz
Scientific Collaborator

UNIVERSITÉ DU LUXEMBOURG

LUXEMBOURG CENTRE FOR SYSTEMS BIOMEDICINE
Campus Belval | House of Biomedicine
7, avenue des Hauts-Fourneaux
L-4362 Esch-sur-Alzette
T +352 46 66 44 6669
F +352 46 66 44 3 6669


yohan.jar...@uni.lu  http://lcsb.uni.lu

LCSB - Accelerating Biomedicine!: https://www.youtube.com/watch?v=oLUE6DjSB7Y
-
This message is confidential and may contain privileged information. It is 
intended for the named recipient only. If you receive it in error please notify 
me and permanently delete the original message and any copies.
-

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Failed to import the Hosted Engine Storage Domain

2017-09-06 Thread Simone Tiraboschi
On Wed, Sep 6, 2017 at 1:22 PM, Arsène Gschwind 
wrote:

>
>
> On 09/05/2017 09:02 AM, Arsène Gschwind wrote:
>
>
>
> On 09/04/2017 07:59 PM, Simone Tiraboschi wrote:
>
>
> On Mon, Sep 4, 2017 at 7:32 PM, Arsène Gschwind  > wrote:
>
>>
>>
>> On 09/04/2017 06:32 PM, Simone Tiraboschi wrote:
>>
>>
>>
>> On Mon, Sep 4, 2017 at 6:24 PM, Arsène Gschwind <
>> arsene.gschw...@unibas.ch> wrote:
>>
>>>
>>>
>>> On 09/04/2017 02:51 PM, Simone Tiraboschi wrote:
>>>
>>>
>>>
>>> On Mon, Sep 4, 2017 at 2:21 PM, Arsène Gschwind <
>>> arsene.gschw...@unibas.ch> wrote:
>>>


 On 09/04/2017 02:01 PM, Simone Tiraboschi wrote:



 On Mon, Sep 4, 2017 at 1:55 PM, Arsène Gschwind <
 arsene.gschw...@unibas.ch> wrote:

>
>
> On 09/04/2017 01:52 PM, Simone Tiraboschi wrote:
>
>
>
> On Mon, Sep 4, 2017 at 12:23 PM, Arsène Gschwind <
> arsene.gschw...@unibas.ch> wrote:
>
>> Hi Simone,
>>
>> On 09/04/2017 11:14 AM, Simone Tiraboschi wrote:
>>
>>
>>
>> On Mon, Sep 4, 2017 at 10:56 AM, Arsène Gschwind <
>> arsene.gschw...@unibas.ch> wrote:
>>
>>> Hi Didi,
>>>
>>> On 09/04/2017 10:15 AM, Yedidyah Bar David wrote:
>>>
>>> On Mon, Sep 4, 2017 at 10:16 AM, Arsène 
>>> Gschwind  wrote:
>>>
>>> Hi all,
>>>
>>> A while ago I had some problem with hosted-engine network which wasn't 
>>> set
>>> correctly at deploy time, so I finally decided to redeploy the hosted 
>>> engine
>>> in the hope the network will be set correctly this time. I've followed 
>>> this
>>> procedure:
>>>
>>> Stop all VMs
>>> Full backup of HE DB and export to safe place
>>> Cleanup HE storage following https://access.redhat.com/solutions/2121581
>>> Reboot Hosts
>>> Re-deploy HE until DB recovery
>>> Recover DB adding the following param:
>>>  --he-remove-storage-vm Removes the hosted-engine storage
>>> domain, all its entities and the hosted-engine VM during restore.
>>>  --he-remove-hosts  Removes all the hosted-engine hosts
>>> during restore.
>>>
>>> Finalize HE deployment.
>>>
>>> Everything did run without errors and I'm able to access Web UI.
>>>
>>> But now I don't see my HE VM and its respective Storage Domain, the logs
>>> says it isn't able to import it. I see all other SD and I'm able to 
>>> manage
>>> my VMs as before.
>>>
>>> Please find attached engine.log
>>>
>>> I think this is your problem:
>>>
>>> 2017-09-04 03:26:14,272+02 INFO
>>> [org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
>>> (org.ovirt.thread.pool-6-thread-24) [2383eaa0] There are existing luns
>>> in the system which are part of VG id
>>> 'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'
>>>
>>> I don't see a VG with this ID, here the IDs I see on the hosts:
>>>
>>>   VG   #PV #LV #SN Attr   VSize
>>> VFree
>>>   6b62cc06-fc44-4c38-af6d-bfd9cbe73246   1  10   0 wz--n- 99.62g
>>> 14.50g
>>>   b0414c06-d984-4001-a998-fd9a2e79fb83   2  70   0 wz--n- 10.00t
>>> 2.31t
>>>   b2e30961-7cff-4cca-83d6-bee3a4f890ee   2  47   0 wz--n-  5.27t
>>> 2.50t
>>>
>>
>>
>> Could you please repeat the command on host adm-kvmh70 ?
>>
>> 2017-09-04 09:04:18,163+02 INFO  [org.ovirt.engine.core.bll.st
>> orage.domain.ImportHostedEngineStorageDomainCommand]
>> (org.ovirt.thread.pool-6-thread-34) [247a3718] Running command:
>> ImportHostedEngineStorageDomainCommand internal: true.
>> 2017-09-04 09:04:18,189+02 INFO  [org.ovirt.engine.core.vdsbro
>> ker.vdsbroker.GetVGInfoVDSCommand] (org.ovirt.thread.pool-6-thread-34)
>> [7d2e6cb2] START, GetVGInfoVDSCommand(HostName = adm-kvmh70,
>> GetVGInfoVDSCommandParameters:{runAsync='true',
>> hostId='acbacabb-6c4a-43fd-a1e2-2d7ff2f6f98b',
>> VGID='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'}), log id: 6693b98a
>> 2017-09-04 09:04:18,232+02 INFO  [org.ovirt.engine.core.vdsbro
>> ker.vdsbroker.GetVGInfoVDSCommand] (org.ovirt.thread.pool-6-thread-34)
>> [7d2e6cb2] FINISH, GetVGInfoVDSCommand, return:
>> [LUNs:{id='repl_HostedEngine', 
>> physicalVolumeId='kYN8Jj-FBDw-MhxI-XcoZ-w1zH-eQL8-IRIgzO',
>> volumeGroupId='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g',
>> serial='SHITACHI_OPEN-V_5048', lunMapping='4',
>> vendorId='HITACHI', productId='OPEN-V', lunConnections='[]',
>> deviceSize='100', pvSize='0', peCount='797', peAllocatedCount='681',
>> vendorName='HITACHI', pathsDictionary='[sdf=true, sdu=true, sdk=true,
>> sdp=true]', pathsCapacity='[sdf=100, sdu=100, sdk=100, sdp=100]',
>> lunType='FCP', status='null', diskId='null', diskAlias='null',
>> 

Re: [ovirt-users] Slow booting host - restart loop

2017-09-06 Thread Martin Perina
On Wed, Sep 6, 2017 at 3:21 PM, Bernardo Juanicó  wrote:

> Hi, the same happens, maybe the password somehow changed, or the file is
> wrong? Is there a way to reset the password?
>
> [root@ovirt-engine ~]#
> ​​
> PGPASSWORD=CKBKMiSQrrPdq0w9RrWDXw psql -U engine -c "update vdc_options
> set option_value = '900' where option_name = '
> FenceQuietTimeBetweenOperationsInSec';" engine
> psql: FATAL:  Peer authentication failed for user "engine"
>
> ​
Hmm, coul you please try the above with specifying hostname and database
name:

​PGPASSWORD="CKBKMiSQrrPdq0w9RrWDXw" psql -h localhost -d engine -U engine
-c "update vdc_options set option_value = '900' where option_name =
'FenceQuietTimeBetweenOperationsInSec';"

​


> this is my 10-setup-database.conf file
>
> [root@ovirt-engine ~]# cat /etc/ovirt-engine/engine.conf.
> d/10-setup-database.conf
> ENGINE_DB_HOST="localhost"
> ENGINE_DB_PORT="5432"
> ENGINE_DB_USER="engine"
> ENGINE_DB_PASSWORD="CKBKMiSQrrPdq0w9RrWDXw"
> ENGINE_DB_DATABASE="engine"
> ENGINE_DB_SECURED="False"
> ENGINE_DB_SECURED_VALIDATION="False"
> ENGINE_DB_DRIVER="org.postgresql.Driver"
> ENGINE_DB_URL="jdbc:postgresql://localhost:5432/engine?sslfactory=org.
> postgresql.ssl.NonValidatingFactory"
>
> Regards,
>
> Bernardo
>
>
> PGP Key 
> Skype: mattraken
>
> 2017-09-06 4:32 GMT-03:00 Eli Mesika :
>
>> Hi
>>
>> try from shell prompt
>>
>> PGPASSWORD= psql -U engine -c "update vdc_options set
>> option_value = '900' where option_name = 
>> 'FenceQuietTimeBetweenOperationsInSec';"
>> engine
>>
>> On Tue, Sep 5, 2017 at 6:34 PM, Bernardo Juanicó 
>> wrote:
>>
>>> Hi Eli,
>>>
>>> I could not access the psql prompt, i tried with the credentials and
>>> values on /etc/ovirt-engine/engine.conf.d/10-setup-database.conf.
>>>
>>> I tried interactively and also with a .pgpass file containing:
>>> localhost:5432:engine:engine:PASSWORD
>>>
>>> And i get the following error:
>>>
>>> psql: FATAL:  Peer authentication failed for user "engine"
>>>
>>> Thanks!
>>>
>>> Bernardo
>>>
>>>
>>> PGP Key 
>>> Skype: mattraken
>>>
>>> 2017-09-05 12:14 GMT-03:00 Eli Mesika :
>>>
 Hi Bernardo

 I would like to suggest a workaround to this problem , can you please
 check that :

 We have a configuration value named FenceQuietTimeBetweenOperation
 sInSec.
 It controls the minimal timeout to wait between fence operation (stop,
 start),
 currently, it is defaulted to 180 sec , The key is not exposed to
 engine-config, so, I would suggest to

 1) Change this key value to 900 by running the following from psql
 prompt :

 update vdc_options set option_value = '900' where option_name =
 'FenceQuietTimeBetweenOperationsInSec';

 2) Restart the engine

 3) Repeat the scenario

 Now, the engine will require 15 min between fencing operations and your
 host can be up again without being fenced again.

 Please let me know if this workaround is working for you

 Thanks

 Eli

 On Tue, Sep 5, 2017 at 4:20 PM, Bernardo Juanicó 
 wrote:

> Martin, thanks for your reply, i was aware of the [1] BUG and the
> implemented solution, changing ServerRebootTimeout to 1200 didnt change a
> thing...
> Now i know about [2] and ill test the fix once it gets released.
>
> Regards,
>
> Bernardo
>
> PGP Key
> 
> Skype: mattraken
>
> 2017-09-05 8:23 GMT-03:00 Martin Perina :
>
>> Hi Bernardo,
>>
>> we have added timeout to wait until host is booted [1] in oVirt
>> 4.1.2. This timeout is by default 5 minutes, but it can be extended using
>> following command:
>>
>>engine-config -s ServerRebootTimeout=NNN
>>
>> where NNN is number of seconds you want to wait until host is booted
>> up.
>>
>> But be aware that you may be affected by [2], which we are currently
>> trying to fix.
>>
>> Regards
>>
>> Martin Perina
>>
>>
>> [1] https://bugzilla.redhat.com/show_bug.cgi?id=1423657
>> [2] https://bugzilla.redhat.com/show_bug.cgi?id=1477700
>>
>>
>> On Fri, Sep 1, 2017 at 7:54 PM, Bernardo Juanicó 
>> wrote:
>>
>>> Hi everyone,
>>>
>>> I installed 2 hosts on a new cluster and the servers take a really
>>> long to boot up (about 8 minutes).
>>>
>>> When a host crashes or is powered off the ovirt-manager starts it
>>> via power management, since the servers takes all that time to boot up 
>>> the
>>> ovirt-manager thinks it failed to start and proceeds to reboot it, 
>>> several
>>> times before giving up, when the server 

Re: [ovirt-users] Slow booting host - restart loop

2017-09-06 Thread Bernardo Juanicó
Hi, the same happens, maybe the password somehow changed, or the file is
wrong? Is there a way to reset the password?

[root@ovirt-engine ~]# PGPASSWORD=CKBKMiSQrrPdq0w9RrWDXw psql -U engine -c
"update vdc_options set option_value = '900' where option_name =
'FenceQuietTimeBetweenOperationsInSec';" engine
psql: FATAL:  Peer authentication failed for user "engine"

this is my 10-setup-database.conf file

[root@ovirt-engine ~]# cat
/etc/ovirt-engine/engine.conf.d/10-setup-database.conf
ENGINE_DB_HOST="localhost"
ENGINE_DB_PORT="5432"
ENGINE_DB_USER="engine"
ENGINE_DB_PASSWORD="CKBKMiSQrrPdq0w9RrWDXw"
ENGINE_DB_DATABASE="engine"
ENGINE_DB_SECURED="False"
ENGINE_DB_SECURED_VALIDATION="False"
ENGINE_DB_DRIVER="org.postgresql.Driver"
ENGINE_DB_URL="jdbc:postgresql://localhost:5432/engine?sslfactory=org.postgresql.ssl.NonValidatingFactory"

Regards,

Bernardo


PGP Key 
Skype: mattraken

2017-09-06 4:32 GMT-03:00 Eli Mesika :

> Hi
>
> try from shell prompt
>
> PGPASSWORD= psql -U engine -c "update vdc_options set
> option_value = '900' where option_name = 
> 'FenceQuietTimeBetweenOperationsInSec';"
> engine
>
> On Tue, Sep 5, 2017 at 6:34 PM, Bernardo Juanicó 
> wrote:
>
>> Hi Eli,
>>
>> I could not access the psql prompt, i tried with the credentials and
>> values on /etc/ovirt-engine/engine.conf.d/10-setup-database.conf.
>>
>> I tried interactively and also with a .pgpass file containing:
>> localhost:5432:engine:engine:PASSWORD
>>
>> And i get the following error:
>>
>> psql: FATAL:  Peer authentication failed for user "engine"
>>
>> Thanks!
>>
>> Bernardo
>>
>>
>> PGP Key 
>> Skype: mattraken
>>
>> 2017-09-05 12:14 GMT-03:00 Eli Mesika :
>>
>>> Hi Bernardo
>>>
>>> I would like to suggest a workaround to this problem , can you please
>>> check that :
>>>
>>> We have a configuration value named FenceQuietTimeBetweenOperation
>>> sInSec.
>>> It controls the minimal timeout to wait between fence operation (stop,
>>> start),
>>> currently, it is defaulted to 180 sec , The key is not exposed to
>>> engine-config, so, I would suggest to
>>>
>>> 1) Change this key value to 900 by running the following from psql
>>> prompt :
>>>
>>> update vdc_options set option_value = '900' where option_name =
>>> 'FenceQuietTimeBetweenOperationsInSec';
>>>
>>> 2) Restart the engine
>>>
>>> 3) Repeat the scenario
>>>
>>> Now, the engine will require 15 min between fencing operations and your
>>> host can be up again without being fenced again.
>>>
>>> Please let me know if this workaround is working for you
>>>
>>> Thanks
>>>
>>> Eli
>>>
>>> On Tue, Sep 5, 2017 at 4:20 PM, Bernardo Juanicó 
>>> wrote:
>>>
 Martin, thanks for your reply, i was aware of the [1] BUG and the
 implemented solution, changing ServerRebootTimeout to 1200 didnt change a
 thing...
 Now i know about [2] and ill test the fix once it gets released.

 Regards,

 Bernardo

 PGP Key
 
 Skype: mattraken

 2017-09-05 8:23 GMT-03:00 Martin Perina :

> Hi Bernardo,
>
> we have added timeout to wait until host is booted [1] in oVirt 4.1.2.
> This timeout is by default 5 minutes, but it can be extended using
> following command:
>
>engine-config -s ServerRebootTimeout=NNN
>
> where NNN is number of seconds you want to wait until host is booted
> up.
>
> But be aware that you may be affected by [2], which we are currently
> trying to fix.
>
> Regards
>
> Martin Perina
>
>
> [1] https://bugzilla.redhat.com/show_bug.cgi?id=1423657
> [2] https://bugzilla.redhat.com/show_bug.cgi?id=1477700
>
>
> On Fri, Sep 1, 2017 at 7:54 PM, Bernardo Juanicó 
> wrote:
>
>> Hi everyone,
>>
>> I installed 2 hosts on a new cluster and the servers take a really
>> long to boot up (about 8 minutes).
>>
>> When a host crashes or is powered off the ovirt-manager starts it via
>> power management, since the servers takes all that time to boot up the
>> ovirt-manager thinks it failed to start and proceeds to reboot it, 
>> several
>> times before giving up, when the server is finally started (about 20
>> minutes after the failure)
>>
>> I changed some engine variables with engine-config trying to set a
>> higher timeout, but the problem persists.
>>
>> Any ideas??
>>
>>
>> Regards,
>> Bernardo
>>
>>
>> PGP Key
>> 
>> Skype: mattraken
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> 

[ovirt-users] VMs going in to non-responding state

2017-09-06 Thread Satheesaran Sundaramoorthi
Hi All,

I have created a converged setup with cluster having both virt and gluster
capability. There are three hosts in this cluster, and this cluster also
has enabled 'native access to gluster domain' which enables VM to use
libgfapi access mechanism.

With this setup, I see VMs created landing up in non-reponding state after
sometime.

I have raised bug[1] for this issue.
Request for help with this issue

[1] - https://bugzilla.redhat.com/show_bug.cgi?id=1488863

Advance Thanks.

-- Satheesaran S ( sas )
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Cannot create snapshot_illegal disk

2017-09-06 Thread Benny Zlotnik
Hi Terry,

The disk in the snapshot appears to be in an illegal state. How long has it
been like this? Do you have logs from when it happened?

On Tue, Sep 5, 2017 at 8:52 PM, Terry hey  wrote:

> Dear all,
> Thank you for your time to read this post first.
> In the same host, there are four virtual machines. Only one virtual
> machine that i cannot create a snapshot for backup usage. I have no idea
> what is happening and how to solve it. Would you guys help me and give some
> suggestion to me?
>
> Please kindly check the following links. I think those information
> including two picture of oVirt web manager, engine.log and vdsm.log, would
> be useful for finding the issue.
>
> *engine.log*
> https://drive.google.com/open?id=0B8dxamAkVEYdeFBOdlFUbnY1cFU
> *vdsm.log*
> https://drive.google.com/open?id=0B8dxamAkVEYddFh6dzc0VkVRMk0
> *Two pictures of oVirt web manger:*
> *Before created the snapshot:*
> https://drive.google.com/open?id=0B8dxamAkVEYda2d0SzVvc3ZMTFU
> *After created the snapshot:*
> https://drive.google.com/open?id=0B8dxamAkVEYdV3lUVUJ0Zy1LeGc
>
> I am really looking forward for you guys' reply. Thank you again for your
> time to assist me.
>
> Regards,
> Terry
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] oVirt selft-hosted with NFS on top gluster

2017-09-06 Thread Abi Askushi
Hi All,

I've playing with ovirt self hosted engine setup and I even use it to
production for several VM. The setup I have is 3 server with gluster
storage in replica 2+1 (1 arbiter).
The data storage domain where VMs are stored is mounted with gluster
through ovirt. The performance I get for the VMs is very low and I was
thinking to switch and mount the same storage through NFS instead of
glusterfs.

The only think I am hesitant is how can I ensure high availability of the
storage when I loose one server? I was thinking to have at /etc/hosts sth
like below:

10.100.100.1 nfsmount
10.100.100.2 nfsmount
10.100.100.3 nfsmount

then use nfsmount as the server name when adding this domain through ovirt
GUI.
Are there any other more elegant solutions? What do you do for such cases?
Note: gluster has the back-vol-file option which provides a lean way to
have redundancy on the mount point and I am using this when mounting with
glusterfs.

Thanx
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] cpu, core and thread mappings

2017-09-06 Thread Andrew Nesbit
> Are there any documents going more deeply in these sort of considerations?

This is exactly what I have been searching for lately, too.  Please let me know 
if you find anything (or blog posts, forums, books, etc).  Thank you!

Sent from my mobile phone

> On 6 Sep 2017, at 12:47, Gianluca Cecchi  wrote:
> 
> Hello,
> I was talking with a guy expert in VMware and discussing performance of VMs 
> in respect of virtual cpus assigned to them in relation with mapping with the 
> real hw of the hypervisor underneath.
> 
> One of the topics was numa usage and its overheads in case of a "too" big VM, 
> in terms of both number of vcpus and memory amount.
> Eg: 
> suppose host has 2 intel based sockets, with 6 cores and HT enabled and has 
> 96Gb of ram (distributed 48+48 between the 2 processors)
> suppose I configure a VM with 16 vcpus (2:4:2): would be the mapping 
> respected at physical level or only a sort of "hint" for the hypervisor?
> Can I say that it would perform better if I configure it 12 vcpus and mapping 
> 1:6:2, because it can stay all inside one cpu?
> 
> And what if I define a VM with 52Gb of ram? Can I say that it would perform 
> in general better if I try to get it all in one cpu related memory slots (eg 
> not more than 48Gb in my example)?
> 
> Are there any documents going more deeply in these sort of considerations?
> 
> Also, if one goes and sizes so that the biggest VM is able to all-stay inside 
> one cpu-memory, does it make sense to say that it will perform better in this 
> scenario a cluster composed by 4 nodes, each one with 1 socket and 48Gb of 
> memory instead of a cluster of 2 nodes, each one with 2 sockets and 96Gb of 
> ram?
> 
> Hope I have clarified my questions/doubts.
> 
> 
> Thanks in advance for any insight,
> Gianluca  
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] cpu, core and thread mappings

2017-09-06 Thread Gianluca Cecchi
Hello,
I was talking with a guy expert in VMware and discussing performance of VMs
in respect of virtual cpus assigned to them in relation with mapping with
the real hw of the hypervisor underneath.

One of the topics was numa usage and its overheads in case of a "too" big
VM, in terms of both number of vcpus and memory amount.
Eg:
suppose host has 2 intel based sockets, with 6 cores and HT enabled and has
96Gb of ram (distributed 48+48 between the 2 processors)
suppose I configure a VM with 16 vcpus (2:4:2): would be the mapping
respected at physical level or only a sort of "hint" for the hypervisor?
Can I say that it would perform better if I configure it 12 vcpus and
mapping 1:6:2, because it can stay all inside one cpu?

And what if I define a VM with 52Gb of ram? Can I say that it would perform
in general better if I try to get it all in one cpu related memory slots
(eg not more than 48Gb in my example)?

Are there any documents going more deeply in these sort of considerations?

Also, if one goes and sizes so that the biggest VM is able to all-stay
inside one cpu-memory, does it make sense to say that it will perform
better in this scenario a cluster composed by 4 nodes, each one with 1
socket and 48Gb of memory instead of a cluster of 2 nodes, each one with 2
sockets and 96Gb of ram?

Hope I have clarified my questions/doubts.


Thanks in advance for any insight,
Gianluca
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Failed to import the Hosted Engine Storage Domain

2017-09-06 Thread Arsène Gschwind



On 09/05/2017 09:02 AM, Arsène Gschwind wrote:




On 09/04/2017 07:59 PM, Simone Tiraboschi wrote:


On Mon, Sep 4, 2017 at 7:32 PM, Arsène Gschwind 
> wrote:




On 09/04/2017 06:32 PM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 6:24 PM, Arsène Gschwind
>
wrote:



On 09/04/2017 02:51 PM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 2:21 PM, Arsène Gschwind
> wrote:



On 09/04/2017 02:01 PM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 1:55 PM, Arsène Gschwind
> wrote:



On 09/04/2017 01:52 PM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 12:23 PM, Arsène Gschwind
> wrote:

Hi Simone,


On 09/04/2017 11:14 AM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 10:56 AM, Arsène
Gschwind > wrote:

Hi Didi,


On 09/04/2017 10:15 AM, Yedidyah Bar
David wrote:

On Mon, Sep 4, 2017 at 10:16 AM, Arsène Gschwind

  wrote:

Hi all,

A while ago I had some problem with hosted-engine 
network which wasn't set
correctly at deploy time, so I finally decided to 
redeploy the hosted engine
in the hope the network will be set correctly this 
time. I've followed this
procedure:

Stop all VMs
Full backup of HE DB and export to safe place
Cleanup HE storage 
followinghttps://access.redhat.com/solutions/2121581

Reboot Hosts
Re-deploy HE until DB recovery
Recover DB adding the following param:
  --he-remove-storage-vm Removes the 
hosted-engine storage
domain, all its entities and the hosted-engine VM 
during restore.
  --he-remove-hosts  Removes all the 
hosted-engine hosts
during restore.

Finalize HE deployment.

Everything did run without errors and I'm able to 
access Web UI.

But now I don't see my HE VM and its respective Storage 
Domain, the logs
says it isn't able to import it. I see all other SD and 
I'm able to manage
my VMs as before.

Please find attached engine.log

I think this is your problem:

2017-09-04 03:26:14,272+02 INFO

[org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-24) [2383eaa0] There 
are existing luns
in the system which are part of VG id
'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'

I don't see a VG with this ID, here the
IDs I see on the hosts:

VG #PV #LV #SN Attr   VSize VFree
6b62cc06-fc44-4c38-af6d-bfd9cbe73246 1 
10   0 wz--n- 99.62g 14.50g
b0414c06-d984-4001-a998-fd9a2e79fb83 2 
70   0 wz--n- 10.00t 2.31t
b2e30961-7cff-4cca-83d6-bee3a4f890ee 2 
47   0 wz--n-  5.27t 2.50t



Could you please repeat the command on host
adm-kvmh70 ?

2017-09-04 09:04:18,163+02 INFO
 [org.ovirt.engine.core.bll.st

orage.domain.ImportHostedEngineStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-34)
[247a3718] Running command:
ImportHostedEngineStorageDomainCommand
internal: true.
2017-09-04 09:04:18,189+02 INFO
 
[org.ovirt.engine.core.vdsbroker.vdsbroker.GetVGInfoVDSCommand]
(org.ovirt.thread.pool-6-thread-34)
[7d2e6cb2] START,

Re: [ovirt-users] Changes in Virt SIG in upcoming CentOS 1708

2017-09-06 Thread Eyal Edri
On Wed, Sep 6, 2017 at 10:33 AM, Sandro Bonazzola 
wrote:

> Hi,
> with upcoming CentOS 1708 all the SIGs have been asked to do some cleanup
> of old / unsupported stuff.
> As part of the cleanup, ovirt 3.5, 3.6 and 4.0 will be dropped from
> http://mirror.centos.org/centos/7/virt/x86_64/
>

This might break the upgrade suite from 4.0->master ( upstream from
prev-release ), should we remove it from CI and not test upgrades from
4.0->master?
It will also break upgrade from 4.0->4.1 suite, should we remove it as
well?


>
>
> Also, as part of the cleanup a few packages previously tagged in CentOS
> VIrt SIG repos and now available in CentOS base / optional / extras repo
> have been untagged and won't be in http://mirror.centos.org/
> centos/7/virt/x86_64/ when the repo will be re-generated from tags.
> As a side effect, you may see cockpit related packages not available if
> you're running oVirt on RHEL 7.4 instead of CentOS 7.1708 because on CentOS
> 7 extras channel is enabled by default while on RHEL you need to enable it
> manually.
>
> Thanks,
> --
>
> SANDRO BONAZZOLA
>
> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R
>
> Red Hat EMEA 
> 
> TRIED. TESTED. TRUSTED. 
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>


-- 

Eyal edri


ASSOCIATE MANAGER

RHV DevOps

EMEA VIRTUALIZATION R


Red Hat EMEA 
 TRIED. TESTED. TRUSTED. 
phone: +972-9-7692018
irc: eedri (on #tlv #rhev-dev #rhev-integ)
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] [QA] fluentd-0.12.39-1.el7

2017-09-06 Thread Sandro Bonazzola
Hi,
CentOS OpsTools SIG tagged fluentd-0.12.39-1.el7
 and
rubygem-cool.io-1.5.1-1.el7
 for testing and will
probably move soon to release repo.
Before it happens, if you're already using it can you please provide
feedback on its goodness on oVirt environment?
I don't see critical changes from 0.12.37 but some feedback would help.

Testing yum repository is
https://buildlogs.centos.org/centos/7/opstools/x86_64/ and is already used
in our CI without any breakage reported so far.

Thanks to whoever volunteer for this testing.

-- 

SANDRO BONAZZOLA

ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R

Red Hat EMEA 

TRIED. TESTED. TRUSTED. 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Cannot create snapshot_illegal disk

2017-09-06 Thread Terry hey
Dear all,
Thank you for your time to read this post first.
In the same host, there are four virtual machines. Only one virtual machine
that i cannot create a snapshot for backup usage. I have no idea what is
happening and how to solve it. Would you guys help me and give some
suggestion to me?

Please kindly check the following links. I think those information
including two picture of oVirt web manager, engine.log and vdsm.log, would
be useful for finding the issue.

*engine.log*
https://drive.google.com/open?id=0B8dxamAkVEYdeFBOdlFUbnY1cFU
*vdsm.log*
https://drive.google.com/open?id=0B8dxamAkVEYddFh6dzc0VkVRMk0
*Two pictures of oVirt web manger:*
*Before created the snapshot:*
https://drive.google.com/open?id=0B8dxamAkVEYda2d0SzVvc3ZMTFU
*After created the snapshot:*
https://drive.google.com/open?id=0B8dxamAkVEYdV3lUVUJ0Zy1LeGc

I am really looking forward for you guys' reply. Thank you again for your
time to assist me.

Regards,
Terry
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Changes in Virt SIG in upcoming CentOS 1708

2017-09-06 Thread Sandro Bonazzola
Hi,
with upcoming CentOS 1708 all the SIGs have been asked to do some cleanup
of old / unsupported stuff.
As part of the cleanup, ovirt 3.5, 3.6 and 4.0 will be dropped from
http://mirror.centos.org/centos/7/virt/x86_64/

Also, as part of the cleanup a few packages previously tagged in CentOS
VIrt SIG repos and now available in CentOS base / optional / extras repo
have been untagged and won't be in
http://mirror.centos.org/centos/7/virt/x86_64/ when the repo will be
re-generated from tags.
As a side effect, you may see cockpit related packages not available if
you're running oVirt on RHEL 7.4 instead of CentOS 7.1708 because on CentOS
7 extras channel is enabled by default while on RHEL you need to enable it
manually.

Thanks,
-- 

SANDRO BONAZZOLA

ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R

Red Hat EMEA 

TRIED. TESTED. TRUSTED. 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Slow booting host - restart loop

2017-09-06 Thread Eli Mesika
Hi

try from shell prompt

PGPASSWORD= psql -U engine -c "update vdc_options set
option_value = '900' where option_name =
'FenceQuietTimeBetweenOperationsInSec';"
engine

On Tue, Sep 5, 2017 at 6:34 PM, Bernardo Juanicó  wrote:

> Hi Eli,
>
> I could not access the psql prompt, i tried with the credentials and
> values on /etc/ovirt-engine/engine.conf.d/10-setup-database.conf.
>
> I tried interactively and also with a .pgpass file containing:
> localhost:5432:engine:engine:PASSWORD
>
> And i get the following error:
>
> psql: FATAL:  Peer authentication failed for user "engine"
>
> Thanks!
>
> Bernardo
>
>
> PGP Key 
> Skype: mattraken
>
> 2017-09-05 12:14 GMT-03:00 Eli Mesika :
>
>> Hi Bernardo
>>
>> I would like to suggest a workaround to this problem , can you please
>> check that :
>>
>> We have a configuration value named FenceQuietTimeBetweenOperationsInSec.
>> It controls the minimal timeout to wait between fence operation (stop,
>> start),
>> currently, it is defaulted to 180 sec , The key is not exposed to
>> engine-config, so, I would suggest to
>>
>> 1) Change this key value to 900 by running the following from psql prompt
>> :
>>
>> update vdc_options set option_value = '900' where option_name =
>> 'FenceQuietTimeBetweenOperationsInSec';
>>
>> 2) Restart the engine
>>
>> 3) Repeat the scenario
>>
>> Now, the engine will require 15 min between fencing operations and your
>> host can be up again without being fenced again.
>>
>> Please let me know if this workaround is working for you
>>
>> Thanks
>>
>> Eli
>>
>> On Tue, Sep 5, 2017 at 4:20 PM, Bernardo Juanicó 
>> wrote:
>>
>>> Martin, thanks for your reply, i was aware of the [1] BUG and the
>>> implemented solution, changing ServerRebootTimeout to 1200 didnt change a
>>> thing...
>>> Now i know about [2] and ill test the fix once it gets released.
>>>
>>> Regards,
>>>
>>> Bernardo
>>>
>>> PGP Key 
>>> Skype: mattraken
>>>
>>> 2017-09-05 8:23 GMT-03:00 Martin Perina :
>>>
 Hi Bernardo,

 we have added timeout to wait until host is booted [1] in oVirt 4.1.2.
 This timeout is by default 5 minutes, but it can be extended using
 following command:

engine-config -s ServerRebootTimeout=NNN

 where NNN is number of seconds you want to wait until host is booted up.

 But be aware that you may be affected by [2], which we are currently
 trying to fix.

 Regards

 Martin Perina


 [1] https://bugzilla.redhat.com/show_bug.cgi?id=1423657
 [2] https://bugzilla.redhat.com/show_bug.cgi?id=1477700


 On Fri, Sep 1, 2017 at 7:54 PM, Bernardo Juanicó 
 wrote:

> Hi everyone,
>
> I installed 2 hosts on a new cluster and the servers take a really
> long to boot up (about 8 minutes).
>
> When a host crashes or is powered off the ovirt-manager starts it via
> power management, since the servers takes all that time to boot up the
> ovirt-manager thinks it failed to start and proceeds to reboot it, several
> times before giving up, when the server is finally started (about 20
> minutes after the failure)
>
> I changed some engine variables with engine-config trying to set a
> higher timeout, but the problem persists.
>
> Any ideas??
>
>
> Regards,
> Bernardo
>
>
> PGP Key
> 
> Skype: mattraken
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>

>>>
>>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Slow booting host - restart loop

2017-09-06 Thread Artyom Lukianov
It can be a result of the bug
https://bugzilla.redhat.com/show_bug.cgi?id=1477700.

Best Regards

On Tue, Sep 5, 2017 at 6:34 PM, Bernardo Juanicó  wrote:

> Hi Eli,
>
> I could not access the psql prompt, i tried with the credentials and
> values on /etc/ovirt-engine/engine.conf.d/10-setup-database.conf.
>
> I tried interactively and also with a .pgpass file containing:
> localhost:5432:engine:engine:PASSWORD
>
> And i get the following error:
>
> psql: FATAL:  Peer authentication failed for user "engine"
>
> Thanks!
>
> Bernardo
>
>
> PGP Key 
> Skype: mattraken
>
> 2017-09-05 12:14 GMT-03:00 Eli Mesika :
>
>> Hi Bernardo
>>
>> I would like to suggest a workaround to this problem , can you please
>> check that :
>>
>> We have a configuration value named FenceQuietTimeBetweenOperationsInSec.
>> It controls the minimal timeout to wait between fence operation (stop,
>> start),
>> currently, it is defaulted to 180 sec , The key is not exposed to
>> engine-config, so, I would suggest to
>>
>> 1) Change this key value to 900 by running the following from psql prompt
>> :
>>
>> update vdc_options set option_value = '900' where option_name =
>> 'FenceQuietTimeBetweenOperationsInSec';
>>
>> 2) Restart the engine
>>
>> 3) Repeat the scenario
>>
>> Now, the engine will require 15 min between fencing operations and your
>> host can be up again without being fenced again.
>>
>> Please let me know if this workaround is working for you
>>
>> Thanks
>>
>> Eli
>>
>> On Tue, Sep 5, 2017 at 4:20 PM, Bernardo Juanicó 
>> wrote:
>>
>>> Martin, thanks for your reply, i was aware of the [1] BUG and the
>>> implemented solution, changing ServerRebootTimeout to 1200 didnt change a
>>> thing...
>>> Now i know about [2] and ill test the fix once it gets released.
>>>
>>> Regards,
>>>
>>> Bernardo
>>>
>>> PGP Key 
>>> Skype: mattraken
>>>
>>> 2017-09-05 8:23 GMT-03:00 Martin Perina :
>>>
 Hi Bernardo,

 we have added timeout to wait until host is booted [1] in oVirt 4.1.2.
 This timeout is by default 5 minutes, but it can be extended using
 following command:

engine-config -s ServerRebootTimeout=NNN

 where NNN is number of seconds you want to wait until host is booted up.

 But be aware that you may be affected by [2], which we are currently
 trying to fix.

 Regards

 Martin Perina


 [1] https://bugzilla.redhat.com/show_bug.cgi?id=1423657
 [2] https://bugzilla.redhat.com/show_bug.cgi?id=1477700


 On Fri, Sep 1, 2017 at 7:54 PM, Bernardo Juanicó 
 wrote:

> Hi everyone,
>
> I installed 2 hosts on a new cluster and the servers take a really
> long to boot up (about 8 minutes).
>
> When a host crashes or is powered off the ovirt-manager starts it via
> power management, since the servers takes all that time to boot up the
> ovirt-manager thinks it failed to start and proceeds to reboot it, several
> times before giving up, when the server is finally started (about 20
> minutes after the failure)
>
> I changed some engine variables with engine-config trying to set a
> higher timeout, but the problem persists.
>
> Any ideas??
>
>
> Regards,
> Bernardo
>
>
> PGP Key
> 
> Skype: mattraken
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>

>>>
>>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users