Re: [ovirt-users] Issue with hosted engine restore.

2017-11-13 Thread Yedidyah Bar David
On Mon, Nov 13, 2017 at 11:58 PM, Krzysztof Wajda  wrote:
> Hello,
>
> I have to restore Hosted Engine on another host (completely new hardware).
> Based on this
> https://www.ovirt.org/documentation/self-hosted/chap-Backing_up_and_Restoring_an_EL-Based_Self-Hosted_Environment/
> is not clear for me if vm's will be rebooted during synchronization hosts
> with engine ?

They should not be rebooted automatically, but you might need to do
this yourself, see below.

>
> I have 5 hosts + 1 completely fresh. On host1 I have HE and there is no vm's
> on other 4 (host1-4) there are around 400 vm which can't be rebooted. Host5
> for restore HE.

Please provide more details about your backup/restore flow.
What died (storage? hosts? data?), what are you going to restore,
how, etc.

Which hosts are hosted-engine hosts. Do they have running VMs.

We are working on updating the documentation, but it will take some time.

For now, you should assume that the safest way is to pass during restore,
to engine-backup, '--he-remove-storage-vm' and '--he-remove-hosts'. This
will remove from the engine all the hosted-engine hosts and storage. So
when you add the hosts back, you'll have to somehow power off the VMs
there - the engine will refuse to add them with running VMs. If you do
not want to use these options, you should plan carefully and test.

See also:

https://bugzilla.redhat.com/show_bug.cgi?id=1235200
https://bugzilla.redhat.com/show_bug.cgi?id=1240466
https://bugzilla.redhat.com/show_bug.cgi?id=1441322

Best regards,
-- 
Didi
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] 4.1 engine-iso-uploader / root password glitch

2017-11-13 Thread Yedidyah Bar David
On Mon, Nov 13, 2017 at 11:47 PM, andre...@starlett.lv
 wrote:
> Hi,
>
> Here are logs:
>
> engine-iso-uploader list
> Please provide the REST API password for the admin@internal oVirt Engine
> user (CTRL+D to abort):
> ISO Storage Domain Name   | ISO Domain Status
> iso   | ok
>
> engine-iso-uploader --ssh-user=root --iso-domain=iso upload

Can you please run it with --verbose? Thanks.

> /vmhosts/virtio-win.iso
> Please provide the REST API password for the admin@internal oVirt Engine
> user (CTRL+D to abort):
> Uploading, please wait...
> INFO: Start uploading /vmhosts/virtio-win.iso
> r...@vm-hostengine.com's password:
> r...@vm-hostengine.com's password: (endless)
>
> iso-uploader log:
> 2017-11-13 23:38:16::INFO::engine-iso-uploader::1033::root:: Start
> uploading /vmhosts/virtio-win.iso
> ... and nothing else
>
> tail -n 5000 messages | grep ssh
> Gives nothing useful except
> vm-hostengine ovirt-vmconsole-proxy-sshd:
> /usr/share/ovirt-vmconsole/ovirt-vmconsole-proxy/ovirt-vmconsole-proxy-sshd/sshd_config
> line 22: Deprecated option RSAAuthentication
> which have no relation to upload.
>
>
> On 11/13/2017 09:54 AM, Yedidyah Bar David wrote:
>> On Mon, Nov 13, 2017 at 8:57 AM, andre...@starlett.lv
>>  wrote:
>>> On 11/13/2017 05:36 AM, Yihui Zhao wrote:
>>>
>>> can you try the admin password?
>>>
>>>
>>> already did, same result.
>>>
>>> On Mon, Nov 13, 2017 at 3:10 AM, andre...@starlett.lv 
>>> wrote:
 Hi,

 I'm trying to upload iso with this coomand.
 engine-iso-uploader --ssh-user=root --iso-domain=iso upload suse.iso

 Please provide the REST API password for the admin@internal oVirt Engine
 user (CTRL+D to abort):
 This go OK.

 However, then it asks root password, I enter it, then it asks again and
 again. Root password is correct for sure, becuase I can coonect vis ssh
 from terminal.

 How to fix this problem?
>> Can you please share the log? Thanks.
>>
 May be its possible just to copy files manually?
>> It is. Locate the iso domain on your storage server.
>> Inside it you'll find a directory whose name is a random
>> uuid, inside it 'images', and inside it a directory named:
>> '----'.
>> You can put your iso files inside that one. Make sure they
>> are readable by user:group 36:36.
>>
 Thanks in advance.
 Andrei
 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users
>>>
>>>
>>>
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>>
>>
>>
>



-- 
Didi
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Enabling libgfapi disk access with oVirt 4.2

2017-11-13 Thread Kasturi Narra
yes, it  does work with 4.1.7.6 version

On Tue, Nov 14, 2017 at 4:49 AM, Bryan Sockel  wrote:

> Is libgfapi supposed to be working in 4.1.7.6?
> Bryan
>
> -Original Message-
> From: Alessandro De Salvo 
> To: users@ovirt.org
> Date: Thu, 9 Nov 2017 09:35:01 +0100
> Subject: Re: [ovirt-users] Enabling libgfapi disk access with oVirt 4.2
>
> Hi again,
> OK, tried to stop all the vms, except the engine, set engine-config -s
> LibgfApiSupported=true (for 4.2 only) and restarted the engine.
> When I tried restarting the VMs they are still not using gfapi, so it does
> not seem to help.
> Cheers,
>
> Alessandro
>
>
> Il 09/11/17 09:12, Alessandro De Salvo ha scritto:
>
>
> Hi,
> where should I enable gfapi via the UI?
> The only command I tried was engine-config -s LibgfApiSupported=true but
> the result is what is shown in my output below, so it’s set to true for
> v4.2. Is it enough?
> I’ll try restarting the engine. Is it really needed to stop all the VMs
> and restart them all? Of course this is a test setup and I can do it, but
> for production clusters in the future it may be a problem.
> Thanks,
>
>Alessandro
>
> Il giorno 09 nov 2017, alle ore 07:23, Kasturi Narra 
> ha scritto:
>
>
> Hi ,
>
> The procedure to enable gfapi is below.
>
> 1) stop all the vms running
> 2) Enable gfapi via UI or using engine-config command
> 3) Restart ovirt-engine service
> 4) start the vms.
>
> Hope you have not missed any !!
>
> Thanks
> kasturi
>
> On Wed, Nov 8, 2017 at 11:58 PM, Alessandro De Salvo <
> alessandro.desa...@roma1.infn.it> wrote:
>>
>> Hi,
>>
>> I'm using the latest 4.2 beta release and want to try the gfapi access,
>> but I'm currently failing to use it.
>>
>> My test setup has an external glusterfs cluster v3.12, not managed by
>> oVirt.
>>
>> The compatibility flag is correctly showing gfapi should be enabled with
>> 4.2:
>>
>> # engine-config -g LibgfApiSupported
>> LibgfApiSupported: false version: 3.6
>> LibgfApiSupported: false version: 4.0
>> LibgfApiSupported: false version: 4.1
>> LibgfApiSupported: true version: 4.2
>>
>> The data center and cluster have the 4.2 compatibility flags as well.
>>
>> However, when starting a VM with a disk on gluster I can still see the
>> disk is mounted via fuse.
>>
>> Any clue of what I'm still missing?
>>
>> Thanks,
>>
>>
>>Alessandro
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>
>
>
> ___
> Users mailing listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] how to redo oVirt cluster?

2017-11-13 Thread Luca 'remix_tj' Lorenzetto
Hello Rudi,

I think that uninstalling ovirt-engine and removing vdsm from hosts should
be enough. Pay attention to cleaning up engine db which contains all the
engine data.

Luca


Il 14 nov 2017 7:20 AM, "Rudi Ahlers"  ha scritto:

> Hi,
>
> I have setup an oVirt cluster and did some tests. But how do I redo
> everything, without reinstalling CentOS as well?
> Would it be as simple as uninstalling all the ovirt? Or do I need to
> manually delete some config files and other traces off the install as well?
>
> --
> Kind Regards
> Rudi Ahlers
> Website: http://www.rudiahlers.co.za
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] how to redo oVirt cluster?

2017-11-13 Thread Rudi Ahlers
Hi,

I have setup an oVirt cluster and did some tests. But how do I redo
everything, without reinstalling CentOS as well?
Would it be as simple as uninstalling all the ovirt? Or do I need to
manually delete some config files and other traces off the install as well?

-- 
Kind Regards
Rudi Ahlers
Website: http://www.rudiahlers.co.za
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Unable to clone a vm

2017-11-13 Thread Budur Nagaraju
HI

When I clone a vm from the existing vm , getting the error "
https://pastebin.com/2p90TU2y;.
Below is the attached logs ,can some one help me on this ?

https://pastebin.com/2p90TU2y

Thanks,
Nagaraju
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Error during SSO authentication Cannot authenticate user 'admin@internal'

2017-11-13 Thread Sverker Abrahamsson
Since upgrading my test lab to ovirt 4.2 I can't get ovirt-provider-ovn 
to work. From ovirt-provider-ovn.log:


2017-11-14 00:40:15,795   Request: POST : /v2.0///tokens
2017-11-14 00:40:15,795   Request body:
{
  "auth" : {
    "passwordCredentials" : {
  "username" : "admin@internal",
  "password" : "x"
    }
  }
}
2017-11-14 00:40:15,819   Starting new HTTPS connection (1): h2-int
2017-11-14 00:40:20,829   "POST /ovirt-engine/sso/oauth/token HTTP/1.1" 
400 118
2017-11-14 00:40:20,830   Error during SSO authentication Cannot 
authenticate user 'admin@internal': The username or password is 
incorrect.. : access_deniedNone

Traceback (most recent call last):
  File "/usr/share/ovirt-provider-ovn/handlers/base_handler.py", line 
119, in _handle_request

    method, path_parts, content)
  File "/usr/share/ovirt-provider-ovn/handlers/selecting_handler.py", 
line 177, in handle_request

    handler, content, parameters
  File "/usr/share/ovirt-provider-ovn/handlers/keystone.py", line 28, 
in call_response_handler

    return response_handler(content, parameters)
  File "/usr/share/ovirt-provider-ovn/handlers/keystone_responses.py", 
line 58, in post_tokens

    user_password=user_password)
  File "/usr/share/ovirt-provider-ovn/auth/plugin_facade.py", line 26, 
in create_token

    return auth.core.plugin.create_token(user_at_domain, user_password)
  File "/usr/share/ovirt-provider-ovn/auth/plugins/ovirt/plugin.py", 
line 48, in create_token

    timeout=self._timeout())
  File "/usr/share/ovirt-provider-ovn/auth/plugins/ovirt/sso.py", line 
62, in create_token

    username, password, engine_url, ca_file, timeout)
  File "/usr/share/ovirt-provider-ovn/auth/plugins/ovirt/sso.py", line 
54, in wrapper

    _check_for_error(response)
  File "/usr/share/ovirt-provider-ovn/auth/plugins/ovirt/sso.py", line 
168, in _check_for_error

    result['error'], details))
Unauthorized: Error during SSO authentication Cannot authenticate user 
'admin@internal': The username or password is incorrect.. : 
access_deniedNone


And in engine.log:

2017-11-14 00:40:20,828+01 ERROR 
[org.ovirt.engine.core.sso.utils.SsoUtils] (default task-16) [] 
OAuthException access_denied: Cannot authenticate user 'admin@internal': 
The username or password is incorrect..


The password in the request is the same as used to log in to the admin 
portal and works fine there.


/Sverker

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Enabling libgfapi disk access with oVirt 4.2

2017-11-13 Thread Bryan Sockel
Is libgfapi supposed to be working in 4.1.7.6?
Bryan 
-Original Message-
From: Alessandro De Salvo 
To: users@ovirt.org
Date: Thu, 9 Nov 2017 09:35:01 +0100
Subject: Re: [ovirt-users] Enabling libgfapi disk access with oVirt 4.2

Hi again,

OK, tried to stop all the vms, except the engine, set engine-config -s 
LibgfApiSupported=true (for 4.2 only) and restarted the engine.

When I tried restarting the VMs they are still not using gfapi, so it does 
not seem to help.

Cheers,



Alessandro




Il 09/11/17 09:12, Alessandro De Salvo ha scritto:

Hi,
where should I enable gfapi via the UI?
The only command I tried was engine-config -s LibgfApiSupported=true but the 
result is what is shown in my output below, so it’s set to true for v4.2. 
Is it enough?
I’ll try restarting the engine. Is it really needed to stop all the VMs 
and restart them all? Of course this is a test setup and I can do it, but 
for production clusters in the future it may be a problem.
Thanks,

   Alessandro

Il giorno 09 nov 2017, alle ore 07:23, Kasturi Narra  ha 
scritto:

Hi ,

The procedure to enable gfapi is below.

1) stop all the vms running
2) Enable gfapi via UI or using engine-config command
3) Restart ovirt-engine service
4) start the vms.

Hope you have not missed any !!

Thanks
kasturi 

On Wed, Nov 8, 2017 at 11:58 PM, Alessandro De Salvo 
 wrote:
Hi,

I'm using the latest 4.2 beta release and want to try the gfapi access, but 
I'm currently failing to use it.

My test setup has an external glusterfs cluster v3.12, not managed by oVirt.

The compatibility flag is correctly showing gfapi should be enabled with 
4.2:

# engine-config -g LibgfApiSupported
LibgfApiSupported: false version: 3.6
LibgfApiSupported: false version: 4.0
LibgfApiSupported: false version: 4.1
LibgfApiSupported: true version: 4.2

The data center and cluster have the 4.2 compatibility flags as well.

However, when starting a VM with a disk on gluster I can still see the disk 
is mounted via fuse.

Any clue of what I'm still missing?

Thanks,


   Alessandro

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] how to clean stuck task

2017-11-13 Thread Gianluca Cecchi
On Mon, Nov 13, 2017 at 8:53 PM, Fred Rolland  wrote:

> For cleaning these "Tasks" entries (which are actually Jobs in the DB),
> you will need to use taskcleaner.sh.
> The unlock_entity.sh is for removing locks on entities.
>

ok. Thanks for clarifying the concept.
But from the inline help of taskcleaner.sh it is not so clear to me which
option makes it display vs remove...
Perhaps without options it only display? or with "-T" it first displays
asking if I want to remove "all comands that have running tasks"?
How can I only display?


> Before running the task cleaner, you need to be sure you actually don't
> have any tasks running in the SPM.
>

How can I check this?
Is still valid this command I found on a previous (around 2014) thread and
executed on the SPM (as shown in web admin portal) host right now?

[root@ov300 vdsm]# vdsClient -s 0 getAllTasksStatuses
{'status': {'message': 'OK', 'code': 0}, 'allTasksStatus': {}}

[root@ov300 vdsm]#

Does it mean no task?



> Regarding the snapshot creation on 'apex", can you provide the logs around
> the time of the operation?
>

engine.log is here:
https://drive.google.com/file/d/1nD0p90ZvwphrEafgKZkCCquNu_
vVY8p8/view?usp=sharing

vdsm log of the SPM host in .xz format here:
https://drive.google.com/file/d/1Zzk04W6400wPh1sM8O4WyFDAGQaqa
UZK/view?usp=sharing

Thanks,
Gianluca
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Issue with hosted engine restore.

2017-11-13 Thread Krzysztof Wajda
Hello,

I have to restore Hosted Engine on another host (completely new hardware).
Based on this
https://www.ovirt.org/documentation/self-hosted/chap-Backing_up_and_Restoring_an_EL-Based_Self-Hosted_Environment/
is not clear for me if vm's will be rebooted during synchronization hosts
with engine ?

I have 5 hosts + 1 completely fresh. On host1 I have HE and there is no
vm's on other 4 (host1-4) there are around 400 vm which can't be rebooted.
Host5 for restore HE.

Kris
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] 4.1 engine-iso-uploader / root password glitch

2017-11-13 Thread andre...@starlett.lv
Hi,

Here are logs:

engine-iso-uploader list
Please provide the REST API password for the admin@internal oVirt Engine
user (CTRL+D to abort):
ISO Storage Domain Name   | ISO Domain Status
iso   | ok

engine-iso-uploader --ssh-user=root --iso-domain=iso upload
/vmhosts/virtio-win.iso
Please provide the REST API password for the admin@internal oVirt Engine
user (CTRL+D to abort):
Uploading, please wait...
INFO: Start uploading /vmhosts/virtio-win.iso
r...@vm-hostengine.com's password:
r...@vm-hostengine.com's password: (endless)

iso-uploader log:
2017-11-13 23:38:16::INFO::engine-iso-uploader::1033::root:: Start
uploading /vmhosts/virtio-win.iso
... and nothing else

tail -n 5000 messages | grep ssh
Gives nothing useful except
vm-hostengine ovirt-vmconsole-proxy-sshd:
/usr/share/ovirt-vmconsole/ovirt-vmconsole-proxy/ovirt-vmconsole-proxy-sshd/sshd_config
line 22: Deprecated option RSAAuthentication
which have no relation to upload.


On 11/13/2017 09:54 AM, Yedidyah Bar David wrote:
> On Mon, Nov 13, 2017 at 8:57 AM, andre...@starlett.lv
>  wrote:
>> On 11/13/2017 05:36 AM, Yihui Zhao wrote:
>>
>> can you try the admin password?
>>
>>
>> already did, same result.
>>
>> On Mon, Nov 13, 2017 at 3:10 AM, andre...@starlett.lv 
>> wrote:
>>> Hi,
>>>
>>> I'm trying to upload iso with this coomand.
>>> engine-iso-uploader --ssh-user=root --iso-domain=iso upload suse.iso
>>>
>>> Please provide the REST API password for the admin@internal oVirt Engine
>>> user (CTRL+D to abort):
>>> This go OK.
>>>
>>> However, then it asks root password, I enter it, then it asks again and
>>> again. Root password is correct for sure, becuase I can coonect vis ssh
>>> from terminal.
>>>
>>> How to fix this problem?
> Can you please share the log? Thanks.
>
>>> May be its possible just to copy files manually?
> It is. Locate the iso domain on your storage server.
> Inside it you'll find a directory whose name is a random
> uuid, inside it 'images', and inside it a directory named:
> '----'.
> You can put your iso files inside that one. Make sure they
> are readable by user:group 36:36.
>
>>> Thanks in advance.
>>> Andrei
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>
>

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Snapshot or not?

2017-11-13 Thread Demeter Tibor
Dear Users, 

I have a disk of a vm, that is have a snapshot. It is very interesting, because 
there are two other disk of that VM, but there are no snapshots of them. 
I found this while I've try to migrate a storage-domain between two datacenter. 
Because, I didn't import that vm from the storage domain, I did an another 
similar VM with exactly same sized thin-provisioned disks. I have renamed, 
copied to here my originals. 

The VM started successfully, but the disk that contain a snapshot did not 
recognized by the os. I can see the whole disk as raw. (disk id, format in 
ovirt, filenames of images, etc) . I think ovirt don't know that is a 
snapshotted image and use as raw. Is it possible? 
I don't see any snapshot in snapshots. Also I have try to list snapshots with 
qemu-img info and qemu-img snapshot -l , but it does not see any snapshots in 
the image. 

Really, I don't know how is possible this. 

[root@storage1 8d815282-6957-41c0-bb3e-6c8f4a23a64b]# qemu-img info 
5974fd33-af4c-4e3b-aadb-bece6054eb6b 
image: 5974fd33-af4c-4e3b-aadb-bece6054eb6b 
file format: qcow2 
virtual size: 13T (13958643712000 bytes) 
disk size: 12T 
cluster_size: 65536 
backing file: 
../8d815282-6957-41c0-bb3e-6c8f4a23a64b/723ad5aa-02f6-4067-ac75-0ce0a761627f 
backing file format: raw 
Format specific information: 
compat: 0.10 

[root@storage1 8d815282-6957-41c0-bb3e-6c8f4a23a64b]# qemu-img info 
723ad5aa-02f6-4067-ac75-0ce0a761627f 
image: 723ad5aa-02f6-4067-ac75-0ce0a761627f 
file format: raw 
virtual size: 2.0T (2147483648000 bytes) 
disk size: 244G 

[root@storage1 8d815282-6957-41c0-bb3e-6c8f4a23a64b]# ll 
total 13096987560 
-rw-rw. 1 36 36 13149448896512 Nov 13 13:42 
5974fd33-af4c-4e3b-aadb-bece6054eb6b 
-rw-rw. 1 36 36 1048576 Nov 13 19:34 
5974fd33-af4c-4e3b-aadb-bece6054eb6b.lease 
-rw-r--r--. 1 36 36 262 Nov 13 19:54 5974fd33-af4c-4e3b-aadb-bece6054eb6b.meta 
-rw-rw. 1 36 36 2147483648000 Jul 8 2016 
723ad5aa-02f6-4067-ac75-0ce0a761627f 
-rw-rw. 1 36 36 1048576 Jul 7 2016 
723ad5aa-02f6-4067-ac75-0ce0a761627f.lease 
-rw-r--r--. 1 36 36 335 Nov 13 19:52 723ad5aa-02f6-4067-ac75-0ce0a761627f.meta 

qemu-img snapshot -l 5974fd33-af4c-4e3b-aadb-bece6054eb6b 

(nothing) 

Because it is a very big (13 TB) disk I can't migrate to an another image, 
because I don't have enough free space. So I just would like to use it in ovirt 
like in the past. 

I have a very old ovirt (3.5) 

How can I use this disk? 

Thanks in advance, 

Regards, 



Tibor 


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] how to clean stuck task

2017-11-13 Thread Fred Rolland
For cleaning these "Tasks" entries (which are actually Jobs in the DB), you
will need to use taskcleaner.sh.
The unlock_entity.sh is for removing locks on entities.

Before running the task cleaner, you need to be sure you actually don't
have any tasks running in the SPM.

Regarding the snapshot creation on 'apex", can you provide the logs around
the time of the operation?



On Mon, Nov 13, 2017 at 8:06 PM, Gianluca Cecchi 
wrote:

> On Mon, Nov 13, 2017 at 7:01 PM, Gianluca Cecchi <
> gianluca.cec...@gmail.com> wrote:
>
>>
>>
>>
>> I'd prefer to avoid direct manage of the rdbms...
>>
>> As it happened on Friday, after running again the command
>>
>> PGPASSWORD=... /usr/share/ovirt-engine/setup/dbutils/unlock_entity.sh -t
>> all -u engine
>>
>> now the gui after some time shows no active task
>>
>> but the vm still shows as locked and in the event pane I only have the
>> message
>> Snapshot '13-11-2017-16-30-Danilo' creation for VM 'apex' was initiated
>> by user1@internal-authz.
>> but no message about its correct completion
>>
>> This is my disks' subtab situation for the vm, where it seems all ok:
>> https://drive.google.com/file/d/1tF3VdA8gzMVvtn2x4HTKOmltRmU
>> R7cTG/view?usp=sharing
>>
>> And this the snapshots' subtab where it seems all ok:
>> https://drive.google.com/file/d/1qbEhdS6eOZL-f8y5xCnuc_4_HsY
>> M4mAa/view?usp=sharing
>>
>> But the VM has the lock symbol in side of itself.
>> Any guidance appreciated, both to solve the problem and to understand if
>> the snapshot is consistent, as a user should to changes to the VM with a
>> chance to have to come back
>>
>> Gianluca
>>
>
> I have just verified that if I run a new task (in my case has been the
> removal of the snaptest VM), suddenly I see from 0 to 3 tasks, with the
> deletion task completing with success, while the 2 stale tasks appearing
> again in the task pane where they disappeared
> see here:
> https://drive.google.com/file/d/1xcbJsI02iSEAHYEiFYkRWvz9SW_
> ElsIY/view?usp=sharing
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] how to clean stuck task

2017-11-13 Thread Gianluca Cecchi
On Mon, Nov 13, 2017 at 7:01 PM, Gianluca Cecchi 
wrote:

>
>
>
> I'd prefer to avoid direct manage of the rdbms...
>
> As it happened on Friday, after running again the command
>
> PGPASSWORD=... /usr/share/ovirt-engine/setup/dbutils/unlock_entity.sh -t
> all -u engine
>
> now the gui after some time shows no active task
>
> but the vm still shows as locked and in the event pane I only have the
> message
> Snapshot '13-11-2017-16-30-Danilo' creation for VM 'apex' was initiated by
> user1@internal-authz.
> but no message about its correct completion
>
> This is my disks' subtab situation for the vm, where it seems all ok:
> https://drive.google.com/file/d/1tF3VdA8gzMVvtn2x4HTKOmltRmUR7
> cTG/view?usp=sharing
>
> And this the snapshots' subtab where it seems all ok:
> https://drive.google.com/file/d/1qbEhdS6eOZL-f8y5xCnuc_4_
> HsYM4mAa/view?usp=sharing
>
> But the VM has the lock symbol in side of itself.
> Any guidance appreciated, both to solve the problem and to understand if
> the snapshot is consistent, as a user should to changes to the VM with a
> chance to have to come back
>
> Gianluca
>

I have just verified that if I run a new task (in my case has been the
removal of the snaptest VM), suddenly I see from 0 to 3 tasks, with the
deletion task completing with success, while the 2 stale tasks appearing
again in the task pane where they disappeared
see here:
https://drive.google.com/file/d/1xcbJsI02iSEAHYEiFYkRWvz9SW_ElsIY/view?usp=sharing
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] how to clean stuck task

2017-11-13 Thread Gianluca Cecchi
On Mon, Nov 13, 2017 at 5:28 PM, Wesley Stewart  wrote:

> Is the VM showing as locked?  You could try to manually unlocking it.  I
> did this in a test environment a while back with some success after a vm
> was stuck in a "locked" state.  Of course the gui and engine should handle
> most of this for you, manually mucking around the internal DB can cause
> some pretty serious issues if you are not careful...
>
> sudo su postgres
>
> psql -d engine -U postgres
>
> SELECT vm_guid, vm_name FROM vm_static WHERE vm_name='*VM_Name_Here*';
> This should return a string such as: "0ec20854-e1ca-4e49-be87-
> a6cd36d40c18"
>
> Reset the lock:
> update vm_dynamic SET status=0 where vm_guid='0ec20854-e1ca-4e49-
> be87-a6cd36d40c18';
>
>
>
>
I'd prefer to avoid direct manage of the rdbms...

As it happened on Friday, after running again the command

PGPASSWORD=... /usr/share/ovirt-engine/setup/dbutils/unlock_entity.sh -t
all -u engine

now the gui after some time shows no active task

but the vm still shows as locked and in the event pane I only have the
message
Snapshot '13-11-2017-16-30-Danilo' creation for VM 'apex' was initiated by
user1@internal-authz.
but no message about its correct completion

This is my disks' subtab situation for the vm, where it seems all ok:
https://drive.google.com/file/d/1tF3VdA8gzMVvtn2x4HTKOmltRmUR7cTG/view?usp=sharing

And this the snapshots' subtab where it seems all ok:
https://drive.google.com/file/d/1qbEhdS6eOZL-f8y5xCnuc_4_HsYM4mAa/view?usp=sharing

But the VM has the lock symbol in side of itself.
Any guidance appreciated, both to solve the problem and to understand if
the snapshot is consistent, as a user should to changes to the VM with a
chance to have to come back

Gianluca
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] how to clean stuck task

2017-11-13 Thread Wesley Stewart
Is the VM showing as locked?  You could try to manually unlocking it.  I
did this in a test environment a while back with some success after a vm
was stuck in a "locked" state.  Of course the gui and engine should handle
most of this for you, manually mucking around the internal DB can cause
some pretty serious issues if you are not careful...

sudo su postgres

psql -d engine -U postgres

SELECT vm_guid, vm_name FROM vm_static WHERE vm_name='*VM_Name_Here*';
This should return a string such as: "0ec20854-e1ca-4e49-be87-a6cd36d40c18"

Reset the lock:
update vm_dynamic SET status=0 where vm_guid='
0ec20854-e1ca-4e49-be87-a6cd36d40c18';



On Mon, Nov 13, 2017 at 11:01 AM, Gianluca Cecchi  wrote:

> Hello,
>
> it seems the same lock (with the same date) has come again in the list of
> tasks.
> Now I have two things in "supposed running" tasks:
> - the "Restoring VM Snapshot Active VM before the preview of VM snaptest"
> of 8/11
> - the "Creating VM Snapshot 13-11-2017-16-30-Danilo for VM apex"  (now
> running since half an hour for a vm that has only one 90Gb disk on SAN FC
> storage domain)
>
> It was not me to run the "create snapshot" command today, so I'm not sure
> if it was the snapshot creation job responsible of letting come out again
> the former hung task... or if it was already there before...
> Any further hint?
> How can I manually verify the status of the snapshot creation task?
>
> Gianluca
>
>
> On Fri, Nov 10, 2017 at 5:01 PM, Gianluca Cecchi <
> gianluca.cec...@gmail.com> wrote:
>
>> On Fri, Nov 10, 2017 at 3:48 PM,  wrote:
>>
>>>

>>> I've seen this behavior too. IIRC the stale cleaning was not instant, it
>>> took some time to be applied.
>>>
>>> Regards.
>>>
>>> Gianluca

>>>
>> Confirmed.
>> Quite soon after the command I saw that the status of the "Current"
>> snapshot line changed from Locked (it was so since 8/11) to OK, but the
>> task remained at least for half an hour.
>> Now, after about one hour and a half I connected again to the web admin
>> gui and I see 0 Tasks, so the problem has been resolved.
>>
>> Thanks again,
>> Gianluca
>>
>>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] how to clean stuck task

2017-11-13 Thread Gianluca Cecchi
Hello,

it seems the same lock (with the same date) has come again in the list of
tasks.
Now I have two things in "supposed running" tasks:
- the "Restoring VM Snapshot Active VM before the preview of VM snaptest"
of 8/11
- the "Creating VM Snapshot 13-11-2017-16-30-Danilo for VM apex"  (now
running since half an hour for a vm that has only one 90Gb disk on SAN FC
storage domain)

It was not me to run the "create snapshot" command today, so I'm not sure
if it was the snapshot creation job responsible of letting come out again
the former hung task... or if it was already there before...
Any further hint?
How can I manually verify the status of the snapshot creation task?

Gianluca


On Fri, Nov 10, 2017 at 5:01 PM, Gianluca Cecchi 
wrote:

> On Fri, Nov 10, 2017 at 3:48 PM,  wrote:
>
>>
>>>
>> I've seen this behavior too. IIRC the stale cleaning was not instant, it
>> took some time to be applied.
>>
>> Regards.
>>
>> Gianluca
>>>
>>
> Confirmed.
> Quite soon after the command I saw that the status of the "Current"
> snapshot line changed from Locked (it was so since 8/11) to OK, but the
> task remained at least for half an hour.
> Now, after about one hour and a half I connected again to the web admin
> gui and I see 0 Tasks, so the problem has been resolved.
>
> Thanks again,
> Gianluca
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] iSCSI domain on 4kn drives

2017-11-13 Thread David Teigland
On Sat, Nov 11, 2017 at 12:24:25AM +, Nir Soffer wrote:
> David, do you know if 4k disks over NFS works for sanlock?

When using files, sanlock always does 512 byte i/o.  This can be a problem
when there are 4k disks used under NFS.  On disks, sanlock detects the
sector size (with libblkid) and uses 512/4k accordingly.

If vdsm knows when to use 4k i/o over files, I can add a sanlock flag that
vdsm can use to create 4k sanlock leases.

Dave

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Host in kdumping state

2017-11-13 Thread Davide Ferrari
Sorry for not following-up...well as strange as it sounds, after replacing
the faulty HW while keeping the same old disks (basically putting the disks
in a new identical server) the kdumping error disappeared and the host is
back working as usual.
Maybe "kdumping" is displayed after a kernel panic reboot?

Thanks

2017-11-06 6:56 GMT+01:00 Oved Ourfali :

> Hosts should not be stuck in that status.
> Can you please attach the engine logs + the relevant host's logs?
>
> Also, are you using the latest 4.1?
>
> On Fri, Nov 3, 2017 at 3:52 PM, Davide Ferrari 
> wrote:
>
>>
>>
>> On 02/11/17 12:00, Davide Ferrari wrote:
>>
>>> I've got a faulty host that keeps rebooting itself from time to time
>>> (due to HW issues), that is/was part of the 3 hosts group hosting the
>>> HostedEngine, and now it always appears as "Kdumping" in the web
>>> administration panel.
>>>
>>
>> Hello again
>>
>> no idea anybody about at least how to reset this "kdumping" status?
>>
>> Thanks
>>
>>
>> --
>> Davide Ferrari
>> Lead System Engineer
>> Billy Performance Network
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>
>


-- 
Davide Ferrari
Senior Systems Engineer
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] oVirt 4.1 - "Create virtual disk - thin provisioning" makes RAW by default not QCOW2

2017-11-13 Thread andreil1
Hi,

Just got test setup of oVirt 4.1 running. 
Seem to be really brilliant piece of software.


1) Defined local storage domain.

Create virtual disk - thin provisioning makes RAW images by default instead of 
QCOW2.
It preallocated all space in advance.
I looked inside META file in the same folder as disk image - it is marked as 
RAW.

May be its converted to RAW because I formatted it as 3 partitions: SWAP + Boot 
Ext4 + Root Ext4 ?
Should I use single Ext4 partition per virtual disk image to keep it QCOW2 ?

Is there any way to change this behaviour, or convert RAW into QCOW2 within 
oVirt? Certainly I can do it with QEMU, but image become unregistered, or may 
be some functionality become broken.


2) SWAP partition - is it better to create small 1GB SWAP as RAM disk or use 
4GB partition on virtual RAW disk, or I’m wrong here ?


Thanks in advance.
Andrei
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] recommendations for best performance and reliability

2017-11-13 Thread Eduardo Mayoral
Just curious. Would Gluster with a JBOD configuration make sense in this
scenario? Maybe with dm-cache on the SSDs?

In general, does a Gluster JBOD configuration make sense as a backend
for oVirt?

Eduardo Mayoral Jimeno (emayo...@arsys.es)
Administrador de sistemas. Departamento de Plataformas. Arsys internet.
+34 941 620 145 ext. 5153

On 13/11/17 12:47, FERNANDO FREDIANI wrote:
>
> Helli Rudi
>
> Nice specs.
>
> I wouldn't use GlusterFS for this setup with the third server in a
> different location. Just have this server as an Standalone and
> replicate the VMs there. You won't have real time replication, but
> much less hassle and probably to have constant failures, specially
> knowing you have a wireless link.
>
> For the SSDs I have been using bcache with success. Relatively simple
> to setup and pretty good performance.
>
> For your specs as you have 4 mechanical disks I would recommend you to
> have a RAID 5 between them (4 disks is my limit for RAID 5) and a RAID
> 0 made of SSDs for the bcache device. If the RAID 0 fails for any
> reason it will fall back directly to the mechanical disks and you can
> do maintenance on the Node doing live migration in order to replace
> the failed disks.
>
> However as you have you have 2 remaining server to create your cluster
> you may need to consider GlusterFS on the top of this RAID to have the
> replication and Highavaibility.
>
> Hope it helps.
>
> Fernando
>
>
> On 13/11/2017 08:03, Rudi Ahlers wrote:
>> Hi, 
>>
>> Can someone please give me some pointers, what would be the best
>> setup for performance and reliability?
>>
>> We have the following hardware setup:
>>
>> 3x Supermicro server with following features per server:
>> 128GB RAM
>> 4x 8TB SATA HDD
>> 2x SSD drives (intel_ssdsc2ba400g4 - 400GB DC S3710)
>> 2x 12 core CPU (Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz
>> Quad port 10Gbe Inter NIC
>> 2x 10GB Cisco switches (to isolate storage network from LAN)
>>
>> One of the servers will be in another office, with a 600Mb wireless
>> link for Disaster Recovery. 
>>
>> What is recommended for the best setup in terms of redundancy and speed?
>>
>> I am guessing GlusterFS with a Distributed Striped Replicated Volume
>> across 3 of the servers. 
>>
>> For added performance I want to use the SSD drives, perhaps with
>> dm-cache?
>>
>> Should I combine the 4x HDD's using LVM on each host node?
>> What about RAID 6?
>>
>>
>>
>> Virtual Machines will then reside on the oVirt Cluster and any one of
>> the 3 host nodes can fail, or any single HDD can fail and all should
>> still work, right/? 
>>
>>
>>
>>
>> -- 
>> Kind Regards
>> Rudi Ahlers
>> Website: http://www.rudiahlers.co.za
>>
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] recommendations for best performance and reliability

2017-11-13 Thread FERNANDO FREDIANI

Hello Rudi

If you have a 4th server it may work,but I am not knowledgeable about 
Gluster's GEOreplication. Perhaps someone else can advise.


With regards RAID 5 for 4 disks this is intended for capacity and as 
mentioned 4 disks is the maximum I would use for RAID5. As you have SSDs 
and intend to use a caching technique like bcache ou dm-cache this 
should cover the performance hit of RAID 5 write.
If you consider using RAID 6 in only 4 disks you are much better of 
using RAID 10 then as you will have double write performance and same 
capacity. I normally use RAID 10 or in ZFS configurations (not this 
case) multiple vdevs of RAID6.


For RAID in Linux I have been using mdraid. I know LVM does some Raid 
but I personally never done myself so can't advise.


Regards.
Fernando


On 13/11/2017 10:19, Rudi Ahlers wrote:

Hi Fernando,

Thanx.

I meant to say, the 4th server will be in another office. It's about 
3Km away and I was thinking of using Gluster's GEOreplication for this 
purpose.


I am not a fond user of RAID5 at all. But this raises the question: 
does RAID add any unnecessary overhead? I would rather run RAID 6 or 
RAI10.
And then, if RAID is the preferred way (over LVM?), as I don't have 
dedicated hardware RAID cards, would mdraid add any benefit?


On Mon, Nov 13, 2017 at 1:47 PM, FERNANDO FREDIANI 
> wrote:


Helli Rudi

Nice specs.

I wouldn't use GlusterFS for this setup with the third server in a
different location. Just have this server as an Standalone and
replicate the VMs there. You won't have real time replication, but
much less hassle and probably to have constant failures, specially
knowing you have a wireless link.

For the SSDs I have been using bcache with success. Relatively
simple to setup and pretty good performance.

For your specs as you have 4 mechanical disks I would recommend
you to have a RAID 5 between them (4 disks is my limit for RAID 5)
and a RAID 0 made of SSDs for the bcache device. If the RAID 0
fails for any reason it will fall back directly to the mechanical
disks and you can do maintenance on the Node doing live migration
in order to replace the failed disks.

However as you have you have 2 remaining server to create your
cluster you may need to consider GlusterFS on the top of this RAID
to have the replication and Highavaibility.

Hope it helps.

Fernando


On 13/11/2017 08:03, Rudi Ahlers wrote:

Hi,

Can someone please give me some pointers, what would be the best
setup for performance and reliability?

We have the following hardware setup:

3x Supermicro server with following features per server:
128GB RAM
4x 8TB SATA HDD
2x SSD drives (intel_ssdsc2ba400g4 - 400GB DC S3710)
2x 12 core CPU (Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz
Quad port 10Gbe Inter NIC
2x 10GB Cisco switches (to isolate storage network from LAN)

One of the servers will be in another office, with a 600Mb
wireless link for Disaster Recovery.

What is recommended for the best setup in terms of redundancy and
speed?

I am guessing GlusterFS with a Distributed Striped Replicated
Volume across 3 of the servers.

For added performance I want to use the SSD drives, perhaps with
dm-cache?

Should I combine the 4x HDD's using LVM on each host node?
What about RAID 6?



Virtual Machines will then reside on the oVirt Cluster and any
one of the 3 host nodes can fail, or any single HDD can fail and
all should still work, right/?




-- 
Kind Regards

Rudi Ahlers
Website: http://www.rudiahlers.co.za


___
Users mailing list
Users@ovirt.org 
http://lists.ovirt.org/mailman/listinfo/users




___
Users mailing list
Users@ovirt.org 
http://lists.ovirt.org/mailman/listinfo/users





--
Kind Regards
Rudi Ahlers
Website: http://www.rudiahlers.co.za


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Installation of oVirt 4.1, Gluster Storage and Hosted Engine

2017-11-13 Thread Kasturi Narra
Hello,

   From the output you have pasted looks like grafton-sanity.check.sh is
passing and disable-multipath.sh script is failing if i understand
correctly. Can you please copy paste the file path and content of the file ?

Thanks
kasturi

On Mon, Nov 13, 2017 at 4:11 PM, Open tech  wrote:

> Hi Kasturi,
>Thanks a lot for taking a look at this. I think its
> "grafton-sanity-check.sh" . Following is the complete output from the
> install attempt. Ansible ver is 2.4. Gdeploy is 2.0.2.
>
> Do you have a tested step by step for 4.1.6/7 ?. Would be great if you can
> share it.
>
>
> PLAY [gluster_servers] **
> ***
>
> TASK [Run a shell script] **
> 
> changed: [ovirt2] => (item=/usr/share/gdeploy/
> scripts/grafton-sanity-check.sh -d sda1 -h ovirt1,ovirt2,ovirt3)
> changed: [ovirt3] => (item=/usr/share/gdeploy/
> scripts/grafton-sanity-check.sh -d sda1 -h ovirt1,ovirt2,ovirt3)
> changed: [ovirt1] => (item=/usr/share/gdeploy/
> scripts/grafton-sanity-check.sh -d sda1 -h ovirt1,ovirt2,ovirt3)
>
> PLAY RECAP 
> *
> ovirt1 : ok=1changed=1unreachable=0
>  failed=0
> ovirt2 : ok=1changed=1unreachable=0
>  failed=0
> ovirt3 : ok=1changed=1unreachable=0
>  failed=0
>
>
> PLAY [gluster_servers] **
> ***
>
> TASK [Enable or disable services] **
> 
> ok: [ovirt1] => (item=chronyd)
> ok: [ovirt3] => (item=chronyd)
> ok: [ovirt2] => (item=chronyd)
>
> PLAY RECAP 
> *
> ovirt1 : ok=1changed=0unreachable=0
>  failed=0
> ovirt2 : ok=1changed=0unreachable=0
>  failed=0
> ovirt3 : ok=1changed=0unreachable=0
>  failed=0
>
>
> PLAY [gluster_servers] **
> ***
>
> TASK [start/stop/restart/reload services] **
> 
> changed: [ovirt3] => (item=chronyd)
> changed: [ovirt1] => (item=chronyd)
> changed: [ovirt2] => (item=chronyd)
>
> PLAY RECAP 
> *
> ovirt1 : ok=1changed=1unreachable=0
>  failed=0
> ovirt2 : ok=1changed=1unreachable=0
>  failed=0
> ovirt3 : ok=1changed=1unreachable=0
>  failed=0
>
>
> PLAY [gluster_servers] **
> ***
>
> TASK [Run a command in the shell] **
> 
> changed: [ovirt2] => (item=vdsm-tool configure --force)
> changed: [ovirt3] => (item=vdsm-tool configure --force)
> changed: [ovirt1] => (item=vdsm-tool configure --force)
>
> PLAY RECAP 
> *
> ovirt1 : ok=1changed=1unreachable=0
>  failed=0
> ovirt2 : ok=1changed=1unreachable=0
>  failed=0
> ovirt3 : ok=1changed=1unreachable=0
>  failed=0
>
>
> PLAY [gluster_servers] **
> ***
>
> TASK [Run a shell script] **
> 
> fatal: [ovirt2]: FAILED! => {"failed": true, "msg": "The conditional check
> 'result.rc != 0' failed. The error was: error while evaluating conditional
> (result.rc != 0): 'dict object' has no attribute 'rc'"}
> fatal: [ovirt3]: FAILED! => {"failed": true, "msg": "The conditional check
> 'result.rc != 0' failed. The error was: error while evaluating conditional
> (result.rc != 0): 'dict object' has no attribute 'rc'"}
> fatal: [ovirt1]: FAILED! => {"failed": true, "msg": "The conditional check
> 'result.rc != 0' failed. The error was: error while evaluating conditional
> (result.rc != 0): 'dict object' has no attribute 'rc'"}
> to retry, use: --limit @/tmp/tmpEkEkpR/run-script.retry
>
> PLAY RECAP 
> *
> ovirt1 : ok=0changed=0unreachable=0
>  failed=1
> ovirt2 : ok=0changed=0unreachable=0
>  failed=1
> ovirt3 : ok=0changed=0unreachable=0
>  failed=1
>
> Error: Ansible(>= 2.2) is not installed.
> Some of the features might not work if not installed.
>
>
> [root@ovirt2 ~]# yum info ansible
>
> Loaded plugins: fastestmirror, imgbased-persist
>
> Loading mirror speeds from cached hostfile
>
>  * epel: mirror01.idc.hinet.net
>
>  * ovirt-4.1: ftp.nluug.nl
>
>  * ovirt-4.1-epel: mirror01.idc.hinet.net
>
> Installed Packages
>
> Name: *ansible*
>
> Arch: noarch
>
> Version : 2.4.0.0
>
> 

Re: [ovirt-users] recommendations for best performance and reliability

2017-11-13 Thread FERNANDO FREDIANI

Helli Rudi

Nice specs.

I wouldn't use GlusterFS for this setup with the third server in a 
different location. Just have this server as an Standalone and replicate 
the VMs there. You won't have real time replication, but much less 
hassle and probably to have constant failures, specially knowing you 
have a wireless link.


For the SSDs I have been using bcache with success. Relatively simple to 
setup and pretty good performance.


For your specs as you have 4 mechanical disks I would recommend you to 
have a RAID 5 between them (4 disks is my limit for RAID 5) and a RAID 0 
made of SSDs for the bcache device. If the RAID 0 fails for any reason 
it will fall back directly to the mechanical disks and you can do 
maintenance on the Node doing live migration in order to replace the 
failed disks.


However as you have you have 2 remaining server to create your cluster 
you may need to consider GlusterFS on the top of this RAID to have the 
replication and Highavaibility.


Hope it helps.

Fernando


On 13/11/2017 08:03, Rudi Ahlers wrote:

Hi,

Can someone please give me some pointers, what would be the best setup 
for performance and reliability?


We have the following hardware setup:

3x Supermicro server with following features per server:
128GB RAM
4x 8TB SATA HDD
2x SSD drives (intel_ssdsc2ba400g4 - 400GB DC S3710)
2x 12 core CPU (Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz
Quad port 10Gbe Inter NIC
2x 10GB Cisco switches (to isolate storage network from LAN)

One of the servers will be in another office, with a 600Mb wireless 
link for Disaster Recovery.


What is recommended for the best setup in terms of redundancy and speed?

I am guessing GlusterFS with a Distributed Striped Replicated Volume 
across 3 of the servers.


For added performance I want to use the SSD drives, perhaps with dm-cache?

Should I combine the 4x HDD's using LVM on each host node?
What about RAID 6?



Virtual Machines will then reside on the oVirt Cluster and any one of 
the 3 host nodes can fail, or any single HDD can fail and all should 
still work, right/?





--
Kind Regards
Rudi Ahlers
Website: http://www.rudiahlers.co.za


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] oVirt 4.1.9 will be the last 4.1 release

2017-11-13 Thread Sandro Bonazzola
Hi,
we are now targeting oVirt 4.2.0 GA around December 12th and oVirt 4.1.8
around December 5th.
Against tradition that would have set 4.1.8 as last 4.1 release, this round
we'll release also oVIrt 4.1.9 being it already in development just before
4.2.0 GA.

We suggest to be prepared to upgrade to 4.2 since 4.1 won't receive any
further update after 4.1.9 release.

-- 

SANDRO BONAZZOLA

ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R

Red Hat EMEA 

TRIED. TESTED. TRUSTED. 

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted Engine is down and won't start

2017-11-13 Thread Simone Tiraboschi
On Fri, Nov 10, 2017 at 9:42 AM, Kasturi Narra  wrote:

> Hello Logan,
>
>One reason the liveliness check fails is host cannot ping your hosted
> engine vm. you can try connecting to HE vm using remote-viewer
> vnc://hypervisor-ip:5900 and from the hosted-engine --vm-status output
> looks like the HE vm is up and running fine.
>
>
Hi,
just a small addition:
we can deploy hosted-engine choosing vnc or spice as the graphical console
protocol so you have to fix the remote viewer command according to what you
are using.
And the tcp post is not always 5900 but it depends on the VMs starting
order.

To get the actual VNC port number you could use:
. /etc/ovirt-hosted-engine/hosted-engine.conf
vdsm-client VM getInfo vmID=$vmid | jq -r '.devices[] | select(.device |
contains("vnc")).port'

An alternative is to use the serial console with:
hosted-engine --console


>
>- Please check internal dns setting like resolv.conf setting
>- Can not resolve virtual host name or ip address.
>
> Thanks
> kasturi
>
>
> On Fri, Nov 10, 2017 at 12:56 PM, Logan Kuhn 
> wrote:
>
>> We lost the backend storage that hosts our self hosted engine tonight.
>> We've recovered it and there was no data corruption on the volume
>> containing the HE disk.  However, when we try to start the HE it doesn't
>> give an error, but it also doesn't start.
>>
>> The VM isn't pingable and the liveliness check always fails.
>>
>>  [root@ovirttest1 ~]# hosted-engine --vm-status | grep -A20 ovirttest1
>> Hostname   : ovirttest1.wolfram.com
>> Host ID: 1
>> Engine status  : {"reason": "failed liveliness
>> check", "health": "bad", "vm": "up", "detail": "up"}
>> Score  : 3400
>> stopped: False
>> Local maintenance  : False
>> crc32  : 2c2f3ec9
>> local_conf_timestamp   : 18980042
>> Host timestamp : 18980039
>> Extra metadata (valid at timestamp):
>>metadata_parse_version=1
>>metadata_feature_version=1
>>timestamp=18980039 (Fri Nov 10 01:17:59 2017)
>>host-id=1
>>score=3400
>>vm_conf_refresh_time=18980042 (Fri Nov 10 01:18:03 2017)
>>conf_on_shared_storage=True
>>maintenance=False
>>state=GlobalMaintenance
>>stopped=False
>>
>> The environment is in Global Maintenance so that we can isolate it to
>> starting on a specific host to eliminate as many variables as possible.
>> I've attached the agent and broker logs
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Installation of oVirt 4.1, Gluster Storage and Hosted Engine

2017-11-13 Thread Kasturi Narra
Hello,

Can you please let me know which is the script it is failing  and
ansible and gdeploy versions?

Thanks
kasturi

On Mon, Nov 13, 2017 at 2:54 PM, Open tech  wrote:

> Hi All,
>I am new to Ovirt. I am hitting the exact same error while trying a new
> install in a nested virtualization setup on esxi 6.5.
> I am following this tutorial as well. Have three nodes on esxi with dual
> networks & passwordless ssh enabled.
> https://www.ovirt.org/blog/2017/04/up-and-running-with-
> ovirt-4.1-and-gluster-storage/
>
> Node install goes through without issue. Run into this error when i hit
> deploy.
>
> TASK [Run a shell script] **
> 
> fatal: [ovirt3]: FAILED! => {"failed": true, "msg": "The conditional check
> 'result.rc != 0' failed. The error was: error while evaluating conditional
> (result.rc != 0): 'dict object' has no attribute 'rc'"}
> fatal: [ovirt1]: FAILED! => {"failed": true, "msg": "The conditional check
> 'result.rc != 0' failed. The error was: error while evaluating conditional
> (result.rc != 0): 'dict object' has no attribute 'rc'"}
> fatal: [ovirt2]: FAILED! => {"failed": true, "msg": "The conditional check
> 'result.rc != 0' failed. The error was: error while evaluating conditional
> (result.rc != 0): 'dict object' has no attribute 'rc'"}
> to retry, use: --limit @/tmp/tmpbDBjAt/run-script.retry
>
>
> @Simone Marchioni were you able to find a solution ???.
>
> Thanks
> hk
>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted Engine is down

2017-11-13 Thread Martin Sivak
Hi,

Following my answer, this is the bug you opened to track this issue right?

https://bugzilla.redhat.com/show_bug.cgi?id=1511788

You said in comment #2 of that bug that all is well now. Should we
close the bug then?

Best regards

Martin Sivak

On Fri, Nov 10, 2017 at 8:22 AM, Logan Kuhn  wrote:
> We lost the backend storage that hosts our self hosted engine tonight.
> We've recovered it and there was no data corruption on the volume containing
> the HE disk.  However, when we try to start the HE it doesn't give an error,
> but it also doesn't start.
>
> The VM isn't pingable and the liveliness check always fails.
>
>  [root@ovirttest1 ~]# hosted-engine --vm-status | grep -A20 ovirttest1
> Hostname   : ovirttest1.wolfram.com
> Host ID: 1
> Engine status  : {"reason": "failed liveliness check",
> "health": "bad", "vm": "up", "detail": "up"}
> Score  : 3400
> stopped: False
> Local maintenance  : False
> crc32  : 2c2f3ec9
> local_conf_timestamp   : 18980042
> Host timestamp : 18980039
> Extra metadata (valid at timestamp):
>metadata_parse_version=1
>metadata_feature_version=1
>timestamp=18980039 (Fri Nov 10 01:17:59 2017)
>host-id=1
>score=3400
>vm_conf_refresh_time=18980042 (Fri Nov 10 01:18:03 2017)
>conf_on_shared_storage=True
>maintenance=False
>state=GlobalMaintenance
>stopped=False
>
> The environment is in Global Maintenance so that we can isolate it to
> starting on a specific host to eliminate as many variables as possible.
> I've attached the agent and broker logs
>
> Regards,
> Logan Kuhn
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] recommendations for best performance and reliability

2017-11-13 Thread Rudi Ahlers
Hi,

Can someone please give me some pointers, what would be the best setup for
performance and reliability?

We have the following hardware setup:

3x Supermicro server with following features per server:
128GB RAM
4x 8TB SATA HDD
2x SSD drives (intel_ssdsc2ba400g4 - 400GB DC S3710)
2x 12 core CPU (Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz
Quad port 10Gbe Inter NIC
2x 10GB Cisco switches (to isolate storage network from LAN)

One of the servers will be in another office, with a 600Mb wireless link
for Disaster Recovery.

What is recommended for the best setup in terms of redundancy and speed?

I am guessing GlusterFS with a Distributed Striped Replicated Volume across
3 of the servers.

For added performance I want to use the SSD drives, perhaps with dm-cache?

Should I combine the 4x HDD's using LVM on each host node?
What about RAID 6?



Virtual Machines will then reside on the oVirt Cluster and any one of the 3
host nodes can fail, or any single HDD can fail and all should still work,
right/?




-- 
Kind Regards
Rudi Ahlers
Website: http://www.rudiahlers.co.za
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Installation of oVirt 4.1, Gluster Storage and Hosted Engine

2017-11-13 Thread Open tech
Hi All,
   I am new to Ovirt. I am hitting the exact same error while trying a new
install in a nested virtualization setup on esxi 6.5.
I am following this tutorial as well. Have three nodes on esxi with dual
networks & passwordless ssh enabled.
https://www.ovirt.org/blog/2017/04/up-and-running-with-ovirt-4.1-and-gluster-storage/

Node install goes through without issue. Run into this error when i hit
deploy.

TASK [Run a shell script]
**
fatal: [ovirt3]: FAILED! => {"failed": true, "msg": "The conditional check
'result.rc != 0' failed. The error was: error while evaluating conditional
(result.rc != 0): 'dict object' has no attribute 'rc'"}
fatal: [ovirt1]: FAILED! => {"failed": true, "msg": "The conditional check
'result.rc != 0' failed. The error was: error while evaluating conditional
(result.rc != 0): 'dict object' has no attribute 'rc'"}
fatal: [ovirt2]: FAILED! => {"failed": true, "msg": "The conditional check
'result.rc != 0' failed. The error was: error while evaluating conditional
(result.rc != 0): 'dict object' has no attribute 'rc'"}
to retry, use: --limit @/tmp/tmpbDBjAt/run-script.retry


@Simone Marchioni were you able to find a solution ???.

Thanks
hk
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted Engine is down

2017-11-13 Thread Kasturi Narra
Hi Logan,

When i look at the hosted-engine --vm-status i see that vm is up but it
is health is bad. Can you try connecting to the vm using remote-viewer
using the command below ?

remote-viewer vnc://ovirttest1.wolfram.com:5900

Thanks
kasturi

On Fri, Nov 10, 2017 at 12:52 PM, Logan Kuhn  wrote:

> We lost the backend storage that hosts our self hosted engine tonight.
> We've recovered it and there was no data corruption on the volume
> containing the HE disk.  However, when we try to start the HE it doesn't
> give an error, but it also doesn't start.
>
> The VM isn't pingable and the liveliness check always fails.
>
>  [root@ovirttest1 ~]# hosted-engine --vm-status | grep -A20 ovirttest1
> Hostname   : ovirttest1.wolfram.com
> Host ID: 1
> Engine status  : {"reason": "failed liveliness check",
> "health": "bad", "vm": "up", "detail": "up"}
> Score  : 3400
> stopped: False
> Local maintenance  : False
> crc32  : 2c2f3ec9
> local_conf_timestamp   : 18980042
> Host timestamp : 18980039
> Extra metadata (valid at timestamp):
>metadata_parse_version=1
>metadata_feature_version=1
>timestamp=18980039 (Fri Nov 10 01:17:59 2017)
>host-id=1
>score=3400
>vm_conf_refresh_time=18980042 (Fri Nov 10 01:18:03 2017)
>conf_on_shared_storage=True
>maintenance=False
>state=GlobalMaintenance
>stopped=False
>
> The environment is in Global Maintenance so that we can isolate it to
> starting on a specific host to eliminate as many variables as possible.
> I've attached the agent and broker logs
>
> Regards,
> Logan Kuhn
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users