[ovirt-users] Manual transfer of VMs from DC to DC

2017-09-15 Thread Abi Askushi
Hi all,

Is there any way of transferring VMs manually from DC to DC, without the
DCs having connectivity with each other?

I was thinking to backup all the export domain directory, and then later
rsync this directory VMs to a new NFS share, then import this NFS share as
an export domain on the other DC.

What do you think?

Thanx,
Alex
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Recommended gluster setup

2017-09-15 Thread Abi Askushi
Hi all,

I see in the ovirt guides that a gluster volume replica 3 with 1 arbiter is
recommended.
Why not simple replica 3? Is it due to the higher replication data that
would cause performance issues?

What I am observing is that a VM running on the server which has the
arbiter brick has slower read performance then when the same VM runs on
another server with a normal brick. Has anyone observed this? Is it because
the arbiter does not have the real data on it?

Thanx,
Alex
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt selft-hosted with NFS on top gluster

2017-09-08 Thread Abi Askushi
Hi all,

Seems that I am seeing light in he tunnel!

I have done the below:

Added the option
/etc/glusterfs/glusterd.vol :
option rpc-auth-allow-insecure on

restarted glusterd

Then set the volume option:
gluster volume set vms server.allow-insecure on

Then disabled the option:
gluster volume set vms performance.readdir-ahead off

which seems to have been enabled when desperately testing gluster options.

then I added again the storage domain with glusterfs (not NFS).
This lead me to have really high performance boost on writes of VMs.

The dd went from 10MB/s to 60MB/s!
And the VM disk benchmarks from few MB to 80 - 100MB/s!

Seems that the above two options which were missing did the trick for the
gluster.

When checking the VM XML I still see the disk defines as file and not
network. I am not sure that ligfapi is used from ovirt.
I will set also a new VM just in case to check the XML again.
Is there any way to confirm use of ligfapi?

On Fri, Sep 8, 2017 at 10:08 AM, Abi Askushi <rightkickt...@gmail.com>
wrote:

> I don't see any other bottleneck. CPUs are quite idle. Seems that the load
> is mostly due to high latency on IO.
>
> Reading further the gluster docs:
>
> https://github.com/gluster/glusterfs-specs/blob/master/
> done/GlusterFS%203.5/libgfapi%20with%20qemu%20libvirt.md
>
> I see that I am missing the following options:
>
> /etc/glusterfs/glusterd.vol :
> option rpc-auth-allow-insecure on
>
> gluster volume set vms gluster server.allow-insecure on
>
> It says that the above allow qemu to use libgfapi.
> When checking the VM XML, I don't see any gluster protocol at the disk
> drive:
>
> 
>io='threads'/>
>   
>   
>   222a1312-5efa-4731-8914-9a9d24dccba5
>   
>   
> 
>
>
>
> While at gluster docs it mentions the below type of disk:
>
> 
>
>
> 
> 
>
> function='0x0'/>
> 
>
>
> Does the above indicate that ovirt/qemu is not using libgfapi but FUSE only?
> This could be the reason of such slow perf.
>
>
> On Thu, Sep 7, 2017 at 1:47 PM, Yaniv Kaul <yk...@redhat.com> wrote:
>
>>
>>
>> On Thu, Sep 7, 2017 at 12:52 PM, Abi Askushi <rightkickt...@gmail.com>
>> wrote:
>>
>>>
>>>
>>> On Thu, Sep 7, 2017 at 10:30 AM, Yaniv Kaul <yk...@redhat.com> wrote:
>>>
>>>>
>>>>
>>>> On Thu, Sep 7, 2017 at 10:06 AM, Yaniv Kaul <yk...@redhat.com> wrote:
>>>>
>>>>>
>>>>>
>>>>> On Wed, Sep 6, 2017 at 6:08 PM, Abi Askushi <rightkickt...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> For a first idea I use:
>>>>>>
>>>>>> dd if=/dev/zero of=testfile bs=1GB count=1
>>>>>>
>>>>>
>>>>> This is an incorrect way to test performance, for various reasons:
>>>>> 1. You are not using oflag=direct , thus not using DirectIO, but using
>>>>> cache.
>>>>> 2. It's unrealistic - it is very uncommon to write large blocks of
>>>>> zeros (sometimes during FS creation or wiping). Certainly not 1GB
>>>>> 3. It is a single thread of IO - again, unrealistic for VM's IO.
>>>>>
>>>>> I forgot to mention that I include oflag=direct in my tests. I agree
>>> though that dd is not the correct way to test, hence I mentioned I just use
>>> it to get a first feel. More tests are done within the VM benchmarking its
>>> disk IO (with tools like IOmeter).
>>>
>>> I suggest using fio and such. See https://github.com/pcuzner/fio-tools
>>>>> for example.
>>>>>
>>>> Do you have any recommended config file to use for VM workload?
>>>
>>
>> Desktops and Servers VMs behave quite differently, so not really. But the
>> 70/30 job is typically a good baseline.
>>
>>
>>>
>>>
>>>>>
>>>>>>
>>>>>> When testing on the gluster mount point using above command I hardly
>>>>>> get 10MB/s. (On the same time the network traffic hardly reaches 
>>>>>> 100Mbit).
>>>>>>
>>>>>> When testing our of the gluster (for example at /root) I get 600 -
>>>>>> 700MB/s.
>>>>>>
>>>>>
>>>>> That's very fast - from 4 disks doing RAID5? Impressive (unless you
>>>>> use caching!). Are those HDDs or SSDs/NVMe?
>>>>>
>>>>>
>>>> These are SAS disks. But there is also

Re: [ovirt-users] oVirt selft-hosted with NFS on top gluster

2017-09-08 Thread Abi Askushi
I don't see any other bottleneck. CPUs are quite idle. Seems that the load
is mostly due to high latency on IO.

Reading further the gluster docs:

https://github.com/gluster/glusterfs-specs/blob/master/done/GlusterFS%203.5/libgfapi%20with%20qemu%20libvirt.md

I see that I am missing the following options:

/etc/glusterfs/glusterd.vol :
option rpc-auth-allow-insecure on

gluster volume set vms gluster server.allow-insecure on

It says that the above allow qemu to use libgfapi.
When checking the VM XML, I don't see any gluster protocol at the disk
drive:


  
  
  
  222a1312-5efa-4731-8914-9a9d24dccba5
  
  




While at gluster docs it mentions the below type of disk:


   
   


   
   



Does the above indicate that ovirt/qemu is not using libgfapi but FUSE only?
This could be the reason of such slow perf.


On Thu, Sep 7, 2017 at 1:47 PM, Yaniv Kaul <yk...@redhat.com> wrote:

>
>
> On Thu, Sep 7, 2017 at 12:52 PM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
>
>>
>>
>> On Thu, Sep 7, 2017 at 10:30 AM, Yaniv Kaul <yk...@redhat.com> wrote:
>>
>>>
>>>
>>> On Thu, Sep 7, 2017 at 10:06 AM, Yaniv Kaul <yk...@redhat.com> wrote:
>>>
>>>>
>>>>
>>>> On Wed, Sep 6, 2017 at 6:08 PM, Abi Askushi <rightkickt...@gmail.com>
>>>> wrote:
>>>>
>>>>> For a first idea I use:
>>>>>
>>>>> dd if=/dev/zero of=testfile bs=1GB count=1
>>>>>
>>>>
>>>> This is an incorrect way to test performance, for various reasons:
>>>> 1. You are not using oflag=direct , thus not using DirectIO, but using
>>>> cache.
>>>> 2. It's unrealistic - it is very uncommon to write large blocks of
>>>> zeros (sometimes during FS creation or wiping). Certainly not 1GB
>>>> 3. It is a single thread of IO - again, unrealistic for VM's IO.
>>>>
>>>> I forgot to mention that I include oflag=direct in my tests. I agree
>> though that dd is not the correct way to test, hence I mentioned I just use
>> it to get a first feel. More tests are done within the VM benchmarking its
>> disk IO (with tools like IOmeter).
>>
>> I suggest using fio and such. See https://github.com/pcuzner/fio-tools
>>>> for example.
>>>>
>>> Do you have any recommended config file to use for VM workload?
>>
>
> Desktops and Servers VMs behave quite differently, so not really. But the
> 70/30 job is typically a good baseline.
>
>
>>
>>
>>>>
>>>>>
>>>>> When testing on the gluster mount point using above command I hardly
>>>>> get 10MB/s. (On the same time the network traffic hardly reaches 100Mbit).
>>>>>
>>>>> When testing our of the gluster (for example at /root) I get 600 -
>>>>> 700MB/s.
>>>>>
>>>>
>>>> That's very fast - from 4 disks doing RAID5? Impressive (unless you use
>>>> caching!). Are those HDDs or SSDs/NVMe?
>>>>
>>>>
>>> These are SAS disks. But there is also a RAID controller with 1GB cache.
>>
>>
>>>>> When I mount the gluster volume with NFS and test on it I get 90 - 100
>>>>> MB/s, (almost 10x from gluster results) which is the max I can get
>>>>> considering I have only 1 Gbit network for the storage.
>>>>>
>>>>> Also, when using glusterfs the general VM performance is very poor and
>>>>> disk write benchmarks show that is it at least 4 times slower then when 
>>>>> the
>>>>> VM is hosted on the same data store when NFS mounted.
>>>>>
>>>>> I don't know why I hitting such a significant performance penalty, and
>>>>> every possible tweak that I was able to find out there did not make any
>>>>> difference on the performance.
>>>>>
>>>>> The hardware I am using is pretty decent for the purposes intended:
>>>>> 3 nodes, each node having with 32 MB of RAM, 16 physical CPU cores, 2
>>>>> TB of storage in RAID5 (4 disks), of which 1.5 TB are sliced for the data
>>>>> store of ovirt where VMs are stored.
>>>>>
>>>>
>>> I forgot to ask why are you using RAID 5 with 4 disks and not RAID 10?
>>> Same usable capacity, higher performance, same protection and faster
>>> recovery, I believe.
>>>
>> Correction: there are 5 disks of 600GB each. The main reason going with
>> RA

Re: [ovirt-users] oVirt selft-hosted with NFS on top gluster

2017-09-07 Thread Abi Askushi
On Thu, Sep 7, 2017 at 10:30 AM, Yaniv Kaul <yk...@redhat.com> wrote:

>
>
> On Thu, Sep 7, 2017 at 10:06 AM, Yaniv Kaul <yk...@redhat.com> wrote:
>
>>
>>
>> On Wed, Sep 6, 2017 at 6:08 PM, Abi Askushi <rightkickt...@gmail.com>
>> wrote:
>>
>>> For a first idea I use:
>>>
>>> dd if=/dev/zero of=testfile bs=1GB count=1
>>>
>>
>> This is an incorrect way to test performance, for various reasons:
>> 1. You are not using oflag=direct , thus not using DirectIO, but using
>> cache.
>> 2. It's unrealistic - it is very uncommon to write large blocks of zeros
>> (sometimes during FS creation or wiping). Certainly not 1GB
>> 3. It is a single thread of IO - again, unrealistic for VM's IO.
>>
>> I forgot to mention that I include oflag=direct in my tests. I agree
though that dd is not the correct way to test, hence I mentioned I just use
it to get a first feel. More tests are done within the VM benchmarking its
disk IO (with tools like IOmeter).

I suggest using fio and such. See https://github.com/pcuzner/fio-tools for
>> example.
>>
> Do you have any recommended config file to use for VM workload?


>>
>>>
>>> When testing on the gluster mount point using above command I hardly get
>>> 10MB/s. (On the same time the network traffic hardly reaches 100Mbit).
>>>
>>> When testing our of the gluster (for example at /root) I get 600 -
>>> 700MB/s.
>>>
>>
>> That's very fast - from 4 disks doing RAID5? Impressive (unless you use
>> caching!). Are those HDDs or SSDs/NVMe?
>>
>>
> These are SAS disks. But there is also a RAID controller with 1GB cache.


>>> When I mount the gluster volume with NFS and test on it I get 90 - 100
>>> MB/s, (almost 10x from gluster results) which is the max I can get
>>> considering I have only 1 Gbit network for the storage.
>>>
>>> Also, when using glusterfs the general VM performance is very poor and
>>> disk write benchmarks show that is it at least 4 times slower then when the
>>> VM is hosted on the same data store when NFS mounted.
>>>
>>> I don't know why I hitting such a significant performance penalty, and
>>> every possible tweak that I was able to find out there did not make any
>>> difference on the performance.
>>>
>>> The hardware I am using is pretty decent for the purposes intended:
>>> 3 nodes, each node having with 32 MB of RAM, 16 physical CPU cores, 2 TB
>>> of storage in RAID5 (4 disks), of which 1.5 TB are sliced for the data
>>> store of ovirt where VMs are stored.
>>>
>>
> I forgot to ask why are you using RAID 5 with 4 disks and not RAID 10?
> Same usable capacity, higher performance, same protection and faster
> recovery, I believe.
>
Correction: there are 5 disks of 600GB each. The main reason going with
RAID 5 was the capacity. With RAID 10 I can use only 4 of them and get only
1.1 TB usable, with RAID 5 I get 2.2 TB usable. I agree going with RAID 10
(+ one additional drive to go with 6 drives) would be better but this is
what I have now.

Y.
>
>
>> You have not mentioned your NIC speeds. Please ensure all work well, with
>> 10g.
>> Is the network dedicated for Gluster traffic? How are they connected?
>>
>>
> I have mentioned that I have 1 Gbit dedicated for the storage. A different
network is used for this and a dedicated 1Gbit switch. The throughput has
been confirmed between all nodes with iperf.
I know 10Gbit would be better, but when using native gluster at ovirt the
network pipe was hardly reaching 100Mbps thus the bottleneck was gluster
and not the network. If I can saturate 1Gbit and I still have performance
issues then I may think to go with 10Gbit. With NFS on top gluster I see
traffic reaching 800Mbit when testing with dd which is much better.


>>> The gluster configuration is the following:
>>>
>>
>> Which version of Gluster are you using?
>>
>>
> The version is  3.8.12

>
>>> Volume Name: vms
>>> Type: Replicate
>>> Volume ID: 4513340d-7919-498b-bfe0-d836b5cea40b
>>> Status: Started
>>> Snapshot Count: 0
>>> Number of Bricks: 1 x (2 + 1) = 3
>>> Transport-type: tcp
>>> Bricks:
>>> Brick1: gluster0:/gluster/vms/brick
>>> Brick2: gluster1:/gluster/vms/brick
>>> Brick3: gluster2:/gluster/vms/brick (arbiter)
>>> Options Reconfigured:
>>> nfs.export-volumes: on
>>> nfs.disable: off
>>> performance.readdir-ahead: on
>>> transport.address-family: inet
>>> performance.quick-read: off
>

Re: [ovirt-users] oVirt selft-hosted with NFS on top gluster

2017-09-06 Thread Abi Askushi
For a first idea I use:

dd if=/dev/zero of=testfile bs=1GB count=1

When testing on the gluster mount point using above command I hardly get
10MB/s. (On the same time the network traffic hardly reaches 100Mbit).

When testing our of the gluster (for example at /root) I get 600 - 700MB/s.

When I mount the gluster volume with NFS and test on it I get 90 - 100
MB/s, (almost 10x from gluster results) which is the max I can get
considering I have only 1 Gbit network for the storage.

Also, when using glusterfs the general VM performance is very poor and disk
write benchmarks show that is it at least 4 times slower then when the VM
is hosted on the same data store when NFS mounted.

I don't know why I hitting such a significant performance penalty, and
every possible tweak that I was able to find out there did not make any
difference on the performance.

The hardware I am using is pretty decent for the purposes intended:
3 nodes, each node having with 32 MB of RAM, 16 physical CPU cores, 2 TB of
storage in RAID5 (4 disks), of which 1.5 TB are sliced for the data store
of ovirt where VMs are stored.

The gluster configuration is the following:

Volume Name: vms
Type: Replicate
Volume ID: 4513340d-7919-498b-bfe0-d836b5cea40b
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x (2 + 1) = 3
Transport-type: tcp
Bricks:
Brick1: gluster0:/gluster/vms/brick
Brick2: gluster1:/gluster/vms/brick
Brick3: gluster2:/gluster/vms/brick (arbiter)
Options Reconfigured:
nfs.export-volumes: on
nfs.disable: off
performance.readdir-ahead: on
transport.address-family: inet
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: on
performance.low-prio-threads: 32
network.remote-dio: off
cluster.eager-lock: off
cluster.quorum-type: auto
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 1
features.shard: on
user.cifs: off
storage.owner-uid: 36
storage.owner-gid: 36
network.ping-timeout: 30
performance.strict-o-direct: on
cluster.granular-entry-heal: enable
features.shard-block-size: 64MB
performance.client-io-threads: on
client.event-threads: 4
server.event-threads: 4
performance.write-behind-window-size: 4MB
performance.cache-size: 1GB

In case I can provide any other details let me know.
At the moment I already switched to gluster based NFS but I have a similar
setup with 2 nodes  where the data store is mounted through gluster (and
again relatively good hardware) where I might check any tweaks or
improvements on this setup.

Thanx


On Wed, Sep 6, 2017 at 5:32 PM, Yaniv Kaul <yk...@redhat.com> wrote:

>
>
> On Wed, Sep 6, 2017 at 3:32 PM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
>
>> Hi All,
>>
>> I've playing with ovirt self hosted engine setup and I even use it to
>> production for several VM. The setup I have is 3 server with gluster
>> storage in replica 2+1 (1 arbiter).
>> The data storage domain where VMs are stored is mounted with gluster
>> through ovirt. The performance I get for the VMs is very low and I was
>> thinking to switch and mount the same storage through NFS instead of
>> glusterfs.
>>
>
> I don't see how it'll improve performance.
> I suggest you share the gluster configuration (as well as the storage HW)
> so we can understand why the performance is low.
> Y.
>
>
>>
>> The only think I am hesitant is how can I ensure high availability of the
>> storage when I loose one server? I was thinking to have at /etc/hosts sth
>> like below:
>>
>> 10.100.100.1 nfsmount
>> 10.100.100.2 nfsmount
>> 10.100.100.3 nfsmount
>>
>> then use nfsmount as the server name when adding this domain through
>> ovirt GUI.
>> Are there any other more elegant solutions? What do you do for such cases?
>> Note: gluster has the back-vol-file option which provides a lean way to
>> have redundancy on the mount point and I am using this when mounting with
>> glusterfs.
>>
>> Thanx
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] oVirt selft-hosted with NFS on top gluster

2017-09-06 Thread Abi Askushi
Hi All,

I've playing with ovirt self hosted engine setup and I even use it to
production for several VM. The setup I have is 3 server with gluster
storage in replica 2+1 (1 arbiter).
The data storage domain where VMs are stored is mounted with gluster
through ovirt. The performance I get for the VMs is very low and I was
thinking to switch and mount the same storage through NFS instead of
glusterfs.

The only think I am hesitant is how can I ensure high availability of the
storage when I loose one server? I was thinking to have at /etc/hosts sth
like below:

10.100.100.1 nfsmount
10.100.100.2 nfsmount
10.100.100.3 nfsmount

then use nfsmount as the server name when adding this domain through ovirt
GUI.
Are there any other more elegant solutions? What do you do for such cases?
Note: gluster has the back-vol-file option which provides a lean way to
have redundancy on the mount point and I am using this when mounting with
glusterfs.

Thanx
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Package dependency issue

2017-08-24 Thread Abi Askushi
I confirm that issue has been resolved now.

On Thu, Aug 24, 2017 at 12:34 PM, Abi Askushi <rightkickt...@gmail.com>
wrote:

> Thanx!
>
> On Thu, Aug 24, 2017 at 11:41 AM, Yedidyah Bar David <d...@redhat.com>
> wrote:
>
>> On Thu, Aug 24, 2017 at 10:53 AM, Abi Askushi <rightkickt...@gmail.com>
>> wrote:
>> > Hi all,
>> >
>> > I have ovirt 4.1 with self hosted engine.
>> > I have performed lately yum update -y on the hosts and at next yum
>> update at
>> > both hosts I get the following dependency issue:
>> >
>> > Error: Package: rubygem-fluent-plugin-secure-forward-0.4.5-1.el7.noarch
>> > (centos-opstools-release)
>> >Requires: rubygem(proxifier)
>> >
>> > The rubygem-proxifier dependency is already installed:
>> >
>> > rubygem-proxifier-1.0.3-1.el7.noarch
>> >
>> > I have tried yum update --skip-broken and rpm -Va --nofiles --nodigest
>> to
>> > see if this resolves this but the issue remains.
>> >
>> > Any idea how to resolve this?
>>
>> The missing package has been tagged for release and will be on CentOS
>> mirrors in a few hours.
>>
>> See also: https://bugzilla.redhat.com/show_bug.cgi?id=1484534
>>
>> Best,
>>
>> >
>> > Thanx,
>> > Abi
>> >
>> > ___
>> > Users mailing list
>> > Users@ovirt.org
>> > http://lists.ovirt.org/mailman/listinfo/users
>> >
>>
>>
>>
>> --
>> Didi
>>
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Package dependency issue

2017-08-24 Thread Abi Askushi
Thanx!

On Thu, Aug 24, 2017 at 11:41 AM, Yedidyah Bar David <d...@redhat.com>
wrote:

> On Thu, Aug 24, 2017 at 10:53 AM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
> > Hi all,
> >
> > I have ovirt 4.1 with self hosted engine.
> > I have performed lately yum update -y on the hosts and at next yum
> update at
> > both hosts I get the following dependency issue:
> >
> > Error: Package: rubygem-fluent-plugin-secure-forward-0.4.5-1.el7.noarch
> > (centos-opstools-release)
> >Requires: rubygem(proxifier)
> >
> > The rubygem-proxifier dependency is already installed:
> >
> > rubygem-proxifier-1.0.3-1.el7.noarch
> >
> > I have tried yum update --skip-broken and rpm -Va --nofiles --nodigest to
> > see if this resolves this but the issue remains.
> >
> > Any idea how to resolve this?
>
> The missing package has been tagged for release and will be on CentOS
> mirrors in a few hours.
>
> See also: https://bugzilla.redhat.com/show_bug.cgi?id=1484534
>
> Best,
>
> >
> > Thanx,
> > Abi
> >
> > ___
> > Users mailing list
> > Users@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> >
>
>
>
> --
> Didi
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Package dependency issue

2017-08-24 Thread Abi Askushi
Hi all,

I have ovirt 4.1 with self hosted engine.
I have performed lately yum update -y on the hosts and at next yum update
at both hosts I get the following dependency issue:

Error: Package: rubygem-fluent-plugin-secure-forward-0.4.5-1.el7.noarch
(centos-opstools-release)
   Requires: rubygem(proxifier)

The rubygem-proxifier dependency is already installed:

rubygem-proxifier-1.0.3-1.el7.noarch

I have tried yum update --skip-broken and rpm -Va --nofiles --nodigest to
see if this resolves this but the issue remains.

Any idea how to resolve this?

Thanx,
Abi
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt VM backups

2017-07-27 Thread Abi Askushi
Hi,

Yes, the python script is using this API.
I was thinking to have in the GUI for VM backups sth similar as the gluster
volume snapshot schedules.


On Thu, Jul 27, 2017 at 2:09 PM, Maor Lipchuk <mlipc...@redhat.com> wrote:

> On Thu, Jul 27, 2017 at 1:14 PM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
> > Hi All,
> >
> > For VM backups I am using some python script to automate the snapshot ->
> > clone -> export -> delete steps (although with some issues when trying to
> > backups a Windows 10 VM)
> >
> > I was wondering if there is there any plan to integrate VM backups in the
> > GUI or what other recommended ways exist out there.
>
> Hi Abi,
>
> Don't you want to use the backup API feature for that:
> https://www.ovirt.org/develop/api/design/backup-api/
>
> Regards,
> Maor
>
> >
> > Thanx,
> > Abi
> >
> > ___
> > Users mailing list
> > Users@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> >
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] oVirt VM backups

2017-07-27 Thread Abi Askushi
Hi All,

For VM backups I am using some python script to automate the snapshot ->
clone -> export -> delete steps (although with some issues when trying to
backups a Windows 10 VM)

I was wondering if there is there any plan to integrate VM backups in the
GUI or what other recommended ways exist out there.

Thanx,
Abi
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] test email

2017-07-19 Thread Abi Askushi
several days without receiving any email from this list.
please test back.

Abi
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Windows USB Redirection

2017-07-14 Thread Abi Askushi
Hi Michal,

Thank you for your tip.
Do you have any example how this attribute can be extended to avoid
changing the default?

Thanx
Abi

On Wed, Jul 12, 2017 at 3:34 PM, Michal Skrivanek <
michal.skriva...@redhat.com> wrote:

>
> On 11 Jul 2017, at 10:09, Abi Askushi <rightkickt...@gmail.com> wrote:
>
> Hi All,
>
> Finally I was able to pin-point the issue.
> Just in case ones needs this.
>
>
> ovirt by default enables USB 1.0 controller.
>
> To enable USB 2.0 controller go to
>
> At ovirt engine at config file below, change the controller from piix3-uhci
> to piix4-uhci as below:
>
>
> piix4-uhci is still USB1 controller
> the one we use is still the most compatible default, we do not test any
> other (other than set of USB2 and USB1 companions when using SPICE)
> you’re welcome to overwrite it (not by changing the defaults which gets
> overwritten on upgrade, but rather extending it by your custom properties,
> ideally per OS type where you need that change)
> You can use all the QEMU supported controllers, e.g. xhci if you want USB3
>
> Thanks,
> michal
>
>
> /etc/ovirt-engine/osinfo.conf.d/00-defaults.properties
>
>
> os.other.devices.usb.controller.value = piix4-uhci
>
>
> then restart engine:
> systemctl restart ovirt-engine
>
>
> USB device then is attached successfully at Windows VM.
>
> On Wed, Jun 28, 2017 at 12:44 PM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
>
>>
>> Hi All,
>>
>> I have Ovirt 4.1 with 3 nodes on top glusterfs.
>>
>> I Have 2 VMs: Windows 2016 64bit and Windows 10 64bit
>>
>> When I attach a USB flash disk to a VM (from host devices) the VM cannot
>> see the USB drive and report driver issue at device manager (see attached).
>> This happens with both VMs when tested.
>>
>> When testing with Windows 7 or Windows XP USB is attached and accessed
>> normally.
>> Have you encountered such issue?
>> I have installed latest guest tools on both VMs.
>>
>> Many thanx
>>
>>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Windows USB Redirection

2017-07-11 Thread Abi Askushi
Hi All,

Finally I was able to pin-point the issue.
Just in case ones needs this.


ovirt by default enables USB 1.0 controller.

To enable USB 2.0 controller go to

At ovirt engine at config file below, change the controller from piix3-uhci
to piix4-uhci as below:

/etc/ovirt-engine/osinfo.conf.d/00-defaults.properties


os.other.devices.usb.controller.value = piix4-uhci


then restart engine:
systemctl restart ovirt-engine


USB device then is attached successfully at Windows VM.

On Wed, Jun 28, 2017 at 12:44 PM, Abi Askushi <rightkickt...@gmail.com>
wrote:

>
> Hi All,
>
> I have Ovirt 4.1 with 3 nodes on top glusterfs.
>
> I Have 2 VMs: Windows 2016 64bit and Windows 10 64bit
>
> When I attach a USB flash disk to a VM (from host devices) the VM cannot
> see the USB drive and report driver issue at device manager (see attached).
> This happens with both VMs when tested.
>
> When testing with Windows 7 or Windows XP USB is attached and accessed
> normally.
> Have you encountered such issue?
> I have installed latest guest tools on both VMs.
>
> Many thanx
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Fwd: Windows USB Redirection

2017-06-28 Thread Abi Askushi
Hi All,

I have Ovirt 4.1 with 3 nodes on top glusterfs.

I Have 2 VMs: Windows 2016 64bit and Windows 10 64bit

When I attach a USB flash disk to a VM (from host devices) the VM cannot
see the USB drive and report driver issue at device manager (see attached).
This happens with both VMs when tested.

When testing with Windows 7 or Windows XP USB is attached and accessed
normally.
Have you encountered such issue?
I have installed latest guest tools on both VMs.

Many thanx
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Ovirt remove network from hosted engine

2017-06-28 Thread Abi Askushi
Hi,

I was trying to remove the nic from the edit menu of the VM and that failed.

I had to unplug the nic first to be able to remove it from the network
interfaces submenu of the vm.

On Jun 28, 2017 08:29, "knarra" <kna...@redhat.com> wrote:

> On 06/27/2017 09:49 PM, Abi Askushi wrote:
>
> Hi all,
>
> Just in case ones needs it, in order to remove the secondary network
> interface from engine, you can go to:
> Virtual Machines -> Hostedengine -> Network Interfaces -> edit -> unplug
> it -> confirm -> remove it.
>
> cool. But in your previous mail you did mention that it fails for you
> since the engine is running. Instead of remove you tried unplug here ?
>
>
> It was simple...
>
>
> On Tue, Jun 27, 2017 at 4:54 PM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
>
>> Hi Knarra,
>>
>> Then I had already enabled NFS on ISO gluster volume.
>> Maybe i had some networking issue then. I need to remove the secondary
>> interface in order to test that again.
>>
>>
>>
>> On Tue, Jun 27, 2017 at 4:25 PM, knarra <kna...@redhat.com> wrote:
>>
>>> On 06/27/2017 06:34 PM, Abi Askushi wrote:
>>>
>>> Hi Knarra,
>>>
>>> The ISO domain is of type gluster though I had nfs enabled on that
>>> volume.
>>>
>>> you need to have nfs enabled on the volume. what i meant is nfs.disable
>>> off which means nfs is on.
>>>
>>> For more info please refer to bug https://bugzilla.redhat.com/sh
>>> ow_bug.cgi?id=1437799
>>>
>>> I will disable the nfs and try. Though in order to try I need first to
>>> remove that second interface from engine.
>>> Is there a way I can remove the secondary storage network interface from
>>> the engine?
>>>
>>> I am not sure how to do that, but   you may shutdown the vm using the
>>> command hosted-engine --vm-shutdown which will power off the vm and try to
>>> remove the networks using vdsclient. (not sure if this is right, but
>>> suggesting a way)
>>>
>>>
>>> Thanx
>>>
>>>
>>>
>>> On Tue, Jun 27, 2017 at 3:32 PM, knarra <kna...@redhat.com> wrote:
>>>
>>>> On 06/27/2017 05:41 PM, Abi Askushi wrote:
>>>>
>>>> Hi all,
>>>>
>>>> When setting up hosted engine setup on top gluster with 3 nodes, I had
>>>> gluster configured on a separate network interface, as recommended. When I
>>>> tried later to upload ISO from engine to ISO domain, the engine was not
>>>> able to upload it since the VM did not have access to the separate storage
>>>> network. I then added the storage network interface to the hosted engine
>>>> and ISO upload succeeded.
>>>>
>>>> May i know what was the volume type created and added as ISO domain ?
>>>>
>>>> If you plan to use a glusterfs volume below is the procedure :
>>>>
>>>> 1) Create a glusterfs volume.
>>>> 2) While adding storage domain select Domain Function as 'ISO' and
>>>> Storage Type as 'glusterfs' .
>>>> 3) You can either use 'use managed gluster volume' check box and select
>>>> the gluster volume which you have created for storing ISO's or you can type
>>>> the full path of the volume.
>>>> 4) Once this is added please make sure to set the option nfs.disable
>>>> off.
>>>> 5) Now you can go to HE engine and run the command engine-iso-uploader
>>>> upload -i  
>>>>
>>>> Iso gets uploaded successfully.
>>>>
>>>>
>>>> 1st question: do I need to add the network interface to engine in order
>>>> to upload ISOs? does there exist any alternate way?
>>>>
>>>> AFAIK, this is not required when glusterfs volume is used.
>>>>
>>>> Attached is the screenshot where i have only one network attached to my
>>>> HE which is ovirtmgmt.
>>>>
>>>>
>>>> Then I proceeded to configure bonding for the storage domain, bonding 2
>>>> NICs at each server. When trying to set a custom bond of mode=6 (as
>>>> recommended from gluster) I received a warning that mode0, 5 and 6 cannot
>>>> be configured since the interface is used from VMs. I also understood that
>>>> having the storage network assigned to VMs makes it a bridge which
>>>> decreases performance of networking. When trying to remove the network
>>>> interface from engine it cannot be done, since the engine is running.
>>>>
>>>> 2nd question: Is there a way I can remove the secondary storage network
>>>> interface from the engine?
>>>>
>>>> Many thanx
>>>>
>>>>
>>>> ___
>>>> Users mailing 
>>>> listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>>>>
>>>>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Ovirt remove network from hosted engine

2017-06-27 Thread Abi Askushi
Hi all,

Just in case ones needs it, in order to remove the secondary network
interface from engine, you can go to:
Virtual Machines -> Hostedengine -> Network Interfaces -> edit -> unplug it
-> confirm -> remove it.

It was simple...


On Tue, Jun 27, 2017 at 4:54 PM, Abi Askushi <rightkickt...@gmail.com>
wrote:

> Hi Knarra,
>
> Then I had already enabled NFS on ISO gluster volume.
> Maybe i had some networking issue then. I need to remove the secondary
> interface in order to test that again.
>
>
>
> On Tue, Jun 27, 2017 at 4:25 PM, knarra <kna...@redhat.com> wrote:
>
>> On 06/27/2017 06:34 PM, Abi Askushi wrote:
>>
>> Hi Knarra,
>>
>> The ISO domain is of type gluster though I had nfs enabled on that
>> volume.
>>
>> you need to have nfs enabled on the volume. what i meant is nfs.disable
>> off which means nfs is on.
>>
>> For more info please refer to bug https://bugzilla.redhat.com/sh
>> ow_bug.cgi?id=1437799
>>
>> I will disable the nfs and try. Though in order to try I need first to
>> remove that second interface from engine.
>> Is there a way I can remove the secondary storage network interface from
>> the engine?
>>
>> I am not sure how to do that, but   you may shutdown the vm using the
>> command hosted-engine --vm-shutdown which will power off the vm and try to
>> remove the networks using vdsclient. (not sure if this is right, but
>> suggesting a way)
>>
>>
>> Thanx
>>
>>
>>
>> On Tue, Jun 27, 2017 at 3:32 PM, knarra <kna...@redhat.com> wrote:
>>
>>> On 06/27/2017 05:41 PM, Abi Askushi wrote:
>>>
>>> Hi all,
>>>
>>> When setting up hosted engine setup on top gluster with 3 nodes, I had
>>> gluster configured on a separate network interface, as recommended. When I
>>> tried later to upload ISO from engine to ISO domain, the engine was not
>>> able to upload it since the VM did not have access to the separate storage
>>> network. I then added the storage network interface to the hosted engine
>>> and ISO upload succeeded.
>>>
>>> May i know what was the volume type created and added as ISO domain ?
>>>
>>> If you plan to use a glusterfs volume below is the procedure :
>>>
>>> 1) Create a glusterfs volume.
>>> 2) While adding storage domain select Domain Function as 'ISO' and
>>> Storage Type as 'glusterfs' .
>>> 3) You can either use 'use managed gluster volume' check box and select
>>> the gluster volume which you have created for storing ISO's or you can type
>>> the full path of the volume.
>>> 4) Once this is added please make sure to set the option nfs.disable off.
>>> 5) Now you can go to HE engine and run the command engine-iso-uploader
>>> upload -i  
>>>
>>> Iso gets uploaded successfully.
>>>
>>>
>>> 1st question: do I need to add the network interface to engine in order
>>> to upload ISOs? does there exist any alternate way?
>>>
>>> AFAIK, this is not required when glusterfs volume is used.
>>>
>>> Attached is the screenshot where i have only one network attached to my
>>> HE which is ovirtmgmt.
>>>
>>>
>>> Then I proceeded to configure bonding for the storage domain, bonding 2
>>> NICs at each server. When trying to set a custom bond of mode=6 (as
>>> recommended from gluster) I received a warning that mode0, 5 and 6 cannot
>>> be configured since the interface is used from VMs. I also understood that
>>> having the storage network assigned to VMs makes it a bridge which
>>> decreases performance of networking. When trying to remove the network
>>> interface from engine it cannot be done, since the engine is running.
>>>
>>> 2nd question: Is there a way I can remove the secondary storage network
>>> interface from the engine?
>>>
>>> Many thanx
>>>
>>>
>>> ___
>>> Users mailing 
>>> listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>>>
>>>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Ovirt remove network from hosted engine

2017-06-27 Thread Abi Askushi
Hi Knarra,

Then I had already enabled NFS on ISO gluster volume.
Maybe i had some networking issue then. I need to remove the secondary
interface in order to test that again.



On Tue, Jun 27, 2017 at 4:25 PM, knarra <kna...@redhat.com> wrote:

> On 06/27/2017 06:34 PM, Abi Askushi wrote:
>
> Hi Knarra,
>
> The ISO domain is of type gluster though I had nfs enabled on that volume.
>
> you need to have nfs enabled on the volume. what i meant is nfs.disable
> off which means nfs is on.
>
> For more info please refer to bug https://bugzilla.redhat.com/
> show_bug.cgi?id=1437799
>
> I will disable the nfs and try. Though in order to try I need first to
> remove that second interface from engine.
> Is there a way I can remove the secondary storage network interface from
> the engine?
>
> I am not sure how to do that, but   you may shutdown the vm using the
> command hosted-engine --vm-shutdown which will power off the vm and try to
> remove the networks using vdsclient. (not sure if this is right, but
> suggesting a way)
>
>
> Thanx
>
>
>
> On Tue, Jun 27, 2017 at 3:32 PM, knarra <kna...@redhat.com> wrote:
>
>> On 06/27/2017 05:41 PM, Abi Askushi wrote:
>>
>> Hi all,
>>
>> When setting up hosted engine setup on top gluster with 3 nodes, I had
>> gluster configured on a separate network interface, as recommended. When I
>> tried later to upload ISO from engine to ISO domain, the engine was not
>> able to upload it since the VM did not have access to the separate storage
>> network. I then added the storage network interface to the hosted engine
>> and ISO upload succeeded.
>>
>> May i know what was the volume type created and added as ISO domain ?
>>
>> If you plan to use a glusterfs volume below is the procedure :
>>
>> 1) Create a glusterfs volume.
>> 2) While adding storage domain select Domain Function as 'ISO' and
>> Storage Type as 'glusterfs' .
>> 3) You can either use 'use managed gluster volume' check box and select
>> the gluster volume which you have created for storing ISO's or you can type
>> the full path of the volume.
>> 4) Once this is added please make sure to set the option nfs.disable off.
>> 5) Now you can go to HE engine and run the command engine-iso-uploader
>> upload -i  
>>
>> Iso gets uploaded successfully.
>>
>>
>> 1st question: do I need to add the network interface to engine in order
>> to upload ISOs? does there exist any alternate way?
>>
>> AFAIK, this is not required when glusterfs volume is used.
>>
>> Attached is the screenshot where i have only one network attached to my
>> HE which is ovirtmgmt.
>>
>>
>> Then I proceeded to configure bonding for the storage domain, bonding 2
>> NICs at each server. When trying to set a custom bond of mode=6 (as
>> recommended from gluster) I received a warning that mode0, 5 and 6 cannot
>> be configured since the interface is used from VMs. I also understood that
>> having the storage network assigned to VMs makes it a bridge which
>> decreases performance of networking. When trying to remove the network
>> interface from engine it cannot be done, since the engine is running.
>>
>> 2nd question: Is there a way I can remove the secondary storage network
>> interface from the engine?
>>
>> Many thanx
>>
>>
>> ___
>> Users mailing 
>> listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>>
>>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Ovirt remove network from hosted engine

2017-06-27 Thread Abi Askushi
Hi Knarra,

The ISO domain is of type gluster though I had nfs enabled on that volume.
I will disable the nfs and try. Though in order to try I need first to
remove that second interface from engine.
Is there a way I can remove the secondary storage network interface from
the engine?

Thanx

On Tue, Jun 27, 2017 at 3:32 PM, knarra <kna...@redhat.com> wrote:

> On 06/27/2017 05:41 PM, Abi Askushi wrote:
>
> Hi all,
>
> When setting up hosted engine setup on top gluster with 3 nodes, I had
> gluster configured on a separate network interface, as recommended. When I
> tried later to upload ISO from engine to ISO domain, the engine was not
> able to upload it since the VM did not have access to the separate storage
> network. I then added the storage network interface to the hosted engine
> and ISO upload succeeded.
>
> May i know what was the volume type created and added as ISO domain ?
>
> If you plan to use a glusterfs volume below is the procedure :
>
> 1) Create a glusterfs volume.
> 2) While adding storage domain select Domain Function as 'ISO' and Storage
> Type as 'glusterfs' .
> 3) You can either use 'use managed gluster volume' check box and select
> the gluster volume which you have created for storing ISO's or you can type
> the full path of the volume.
> 4) Once this is added please make sure to set the option nfs.disable off.
> 5) Now you can go to HE engine and run the command engine-iso-uploader
> upload -i  
>
> Iso gets uploaded successfully.
>
>
> 1st question: do I need to add the network interface to engine in order to
> upload ISOs? does there exist any alternate way?
>
> AFAIK, this is not required when glusterfs volume is used.
>
> Attached is the screenshot where i have only one network attached to my HE
> which is ovirtmgmt.
>
>
> Then I proceeded to configure bonding for the storage domain, bonding 2
> NICs at each server. When trying to set a custom bond of mode=6 (as
> recommended from gluster) I received a warning that mode0, 5 and 6 cannot
> be configured since the interface is used from VMs. I also understood that
> having the storage network assigned to VMs makes it a bridge which
> decreases performance of networking. When trying to remove the network
> interface from engine it cannot be done, since the engine is running.
>
> 2nd question: Is there a way I can remove the secondary storage network
> interface from the engine?
>
> Many thanx
>
>
> ___
> Users mailing listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Ovirt remove network from hosted engine

2017-06-27 Thread Abi Askushi
Hi all,

When setting up hosted engine setup on top gluster with 3 nodes, I had
gluster configured on a separate network interface, as recommended. When I
tried later to upload ISO from engine to ISO domain, the engine was not
able to upload it since the VM did not have access to the separate storage
network. I then added the storage network interface to the hosted engine
and ISO upload succeeded.

1st question: do I need to add the network interface to engine in order to
upload ISOs? does there exist any alternate way?

Then I proceeded to configure bonding for the storage domain, bonding 2
NICs at each server. When trying to set a custom bond of mode=6 (as
recommended from gluster) I received a warning that mode0, 5 and 6 cannot
be configured since the interface is used from VMs. I also understood that
having the storage network assigned to VMs makes it a bridge which
decreases performance of networking. When trying to remove the network
interface from engine it cannot be done, since the engine is running.

2nd question: Is there a way I can remove the secondary storage network
interface from the engine?

Many thanx
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt split brain resolution

2017-06-27 Thread Abi Askushi
Hi Satheesaran,

gluster volume info engine

Volume Name: engine
Type: Replicate
Volume ID: 3caae601-74dd-40d1-8629-9a61072bec0f
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x (2 + 1) = 3
Transport-type: tcp
Bricks:
Brick1: gluster0:/gluster/engine/brick
Brick2: gluster1:/gluster/engine/brick
Brick3: gluster2:/gluster/engine/brick (arbiter)
Options Reconfigured:
nfs.disable: on
performance.readdir-ahead: on
transport.address-family: inet
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: off
performance.low-prio-threads: 32
network.remote-dio: off
cluster.eager-lock: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 1
features.shard: on
user.cifs: off
storage.owner-uid: 36
storage.owner-gid: 36
network.ping-timeout: 30
performance.strict-o-direct: on
cluster.granular-entry-heal: enable
nfs.export-volumes: on

As per my previous, i have resolved this by following the steps described.


On Tue, Jun 27, 2017 at 1:42 PM, Satheesaran Sundaramoorthi <
sasun...@redhat.com> wrote:

> On Sat, Jun 24, 2017 at 3:17 PM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
>
>> Hi all,
>>
>> For the records, I had to remove manually the conflicting directory and
>> ts respective gfid from the arbiter volume:
>>
>>  getfattr -m . -d -e hex e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
>>
>> That gave me the gfid: 0x277c9caa9dce4a17a2a93775357befd5
>>
>> Then cd .glusterfs/27/7c
>>
>> rm -rf 277c9caa-9dce-4a17-a2a9-3775357befd5 (or move it out of there)
>>
>> Triggerred heal: gluster volume heal engine
>>
>> Then all ok:
>>
>> gluster volume heal engine info
>> Brick gluster0:/gluster/engine/brick
>> Status: Connected
>> Number of entries: 0
>>
>> Brick gluster1:/gluster/engine/brick
>> Status: Connected
>> Number of entries: 0
>>
>> Brick gluster2:/gluster/engine/brick
>> Status: Connected
>> Number of entries: 0
>>
>> Thanx.
>>
>
> ​Hi Abi,
>
> What is the volume type of 'engine' volume ?
> Could you also provide the output of 'gluster volume info engine' to get
> to the
> closer look at the problem
>
> ​-- sas​
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt split brain resolution

2017-06-24 Thread Abi Askushi
Hi all,

For the records, I had to remove manually the conflicting directory and ts
respective gfid from the arbiter volume:

 getfattr -m . -d -e hex e1c80750-b880-495e-9609-b8bc7760d101/ha_agent

That gave me the gfid: 0x277c9caa9dce4a17a2a93775357befd5

Then cd .glusterfs/27/7c

rm -rf 277c9caa-9dce-4a17-a2a9-3775357befd5 (or move it out of there)

Triggerred heal: gluster volume heal engine

Then all ok:

gluster volume heal engine info
Brick gluster0:/gluster/engine/brick
Status: Connected
Number of entries: 0

Brick gluster1:/gluster/engine/brick
Status: Connected
Number of entries: 0

Brick gluster2:/gluster/engine/brick
Status: Connected
Number of entries: 0

Thanx.

On Fri, Jun 23, 2017 at 7:21 PM, Abi Askushi <rightkickt...@gmail.com>
wrote:

> Hi Denis,
>
> I receive permission denied as below:
>
> gluster volume heal engine split-brain latest-mtime
> /e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
> Healing /e1c80750-b880-495e-9609-b8bc7760d101/ha_agent failed:Operation
> not permitted.
> Volume heal failed.
>
>
> When I shutdown host3 then no split brain is reported from the remaining
> two hosts. When I power up host3 then I receive the mentioned split brain
> and host3 logs the following at ovirt-hosted-engine-ha/agent.log
>
> MainThread::INFO::2017-06-23 16:18:06,067::hosted_engine::
> 594::ovirt_hosted_engine_ha.agent.hosted_engine.
> HostedEngine::(_initialize_broker) Failed set the storage domain: 'Failed
> to set storage domain VdsmBackend, options {'hosted-engine.lockspace': '
> 7B22696D6167655F75756964223A202238323132626637382D66392D
> 346465652D61672D346265633734353035366235222C202270617468
> 223A206E756C6C2C2022766F6C756D655F75756964223A20223632373930
> 3162652D666261332D346263342D393037632D393931356138333632633537227D',
> 'sp_uuid': '----', 'dom_type':
> 'glusterfs', 'hosted-engine.metadata': '7B22696D6167655F75756964223A20
> 2263353930633034372D613462322D346539312D613832362D6434386239
> 61643537323330222C202270617468223A206E756C6C2C2022766F6C756D
> 655F75756964223A202230353166653865612D39632D346134302D38
> 3438382D386335313138666438373238227D', 'sd_uuid':
> 'e1c80750-b880-495e-9609-b8bc7760d101'}: Request failed:  'exceptions.OSError'>'. Waiting '5's before the next attempt
>
> and the following at /var/log/messages:
> Jun 23 16:19:43 v2 journal: vdsm root ERROR failed to retrieve Hosted
> Engine HA info#012Traceback (most recent call last):#012  File
> "/usr/lib/python2.7/site-packages/vdsm/host/api.py", line 231, in
> _getHaInfo#012stats = instance.get_all_stats()#012  File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py",
> line 105, in get_all_stats#012stats = 
> broker.get_stats_from_storage(service)#012
> File 
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
> line 233, in get_stats_from_storage#012result =
> self._checked_communicate(request)#012  File "/usr/lib/python2.7/site-
> packages/ovirt_hosted_engine_ha/lib/brokerlink.py", line 261, in
> _checked_communicate#012.format(message or response))#012RequestError:
> Request failed: failed to read metadata: [Errno 5] Input/output error:
> '/rhev/data-center/mnt/glusterSD/10.100.100.1:_engine/e1c80750-b880-495e-
> 9609-b8bc7760d101/ha_agent/hosted-engine.metadata'
>
> Thanx
>
>
> On Fri, Jun 23, 2017 at 6:05 PM, Denis Chaplygin <dchap...@redhat.com>
> wrote:
>
>> Hello Abi,
>>
>> On Fri, Jun 23, 2017 at 4:47 PM, Abi Askushi <rightkickt...@gmail.com>
>> wrote:
>>
>>> Hi All,
>>>
>>> I have a 3 node ovirt 4.1 setup. I lost one node due to raid controller
>>> issues. Upon restoration I have the following split brain, although the
>>> hosts have mounted the storage domains:
>>>
>>> gluster volume heal engine info split-brain
>>> Brick gluster0:/gluster/engine/brick
>>> /e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
>>> Status: Connected
>>> Number of entries in split-brain: 1
>>>
>>> Brick gluster1:/gluster/engine/brick
>>> /e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
>>> Status: Connected
>>> Number of entries in split-brain: 1
>>>
>>> Brick gluster2:/gluster/engine/brick
>>> /e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
>>> Status: Connected
>>> Number of entries in split-brain: 1
>>>
>>>
>>>
>> It is definitely on gluster side. You could try to use
>>
>> gluster volume heal engine split-brain latest-mtime /e1c80750-b880-49
>> 5e-9609-b8bc7760d101/ha_agent
>>
>>
>> I also added gluster developers to that thread, so they may provide you
>> with better advices.
>>
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt split brain resolution

2017-06-23 Thread Abi Askushi
Hi Denis,

I receive permission denied as below:

gluster volume heal engine split-brain latest-mtime
/e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
Healing /e1c80750-b880-495e-9609-b8bc7760d101/ha_agent failed:Operation not
permitted.
Volume heal failed.


When I shutdown host3 then no split brain is reported from the remaining
two hosts. When I power up host3 then I receive the mentioned split brain
and host3 logs the following at ovirt-hosted-engine-ha/agent.log

MainThread::INFO::2017-06-23
16:18:06,067::hosted_engine::594::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_broker)
Failed set the storage domain: 'Failed to set storage domain VdsmBackend,
options {'hosted-engine.lockspace':
'7B22696D6167655F75756964223A202238323132626637382D66392D346465652D61672D346265633734353035366235222C202270617468223A206E756C6C2C2022766F6C756D655F75756964223A202236323739303162652D666261332D346263342D393037632D393931356138333632633537227D',
'sp_uuid': '----', 'dom_type': 'glusterfs',
'hosted-engine.metadata':
'7B22696D6167655F75756964223A202263353930633034372D613462322D346539312D613832362D643438623961643537323330222C202270617468223A206E756C6C2C2022766F6C756D655F75756964223A202230353166653865612D39632D346134302D383438382D386335313138666438373238227D',
'sd_uuid': 'e1c80750-b880-495e-9609-b8bc7760d101'}: Request failed: '. Waiting '5's before the next attempt

and the following at /var/log/messages:
Jun 23 16:19:43 v2 journal: vdsm root ERROR failed to retrieve Hosted
Engine HA info#012Traceback (most recent call last):#012  File
"/usr/lib/python2.7/site-packages/vdsm/host/api.py", line 231, in
_getHaInfo#012stats = instance.get_all_stats()#012  File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py",
line 105, in get_all_stats#012stats =
broker.get_stats_from_storage(service)#012  File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
line 233, in get_stats_from_storage#012result =
self._checked_communicate(request)#012  File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
line 261, in _checked_communicate#012.format(message or
response))#012RequestError: Request failed: failed to read metadata: [Errno
5] Input/output error: '/rhev/data-center/mnt/glusterSD/10.100.100.1:
_engine/e1c80750-b880-495e-9609-b8bc7760d101/ha_agent/hosted-engine.metadata'

Thanx


On Fri, Jun 23, 2017 at 6:05 PM, Denis Chaplygin <dchap...@redhat.com>
wrote:

> Hello Abi,
>
> On Fri, Jun 23, 2017 at 4:47 PM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
>
>> Hi All,
>>
>> I have a 3 node ovirt 4.1 setup. I lost one node due to raid controller
>> issues. Upon restoration I have the following split brain, although the
>> hosts have mounted the storage domains:
>>
>> gluster volume heal engine info split-brain
>> Brick gluster0:/gluster/engine/brick
>> /e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
>> Status: Connected
>> Number of entries in split-brain: 1
>>
>> Brick gluster1:/gluster/engine/brick
>> /e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
>> Status: Connected
>> Number of entries in split-brain: 1
>>
>> Brick gluster2:/gluster/engine/brick
>> /e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
>> Status: Connected
>> Number of entries in split-brain: 1
>>
>>
>>
> It is definitely on gluster side. You could try to use
>
> gluster volume heal engine split-brain latest-mtime /e1c80750-b880-
> 495e-9609-b8bc7760d101/ha_agent
>
>
> I also added gluster developers to that thread, so they may provide you
> with better advices.
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] oVirt split brain resolution

2017-06-23 Thread Abi Askushi
Hi All,

I have a 3 node ovirt 4.1 setup. I lost one node due to raid controller
issues. Upon restoration I have the following split brain, although the
hosts have mounted the storage domains:

gluster volume heal engine info split-brain
Brick gluster0:/gluster/engine/brick
/e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
Status: Connected
Number of entries in split-brain: 1

Brick gluster1:/gluster/engine/brick
/e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
Status: Connected
Number of entries in split-brain: 1

Brick gluster2:/gluster/engine/brick
/e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
Status: Connected
Number of entries in split-brain: 1


Hosted engine status gives the following:

hosted-engine --vm-status
Traceback (most recent call last):
  File "/usr/lib64/python2.7/runpy.py", line 162, in _run_module_as_main
"__main__", fname, loader, pkg_name)
  File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code
exec code in run_globals
  File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_setup/vm_status.py",
line 173, in 
if not status_checker.print_status():
  File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_setup/vm_status.py",
line 103, in print_status
all_host_stats = self._get_all_host_stats()
  File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_setup/vm_status.py",
line 73, in _get_all_host_stats
all_host_stats = ha_cli.get_all_host_stats()
  File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py",
line 160, in get_all_host_stats
return self.get_all_stats(self.StatModes.HOST)
  File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py",
line 105, in get_all_stats
stats = broker.get_stats_from_storage(service)
  File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
line 233, in get_stats_from_storage
result = self._checked_communicate(request)
  File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
line 261, in _checked_communicate
.format(message or response))
ovirt_hosted_engine_ha.lib.exceptions.RequestError: Request failed: failed
to read metadata: [Errno 5] Input/output error:
'/rhev/data-center/mnt/glusterSD/10.100.100.1:
_engine/e1c80750-b880-495e-9609-b8bc7760d101/ha_agent/hosted-engine.metadata'

Any idea on how to resolve this split brain?

Thanx,
Alex
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Win server 2016 guest tools

2017-06-20 Thread Abi Askushi
Thank you Lev.
I confirm that these tools install successfully.

Alex

On Tue, Jun 20, 2017 at 4:04 PM, Lev Veyde <lve...@redhat.com> wrote:

> Hi,
>
> You can use the tools from: http://plain.resources.ovirt.
> org/pub/ovirt-4.1/iso/oVirt-toolsSetup/oVirt-toolsSetup-4.1-5.fc24.iso
>
> It should support MS Windows Server 2016.
>
> Thanks in advance,
>
> On Tue, Jun 20, 2017 at 11:26 AM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
>
>> Hi all,
>>
>> Are there any windows server 2016 guest tools? At the site it doesn't
>> list this OS and during installation from ISO that is included in
>> ovirt-guest-tools-iso package i receive the error that the OS is not
>> supported.
>>
>> If these tools are not installed and I install only virtio tools are
>> there any performance implications for the VM?
>>
>> Thanx,
>> Alex
>>
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>
>
> --
>
> Lev Veyde
>
> Software Engineer, RHCE | RHCVA | MCITP
>
> Red Hat Israel
>
> <https://www.redhat.com>
>
> l...@redhat.com | lve...@redhat.com
> <https://red.ht/sig>
> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Win server 2016 guest tools

2017-06-20 Thread Abi Askushi
Hi all,

Are there any windows server 2016 guest tools? At the site it doesn't list
this OS and during installation from ISO that is included in
ovirt-guest-tools-iso package i receive the error that the OS is not
supported.

If these tools are not installed and I install only virtio tools are there
any performance implications for the VM?

Thanx,
Alex
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt gluster sanlock issue

2017-06-08 Thread Abi Askushi
Filed the *Bug 1459855*
<https://bugzilla.redhat.com/show_bug.cgi?id=1459855>

Alex

On Thu, Jun 8, 2017 at 1:16 PM, Abi Askushi <rightkickt...@gmail.com> wrote:

> Hi Denis,
>
> Ok I will file a bug for this.
> I am not sure if I will be able to provide troubleshooting info for much
> long as I already have put forward the replacement of disks with 512 ones.
>
> Alex
>
> On Thu, Jun 8, 2017 at 11:48 AM, Denis Chaplygin <dchap...@redhat.com>
> wrote:
>
>> Hello Alex,
>>
>>
>> On Wed, Jun 7, 2017 at 11:39 AM, Abi Askushi <rightkickt...@gmail.com>
>> wrote:
>>
>>> Hi Sahina,
>>>
>>> Did you have the chance to check the logs and have any idea how this may
>>> be addressed?
>>>
>>
>>
>> It seems to be a VDSM issue, as VDSM uses direct IO (and id actualy calls
>> dd) and assumes that block size is 512. I see in the code, that block size
>> is defined as a constant, so it probably may be adjusted, but i think it
>> would be better if we ask some one who knows that part better.
>>
>> Anyway, could you please file a bug on that issue? Thanks in advance.
>>
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt gluster sanlock issue

2017-06-08 Thread Abi Askushi
Hi Denis,

Ok I will file a bug for this.
I am not sure if I will be able to provide troubleshooting info for much
long as I already have put forward the replacement of disks with 512 ones.

Alex

On Thu, Jun 8, 2017 at 11:48 AM, Denis Chaplygin <dchap...@redhat.com>
wrote:

> Hello Alex,
>
>
> On Wed, Jun 7, 2017 at 11:39 AM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
>
>> Hi Sahina,
>>
>> Did you have the chance to check the logs and have any idea how this may
>> be addressed?
>>
>
>
> It seems to be a VDSM issue, as VDSM uses direct IO (and id actualy calls
> dd) and assumes that block size is 512. I see in the code, that block size
> is defined as a constant, so it probably may be adjusted, but i think it
> would be better if we ask some one who knows that part better.
>
> Anyway, could you please file a bug on that issue? Thanks in advance.
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt gluster sanlock issue

2017-06-07 Thread Abi Askushi
Hi Sahina,

Did you have the chance to check the logs and have any idea how this may be
addressed?


Thanx,
Alex

On Mon, Jun 5, 2017 at 12:14 PM, Sahina Bose <sab...@redhat.com> wrote:

> Can we have the gluster mount logs and brick logs to check if it's the
> same issue?
>
> On Sun, Jun 4, 2017 at 11:21 PM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
>
>> I clean installed everything and ran into the same.
>> I then ran gdeploy and encountered the same issue when deploying engine.
>> Seems that gluster (?) doesn't like 4K sector drives. I am not sure if it
>> has to do with alignment. The weird thing is that gluster volumes are all
>> ok, replicating normally and no split brain is reported.
>>
>> The solution to the mentioned bug (1386443
>> <https://bugzilla.redhat.com/show_bug.cgi?id=1386443>) was to format
>> with 512 sector size, which for my case is not an option:
>>
>> mkfs.xfs -f -i size=512 -s size=512 /dev/gluster/engine
>> illegal sector size 512; hw sector is 4096
>>
>> Is there any workaround to address this?
>>
>> Thanx,
>> Alex
>>
>>
>> On Sun, Jun 4, 2017 at 5:48 PM, Abi Askushi <rightkickt...@gmail.com>
>> wrote:
>>
>>> Hi Maor,
>>>
>>> My disk are of 4K block size and from this bug seems that gluster
>>> replica needs 512B block size.
>>> Is there a way to make gluster function with 4K drives?
>>>
>>> Thank you!
>>>
>>> On Sun, Jun 4, 2017 at 2:34 PM, Maor Lipchuk <mlipc...@redhat.com>
>>> wrote:
>>>
>>>> Hi Alex,
>>>>
>>>> I saw a bug that might be related to the issue you encountered at
>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1386443
>>>>
>>>> Sahina, maybe you have any advise? Do you think that BZ1386443is
>>>> related?
>>>>
>>>> Regards,
>>>> Maor
>>>>
>>>> On Sat, Jun 3, 2017 at 8:45 PM, Abi Askushi <rightkickt...@gmail.com>
>>>> wrote:
>>>> > Hi All,
>>>> >
>>>> > I have installed successfully several times oVirt (version 4.1) with
>>>> 3 nodes
>>>> > on top glusterfs.
>>>> >
>>>> > This time, when trying to configure the same setup, I am facing the
>>>> > following issue which doesn't seem to go away. During installation i
>>>> get the
>>>> > error:
>>>> >
>>>> > Failed to execute stage 'Misc configuration': Cannot acquire host id:
>>>> > (u'a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922', SanlockException(22,
>>>> 'Sanlock
>>>> > lockspace add failure', 'Invalid argument'))
>>>> >
>>>> > The only different in this setup is that instead of standard
>>>> partitioning i
>>>> > have GPT partitioning and the disks have 4K block size instead of 512.
>>>> >
>>>> > The /var/log/sanlock.log has the following lines:
>>>> >
>>>> > 2017-06-03 19:21:15+0200 23450 [943]: s9 lockspace
>>>> > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047:250:/rhev/data-center/m
>>>> nt/_var_lib_ovirt-hosted-engin-setup_tmptjkIDI/ba6bd862-c2b8
>>>> -46e7-b2c8-91e4a5bb2047/dom_md/ids:0
>>>> > 2017-06-03 19:21:36+0200 23471 [944]: s9:r5 resource
>>>> > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047:SDM:/rhev/data-center/m
>>>> nt/_var_lib_ovirt-hosted-engine-setup_tmptjkIDI/ba6bd862-c2b
>>>> 8-46e7-b2c8-91e4a5bb2047/dom_md/leases:1048576
>>>> > for 2,9,23040
>>>> > 2017-06-03 19:21:36+0200 23471 [943]: s10 lockspace
>>>> > a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922:250:/rhev/data-center/m
>>>> nt/glusterSD/10.100.100.1:_engine/a5a6b0e7-fc3f-4838-8e26-c8
>>>> b4d5e5e922/dom_md/ids:0
>>>> > 2017-06-03 19:21:36+0200 23471 [23522]: a5a6b0e7 aio collect RD
>>>> > 0x7f59b8c0:0x7f59b8d0:0x7f59b0101000 result -22:0 match res
>>>> > 2017-06-03 19:21:36+0200 23471 [23522]: read_sectors delta_leader
>>>> offset
>>>> > 127488 rv -22
>>>> > /rhev/data-center/mnt/glusterSD/10.100.100.1:_engine/a5a6b0e
>>>> 7-fc3f-4838-8e26-c8b4d5e5e922/dom_md/ids
>>>> > 2017-06-03 19:21:37+0200 23472 [930]: s9 host 250 1 23450
>>>> > 88c2244c-a782-40ed-9560-6cfa4d46f853.v0.neptune
>>>> > 2017-06-03 19:21:37+0200 23472 [943]: s10 add_lockspace fail result
>>>> -22
>>>> >

Re: [ovirt-users] oVirt gluster sanlock issue

2017-06-06 Thread Abi Askushi
Just to note that the mentioned logs below are from the dd with bs=512,
which were failing.
Attached the full logs from mount and brick.

Alex

On Tue, Jun 6, 2017 at 3:18 PM, Abi Askushi <rightkickt...@gmail.com> wrote:

> Hi Krutika,
>
> My comments inline.
>
> Also attached the strace of:
> strace -y -ff -o /root/512-trace-on-root.log dd if=/dev/zero
> of=/mnt/test2.img oflag=direct bs=512 count=1
>
> and of:
> strace -y -ff -o /root/4096-trace-on-root.log dd if=/dev/zero
> of=/mnt/test2.img oflag=direct bs=4096 count=16
>
> I have mounted gluster volume at /mnt.
> The dd with bs=4096 is successful.
>
> The gluster mount log gives only the following:
> [2017-06-06 12:04:54.102576] W [MSGID: 114031] 
> [client-rpc-fops.c:854:client3_3_writev_cbk]
> 0-engine-client-0: remote operation failed [Invalid argument]
> [2017-06-06 12:04:54.102591] W [MSGID: 114031] 
> [client-rpc-fops.c:854:client3_3_writev_cbk]
> 0-engine-client-1: remote operation failed [Invalid argument]
> [2017-06-06 12:04:54.103355] W [fuse-bridge.c:2312:fuse_writev_cbk]
> 0-glusterfs-fuse: 205: WRITE => -1 gfid=075ab3a5-0274-4f07-a075-2748c3b4d394
> fd=0x7faf1d08706c (Transport endpoint is not connected)
>
> The gluster brick log gives:
> [2017-06-06 12:07:03.793080] E [MSGID: 113072] [posix.c:3453:posix_writev]
> 0-engine-posix: write failed: offset 0, [Invalid argument]
> [2017-06-06 12:07:03.793172] E [MSGID: 115067] 
> [server-rpc-fops.c:1346:server_writev_cbk]
> 0-engine-server: 291: WRITEV 0 (075ab3a5-0274-4f07-a075-2748c3b4d394) ==>
> (Invalid argument) [Invalid argument]
>
>
>
> On Tue, Jun 6, 2017 at 12:50 PM, Krutika Dhananjay <kdhan...@redhat.com>
> wrote:
>
>> OK.
>>
>> So for the 'Transport endpoint is not connected' issue, could you share
>> the mount and brick logs?
>>
>> Hmmm.. 'Invalid argument' error even on the root partition. What if you
>> change bs to 4096 and run?
>>
> If I use bs=4096 the dd is successful on /root and at gluster mounted
> volume.
>
>>
>> The logs I showed in my earlier mail shows that gluster is merely
>> returning the error it got from the disk file system where the
>> brick is hosted. But you're right about the fact that the offset 127488
>> is not 4K-aligned.
>>
>> If the dd on /root worked for you with bs=4096, could you try the same
>> directly on gluster mount point on a dummy file and capture the strace
>> output of dd?
>> You can perhaps reuse your existing gluster volume by mounting it at
>> another location and doing the dd.
>> Here's what you need to execute:
>>
>> strace -ff -T -p  -o 
>> `
>>
>> FWIW, here's something I found in man(2) open:
>>
>>
>>
>>
>> *Under  Linux  2.4,  transfer  sizes,  and  the alignment of the user
>> buffer and the file offset must all be multiples of the logical block size
>> of the filesystem.  Since Linux 2.6.0, alignment to the logical block size
>> of the   underlying storage (typically 512 bytes) suffices.  The
>> logical block size can be determined using the ioctl(2) BLKSSZGET operation
>> or from the shell using the command:   blockdev --getss*
>>
> Please note also that the physical disks are of 4K sector size (native).
> Thus OS is having 4096/4096 local/physical sector size.
> [root@v0 ~]# blockdev --getss /dev/sda
> 4096
> [root@v0 ~]# blockdev --getpbsz /dev/sda
> 4096
>
>>
>>
>> -Krutika
>>
>>
>> On Tue, Jun 6, 2017 at 1:18 AM, Abi Askushi <rightkickt...@gmail.com>
>> wrote:
>>
>>> Also when testing with dd i get the following:
>>>
>>> *Testing on the gluster mount: *
>>> dd if=/dev/zero of=/rhev/data-center/mnt/glusterSD/10.100.100.1:
>>> _engine/test2.img oflag=direct bs=512 count=1
>>> dd: error writing β/rhev/data-center/mnt/glusterSD/10.100.100.1:
>>> _engine/test2.imgβ: *Transport endpoint is not connected*
>>> 1+0 records in
>>> 0+0 records out
>>> 0 bytes (0 B) copied, 0.00336755 s, 0.0 kB/s
>>>
>>> *Testing on the /root directory (XFS): *
>>> dd if=/dev/zero of=/test2.img oflag=direct bs=512 count=1
>>> dd: error writing β/test2.imgβ:* Invalid argument*
>>> 1+0 records in
>>> 0+0 records out
>>> 0 bytes (0 B) copied, 0.000321239 s, 0.0 kB/s
>>>
>>> Seems that the gluster is trying to do the same and fails.
>>>
>>>
>>>
>>> On Mon, Jun 5, 2017 at 10:10 PM, Abi Askushi <rightkickt...@gmail.com>
>>> wrote:
>>>
>>>> The question that rises is what is needed 

Re: [ovirt-users] oVirt gluster sanlock issue

2017-06-06 Thread Abi Askushi
Hi Krutika,

My comments inline.

Also attached the strace of:
strace -y -ff -o /root/512-trace-on-root.log dd if=/dev/zero
of=/mnt/test2.img oflag=direct bs=512 count=1

and of:
strace -y -ff -o /root/4096-trace-on-root.log dd if=/dev/zero
of=/mnt/test2.img oflag=direct bs=4096 count=16

I have mounted gluster volume at /mnt.
The dd with bs=4096 is successful.

The gluster mount log gives only the following:
[2017-06-06 12:04:54.102576] W [MSGID: 114031]
[client-rpc-fops.c:854:client3_3_writev_cbk] 0-engine-client-0: remote
operation failed [Invalid argument]
[2017-06-06 12:04:54.102591] W [MSGID: 114031]
[client-rpc-fops.c:854:client3_3_writev_cbk] 0-engine-client-1: remote
operation failed [Invalid argument]
[2017-06-06 12:04:54.103355] W [fuse-bridge.c:2312:fuse_writev_cbk]
0-glusterfs-fuse: 205: WRITE => -1
gfid=075ab3a5-0274-4f07-a075-2748c3b4d394 fd=0x7faf1d08706c (Transport
endpoint is not connected)

The gluster brick log gives:
[2017-06-06 12:07:03.793080] E [MSGID: 113072] [posix.c:3453:posix_writev]
0-engine-posix: write failed: offset 0, [Invalid argument]
[2017-06-06 12:07:03.793172] E [MSGID: 115067]
[server-rpc-fops.c:1346:server_writev_cbk] 0-engine-server: 291: WRITEV 0
(075ab3a5-0274-4f07-a075-2748c3b4d394) ==> (Invalid argument) [Invalid
argument]



On Tue, Jun 6, 2017 at 12:50 PM, Krutika Dhananjay <kdhan...@redhat.com>
wrote:

> OK.
>
> So for the 'Transport endpoint is not connected' issue, could you share
> the mount and brick logs?
>
> Hmmm.. 'Invalid argument' error even on the root partition. What if you
> change bs to 4096 and run?
>
If I use bs=4096 the dd is successful on /root and at gluster mounted
volume.

>
> The logs I showed in my earlier mail shows that gluster is merely
> returning the error it got from the disk file system where the
> brick is hosted. But you're right about the fact that the offset 127488 is
> not 4K-aligned.
>
> If the dd on /root worked for you with bs=4096, could you try the same
> directly on gluster mount point on a dummy file and capture the strace
> output of dd?
> You can perhaps reuse your existing gluster volume by mounting it at
> another location and doing the dd.
> Here's what you need to execute:
>
> strace -ff -T -p  -o 
> `
>
> FWIW, here's something I found in man(2) open:
>
>
>
>
> *Under  Linux  2.4,  transfer  sizes,  and  the alignment of the user
> buffer and the file offset must all be multiples of the logical block size
> of the filesystem.  Since Linux 2.6.0, alignment to the logical block size
> of the   underlying storage (typically 512 bytes) suffices.  The
> logical block size can be determined using the ioctl(2) BLKSSZGET operation
> or from the shell using the command:   blockdev --getss*
>
Please note also that the physical disks are of 4K sector size (native).
Thus OS is having 4096/4096 local/physical sector size.
[root@v0 ~]# blockdev --getss /dev/sda
4096
[root@v0 ~]# blockdev --getpbsz /dev/sda
4096

>
>
> -Krutika
>
>
> On Tue, Jun 6, 2017 at 1:18 AM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
>
>> Also when testing with dd i get the following:
>>
>> *Testing on the gluster mount: *
>> dd if=/dev/zero 
>> of=/rhev/data-center/mnt/glusterSD/10.100.100.1:_engine/test2.img
>> oflag=direct bs=512 count=1
>> dd: error writing β/rhev/data-center/mnt/glusterSD/10.100.100.1:
>> _engine/test2.imgβ: *Transport endpoint is not connected*
>> 1+0 records in
>> 0+0 records out
>> 0 bytes (0 B) copied, 0.00336755 s, 0.0 kB/s
>>
>> *Testing on the /root directory (XFS): *
>> dd if=/dev/zero of=/test2.img oflag=direct bs=512 count=1
>> dd: error writing β/test2.imgβ:* Invalid argument*
>> 1+0 records in
>> 0+0 records out
>> 0 bytes (0 B) copied, 0.000321239 s, 0.0 kB/s
>>
>> Seems that the gluster is trying to do the same and fails.
>>
>>
>>
>> On Mon, Jun 5, 2017 at 10:10 PM, Abi Askushi <rightkickt...@gmail.com>
>> wrote:
>>
>>> The question that rises is what is needed to make gluster aware of the
>>> 4K physical sectors presented to it (the logical sector is also 4K). The
>>> offset (127488) at the log does not seem aligned at 4K.
>>>
>>> Alex
>>>
>>> On Mon, Jun 5, 2017 at 2:47 PM, Abi Askushi <rightkickt...@gmail.com>
>>> wrote:
>>>
>>>> Hi Krutika,
>>>>
>>>> I am saying that I am facing this issue with 4k drives. I never
>>>> encountered this issue with 512 drives.
>>>>
>>>> Alex
>>>>
>>>> On Jun 5, 2017 14:26, "Krutika Dhananjay" <kdhan...@redhat.com> wrote:
>>>>
>>>>> This se

Re: [ovirt-users] oVirt gluster sanlock issue

2017-06-05 Thread Abi Askushi
Also when testing with dd i get the following:

*Testing on the gluster mount: *
dd if=/dev/zero
of=/rhev/data-center/mnt/glusterSD/10.100.100.1:_engine/test2.img
oflag=direct bs=512 count=1
dd: error writing
β/rhev/data-center/mnt/glusterSD/10.100.100.1:_engine/test2.imgβ:
*Transport endpoint is not connected*
1+0 records in
0+0 records out
0 bytes (0 B) copied, 0.00336755 s, 0.0 kB/s

*Testing on the /root directory (XFS): *
dd if=/dev/zero of=/test2.img oflag=direct bs=512 count=1
dd: error writing β/test2.imgβ:* Invalid argument*
1+0 records in
0+0 records out
0 bytes (0 B) copied, 0.000321239 s, 0.0 kB/s

Seems that the gluster is trying to do the same and fails.



On Mon, Jun 5, 2017 at 10:10 PM, Abi Askushi <rightkickt...@gmail.com>
wrote:

> The question that rises is what is needed to make gluster aware of the 4K
> physical sectors presented to it (the logical sector is also 4K). The
> offset (127488) at the log does not seem aligned at 4K.
>
> Alex
>
> On Mon, Jun 5, 2017 at 2:47 PM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
>
>> Hi Krutika,
>>
>> I am saying that I am facing this issue with 4k drives. I never
>> encountered this issue with 512 drives.
>>
>> Alex
>>
>> On Jun 5, 2017 14:26, "Krutika Dhananjay" <kdhan...@redhat.com> wrote:
>>
>>> This seems like a case of O_DIRECT reads and writes gone wrong, judging
>>> by the 'Invalid argument' errors.
>>>
>>> The two operations that have failed on gluster bricks are:
>>>
>>> [2017-06-05 09:40:39.428979] E [MSGID: 113072]
>>> [posix.c:3453:posix_writev] 0-engine-posix: write failed: offset 0,
>>> [Invalid argument]
>>> [2017-06-05 09:41:00.865760] E [MSGID: 113040]
>>> [posix.c:3178:posix_readv] 0-engine-posix: read failed on
>>> gfid=8c94f658-ac3c-4e3a-b368-8c038513a914, fd=0x7f408584c06c,
>>> offset=127488 size=512, buf=0x7f4083c0b000 [Invalid argument]
>>>
>>> But then, both the write and the read have 512byte-aligned offset, size
>>> and buf address (which is correct).
>>>
>>> Are you saying you don't see this issue with 4K block-size?
>>>
>>> -Krutika
>>>
>>> On Mon, Jun 5, 2017 at 3:21 PM, Abi Askushi <rightkickt...@gmail.com>
>>> wrote:
>>>
>>>> Hi Sahina,
>>>>
>>>> Attached are the logs. Let me know if sth else is needed.
>>>>
>>>> I have 5 disks (with 4K physical sector) in RAID5. The RAID has 64K
>>>> stripe size at the moment.
>>>> I have prepared the storage as below:
>>>>
>>>> pvcreate --dataalignment 256K /dev/sda4
>>>> vgcreate --physicalextentsize 256K gluster /dev/sda4
>>>>
>>>> lvcreate -n engine --size 120G gluster
>>>> mkfs.xfs -f -i size=512 /dev/gluster/engine
>>>>
>>>> Thanx,
>>>> Alex
>>>>
>>>> On Mon, Jun 5, 2017 at 12:14 PM, Sahina Bose <sab...@redhat.com> wrote:
>>>>
>>>>> Can we have the gluster mount logs and brick logs to check if it's the
>>>>> same issue?
>>>>>
>>>>> On Sun, Jun 4, 2017 at 11:21 PM, Abi Askushi <rightkickt...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> I clean installed everything and ran into the same.
>>>>>> I then ran gdeploy and encountered the same issue when deploying
>>>>>> engine.
>>>>>> Seems that gluster (?) doesn't like 4K sector drives. I am not sure
>>>>>> if it has to do with alignment. The weird thing is that gluster volumes 
>>>>>> are
>>>>>> all ok, replicating normally and no split brain is reported.
>>>>>>
>>>>>> The solution to the mentioned bug (1386443
>>>>>> <https://bugzilla.redhat.com/show_bug.cgi?id=1386443>) was to format
>>>>>> with 512 sector size, which for my case is not an option:
>>>>>>
>>>>>> mkfs.xfs -f -i size=512 -s size=512 /dev/gluster/engine
>>>>>> illegal sector size 512; hw sector is 4096
>>>>>>
>>>>>> Is there any workaround to address this?
>>>>>>
>>>>>> Thanx,
>>>>>> Alex
>>>>>>
>>>>>>
>>>>>> On Sun, Jun 4, 2017 at 5:48 PM, Abi Askushi <rightkickt...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi Maor,
>>>>>>>
>>>>>>&g

Re: [ovirt-users] oVirt gluster sanlock issue

2017-06-05 Thread Abi Askushi
The question that rises is what is needed to make gluster aware of the 4K
physical sectors presented to it (the logical sector is also 4K). The
offset (127488) at the log does not seem aligned at 4K.

Alex

On Mon, Jun 5, 2017 at 2:47 PM, Abi Askushi <rightkickt...@gmail.com> wrote:

> Hi Krutika,
>
> I am saying that I am facing this issue with 4k drives. I never
> encountered this issue with 512 drives.
>
> Alex
>
> On Jun 5, 2017 14:26, "Krutika Dhananjay" <kdhan...@redhat.com> wrote:
>
>> This seems like a case of O_DIRECT reads and writes gone wrong, judging
>> by the 'Invalid argument' errors.
>>
>> The two operations that have failed on gluster bricks are:
>>
>> [2017-06-05 09:40:39.428979] E [MSGID: 113072]
>> [posix.c:3453:posix_writev] 0-engine-posix: write failed: offset 0,
>> [Invalid argument]
>> [2017-06-05 09:41:00.865760] E [MSGID: 113040] [posix.c:3178:posix_readv]
>> 0-engine-posix: read failed on gfid=8c94f658-ac3c-4e3a-b368-8c038513a914,
>> fd=0x7f408584c06c, offset=127488 size=512, buf=0x7f4083c0b000 [Invalid
>> argument]
>>
>> But then, both the write and the read have 512byte-aligned offset, size
>> and buf address (which is correct).
>>
>> Are you saying you don't see this issue with 4K block-size?
>>
>> -Krutika
>>
>> On Mon, Jun 5, 2017 at 3:21 PM, Abi Askushi <rightkickt...@gmail.com>
>> wrote:
>>
>>> Hi Sahina,
>>>
>>> Attached are the logs. Let me know if sth else is needed.
>>>
>>> I have 5 disks (with 4K physical sector) in RAID5. The RAID has 64K
>>> stripe size at the moment.
>>> I have prepared the storage as below:
>>>
>>> pvcreate --dataalignment 256K /dev/sda4
>>> vgcreate --physicalextentsize 256K gluster /dev/sda4
>>>
>>> lvcreate -n engine --size 120G gluster
>>> mkfs.xfs -f -i size=512 /dev/gluster/engine
>>>
>>> Thanx,
>>> Alex
>>>
>>> On Mon, Jun 5, 2017 at 12:14 PM, Sahina Bose <sab...@redhat.com> wrote:
>>>
>>>> Can we have the gluster mount logs and brick logs to check if it's the
>>>> same issue?
>>>>
>>>> On Sun, Jun 4, 2017 at 11:21 PM, Abi Askushi <rightkickt...@gmail.com>
>>>> wrote:
>>>>
>>>>> I clean installed everything and ran into the same.
>>>>> I then ran gdeploy and encountered the same issue when deploying
>>>>> engine.
>>>>> Seems that gluster (?) doesn't like 4K sector drives. I am not sure if
>>>>> it has to do with alignment. The weird thing is that gluster volumes are
>>>>> all ok, replicating normally and no split brain is reported.
>>>>>
>>>>> The solution to the mentioned bug (1386443
>>>>> <https://bugzilla.redhat.com/show_bug.cgi?id=1386443>) was to format
>>>>> with 512 sector size, which for my case is not an option:
>>>>>
>>>>> mkfs.xfs -f -i size=512 -s size=512 /dev/gluster/engine
>>>>> illegal sector size 512; hw sector is 4096
>>>>>
>>>>> Is there any workaround to address this?
>>>>>
>>>>> Thanx,
>>>>> Alex
>>>>>
>>>>>
>>>>> On Sun, Jun 4, 2017 at 5:48 PM, Abi Askushi <rightkickt...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hi Maor,
>>>>>>
>>>>>> My disk are of 4K block size and from this bug seems that gluster
>>>>>> replica needs 512B block size.
>>>>>> Is there a way to make gluster function with 4K drives?
>>>>>>
>>>>>> Thank you!
>>>>>>
>>>>>> On Sun, Jun 4, 2017 at 2:34 PM, Maor Lipchuk <mlipc...@redhat.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi Alex,
>>>>>>>
>>>>>>> I saw a bug that might be related to the issue you encountered at
>>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1386443
>>>>>>>
>>>>>>> Sahina, maybe you have any advise? Do you think that BZ1386443is
>>>>>>> related?
>>>>>>>
>>>>>>> Regards,
>>>>>>> Maor
>>>>>>>
>>>>>>> On Sat, Jun 3, 2017 at 8:45 PM, Abi Askushi <rightkickt...@gmail.com>
>>>>>>> wrote:
>>>>>>> > Hi All,
>&

Re: [ovirt-users] oVirt gluster sanlock issue

2017-06-05 Thread Abi Askushi
Hi Krutika,

I am saying that I am facing this issue with 4k drives. I never encountered
this issue with 512 drives.

Alex

On Jun 5, 2017 14:26, "Krutika Dhananjay" <kdhan...@redhat.com> wrote:

> This seems like a case of O_DIRECT reads and writes gone wrong, judging by
> the 'Invalid argument' errors.
>
> The two operations that have failed on gluster bricks are:
>
> [2017-06-05 09:40:39.428979] E [MSGID: 113072] [posix.c:3453:posix_writev]
> 0-engine-posix: write failed: offset 0, [Invalid argument]
> [2017-06-05 09:41:00.865760] E [MSGID: 113040] [posix.c:3178:posix_readv]
> 0-engine-posix: read failed on gfid=8c94f658-ac3c-4e3a-b368-8c038513a914,
> fd=0x7f408584c06c, offset=127488 size=512, buf=0x7f4083c0b000 [Invalid
> argument]
>
> But then, both the write and the read have 512byte-aligned offset, size
> and buf address (which is correct).
>
> Are you saying you don't see this issue with 4K block-size?
>
> -Krutika
>
> On Mon, Jun 5, 2017 at 3:21 PM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
>
>> Hi Sahina,
>>
>> Attached are the logs. Let me know if sth else is needed.
>>
>> I have 5 disks (with 4K physical sector) in RAID5. The RAID has 64K
>> stripe size at the moment.
>> I have prepared the storage as below:
>>
>> pvcreate --dataalignment 256K /dev/sda4
>> vgcreate --physicalextentsize 256K gluster /dev/sda4
>>
>> lvcreate -n engine --size 120G gluster
>> mkfs.xfs -f -i size=512 /dev/gluster/engine
>>
>> Thanx,
>> Alex
>>
>> On Mon, Jun 5, 2017 at 12:14 PM, Sahina Bose <sab...@redhat.com> wrote:
>>
>>> Can we have the gluster mount logs and brick logs to check if it's the
>>> same issue?
>>>
>>> On Sun, Jun 4, 2017 at 11:21 PM, Abi Askushi <rightkickt...@gmail.com>
>>> wrote:
>>>
>>>> I clean installed everything and ran into the same.
>>>> I then ran gdeploy and encountered the same issue when deploying
>>>> engine.
>>>> Seems that gluster (?) doesn't like 4K sector drives. I am not sure if
>>>> it has to do with alignment. The weird thing is that gluster volumes are
>>>> all ok, replicating normally and no split brain is reported.
>>>>
>>>> The solution to the mentioned bug (1386443
>>>> <https://bugzilla.redhat.com/show_bug.cgi?id=1386443>) was to format
>>>> with 512 sector size, which for my case is not an option:
>>>>
>>>> mkfs.xfs -f -i size=512 -s size=512 /dev/gluster/engine
>>>> illegal sector size 512; hw sector is 4096
>>>>
>>>> Is there any workaround to address this?
>>>>
>>>> Thanx,
>>>> Alex
>>>>
>>>>
>>>> On Sun, Jun 4, 2017 at 5:48 PM, Abi Askushi <rightkickt...@gmail.com>
>>>> wrote:
>>>>
>>>>> Hi Maor,
>>>>>
>>>>> My disk are of 4K block size and from this bug seems that gluster
>>>>> replica needs 512B block size.
>>>>> Is there a way to make gluster function with 4K drives?
>>>>>
>>>>> Thank you!
>>>>>
>>>>> On Sun, Jun 4, 2017 at 2:34 PM, Maor Lipchuk <mlipc...@redhat.com>
>>>>> wrote:
>>>>>
>>>>>> Hi Alex,
>>>>>>
>>>>>> I saw a bug that might be related to the issue you encountered at
>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1386443
>>>>>>
>>>>>> Sahina, maybe you have any advise? Do you think that BZ1386443is
>>>>>> related?
>>>>>>
>>>>>> Regards,
>>>>>> Maor
>>>>>>
>>>>>> On Sat, Jun 3, 2017 at 8:45 PM, Abi Askushi <rightkickt...@gmail.com>
>>>>>> wrote:
>>>>>> > Hi All,
>>>>>> >
>>>>>> > I have installed successfully several times oVirt (version 4.1)
>>>>>> with 3 nodes
>>>>>> > on top glusterfs.
>>>>>> >
>>>>>> > This time, when trying to configure the same setup, I am facing the
>>>>>> > following issue which doesn't seem to go away. During installation
>>>>>> i get the
>>>>>> > error:
>>>>>> >
>>>>>> > Failed to execute stage 'Misc configuration': Cannot acquire host
>>>>>> id:
>>>>>> > (u'a5a6b0e7-fc3f-4838-8e26-c8b4d5e

Re: [ovirt-users] oVirt gluster sanlock issue

2017-06-05 Thread Abi Askushi
Hi Sahina,

Attached are the logs. Let me know if sth else is needed.

I have 5 disks (with 4K physical sector) in RAID5. The RAID has 64K stripe
size at the moment.
I have prepared the storage as below:

pvcreate --dataalignment 256K /dev/sda4
vgcreate --physicalextentsize 256K gluster /dev/sda4

lvcreate -n engine --size 120G gluster
mkfs.xfs -f -i size=512 /dev/gluster/engine

Thanx,
Alex

On Mon, Jun 5, 2017 at 12:14 PM, Sahina Bose <sab...@redhat.com> wrote:

> Can we have the gluster mount logs and brick logs to check if it's the
> same issue?
>
> On Sun, Jun 4, 2017 at 11:21 PM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
>
>> I clean installed everything and ran into the same.
>> I then ran gdeploy and encountered the same issue when deploying engine.
>> Seems that gluster (?) doesn't like 4K sector drives. I am not sure if it
>> has to do with alignment. The weird thing is that gluster volumes are all
>> ok, replicating normally and no split brain is reported.
>>
>> The solution to the mentioned bug (1386443
>> <https://bugzilla.redhat.com/show_bug.cgi?id=1386443>) was to format
>> with 512 sector size, which for my case is not an option:
>>
>> mkfs.xfs -f -i size=512 -s size=512 /dev/gluster/engine
>> illegal sector size 512; hw sector is 4096
>>
>> Is there any workaround to address this?
>>
>> Thanx,
>> Alex
>>
>>
>> On Sun, Jun 4, 2017 at 5:48 PM, Abi Askushi <rightkickt...@gmail.com>
>> wrote:
>>
>>> Hi Maor,
>>>
>>> My disk are of 4K block size and from this bug seems that gluster
>>> replica needs 512B block size.
>>> Is there a way to make gluster function with 4K drives?
>>>
>>> Thank you!
>>>
>>> On Sun, Jun 4, 2017 at 2:34 PM, Maor Lipchuk <mlipc...@redhat.com>
>>> wrote:
>>>
>>>> Hi Alex,
>>>>
>>>> I saw a bug that might be related to the issue you encountered at
>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1386443
>>>>
>>>> Sahina, maybe you have any advise? Do you think that BZ1386443is
>>>> related?
>>>>
>>>> Regards,
>>>> Maor
>>>>
>>>> On Sat, Jun 3, 2017 at 8:45 PM, Abi Askushi <rightkickt...@gmail.com>
>>>> wrote:
>>>> > Hi All,
>>>> >
>>>> > I have installed successfully several times oVirt (version 4.1) with
>>>> 3 nodes
>>>> > on top glusterfs.
>>>> >
>>>> > This time, when trying to configure the same setup, I am facing the
>>>> > following issue which doesn't seem to go away. During installation i
>>>> get the
>>>> > error:
>>>> >
>>>> > Failed to execute stage 'Misc configuration': Cannot acquire host id:
>>>> > (u'a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922', SanlockException(22,
>>>> 'Sanlock
>>>> > lockspace add failure', 'Invalid argument'))
>>>> >
>>>> > The only different in this setup is that instead of standard
>>>> partitioning i
>>>> > have GPT partitioning and the disks have 4K block size instead of 512.
>>>> >
>>>> > The /var/log/sanlock.log has the following lines:
>>>> >
>>>> > 2017-06-03 19:21:15+0200 23450 [943]: s9 lockspace
>>>> > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047:250:/rhev/data-center/m
>>>> nt/_var_lib_ovirt-hosted-engin-setup_tmptjkIDI/ba6bd862-c2b8
>>>> -46e7-b2c8-91e4a5bb2047/dom_md/ids:0
>>>> > 2017-06-03 19:21:36+0200 23471 [944]: s9:r5 resource
>>>> > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047:SDM:/rhev/data-center/m
>>>> nt/_var_lib_ovirt-hosted-engine-setup_tmptjkIDI/ba6bd862-c2b
>>>> 8-46e7-b2c8-91e4a5bb2047/dom_md/leases:1048576
>>>> > for 2,9,23040
>>>> > 2017-06-03 19:21:36+0200 23471 [943]: s10 lockspace
>>>> > a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922:250:/rhev/data-center/m
>>>> nt/glusterSD/10.100.100.1:_engine/a5a6b0e7-fc3f-4838-8e26-c8
>>>> b4d5e5e922/dom_md/ids:0
>>>> > 2017-06-03 19:21:36+0200 23471 [23522]: a5a6b0e7 aio collect RD
>>>> > 0x7f59b8c0:0x7f59b8d0:0x7f59b0101000 result -22:0 match res
>>>> > 2017-06-03 19:21:36+0200 23471 [23522]: read_sectors delta_leader
>>>> offset
>>>> > 127488 rv -22
>>>> > /rhev/data-center/mnt/glusterSD/10.100.100.1:_engine/a5a6b0e
>>>> 7-fc3f-4838-8e26-c8b4d5e5e922/dom_

Re: [ovirt-users] oVirt gluster sanlock issue

2017-06-04 Thread Abi Askushi
I clean installed everything and ran into the same.
I then ran gdeploy and encountered the same issue when deploying engine.
Seems that gluster (?) doesn't like 4K sector drives. I am not sure if it
has to do with alignment. The weird thing is that gluster volumes are all
ok, replicating normally and no split brain is reported.

The solution to the mentioned bug (1386443
<https://bugzilla.redhat.com/show_bug.cgi?id=1386443>) was to format with
512 sector size, which for my case is not an option:

mkfs.xfs -f -i size=512 -s size=512 /dev/gluster/engine
illegal sector size 512; hw sector is 4096

Is there any workaround to address this?

Thanx,
Alex


On Sun, Jun 4, 2017 at 5:48 PM, Abi Askushi <rightkickt...@gmail.com> wrote:

> Hi Maor,
>
> My disk are of 4K block size and from this bug seems that gluster replica
> needs 512B block size.
> Is there a way to make gluster function with 4K drives?
>
> Thank you!
>
> On Sun, Jun 4, 2017 at 2:34 PM, Maor Lipchuk <mlipc...@redhat.com> wrote:
>
>> Hi Alex,
>>
>> I saw a bug that might be related to the issue you encountered at
>> https://bugzilla.redhat.com/show_bug.cgi?id=1386443
>>
>> Sahina, maybe you have any advise? Do you think that BZ1386443is related?
>>
>> Regards,
>> Maor
>>
>> On Sat, Jun 3, 2017 at 8:45 PM, Abi Askushi <rightkickt...@gmail.com>
>> wrote:
>> > Hi All,
>> >
>> > I have installed successfully several times oVirt (version 4.1) with 3
>> nodes
>> > on top glusterfs.
>> >
>> > This time, when trying to configure the same setup, I am facing the
>> > following issue which doesn't seem to go away. During installation i
>> get the
>> > error:
>> >
>> > Failed to execute stage 'Misc configuration': Cannot acquire host id:
>> > (u'a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922', SanlockException(22, 'Sanlock
>> > lockspace add failure', 'Invalid argument'))
>> >
>> > The only different in this setup is that instead of standard
>> partitioning i
>> > have GPT partitioning and the disks have 4K block size instead of 512.
>> >
>> > The /var/log/sanlock.log has the following lines:
>> >
>> > 2017-06-03 19:21:15+0200 23450 [943]: s9 lockspace
>> > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047:250:/rhev/data-center/
>> mnt/_var_lib_ovirt-hosted-engin-setup_tmptjkIDI/ba6bd862
>> -c2b8-46e7-b2c8-91e4a5bb2047/dom_md/ids:0
>> > 2017-06-03 19:21:36+0200 23471 [944]: s9:r5 resource
>> > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047:SDM:/rhev/data-center/
>> mnt/_var_lib_ovirt-hosted-engine-setup_tmptjkIDI/ba6bd86
>> 2-c2b8-46e7-b2c8-91e4a5bb2047/dom_md/leases:1048576
>> > for 2,9,23040
>> > 2017-06-03 19:21:36+0200 23471 [943]: s10 lockspace
>> > a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922:250:/rhev/data-center/
>> mnt/glusterSD/10.100.100.1:_engine/a5a6b0e7-fc3f-4838-
>> 8e26-c8b4d5e5e922/dom_md/ids:0
>> > 2017-06-03 19:21:36+0200 23471 [23522]: a5a6b0e7 aio collect RD
>> > 0x7f59b8c0:0x7f59b8d0:0x7f59b0101000 result -22:0 match res
>> > 2017-06-03 19:21:36+0200 23471 [23522]: read_sectors delta_leader offset
>> > 127488 rv -22
>> > /rhev/data-center/mnt/glusterSD/10.100.100.1:_engine/
>> a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922/dom_md/ids
>> > 2017-06-03 19:21:37+0200 23472 [930]: s9 host 250 1 23450
>> > 88c2244c-a782-40ed-9560-6cfa4d46f853.v0.neptune
>> > 2017-06-03 19:21:37+0200 23472 [943]: s10 add_lockspace fail result -22
>> >
>> > And /var/log/vdsm/vdsm.log says:
>> >
>> > 2017-06-03 19:19:38,176+0200 WARN  (jsonrpc/3)
>> > [storage.StorageServer.MountConnection] Using user specified
>> > backup-volfile-servers option (storageServer:253)
>> > 2017-06-03 19:21:12,379+0200 WARN  (periodic/1) [throttled] MOM not
>> > available. (throttledlog:105)
>> > 2017-06-03 19:21:12,380+0200 WARN  (periodic/1) [throttled] MOM not
>> > available, KSM stats will be missing. (throttledlog:105)
>> > 2017-06-03 19:21:14,714+0200 WARN  (jsonrpc/1)
>> > [storage.StorageServer.MountConnection] Using user specified
>> > backup-volfile-servers option (storageServer:253)
>> > 2017-06-03 19:21:15,515+0200 ERROR (jsonrpc/4) [storage.initSANLock]
>> Cannot
>> > initialize SANLock for domain a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922
>> > (clusterlock:238)
>> > Traceback (most recent call last):
>> >   File "/usr/lib/python2.7/site-packages/vdsm/storage/clusterlock.py",
>> line
>> > 234, in initSANLock
>> > sanlock.init_lockspace(sdUUID

Re: [ovirt-users] oVirt gluster sanlock issue

2017-06-04 Thread Abi Askushi
Hi Maor,

My disk are of 4K block size and from this bug seems that gluster replica
needs 512B block size.
Is there a way to make gluster function with 4K drives?

Thank you!

On Sun, Jun 4, 2017 at 2:34 PM, Maor Lipchuk <mlipc...@redhat.com> wrote:

> Hi Alex,
>
> I saw a bug that might be related to the issue you encountered at
> https://bugzilla.redhat.com/show_bug.cgi?id=1386443
>
> Sahina, maybe you have any advise? Do you think that BZ1386443is related?
>
> Regards,
> Maor
>
> On Sat, Jun 3, 2017 at 8:45 PM, Abi Askushi <rightkickt...@gmail.com>
> wrote:
> > Hi All,
> >
> > I have installed successfully several times oVirt (version 4.1) with 3
> nodes
> > on top glusterfs.
> >
> > This time, when trying to configure the same setup, I am facing the
> > following issue which doesn't seem to go away. During installation i get
> the
> > error:
> >
> > Failed to execute stage 'Misc configuration': Cannot acquire host id:
> > (u'a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922', SanlockException(22, 'Sanlock
> > lockspace add failure', 'Invalid argument'))
> >
> > The only different in this setup is that instead of standard
> partitioning i
> > have GPT partitioning and the disks have 4K block size instead of 512.
> >
> > The /var/log/sanlock.log has the following lines:
> >
> > 2017-06-03 19:21:15+0200 23450 [943]: s9 lockspace
> > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047:250:/rhev/data-
> center/mnt/_var_lib_ovirt-hosted-engin-setup_tmptjkIDI/
> ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047/dom_md/ids:0
> > 2017-06-03 19:21:36+0200 23471 [944]: s9:r5 resource
> > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047:SDM:/rhev/data-
> center/mnt/_var_lib_ovirt-hosted-engine-setup_tmptjkIDI/
> ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047/dom_md/leases:1048576
> > for 2,9,23040
> > 2017-06-03 19:21:36+0200 23471 [943]: s10 lockspace
> > a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922:250:/rhev/data-
> center/mnt/glusterSD/10.100.100.1:_engine/a5a6b0e7-fc3f-
> 4838-8e26-c8b4d5e5e922/dom_md/ids:0
> > 2017-06-03 19:21:36+0200 23471 [23522]: a5a6b0e7 aio collect RD
> > 0x7f59b8c0:0x7f59b8d0:0x7f59b0101000 result -22:0 match res
> > 2017-06-03 19:21:36+0200 23471 [23522]: read_sectors delta_leader offset
> > 127488 rv -22
> > /rhev/data-center/mnt/glusterSD/10.100.100.1:_engine/a5a6b0e7-fc3f-4838-
> 8e26-c8b4d5e5e922/dom_md/ids
> > 2017-06-03 19:21:37+0200 23472 [930]: s9 host 250 1 23450
> > 88c2244c-a782-40ed-9560-6cfa4d46f853.v0.neptune
> > 2017-06-03 19:21:37+0200 23472 [943]: s10 add_lockspace fail result -22
> >
> > And /var/log/vdsm/vdsm.log says:
> >
> > 2017-06-03 19:19:38,176+0200 WARN  (jsonrpc/3)
> > [storage.StorageServer.MountConnection] Using user specified
> > backup-volfile-servers option (storageServer:253)
> > 2017-06-03 19:21:12,379+0200 WARN  (periodic/1) [throttled] MOM not
> > available. (throttledlog:105)
> > 2017-06-03 19:21:12,380+0200 WARN  (periodic/1) [throttled] MOM not
> > available, KSM stats will be missing. (throttledlog:105)
> > 2017-06-03 19:21:14,714+0200 WARN  (jsonrpc/1)
> > [storage.StorageServer.MountConnection] Using user specified
> > backup-volfile-servers option (storageServer:253)
> > 2017-06-03 19:21:15,515+0200 ERROR (jsonrpc/4) [storage.initSANLock]
> Cannot
> > initialize SANLock for domain a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922
> > (clusterlock:238)
> > Traceback (most recent call last):
> >   File "/usr/lib/python2.7/site-packages/vdsm/storage/clusterlock.py",
> line
> > 234, in initSANLock
> > sanlock.init_lockspace(sdUUID, idsPath)
> > SanlockException: (107, 'Sanlock lockspace init failure', 'Transport
> > endpoint is not connected')
> > 2017-06-03 19:21:15,515+0200 WARN  (jsonrpc/4)
> > [storage.StorageDomainManifest] lease did not initialize successfully
> > (sd:557)
> > Traceback (most recent call last):
> >   File "/usr/share/vdsm/storage/sd.py", line 552, in initDomainLock
> > self._domainLock.initLock(self.getDomainLease())
> >   File "/usr/lib/python2.7/site-packages/vdsm/storage/clusterlock.py",
> line
> > 271, in initLock
> > initSANLock(self._sdUUID, self._idsPath, lease)
> >   File "/usr/lib/python2.7/site-packages/vdsm/storage/clusterlock.py",
> line
> > 239, in initSANLock
> > raise se.ClusterLockInitError()
> > ClusterLockInitError: Could not initialize cluster lock: ()
> > 2017-06-03 19:21:37,867+0200 ERROR (jsonrpc/2) [storage.StoragePool]
> Create
> > pool hosted_datacenter canceled  (sp:655)
> > Traceback (most recen

[ovirt-users] oVirt gluster sanlock issue

2017-06-03 Thread Abi Askushi
Hi All,

I have installed successfully several times oVirt (version 4.1) with 3
nodes on top glusterfs.

This time, when trying to configure the same setup, I am facing the
following issue which doesn't seem to go away. During installation i get
the error:

Failed to execute stage 'Misc configuration': Cannot acquire host id:
(u'a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922', SanlockException(22, 'Sanlock
lockspace add failure', 'Invalid argument'))

The only different in this setup is that instead of standard partitioning i
have GPT partitioning and the disks have 4K block size instead of 512.

The /var/log/sanlock.log has the following lines:

2017-06-03 19:21:15+0200 23450 [943]: s9 lockspace
ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047:250:/rhev/data-center/mnt/_var_lib_ovirt-hosted-engin-setup_tmptjkIDI/ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047/dom_md/ids:0
2017-06-03 19:21:36+0200 23471 [944]: s9:r5 resource
ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047:SDM:/rhev/data-center/mnt/_var_lib_ovirt-hosted-engine-setup_tmptjkIDI/ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047/dom_md/leases:1048576
for 2,9,23040
2017-06-03 19:21:36+0200 23471 [943]: s10 lockspace
a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922:250:/rhev/data-center/mnt/glusterSD/10.100.100.1:
_engine/a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922/dom_md/ids:0
2017-06-03 19:21:36+0200 23471 [23522]: a5a6b0e7 aio collect RD
0x7f59b8c0:0x7f59b8d0:0x7f59b0101000 result -22:0 match res
2017-06-03 19:21:36+0200 23471 [23522]: read_sectors delta_leader offset
127488 rv -22 /rhev/data-center/mnt/glusterSD/10.100.100.1:
_engine/a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922/dom_md/ids
2017-06-03 19:21:37+0200 23472 [930]: s9 host 250 1 23450
88c2244c-a782-40ed-9560-6cfa4d46f853.v0.neptune
2017-06-03 19:21:37+0200 23472 [943]: s10 add_lockspace fail result -22

And /var/log/vdsm/vdsm.log says:

2017-06-03 19:19:38,176+0200 WARN  (jsonrpc/3)
[storage.StorageServer.MountConnection] Using user specified
backup-volfile-servers option (storageServer:253)
2017-06-03 19:21:12,379+0200 WARN  (periodic/1) [throttled] MOM not
available. (throttledlog:105)
2017-06-03 19:21:12,380+0200 WARN  (periodic/1) [throttled] MOM not
available, KSM stats will be missing. (throttledlog:105)
2017-06-03 19:21:14,714+0200 WARN  (jsonrpc/1)
[storage.StorageServer.MountConnection] Using user specified
backup-volfile-servers option (storageServer:253)
2017-06-03 19:21:15,515+0200 ERROR (jsonrpc/4) [storage.initSANLock] Cannot
initialize SANLock for domain a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922
(clusterlock:238)
Traceback (most recent call last):
  File "/usr/lib/python2.7/site-packages/vdsm/storage/clusterlock.py", line
234, in initSANLock
sanlock.init_lockspace(sdUUID, idsPath)
SanlockException: (107, 'Sanlock lockspace init failure', 'Transport
endpoint is not connected')
2017-06-03 19:21:15,515+0200 WARN  (jsonrpc/4)
[storage.StorageDomainManifest] lease did not initialize successfully
(sd:557)
Traceback (most recent call last):
  File "/usr/share/vdsm/storage/sd.py", line 552, in initDomainLock
self._domainLock.initLock(self.getDomainLease())
  File "/usr/lib/python2.7/site-packages/vdsm/storage/clusterlock.py", line
271, in initLock
initSANLock(self._sdUUID, self._idsPath, lease)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/clusterlock.py", line
239, in initSANLock
raise se.ClusterLockInitError()
ClusterLockInitError: Could not initialize cluster lock: ()
2017-06-03 19:21:37,867+0200 ERROR (jsonrpc/2) [storage.StoragePool] Create
pool hosted_datacenter canceled  (sp:655)
Traceback (most recent call last):
  File "/usr/share/vdsm/storage/sp.py", line 652, in create
self.attachSD(sdUUID)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", line
79, in wrapper
return method(self, *args, **kwargs)
  File "/usr/share/vdsm/storage/sp.py", line 971, in attachSD
dom.acquireHostId(self.id)
  File "/usr/share/vdsm/storage/sd.py", line 790, in acquireHostId
self._manifest.acquireHostId(hostId, async)
  File "/usr/share/vdsm/storage/sd.py", line 449, in acquireHostId
self._domainLock.acquireHostId(hostId, async)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/clusterlock.py", line
297, in acquireHostId
raise se.AcquireHostIdFailure(self._sdUUID, e)
AcquireHostIdFailure: Cannot acquire host id:
(u'a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922', SanlockException(22, 'Sanlock
lockspace add failure', 'Invalid argument'))
2017-06-03 19:21:37,870+0200 ERROR (jsonrpc/2) [storage.StoragePool] Domain
ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047 detach from MSD
ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047 Ver 1 failed. (sp:528)
Traceback (most recent call last):
  File "/usr/share/vdsm/storage/sp.py", line 525, in __cleanupDomains
self.detachSD(sdUUID)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", line
79, in wrapper
return method(self, *args, **kwargs)
  File "/usr/share/vdsm/storage/sp.py", line 1046, in detachSD
raise se.CannotDetachMasterStorageDomain(sdUUID)

[ovirt-users] oVirt Power Management with Cisco UCS C220 M4S

2017-05-20 Thread Abi Askushi
Hi All,

For anyone that might stumble on a Cisco UCS C220 M4S and wondering how to
configure power management. below are the steps to configure it, as it took
me some hours to figure it out...

1. enable IPMI on server. (Cisco has this documented)

2. at ovirt GUI, edit host -> power management, then select "ipmilan" and
add *lanplus=1* as an option. (the bold one was the tricky part)

To test from command line:
 ipmitool -I lanplus -H  -U admin -P somepass -v chassis power
status
It will give the response: "Chassis Power is on"

Alex
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] oVirt global utilization trend

2017-03-07 Thread Abi Askushi
Hi,

I see that the global utilization 24h trend graph (CPU, RAM, Disk) at
dashboard is showing only the last 3 hours. The line for previous hours is
flat (see attached). I remember that on a previous installation I had
tested it was showing the full 24 hours. Tried to run again engine-setup
but issue remains.
Any suggestions?

Thanx,
Alex
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users