Re: [ovirt-users] problem while updating oVirt from 3.6 to 4.0

2017-05-13 Thread Yedidyah Bar David
On Thu, May 11, 2017 at 7:11 PM, Nelson Lameiras <
nelson.lamei...@lyra-network.com> wrote:

> Hello,
>
> Not sure if I should post here, or on the dev mailling list.
>
> I had a serious problem when upgrading an oVirt cluster from 3.6 to 4.0
> (before updating to 4.1)
>
> The error on engine-setup log was :
> "schema.sh: FATAL: Cannot execute sql command: --file=/usr/share/ovirt-
> engine/dbscripts/upgrade/04_00_0930_disallow_boot_protocol_to_be_null.sql"
>
> After long hours of debuguing and analysing I found out the problem. On
> database, some entries on table network_attachements had a NULL value which
> were incompatible with the sql file above.
> I'm guessing this is an "exotic situation" which shouldn't exist... in
> theory ;)
>
> Anyway, I'm fairly sure that the solution for this problem would be to
> patch the following file 'ovirt-engine/packaging/
> dbscripts/upgrade/04_00_0910_do_not_allow_null_boot_protocol.sql'
> by adding the following line "update network_attachments set boot_protocol
> ='NONE' where boot_protocol is null;"
>
> If a dev want to discuss it further, I'll be happy to explain my thoughts
> on this matter.
>

Doing the above might be enough for engine-setup to finish, but it might
introduce a risk.
Ideally we should find the root cause leading to this state and:
1. Make sure it can't happen anymore
2. Fully fix the data - might require more than just the above update.
3. Alternatively, add a check earlier and notify the user to fix this
manually.
Adding Dan for further analysis.

Thanks for the report!

Best,


>
> cordialement, regards,
>
> 
> Nelson LAMEIRAS
> Ingénieur Systèmes et Réseaux / Systems and Networks engineer
> Tel: +33 5 32 09 09 70 <+33%205%2032%2009%2009%2070>
> nelson.lamei...@lyra-network.com
> www.lyra-network.com | www.payzen.eu 
> 
> 
> 
> 
> --
> Lyra Network, 109 rue de l'innovation, 31670 Labège, FRANCE
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>


-- 
Didi
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] High latency on storage domains and sanlock renewal error

2017-05-13 Thread Stefano Bovina
It's FC/FcoE.

This is with configuration suggested by emc/redhat

360060160a62134002818778f949de411 dm-5 DGC,VRAID
size=11T features='2 queue_if_no_path retain_attached_hw_handler'
hwhandler='1 alua' wp=rw
|-+- policy='round-robin 0' prio=50 status=active
| |- 1:0:1:2 sdr  65:16  active ready running
| `- 2:0:1:2 sdy  65:128 active ready running
`-+- policy='round-robin 0' prio=10 status=enabled
  |- 1:0:0:2 sdd  8:48   active ready running
  `- 2:0:0:2 sdk  8:160  active ready running
360060160a6213400e622de69949de411 dm-2 DGC,VRAID
size=6.0T features='2 queue_if_no_path retain_attached_hw_handler'
hwhandler='1 alua' wp=rw
|-+- policy='round-robin 0' prio=50 status=active
| |- 1:0:1:0 sdp  8:240  active ready running
| `- 2:0:1:0 sdw  65:96  active ready running
`-+- policy='round-robin 0' prio=10 status=enabled
  |- 1:0:0:0 sdb  8:16   active ready running
  `- 2:0:0:0 sdi  8:128  active ready running
360060160a6213400cce46e40949de411 dm-4 DGC,VRAID
size=560G features='2 queue_if_no_path retain_attached_hw_handler'
hwhandler='1 alua' wp=rw
|-+- policy='round-robin 0' prio=50 status=active
| |- 1:0:1:3 sds  65:32  active ready running
| `- 2:0:1:3 sdz  65:144 active ready running
`-+- policy='round-robin 0' prio=10 status=enabled
  |- 1:0:0:3 sde  8:64   active ready running
  `- 2:0:0:3 sdl  8:176  active ready running
360060160a6213400c4b39e80949de411 dm-3 DGC,VRAID
size=500G features='2 queue_if_no_path retain_attached_hw_handler'
hwhandler='1 alua' wp=rw
|-+- policy='round-robin 0' prio=50 status=active
| |- 1:0:1:1 sdq  65:0   active ready running
| `- 2:0:1:1 sdx  65:112 active ready running
`-+- policy='round-robin 0' prio=10 status=enabled
  |- 1:0:0:1 sdc  8:32   active ready running
  `- 2:0:0:1 sdj  8:144  active ready running
360060160a6213400fa2d31acbbfce511 dm-8 DGC,RAID 5
size=5.4T features='2 queue_if_no_path retain_attached_hw_handler'
hwhandler='1 alua' wp=rw
|-+- policy='round-robin 0' prio=50 status=active
| |- 1:0:0:6 sdh  8:112  active ready running
| `- 2:0:0:6 sdo  8:224  active ready running
`-+- policy='round-robin 0' prio=10 status=enabled
  |- 1:0:1:6 sdv  65:80  active ready running
  `- 2:0:1:6 sdac 65:192 active ready running
360060160a621340040652b7582f5e511 dm-7 DGC,RAID 5
size=3.6T features='2 queue_if_no_path retain_attached_hw_handler'
hwhandler='1 alua' wp=rw
|-+- policy='round-robin 0' prio=50 status=active
| |- 1:0:0:4 sdf  8:80   active ready running
| `- 2:0:0:4 sdm  8:192  active ready running
`-+- policy='round-robin 0' prio=10 status=enabled
  |- 1:0:1:4 sdt  65:48  active ready running
  `- 2:0:1:4 sdaa 65:160 active ready running
360060160a621340064b1034cbbfce511 dm-6 DGC,RAID 5
size=1.0T features='2 queue_if_no_path retain_attached_hw_handler'
hwhandler='1 alua' wp=rw
|-+- policy='round-robin 0' prio=50 status=active
| |- 1:0:1:5 sdu  65:64  active ready running
| `- 2:0:1:5 sdab 65:176 active ready running
`-+- policy='round-robin 0' prio=10 status=enabled
  |- 1:0:0:5 sdg  8:96   active ready running
  `- 2:0:0:5 sdn  8:208  active ready running


This is with ovirt default conf:

360060160a6213400848e60af82f5e511 dm-3 DGC ,RAID 5
size=3.6T features='1 retain_attached_hw_handler' hwhandler='1 alua' wp=rw
|-+- policy='service-time 0' prio=50 status=active
| |- 12:0:0:4 sdj 8:144 active ready  running
| `- 13:0:1:4 sdd 8:48  active ready  running
`-+- policy='service-time 0' prio=10 status=enabled
  |- 12:0:1:4 sdf 8:80  active ready  running
  `- 13:0:0:4 sdh 8:112 active ready  running
360060160a621345e425b6b10e611 dm-2 DGC ,RAID 10
size=4.2T features='1 retain_attached_hw_handler' hwhandler='1 alua' wp=rw
|-+- policy='service-time 0' prio=50 status=active
| |- 12:0:1:0 sde 8:64  active ready  running
| `- 13:0:0:0 sdg 8:96  active ready  running
`-+- policy='service-time 0' prio=10 status=enabled
  |- 13:0:1:0 sdc 8:32  active ready  running
  `- 12:0:0:0 sdi 8:128 active ready  running


2017-05-13 18:50 GMT+02:00 Juan Pablo :

> can you please give the output of:
> multipath -ll
> and
> iscsiadm -m session -P3
>
> JP
>
> 2017-05-13 6:48 GMT-03:00 Stefano Bovina :
>
>> Hi,
>>
>> 2.6.32-696.1.1.el6.x86_64
>> 3.10.0-514.10.2.el7.x86_64
>>
>> I tried ioping test from different group of servers using multipath,
>> members of different storage group (different lun, different raid ecc), and
>> everyone report latency.
>> I tried the same test (ioping) on a server with powerpath instead of
>> multipath, with a dedicated raid group and ioping don't report latency.
>>
>>
>> 2017-05-13 2:00 GMT+02:00 Juan Pablo :
>>
>>> sorry to jump in, but what kernel version are you using? had similar
>>> issue with kernel 4.10/4.11
>>>
>>>
>>> 2017-05-12 16:36 GMT-03:00 Stefano Bovina :
>>>
 Hi,
 a little update:

 The command multipath -ll hung when executed on the host while the
 problem occur (nothing logged in /var/log/messages or dmesg).

 I 

Re: [ovirt-users] User story

2017-05-13 Thread Yaniv Kaul
Hi,

First of all, thanks for sharing. It's always good to get feedback,
especially when it's balanced and with specific examples and comparisons.

Secondly, I do believe you have touched on what I believe is a conceptual
difference oVirt has, which translates to a gap in the experience you have
described: when managing 2-3 hosts, it is more intuitive and easier to just
configure each separately (and there's very little to configure anyway, and
the number of hosts is low), then to configure on a higher level (in oVirt
case, data center and cluster level) and apply - who needs either when you
have 2-3 hosts, right?

In a sense, the hyper-converged (gdeploy - see
http://www.ovirt.org/blog/2017/04/up-and-running-with-ovirt-4-1-and-gluster-storage/
) provides a good 'day 1' experience I believe - but is indeed limited to
the hyper-converged deployment type. It'd be a good idea to expand it to
the general case of 2-3 hosts, I reckon.

Perhaps we need to go further and somehow hide both data center and cluster
(for X hosts, where X is lower than... 5?) assuming you'd have only a
single DC and a single cluster - and present their options as 'global'?
Once you go above 5 hosts we'll expand the options and display the bigger
hierarchy?

We've had the idea of 'ovirt-lite' years ago, and it never really
materialized - perhaps we should revisit it. I think it's easy
technologically, a bit more challenging to get right the improved user
experience. I can certainly see the use cases of both small labs, remote
offices and proof-of-concept setups.


As for the installation, I would really like to see:
1. Install an OS -or- install oVirt node
2. Go to http://
3. Installation wizard.

This is exactly (again) what gdeploy provides, as well as hosted-engine -
but we probably need to streamline further more and add regular engine
setup to it.

Thanks again,
Y.



On Sat, May 13, 2017 at 9:04 PM, Johannes Spanier  wrote:

> Hi oVirt community.
>
> I did a short series for tweets @jospanier judging my first time user
> experience with several virtualization platforms and was asked by Sandro
> Bonazzola to elaborate a bit further than what fits into 140 chars.
>
> I had a specific use case: The small-ish learning lab with only 2-3 nodes
> and it needs to be free. I also wanted live migration to stay flexible with
> my hosts.
>
> I currently use my lab for to run ~10 virtual CSR1000V routers on free
> ESXi in addition to some real router hardware. I want to expand the lab to
> be able to explore some other technologies as well like network automation,
> SDN, infrastructure as code and the likes.
>
> The lineup for the PoC was oVirt, ESXi, Openstack and Proxmox VE.
>
> I my tweets I was referring to a) the install procedure and b) the
> operational experience.
>
> Here is what I found. These findings are highly subjective and debatable.
> I am aware of that.
>
> Both ESXi and Proxmox VE is trivial to install. You grab the ISO image,
> use a tool like Rufus to make an bootable USB stick or use iLO virtual CD
> functionality and off you go. Both installers do not ask many questions and
> just do their job. After installation ESXi is all ready to run. Just open
> the WebGui and start deploying your first node. With Proxmox VE you get a
> TUI wizard guiding you though the last steps. After that the WebGui is
> ready and you can deploy your first VM immediately.
>
> I found oVirt a bit more involved to install. You have to install the
> Engine on one node and then register the other hosts with it. While that
> process is easy to handle it is a bit more work. A big thing for me was
> that at first glance there did no seem to be a "single node" install. My
> fist impression was that I needed a minimum of two servers. Of course later
> I learned about the Hosted Engine and the All-In-One install.
>
> Do not get me wrong. First time oVirt installation is still easy to handle
> on a quiet afternoon.
>
> Openstack installation compared to that is a PITA nightmare. I tried both
> RDO (TripleO) and Fuel for setup but gave up after two days for both,
> confused about what I actually need to do for a start. Got some nodes
> running with Fuel but was not satisfied. I then followed the Openstack
> manual Install Guide. I have a day job, so it took me about 5 days to get
> through the whole procedure, but a least I understood what was going on and
> what I needed to do.
>
> So that was my "first day" experience with those.
> Now for the "second day" i.e. operation.
>
> ESXi and Proxmox VE are both very simple to understand. You usually do not
> need a manual to find you way around. Deploying a VM is a breeze. oVirt is
> pretty simple to understand too. But you have to wrap your head around the
> Data Center principle underpinning everything. Its just a bit more
> complicated. On one or two occasions while playing around it was unclear at
> first why my datacenter was offline and I had to consult the manual for
> that. One can 

[ovirt-users] User story

2017-05-13 Thread Johannes Spanier

Hi oVirt community.

I did a short series for tweets @jospanier judging my first time user 
experience with several virtualization platforms and was asked by Sandro 
Bonazzola to elaborate a bit further than what fits into 140 chars.


I had a specific use case: The small-ish learning lab with only 2-3 
nodes and it needs to be free. I also wanted live migration to stay 
flexible with my hosts.


I currently use my lab for to run ~10 virtual CSR1000V routers on free 
ESXi in addition to some real router hardware. I want to expand the lab 
to be able to explore some other technologies as well like network 
automation, SDN, infrastructure as code and the likes.


The lineup for the PoC was oVirt, ESXi, Openstack and Proxmox VE.

I my tweets I was referring to a) the install procedure and b) the 
operational experience.


Here is what I found. These findings are highly subjective and 
debatable. I am aware of that.


Both ESXi and Proxmox VE is trivial to install. You grab the ISO image, 
use a tool like Rufus to make an bootable USB stick or use iLO virtual 
CD functionality and off you go. Both installers do not ask many 
questions and just do their job. After installation ESXi is all ready to 
run. Just open the WebGui and start deploying your first node. With 
Proxmox VE you get a TUI wizard guiding you though the last steps. After 
that the WebGui is ready and you can deploy your first VM immediately.


I found oVirt a bit more involved to install. You have to install the 
Engine on one node and then register the other hosts with it. While that 
process is easy to handle it is a bit more work. A big thing for me was 
that at first glance there did no seem to be a "single node" install. My 
fist impression was that I needed a minimum of two servers. Of course 
later I learned about the Hosted Engine and the All-In-One install.


Do not get me wrong. First time oVirt installation is still easy to 
handle on a quiet afternoon.


Openstack installation compared to that is a PITA nightmare. I tried 
both RDO (TripleO) and Fuel for setup but gave up after two days for 
both, confused about what I actually need to do for a start. Got some 
nodes running with Fuel but was not satisfied. I then followed the 
Openstack manual Install Guide. I have a day job, so it took me about 5 
days to get through the whole procedure, but a least I understood what 
was going on and what I needed to do.


So that was my "first day" experience with those.
Now for the "second day" i.e. operation.

ESXi and Proxmox VE are both very simple to understand. You usually do 
not need a manual to find you way around. Deploying a VM is a breeze. 
oVirt is pretty simple to understand too. But you have to wrap your head 
around the Data Center principle underpinning everything. Its just a bit 
more complicated. On one or two occasions while playing around it was 
unclear at first why my datacenter was offline and I had to consult the 
manual for that. One can immediately feel that multi-tenancy is a big 
benefit of oVirt that is not so obvious in ESXi and Proxmox. But it 
comes with the price of added complexity.


The new WebGui in the ESXi 6.5 I used is slw but some functions are 
not available with the vSphere Client Tool any more.
Proxmox and oVirt WebGui are very responsive and "feel" way better. The 
Openstack WebGui is also quite nice.


If you want to build a private multi-teneant scale-out cloud OpenStack 
is probably your best horse in the stable. But for my purpose it was 
just overkill. The process of deploying a simple VM was the most work 
from all four tools.


Wrap-up
For the specific use-case (see above) Proxmox VE was the best fit for 
me, closely followed by oVirt due to the more involved installation. 
ESXi comes in lagging in third place as it does not offer live migration 
in the free version and due to the poor WebGui experience I got. Also 
you have to pay big bucks to get the same functionality as in the other 
three. Openstack is placed at a distant fourth place for this use case.


Hope that was interesting. Do reach out to me on twitter if you have any 
further questions or suggestions.


Regards
Johannes
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Setting up GeoReplication

2017-05-13 Thread Jim Kusznir
Hi All:

I've been trying to set up georeplication for a while now, but can't seem
to make it work.  I've found documentation on the web (mostly
https://gluster.readthedocs.io/en/refactor/Administrator%20Guide/Geo%20Replication/),
and I found
http://blog.gluster.org/2015/09/introducing-georepsetup-gluster-geo-replication-setup-tool/

Unfortunately, it seems that some critical steps are missing from both, and
I can't figure out for sure what they are.

My environment:

Production: replica 2 + arbitrator running on my 3-node oVirt cluster, 3
volumes (engine, data, iso).

New geo-replication: Raspberry Pi3 with USB hard drive shoved in some other
data closet off-site.

I've installed rasbian-lite, and after much fighting, got
glusterfs-*-3.8.11 installed.  I've created my mountpoint (USB hard drive,
much larger than my gluster volumes), and then ran the command.  I get this
far:

[OK] georep.nwfiber.com is Reachable(Port 22)
[OK] SSH Connection established r...@georep.nwfiber.com
[OK] Master Volume and Slave Volume are compatible (Version: 3.8.11)
[NOT OK] Unable to Mount Gluster Volume georep.nwfiber.com:engine-rep

Trying it with the steps in the gluster docs also has the same problem.  No
long files are generated on the slave.  Log files on the master include:

[root@ovirt1 geo-replication]# more georepsetup.mount.log
[2017-05-13 17:26:27.318599] I [MSGID: 100030] [glusterfsd.c:2454:main]
0-glusterfs: Started running glusterfs version 3.8.11 (args:
 glusterfs --xlator-option="*dht.lookup-unhashed=off" --volfile-server
localhost --volfile-id engine -l /var/log/glusterfs/geo-repli
cation/georepsetup.mount.log --client-pid=-1 /tmp/georepsetup_wZtfkN)
[2017-05-13 17:26:27.341170] I [MSGID: 101190]
[event-epoll.c:628:event_dispatch_epoll_worker] 0-epoll: Started thread
with index 1
[2017-05-13 17:26:27.341260] E [socket.c:2309:socket_connect_finish]
0-glusterfs: connection to ::1:24007 failed (Connection refused
)
[2017-05-13 17:26:27.341846] E [glusterfsd-mgmt.c:1908:mgmt_rpc_notify]
0-glusterfsd-mgmt: failed to connect with remote-host: local
host (Transport endpoint is not connected)
[2017-05-13 17:26:31.335849] I [MSGID: 101190]
[event-epoll.c:628:event_dispatch_epoll_worker] 0-epoll: Started thread
with index 2
[2017-05-13 17:26:31.337545] I [MSGID: 114020] [client.c:2356:notify]
0-engine-client-0: parent translators are ready, attempting co
nnect on transport
[2017-05-13 17:26:31.344485] I [MSGID: 114020] [client.c:2356:notify]
0-engine-client-1: parent translators are ready, attempting co
nnect on transport
[2017-05-13 17:26:31.345146] I [rpc-clnt.c:1965:rpc_clnt_reconfig]
0-engine-client-0: changing port to 49157 (from 0)
[2017-05-13 17:26:31.350868] I [MSGID: 114020] [client.c:2356:notify]
0-engine-client-2: parent translators are ready, attempting co
nnect on transport
[2017-05-13 17:26:31.355946] I [MSGID: 114057]
[client-handshake.c:1440:select_server_supported_programs]
0-engine-client-0: Using P
rogram GlusterFS 3.3, Num (1298437), Version (330)
[2017-05-13 17:26:31.356280] I [rpc-clnt.c:1965:rpc_clnt_reconfig]
0-engine-client-1: changing port to 49157 (from 0)
Final graph:
+--+
  1: volume engine-client-0
  2: type protocol/client
  3: option clnt-lk-version 1
  4: option volfile-checksum 0
  5: option volfile-key engine
  6: option client-version 3.8.11
  7: option process-uuid
ovirt1.nwfiber.com-25660-2017/05/13-17:26:27:311929-engine-client-0-0-0
  8: option fops-version 1298437
  9: option ping-timeout 30
 10: option remote-host ovirt1.nwfiber.com
 11: option remote-subvolume /gluster/brick1/engine
 12: option transport-type socket
 13: option username 028984cf-0399-42e6-b04b-bb9b1685c536
 14: option password eae737cc-9659-405f-865e-9a7ef97a3307
 15: option filter-O_DIRECT off
 16: option send-gids true
 17: end-volume
 18:
 19: volume engine-client-1
 20: type protocol/client
 21: option ping-timeout 30
 22: option remote-host ovirt2.nwfiber.com
 23: option remote-subvolume /gluster/brick1/engine
 24: option transport-type socket
 25: option username 028984cf-0399-42e6-b04b-bb9b1685c536
 26: option password eae737cc-9659-405f-865e-9a7ef97a3307
 27: option filter-O_DIRECT off
 28: option send-gids true
 29: end-volume
 30:
 31: volume engine-client-2
 32: type protocol/client
 33: option ping-timeout 30
 34: option remote-host ovirt3.nwfiber.com
 35: option remote-subvolume /gluster/brick1/engine
 36: option transport-type socket
 37: option username 028984cf-0399-42e6-b04b-bb9b1685c536
 38: option password eae737cc-9659-405f-865e-9a7ef97a3307
 39: option filter-O_DIRECT off
 40: option send-gids true
 41: end-volume
 42:
 43: volume engine-replicate-0
 44: type cluster/replicate
 45: option arbiter-count 1
 46: option data-self-heal-algorithm full
 47: option 

[ovirt-users] can i Extend volume from replica 2 to replica 3 with arbiter

2017-05-13 Thread Khalid Jamal

Dear Teami need you advice for convert our volume replica 2 to replica 3 with arbiter but important thing it's production environments , i try to convert the same name of volume in replica 2 to convert is to the replica 3 to avoid losing vm's or any data that's what i do it :# gluster volume create gfs1 replica 3 arbiter 1 s1:/export/sdb1/br1 s2:/export/sdb1/br1 s3:/export/sdb1/br1 s1:/export/sdc1/br1 s2:/export/sdc1/br1 s3:/export/sdc1/br1 s4:/export/sdb1/br1 s5:/export/sdb1/br1 s6:/export/sdb1/br1 s4:/export/sdc1/br1 s5:/export/sdc1/br1 s6:/export/sdc1/br1that's the result :volume create: gfs1: failed: Volume gfs1 already existsi try to change the name but without any success what shall i do.best regardsEng khalid jamalSystem Admin@IT DepartmentEarthlink TelecomEmail: khalid.ja...@earthlinktele.comNo: 3355skype: engkhalid21986NO : 07704268321
 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] High latency on storage domains and sanlock renewal error

2017-05-13 Thread Juan Pablo
can you please give the output of:
multipath -ll
and
iscsiadm -m session -P3

JP

2017-05-13 6:48 GMT-03:00 Stefano Bovina :

> Hi,
>
> 2.6.32-696.1.1.el6.x86_64
> 3.10.0-514.10.2.el7.x86_64
>
> I tried ioping test from different group of servers using multipath,
> members of different storage group (different lun, different raid ecc), and
> everyone report latency.
> I tried the same test (ioping) on a server with powerpath instead of
> multipath, with a dedicated raid group and ioping don't report latency.
>
>
> 2017-05-13 2:00 GMT+02:00 Juan Pablo :
>
>> sorry to jump in, but what kernel version are you using? had similar
>> issue with kernel 4.10/4.11
>>
>>
>> 2017-05-12 16:36 GMT-03:00 Stefano Bovina :
>>
>>> Hi,
>>> a little update:
>>>
>>> The command multipath -ll hung when executed on the host while the
>>> problem occur (nothing logged in /var/log/messages or dmesg).
>>>
>>> I tested latency with ioping:
>>> ioping /dev/6a386652-629d-4045-835b-21d2f5c104aa/metadata
>>>
>>> Usually it return "time=15.6 ms", sometimes return "time=19 s" (yes,
>>> seconds)
>>>
>>> Systems are up to date and I tried both path_checker (emc_clariion and
>>> directio), without results.
>>> (https://access.redhat.com/solutions/139193, it refers to the Rev A31
>>> of EMC document; last is A42 and suggest emc_clariion).
>>>
>>> Any idea or suggestion?
>>>
>>> Thanks,
>>>
>>> Stefano
>>>
>>> 2017-05-08 11:56 GMT+02:00 Yaniv Kaul :
>>>


 On Mon, May 8, 2017 at 11:50 AM, Stefano Bovina 
 wrote:

> Yes,
> this configuration is the one suggested by EMC for EL7.
>

 https://access.redhat.com/solutions/139193 suggest that for alua, the
 patch checker needs to be different.

 Anyway, it is very likely that you have storage issues - they need to
 be resolved first and I believe they have little to do with oVirt at the
 moment.
 Y.


>
> By the way,
> "The parameters rr_min_io vs. rr_min_io_rq mean the same thing but are
> used for device-mapper-multipath on differing kernel versions." and
> rr_min_io_rq default value is 1, rr_min_io default value is 1000, so it
> should be fine.
>
>
> 2017-05-08 9:39 GMT+02:00 Yaniv Kaul :
>
>>
>> On Sun, May 7, 2017 at 1:27 PM, Stefano Bovina 
>> wrote:
>>
>>> Sense data are 0x0/0x0/0x0
>>
>>
>> Interesting - first time I'm seeing 0/0/0. The 1st is usually 0x2
>> (see [1]), and then the rest [2], [3] make sense.
>>
>> A google search found another user with Clarion with the exact same
>> error[4], so I'm leaning toward misconfiguration of multipathing/clarion
>> here.
>>
>> Is your multipathing configuration working well for you?
>> Are you sure it's a EL7 configuration? For example, I believe you
>> should have rr_min_io_rq and not rr_min_io .
>> Y.
>>
>> [1] http://www.t10.org/lists/2status.htm
>> [2] http://www.t10.org/lists/2sensekey.htm
>> [3] http://www.t10.org/lists/asc-num.htm
>> [4] http://www.linuxquestions.org/questions/centos-111/multi
>> path-problems-4175544908/
>>
>
>

>>>
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>>
>>>
>>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] High latency on storage domains and sanlock renewal error

2017-05-13 Thread Stefano Bovina
Hi,

2.6.32-696.1.1.el6.x86_64
3.10.0-514.10.2.el7.x86_64

I tried ioping test from different group of servers using multipath,
members of different storage group (different lun, different raid ecc), and
everyone report latency.
I tried the same test (ioping) on a server with powerpath instead of
multipath, with a dedicated raid group and ioping don't report latency.


2017-05-13 2:00 GMT+02:00 Juan Pablo :

> sorry to jump in, but what kernel version are you using? had similar issue
> with kernel 4.10/4.11
>
>
> 2017-05-12 16:36 GMT-03:00 Stefano Bovina :
>
>> Hi,
>> a little update:
>>
>> The command multipath -ll hung when executed on the host while the
>> problem occur (nothing logged in /var/log/messages or dmesg).
>>
>> I tested latency with ioping:
>> ioping /dev/6a386652-629d-4045-835b-21d2f5c104aa/metadata
>>
>> Usually it return "time=15.6 ms", sometimes return "time=19 s" (yes,
>> seconds)
>>
>> Systems are up to date and I tried both path_checker (emc_clariion and
>> directio), without results.
>> (https://access.redhat.com/solutions/139193, it refers to the Rev A31 of
>> EMC document; last is A42 and suggest emc_clariion).
>>
>> Any idea or suggestion?
>>
>> Thanks,
>>
>> Stefano
>>
>> 2017-05-08 11:56 GMT+02:00 Yaniv Kaul :
>>
>>>
>>>
>>> On Mon, May 8, 2017 at 11:50 AM, Stefano Bovina 
>>> wrote:
>>>
 Yes,
 this configuration is the one suggested by EMC for EL7.

>>>
>>> https://access.redhat.com/solutions/139193 suggest that for alua, the
>>> patch checker needs to be different.
>>>
>>> Anyway, it is very likely that you have storage issues - they need to be
>>> resolved first and I believe they have little to do with oVirt at the
>>> moment.
>>> Y.
>>>
>>>

 By the way,
 "The parameters rr_min_io vs. rr_min_io_rq mean the same thing but are
 used for device-mapper-multipath on differing kernel versions." and
 rr_min_io_rq default value is 1, rr_min_io default value is 1000, so it
 should be fine.


 2017-05-08 9:39 GMT+02:00 Yaniv Kaul :

>
> On Sun, May 7, 2017 at 1:27 PM, Stefano Bovina 
> wrote:
>
>> Sense data are 0x0/0x0/0x0
>
>
> Interesting - first time I'm seeing 0/0/0. The 1st is usually 0x2 (see
> [1]), and then the rest [2], [3] make sense.
>
> A google search found another user with Clarion with the exact same
> error[4], so I'm leaning toward misconfiguration of multipathing/clarion
> here.
>
> Is your multipathing configuration working well for you?
> Are you sure it's a EL7 configuration? For example, I believe you
> should have rr_min_io_rq and not rr_min_io .
> Y.
>
> [1] http://www.t10.org/lists/2status.htm
> [2] http://www.t10.org/lists/2sensekey.htm
> [3] http://www.t10.org/lists/asc-num.htm
> [4] http://www.linuxquestions.org/questions/centos-111/multi
> path-problems-4175544908/
>


>>>
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users