Okay, both the BZ and ML posts are interesting and helpful. I'm kind of
surprised there seems to be so much trouble and confusion for what I
would have thought to be a very common setup. Are most people using
something else?
I think this gives me what I need for my next stab at doing this but I"m
still puzzled on how to tear down what I have in oVirt so that I can
redo it. Specifically, I didn't see how to delete the existing iSCSI
connections. I've read that this can only be done through the REST
API. I have managed to redo the interfaces on my Hosts so that
everything is now on just 2 NICs each, leaving 2 NICs free for a
foothold on a new setup. From all of my experimentation, it would
appear that my only option is to create a new storage domain and
export/import each disk volume one by one. Maybe there's a migration
option I have yet to see, but I don't see any way around creating a new
storage domain here.
On 1/21/19 7:12 AM, Vinícius Ferrão wrote:
Hello people, in the past Maor Lipchuk (from RH) tried very hard to
help me and Uwe but we was unable to converge on the solution.
This was discussed a year ago and on my understanding it still and
oVirt bug. As today, if you simple “DuckDuckGo” for “ovirt iscsi
multipath not working” the third link points to this bugzilla:
https://bugzilla.redhat.com/show_bug.cgi?id=1474904
Which is the one I’ve mentioned and it’s extremely similar to John
Florian case, which was my case too.
@John, take a look at the bugzilla link and see if the desired
topology match with your case.
Regards,
On 21 Jan 2019, at 05:21, Eitan Raviv <[email protected]
<mailto:[email protected]>> wrote:
Shani,
Can you help here with iSCSI bonding?
Thanks
On Mon, Jan 21, 2019 at 7:51 AM Uwe Laverenz <[email protected]
<mailto:[email protected]>> wrote:
Hi John,
Am 20.01.19 um 18:32 schrieb John Florian:
As for how to get there, whatever exactly that might look like, I'm
also
having troubles figuring that out. I figured I would transform the
setup described below into one where each host has:
* 2 NICs bonded with LACP for my ovirtmgmt and "main" net
* 1 NIC for my 1st storage net
* 1 NIC for my 2nd storage net
This is exactly the setup I use. I have run this successfully with
CentOS/LIO and FreeNAS iSCSI targets with good performance.
In short:
- 2 separate, isolated networks for iSCSI with dedicated adapters
on hosts and iSCSI target
- jumbo frames enabled
- no VLANs config needed on hosts, untagged VLANs on switch
- do _not_ use LACP, let multipathd handle failovers
Same experience as Vinicius: what did _not_ work for me is the
iSCSI-Bonding in OVirt. It seems to require that all storage IPs are
reachable from all other IPs, which is not the case in every setup.
To get multipathing to work I use multipath directly:
https://www.mail-archive.com/[email protected]/msg42735.html
I will post a bonnie++ result later. If you need more details please let
me know.
cu,
Uwe
_______________________________________________
Users mailing list -- [email protected] <mailto:[email protected]>
To unsubscribe send an email to [email protected]
<mailto:[email protected]>
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/[email protected]/message/E2QKV7CZR27NT6MRSNL352KLOQ5OAGDR/
_______________________________________________
Users mailing list [email protected]
To unsubscribe send an email [email protected]
Privacy Statement:https://www.ovirt.org/site/privacy-policy/
oVirt Code of
Conduct:https://www.ovirt.org/community/about/community-guidelines/
List
Archives:https://lists.ovirt.org/archives/list/[email protected]/message/NSE5BCLJSIFDX2VDZRBRLODEH3ZCPYWN/
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/[email protected]/message/G6GMNF3RU5IDEDM4OWG4RDXAFY5BHDSV/