Hi Vincent,

nice to hear the news :-)
 
I have read the BZ and see you run into NFS trouble and solved it now.

I took a look on my centos server for nfs data domains and
see server running V4 and the clients (node mounts with protocol vers=4.1)

I run the latest (and greatest) ovirt stable 4.2.6.4-1 on centos 7.5+ with
engine installed and a ovirt-node 4.2.6.4.

* If you can migrate your running VMs and can switch your SPM, 
  i would upgrade and reboot the hosts one by one, now.

* reboot seem to be a minimum, remember you do that `virt. thing´,
  therefor you can access and boot your bare metal and host os ;-)


ok, back to iSCSI, i have also a EQUALOGIC running as iSCSI target over years.

* I have allowed multi host access to the volumes which ovirt uses.
  Access control lists contains raw IP addresses from my ovirt-hosts.

  ovirt handles the volume access virtuos with multipathd and lvm vg’s and lv’s.
  unused lvs are offline (host specific) and released volumes are deactivated.
  
* Also it’s possible you have to reinstall (from GUI) your hosts,
  to upgrade or install the needed packages, which handles iSCSI Client access.

* If you then free from errors and your iscsi data domain still missing, we talk
  about vg activation and domain import.

Sheers

Oliver


> Am 04.10.2018 um 22:00 schrieb Vincent Royer <[email protected]>:
> 
> Ok, getting somewhere here. 
> 
> did a rpcinfo -p and found no nfs entries in portmap. 
> 
> systemctl stop nfs
> systemctl start nfs
> 
> Suddenly shares are mounted and datacenter is up again. 
> 
> was able to add export domain over NFS.  
> 
> Why would nfs shit the bed?  
> 
> still can't seem to get iscsi mounted properly now, and that's where all the 
> disks are located :/
> 
> 
> On Thu, Oct 4, 2018 at 11:00 AM Vincent Royer <[email protected] 
> <mailto:[email protected]>> wrote:
> Thanks for your help Oliver, 
> 
> To give you some background here:
> 
> Host 1 on Ovirt 4.2 attached to NFS storage
> Host 2 I upgraded to Ovirt 4.2.5 and then 4.2.6, since then it has had 
> troubles with NFS due to this bug 
> https://bugzilla.redhat.com/show_bug.cgi?id=1595549 
> <https://bugzilla.redhat.com/show_bug.cgi?id=1595549>.   The host was up and 
> could run the hosted engine, but could not migrate any VMs to it. 
> 
> I decided to switch from NFS to ISCSI so that I could stay on current 
> releases.   So I began the work of attaching iscsi domain. 
> 
> The iscsi domain attached, and I transferred most of the disks to it.  Then 
> it started melting down saying that Host 1 could not mount it, and the whole 
> DC went down. 
> 
> Current status is data center "non responsive".  Keeps trying "Reconstructing 
> master domain on Data Center"  over and over again but always fails.  Master 
> domain status is "inactive".  Clicking activate fails.  The new ISCSI domain, 
> I put in maintenance until I figure the rest out.  I can't add or remove any 
> other domains, Ovirt says I need to attach the master first. 
> 
> Both hosts are "UP".   Host 1 health is "bad"   Host 2 health is "ok", and it 
> is running HE.  Host 1 (the 4.2 host) says "this host needs to be 
> reinstalled".  But the reinstall option is grayed out. 
> 
> I am weary about updating host1, because of the NFS storage bug... I fear it 
> won't ever be able to attach the old domain again. 
> 
> If I try mounting the NFS shares in cockpit from either node, they say 
> "mount.nfs: Remote I/O error".   However on another blank centos machine 
> sitting on the same network, I can mount the shares normally. 
> 
> Vincent Royer
> 778-825-1057
> 
> 
>  <http://www.epicenergy.ca/>
> SUSTAINABLE MOBILE ENERGY SOLUTIONS
> 
> 
> 
> 
> 
> On Thu, Oct 4, 2018 at 1:04 AM Oliver Riesener <[email protected] 
> <mailto:[email protected]>> wrote:
> When your hosts are up and running and your Domain didn't go active within 
> minutes
> * Activate your Storage Domain under:
> 
> Storage -> Storage Domain -> (Open your Domain)  -> Data Center -> (Right 
> Click Your Data Center Name) -> Activate.
> On 10/4/18 9:50 AM, Oliver Riesener wrote:
>> Hi Vincent,
>> 
>> OK you master domain, isn't avail a the moment, but no panic.
>> First off all we need the status from your hosts. No HOSTS -> No Storage !
>> * Do you reboot them hard, without Confirm "Host has been rebooted"
>> 
>> * Are they actived in the DataCenter / Cluster ? Green Arrow ?
>> 
>> 
>> On 10/4/18 7:46 AM, Vincent Royer wrote:
>>> I was attempting to migrate from nfs to iscsi storage domains.  I have 
>>> reached a state where I can no longer activate the old master storage 
>>> domain, and thus no others will activate either. 
>>> 
>>> I'm ready to give up on the installation and just move to an HCI deployment 
>>> instead.  Wipe all the hosts clean and start again. 
>>> 
>>> My plan was to create and use an export domain, then wipe the nodes and set 
>>> them up HCI where I could re-import.  But without being able to activate a 
>>> master domain, I can't create the export domain.
>>> 
>>> I'm not sure why it can't find the master anymore, as nothing has happened 
>>> to the NFS storage, but the error in vdsm says it just can't find it:
>>> 
>>> StoragePoolMasterNotFound: Cannot find master domain: 
>>> u'spUUID=5a77bed1-0238-030c-0122-0000000003b3, 
>>> msdUUID=d3165759-07c2-46ae-b7b8-b6226a929d68'
>>> 2018-10-03 22:40:33,751-0700 INFO  (jsonrpc/3) [storage.TaskManager.Task] 
>>> (Task='83f33db5-90f3-4064-87df-0512ab9b6378') aborting: Task is aborted: 
>>> "Cannot find master domain: u'spUUID=5a77bed1-0238-030c-0122-0000000003b3, 
>>> msdUUID=d3165759-07c2-46ae-b7b8-b6226a929d68'" - code 304 (task:1181)
>>> 2018-10-03 22:40:33,751-0700 ERROR (jsonrpc/3) [storage.Dispatcher] FINISH 
>>> connectStoragePool error=Cannot find master domain: 
>>> u'spUUID=5a77bed1-0238-030c-0122-0000000003b3, 
>>> msdUUID=d3165759-07c2-46ae-b7b8-b6226a929d68' (dispatcher:82)
>>> 2018-10-03 22:40:33,751-0700 INFO  (jsonrpc/3) [jsonrpc.JsonRpcServer] RPC 
>>> call StoragePool.connect failed (error 304) in 0.17 seconds (__init__:573)
>>> 2018-10-03 22:40:34,200-0700 INFO  (jsonrpc/1) [api.host] START getStats() 
>>> from=::ffff:172.16.100.13,39028 (api:46)
>>> 
>>> When I look in cockpit on the hosts, the storage domain is mounted and 
>>> seems fine. 
>>> 
>>> 
>>> 
>>> 
>>> _______________________________________________
>>> Users mailing list -- [email protected] <mailto:[email protected]>
>>> To unsubscribe send an email to [email protected] 
>>> <mailto:[email protected]>
>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ 
>>> <https://www.ovirt.org/site/privacy-policy/>
>>> oVirt Code of Conduct: 
>>> https://www.ovirt.org/community/about/community-guidelines/ 
>>> <https://www.ovirt.org/community/about/community-guidelines/>
>>> List Archives: 
>>> https://lists.ovirt.org/archives/list/[email protected]/message/LTZ6SIFYDFEMSZ4ACUNVC5KETWG7BBIZ/
>>>  
>>> <https://lists.ovirt.org/archives/list/[email protected]/message/LTZ6SIFYDFEMSZ4ACUNVC5KETWG7BBIZ/>
>> -- 
>> Mit freundlichem Gruß
>> 
>> 
>> Oliver Riesener
>> 
>> --
>> Hochschule Bremen
>> Elektrotechnik und Informatik
>> Oliver Riesener
>> Neustadtswall 30
>> D-28199 Bremen
>> 
>> Tel: 0421 5905-2405, Fax: -2400
>> e-mail:[email protected] 
>> <mailto:e-mail:[email protected]>
> Tel: 0421 5905-2405, Fax: -2400
> e-mail:[email protected] 
> <mailto:e-mail:[email protected]>
> _______________________________________________
> Users mailing list -- [email protected] <mailto:[email protected]>
> To unsubscribe send an email to [email protected] 
> <mailto:[email protected]>
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ 
> <https://www.ovirt.org/site/privacy-policy/>
> oVirt Code of Conduct: 
> https://www.ovirt.org/community/about/community-guidelines/ 
> <https://www.ovirt.org/community/about/community-guidelines/>
> List Archives: 
> https://lists.ovirt.org/archives/list/[email protected]/message/V72KMULZJAT3XIR3GBTOCA5RLACVQSRC/
>  
> <https://lists.ovirt.org/archives/list/[email protected]/message/V72KMULZJAT3XIR3GBTOCA5RLACVQSRC/>
> _______________________________________________
> Users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct: 
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives: 
> https://lists.ovirt.org/archives/list/[email protected]/message/OZ3TOO554D4YAZJ6FPGI4SAJ6CKWZRFH/

_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/CIOFZOSWGDMK6S6ORD3SVCQZQEHZ6L7E/

Reply via email to