thanks, I'm trying to move away from NFS due to the poor vw performance and problems I had with it. Hence wanting to move the master domain and HE domains over to iscsi domains. I have now an export domain, an ISO domain, and HostedEngine domain. From my reading, to switch master, you need to have only 2 data domains, then put the old one into maint and it will switch to the new one. Only then can you remove the old one... is that not correct?
I gained access to that article by creating a dev account, didn't realize it was free ;) On Fri, Oct 5, 2018 at 1:22 AM Oliver Riesener <oliver.riese...@hs-bremen.de> wrote: > > On 10/5/18 8:10 AM, Vincent Royer wrote: > > Alright, some combination of messing around got me mostly back up and > running here. Iscsi is working reliably, VMs are running on the new iscsi > domain. the old NFS domain is empty save a couple of OVF_STORE disks. > > good job > > The procedure now, as I understand it, is to shut down all vms, > > why ? > > put all storage domains into maintenance > > which additional domain you have ? > > except the old NFS and new SCSI domain, then push the NFS one into > maintenance, which *should* promote the ISCSI domain to master. > > it's going to happen. but it's not importend for ovirt where the master > lives. > > After that, I'll try to move the HE storage to iscsi too, although it > would be very helpful to have access to this document > <https://access.redhat.com/solutions/2998291> for that procedure. > > Sorry i have also no access, but it sounds like backup/restore ... > > RHEV Documentation is <a href= > "https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.2/" > <https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.2/> > >here</a>. > > I think it's like E(BareMetal) <a href= > "https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.2/html/self-hosted_engine_guide/chap-migrating_from_bare_metal_to_a_rhel-based_self-hosted_environment" > <https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.2/html/self-hosted_engine_guide/chap-migrating_from_bare_metal_to_a_rhel-based_self-hosted_environment>>(see > here)</a>to HE(ISCSI), but you want HE(NFS) to HE(iSCSI). > > It's time for a new mailing thread. > > cu > > Oliver > > > thanks for your assistance! > > On Thu, Oct 4, 2018 at 2:23 PM Oliver Riesener < > oliver.riese...@hs-bremen.de> wrote: > >> Hi Vincent, >> >> nice to hear the news :-) >> >> I have read the BZ and see you run into NFS trouble and solved it now. >> >> I took a look on my centos server for nfs data domains and >> see server running V4 and the clients (node mounts with protocol vers=4.1) >> >> I run the latest (and greatest) ovirt stable 4.2.6.4-1 on centos 7.5+ with >> engine installed and a ovirt-node 4.2.6.4. >> >> * If you can migrate your running VMs and can switch your SPM, >> i would upgrade and reboot the hosts one by one, now. >> >> * reboot seem to be a minimum, remember you do that `virt. thing´, >> therefor you can access and boot your bare metal and host os ;-) >> >> >> ok, back to iSCSI, i have also a EQUALOGIC running as iSCSI target over >> years. >> >> * I have allowed multi host access to the volumes which ovirt uses. >> Access control lists contains raw IP addresses from my ovirt-hosts. >> >> ovirt handles the volume access virtuos with multipathd and lvm vg’s >> and lv’s. >> unused lvs are offline (host specific) and released volumes are >> deactivated. >> >> * Also it’s possible you have to reinstall (from GUI) your hosts, >> to upgrade or install the needed packages, which handles iSCSI Client >> access. >> >> * If you then free from errors and your iscsi data domain still missing, >> we talk >> about vg activation and domain import. >> >> Sheers >> >> Oliver >> >> >> Am 04.10.2018 um 22:00 schrieb Vincent Royer <vinc...@epicenergy.ca>: >> >> Ok, getting somewhere here. >> >> did a rpcinfo -p and found no nfs entries in portmap. >> >> systemctl stop nfs >> systemctl start nfs >> >> Suddenly shares are mounted and datacenter is up again. >> >> was able to add export domain over NFS. >> >> Why would nfs shit the bed? >> >> still can't seem to get iscsi mounted properly now, and that's where all >> the disks are located :/ >> >> >> On Thu, Oct 4, 2018 at 11:00 AM Vincent Royer <vinc...@epicenergy.ca> >> wrote: >> >>> Thanks for your help Oliver, >>> >>> To give you some background here: >>> >>> Host 1 on Ovirt 4.2 attached to NFS storage >>> Host 2 I upgraded to Ovirt 4.2.5 and then 4.2.6, since then it has had >>> troubles with NFS due to this bug >>> https://bugzilla.redhat.com/show_bug.cgi?id=1595549. The host was up >>> and could run the hosted engine, but could not migrate any VMs to it. >>> >>> I decided to switch from NFS to ISCSI so that I could stay on current >>> releases. So I began the work of attaching iscsi domain. >>> >>> The iscsi domain attached, and I transferred most of the disks to it. >>> Then it started melting down saying that Host 1 could not mount it, and the >>> whole DC went down. >>> >>> Current status is data center "non responsive". Keeps trying >>> "Reconstructing master domain on Data Center" over and over again but >>> always fails. Master domain status is "inactive". Clicking activate >>> fails. The new ISCSI domain, I put in maintenance until I figure the rest >>> out. I can't add or remove any other domains, Ovirt says I need to attach >>> the master first. >>> >>> Both hosts are "UP". Host 1 health is "bad" Host 2 health is "ok", >>> and it is running HE. Host 1 (the 4.2 host) says "this host needs to be >>> reinstalled". But the reinstall option is grayed out. >>> >>> I am weary about updating host1, because of the NFS storage bug... I >>> fear it won't ever be able to attach the old domain again. >>> >>> If I try mounting the NFS shares in cockpit from either node, they say >>> "mount.nfs: Remote I/O error". However on another blank centos machine >>> sitting on the same network, I can mount the shares normally. >>> >>> *Vincent Royer* >>> *778-825-1057* >>> >>> >>> <http://www.epicenergy.ca/> >>> *SUSTAINABLE MOBILE ENERGY SOLUTIONS* >>> >>> >>> >>> >>> >>> On Thu, Oct 4, 2018 at 1:04 AM Oliver Riesener < >>> oliver.riese...@hs-bremen.de> wrote: >>> >>>> When your hosts are up and running and your Domain didn't go active >>>> within minutes >>>> >>>> * Activate your Storage Domain under: >>>> >>>> Storage -> Storage Domain -> (Open your Domain) -> Data Center -> >>>> (Right Click Your Data Center Name) -> Activate. >>>> On 10/4/18 9:50 AM, Oliver Riesener wrote: >>>> >>>> Hi Vincent, >>>> >>>> OK you master domain, isn't avail a the moment, but no panic. >>>> >>>> First off all we need the status from your hosts. No HOSTS -> No >>>> Storage ! >>>> >>>> * Do you reboot them hard, without Confirm "Host has been rebooted" >>>> >>>> * Are they actived in the DataCenter / Cluster ? Green Arrow ? >>>> >>>> >>>> On 10/4/18 7:46 AM, Vincent Royer wrote: >>>> >>>> I was attempting to migrate from nfs to iscsi storage domains. I have >>>> reached a state where I can no longer activate the old master storage >>>> domain, and thus no others will activate either. >>>> >>>> I'm ready to give up on the installation and just move to an HCI >>>> deployment instead. Wipe all the hosts clean and start again. >>>> >>>> My plan was to create and use an export domain, then wipe the nodes and >>>> set them up HCI where I could re-import. But without being able to >>>> activate a master domain, I can't create the export domain. >>>> >>>> I'm not sure why it can't find the master anymore, as nothing has >>>> happened to the NFS storage, but the error in vdsm says it just can't find >>>> it: >>>> >>>> StoragePoolMasterNotFound: Cannot find master domain: >>>> u'spUUID=5a77bed1-0238-030c-0122-0000000003b3, >>>> msdUUID=d3165759-07c2-46ae-b7b8-b6226a929d68' >>>> 2018-10-03 22:40:33,751-0700 INFO (jsonrpc/3) >>>> [storage.TaskManager.Task] (Task='83f33db5-90f3-4064-87df-0512ab9b6378') >>>> aborting: Task is aborted: "Cannot find master domain: >>>> u'spUUID=5a77bed1-0238-030c-0122-0000000003b3, >>>> msdUUID=d3165759-07c2-46ae-b7b8-b6226a929d68'" - code 304 (task:1181) >>>> 2018-10-03 22:40:33,751-0700 ERROR (jsonrpc/3) [storage.Dispatcher] >>>> FINISH connectStoragePool error=Cannot find master domain: >>>> u'spUUID=5a77bed1-0238-030c-0122-0000000003b3, >>>> msdUUID=d3165759-07c2-46ae-b7b8-b6226a929d68' (dispatcher:82) >>>> 2018-10-03 22:40:33,751-0700 INFO (jsonrpc/3) [jsonrpc.JsonRpcServer] >>>> RPC call StoragePool.connect failed (error 304) in 0.17 seconds >>>> (__init__:573) >>>> 2018-10-03 22:40:34,200-0700 INFO (jsonrpc/1) [api.host] START >>>> getStats() from=::ffff:172.16.100.13,39028 (api:46) >>>> >>>> When I look in cockpit on the hosts, the storage domain is mounted and >>>> seems fine. >>>> >>>> >>>> >>>> _______________________________________________ >>>> Users mailing list -- users@ovirt.org >>>> To unsubscribe send an email to users-le...@ovirt.org >>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>> oVirt Code of Conduct: >>>> https://www.ovirt.org/community/about/community-guidelines/ >>>> List Archives: >>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/LTZ6SIFYDFEMSZ4ACUNVC5KETWG7BBIZ/ >>>> >>>> -- >>>> Mit freundlichem Gruß >>>> >>>> >>>> Oliver Riesener >>>> >>>> -- >>>> Hochschule Bremen >>>> Elektrotechnik und Informatik >>>> Oliver Riesener >>>> Neustadtswall 30 >>>> D-28199 Bremen >>>> >>>> Tel: 0421 5905-2405, Fax: -2400e-mail:oliver.riese...@hs-bremen.de >>>> >>>> Tel: 0421 5905-2405, Fax: -2400e-mail:oliver.riese...@hs-bremen.de >>>> >>>> _______________________________________________ >>>> Users mailing list -- users@ovirt.org >>>> To unsubscribe send an email to users-le...@ovirt.org >>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>> oVirt Code of Conduct: >>>> https://www.ovirt.org/community/about/community-guidelines/ >>>> List Archives: >>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/V72KMULZJAT3XIR3GBTOCA5RLACVQSRC/ >>>> >>> _______________________________________________ >> Users mailing list -- users@ovirt.org >> To unsubscribe send an email to users-le...@ovirt.org >> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >> oVirt Code of Conduct: >> https://www.ovirt.org/community/about/community-guidelines/ >> List Archives: >> https://lists.ovirt.org/archives/list/users@ovirt.org/message/OZ3TOO554D4YAZJ6FPGI4SAJ6CKWZRFH/ >> >> >> > _______________________________________________ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/XBVJAWRSA2NCNAWEOFKCONJBP7NTPSF2/ > > -- > Mit freundlichem Gruß > > > Oliver Riesener > > -- > Hochschule Bremen > Elektrotechnik und Informatik > Oliver Riesener > Neustadtswall 30 > D-28199 Bremen > > Tel: 0421 5905-2405, Fax: -2400e-mail:oliver.riese...@hs-bremen.de > >
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/ROT3YZ446QC6BKYD4ZW3VDNOYO3KMEK3/