Right now the HE domain is old (NFS). I have created a new one (iSCSI), nothing on it yet.
The master domain is old (NFS) but all my disks are migrated to the new one (iSCSI). I don't really care which of the iSCSI domains becomes the master. Just want off of NFS. On Fri, Oct 5, 2018 at 2:31 AM Oliver Riesener <oliver.riese...@hs-bremen.de> wrote: > (Master) is only the flag which OVF-STORE files are active used to store > runtime and configuration storage options. > > What's the storage type of your Hosted Engine Domain ? > > To remove a domain, you have to put it first into maintenance, then the > (Master) status automatically switch to another domain. > If you really want to determine with destination for master domain you > got, two domains are correct, > but it's not importend where the master lives, expect the "to be removed" > domain. > > Storage -> Storage Domains -> Open (your domain) -> DataCenter -> > Centername -> Right Click -> Maintenance > > Now you have two Options: Detach or Destroy > > Detach: OVF-FILES and Disks are not really physical removed and can't > imported later to another DC. > > Storage -> Storage Domains -> Open (your domain) -> DataCenter -> > Centername -> Right Click -> Detach > > Detroy: Storage would be blanked, OVF-Store Disks, thats what you want > here. > > !!!! Warning migrate your all your disks first! !!!! > > Storage -> Storage Domains -> Open (your domain) -> DotDotDot menu -> > Detroy > On 10/5/18 10:30 AM, Vincent Royer wrote: > > thanks, > > I'm trying to move away from NFS due to the poor vw performance and > problems I had with it. Hence wanting to move the master domain and HE > domains over to iscsi domains. I have now an export domain, an ISO domain, > and HostedEngine domain. From my reading, to switch master, you need to > have only 2 data domains, then put the old one into maint and it will > switch to the new one. Only then can you remove the old one... is that > not correct? > > I gained access to that article by creating a dev account, didn't realize > it was free ;) > > > > On Fri, Oct 5, 2018 at 1:22 AM Oliver Riesener < > oliver.riese...@hs-bremen.de> wrote: > >> >> On 10/5/18 8:10 AM, Vincent Royer wrote: >> >> Alright, some combination of messing around got me mostly back up and >> running here. Iscsi is working reliably, VMs are running on the new iscsi >> domain. the old NFS domain is empty save a couple of OVF_STORE disks. >> >> good job >> >> The procedure now, as I understand it, is to shut down all vms, >> >> why ? >> >> put all storage domains into maintenance >> >> which additional domain you have ? >> >> except the old NFS and new SCSI domain, then push the NFS one into >> maintenance, which *should* promote the ISCSI domain to master. >> >> it's going to happen. but it's not importend for ovirt where the master >> lives. >> >> After that, I'll try to move the HE storage to iscsi too, although it >> would be very helpful to have access to this document >> <https://access.redhat.com/solutions/2998291> for that procedure. >> >> Sorry i have also no access, but it sounds like backup/restore ... >> >> RHEV Documentation is <a href= >> "https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.2/" >> <https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.2/> >> >here</a>. >> >> I think it's like E(BareMetal) <a href= >> "https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.2/html/self-hosted_engine_guide/chap-migrating_from_bare_metal_to_a_rhel-based_self-hosted_environment" >> <https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.2/html/self-hosted_engine_guide/chap-migrating_from_bare_metal_to_a_rhel-based_self-hosted_environment>>(see >> here)</a>to HE(ISCSI), but you want HE(NFS) to HE(iSCSI). >> >> It's time for a new mailing thread. >> >> cu >> >> Oliver >> >> >> thanks for your assistance! >> >> On Thu, Oct 4, 2018 at 2:23 PM Oliver Riesener < >> oliver.riese...@hs-bremen.de> wrote: >> >>> Hi Vincent, >>> >>> nice to hear the news :-) >>> >>> I have read the BZ and see you run into NFS trouble and solved it now. >>> >>> I took a look on my centos server for nfs data domains and >>> see server running V4 and the clients (node mounts with protocol >>> vers=4.1) >>> >>> I run the latest (and greatest) ovirt stable 4.2.6.4-1 on centos 7.5+ >>> with >>> engine installed and a ovirt-node 4.2.6.4. >>> >>> * If you can migrate your running VMs and can switch your SPM, >>> i would upgrade and reboot the hosts one by one, now. >>> >>> * reboot seem to be a minimum, remember you do that `virt. thing´, >>> therefor you can access and boot your bare metal and host os ;-) >>> >>> >>> ok, back to iSCSI, i have also a EQUALOGIC running as iSCSI target over >>> years. >>> >>> * I have allowed multi host access to the volumes which ovirt uses. >>> Access control lists contains raw IP addresses from my ovirt-hosts. >>> >>> ovirt handles the volume access virtuos with multipathd and lvm vg’s >>> and lv’s. >>> unused lvs are offline (host specific) and released volumes are >>> deactivated. >>> >>> * Also it’s possible you have to reinstall (from GUI) your hosts, >>> to upgrade or install the needed packages, which handles iSCSI Client >>> access. >>> >>> * If you then free from errors and your iscsi data domain still missing, >>> we talk >>> about vg activation and domain import. >>> >>> Sheers >>> >>> Oliver >>> >>> >>> Am 04.10.2018 um 22:00 schrieb Vincent Royer <vinc...@epicenergy.ca>: >>> >>> Ok, getting somewhere here. >>> >>> did a rpcinfo -p and found no nfs entries in portmap. >>> >>> systemctl stop nfs >>> systemctl start nfs >>> >>> Suddenly shares are mounted and datacenter is up again. >>> >>> was able to add export domain over NFS. >>> >>> Why would nfs shit the bed? >>> >>> still can't seem to get iscsi mounted properly now, and that's where all >>> the disks are located :/ >>> >>> >>> On Thu, Oct 4, 2018 at 11:00 AM Vincent Royer <vinc...@epicenergy.ca> >>> wrote: >>> >>>> Thanks for your help Oliver, >>>> >>>> To give you some background here: >>>> >>>> Host 1 on Ovirt 4.2 attached to NFS storage >>>> Host 2 I upgraded to Ovirt 4.2.5 and then 4.2.6, since then it has had >>>> troubles with NFS due to this bug >>>> https://bugzilla.redhat.com/show_bug.cgi?id=1595549. The host was up >>>> and could run the hosted engine, but could not migrate any VMs to it. >>>> >>>> I decided to switch from NFS to ISCSI so that I could stay on current >>>> releases. So I began the work of attaching iscsi domain. >>>> >>>> The iscsi domain attached, and I transferred most of the disks to it. >>>> Then it started melting down saying that Host 1 could not mount it, and the >>>> whole DC went down. >>>> >>>> Current status is data center "non responsive". Keeps trying >>>> "Reconstructing master domain on Data Center" over and over again but >>>> always fails. Master domain status is "inactive". Clicking activate >>>> fails. The new ISCSI domain, I put in maintenance until I figure the rest >>>> out. I can't add or remove any other domains, Ovirt says I need to attach >>>> the master first. >>>> >>>> Both hosts are "UP". Host 1 health is "bad" Host 2 health is "ok", >>>> and it is running HE. Host 1 (the 4.2 host) says "this host needs to be >>>> reinstalled". But the reinstall option is grayed out. >>>> >>>> I am weary about updating host1, because of the NFS storage bug... I >>>> fear it won't ever be able to attach the old domain again. >>>> >>>> If I try mounting the NFS shares in cockpit from either node, they say >>>> "mount.nfs: Remote I/O error". However on another blank centos machine >>>> sitting on the same network, I can mount the shares normally. >>>> >>>> *Vincent Royer* >>>> *778-825-1057* >>>> >>>> >>>> <http://www.epicenergy.ca/> >>>> *SUSTAINABLE MOBILE ENERGY SOLUTIONS* >>>> >>>> >>>> >>>> >>>> >>>> On Thu, Oct 4, 2018 at 1:04 AM Oliver Riesener < >>>> oliver.riese...@hs-bremen.de> wrote: >>>> >>>>> When your hosts are up and running and your Domain didn't go active >>>>> within minutes >>>>> >>>>> * Activate your Storage Domain under: >>>>> >>>>> Storage -> Storage Domain -> (Open your Domain) -> Data Center -> >>>>> (Right Click Your Data Center Name) -> Activate. >>>>> On 10/4/18 9:50 AM, Oliver Riesener wrote: >>>>> >>>>> Hi Vincent, >>>>> >>>>> OK you master domain, isn't avail a the moment, but no panic. >>>>> >>>>> First off all we need the status from your hosts. No HOSTS -> No >>>>> Storage ! >>>>> >>>>> * Do you reboot them hard, without Confirm "Host has been rebooted" >>>>> >>>>> * Are they actived in the DataCenter / Cluster ? Green Arrow ? >>>>> >>>>> >>>>> On 10/4/18 7:46 AM, Vincent Royer wrote: >>>>> >>>>> I was attempting to migrate from nfs to iscsi storage domains. I have >>>>> reached a state where I can no longer activate the old master storage >>>>> domain, and thus no others will activate either. >>>>> >>>>> I'm ready to give up on the installation and just move to an HCI >>>>> deployment instead. Wipe all the hosts clean and start again. >>>>> >>>>> My plan was to create and use an export domain, then wipe the nodes >>>>> and set them up HCI where I could re-import. But without being able to >>>>> activate a master domain, I can't create the export domain. >>>>> >>>>> I'm not sure why it can't find the master anymore, as nothing has >>>>> happened to the NFS storage, but the error in vdsm says it just can't find >>>>> it: >>>>> >>>>> StoragePoolMasterNotFound: Cannot find master domain: >>>>> u'spUUID=5a77bed1-0238-030c-0122-0000000003b3, >>>>> msdUUID=d3165759-07c2-46ae-b7b8-b6226a929d68' >>>>> 2018-10-03 22:40:33,751-0700 INFO (jsonrpc/3) >>>>> [storage.TaskManager.Task] (Task='83f33db5-90f3-4064-87df-0512ab9b6378') >>>>> aborting: Task is aborted: "Cannot find master domain: >>>>> u'spUUID=5a77bed1-0238-030c-0122-0000000003b3, >>>>> msdUUID=d3165759-07c2-46ae-b7b8-b6226a929d68'" - code 304 (task:1181) >>>>> 2018-10-03 22:40:33,751-0700 ERROR (jsonrpc/3) [storage.Dispatcher] >>>>> FINISH connectStoragePool error=Cannot find master domain: >>>>> u'spUUID=5a77bed1-0238-030c-0122-0000000003b3, >>>>> msdUUID=d3165759-07c2-46ae-b7b8-b6226a929d68' (dispatcher:82) >>>>> 2018-10-03 22:40:33,751-0700 INFO (jsonrpc/3) [jsonrpc.JsonRpcServer] >>>>> RPC call StoragePool.connect failed (error 304) in 0.17 seconds >>>>> (__init__:573) >>>>> 2018-10-03 22:40:34,200-0700 INFO (jsonrpc/1) [api.host] START >>>>> getStats() from=::ffff:172.16.100.13,39028 (api:46) >>>>> >>>>> When I look in cockpit on the hosts, the storage domain is mounted and >>>>> seems fine. >>>>> >>>>> >>>>> >>>>> _______________________________________________ >>>>> Users mailing list -- users@ovirt.org >>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>> oVirt Code of Conduct: >>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>> List Archives: >>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/LTZ6SIFYDFEMSZ4ACUNVC5KETWG7BBIZ/ >>>>> >>>>> -- >>>>> Mit freundlichem Gruß >>>>> >>>>> >>>>> Oliver Riesener >>>>> >>>>> -- >>>>> Hochschule Bremen >>>>> Elektrotechnik und Informatik >>>>> Oliver Riesener >>>>> Neustadtswall 30 >>>>> D-28199 Bremen >>>>> >>>>> Tel: 0421 5905-2405, Fax: -2400e-mail:oliver.riese...@hs-bremen.de >>>>> >>>>> Tel: 0421 5905-2405, Fax: -2400e-mail:oliver.riese...@hs-bremen.de >>>>> >>>>> _______________________________________________ >>>>> Users mailing list -- users@ovirt.org >>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>> oVirt Code of Conduct: >>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>> List Archives: >>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/V72KMULZJAT3XIR3GBTOCA5RLACVQSRC/ >>>>> >>>> _______________________________________________ >>> Users mailing list -- users@ovirt.org >>> To unsubscribe send an email to users-le...@ovirt.org >>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>> oVirt Code of Conduct: >>> https://www.ovirt.org/community/about/community-guidelines/ >>> List Archives: >>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/OZ3TOO554D4YAZJ6FPGI4SAJ6CKWZRFH/ >>> >>> >>> >> _______________________________________________ >> Users mailing list -- users@ovirt.org >> To unsubscribe send an email to users-le...@ovirt.org >> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >> oVirt Code of Conduct: >> https://www.ovirt.org/community/about/community-guidelines/ >> List Archives: >> https://lists.ovirt.org/archives/list/users@ovirt.org/message/XBVJAWRSA2NCNAWEOFKCONJBP7NTPSF2/ >> >> -- >> Mit freundlichem Gruß >> >> >> Oliver Riesener >> >> -- >> Hochschule Bremen >> Elektrotechnik und Informatik >> Oliver Riesener >> Neustadtswall 30 >> D-28199 Bremen >> >> Tel: 0421 5905-2405, Fax: -2400e-mail:oliver.riese...@hs-bremen.de >> >> > _______________________________________________ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/ROT3YZ446QC6BKYD4ZW3VDNOYO3KMEK3/ > > -- > Mit freundlichem Gruß > > > Oliver Riesener > > -- > Hochschule Bremen > Elektrotechnik und Informatik > Oliver Riesener > Neustadtswall 30 > D-28199 Bremen > > Tel: 0421 5905-2405, Fax: -2400e-mail:oliver.riese...@hs-bremen.de > > _______________________________________________ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/Z7QD5VVNPHSWPLPVNDXAF4HCMWFKDORE/ >
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/2JGANPK6OFKPMCPX7NHARBC3JBYGTLWS/