Hi Charles, One question, while configuring a storage domain you are saying "host to use: " node1, then in the connection details you say node1:/data. What about the backup-volfile-servers option in the UI while configuring storage domain? Are you specifying that too?
Thanks kasturi On Fri, Sep 1, 2017 at 5:52 PM, Charles Kozler <ckozler...@gmail.com> wrote: > @ Jim - you have only two data volumes and lost quorum. Arbitrator only > stores metadata, no actual files. So yes, you were running in degraded mode > so some operations were hindered. > > @ Sahina - Yes, this actually worked fine for me once I did that. However, > the issue I am still facing, is when I go to create a new gluster storage > domain (replica 3, hyperconverged) and I tell it "Host to use" and I select > that host. If I fail that host, all VMs halt. I do not recall this in 3.6 > or early 4.0. This to me makes it seem like this is "pinning" a node to a > volume and vice versa like you could, for instance, for a singular > hyperconverged to ex: export a local disk via NFS and then mount it via > ovirt domain. But of course, this has its caveats. To that end, I am using > gluster replica 3, when configuring it I say "host to use: " node 1, then > in the connection details I give it node1:/data. I fail node1, all VMs > halt. Did I miss something? > > On Fri, Sep 1, 2017 at 2:13 AM, Sahina Bose <sab...@redhat.com> wrote: > >> To the OP question, when you set up a gluster storage domain, you need to >> specify backup-volfile-servers=<server2>:<server3> where server2 and >> server3 also have bricks running. When server1 is down, and the volume is >> mounted again - server2 or server3 are queried to get the gluster volfiles. >> >> @Jim, if this does not work, are you using 4.1.5 build with libgfapi >> access? If not, please provide the vdsm and gluster mount logs to analyse >> >> If VMs go to paused state - this could mean the storage is not available. >> You can check "gluster volume status <volname>" to see if atleast 2 bricks >> are running. >> >> On Fri, Sep 1, 2017 at 11:31 AM, Johan Bernhardsson <jo...@kafit.se> >> wrote: >> >>> If gluster drops in quorum so that it has less votes than it should it >>> will stop file operations until quorum is back to normal.If i rember it >>> right you need two bricks to write for quorum to be met and that the >>> arbiter only is a vote to avoid split brain. >>> >>> >>> Basically what you have is a raid5 solution without a spare. And when >>> one disk dies it will run in degraded mode. And some raid systems will stop >>> the raid until you have removed the disk or forced it to run anyway. >>> >>> You can read up on it here: https://gluster.readthed >>> ocs.io/en/latest/Administrator%20Guide/arbiter-volumes-and-quorum/ >>> >>> /Johan >>> >>> On Thu, 2017-08-31 at 22:33 -0700, Jim Kusznir wrote: >>> >>> Hi all: >>> >>> Sorry to hijack the thread, but I was about to start essentially the >>> same thread. >>> >>> I have a 3 node cluster, all three are hosts and gluster nodes (replica >>> 2 + arbitrar). I DO have the mnt_options=backup-volfile-servers= set: >>> >>> storage=192.168.8.11:/engine >>> mnt_options=backup-volfile-servers=192.168.8.12:192.168.8.13 >>> >>> I had an issue today where 192.168.8.11 went down. ALL VMs immediately >>> paused, including the engine (all VMs were running on host2:192.168.8.12). >>> I couldn't get any gluster stuff working until host1 (192.168.8.11) was >>> restored. >>> >>> What's wrong / what did I miss? >>> >>> (this was set up "manually" through the article on setting up >>> self-hosted gluster cluster back when 4.0 was new..I've upgraded it to 4.1 >>> since). >>> >>> Thanks! >>> --Jim >>> >>> >>> On Thu, Aug 31, 2017 at 12:31 PM, Charles Kozler <ckozler...@gmail.com> >>> wrote: >>> >>> Typo..."Set it up and then failed that **HOST**" >>> >>> And upon that host going down, the storage domain went down. I only have >>> hosted storage domain and this new one - is this why the DC went down and >>> no SPM could be elected? >>> >>> I dont recall this working this way in early 4.0 or 3.6 >>> >>> On Thu, Aug 31, 2017 at 3:30 PM, Charles Kozler <ckozler...@gmail.com> >>> wrote: >>> >>> So I've tested this today and I failed a node. Specifically, I setup a >>> glusterfs domain and selected "host to use: node1". Set it up and then >>> failed that VM >>> >>> However, this did not work and the datacenter went down. My engine >>> stayed up, however, it seems configuring a domain to pin to a host to use >>> will obviously cause it to fail >>> >>> This seems counter-intuitive to the point of glusterfs or any redundant >>> storage. If a single host has to be tied to its function, this introduces a >>> single point of failure >>> >>> Am I missing something obvious? >>> >>> On Thu, Aug 31, 2017 at 9:43 AM, Kasturi Narra <kna...@redhat.com> >>> wrote: >>> >>> yes, right. What you can do is edit the hosted-engine.conf file and >>> there is a parameter as shown below [1] and replace h2 and h3 with your >>> second and third storage servers. Then you will need to restart >>> ovirt-ha-agent and ovirt-ha-broker services in all the nodes . >>> >>> [1] 'mnt_options=backup-volfile-servers=<h2>:<h3>' >>> >>> On Thu, Aug 31, 2017 at 5:54 PM, Charles Kozler <ckozler...@gmail.com> >>> wrote: >>> >>> Hi Kasturi - >>> >>> Thanks for feedback >>> >>> > If cockpit+gdeploy plugin would be have been used then that would >>> have automatically detected glusterfs replica 3 volume created during >>> Hosted Engine deployment and this question would not have been asked >>> >>> Actually, doing hosted-engine --deploy it too also auto detects >>> glusterfs. I know glusterfs fuse client has the ability to failover >>> between all nodes in cluster, but I am still curious given the fact that I >>> see in ovirt config node1:/engine (being node1 I set it to in hosted-engine >>> --deploy). So my concern was to ensure and find out exactly how engine >>> works when one node goes away and the fuse client moves over to the other >>> node in the gluster cluster >>> >>> But you did somewhat answer my question, the answer seems to be no (as >>> default) and I will have to use hosted-engine.conf and change the parameter >>> as you list >>> >>> So I need to do something manual to create HA for engine on gluster? Yes? >>> >>> Thanks so much! >>> >>> On Thu, Aug 31, 2017 at 3:03 AM, Kasturi Narra <kna...@redhat.com> >>> wrote: >>> >>> Hi, >>> >>> During Hosted Engine setup question about glusterfs volume is being >>> asked because you have setup the volumes yourself. If cockpit+gdeploy >>> plugin would be have been used then that would have automatically detected >>> glusterfs replica 3 volume created during Hosted Engine deployment and this >>> question would not have been asked. >>> >>> During new storage domain creation when glusterfs is selected there >>> is a feature called 'use managed gluster volumes' and upon checking this >>> all glusterfs volumes managed will be listed and you could choose the >>> volume of your choice from the dropdown list. >>> >>> There is a conf file called /etc/hosted-engine/hosted-engine.conf >>> where there is a parameter called backup-volfile-servers="h1:h2" and if one >>> of the gluster node goes down engine uses this parameter to provide ha / >>> failover. >>> >>> Hope this helps !! >>> >>> Thanks >>> kasturi >>> >>> >>> >>> On Wed, Aug 30, 2017 at 8:09 PM, Charles Kozler <ckozler...@gmail.com> >>> wrote: >>> >>> Hello - >>> >>> I have successfully created a hyperconverged hosted engine setup >>> consisting of 3 nodes - 2 for VM's and the third purely for storage. I >>> manually configured it all, did not use ovirt node or anything. Built the >>> gluster volumes myself >>> >>> However, I noticed that when setting up the hosted engine and even when >>> adding a new storage domain with glusterfs type, it still asks for >>> hostname:/volumename >>> >>> This leads me to believe that if that one node goes down (ex: >>> node1:/data), then ovirt engine wont be able to communicate with that >>> volume because its trying to reach it on node 1 and thus, go down >>> >>> I know glusterfs fuse client can connect to all nodes to provide >>> failover/ha but how does the engine handle this? >>> >>> _______________________________________________ >>> Users mailing list >>> Users@ovirt.org >>> http://lists.ovirt.org/mailman/listinfo/users >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> _______________________________________________ >>> Users mailing list >>> Users@ovirt.org >>> http://lists.ovirt.org/mailman/listinfo/users >>> >>> >>> _______________________________________________ >>> Users mailing >>> listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users >>> >>> >>> _______________________________________________ >>> Users mailing list >>> Users@ovirt.org >>> http://lists.ovirt.org/mailman/listinfo/users >>> >>> >> > > _______________________________________________ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users > >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users