Le 06/08/2015 16:36, Tim Macy a écrit :
Nicolas,  I have the same setup dedicated physical system running engine
on CentOS 6.6 three hosts running CentOS 7.1 with Gluster and KVM, and
firewall is disabled on all hosts.  I also followed the same documents
to build my environment so I assume they are very similar.  I have on
occasion had the same errors and have also found that "ctdb rebalanceip
<floating ip>" is the only way to resolve the problem.  I intend to
remove ctdb since it is not needed with the configuration we are
running.  CTDB is only needed for hosted engine on a floating NFS mount,
so you should be able change the gluster storage domain mount paths to
"localhost:<name>".  The only thing that has prevented me from making
this change is that my environment is live with running VM's.  Please
let me know if you go this route.
>
> Thank you,
> Tim Macy

This week, I eventually took the time to change this, as this DC is not in production.

- Our big NFS storage domain was the master, it contained some VMs
- I wiped all my VMs
- I created a very small temporary NFS master domain, because I did not want to bother with any issue related with erasing the last master storage domain
- I removed the big NFS SD
- I wiped all that was inside, on a filesystem level
[
- I disabled ctdb, removed the "meta" gluster volume that ctdb used for its locks
]
- I added a new storage domain, using your advice :
  - gluster type
  - localhost:<name>
- I removed the temp SD, and all switched correctly on the big glusterFS

I then spent some time playing with P2V, and storing new VMs on this new style glusterFS storage domain. I'm watching the CPU and i/o on the hosts, and yes, they are working, but that keeps sane.

On this particular change (NFS to glusterFS), everything was very smooth.

Regards,

--
Nicolas ECARNOT
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to