Hi! AFAIK, you'll have to have the same device name on every cluster node, and that one device should be actually be the same shared disk.
Regards, Ulrich >>> Fredrik Hudner <[email protected]> schrieb am 28.03.2013 um 10:52 in Nachricht <CAFtwCgP+k83ebWkTrQGjrSY8QLc2LWVNMYfFOE5yp6=u_8+...@mail.gmail.com>: > On Tue, Mar 26, 2013 at 4:42 PM, Lars Marowsky-Bree <[email protected]> wrote: > > > On 2013-03-26T08:10:06, Fredrik Hudner <[email protected]> wrote: > > > > > Hi, > > > I have a question about the setup sbd which I think belong to this forum > > > > > > I have 3 nodes. Two active pacemaker nodes and one kind of quorum node. > > > I would like to setup sbd as fencing device between these 3 nodes that > > are > > > running in VMware instances. > > > > > > Best demonstrated practise is the use of 3 devices (disks) but I'm not > > sure > > > about the actual setup. > > > > You must map the same 1 to 3 block devices to the VMs so that concurrent > > shared read/write access is allowed. > > > > Also, if the block devices are not independent but partitions on the > > same disk, that's not worth bothering with. > > > > If they're all on the same host, that's a bit pointless, but you knew > > that ;-) > > > > How are you backing the devices? SAN, FC, iSCSI? If it's iSCSI, you > > could decide to directly run the iSCSI initiator in the guest. > > > > > I have read the http://linux-ha.org/wiki/SBD_Fencing but it only help > > me > > > so far > > > > That's because it describes how to configure sbd, not your storage layer > > ;-) > > > > Thanks Lars, > > > > The backing device is over SAN and the devices are all independent. > > > > > > So when you say, map the same 1 to 3 block devices my approach should be > > correct ? > > > > 1/ add 1 block device on each of the 3 nodes, e.g /dev/sdd, /dev/sde > and /dev/sdf > 2/ map each block device to the other VM's > > When I run *# sbd -d /dev/sdd list *(same for all devices) it only > comes back with all the options I can use. > I then run *# sbd -d /dev/sde allocate <nodename>* for each device and > I can see e.g > on testclu01 (node1) > *# sbd -d /dev/sde list > * > * 0 testclu02 test testclu01* > *but on /dev/sdd > # sbd -d /dev/sdd list > 0 testclu01 clear* > *# sbd -d /dev/sdf list* > (empty) > > on testclu03 (node3): > *# sbd -d /dev/sde list* > *0 testclu01 clear > # sbd -d /dev/sdf list > **0 testclu02 test testclu01* > *# sbd -d /dev/sdd list > *(empty)* > * > So I'm not sure if this is ok or not ? > > > > > Architect Storage/HA > > SUSE LINUX Products GmbH, GF: Jeff Hawn, Jennifer Guild, Felix > > Imendörffer, HRB 21284 (AG Nürnberg) > > "Experience is the name everyone gives to their mistakes." -- Oscar Wilde > > > > _______________________________________________ > > Linux-HA mailing list > > [email protected] > > http://lists.linux-ha.org/mailman/listinfo/linux-ha > > See also: http://linux-ha.org/ReportingProblems > > > > _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
