[Lustre-discuss] OST Failover Configuration (Active/Active) verification

2014-01-30 Thread Peter Mistich
hello,

anyone here can answer a questions about OST Failover Configuration 
(Active/Active) I think I understand but want to make sure.

I configure 2 oss  servernames = node1 and node2  with 2 shared drives 
/dev/sdb and /dev/sdc and  on node1

I run the command on node1 mkfs.lustre --fsname=testfs --ost 
--failnode=node2 --mgsnode=msg /dev/sdb

I run the command on node2 mkfs.lustre --fsname=testfs --ost 
--failnode=node1 --mgsnode=msg /dev/sdc

I mount  /dev/sdb on node 1 and  mount /dev/sdc on node2

if node1 fails then I just mount  /dev/sdb on node2 and that is how 
active/active works

is this correct ?

Thanks,
Pete
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] OST Failover Configuration (Active/Active) verification

2014-01-30 Thread Jeff Johnson
Peter,

You have it about half right. Since you are dealing with active filesystems
on the OST shared storage devices and you can never truly predict the type
of failure a system will have you need to add a software command/control
layer that will manage failover, remounting storage and cutting power to
the failing node. Failure to do that could result in a split brain
situation where your OST backend filesystem gets corrupted.

You have the mkfs.lustre part right. You need to add heartbeat/corosync to
the configuration and configure it so the two systems monitor each other
with a watchdog heartbeat. Failing machine gets sensed by healthy machine
and the healthy machine shoots it (STONITH: shoot the other node in the
head) via ipmi power control or a smart rack PDU like an APC ethernet
managed PDU.

The heartbeat/corosync config takes your existing config and adds automated
directives like:
node1 mount sdb : node2 mounts sdc
if node1 dies node2 mounts sdb
if node2 dies node1 mounts sdc
if surviving node senses restoration of heartbeat local defined storage
gets remounted to owning node

Intel's IEEL Lustre distribution does all of this sort of thing
automagically. Or you can manually install Lustre and the corosync app
packages and configure it manually.

--Jeff


On Thu, Jan 30, 2014 at 9:21 PM, Peter Mistich
peter.mist...@rackspace.comwrote:

 hello,

 anyone here can answer a questions about OST Failover Configuration
 (Active/Active) I think I understand but want to make sure.

 I configure 2 oss  servernames = node1 and node2  with 2 shared drives
 /dev/sdb and /dev/sdc and  on node1

 I run the command on node1 mkfs.lustre --fsname=testfs --ost
 --failnode=node2 --mgsnode=msg /dev/sdb

 I run the command on node2 mkfs.lustre --fsname=testfs --ost
 --failnode=node1 --mgsnode=msg /dev/sdc

 I mount  /dev/sdb on node 1 and  mount /dev/sdc on node2

 if node1 fails then I just mount  /dev/sdb on node2 and that is how
 active/active works

 is this correct ?

 Thanks,
 Pete
 ___
 Lustre-discuss mailing list
 Lustre-discuss@lists.lustre.org
 http://lists.lustre.org/mailman/listinfo/lustre-discuss




-- 
--
Jeff Johnson
Co-Founder
Aeon Computing

jeff.john...@aeoncomputing.com
www.aeoncomputing.com
t: 858-412-3810 x1001   f: 858-412-3845
m: 619-204-9061

4170 Morena Boulevard, Suite D - San Diego, CA 92117

High-Performance Computing / Lustre Filesystems / Scale-out Storage
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] OST Failover Configuration (Active/Active) verification

2014-01-30 Thread Peter Mistich

Jeff,


Thanks for the verification.

I understand I need something for the software command/control layer I 
was planning to use redhat cluster.


I was checking to make sure I had lustre fail-over correct.

Thank you again for verifying. I am testing it now.

Pete

On 01/30/2014 11:57 PM, Jeff Johnson wrote:

Peter,

You have it about half right. Since you are dealing with active 
filesystems on the OST shared storage devices and you can never truly 
predict the type of failure a system will have you need to add a 
software command/control layer that will manage failover, remounting 
storage and cutting power to the failing node. Failure to do that 
could result in a split brain situation where your OST backend 
filesystem gets corrupted.


You have the mkfs.lustre part right. You need to add 
heartbeat/corosync to the configuration and configure it so the two 
systems monitor each other with a watchdog heartbeat. Failing machine 
gets sensed by healthy machine and the healthy machine shoots it 
(STONITH: shoot the other node in the head) via ipmi power control or 
a smart rack PDU like an APC ethernet managed PDU.


The heartbeat/corosync config takes your existing config and adds 
automated directives like:

node1 mount sdb : node2 mounts sdc
if node1 dies node2 mounts sdb
if node2 dies node1 mounts sdc
if surviving node senses restoration of heartbeat local defined 
storage gets remounted to owning node


Intel's IEEL Lustre distribution does all of this sort of thing 
automagically. Or you can manually install Lustre and the corosync app 
packages and configure it manually.


--Jeff


On Thu, Jan 30, 2014 at 9:21 PM, Peter Mistich 
peter.mist...@rackspace.com mailto:peter.mist...@rackspace.com wrote:


hello,

anyone here can answer a questions about OST Failover Configuration
(Active/Active) I think I understand but want to make sure.

I configure 2 oss  servernames = node1 and node2  with 2 shared drives
/dev/sdb and /dev/sdc and  on node1

I run the command on node1 mkfs.lustre --fsname=testfs --ost
--failnode=node2 --mgsnode=msg /dev/sdb

I run the command on node2 mkfs.lustre --fsname=testfs --ost
--failnode=node1 --mgsnode=msg /dev/sdc

I mount  /dev/sdb on node 1 and  mount /dev/sdc on node2

if node1 fails then I just mount  /dev/sdb on node2 and that is how
active/active works

is this correct ?

Thanks,
Pete
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
mailto:Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss




--
--
Jeff Johnson
Co-Founder
Aeon Computing

jeff.john...@aeoncomputing.com mailto:jeff.john...@aeoncomputing.com
www.aeoncomputing.com http://www.aeoncomputing.com
t: 858-412-3810 x1001   f: 858-412-3845
m: 619-204-9061

4170 Morena Boulevard, Suite D - San Diego, CA 92117

High-Performance Computing / Lustre Filesystems / Scale-out Storage


___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss