[Lustre-discuss] Mount 2 clusters, different networks - LNET tcp1-tcp2-o2ib

2011-06-14 Thread Thomas Roth
Hi all,

I'd like to mount two Lustre filesystems on one client. Issues with more than 
one MGS set aside,
the point here is that one of them is an Infiniband-cluster, the other is 
ethernet-based.
And my client is on the ethernet.
I have managed to mount the o2ib-fs by setting up an LNET router, but now this 
client's LNET doesn;t 
known how to reach the ethernet-fs.

So the basic modprobe.conf reads
  options lnet networks=tcp1(eth0) routes=o2ib LNET-Router-IP@tcp1
This mounts the MGS on the o2ib network.

What do I have to add to get to the MGS on the tpc network?

Meanwhile I have studied more posts here and came up with
  options lnet networks=tcp1(eth0),tcp2(eth0:0) routes=o2ib 
  LNET-Router-IP@tcp1; tcp 
Default-Gateway-IP@tcp2

Doesn't work either, but I see in the log of the (tcp-)MGS:
  LustreError: 120-3: Refusing connection from Client-IP for MGS-IP@tcp2: No 
  matching NI

Somethings getting through ...

Any ideas?

Regards,
Thomas
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] Mount 2 clusters, different networks - LNET tcp1-tcp2-o2ib

2011-06-14 Thread Michael Shuey
Is your ethernet FS in tcp1, or tcp0?  Your config bits indicate the
client is in tcp1 - do the servers agree?

--
Mike Shuey



On Tue, Jun 14, 2011 at 12:23 PM, Thomas Roth t.r...@gsi.de wrote:
 Hi all,

 I'd like to mount two Lustre filesystems on one client. Issues with more than 
 one MGS set aside,
 the point here is that one of them is an Infiniband-cluster, the other is 
 ethernet-based.
 And my client is on the ethernet.
 I have managed to mount the o2ib-fs by setting up an LNET router, but now 
 this client's LNET doesn;t
 known how to reach the ethernet-fs.

 So the basic modprobe.conf reads
   options lnet networks=tcp1(eth0) routes=o2ib LNET-Router-IP@tcp1
 This mounts the MGS on the o2ib network.

 What do I have to add to get to the MGS on the tpc network?

 Meanwhile I have studied more posts here and came up with
   options lnet networks=tcp1(eth0),tcp2(eth0:0) routes=o2ib 
 LNET-Router-IP@tcp1; tcp
 Default-Gateway-IP@tcp2

 Doesn't work either, but I see in the log of the (tcp-)MGS:
   LustreError: 120-3: Refusing connection from Client-IP for MGS-IP@tcp2: No 
 matching NI

 Somethings getting through ...

 Any ideas?

 Regards,
 Thomas
 ___
 Lustre-discuss mailing list
 Lustre-discuss@lists.lustre.org
 http://lists.lustre.org/mailman/listinfo/lustre-discuss

___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] Mount 2 clusters, different networks - LNET tcp1-tcp2-o2ib

2011-06-14 Thread Thomas Roth
Hm, the ethernet FS is in tcp0 - MGS says its nids are MGS-IP@tcp.
So not surprising it refuses that connection.
On the other hand,
  options lnet networks=tcp1(eth0),tcp(eth0:0) routes=o2ib 
  LNET-Router-IP@tcp1; tcp 
Default-Gateway-IP@tcp

results in
  Can't create route to tcp via Gateway-IP@tcp

Cheers,
Thomas


On 06/14/2011 07:00 PM, Michael Shuey wrote:
 Is your ethernet FS in tcp1, or tcp0? Your config bits indicate the
 client is in tcp1 - do the servers agree?

 --
 Mike Shuey



 On Tue, Jun 14, 2011 at 12:23 PM, Thomas Roth t.r...@gsi.de wrote:
   Hi all,
  
   I'd like to mount two Lustre filesystems on one client. Issues with more 
 than one MGS set aside,
   the point here is that one of them is an Infiniband-cluster, the other is 
 ethernet-based.
   And my client is on the ethernet.
   I have managed to mount the o2ib-fs by setting up an LNET router, but now 
 this client's LNET doesn;t
   known how to reach the ethernet-fs.
  
   So the basic modprobe.conf reads
options lnet networks=tcp1(eth0) routes=o2ib LNET-Router-IP@tcp1
   This mounts the MGS on the o2ib network.
  
   What do I have to add to get to the MGS on the tpc network?
  
   Meanwhile I have studied more posts here and came up with
options lnet networks=tcp1(eth0),tcp2(eth0:0) routes=o2ib 
 LNET-Router-IP@tcp1; tcp
   Default-Gateway-IP@tcp2
  
   Doesn't work either, but I see in the log of the (tcp-)MGS:
LustreError: 120-3: Refusing connection from Client-IP for MGS-IP@tcp2: 
 No matching NI
  
   Somethings getting through ...
  
   Any ideas?
  
   Regards,
   Thomas
   ___
   Lustre-discuss mailing list
   Lustre-discuss@lists.lustre.org
   http://lists.lustre.org/mailman/listinfo/lustre-discuss
  



___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] Mount 2 clusters, different networks - LNET tcp1-tcp2-o2ib

2011-06-14 Thread Michael Shuey
That may be because your gateway doesn't have an interface on tcp (aka
tcp0).  I suspect you want to keep your ethernet clients in tcp0, your
IB clients in o2ib0, and your router in both.  Personally, I find it
easiest to just give different module options on each system (rather
than try ip2nets stuff).

On the ether clients, I'd try:

options lnet networks=tcp0(eth0) routes=o2ib0
LNET-router-eth_IP@tcp0 dead_router_check_interval=300

On IB clients:

options lnet networks=o2ib0(ib0) routes=tcp0 LNET-router-IB_IP@ib0
dead_router_check_interval=300

then on the router:

options lnet networks=tcp0(eth0),o2ib0(ib0) forwarding=enabled accept_timeout=15

Obviously, your file servers will need to have lnet options similar to
the clients:

options lnet networks=tcp0(eth0) routes=o2ib0
LNET-router-eth_IP@tcp0 dead_router_check_interval=300
options lnet networks=o2ib0(ib0) routes=tcp0 LNET-router-IB_IP@o2ib0
dead_router_check_interval=300

That's just a guess, your mileage may vary, etc., but I think it's
close to what you want.  Note that you really want the
dead_router_check_interval if you're using lnet routers.  Without that
parameter, the lustre client will automatically mark a router as
failed when it's unavailable but will not check to see if it ever
comes back.  With this param, it checks every 300 seconds (and
re-enables it if found).

Hope this helps.

--
Mike Shuey



On Tue, Jun 14, 2011 at 1:26 PM, Thomas Roth t.r...@gsi.de wrote:
 Hm, the ethernet FS is in tcp0 - MGS says its nids are MGS-IP@tcp.
 So not surprising it refuses that connection.
 On the other hand,
 options lnet networks=tcp1(eth0),tcp(eth0:0) routes=o2ib
 LNET-Router-IP@tcp1; tcp Default-Gateway-IP@tcp

 results in
 Can't create route to tcp via Gateway-IP@tcp

 Cheers,
 Thomas


 On 06/14/2011 07:00 PM, Michael Shuey wrote:

 Is your ethernet FS in tcp1, or tcp0? Your config bits indicate the
 client is in tcp1 - do the servers agree?

 --
 Mike Shuey



 On Tue, Jun 14, 2011 at 12:23 PM, Thomas Roth t.r...@gsi.de wrote:
   Hi all,
  
   I'd like to mount two Lustre filesystems on one client. Issues with
 more than one MGS set aside,
   the point here is that one of them is an Infiniband-cluster, the other
 is ethernet-based.
   And my client is on the ethernet.
   I have managed to mount the o2ib-fs by setting up an LNET router, but
 now this client's LNET doesn;t
   known how to reach the ethernet-fs.
  
   So the basic modprobe.conf reads
    options lnet networks=tcp1(eth0) routes=o2ib LNET-Router-IP@tcp1
   This mounts the MGS on the o2ib network.
  
   What do I have to add to get to the MGS on the tpc network?
  
   Meanwhile I have studied more posts here and came up with
    options lnet networks=tcp1(eth0),tcp2(eth0:0) routes=o2ib
 LNET-Router-IP@tcp1; tcp
   Default-Gateway-IP@tcp2
  
   Doesn't work either, but I see in the log of the (tcp-)MGS:
    LustreError: 120-3: Refusing connection from Client-IP for
 MGS-IP@tcp2: No matching NI
  
   Somethings getting through ...
  
   Any ideas?
  
   Regards,
   Thomas
   ___
   Lustre-discuss mailing list
   Lustre-discuss@lists.lustre.org
   http://lists.lustre.org/mailman/listinfo/lustre-discuss
  




___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] Mount 2 clusters, different networks - LNET tcp1-tcp2-o2ib

2011-06-14 Thread Thomas Roth
Thanks, Michael.
I'll certainly put in the check_interval, that will be needed.

However, what I tried was to have an ethernet client that mounts one FS via the 
LNET router 
(Infiniband-FS behind it) and simultaneously mounts the other FS, which is on 
tcp0  - via its default 
route. So actually I don't have any IB clients (except for the LNET routers).

Probably I messed up the tcpX-network names.

Cheers,
Thomas

On 06/14/2011 08:04 PM, Michael Shuey wrote:
 That may be because your gateway doesn't have an interface on tcp (aka
 tcp0). I suspect you want to keep your ethernet clients in tcp0, your
 IB clients in o2ib0, and your router in both. Personally, I find it
 easiest to just give different module options on each system (rather
 than try ip2nets stuff).

 On the ether clients, I'd try:

 options lnet networks=tcp0(eth0) routes=o2ib0
 LNET-router-eth_IP@tcp0 dead_router_check_interval=300

 On IB clients:

 options lnet networks=o2ib0(ib0) routes=tcp0 LNET-router-IB_IP@ib0
 dead_router_check_interval=300

 then on the router:

 options lnet networks=tcp0(eth0),o2ib0(ib0) forwarding=enabled 
 accept_timeout=15

 Obviously, your file servers will need to have lnet options similar to
 the clients:

 options lnet networks=tcp0(eth0) routes=o2ib0
 LNET-router-eth_IP@tcp0 dead_router_check_interval=300
 options lnet networks=o2ib0(ib0) routes=tcp0 LNET-router-IB_IP@o2ib0
 dead_router_check_interval=300

 That's just a guess, your mileage may vary, etc., but I think it's
 close to what you want. Note that you really want the
 dead_router_check_interval if you're using lnet routers. Without that
 parameter, the lustre client will automatically mark a router as
 failed when it's unavailable but will not check to see if it ever
 comes back. With this param, it checks every 300 seconds (and
 re-enables it if found).

 Hope this helps.

 --
 Mike Shuey



 On Tue, Jun 14, 2011 at 1:26 PM, Thomas Roth t.r...@gsi.de wrote:
   Hm, the ethernet FS is in tcp0 - MGS says its nids are MGS-IP@tcp.
   So not surprising it refuses that connection.
   On the other hand,
   options lnet networks=tcp1(eth0),tcp(eth0:0) routes=o2ib
   LNET-Router-IP@tcp1; tcp Default-Gateway-IP@tcp
  
   results in
   Can't create route to tcp via Gateway-IP@tcp
  
   Cheers,
   Thomas
  
  
   On 06/14/2011 07:00 PM, Michael Shuey wrote:
  
   Is your ethernet FS in tcp1, or tcp0? Your config bits indicate the
   client is in tcp1 - do the servers agree?
  
   --
   Mike Shuey
  
  
  
   On Tue, Jun 14, 2011 at 12:23 PM, Thomas Roth t.r...@gsi.de wrote:
Hi all,
   
I'd like to mount two Lustre filesystems on one client. Issues with
   more than one MGS set aside,
the point here is that one of them is an Infiniband-cluster, the other
   is ethernet-based.
And my client is on the ethernet.
I have managed to mount the o2ib-fs by setting up an LNET router, but
   now this client's LNET doesn;t
known how to reach the ethernet-fs.
   
So the basic modprobe.conf reads
 options lnet networks=tcp1(eth0) routes=o2ib LNET-Router-IP@tcp1
This mounts the MGS on the o2ib network.
   
What do I have to add to get to the MGS on the tpc network?
   
Meanwhile I have studied more posts here and came up with
 options lnet networks=tcp1(eth0),tcp2(eth0:0) routes=o2ib
   LNET-Router-IP@tcp1; tcp
Default-Gateway-IP@tcp2
   
Doesn't work either, but I see in the log of the (tcp-)MGS:
 LustreError: 120-3: Refusing connection from Client-IP for
   MGS-IP@tcp2: No matching NI
   
Somethings getting through ...
   
Any ideas?
   
Regards,
Thomas
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss
   
  
  
  
  



-- 

Thomas Roth
Department: Informationstechnologie
Location: SB3 1.262
Phone: +49-6159-71 1453  Fax: +49-6159-71 2986

GSI Helmholtzzentrum für Schwerionenforschung GmbH
Planckstraße 1
64291 Darmstadt
www.gsi.de

Gesellschaft mit beschränkter Haftung
Sitz der Gesellschaft: Darmstadt
Handelsregister: Amtsgericht Darmstadt, HRB 1528

Geschäftsführung: Professor Dr. Dr. h.c. Horst Stöcker,
Dr. Hartmut Eickhoff

Vorsitzende des Aufsichtsrates: Dr. Beatrix Vierkorn-Rudolph
Stellvertreter: Ministerialdirigent Dr. Rolf Bernhardt

___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] Mount 2 clusters, different networks - LNET tcp1-tcp2-o2ib - solved?

2011-06-14 Thread Thomas Roth
Hi all,

this seems to work with the correct IPs and correct network names ;-[]
I now have the following modprobe on my ethernet client:
  options lnet networks=tcp1(eth0),tcp0(eth0:0) routes=o2ib LNET-Router@tcp1; 
  tcp Default-Route@tcp1

With these options, loading the modules gives me
  Jun 14 20:12:55 kernel: Lustre: Added LNI 10.12.70.183@tcp1 [8/256/0/180]
  Jun 14 20:12:55 kernel: Lustre: Added LNI 10.12.0.21@tcp [8/256/0/180]
which are the IPs of eth0 and eth0:0.


Now I still wonder why the alias interface eth0:0 is necessary (if left out, 
the whole endeavor 
fails). The routes=statement seems to say: If you have data for tcp, use the 
Default-Router-IP and go 
via the interace that is on network tcp1.

Oh well, I should probably take some networking lectures...

Regards,
Thomas


On 06/14/2011 06:23 PM, Thomas Roth wrote:
 Hi all,

 I'd like to mount two Lustre filesystems on one client. Issues with more than 
 one MGS set aside,
 the point here is that one of them is an Infiniband-cluster, the other is 
 ethernet-based.
 And my client is on the ethernet.
 I have managed to mount the o2ib-fs by setting up an LNET router, but now 
 this client's LNET doesn;t
 known how to reach the ethernet-fs.

 So the basic modprobe.conf reads
 options lnet networks=tcp1(eth0) routes=o2ib LNET-Router-IP@tcp1
 This mounts the MGS on the o2ib network.

 What do I have to add to get to the MGS on the tpc network?

 Meanwhile I have studied more posts here and came up with
 options lnet networks=tcp1(eth0),tcp2(eth0:0) routes=o2ib 
 LNET-Router-IP@tcp1; tcp
 Default-Gateway-IP@tcp2

 Doesn't work either, but I see in the log of the (tcp-)MGS:
 LustreError: 120-3: Refusing connection from Client-IP for MGS-IP@tcp2: 
 No matching NI

 Somethings getting through ...

 Any ideas?

 Regards,
 Thomas
 ___
 Lustre-discuss mailing list
 Lustre-discuss@lists.lustre.org
 http://lists.lustre.org/mailman/listinfo/lustre-discuss


___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] Enabling mds failover after filesystem creation

2011-06-14 Thread Jeff Johnson
Greetings,

I am attempting to add mds failover operation to an existing v1.8.4 
filesystem. I have heartbeat/stonith configured on the mds nodes. What 
is unclear is what to change in the lustre parameters. I have read over 
the 1.8.x and 2.0 manuals and they are unclear as exactly how to enable 
failover mds operation on an existing filesystem.

Do I simply run the following on the primary mds node and specify the 
NID of the secondary mds node?

tunefs.lustre --param=failover.node=10.0.1.3@o2ib /dev/mdt device

where: 10.0.1.2=primary mds, 10.0.1.3=secondary mds

All of the examples for enabling failover via tunefs.lustre are for OSTs 
and I want to be sure that there isn't a different procedure for the MDS 
since it can only be active/passive.

Thanks,

--Jeff

--
Jeff Johnson
Aeon Computing

www.aeoncomputing.com
4905 Morena Boulevard, Suite 1313 - San Diego, CA 92117

___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] Enabling mds failover after filesystem creation

2011-06-14 Thread Cliff White
It depends - are you using a combined MGS/MDS?
If so, you will have to update the mgsnid on all servers to reflect the
failover node,
plus change the client mount string to show the failover node.
otherwise, it's the same procedure as with an OST.
cliffw


On Tue, Jun 14, 2011 at 12:06 PM, Jeff Johnson 
jeff.john...@aeoncomputing.com wrote:

 Greetings,

 I am attempting to add mds failover operation to an existing v1.8.4
 filesystem. I have heartbeat/stonith configured on the mds nodes. What
 is unclear is what to change in the lustre parameters. I have read over
 the 1.8.x and 2.0 manuals and they are unclear as exactly how to enable
 failover mds operation on an existing filesystem.

 Do I simply run the following on the primary mds node and specify the
 NID of the secondary mds node?

 tunefs.lustre --param=failover.node=10.0.1.3@o2ib /dev/mdt device

 where: 10.0.1.2=primary mds, 10.0.1.3=secondary mds

 All of the examples for enabling failover via tunefs.lustre are for OSTs
 and I want to be sure that there isn't a different procedure for the MDS
 since it can only be active/passive.

 Thanks,

 --Jeff

 --
 Jeff Johnson
 Aeon Computing

 www.aeoncomputing.com
 4905 Morena Boulevard, Suite 1313 - San Diego, CA 92117

 ___
 Lustre-discuss mailing list
 Lustre-discuss@lists.lustre.org
 http://lists.lustre.org/mailman/listinfo/lustre-discuss




-- 
cliffw
Support Guy
WhamCloud, Inc.
www.whamcloud.com
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] Enabling mds failover after filesystem creation

2011-06-14 Thread Jeff Johnson

Apologies, I should have been more descriptive.

I am running a dedicated MGS node and MGT device. The MDT is a 
standalone RAID-10 shared via SAS between two nodes, one being the 
current MDS and the second being the planned secondary MDS. Heartbeat 
and stonith w/ ipmi control is currently configured but not started 
between the two nodes.



On 6/14/11 12:12 PM, Cliff White wrote:

It depends - are you using a combined MGS/MDS?
If so, you will have to update the mgsnid on all servers to reflect 
the failover node,

plus change the client mount string to show the failover node.
otherwise, it's the same procedure as with an OST.
cliffw


On Tue, Jun 14, 2011 at 12:06 PM, Jeff Johnson 
jeff.john...@aeoncomputing.com 
mailto:jeff.john...@aeoncomputing.com wrote:


Greetings,

I am attempting to add mds failover operation to an existing v1.8.4
filesystem. I have heartbeat/stonith configured on the mds nodes. What
is unclear is what to change in the lustre parameters. I have read
over
the 1.8.x and 2.0 manuals and they are unclear as exactly how to
enable
failover mds operation on an existing filesystem.

Do I simply run the following on the primary mds node and specify the
NID of the secondary mds node?

tunefs.lustre --param=failover.node=10.0.1.3@o2ib /dev/mdt device

where: 10.0.1.2=primary mds, 10.0.1.3=secondary mds

All of the examples for enabling failover via tunefs.lustre are
for OSTs
and I want to be sure that there isn't a different procedure for
the MDS
since it can only be active/passive.

Thanks,

--Jeff

--
Jeff Johnson
Aeon Computing

www.aeoncomputing.com http://www.aeoncomputing.com
4905 Morena Boulevard, Suite 1313 - San Diego, CA 92117

___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
mailto:Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss




--
cliffw
Support Guy
WhamCloud, Inc.
www.whamcloud.com http://www.whamcloud.com




--

Jeff Johnson
Aeon Computing

www.aeoncomputing.com
4905 Morena Boulevard, Suite 1313 - San Diego, CA 92117

___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] Enabling mds failover after filesystem creation

2011-06-14 Thread Cliff White
Then it should be the same as the OST case. The only difference between the
two
is that we never allow two active MDSs on the same filesystem, so MDT is
always active/passive.
cliffw

On Tue, Jun 14, 2011 at 12:18 PM, Jeff Johnson 
jeff.john...@aeoncomputing.com wrote:

  Apologies, I should have been more descriptive.

 I am running a dedicated MGS node and MGT device. The MDT is a standalone
 RAID-10 shared via SAS between two nodes, one being the current MDS and the
 second being the planned secondary MDS. Heartbeat and stonith w/ ipmi
 control is currently configured but not started between the two nodes.



 On 6/14/11 12:12 PM, Cliff White wrote:

 It depends - are you using a combined MGS/MDS?
 If so, you will have to update the mgsnid on all servers to reflect the
 failover node,
 plus change the client mount string to show the failover node.
  otherwise, it's the same procedure as with an OST.
 cliffw


 On Tue, Jun 14, 2011 at 12:06 PM, Jeff Johnson 
 jeff.john...@aeoncomputing.com wrote:

 Greetings,

 I am attempting to add mds failover operation to an existing v1.8.4
 filesystem. I have heartbeat/stonith configured on the mds nodes. What
 is unclear is what to change in the lustre parameters. I have read over
 the 1.8.x and 2.0 manuals and they are unclear as exactly how to enable
 failover mds operation on an existing filesystem.

 Do I simply run the following on the primary mds node and specify the
 NID of the secondary mds node?

 tunefs.lustre --param=failover.node=10.0.1.3@o2ib /dev/mdt device

 where: 10.0.1.2=primary mds, 10.0.1.3=secondary mds

 All of the examples for enabling failover via tunefs.lustre are for OSTs
 and I want to be sure that there isn't a different procedure for the MDS
 since it can only be active/passive.

 Thanks,

 --Jeff

 --
 Jeff Johnson
 Aeon Computing

 www.aeoncomputing.com
 4905 Morena Boulevard, Suite 1313 - San Diego, CA 92117

 ___
 Lustre-discuss mailing list
 Lustre-discuss@lists.lustre.org
 http://lists.lustre.org/mailman/listinfo/lustre-discuss




 --
 cliffw
 Support Guy
 WhamCloud, Inc.
 www.whamcloud.com



 --

 Jeff Johnson
 Aeon Computing
 www.aeoncomputing.com
 4905 Morena Boulevard, Suite 1313 - San Diego, CA 92117


 ___
 Lustre-discuss mailing list
 Lustre-discuss@lists.lustre.org
 http://lists.lustre.org/mailman/listinfo/lustre-discuss




-- 
cliffw
Support Guy
WhamCloud, Inc.
www.whamcloud.com
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] Where should SHARED_DIRECTORY of acc-sm cfg variable set to?

2011-06-14 Thread Andreas Dilger
On 2011-06-14, at 9:57 AM, Surya, Prakash B. wrote:
 Perhaps it would be beneficial to add Andreas's comment to the source to 
 avoid any future confusion?
 
 {code}
# This is used by a small number of tests to share state between the client
# running the tests, or in some cases between the servers (e.g. lfsck.sh).
# It needs to be a non-lustre filesystem that is available on all the nod
export SHARED_DIRECTORY=${SHARED_DIRECTORY:-/tmp}
 {/code}

Sure, a patch would be quickly approved.

 
 From: lustre-discuss-boun...@lists.lustre.org 
 [lustre-discuss-boun...@lists.lustre.org] On Behalf Of Andreas Dilger 
 [adil...@whamcloud.com]
 Sent: Monday, June 13, 2011 9:24 PM
 To: Jay Lan
 Cc: lustre-discuss@lists.lustre.org
 Subject: Re: [Lustre-discuss] Where should SHARED_DIRECTORY of acc-sm cfg 
   variable set to?
 
 On 2011-06-13, at 7:41 PM, Jay Lan wrote:
 To run some test suites of acc-sm, the SHARED_DIRECTORY variable
 needs to be set. Where should it be set to?
 
 Bugzilla 17839 Comment #65 said: the variable must contain the path of
 a shared filesystem.
 
 Does it need to be shared by what nodes? The clients? The clients + OST'es?
 Can it be $MOUNT? Or should it be a non-lustre filesystem that is mounted
 by NFS? Please advise. Thanks!
 
 This is used by a small number of tests to share state between the client
 running the tests, or in some cases between the servers (e.g. lfsck.sh).
 It needs to be a non-lustre filesystem that is available on all the nodes.
 
 Cheers, Andreas
 --
 Andreas Dilger
 Principal Engineer
 Whamcloud, Inc.
 
 
 
 ___
 Lustre-discuss mailing list
 Lustre-discuss@lists.lustre.org
 http://lists.lustre.org/mailman/listinfo/lustre-discuss


Cheers, Andreas
--
Andreas Dilger 
Principal Engineer
Whamcloud, Inc.



___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] lustre from source EXTRAVERSION

2011-06-14 Thread Michael Di Domenico
I'm trying to rebuild the RHEL kernel with the lustre patches, most
everything has gone okay one time through so far, but i ran into an
issue when trying to get OFED to compile against the new code

According to the Whamcloud wiki I am to perform this step

Add a unique build id so we can be certain our kernel is booted. Edit
~build/kernel/rpmbuild/BUILD/kernel-2.6.18/linux-2.6.18.x86_64/Makefile
and modify line 4, the EXTRAVERSION to read:
EXTRAVERSION = -lustre18

And then further down, this one

Go into the kernel source directory and issue the following commands
to build a kernel rpm.
cd /build/kernel/rpmbuild/BUILD/kernel-2.6.18/linux-2.6.18.x86_64
make oldconfig || make menuconfig
make include/asm
make include/linux/version.h
make SUBDIRS=scripts
make include/linux/utsrelease.h
make rpm

Two questions:

1.  No matter what i do, i cannot get the kernel version to come out
as 2.6.18-238.el5.lustre185

I can't seem to get EXTRAVERSION to take dashes, I end up with
2.6.18238.el5.lustre185 or given the example above 2.6.18lustre18, is
there a fix for this?

2.  The resulting rpm does not include an initrd file.  I can
certainly create one on my own, but is there a parameter or step that
will allow 'make rpm' to do that for me and include it in the rpm?

thanks
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss