[Lustre-discuss] Mount 2 clusters, different networks - LNET tcp1-tcp2-o2ib
Hi all, I'd like to mount two Lustre filesystems on one client. Issues with more than one MGS set aside, the point here is that one of them is an Infiniband-cluster, the other is ethernet-based. And my client is on the ethernet. I have managed to mount the o2ib-fs by setting up an LNET router, but now this client's LNET doesn;t known how to reach the ethernet-fs. So the basic modprobe.conf reads options lnet networks=tcp1(eth0) routes=o2ib LNET-Router-IP@tcp1 This mounts the MGS on the o2ib network. What do I have to add to get to the MGS on the tpc network? Meanwhile I have studied more posts here and came up with options lnet networks=tcp1(eth0),tcp2(eth0:0) routes=o2ib LNET-Router-IP@tcp1; tcp Default-Gateway-IP@tcp2 Doesn't work either, but I see in the log of the (tcp-)MGS: LustreError: 120-3: Refusing connection from Client-IP for MGS-IP@tcp2: No matching NI Somethings getting through ... Any ideas? Regards, Thomas ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Mount 2 clusters, different networks - LNET tcp1-tcp2-o2ib
Is your ethernet FS in tcp1, or tcp0? Your config bits indicate the client is in tcp1 - do the servers agree? -- Mike Shuey On Tue, Jun 14, 2011 at 12:23 PM, Thomas Roth t.r...@gsi.de wrote: Hi all, I'd like to mount two Lustre filesystems on one client. Issues with more than one MGS set aside, the point here is that one of them is an Infiniband-cluster, the other is ethernet-based. And my client is on the ethernet. I have managed to mount the o2ib-fs by setting up an LNET router, but now this client's LNET doesn;t known how to reach the ethernet-fs. So the basic modprobe.conf reads options lnet networks=tcp1(eth0) routes=o2ib LNET-Router-IP@tcp1 This mounts the MGS on the o2ib network. What do I have to add to get to the MGS on the tpc network? Meanwhile I have studied more posts here and came up with options lnet networks=tcp1(eth0),tcp2(eth0:0) routes=o2ib LNET-Router-IP@tcp1; tcp Default-Gateway-IP@tcp2 Doesn't work either, but I see in the log of the (tcp-)MGS: LustreError: 120-3: Refusing connection from Client-IP for MGS-IP@tcp2: No matching NI Somethings getting through ... Any ideas? Regards, Thomas ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Mount 2 clusters, different networks - LNET tcp1-tcp2-o2ib
Hm, the ethernet FS is in tcp0 - MGS says its nids are MGS-IP@tcp. So not surprising it refuses that connection. On the other hand, options lnet networks=tcp1(eth0),tcp(eth0:0) routes=o2ib LNET-Router-IP@tcp1; tcp Default-Gateway-IP@tcp results in Can't create route to tcp via Gateway-IP@tcp Cheers, Thomas On 06/14/2011 07:00 PM, Michael Shuey wrote: Is your ethernet FS in tcp1, or tcp0? Your config bits indicate the client is in tcp1 - do the servers agree? -- Mike Shuey On Tue, Jun 14, 2011 at 12:23 PM, Thomas Roth t.r...@gsi.de wrote: Hi all, I'd like to mount two Lustre filesystems on one client. Issues with more than one MGS set aside, the point here is that one of them is an Infiniband-cluster, the other is ethernet-based. And my client is on the ethernet. I have managed to mount the o2ib-fs by setting up an LNET router, but now this client's LNET doesn;t known how to reach the ethernet-fs. So the basic modprobe.conf reads options lnet networks=tcp1(eth0) routes=o2ib LNET-Router-IP@tcp1 This mounts the MGS on the o2ib network. What do I have to add to get to the MGS on the tpc network? Meanwhile I have studied more posts here and came up with options lnet networks=tcp1(eth0),tcp2(eth0:0) routes=o2ib LNET-Router-IP@tcp1; tcp Default-Gateway-IP@tcp2 Doesn't work either, but I see in the log of the (tcp-)MGS: LustreError: 120-3: Refusing connection from Client-IP for MGS-IP@tcp2: No matching NI Somethings getting through ... Any ideas? Regards, Thomas ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Mount 2 clusters, different networks - LNET tcp1-tcp2-o2ib
That may be because your gateway doesn't have an interface on tcp (aka tcp0). I suspect you want to keep your ethernet clients in tcp0, your IB clients in o2ib0, and your router in both. Personally, I find it easiest to just give different module options on each system (rather than try ip2nets stuff). On the ether clients, I'd try: options lnet networks=tcp0(eth0) routes=o2ib0 LNET-router-eth_IP@tcp0 dead_router_check_interval=300 On IB clients: options lnet networks=o2ib0(ib0) routes=tcp0 LNET-router-IB_IP@ib0 dead_router_check_interval=300 then on the router: options lnet networks=tcp0(eth0),o2ib0(ib0) forwarding=enabled accept_timeout=15 Obviously, your file servers will need to have lnet options similar to the clients: options lnet networks=tcp0(eth0) routes=o2ib0 LNET-router-eth_IP@tcp0 dead_router_check_interval=300 options lnet networks=o2ib0(ib0) routes=tcp0 LNET-router-IB_IP@o2ib0 dead_router_check_interval=300 That's just a guess, your mileage may vary, etc., but I think it's close to what you want. Note that you really want the dead_router_check_interval if you're using lnet routers. Without that parameter, the lustre client will automatically mark a router as failed when it's unavailable but will not check to see if it ever comes back. With this param, it checks every 300 seconds (and re-enables it if found). Hope this helps. -- Mike Shuey On Tue, Jun 14, 2011 at 1:26 PM, Thomas Roth t.r...@gsi.de wrote: Hm, the ethernet FS is in tcp0 - MGS says its nids are MGS-IP@tcp. So not surprising it refuses that connection. On the other hand, options lnet networks=tcp1(eth0),tcp(eth0:0) routes=o2ib LNET-Router-IP@tcp1; tcp Default-Gateway-IP@tcp results in Can't create route to tcp via Gateway-IP@tcp Cheers, Thomas On 06/14/2011 07:00 PM, Michael Shuey wrote: Is your ethernet FS in tcp1, or tcp0? Your config bits indicate the client is in tcp1 - do the servers agree? -- Mike Shuey On Tue, Jun 14, 2011 at 12:23 PM, Thomas Roth t.r...@gsi.de wrote: Hi all, I'd like to mount two Lustre filesystems on one client. Issues with more than one MGS set aside, the point here is that one of them is an Infiniband-cluster, the other is ethernet-based. And my client is on the ethernet. I have managed to mount the o2ib-fs by setting up an LNET router, but now this client's LNET doesn;t known how to reach the ethernet-fs. So the basic modprobe.conf reads options lnet networks=tcp1(eth0) routes=o2ib LNET-Router-IP@tcp1 This mounts the MGS on the o2ib network. What do I have to add to get to the MGS on the tpc network? Meanwhile I have studied more posts here and came up with options lnet networks=tcp1(eth0),tcp2(eth0:0) routes=o2ib LNET-Router-IP@tcp1; tcp Default-Gateway-IP@tcp2 Doesn't work either, but I see in the log of the (tcp-)MGS: LustreError: 120-3: Refusing connection from Client-IP for MGS-IP@tcp2: No matching NI Somethings getting through ... Any ideas? Regards, Thomas ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Mount 2 clusters, different networks - LNET tcp1-tcp2-o2ib
Thanks, Michael. I'll certainly put in the check_interval, that will be needed. However, what I tried was to have an ethernet client that mounts one FS via the LNET router (Infiniband-FS behind it) and simultaneously mounts the other FS, which is on tcp0 - via its default route. So actually I don't have any IB clients (except for the LNET routers). Probably I messed up the tcpX-network names. Cheers, Thomas On 06/14/2011 08:04 PM, Michael Shuey wrote: That may be because your gateway doesn't have an interface on tcp (aka tcp0). I suspect you want to keep your ethernet clients in tcp0, your IB clients in o2ib0, and your router in both. Personally, I find it easiest to just give different module options on each system (rather than try ip2nets stuff). On the ether clients, I'd try: options lnet networks=tcp0(eth0) routes=o2ib0 LNET-router-eth_IP@tcp0 dead_router_check_interval=300 On IB clients: options lnet networks=o2ib0(ib0) routes=tcp0 LNET-router-IB_IP@ib0 dead_router_check_interval=300 then on the router: options lnet networks=tcp0(eth0),o2ib0(ib0) forwarding=enabled accept_timeout=15 Obviously, your file servers will need to have lnet options similar to the clients: options lnet networks=tcp0(eth0) routes=o2ib0 LNET-router-eth_IP@tcp0 dead_router_check_interval=300 options lnet networks=o2ib0(ib0) routes=tcp0 LNET-router-IB_IP@o2ib0 dead_router_check_interval=300 That's just a guess, your mileage may vary, etc., but I think it's close to what you want. Note that you really want the dead_router_check_interval if you're using lnet routers. Without that parameter, the lustre client will automatically mark a router as failed when it's unavailable but will not check to see if it ever comes back. With this param, it checks every 300 seconds (and re-enables it if found). Hope this helps. -- Mike Shuey On Tue, Jun 14, 2011 at 1:26 PM, Thomas Roth t.r...@gsi.de wrote: Hm, the ethernet FS is in tcp0 - MGS says its nids are MGS-IP@tcp. So not surprising it refuses that connection. On the other hand, options lnet networks=tcp1(eth0),tcp(eth0:0) routes=o2ib LNET-Router-IP@tcp1; tcp Default-Gateway-IP@tcp results in Can't create route to tcp via Gateway-IP@tcp Cheers, Thomas On 06/14/2011 07:00 PM, Michael Shuey wrote: Is your ethernet FS in tcp1, or tcp0? Your config bits indicate the client is in tcp1 - do the servers agree? -- Mike Shuey On Tue, Jun 14, 2011 at 12:23 PM, Thomas Roth t.r...@gsi.de wrote: Hi all, I'd like to mount two Lustre filesystems on one client. Issues with more than one MGS set aside, the point here is that one of them is an Infiniband-cluster, the other is ethernet-based. And my client is on the ethernet. I have managed to mount the o2ib-fs by setting up an LNET router, but now this client's LNET doesn;t known how to reach the ethernet-fs. So the basic modprobe.conf reads options lnet networks=tcp1(eth0) routes=o2ib LNET-Router-IP@tcp1 This mounts the MGS on the o2ib network. What do I have to add to get to the MGS on the tpc network? Meanwhile I have studied more posts here and came up with options lnet networks=tcp1(eth0),tcp2(eth0:0) routes=o2ib LNET-Router-IP@tcp1; tcp Default-Gateway-IP@tcp2 Doesn't work either, but I see in the log of the (tcp-)MGS: LustreError: 120-3: Refusing connection from Client-IP for MGS-IP@tcp2: No matching NI Somethings getting through ... Any ideas? Regards, Thomas ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss -- Thomas Roth Department: Informationstechnologie Location: SB3 1.262 Phone: +49-6159-71 1453 Fax: +49-6159-71 2986 GSI Helmholtzzentrum für Schwerionenforschung GmbH Planckstraße 1 64291 Darmstadt www.gsi.de Gesellschaft mit beschränkter Haftung Sitz der Gesellschaft: Darmstadt Handelsregister: Amtsgericht Darmstadt, HRB 1528 Geschäftsführung: Professor Dr. Dr. h.c. Horst Stöcker, Dr. Hartmut Eickhoff Vorsitzende des Aufsichtsrates: Dr. Beatrix Vierkorn-Rudolph Stellvertreter: Ministerialdirigent Dr. Rolf Bernhardt ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Mount 2 clusters, different networks - LNET tcp1-tcp2-o2ib - solved?
Hi all, this seems to work with the correct IPs and correct network names ;-[] I now have the following modprobe on my ethernet client: options lnet networks=tcp1(eth0),tcp0(eth0:0) routes=o2ib LNET-Router@tcp1; tcp Default-Route@tcp1 With these options, loading the modules gives me Jun 14 20:12:55 kernel: Lustre: Added LNI 10.12.70.183@tcp1 [8/256/0/180] Jun 14 20:12:55 kernel: Lustre: Added LNI 10.12.0.21@tcp [8/256/0/180] which are the IPs of eth0 and eth0:0. Now I still wonder why the alias interface eth0:0 is necessary (if left out, the whole endeavor fails). The routes=statement seems to say: If you have data for tcp, use the Default-Router-IP and go via the interace that is on network tcp1. Oh well, I should probably take some networking lectures... Regards, Thomas On 06/14/2011 06:23 PM, Thomas Roth wrote: Hi all, I'd like to mount two Lustre filesystems on one client. Issues with more than one MGS set aside, the point here is that one of them is an Infiniband-cluster, the other is ethernet-based. And my client is on the ethernet. I have managed to mount the o2ib-fs by setting up an LNET router, but now this client's LNET doesn;t known how to reach the ethernet-fs. So the basic modprobe.conf reads options lnet networks=tcp1(eth0) routes=o2ib LNET-Router-IP@tcp1 This mounts the MGS on the o2ib network. What do I have to add to get to the MGS on the tpc network? Meanwhile I have studied more posts here and came up with options lnet networks=tcp1(eth0),tcp2(eth0:0) routes=o2ib LNET-Router-IP@tcp1; tcp Default-Gateway-IP@tcp2 Doesn't work either, but I see in the log of the (tcp-)MGS: LustreError: 120-3: Refusing connection from Client-IP for MGS-IP@tcp2: No matching NI Somethings getting through ... Any ideas? Regards, Thomas ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
[Lustre-discuss] Enabling mds failover after filesystem creation
Greetings, I am attempting to add mds failover operation to an existing v1.8.4 filesystem. I have heartbeat/stonith configured on the mds nodes. What is unclear is what to change in the lustre parameters. I have read over the 1.8.x and 2.0 manuals and they are unclear as exactly how to enable failover mds operation on an existing filesystem. Do I simply run the following on the primary mds node and specify the NID of the secondary mds node? tunefs.lustre --param=failover.node=10.0.1.3@o2ib /dev/mdt device where: 10.0.1.2=primary mds, 10.0.1.3=secondary mds All of the examples for enabling failover via tunefs.lustre are for OSTs and I want to be sure that there isn't a different procedure for the MDS since it can only be active/passive. Thanks, --Jeff -- Jeff Johnson Aeon Computing www.aeoncomputing.com 4905 Morena Boulevard, Suite 1313 - San Diego, CA 92117 ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Enabling mds failover after filesystem creation
It depends - are you using a combined MGS/MDS? If so, you will have to update the mgsnid on all servers to reflect the failover node, plus change the client mount string to show the failover node. otherwise, it's the same procedure as with an OST. cliffw On Tue, Jun 14, 2011 at 12:06 PM, Jeff Johnson jeff.john...@aeoncomputing.com wrote: Greetings, I am attempting to add mds failover operation to an existing v1.8.4 filesystem. I have heartbeat/stonith configured on the mds nodes. What is unclear is what to change in the lustre parameters. I have read over the 1.8.x and 2.0 manuals and they are unclear as exactly how to enable failover mds operation on an existing filesystem. Do I simply run the following on the primary mds node and specify the NID of the secondary mds node? tunefs.lustre --param=failover.node=10.0.1.3@o2ib /dev/mdt device where: 10.0.1.2=primary mds, 10.0.1.3=secondary mds All of the examples for enabling failover via tunefs.lustre are for OSTs and I want to be sure that there isn't a different procedure for the MDS since it can only be active/passive. Thanks, --Jeff -- Jeff Johnson Aeon Computing www.aeoncomputing.com 4905 Morena Boulevard, Suite 1313 - San Diego, CA 92117 ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss -- cliffw Support Guy WhamCloud, Inc. www.whamcloud.com ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Enabling mds failover after filesystem creation
Apologies, I should have been more descriptive. I am running a dedicated MGS node and MGT device. The MDT is a standalone RAID-10 shared via SAS between two nodes, one being the current MDS and the second being the planned secondary MDS. Heartbeat and stonith w/ ipmi control is currently configured but not started between the two nodes. On 6/14/11 12:12 PM, Cliff White wrote: It depends - are you using a combined MGS/MDS? If so, you will have to update the mgsnid on all servers to reflect the failover node, plus change the client mount string to show the failover node. otherwise, it's the same procedure as with an OST. cliffw On Tue, Jun 14, 2011 at 12:06 PM, Jeff Johnson jeff.john...@aeoncomputing.com mailto:jeff.john...@aeoncomputing.com wrote: Greetings, I am attempting to add mds failover operation to an existing v1.8.4 filesystem. I have heartbeat/stonith configured on the mds nodes. What is unclear is what to change in the lustre parameters. I have read over the 1.8.x and 2.0 manuals and they are unclear as exactly how to enable failover mds operation on an existing filesystem. Do I simply run the following on the primary mds node and specify the NID of the secondary mds node? tunefs.lustre --param=failover.node=10.0.1.3@o2ib /dev/mdt device where: 10.0.1.2=primary mds, 10.0.1.3=secondary mds All of the examples for enabling failover via tunefs.lustre are for OSTs and I want to be sure that there isn't a different procedure for the MDS since it can only be active/passive. Thanks, --Jeff -- Jeff Johnson Aeon Computing www.aeoncomputing.com http://www.aeoncomputing.com 4905 Morena Boulevard, Suite 1313 - San Diego, CA 92117 ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org mailto:Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss -- cliffw Support Guy WhamCloud, Inc. www.whamcloud.com http://www.whamcloud.com -- Jeff Johnson Aeon Computing www.aeoncomputing.com 4905 Morena Boulevard, Suite 1313 - San Diego, CA 92117 ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Enabling mds failover after filesystem creation
Then it should be the same as the OST case. The only difference between the two is that we never allow two active MDSs on the same filesystem, so MDT is always active/passive. cliffw On Tue, Jun 14, 2011 at 12:18 PM, Jeff Johnson jeff.john...@aeoncomputing.com wrote: Apologies, I should have been more descriptive. I am running a dedicated MGS node and MGT device. The MDT is a standalone RAID-10 shared via SAS between two nodes, one being the current MDS and the second being the planned secondary MDS. Heartbeat and stonith w/ ipmi control is currently configured but not started between the two nodes. On 6/14/11 12:12 PM, Cliff White wrote: It depends - are you using a combined MGS/MDS? If so, you will have to update the mgsnid on all servers to reflect the failover node, plus change the client mount string to show the failover node. otherwise, it's the same procedure as with an OST. cliffw On Tue, Jun 14, 2011 at 12:06 PM, Jeff Johnson jeff.john...@aeoncomputing.com wrote: Greetings, I am attempting to add mds failover operation to an existing v1.8.4 filesystem. I have heartbeat/stonith configured on the mds nodes. What is unclear is what to change in the lustre parameters. I have read over the 1.8.x and 2.0 manuals and they are unclear as exactly how to enable failover mds operation on an existing filesystem. Do I simply run the following on the primary mds node and specify the NID of the secondary mds node? tunefs.lustre --param=failover.node=10.0.1.3@o2ib /dev/mdt device where: 10.0.1.2=primary mds, 10.0.1.3=secondary mds All of the examples for enabling failover via tunefs.lustre are for OSTs and I want to be sure that there isn't a different procedure for the MDS since it can only be active/passive. Thanks, --Jeff -- Jeff Johnson Aeon Computing www.aeoncomputing.com 4905 Morena Boulevard, Suite 1313 - San Diego, CA 92117 ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss -- cliffw Support Guy WhamCloud, Inc. www.whamcloud.com -- Jeff Johnson Aeon Computing www.aeoncomputing.com 4905 Morena Boulevard, Suite 1313 - San Diego, CA 92117 ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss -- cliffw Support Guy WhamCloud, Inc. www.whamcloud.com ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
Re: [Lustre-discuss] Where should SHARED_DIRECTORY of acc-sm cfg variable set to?
On 2011-06-14, at 9:57 AM, Surya, Prakash B. wrote: Perhaps it would be beneficial to add Andreas's comment to the source to avoid any future confusion? {code} # This is used by a small number of tests to share state between the client # running the tests, or in some cases between the servers (e.g. lfsck.sh). # It needs to be a non-lustre filesystem that is available on all the nod export SHARED_DIRECTORY=${SHARED_DIRECTORY:-/tmp} {/code} Sure, a patch would be quickly approved. From: lustre-discuss-boun...@lists.lustre.org [lustre-discuss-boun...@lists.lustre.org] On Behalf Of Andreas Dilger [adil...@whamcloud.com] Sent: Monday, June 13, 2011 9:24 PM To: Jay Lan Cc: lustre-discuss@lists.lustre.org Subject: Re: [Lustre-discuss] Where should SHARED_DIRECTORY of acc-sm cfg variable set to? On 2011-06-13, at 7:41 PM, Jay Lan wrote: To run some test suites of acc-sm, the SHARED_DIRECTORY variable needs to be set. Where should it be set to? Bugzilla 17839 Comment #65 said: the variable must contain the path of a shared filesystem. Does it need to be shared by what nodes? The clients? The clients + OST'es? Can it be $MOUNT? Or should it be a non-lustre filesystem that is mounted by NFS? Please advise. Thanks! This is used by a small number of tests to share state between the client running the tests, or in some cases between the servers (e.g. lfsck.sh). It needs to be a non-lustre filesystem that is available on all the nodes. Cheers, Andreas -- Andreas Dilger Principal Engineer Whamcloud, Inc. ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss Cheers, Andreas -- Andreas Dilger Principal Engineer Whamcloud, Inc. ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss
[Lustre-discuss] lustre from source EXTRAVERSION
I'm trying to rebuild the RHEL kernel with the lustre patches, most everything has gone okay one time through so far, but i ran into an issue when trying to get OFED to compile against the new code According to the Whamcloud wiki I am to perform this step Add a unique build id so we can be certain our kernel is booted. Edit ~build/kernel/rpmbuild/BUILD/kernel-2.6.18/linux-2.6.18.x86_64/Makefile and modify line 4, the EXTRAVERSION to read: EXTRAVERSION = -lustre18 And then further down, this one Go into the kernel source directory and issue the following commands to build a kernel rpm. cd /build/kernel/rpmbuild/BUILD/kernel-2.6.18/linux-2.6.18.x86_64 make oldconfig || make menuconfig make include/asm make include/linux/version.h make SUBDIRS=scripts make include/linux/utsrelease.h make rpm Two questions: 1. No matter what i do, i cannot get the kernel version to come out as 2.6.18-238.el5.lustre185 I can't seem to get EXTRAVERSION to take dashes, I end up with 2.6.18238.el5.lustre185 or given the example above 2.6.18lustre18, is there a fix for this? 2. The resulting rpm does not include an initrd file. I can certainly create one on my own, but is there a parameter or step that will allow 'make rpm' to do that for me and include it in the rpm? thanks ___ Lustre-discuss mailing list Lustre-discuss@lists.lustre.org http://lists.lustre.org/mailman/listinfo/lustre-discuss