[Lustre-discuss] MDS question

2008-08-14 Thread Mag Gam
What is the disadvantage of creating a MDS partition with smaller
inodes per block? Now, its 4k per inode what happens if we go to the
least blocks which is 1024k?  This would let us create more smaller
files which will lead to more inodes used. But what is the downside?

TIA
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] Bug 15912

2008-08-14 Thread Brock Palen
I see it is fixed now as we were being bit by this.
We would like to put the new filesystem into use on Monday thus we  
are trying to get this resolved.

Question, because it is just a parsing problem in mkfs, can after the  
filesystem is created can the problem be corrected?

If not, how can we work around this?  Do I just need to build the  
mkfs out of CVS for 1.6.6 ?


Brock Palen
www.umich.edu/~brockp
Center for Advanced Computing
[EMAIL PROTECTED]
(734)936-1985



___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] drbd async mode

2008-08-14 Thread Andrea Frittoli
I'm trying a Lustre + DRBD set-up, I would like to use DRBD with the
following setup:

on MDS: DRBD emulates an MDT on shared storage. From Lustre
perspective one MDS is active, the other is passive
on OSS: BRBD emulates two OST on shared storage for each OSS couple.
From Lustre perspective each OSS is active for one OST and passive for
the other.

I'm planning to use DRBD in active/passive mode, so that the storage
is mounted only on the active server, and luster is mounted on top.
I'm not sure how to sync the mount/umount of the DRBD partition and
the mount of MDT/OST upon failover.

The target setup shall have 2 MDS (1 couple) and 8 OST (4 couples),
dedicate LANs for heartbeat and drbd.

Do you have any details of your setup you may share about this?

TIA!


On Aug 12, 4:50 am, Mag Gam [EMAIL PROTECTED] wrote:
 Ha yes..I seen this before Sebastein. :-)

 I wanted to see if other people have tried it and their thoughts were.

 thats all
 TIA

 On Mon, Aug 11, 2008 at 2:33 AM, Sébastien Buisson

 [EMAIL PROTECTED] wrote:
  Hello,

  I ran some tests with DRBD and Lustre a few months ago (including in async
  mode). You can find my modest findings here on the Lustre wiki :
 http://wiki.lustre.org/index.php?title=Drbd_And_Lustre

  Cheers,
  Sebastien.

  Mag Gam a écrit :

  Has anyone been using DRBD and Lustre in async mode? I was curious
  what type of speeds you were getting and also what network are you
  using?

  TIA
  ___
  Lustre-discuss mailing list
  [EMAIL PROTECTED]
 http://lists.lustre.org/mailman/listinfo/lustre-discuss

 ___
 Lustre-discuss mailing list
 [EMAIL PROTECTED]://lists.lustre.org/mailman/listinfo/lustre-discuss
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] Log file opens/reads/etc?

2008-08-14 Thread daledude
Is there is a tool that shows what files are being accessed? Sort of
like inotify, but not inotify? I'd like to compile file access
statistics to try and balance the most accessed files across the OST's
better.

Thanks for any tips,
Dale
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] Lustre 1.6.5.1 + RHEL4.7 + Infiniband

2008-08-14 Thread Götz Waschk
Hi everyone,

I'm trying to build lustre 1.6.5.1 against Red Hat's kernel
2.6.9-78.EL for RHEL4, but the configure script doesn't detect the
headers needed for OFED gen2 support (Infiniband). Does anybody know a
fix for that problem?

Regards, Götz Waschk
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] lustre + debian

2008-08-14 Thread rafalak
Hello, I want to use lustre with my debian (etch or lenny)
I try to compile lustre as module with debian kernel and I've got
error

How to run lustre with debian ?
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] diskless booting Debian

2008-08-14 Thread Daire Byrne
Jeff/Troy,

- Jeff Darcy [EMAIL PROTECTED] wrote:

 We've tried loading canned MDT/OST images into memory on some nodes
 and serving from there, and it does seem to work.  There are two
 downsides, though.  One is that the Linux loopback driver is a real
 performance bottleneck, ever since some bright person had the idea to
 make it less multi-threaded than it had been.  Another is that
 booting tends to involve metadata-heavy access patterns which are not exactly
 Lustre's strength - a situation made worse when you have nearly a
 thousand clients doing it at the same time and your MDS is a
 relatively small node like the others.  So far we've found that NBD serves us
 better in the boot/root filesystem role, though that means a
 read-only root which involves its own complexity.  Your mileage will almost
 certainly vary.

A good trick with Lustre to get around the metadata bottleneck is to use disk 
image files on Lustre (e.g. SquashFS) and mount them using the loopback driver 
on each compute node (or lctl attach_device ?). So instead of having to 
bother the MDS you need only seek through the file on an OST. By either 
striping the read-only image across all your OSTs or having a round-robin image 
per OST you can get pretty good scalability.

I tried this with our 700 node compute cluster but to be honest the overall 
booting performance was not that different to a couple of NFS servers serving a 
read-only root so it was not really worth the extra complexity in the end.

We do still use SquashFS on Lustre from time to time when we have a directory 
tree with 30,000 small files in it that needs to be read by every farm machine. 
It's rare but it does happen and traditionally NFS does much better with such 
workloads.

Daire
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] Re-using OST and MDT names

2008-08-14 Thread Ms. Megan Larko
Hello,

As a part of my continuing to benchmark Lustre to ascertain was may be
best-suited for our needs here, I have re-created at the LSI ELP
card level some of my arrays from my earlier benchmark posts.  The
card is now sending /dev/sdf 998999Mb with 128kB stripesize and
/dev/sdg 6992995 Mb with 128kB stripesize to my OSS.  The sdf and sdg
formatted fine with Lustre and mounted without issue on the OSS.
Recycling the MGS MDT's seem to have been a problem.  When I tried to
mount the MDT on the MGS after mounting the new OST's the mounts
performed without error, but the bonnie benchmark test as run before
would hang every time.

Sample of errors in MGS file /var/log/messages:
Aug 13 12:39:30 mds1 kernel: Lustre: crew5-OST0001-osc: Connection to
service crew5-OST000
1 via nid [EMAIL PROTECTED] was lost; in progress operations using this
service will wait for recovery to complete.
  Aug 13 12:39:30 mds1 kernel:
LustreError: 167-0: This client was evicted by crew5-OST0001;
 in progress operations using this service will fail.
   Aug 13 12:39:30 mds1 kernel: Lustre:
crew5-OST0001-osc: Connection restored to service crew5-OST0001 using
nid [EMAIL PROTECTED]
Aug 13 12:39:30 mds1 kernel: Lustre: MDS crew5-MDT:
crew5-OST0001_UUID now active, resetting orphans
 Aug 13 12:42:42
mds1 kernel: Lustre: 3406:0:(ldlm_lib.c:519:target_handle_reconnect())
cre
w5-MDT: 50b043bb-0e8c-7a5b-b0fe-6bdb67d21e0b reconnecting
   Aug 13 12:42:42 mds1 kernel: Lustre:
3406:0:(ldlm_lib.c:519:target_handle_reconnect()) Skipped 24 previous
similar messages
Aug 13 12:42:42 mds1 kernel: Lustre:
3406:0:(ldlm_lib.c:747:target_handle_connect()) crew5-MDT: refuse
reconnection from
[EMAIL PROTECTED]@o2ib to
0x81006994d000; still busy with 2 active RPCs
Aug 13 12:42:42 mds1 kernel: Lustre:
3406:0:(ldlm_lib.c:747:target_handle_connect()) Skipped 24 previous
similar messages
Aug 13 12:42:42 mds1 kernel: LustreError:
3406:0:(ldlm_lib.c:1442:target_send_reply_msg())
 @@@ processing error (-16)  [EMAIL PROTECTED] x600107/t0
o38-[EMAIL PROTECTED]:-1
lens 304/200 ref 0 fl Interpret:/0/0 rc -16/0Aug 13 12:42:42 mds1
kernel: LustreError: 3406:0:(ldlm_lib.c:1442:target_send_reply_msg())
 Skipped 24 previous similar messages
   Aug 13 12:43:40 mds1 kernel: LustreError: 11-0: an
error occurred while communicating with [EMAIL PROTECTED] The
ost_connect operation failed with -19
Aug 13 12:43:40 mds1 kernel: LustreError: Skipped 7 previous similar
messages Aug 13 12:47:50 mds1 kernel: Lustre:
crew5-OST0001-osc: Connection to service crew5-OST0001 via nid
[EMAIL PROTECTED] was lost; in progress operations using this service
will wait f
or recovery to complete.

Sample of errors on OSS file /var/log/messages:
Aug 13 12:39:30 oss4 kernel: Lustre: crew5-OST0001: received MDS
connection from [EMAIL PROTECTED]
Aug 13 12:43:57 oss4 kernel: Lustre: crew5-OST0001: haven't heard from
client crew5-mdtlov_UUID (at [EMAIL PROTECTED]) in 267 seconds. I think
it's dead, and I am evicting it.
Aug 13 12:46:27 oss4 kernel: LustreError: 137-5: UUID
'crew5-OST_UUID' is not available  for connect (no target)
Aug 13 12:46:27 oss4 kernel: LustreError: Skipped 51 previous similar messages
Aug 13 12:46:27 oss4 kernel: LustreError:
4151:0:(ldlm_lib.c:1442:target_send_reply_msg()) @@@ processing error
(-19)  [EMAIL PROTECTED] x600171/t0 o8-?@?:-1 lens 240/0 ref 0
fl Interpret:/0/0 rc -19/0
Aug 13 12:46:27 oss4 kernel: LustreError:
4151:0:(ldlm_lib.c:1442:target_send_reply_msg()) Skipped 52 previous
similar messages
Aug 13 12:47:50 oss4 kernel: Lustre: crew5-OST0001: received MDS
connection from [EMAIL PROTECTED]

In lctl all pings were successful.  Additionally files on Lustre disks
on our live system using the same MGS were all fine; no errors in
logfile.

I thought that maybe changing the disk kB size and reformatting the
OST without reformatting the MDT was a problem.  So I unmounted OST
and MDT and reformatted the MDT on the MGS.  All okay.   The OST
remount without error.The MDT on the MGS will not remount:

[EMAIL PROTECTED] ~]# mount -t lustre /dev/sdg1 /srv/lustre/mds/crew5-MDT
mount.lustre: mount /dev/sdg1 at /srv/lustre/mds/crew5-MDT failed:
Address already in use
The target service's index is already in use. (/dev/sdg1)

Again, the live systems on the MGS are still fine.   A web search for
the error suggested I try  tunefs.lustre --reformat --index=0
--writeconf=/dev/sdg1 but I was unable to get a syntax of that
command that would run for me (I tried various index= and adding
/dev/sdg1 at the end of the line but it failed each time and only
reprinted the help without indicating what about what I typed was
unparseable).

My current thought is that a stripesize of 128kB from the LSI ELP
card is not testable on my Lustre 1.6.4 system.  This does not seem to
be an accurate statement from 

Re: [Lustre-discuss] lustre + debian

2008-08-14 Thread Patrick Winnertz
Hello,

On Thursday 14 August 2008 18:57:52 Robert LeBlanc wrote:
 It's easy and then it's not. The Debian way is to download the Lustre
 kernel packages and the Lustre binary packages. You build a kernel using
 make-kpkg the --with-patches Lustre option, we found a bug when trying to
 use the --append_to_kernel option so unless that got fixed, don't use it,
Thanks for this informations, I'll have a look on it. 

 them up. You are now ready to use Lustre Debian style. If deploying this on
 a cluster, just install your ready made .debs for the kernel and modules.
 None of this was explained in the README in /usr/share/doc/Lustre so we had
 to figure it out ourselves. Hopefully the Debian folks have fixed that.
Could you please have a look on our Readme.Debian. I've attached the current 
version of this README.



 Do not use a vanilla kernel from kernel.org, only use the kernel source
 from the Debian repository as the Lustre kernel packages are tested against
 them.
Vanilla kernels should be supported.. We've only packaged the upstream patches 
and modified them a bit for the debian kernels. So vanilla kernels should work 
as usual. 

Greetings
Patrick Winnertz
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] mounting an OST from another node attached to a fibre channel switch

2008-08-14 Thread Ron
Hi,
We have set up a couple of test systems where the OSTs are mounted on
the same node as the MDT.   The OSTs are LUNS on a SATA Beast
controller accessible from multiple systems attached to a fibre
channel switch,. We would like to umount an OST from the MDT system
and mount it on another system. We've tried doing this and even though
there is network traffic between the new system and the mds system,
the mds system seems to be ignoring the OST mount.  Can the an OSTs
OSS change?
Thanks,
Ron
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] lustre + debian

2008-08-14 Thread Robert LeBlanc
On 8/14/08 12:51 PM, Patrick Winnertz [EMAIL PROTECTED]
wrote:

 Hello,
 
 On Thursday 14 August 2008 18:57:52 Robert LeBlanc wrote:
 It's easy and then it's not. The Debian way is to download the Lustre
 kernel packages and the Lustre binary packages. You build a kernel using
 make-kpkg the --with-patches Lustre option, we found a bug when trying to
 use the --append_to_kernel option so unless that got fixed, don't use it,
 Thanks for this informations, I'll have a look on it.
It was that the build script was very strict on what the kernel was named to
determine what patches to apply, this was an issue with the script from
Lustre. It's been almost a year since I've built Lustre, so this may have
changed.
 
 them up. You are now ready to use Lustre Debian style. If deploying this on
 a cluster, just install your ready made .debs for the kernel and modules.
 None of this was explained in the README in /usr/share/doc/Lustre so we had
 to figure it out ourselves. Hopefully the Debian folks have fixed that.
 Could you please have a look on our Readme.Debian. I've attached the current
 version of this README.
The attachment did not make it, I'm downloading the current Lenny Lustre
packages to view the file.
 
 Do not use a vanilla kernel from kernel.org, only use the kernel source
 from the Debian repository as the Lustre kernel packages are tested against
 them.
 Vanilla kernels should be supported.. We've only packaged the upstream patches
 and modified them a bit for the debian kernels. So vanilla kernels should work
 as usual.
Yes, all the vanilla patches that Lustre have are in the Debian package, the
problem is that Lustre is usually behind Lenny so grabbing the newest kernel
from kernel.org fails no matter what distro. What I've found about Debian is
that the package team has worked hard to patch Lustre for the current kernel
supported by Debian, this is usually a version or two ahead of Lustre. That
is why I said not to use a vanilla kernel, your chance of an error free
build is greater.

-- 
Robert LeBlanc
Life Sciences Computer Support
Brigham Young University
[EMAIL PROTECTED]
(801)422-1882

___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] mounting an OST from another node attached to a fibre channel switch

2008-08-14 Thread Klaus Steden

Yes. There is an entry in the manual on this topic. You'll have to stop
Lustre and update the MGS configuration, but it's a pretty quick operation.

cheers,
Klaus

On 8/14/08 2:29 PM, Ron [EMAIL PROTECTED]did etch on stone tablets:

 Hi,
 We have set up a couple of test systems where the OSTs are mounted on
 the same node as the MDT.   The OSTs are LUNS on a SATA Beast
 controller accessible from multiple systems attached to a fibre
 channel switch,. We would like to umount an OST from the MDT system
 and mount it on another system. We've tried doing this and even though
 there is network traffic between the new system and the mds system,
 the mds system seems to be ignoring the OST mount.  Can the an OSTs
 OSS change?
 Thanks,
 Ron
 ___
 Lustre-discuss mailing list
 Lustre-discuss@lists.lustre.org
 http://lists.lustre.org/mailman/listinfo/lustre-discuss

___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] Startup Sequence 1.6.5.1

2008-08-14 Thread Mike Feuerstein
Is there a recommended startup mounting sequence for 1.6.5.1?

We are internally debating 

MGS lun - MDS luns - OSTs 

vs. 

MGS lun - OSTs - MDS luns

 

Mike 

 

___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss