Re: [Lustre-discuss] Quick writeback cache question

2009-03-04 Thread Andreas Dilger
On Mar 04, 2009 09:22 +, Daire Byrne wrote: - Andreas Dilger adil...@sun.com wrote: Secondly, with small files there is also the overhead of creating the many small files on the MDS, which will likely take as long as the RPCs being sent to the OSTs with data. There is not

Re: [Lustre-discuss] How to start

2009-03-04 Thread Daire Byrne
Rayentray, - Rayentray Tappa rtappa.ascen...@gmail.com wrote: Hi List, I'm working for a small government organization in South America, and in our quest for a large distributed storage solution we ran across Lustre. We have some servers in which we're trying to build a prototype to

Re: [Lustre-discuss] How to start

2009-03-04 Thread Rayentray Tappa
On Wed, 2009-03-04 at 11:09 +, Daire Byrne wrote: Rayentray, - Rayentray Tappa rtappa.ascen...@gmail.com wrote: Hi List, I'm working for a small government organization in South America, and in our quest for a large distributed storage solution we ran across Lustre. We have

Re: [Lustre-discuss] How to start

2009-03-04 Thread Daire Byrne
Rayentray, - Rayentray Tappa rtappa.ascen...@gmail.com wrote: I'm testing this in order to build a large system later, where high availability of the data store is a must - so we do want to check how this works, what involves to configure it and what may happen in case of failure and how

[Lustre-discuss] unable to mount ost

2009-03-04 Thread Rayentray Tappa
Hi, I created an OST today by running: mkfs.lustre --fsname=testfs --ost --mgsnode=192.168.6@tcp /dev/sda3 (ltcl list_nids in the mdt/mgs gave me the 192.168.6@tcp hint) When I want to mount it, I get: mount -t lustre /dev/sda3 /mnt/test/ost0/mount.lustre: mount /dev/sda3 at

Re: [Lustre-discuss] How to read/understand a Lustre error

2009-03-04 Thread Brian J. Murrell
On Wed, 2009-03-04 at 12:11 -0500, Ms. Megan Larko wrote: Greetings, I have a Lustre OSS with eleven (0-11) OSTs. Every once in a while the OSS hosting the OSTs fails with a kernel panic. To be clear, what you are reporting is not a kernel panic. It is a watchdog timeout. Kernel panics

Re: [Lustre-discuss] e2scan MDT back-up

2009-03-04 Thread Daire Byrne
Brian, - Brian J. Murrell brian.murr...@sun.com wrote: On Wed, 2009-03-04 at 17:03 +, Daire Byrne wrote: Once you move to LVM then you can snapshot the MDT and mount that without unmounting the production MDT. You should probably destroy the snapshot afterwards as it affects

Re: [Lustre-discuss] e2scan MDT back-up

2009-03-04 Thread Brian J. Murrell
On Wed, 2009-03-04 at 17:42 +, Daire Byrne wrote: Brian, Hi Daire, I looked at zumastor. Although it didn't get any slower with extra snapshots it took quite a hit (~ x6 drop in writes) with just a single snapshot. Ouch! Far more promising is the port of zumastor exceptions to the

Re: [Lustre-discuss] Problem replacing an OST in 1.6.7

2009-03-04 Thread Nirmal Seenu
Andreas Dilger wrote: On Mar 03, 2009 17:15 -0600, Nirmal Seenu wrote: mkfs.lustre --fsname=lqcdproj --ost --mgsnode=iblust...@tcp1 --mkfsoptions=-m 0 --index= --reformat /dev/md2 I received these error messages when I tried to mount it for the first time: Mar 3 16:19:53 lustre1

[Lustre-discuss] e2scan for cleaning scratch space

2009-03-04 Thread Brock Palen
e2scan will show me all the files that have changed from a date, but I want to know all the files that have not changed sense some date. The goal is to make a system for purging scratch spaces that is fast, and minimum wear on the filesystem. How are groups doing this now? Are you using

Re: [Lustre-discuss] e2scan for cleaning scratch space

2009-03-04 Thread Brock Palen
The e2scan shipped from sun's rpms does not support sqlite3 out of the box: rpm -qf /usr/sbin/e2scan e2fsprogs-1.40.7.sun3-0redhat e2scan: sqlite3 was not detected on configure, database creation is not supported Should I just rebuilt only e2scan? Brock Palen www.umich.edu/~brockp Center

[Lustre-discuss] read performance on ia64

2009-03-04 Thread Michel Dionne
Hello all I am running test between an X4150 and SGI Altix with luster 1.6.6 and I can see a read performance drop on the SGI. Numbers: X4150 dd if=/dev/zero of=/mnt/exec4_client_ib/mike/4/bigfile bs=4096k count=1 (336MB/s) this a folder with a stripe of 4 for all tests dd

Re: [Lustre-discuss] read performance on ia64

2009-03-04 Thread Kevin Van Maren
What page size are you running on IA64? What networking? Kevin Michel Dionne wrote: Hello all I am running test between an X4150 and SGI Altix with luster 1.6.6 and I can see a read performance drop on the SGI. Numbers: X4150 dd if=/dev/zero of=/mnt/exec4_client_ib/mike/4/bigfile

Re: [Lustre-discuss] Job fails opening 24k files and keeps them open during execution

2009-03-04 Thread Brian J. Murrell
On Wed, 2009-03-04 at 16:27 -0500, Osvaldo Rentas wrote: Hello, I am working with a user that has Fortran code that opens 24.000 files and keeps them open during execution. We had to adjust our kernel parameters to allow this to happen, since Linux cuts you off at 1024 by default.

Re: [Lustre-discuss] Job fails opening 24k files and keeps them open during execution

2009-03-04 Thread Wang Yibin
Lustre does not impose maximum number of open files, but practically it depends on amount of RAM on the MDS. There are no tables for open files on the MDS, as they are only linked in a list to a given client's export. Each client process probably has a limit of several thousands of open files