Re: [Lustre-discuss] abnormally long ftruncates on Cray XT4

2010-01-04 Thread Rob Latham
On Fri, Dec 11, 2009 at 10:58:19AM -0800, Mark Howison wrote: > The purple indicates ftruncate, which is only called from task 0 (that > is how MPI_File_set_size is implemented). The salmon at the far right > actually indicates fsyncs, not reads. And the brown (hard to see) is > fclose. Just to ad

[Lustre-discuss] updated lustre adio driver from LiuYing

2010-01-04 Thread Rob Latham
Cc: Bcc: Subject: [emoly@sun.com: some fixes for lustre adio driver] Reply-To: Here is LiuYing's message from last month. If all looks OK, once I've merged the two patch sets, all the known bugs will be fixed. Thanks for the work! ==rob - Forwarded message from "emoly.liu" - S

Re: [Lustre-discuss] [mpich-discuss] adio lustre patch

2010-01-04 Thread Rob Latham
On Mon, Jan 04, 2010 at 11:09:28AM -0700, Martin Pokorny wrote: > Hello, > > I took a look at the ADIO Lustre driver last week in an attempt to > make collective I/O work for my application. I believe that I've > gotten the driver to do the right thing now. With the following > patch all of the te

Re: [Lustre-discuss] MDS crashes daily at the same hour

2010-01-04 Thread Andreas Dilger
On 2010-01-04, at 03:02, David Cohen wrote: > I'm using a mixed environment of 1.8.0.1 MDS and 1.6.6 OSS's (had a > problem > with qlogic drivers and rolled back to 1.6.6). > My MDS get unresponsive each day at 4-5 am local time, no kernel > panic or > error messages before. Judging by the tim

[Lustre-discuss] adio lustre patch

2010-01-04 Thread Martin Pokorny
Hello, I took a look at the ADIO Lustre driver last week in an attempt to make collective I/O work for my application. I believe that I've gotten the driver to do the right thing now. With the following patch all of the tests under romio/test pass in my environment. My application, which requ

Re: [Lustre-discuss] NFS re-exporting lustre

2010-01-04 Thread Frederik Ferner
Aaron Knister wrote: > Just wondering if anybody is successfully NFS exporting their lustre > filesystem. I'm also curious to hear about experiences doing this. I > tried this back with some of the early 1.6 releases and ended up with > the lustre nfs exports freezing after a few minutes of heavy i

Re: [Lustre-discuss] ldlm_cli_cancel_req()) Got rc -108 from cancel RPC: canceling anyway

2010-01-04 Thread Christopher J.Walker
Heiko Schröter wrote: > Am Mittwoch 23 Dezember 2009 12:22:17 schrieb Christopher J. Walker: >>> Dec 22 17:18:49 proof kernelLustreError: 10917:0: >>> (ldlm_request.c:1030:ldlm_cli_cancel_req()) Got rc -108 from cancel RPC: >>> canceling anyway >>> Dec 22 17:18:49 proof kernelLustreError: 10917:0:

Re: [Lustre-discuss] NFS re-exporting lustre

2010-01-04 Thread Todd, Allen
Under lustre 1.6.4.3 we have to use the single threaded user space nfsd in order to not lock up our OSSs. As you may expect, the performance is not very good. The kernel nfsd works fine for us so far in our 1.8 testing. -Original Message- From: lustre-discuss-boun...@lists.lustre.org

[Lustre-discuss] MDS crashes daily at the same hour

2010-01-04 Thread David Cohen
Hi, I'm using a mixed environment of 1.8.0.1 MDS and 1.6.6 OSS's (had a problem with qlogic drivers and rolled back to 1.6.6). My MDS get unresponsive each day at 4-5 am local time, no kernel panic or error messages before. Some errors and an LBUG appear in the log after force booting the MDS and