On Fri, Dec 11, 2009 at 10:58:19AM -0800, Mark Howison wrote:
> The purple indicates ftruncate, which is only called from task 0 (that
> is how MPI_File_set_size is implemented). The salmon at the far right
> actually indicates fsyncs, not reads. And the brown (hard to see) is
> fclose.
Just to ad
Cc:
Bcc:
Subject: [emoly@sun.com: some fixes for lustre adio driver]
Reply-To:
Here is LiuYing's message from last month. If all looks OK, once I've
merged the two patch sets, all the known bugs will be fixed.
Thanks for the work!
==rob
- Forwarded message from "emoly.liu" -
S
On Mon, Jan 04, 2010 at 11:09:28AM -0700, Martin Pokorny wrote:
> Hello,
>
> I took a look at the ADIO Lustre driver last week in an attempt to
> make collective I/O work for my application. I believe that I've
> gotten the driver to do the right thing now. With the following
> patch all of the te
On 2010-01-04, at 03:02, David Cohen wrote:
> I'm using a mixed environment of 1.8.0.1 MDS and 1.6.6 OSS's (had a
> problem
> with qlogic drivers and rolled back to 1.6.6).
> My MDS get unresponsive each day at 4-5 am local time, no kernel
> panic or
> error messages before.
Judging by the tim
Hello,
I took a look at the ADIO Lustre driver last week in an attempt to make
collective I/O work for my application. I believe that I've gotten the
driver to do the right thing now. With the following patch all of the
tests under romio/test pass in my environment. My application, which
requ
Aaron Knister wrote:
> Just wondering if anybody is successfully NFS exporting their lustre
> filesystem. I'm also curious to hear about experiences doing this. I
> tried this back with some of the early 1.6 releases and ended up with
> the lustre nfs exports freezing after a few minutes of heavy i
Heiko Schröter wrote:
> Am Mittwoch 23 Dezember 2009 12:22:17 schrieb Christopher J. Walker:
>>> Dec 22 17:18:49 proof kernelLustreError: 10917:0:
>>> (ldlm_request.c:1030:ldlm_cli_cancel_req()) Got rc -108 from cancel RPC:
>>> canceling anyway
>>> Dec 22 17:18:49 proof kernelLustreError: 10917:0:
Under lustre 1.6.4.3 we have to use the single threaded user space nfsd in
order to not lock up our OSSs. As you may expect, the performance is not very
good. The kernel nfsd works fine for us so far in our 1.8 testing.
-Original Message-
From: lustre-discuss-boun...@lists.lustre.org
Hi,
I'm using a mixed environment of 1.8.0.1 MDS and 1.6.6 OSS's (had a problem
with qlogic drivers and rolled back to 1.6.6).
My MDS get unresponsive each day at 4-5 am local time, no kernel panic or
error messages before.
Some errors and an LBUG appear in the log after force booting the MDS and