Re: [lustre-discuss] Lustre 2.10.7 released

2019-03-26 Thread Götz Waschk
On Mon, Mar 25, 2019 at 10:53 PM Peter Jones  wrote:
> We are pleased to announce that the Lustre 2.10.7 Release has been declared 
> GA and is available for download. You can also grab the source from git.
Hi Peter,

is e2fsprogs v1.42.13.wc6 still the recommended version or are newer
versions fine as well?

Regards, Götz Waschk
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org


[lustre-discuss] Crash of a MDS

2019-03-13 Thread Götz Waschk
Hi everyone,

I'm running a Lustre 2.10.6 file system with two MDS. One of them has
just crashed, I have received these messages:

Message from syslogd@miscmds4 at Mar 13 09:01:24 ...
 kernel:[5584275.388034] LustreError:
321079:0:(lod_object.c:1584:lod_parse_dir_striping()) ASSERTION(
lo->ldo_stripe == ((void *)0) ) failed:

Message from syslogd@miscmds4 at Mar 13 09:01:24 ...
 kernel:[5584275.394146] LustreError:
321079:0:(lod_object.c:1584:lod_parse_dir_striping()) LBUG

Is there a known problem that could cause this?

Regards,
Götz Waschk
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org


[lustre-discuss] Lustre 2.10.0 ZFS version

2017-07-17 Thread Götz Waschk
Hi everyone,

which version of kmod-zfs was the official Lustre 2.10.0 binary
release for CentOS 7.3 built against?

Regards, Götz Waschk
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org


Re: [lustre-discuss] Can't get zfs OST to mount on a freshly built server

2015-11-05 Thread Götz Waschk
On Wed, May 13, 2015 at 4:16 PM, Bob Ball <b...@umich.edu> wrote:
> OK, so, I am seeing EXACTLY the issue reported at the end of LU-6452 8
> minutes after it was closed by Andreas Dilger.
> https://jira.hpdd.intel.com/browse/LU-6452
>
> There is no response.  Is there a solution?

Hi Bob,

I guess my answer is rather for the archive. I had the same problem. I
guess this information is for paying customers only and not found in
the release notes or on Jira.

> This is Lustre 2.7.0 with (now) zfs 0.6.4.1-1, which was current when the
> server was built.  I see a number of recent Emails about updates to Lustre
> sources against this zfs version, but is there a solution for the standard
> set of 2.7.0 Lustre rpms?  Or any solution that will get me un-stuck?


Your version of zfs is too new. The binary rpms of lustre 2.7.0 are
compatible with spl/zfs 0.6.3-1.2 only:
# modinfo zfs|grep version
version:0.6.3-1.2
srcversion: 9888DC55B2F55794F6B5D44
# modinfo osd_zfs|grep version
srcversion: D9F832C8E8804F08B693280
vermagic:   2.6.32-504.8.1.el6_lustre.x86_64 SMP mod_unload modversions


Regards, Götz Waschk
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org


Re: [lustre-discuss] Inodes Quota problem

2015-10-09 Thread Götz Waschk
On Thu, Oct 8, 2015 at 4:45 PM, Mohr Jr, Richard Frank (Rick Mohr)
 wrote:
> Not sure if this is a known bug.  Have you tried regenerating the quota data? 
>  I have had to do this a few times on some OSTs.

Hi,

I haven't tried it yet, I will at the next downtime.

Regards, Götz
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org


[lustre-discuss] Inodes Quota problem

2015-10-08 Thread Götz Waschk
Hi everyone,

I have a quota problem on a file system upgraded from Lustre 1.8 to
2.5.3. The number of inodes reported by lfs getquota is obviously
wrong:

# ls -l
insgesamt 3064024
drwxr-sr-x 3 u248 icecube1830912 28. Aug 2014  Callisto
drwxr-sr-x 2 u248 icecube2301952 26. Aug 2014  CameraOutput
-rw-r--r-- 1 u248 icecube  21542  1. Aug 2014  LogCallistoM1_026202_.log
-rw-r--r-- 1 u248 icecube  439316480 15. Aug 2014  logs25000_35000.tar.gz
-rw-r--r-- 1 u248 icecube  224972800 18. Aug 2014  logs35000_4.tar.gz
-rw-r--r-- 1 u248 icecube  225146880 28. Aug 2014  logscal40k-45k.tar.gz
-rw-r--r-- 1 u248 icecube 1122981634 28. Jul 2014
Pr_za05to36_4_027617_ct1_w0.rfl
-rw-r--r-- 1 u248 icecube 1120948234 28. Jul 2014
Pr_za05to36_4_027617_ct2_w0.rfl
# lfs quota -u u248 .
Disk quotas for user u248 (uid 20056):
 Filesystem  kbytes   quota   limit   grace   files   quota   limit   grace
  . 4709669648   0   0   -   1   0
  0   -


Is this a known bug?

Regards, Götz
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org


Re: [lustre-discuss] free space on ldiskfs vs. zfs

2015-08-25 Thread Götz Waschk
Dear All,

I'm sorry, I cannot provide verbose zpool information anymore. I was a
bit in a hurry to put the file system into production and that's why I
have reformatted the servers with ldiskfs.

On Tue, Aug 25, 2015 at 5:54 AM, Alexander I Kulyavtsev a...@fnal.gov wrote:
 I was assuming the question was about total space as I struggled for some 
 time to understand  why do I have 99 TB total available space per OSS, after 
 installing zfs lustre, while ldiskfs OSTs have 120 TB on the same hardware. 
 The 20% difference was partially (10%) accounted by different raid6 / raidz2 
 configuration. But I was not able to explain the other 10%.

 For question in original post, I can not make 24 TB from available field of 
 df output:
 207 KiB available on his zfs lustre,  198 KiB on ldiskfs lustre.
 At the same time the difference of the total space is
 233548424256 -207693153280 = 25855270976 KiB = 24.09 TB.

 Götz, could you please tell us what did you mean by available ?


I was comparing the Lustre file system size from the two
configurations, the space available for user data. I expected it to be
the same, that is 218T for both file systems.

I understand that you have the same issue.

Regards, Götz Waschk
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org


[lustre-discuss] free space on ldiskfs vs. zfs

2015-08-19 Thread Götz Waschk
Dear Lustre experts,

I have configured two different Lustre instances, both using Lustre
2.5.3, one with ldiskfs on RAID-6 hardware RAID and one using ZFS and
RAID-Z2, using the same type of hardware. I was wondering, why I 24 TB
less space available, when I should have the same amount of parity
used:

 # lfs df
UUID   1K-blocksUsed   Available Use% Mounted on
fs19-MDT_UUID   50322916  47269646494784   1%
/testlustre/fs19[MDT:0]
fs19-OST_UUID51923288320   12672 51923273600   0%
/testlustre/fs19[OST:0]
fs19-OST0001_UUID51923288320   12672 51923273600   0%
/testlustre/fs19[OST:1]
fs19-OST0002_UUID51923288320   12672 51923273600   0%
/testlustre/fs19[OST:2]
fs19-OST0003_UUID51923288320   12672 51923273600   0%
/testlustre/fs19[OST:3]
filesystem summary:  207693153280   50688 207693094400   0% /testlustre/fs19
UUID   1K-blocksUsed   Available Use% Mounted on
fs18-MDT_UUID   47177700  48215243550028   1%
/lustre/fs18[MDT:0]
fs18-OST_UUID58387106064  6014088200 49452733560  11%
/lustre/fs18[OST:0]
fs18-OST0001_UUID58387106064  5919753028 49547068928  11%
/lustre/fs18[OST:1]
fs18-OST0002_UUID58387106064  5944542316 49522279640  11%
/lustre/fs18[OST:2]
fs18-OST0003_UUID58387106064  5906712004 49560109952  11%
/lustre/fs18[OST:3]
filesystem summary:  233548424256 23785095548 198082192080  11% /lustre/fs18

fs18 is using ldiskfs, while fs19 is ZFS:
# zpool list
NAME  SIZE  ALLOC   FREECAP  DEDUP  HEALTH  ALTROOT
lustre-ost165T  18,1M  65,0T 0%  1.00x  ONLINE  -
# zfs list
NAME   USED  AVAIL  REFER  MOUNTPOINT
lustre-ost1   13,6M  48,7T   311K  /lustre-ost1
lustre-ost1/ost1  12,4M  48,7T  12,4M  /lustre-ost1/ost1


Any idea on why my 6TB per OST went?

Regards, Götz Waschk
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org


Re: [lustre-discuss] ldiskfs ost size limit

2015-07-21 Thread Götz Waschk
Thanks Ben,

is there a public document where I could have found this limit?

Regards, Götz

On Tue, Jul 21, 2015 at 4:33 PM, Ben Evans bev...@cray.com wrote:
 128 TB is the current limit

 You can force more than that, but it looks like you won't need to.

 -Ben Evans

 -Original Message-
 From: lustre-discuss [mailto:lustre-discuss-boun...@lists.lustre.org] On 
 Behalf Of Götz Waschk
 Sent: Tuesday, July 21, 2015 10:18 AM
 To: lustre-discuss@lists.lustre.org
 Subject: [lustre-discuss] ldiskfs ost size limit

 Dear Lustre experts,

 I'm in the process of installing a new Lustre file system based on version 
 2.5. What is the size limit for an OST when using ldiskfs?Can I format a 60 
 TB device with ldiskfs?

 Regards,
 Götz Waschk
 ___
 lustre-discuss mailing list
 lustre-discuss@lists.lustre.org
 http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org



-- 
AL I:40: Do what thou wilt shall be the whole of the Law.
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org


[lustre-discuss] ldiskfs ost size limit

2015-07-21 Thread Götz Waschk
Dear Lustre experts,

I'm in the process of installing a new Lustre file system based on
version 2.5. What is the size limit for an OST when using ldiskfs?Can
I format a 60 TB device with ldiskfs?

Regards,
Götz Waschk
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org


[lustre-discuss] Quota problem after upgrade from 2.1.6 to 2.5.3

2015-04-30 Thread Götz Waschk
Dear Lustre experts,

I have upgraded a lustre file system from 2.1.6 to 2.5.3. It went
smoothly, but there's a quota problem. It seems the kbytes used are by
a pretty large value:

du /lustre/fs17/someuser gives me 1695469716. The output of lfs quota
-v -u someuser /lustre/fs17 is:
Disk quotas for user someuser (uid 17300):
 Filesystem  kbytes   quota   limit   grace   files   quota   limit   grace
/lustre/fs17/
2558638436   0   0   -  224494   0
  0   -
fs17-MDT_UUID
  46948   -   0   -  224494   -   0   -
fs17-OST_UUID
494111600   -   0   -   -   -
 -   -
fs17-OST0001_UUID
457182260   -   0   -   -   -
 -   -
fs17-OST0002_UUID
402034508   -   0   -   -   -
 -   -
fs17-OST0003_UUID
277070248   -   0   -   -   -
 -   -
fs17-OST0004_UUID
392850908   -   0   -   -   -
 -   -
fs17-OST0005_UUID
535341964   -   0   -   -   -
 -   -
Total allocated inode limit: 0, total allocated block limit: 0

The number of files matches exactly the output of find -type f, but
the kbytes are wrong. I have checked the obvious, the user doesn't own
any files outside of his directory. Any idea on what to check?

Regards,Götz Waschk
___
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org


Re: [Lustre-discuss] Has anyone built 1.8.5 on Centos 5.6?

2011-06-01 Thread Götz Waschk
On Tue, May 31, 2011 at 10:56 PM, Joe Landman
land...@scalableinformatics.com wrote:
 Are there any gotchas?  Or is it worth staying with the older Centos
 5.4/5.5 based kernels from the download site?
Hi Joseph,

are you talking about the client or the server? The client works fine
with the 2.6.18-238.9.1.el5 kernel, on Scientific Linux 5, that is.

Regards, Götz Waschk
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] Patchless client on RHEL6

2011-05-10 Thread Götz Waschk
Hi everyone,

the 1_8 branch in the Whamcloud git repository has the needed patches
to make Lustre 1.8 build on RHEL6. I have tested it a bit and I
couldn't make it crash yet.

Regards, Götz Waschk
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] git server down?

2011-01-18 Thread Götz Waschk
Hi everyone,

I have tried to access the git server but have received this error:

% git clone git://git.lustre.org/prime/lustre.git
Initialized empty Git repository in /tmp/lustre/.git/
fatal: read error: Connection reset by peer

Is this a known problem, will it be fixed soon?

Regards, Götz Waschk
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] Patchless client on RHEL6

2010-12-10 Thread Götz Waschk
On Fri, Dec 10, 2010 at 2:29 PM, Sébastien Buisson
sebastien.buis...@bull.net wrote:
 Maybe you've hit the problem addressed by attachment
 https://bugzilla.lustre.org/attachment.cgi?id=30289 from bug 22375.
 It was initially designed for Lustre 2.0, but maybe you could adapt it for
 1.8.
Hi,

I have rediffed the patch for Lustre 1.8.5, ran the auto* tools and
after that I got the same build error as above. By the way, git master
of Lustre 2.x fails in the same way.

Regards, Götz Waschk
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] Newbie Q : 16TB OST with 1.8.2

2010-04-22 Thread Götz Waschk
On Thu, Apr 22, 2010 at 4:01 PM, Norberto Meijome numard...@gmail.com wrote:
 I am looking at using Lustre 1.8.2 with Centos 5.3 or 5.4 64bit. I believe
 this should support 16TB ext4/lsfdisk backed OSTs.
 Is this correct ?

Hi Norberto,

yes, it does. I'm using it on Scientific Linux 5.4:

# lfs df -h /lustre/fs4/
UUID bytes  Used Available  Use% Mounted on
fs4-MDT_UUID895.9G471.9M844.3G0% /lustre/fs4[MDT:0]
fs4-OST_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:0]
fs4-OST0001_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:1]
fs4-OST0002_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:2]
fs4-OST0003_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:3]
fs4-OST0004_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:4]
fs4-OST0005_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:5]
fs4-OST0006_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:6]
fs4-OST0007_UUID 11.5T 27.1M 10.9T0% /lustre/fs4[OST:7]

filesystem summary:  92.2T216.5M 87.5T0% /lustre/fs4


Regards, Götz

-- 
AL I:40: Do what thou wilt shall be the whole of the Law.
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] Extremely high load and hanging processes on a Lustre client

2010-03-08 Thread Götz Waschk
2010/3/6 Bernd Schubert bs_li...@aakef.fastmail.fm:
 Do you have any idea what to test before rebooting the machine?
 Don't reboot, but disable LRU resizing.
 for i in /proc/fs/lustre/ldlm/namespaces/*; do echo 800  ${i}/lru_size; done
 At least that helped all the time before when we had that problem. I hoped it
 would be fixed in 1.8.2, but seems it is not. Please open a bug report.
Hi Bernd,

thanks for your help, it worked. I have opened a bug here:

https://bugzilla.lustre.org/show_bug.cgi?id=22276

Regards, Götz Waschk
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] Extremely high load and hanging processes on a Lustre client

2010-03-05 Thread Götz Waschk
Hi everyone,

I have a critical problem on one of my Lustre client machines running
Scientific Linux 5.4 and the patchless Lustre 1.8.2 client. After a
few days of usage, some processes like cp and kswapd0 start to use
100% CPU. Only 180k of swap space are in use though.

Processes that try to access Lustre use a lot of CPU and seem to hang.

There is some output in the kernel log I'll attach to this mail.

Do you have any idea what to test before rebooting the machine?

Regards, Götz Waschk

-- 
AL I:40: Do what thou wilt shall be the whole of the Law.


kernel-log.txt.bz2
Description: BZip2 compressed data


kernel-lustre-log.txt.bz2
Description: BZip2 compressed data
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] problems while installing lustre kernel

2010-03-01 Thread Götz Waschk
On Mon, Mar 1, 2010 at 3:09 PM, SRIVATSA N D shri14...@gmail.com wrote:
 hello,
    I'm experiencing problems while installing lustre kernel on fedora 9 . I
 have attached the screen shot. Please help.

Hi Srivatsa,

you are trying to install a RHEL5 package on Fedora, this is not
supported. You should use RHEL5, but you can also try CentOS or
Scientific Linux.

To work around the error above, just use rpm --force to install the package.

Regards, Götz Waschk
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] Upgrade from Lustre 1.6.4.3 to 1.8.1.1

2010-01-13 Thread Götz Waschk
Hi everyone,

I'm planning a Lustre server upgrade. Currently I'm running 1.6.4.3
and I want to upgrade to 1.8.1.1. The operations manual states that
only upgrades from the latest 1.6.x version are supported. Do you
recommend to upgrade to the latest 1.6.x first or would it be rather
safe to try the direct upgrade? I'm not using quotas on these servers.


Regards, Götz Waschk
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] Building 1.8.0 patchless client

2009-05-19 Thread Götz Waschk
On Tue, May 19, 2009 at 1:24 AM, Martin Pokorny mpoko...@nrao.edu wrote:
 What I ended up doing was to remove the conditional near line 2281 of
 aclocal.m4, and then completed the build with make configure,
 configure --linux-src=..., etc. Does anyone know if this will work, or
 was a risky thing to do? I know this isn't the right thing to do, but
 I'm willing to do tests with the packages I built if there are no
 obvious reasons not to.
Hi Martin,

I had to do the same, building the patchless client for the latest
RHEL5.3 kernel. It didn't cause any problems, this is just a small bug
in the build system and it doesn't affect the built kernel module.

Regards, Götz Waschk
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] tuning max_sectors

2009-04-17 Thread Götz Waschk
Hi everyone,

on starting an OST I get this message in the log:

Lustre: zn_atlas-OST: underlying device cciss/c1d0p1 should be
tuned for larger I/O requests: max_sectors = 1024 could be up to
max_hw_sectors=2048

/dev/cciss/c1d0p1 is a RAID6 on a HP Smart Array P800 controller with
12 750 GB SATA drives. The stripe size is 128 KB according to the
hpacucli script.

The OST was created with
mkfs.lustre --fsname=zn_atlas --ost --mkfsoptions=-E stride=32 -E
stripe-width=320 -J device=/dev/vg00/j-ost0 -i 1048576
--mgsnode=192.168.22...@o2ib1,141.34.21...@tcp0 /dev/cciss/c1d0p1

What can I do?


Regards, Götz Waschk
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] problem mounting two different lustre instances

2009-02-10 Thread Götz Waschk
On Tue, Feb 10, 2009 at 1:44 AM, Isaac Huang he.hu...@sun.com wrote:
 On Mon, Feb 09, 2009 at 04:52:20PM +0100, G?tz Waschk wrote:
 My client has this in modprobe.conf:
 options lnet networks=o2ib,tcp
 I'm trying to mount the remote network with
 mount -t lustre 141.34.228...@tcp0:/atlas /scratch/lustre-1.6/atlas
 and the command just hangs, the error is this:
 LustreError: 2887:0:(events.c:66:request_out_callback()) @@@ type 4,
 status -113  ...@0100dfc2ac00 x7/t0

Hi Isaac,

 The outgoing message failed with -113 (EHOSTUNREACH). What does lctl
 list_nids say on the client?
on that client, the output is:
192.168.224...@o2ib
141.34.216...@tcp

 Also, please:
 echo +neterror  /proc/sys/lnet/printk
 So that more network errors would go onto console.
OK, after the next mount attempt I have this in the log now:

Lustre: OBD class driver, http://www.lustre.org/
Lustre Version: 1.6.6
Build Version:
1.6.6-1970010101-PRISTINE-.usr.src.redhat.BUILD.lustre-1.6.6.kernel-2.6.9-78.0.13.ELsmp
Lustre: Added LNI 192.168.224...@o2ib [8/64]
Lustre: Added LNI 141.34.216...@tcp [8/256]
Lustre: Accept secure, port 988
Lustre: Lustre Client File System; http://www.lustre.org/
Lustre: 2887:0:(o2iblnd_cb.c:2704:kiblnd_cm_callback())
192.168.22...@o2ib: ROUTE ERROR -22
Lustre: 2887:0:(o2iblnd_cb.c:2118:kiblnd_peer_connect_failed())
Deleting messages for 192.168.22...@o2ib: connection failed
LustreError: 2887:0:(events.c:66:request_out_callback()) @@@ type 4,
status -113  r...@010037eeac00 x7/t0
o38-atlas-mdt_u...@192.168.22.32@o2ib:12/10 lens 240/400 e 0 to 5
dl 1234255029 ref 2 fl Rpc:/0/0 rc 0/0
Lustre: 9263:0:(client.c:1199:ptlrpc_expire_one_request()) @@@ network
error (sent at 1234255024, 0s ago)  r...@010037eeac00 x7/t0
o38-atlas-mdt_u...@192.168.22.32@o2ib:12/10 lens 240/400 e 0 to 5
dl 1234255029 ref 1 fl Rpc:/0/0 rc 0/0
Lustre: Request x7 sent from atlas-MDT-mdc-0107fc2ee400 to NID
192.168.22...@o2ib 0s ago has timed out (limit 5s).
Lustre: 9264:0:(import.c:410:import_select_connection())
atlas-MDT-mdc-0107fc2ee400: tried all connections, increasing
latency to 5s
Lustre: 2887:0:(o2iblnd_cb.c:2704:kiblnd_cm_callback())
192.168.22...@o2ib: ROUTE ERROR -22
Lustre: 2887:0:(o2iblnd_cb.c:2118:kiblnd_peer_connect_failed())
Deleting messages for 192.168.22...@o2ib: connection failed
LustreError: 2887:0:(events.c:66:request_out_callback()) @@@ type 4,
status -113  r...@01080325a400 x10/t0
o38-atlas-mdt_u...@192.168.22.32@o2ib:12/10 lens 240/400 e 0 to 5
dl 1234255054 ref 2 fl Rpc:/0/0 rc 0/0
Lustre: 9263:0:(client.c:1199:ptlrpc_expire_one_request()) @@@ network
error (sent at 1234255049, 0s ago)  r...@01080325a400 x10/t0
o38-atlas-mdt_u...@192.168.22.32@o2ib:12/10 lens 240/400 e 0 to 5
dl 1234255054 ref 1 fl Rpc:/0/0 rc 0/0
Lustre: Request x10 sent from atlas-MDT-mdc-0107fc2ee400 to
NID 192.168.22...@o2ib 0s ago has timed out (limit 5s).


Regards, Götz

-- 
AL I:40: Do what thou wilt shall be the whole of the Law.
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] problem mounting two different lustre instances

2009-02-10 Thread Götz Waschk
On Tue, Feb 10, 2009 at 1:45 PM, Johann Lombardi joh...@sun.com wrote:
 38 = MDS_CONNECT. The client tries to reach the MDT via 192.168.22...@o2ib,
 whereas i think it should use tcp to access the lustre filesystem of the 
 remote
 cluster, is my understanding of your configuration correct?
That's right, it should use 141.34.228...@tcp0 instead.

Regards, Götz


-- 
AL I:40: Do what thou wilt shall be the whole of the Law.
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] problem mounting two different lustre instances

2009-02-09 Thread Götz Waschk
Hello everyone,

I have a problem mounting two different lustre instances on one
client. Both lustre instances are configured with o2ib networking for
the local clients and tcp for remote clients.

So I have two MGS instances, 141.34.228...@tcp0 is the remote lustre,
192.168.22...@o2ib0 is the local one.


My client has this in modprobe.conf:
options lnet networks=o2ib,tcp
I'm trying to mount the remote network with
mount -t lustre 141.34.228...@tcp0:/atlas /scratch/lustre-1.6/atlas
and the command just hangs, the error is this:
LustreError: 2887:0:(events.c:66:request_out_callback()) @@@ type 4,
status -113  r...@0100dfc2ac00 x7/t0
o38-atlas-mdt_u...@192.168.22.32@o2ib:12/10 lens 240/400 e 0 to 5
dl 1234194365 ref 2 fl Rpc:/0/0 rc 0/0

I can mount the local lustre just fine:
mount -t lustre 192.168.22...@o2ib0:/lhcb /lustre/lhcb/

On the other client I have reversed the network list in modprobe.conf:
options lnet networks=tcp,o2ib
Now I can mount both lustre instances, but both seem to use the tcp
network, even the one that is local and should use o2ib.

On the local MGS:
lctl list_nids
192.168.22...@o2ib
141.34.21...@tcp
On my client:
lctl which_nid 192.168.22...@o2ib 141.34.21...@tcp
141.34.21...@tcp


What can I do?

Regards, Götz Waschk
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] Lustre 1.6.5.1 + RHEL4.7 + Infiniband

2008-08-14 Thread Götz Waschk
Hi everyone,

I'm trying to build lustre 1.6.5.1 against Red Hat's kernel
2.6.9-78.EL for RHEL4, but the configure script doesn't detect the
headers needed for OFED gen2 support (Infiniband). Does anybody know a
fix for that problem?

Regards, Götz Waschk
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss