This is from the MDS machine: I have rebooted the client, mounted the /lustre 
partition and entered "ls /mnt/lustre". The client is hung and remains so.

(NOTE: much of this log is from earlier today - you can see previous attempts 
to make this process work)

Sep 17 06:16:52 lustre_mds0 kernel: Lustre: Lustre: Build Version: 
RC3--PRISTINE-2.6.32-279.2.1.el6_lustre.gc46c389.x86_
64
Sep 17 06:16:53 lustre_mds0 kernel: Lustre: Added LNI 10.127.24.42@tcp 
[8/256/0/180]
Sep 17 06:16:53 lustre_mds0 kernel: Lustre: Accept secure, port 988
Sep 17 06:16:54 lustre_mds0 kernel: Lustre: Lustre OSC module 
(ffffffffa07f2b20).
Sep 17 06:16:54 lustre_mds0 kernel: Lustre: Lustre LOV module 
(ffffffffa08714c0).
Sep 17 06:16:54 lustre_mds0 kernel: Lustre: Lustre client module 
(ffffffffa095e260).
Sep 17 06:17:17 lustre_mds0 kernel: LDISKFS-fs warning (device loop0): 
ldiskfs_fill_super: extents feature not enabled o
n this filesystem, use tune2fs.
Sep 17 06:17:17 lustre_mds0 kernel: LDISKFS-fs (loop0): warning: maximal mount 
count reached, running e2fsck is recommen
ded
Sep 17 06:17:17 lustre_mds0 kernel: LDISKFS-fs (loop0): mounted filesystem with 
ordered data mode. Opts:
Sep 17 06:18:12 lustre_mds0 kernel: LDISKFS-fs warning (device loop0): 
ldiskfs_fill_super: extents feature not enabled o
n this filesystem, use tune2fs.
Sep 17 06:18:12 lustre_mds0 kernel: LDISKFS-fs (loop0): mounted filesystem with 
ordered data mode. Opts:
Sep 17 06:20:49 lustre_mds0 kernel: LDISKFS-fs warning (device loop0): 
ldiskfs_fill_super: extents feature not enabled o
n this filesystem, use tune2fs.
Sep 17 06:20:49 lustre_mds0 kernel: LDISKFS-fs (loop0): mounted filesystem with 
ordered data mode. Opts:
Sep 17 06:20:49 lustre_mds0 kernel: LDISKFS-fs warning (device loop0): 
ldiskfs_fill_super: extents feature not enabled o
n this filesystem, use tune2fs.
Sep 17 06:20:49 lustre_mds0 kernel: LDISKFS-fs (loop0): mounted filesystem with 
ordered data mode. Opts:
Sep 17 06:20:49 lustre_mds0 kernel: Lustre: MGS MGS started
Sep 17 06:20:49 lustre_mds0 kernel: Lustre: 
1228:0:(ldlm_lib.c:946:target_handle_connect()) MGS: connection from 7923c00
8-a0de-1c87-f21a-4a5ab48abb96@0@lo t0 exp (null) cur 1347888049 last 0
Sep 17 06:20:49 lustre_mds0 kernel: Lustre: MGC10.127.24.42@tcp: Reactivating 
import
Sep 17 06:20:49 lustre_mds0 kernel: Lustre: MGS: Regenerating lustre-MDTffff 
log by user request.
Sep 17 06:20:49 lustre_mds0 kernel: Lustre: Setting parameter 
lustre-MDT0000-mdtlov.lov.stripesize in log lustre-MDT0000

Sep 17 06:20:49 lustre_mds0 kernel: Lustre: Enabling ACL
Sep 17 06:20:49 lustre_mds0 kernel: Lustre: Enabling user_xattr
Sep 17 06:20:50 lustre_mds0 kernel: Lustre: lustre-MDT0000: new disk, 
initializing
Sep 17 06:20:50 lustre_mds0 kernel: Lustre: 
1240:0:(mdt_lproc.c:416:lprocfs_wr_identity_upcall()) lustre-MDT0000: identi
ty upcall set to /usr/sbin/l_getidentity
Sep 17 06:21:04 lustre_mds0 kernel: Lustre: 
1228:0:(ldlm_lib.c:946:target_handle_connect()) MGS: connection from b05bf03
[email protected]@tcp t0 exp (null) cur 1347888064 
last 0
Sep 17 06:21:04 lustre_mds0 kernel: Lustre: MGS: Regenerating lustre-OSTffff 
log by user request.
Sep 17 06:21:10 lustre_mds0 kernel: Lustre: MDS mdd_obd-lustre-MDT0000: 
lustre-OST0000_UUID now active, resetting orphan
s
Sep 17 06:22:44 lustre_mds0 kernel: Lustre: 
1228:0:(ldlm_lib.c:946:target_handle_connect()) MGS: connection from c72ffd0
[email protected]@tcp t0 exp (null) cur 1347888164 
last 0
Sep 17 06:22:44 lustre_mds0 kernel: Lustre: MGS: Regenerating lustre-OSTffff 
log by user request.
Sep 17 06:22:47 lustre_mds0 kernel: Lustre: MDS mdd_obd-lustre-MDT0000: 
lustre-OST0001_UUID now active, resetting orphan
s
Sep 17 06:24:42 lustre_mds0 kernel: LustreError: 152-6: Ignoring deprecated 
mount option 'acl'.
Sep 17 06:24:42 lustre_mds0 kernel: Lustre: 
1253:0:(ldlm_lib.c:946:target_handle_connect()) lustre-MDT0000: connection f
rom 8c097db6-d726-216f-0c84-77cbead03807@0@lo t0 exp (null) cur 1347888282 last 0
Sep 17 06:24:42 lustre_mds0 kernel: Lustre: Mounted lustre-client
Sep 17 06:27:21 lustre_mds0 kernel: Lustre: Modifying parameter 
lustre-MDT0000-mdtlov.lov.stripesize in log lustre-MDT00
00
Sep 17 06:27:21 lustre_mds0 kernel: Lustre: Skipped 4 previous similar messages
Sep 17 06:27:45 lustre_mds0 kernel: Lustre: ctl-lustre-MDT0000: super-sequence 
allocation rc = 0 [0x0000000200000400-0x0
000000240000400):0:0
Sep 17 06:34:24 lustre_mds0 kernel: Lustre: 
1227:0:(ldlm_lib.c:946:target_handle_connect()) MGS: connection from 9077ab0
[email protected]@tcp t0 exp (null) cur 1347888864 
last 0
Sep 17 06:34:24 lustre_mds0 kernel: Lustre: MGS: Regenerating lustre-OSTffff 
log by user request.
Sep 17 06:34:28 lustre_mds0 kernel: Lustre: MDS mdd_obd-lustre-MDT0000: 
lustre-OST0002_UUID now active, resetting orphan
s
Sep 17 06:35:01 lustre_mds0 kernel: LustreError: 
1352:0:(ldlm_request.c:1172:ldlm_cli_cancel_req()) Got rc -108 from can
cel RPC: canceling anyway
Sep 17 06:35:01 lustre_mds0 kernel: LustreError: 
1352:0:(ldlm_request.c:1799:ldlm_cli_cancel_list()) ldlm_cli_cancel_lis
t: -108
Sep 17 06:35:01 lustre_mds0 kernel: Lustre: Unmounted lustre-client
Sep 17 06:35:27 lustre_mds0 kernel: LustreError: 152-6: Ignoring deprecated 
mount option 'acl'.
Sep 17 06:35:27 lustre_mds0 kernel: Lustre: 
1252:0:(ldlm_lib.c:946:target_handle_connect()) lustre-MDT0000: connection f
rom b6b66579-1f44-90e5-ae63-e778d4ed6ac5@0@lo t0 exp (null) cur 1347888927 last 0
Sep 17 06:35:27 lustre_mds0 kernel: Lustre: Mounted lustre-client
Sep 17 06:57:08 lustre_mds0 kernel: Lustre: 
1227:0:(ldlm_lib.c:946:target_handle_connect()) MGS: connection from 35c69b8
[email protected]@tcp t0 exp (null) cur 1347890228 
last 0
Sep 17 07:31:54 lustre_mds0 kernel: hrtimer: interrupt took 10250792 ns
Sep 17 07:34:23 lustre_mds0 kernel: Lustre: 
1328:0:(client.c:1780:ptlrpc_expire_one_request()) @@@ Request  sent has tim
ed out for slow reply: [sent 1347892452/real 1347892452]  req@ffff880078a04800 
x1413362811414993/t0(0) o104->lustre-MDT0
[email protected]@tcp:15/16 lens 296/192 e 0 to 1 dl 1347892463 ref 1 fl 
Rpc:XN/0/ffffffff rc 0/-1
Sep 17 07:34:23 lustre_mds0 kernel: LustreError: 138-a: lustre-MDT0000: A 
client on nid 10.127.24.35@tcp was evicted due
 to a lock blocking callback time out: rc -107
Sep 17 07:34:39 lustre_mds0 kernel: Lustre: MGS: haven't heard from client 
35c69b86-f9d2-e274-7ac5-5380a17f9c32 (at 10.1
27.24.35@tcp) in 50 seconds. I think it's dead, and I am evicting it. exp 
ffff88007bad8000, cur 1347892479 expire 134789
2449 last 1347892429
Sep 17 07:39:56 lustre_mds0 kernel: Lustre: 
1228:0:(ldlm_lib.c:946:target_handle_connect()) MGS: connection from 94b20a9
[email protected]@tcp t0 exp (null) cur 1347892796 
last 0
Sep 17 07:39:56 lustre_mds0 kernel: Lustre: 
1228:0:(ldlm_lib.c:946:target_handle_connect()) Skipped 1 previous similar m
essage
Sep 17 07:41:09 lustre_mds0 kernel: Lustre: MGS: haven't heard from client 
94b20a97-ffc1-174f-b2fa-84612be87730 (at 10.1
27.24.35@tcp) in 48 seconds. I think it's dead, and I am evicting it. exp 
ffff88002f159c00, cur 1347892869 expire 134789
2839 last 1347892821
Sep 17 07:41:13 lustre_mds0 kernel: Lustre: lustre-MDT0000: haven't heard from 
client 98c74908-5348-2392-0f61-6a7f6d3a03
76 (at 10.127.24.35@tcp) in 52 seconds. I think it's dead, and I am evicting 
it. exp ffff8800765d6000, cur 1347892873 ex
pire 1347892843 last 1347892821
Sep 17 07:43:41 lustre_mds0 kernel: Lustre: 
1228:0:(ldlm_lib.c:946:target_handle_connect()) MGS: connection from 272a840
[email protected]@tcp t0 exp (null) cur 1347893021 
last 0
Sep 17 07:43:41 lustre_mds0 kernel: Lustre: 
1228:0:(ldlm_lib.c:946:target_handle_connect()) Skipped 1 previous similar m
essage
Sep 17 08:45:08 lustre_mds0 kernel: Lustre: 
1228:0:(ldlm_lib.c:946:target_handle_connect()) MGS: connection from 5b11484
[email protected]@tcp t0 exp (null) cur 1347896708 
last 0
Sep 17 08:45:08 lustre_mds0 kernel: Lustre: 
1228:0:(ldlm_lib.c:946:target_handle_connect()) Skipped 1 previous similar m
essage
Sep 17 08:45:08 lustre_mds0 kernel: Lustre: 
1252:0:(ldlm_lib.c:946:target_handle_connect()) lustre-MDT0000: connection f
rom [email protected]@tcp t0 exp (null) cur 
1347896708 last 0
[root@lustre_mds0 lustre]#

-----Original Message-----
From: Colin Faber [mailto:[email protected]]
Sent: Monday, September 17, 2012 8:38 AM
To: Jon Yeargers
Cc: [email protected]
Subject: Re: [Lustre-discuss] Client hangs on 'simple' lustre setup

Hi,

Ideally just post the logging directly around the event and not the entire 
syslog =)

/var/log/messages right around the time you try and mount / ls

and

dmesg | tail -n 75

or so, if it's larger, pastebin is a great resource to reference so you're not 
polluting peoples inbox with large logs.

-cf

On 09/17/2012 09:33 AM, Jon Yeargers wrote:
> What's the preferred method for posting large files to this list?
>
> Does Lustre have specific log files that I should look for or just anything 
> in /var/log/messages that seems appropriate?
>
> -----Original Message-----
> From: Colin Faber [mailto:[email protected]]
> Sent: Monday, September 17, 2012 8:05 AM
> To: Jon Yeargers
> Cc: [email protected]
> Subject: Re: [Lustre-discuss] Client hangs on 'simple' lustre setup
>
> Hi Jon,
>
> If you could provide the logging from your clients and mds (as well as 
> possibly oss's) that'll help in determining the problem.
>
> -cf
>
> On 09/17/2012 08:57 AM, Jon Yeargers wrote:
>> Issue: I'm trying to assess the (possible) use of Lustre for our
>> group. To this end I've been trying to create a simple system to
>> explore the nuances. I can't seem to get past the 'llmount.sh' test
>> with any degree of success.
>>
>> What I've done: Each system (throwaway PCs with 70Gb HD, 2Gb RAM) is
>> formatted with CentOS 6.2. I then update everything and install the
>> Lustre kernel from downloads.whamcloud.com and add on the various
>> (appropriate) lustre and e2fs RPM files. Systems are rebooted and
>> tested with 'llmount.sh' (and then cleared with 'llmountcleanup.sh').
>> All is well to this point.
>>
>> First I create an MDS/MDT system via:
>>
>> /usr/sbin/mkfs.lustre --mgs --mdt --fsname=lustre
>> --device-size=2000000 --param sys.timeout=20
>> --mountfsoptions=errors=remount-ro,user_xattr,acl --param
>> lov.stripesize=1048576 --param lov.stripecount=0 --param
>> mdt.identity_upcall=/usr/sbin/l_getidentity --backfstype ldiskfs
>> --reformat /tmp/lustre-mdt1
>>
>> and then
>>
>> mkdir -p /mnt/mds1
>>
>> mount -t lustre -o loop,user_xattr,acl /tmp/lustre-mdt1 /mnt/mds1
>>
>> Next I take 3 systems and create a 2Gb loop mount via:
>>
>> /usr/sbin/mkfs.lustre --ost --fsname=lustre --device-size=2000000
>> --param sys.timeout=20 --mgsnode=lustre_MDS0@tcp --backfstype ldiskfs
>> --reformat /tmp/lustre-ost1
>>
>> mkdir -p /mnt/ost1
>>
>> mount -t lustre -o loop /tmp/lustre-ost1 /mnt/ost1
>>
>> The logs on the MDT box show the OSS boxes connecting up. All appears ok.
>>
>> Last I create a client and attach to the MDT box:
>>
>> mkdir -p /mnt/lustre
>>
>> mount -t lustre -o user_xattr,acl,flock luster_MDS0@tcp:/lustre
>> /mnt/lustre
>>
>> Again, the log on the MDT box shows the client connection. Appears to
>> be successful.
>>
>> Here's where the issues (appear to) start. If I do a 'df -h' on the
>> client it hangs after showing the system drives. If I attempt to
>> create files (via 'dd') on the lustre mount the session hangs and the
>> job can't be killed. Rebooting the client is the only solution.
>>
>> I can create and use a client on the MDS/MSG box. Doing so from any
>> other machine will hang.
>>
>>  From the MDS box:
>>
>> [root@lustre_mds0 lustre]# lctl dl
>>
>> 0 UP mgs MGS MGS 13
>>
>> 1 UP mgc MGC10.127.24.42@tcp 7923c008-a0de-1c87-f21a-4a5ab48abb96 5
>>
>> 2 UP lov lustre-MDT0000-mdtlov lustre-MDT0000-mdtlov_UUID 4
>>
>> 3 UP mdt lustre-MDT0000 lustre-MDT0000_UUID 7
>>
>> 4 UP mds mdd_obd-lustre-MDT0000 mdd_obd_uuid-lustre-MDT0000 3
>>
>> 5 UP osc lustre-OST0000-osc-MDT0000 lustre-MDT0000-mdtlov_UUID 5
>>
>> 6 UP osc lustre-OST0001-osc-MDT0000 lustre-MDT0000-mdtlov_UUID 5
>>
>> 7 UP lov lustre-clilov-ffff8800631c8000
>> b6b66579-1f44-90e5-ae63-e778d4ed6ac5 4
>>
>> 8 UP lmv lustre-clilmv-ffff8800631c8000
>> b6b66579-1f44-90e5-ae63-e778d4ed6ac5 4
>>
>> 9 UP mdc lustre-MDT0000-mdc-ffff8800631c8000
>> b6b66579-1f44-90e5-ae63-e778d4ed6ac5 5
>>
>> 10 UP osc lustre-OST0000-osc-ffff8800631c8000
>> b6b66579-1f44-90e5-ae63-e778d4ed6ac5 5
>>
>> 11 UP osc lustre-OST0001-osc-ffff8800631c8000
>> b6b66579-1f44-90e5-ae63-e778d4ed6ac5 5
>>
>> 12 UP osc lustre-OST0002-osc-ffff8800631c8000
>> b6b66579-1f44-90e5-ae63-e778d4ed6ac5 5
>>
>> 13 UP osc lustre-OST0002-osc-MDT0000 lustre-MDT0000-mdtlov_UUID 5
>>
>> [root@lustre_mds0 lustre]# lfs df -h
>>
>> UUID bytes Used Available Use% Mounted on
>>
>> lustre-MDT0000_UUID 1.4G 83.9M 1.3G 6% /mnt/lustre[MDT:0]
>>
>> lustre-OST0000_UUID 1.9G 1.1G 716.5M 61% /mnt/lustre[OST:0]
>>
>> lustre-OST0001_UUID 1.9G 1.1G 728.5M 60% /mnt/lustre[OST:1]
>>
>> lustre-OST0002_UUID 1.9G 1.1G 728.5M 60% /mnt/lustre[OST:2]
>>
>> filesystem summary: 5.6G 3.2G 2.1G 60% /mnt/lustre
>>
>> All appears normal.
>>
>> Doing this from another (identical) client:
>>
>> [root@lfstest0 lustre]# lctl dl
>>
>> 0 UP mgc MGC10.127.24.42@tcp 272a8405-8512-e9de-f532-feb5b7d6f9b1 5
>>
>> 1 UP lov lustre-clilov-ffff880070eee400
>> 0cb7fd2e-ade0-dab3-c4b9-6b7956ef9720 4
>>
>> 2 UP lmv lustre-clilmv-ffff880070eee400
>> 0cb7fd2e-ade0-dab3-c4b9-6b7956ef9720 4
>>
>> 3 UP mdc lustre-MDT0000-mdc-ffff880070eee400
>> 0cb7fd2e-ade0-dab3-c4b9-6b7956ef9720 5
>>
>> 4 UP osc lustre-OST0000-osc-ffff880070eee400
>> 0cb7fd2e-ade0-dab3-c4b9-6b7956ef9720 5
>>
>> 5 UP osc lustre-OST0001-osc-ffff880070eee400
>> 0cb7fd2e-ade0-dab3-c4b9-6b7956ef9720 5
>>
>> 6 UP osc lustre-OST0002-osc-ffff880070eee400
>> 0cb7fd2e-ade0-dab3-c4b9-6b7956ef9720 5
>>
>> [root@lfstest0 lustre]# lfs df
>>
>> UUID 1K-blocks Used Available Use% Mounted on
>>
>> lustre-MDT0000_UUID 1499596 85888 1313708 6% /mnt/lustre[MDT:0]
>>
>> OST0000 : inactive device
>>
>> lustre-OST0001_UUID 1968528 1122468 745996 60% /mnt/lustre[OST:1]
>>
>> OST0002 : inactive device
>>
>> filesystem summary: 1968528 1122468 745996 60% /mnt/luster
>>
>> Doing a ‘dd’ or ‘touch’ or even ‘df’ from this machine will hang it.
>>
>> EDIT: each system has all other systems defined in /etc/hosts and
>> entries in iptables to provide access.
>>
>> All systems have identical setup:
>>
>> [root@lfstest0 lustre]# rpm -qa | grep lustre
>>
>> lustre-ldiskfs-3.3.0-2.6.32_279.2.1.el6_lustre.gc46c389.x86_64.x86_64
>>
>> lustre-2.1.3-2.6.32_279.2.1.el6_lustre.gc46c389.x86_64.x86_64
>>
>> kernel-2.6.32-279.2.1.el6_lustre.gc46c389.x86_64
>>
>> lustre-modules-2.1.3-2.6.32_279.2.1.el6_lustre.gc46c389.x86_64.x86_64
>>
>> lustre-tests-2.1.3-2.6.32_279.2.1.el6_lustre.gc46c389.x86_64.x86_64
>>
>> [root@lfstest0 lustre]# uname -a
>>
>> Linux lfstest0 2.6.32-279.2.1.el6_lustre.gc46c389.x86_64 #1 SMP Mon
>> Aug 13 11:00:10 PDT 2012 x86_64 x86_64 x86_64 GNU/Linux
>>
>> [root@lfstest0 lustre]# rpm -qa | grep e2fs
>>
>> e2fsprogs-libs-1.41.90.wc2-7.el6.x86_64
>>
>> e2fsprogs-1.41.90.wc2-7.el6.x86_64
>>
>> SO: I'm clearly making several mistakes. Any pointers as to where to
>> start correcting them?
>>
>>
>>
>> _______________________________________________
>> Lustre-discuss mailing list
>> [email protected]
>> http://lists.lustre.org/mailman/listinfo/lustre-discuss
> _______________________________________________
> Lustre-discuss mailing list
> [email protected]
> http://lists.lustre.org/mailman/listinfo/lustre-discuss

_______________________________________________
Lustre-discuss mailing list
[email protected]
http://lists.lustre.org/mailman/listinfo/lustre-discuss

Reply via email to