Maybe someone else in the list can add clarity but I don't believe a
recovery process on mount would keep the MDS read-only or trigger that
trace. Something else may be going on.

I would start from the ground up. Bring your servers up, unmounted. Ensure
lnet is loaded and configured properly. Test lnet using ping or
lnet_selftest from your MDS to all of your OSS nodes. Then mount your
combined MGS/MDT volume on the MDS and see what happens.

Is your MDS in a high-availability pair?
What version of Lustre are you running?

...just a few things readers on the list might want to know.

--Jeff


On Wed, Jun 21, 2023 at 11:21 AM Mike Mosley <mike.mos...@charlotte.edu>
wrote:

> Jeff,
>
> At this point we have the OSS shutdown.  We were coming back from. full
> outage and so we are trying to get the MDS up before starting to bring up
> the OSS.
>
> Mike
>
> On Wed, Jun 21, 2023 at 2:15 PM Jeff Johnson <
> jeff.john...@aeoncomputing.com> wrote:
>
>> Mike,
>>
>> Have you made sure the the o2ib interface on all of your Lustre servers
>> (MDS & OSS) are functioning properly? Are you able to `lctl ping
>> x.x.x.x@o2ib` successfully between MDS and OSS nodes?
>>
>> --Jeff
>>
>>
>> On Wed, Jun 21, 2023 at 10:08 AM Mike Mosley via lustre-discuss <
>> lustre-discuss@lists.lustre.org> wrote:
>>
>>> Rick,
>>> 172.16.100.4 is the IB address of one of the OSS servers.    I
>>>  believe the mgt and mdt0 are the same target.   My understanding is
>>> that we have a single instanceof the MGT which is on the first MDT server
>>> i.e. it was created via a comand similar to:
>>>
>>> # mkfs.lustre --fsname=scratch --index=0 --mdt --mgs --replace /dev/sdb
>>>
>>> Does that make sense.
>>>
>>> On Wed, Jun 21, 2023 at 12:55 PM Mohr, Rick <moh...@ornl.gov> wrote:
>>>
>>>> Which host is 172.16.100.4?  Also, are the mgt and mdt0 on the same
>>>> target or are they two separate targets just on the same host?
>>>>
>>>> --Rick
>>>>
>>>>
>>>> On 6/21/23, 12:52 PM, "Mike Mosley" <mike.mos...@charlotte.edu
>>>> <mailto:mike.mos...@charlotte.edu>> wrote:
>>>>
>>>>
>>>> Hi Rick,
>>>>
>>>>
>>>> The MGS/MDS are combined. The output I posted is from the primary.
>>>>
>>>>
>>>>
>>>>
>>>> THanks,
>>>>
>>>>
>>>>
>>>>
>>>> Mike
>>>>
>>>>
>>>>
>>>>
>>>> On Wed, Jun 21, 2023 at 12:27 PM Mohr, Rick <moh...@ornl.gov <mailto:
>>>> moh...@ornl.gov> <mailto:moh...@ornl.gov <mailto:moh...@ornl.gov>>>
>>>> wrote:
>>>>
>>>>
>>>> Mike,
>>>>
>>>>
>>>> It looks like the mds server is having a problem contacting the mgs
>>>> server. I'm guessing the mgs is a separate host? I would start by looking
>>>> for possible network problems that might explain the LNet timeouts. You can
>>>> try using "lctl ping" to test the LNet connection between nodes, and you
>>>> can also try regular "ping" between the IP addresses on the IB interfaces.
>>>>
>>>>
>>>> --Rick
>>>>
>>>>
>>>>
>>>>
>>>> On 6/21/23, 11:35 AM, "lustre-discuss on behalf of Mike Mosley via
>>>> lustre-discuss" <lustre-discuss-boun...@lists.lustre.org <mailto:
>>>> lustre-discuss-boun...@lists.lustre.org> <_blank> <mailto:
>>>> lustre-discuss-boun...@lists.lustre.org <mailto:
>>>> lustre-discuss-boun...@lists.lustre.org> <_blank>> on behalf of
>>>> lustre-discuss@lists.lustre.org <mailto:lustre-discuss@lists.lustre.org>
>>>> <_blank> <mailto:lustre-discuss@lists.lustre.org <mailto:
>>>> lustre-discuss@lists.lustre.org> <_blank>>> wrote:
>>>>
>>>>
>>>>
>>>>
>>>> Greetings,
>>>>
>>>>
>>>>
>>>>
>>>> We have experienced some type of issue that is causing both of our MDS
>>>> servers to only be able to mount the mdt device in read only mode. Here are
>>>> some of the error messages we are seeing in the log files below. We lost
>>>> our Lustre expert a while back and we are not sure how to proceed to
>>>> troubleshoot this issue. Can anybody provide us guidance on how to proceed?
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Thanks,
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Mike
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Jun 20 15:12:14 hyd-mds1 kernel: INFO: task mount.lustre:4123 blocked
>>>> for more than 120 seconds.
>>>> Jun 20 15:12:14 hyd-mds1 kernel: "echo 0 >
>>>> /proc/sys/kernel/hung_task_timeout_secs" disables this message.
>>>> Jun 20 15:12:14 hyd-mds1 kernel: mount.lustre D ffff9f27a3bc5230 0 4123
>>>> 1 0x00000086
>>>> Jun 20 15:12:14 hyd-mds1 kernel: Call Trace:
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbb585da9>] schedule+0x29/0x70
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbb5838b1>]
>>>> schedule_timeout+0x221/0x2d0
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbaf6b8e5>] ?
>>>> tracing_is_on+0x15/0x30
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbaf6f5bd>] ?
>>>> tracing_record_cmdline+0x1d/0x120
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbaf77d9b>] ?
>>>> probe_sched_wakeup+0x2b/0xa0
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbaed7d15>] ?
>>>> ttwu_do_wakeup+0xb5/0xe0
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbb58615d>]
>>>> wait_for_completion+0xfd/0x140
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbaedb990>] ?
>>>> wake_up_state+0x20/0x20
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0f529a4>]
>>>> llog_process_or_fork+0x244/0x450 [obdclass]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0f52bc4>]
>>>> llog_process+0x14/0x20 [obdclass]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0f85d05>]
>>>> class_config_parse_llog+0x125/0x350 [obdclass]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0a69fc0>]
>>>> mgc_process_cfg_log+0x790/0xc40 [mgc]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0a6d4cc>]
>>>> mgc_process_log+0x3dc/0x8f0 [mgc]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0a6e15f>] ?
>>>> config_recover_log_add+0x13f/0x280 [mgc]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0f8df40>] ?
>>>> class_config_dump_handler+0x7e0/0x7e0 [obdclass]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0a6eb2b>]
>>>> mgc_process_config+0x88b/0x13f0 [mgc]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0f91b58>]
>>>> lustre_process_log+0x2d8/0xad0 [obdclass]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0e5a177>] ?
>>>> libcfs_debug_msg+0x57/0x80 [libcfs]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0f7c8b9>] ?
>>>> lprocfs_counter_add+0xf9/0x160 [obdclass]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0fc08f4>]
>>>> server_start_targets+0x13a4/0x2a20 [obdclass]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0f94bb0>] ?
>>>> lustre_start_mgc+0x260/0x2510 [obdclass]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0f8df40>] ?
>>>> class_config_dump_handler+0x7e0/0x7e0 [obdclass]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0fc303c>]
>>>> server_fill_super+0x10cc/0x1890 [obdclass]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0f97a08>]
>>>> lustre_fill_super+0x468/0x960 [obdclass]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0f975a0>] ?
>>>> lustre_common_put_super+0x270/0x270 [obdclass]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbb0510cf>]
>>>> mount_nodev+0x4f/0xb0
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffc0f8f9a8>]
>>>> lustre_mount+0x38/0x60 [obdclass]
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbb051c4e>]
>>>> mount_fs+0x3e/0x1b0
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbb0707a7>]
>>>> vfs_kern_mount+0x67/0x110
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbb072edf>]
>>>> do_mount+0x1ef/0xd00
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbb049d7a>] ?
>>>> __check_object_size+0x1ca/0x250
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbb0288ec>] ?
>>>> kmem_cache_alloc_trace+0x3c/0x200
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbb073d33>]
>>>> SyS_mount+0x83/0xd0
>>>> Jun 20 15:12:14 hyd-mds1 kernel: [<ffffffffbb592ed2>]
>>>> system_call_fastpath+0x25/0x2a
>>>> Jun 20 15:13:14 hyd-mds1 kernel: LNet:
>>>> 4458:0:(o2iblnd_cb.c:3397:kiblnd_check_conns()) Timed out tx for
>>>> 172.16.100.4@o2ib: 9 seconds
>>>> Jun 20 15:13:14 hyd-mds1 kernel: LNet:
>>>> 4458:0:(o2iblnd_cb.c:3397:kiblnd_check_conns()) Skipped 239 previous
>>>> similar messages
>>>> Jun 20 15:14:14 hyd-mds1 kernel: INFO: task mount.lustre:4123 blocked
>>>> for more than 120 seconds.
>>>> Jun 20 15:14:14 hyd-mds1 kernel: "echo 0 >
>>>> /proc/sys/kernel/hung_task_timeout_secs" disables this message.
>>>> Jun 20 15:14:14 hyd-mds1 kernel: mount.lustre D ffff9f27a3bc5230 0 4123
>>>> 1 0x00000086
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> dumpe2fs seems to show that the file systems are clean i.e.
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> dumpe2fs 1.45.6.wc1 (20-Mar-2020)
>>>> Filesystem volume name: hydra-MDT0000
>>>> Last mounted on: /
>>>> Filesystem UUID: 3ae09231-7f2a-43b3-a4ee-7f36080b5a66
>>>> Filesystem magic number: 0xEF53
>>>> Filesystem revision #: 1 (dynamic)
>>>> Filesystem features: has_journal ext_attr resize_inode dir_index
>>>> filetype mmp flex_bg dirdata sparse_super large_file huge_file uninit_bg
>>>> dir_nlink quota
>>>> Filesystem flags: signed_directory_hash
>>>> Default mount options: user_xattr acl
>>>> Filesystem state: clean
>>>> Errors behavior: Continue
>>>> Filesystem OS type: Linux
>>>> Inode count: 2247671504
>>>> Block count: 1404931944
>>>> Reserved block count: 70246597
>>>> Free blocks: 807627552
>>>> Free inodes: 2100036536
>>>> First block: 0
>>>> Block size: 4096
>>>> Fragment size: 4096
>>>> Reserved GDT blocks: 1024
>>>> Blocks per group: 20472
>>>> Fragments per group: 20472
>>>> Inodes per group: 32752
>>>> Inode blocks per group: 8188
>>>> Flex block group size: 16
>>>> Filesystem created: Thu Aug 8 14:21:01 2019
>>>> Last mount time: Tue Jun 20 15:19:03 2023
>>>> Last write time: Wed Jun 21 10:43:51 2023
>>>> Mount count: 38
>>>> Maximum mount count: -1
>>>> Last checked: Thu Aug 8 14:21:01 2019
>>>> Check interval: 0 (<none>)
>>>> Lifetime writes: 219 TB
>>>> Reserved blocks uid: 0 (user root)
>>>> Reserved blocks gid: 0 (group root)
>>>> First inode: 11
>>>> Inode size: 1024
>>>> Required extra isize: 32
>>>> Desired extra isize: 32
>>>> Journal inode: 8
>>>> Default directory hash: half_md4
>>>> Directory Hash Seed: 2e518531-82d9-4652-9acd-9cf9ca09c399
>>>> Journal backup: inode blocks
>>>> MMP block number: 1851467
>>>> MMP update interval: 5
>>>> User quota inode: 3
>>>> Group quota inode: 4
>>>> Journal features: journal_incompat_revoke
>>>> Journal size: 4096M
>>>> Journal length: 1048576
>>>> Journal sequence: 0x0a280713
>>>> Journal start: 0
>>>> MMP_block:
>>>> mmp_magic: 0x4d4d50
>>>> mmp_check_interval: 6
>>>> mmp_sequence: 0xff4d4d50
>>>> mmp_update_date: Wed Jun 21 10:43:51 2023
>>>> mmp_update_time: 1687358631
>>>> mmp_node_name: hyd-mds1.uncc.edu <_blank> <_blank>
>>>> mmp_device_name: dm-0
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>> lustre-discuss mailing list
>>> lustre-discuss@lists.lustre.org
>>> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>>>
>>
>>
>> --
>> ------------------------------
>> Jeff Johnson
>> Co-Founder
>> Aeon Computing
>>
>> jeff.john...@aeoncomputing.com
>> www.aeoncomputing.com
>> t: 858-412-3810 x1001   f: 858-412-3845
>> m: 619-204-9061
>>
>> 4170 Morena Boulevard, Suite C - San Diego, CA 92117
>>
>> High-Performance Computing / Lustre Filesystems / Scale-out Storage
>>
>

-- 
------------------------------
Jeff Johnson
Co-Founder
Aeon Computing

jeff.john...@aeoncomputing.com
www.aeoncomputing.com
t: 858-412-3810 x1001   f: 858-412-3845
m: 619-204-9061

4170 Morena Boulevard, Suite C - San Diego, CA 92117

High-Performance Computing / Lustre Filesystems / Scale-out Storage
_______________________________________________
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org

Reply via email to