[root@LH20-GPFS1 ~]# mmmount fs_gpfs01 -a Tue Jul 4 13:52:07 IDT 2017: mmmount: Mounting file systems ... LH20-GPFS1: mount: mount fs_gpfs01 on /fs_gpfs01 failed: Wrong medium type mmdsh: LH20-GPFS1 remote shell process had return code 32. LH20-GPFS2: mount: mount fs_gpfs01 on /fs_gpfs01 failed: Stale file handle mmdsh: LH20-GPFS2 remote shell process had return code 32. mmmount: Command failed. Examine previous error messages to determine cause.
[root@LH20-GPFS1 ~]# mmmount -o rs /fs_gpfs01 mmmount: Mount point can not be a relative path name: rs [root@LH20-GPFS1 ~]# mmmount -o rs fs_gpfs01 mmmount: Mount point can not be a relative path name: rs I recieve in "dmesg": [ 18.338044] sd 2:0:0:1: [sdc] Attached SCSI disk [ 141.363422] hvt_cn_callback: unexpected netlink message! [ 141.366153] hvt_cn_callback: unexpected netlink message! [ 4479.292850] tracedev: loading out-of-tree module taints kernel. [ 4479.292888] tracedev: module verification failed: signature and/or required key missing - tainting kernel [ 4482.928413] ------------[ cut here ]------------ [ 4482.928445] WARNING: at fs/xfs/xfs_aops.c:906 xfs_do_writepage+0x537/0x550 [xfs]() [ 4482.928446] Modules linked in: mmfs26(OE) mmfslinux(OE) tracedev(OE) iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi ext4 mbcache jbd2 loop intel_powerclamp iosf_mbi sg pcspkr hv_utils i2c_piix4 i2c_core nfsd auth_rpcgss nfs_acl lockd grace sunrpc binfmt_misc ip_tables xfs libcrc32c sd_mod crc_t10dif crct10dif_generic crct10dif_common ata_generic pata_acpi hv_netvsc hyperv_keyboard hid_hyperv hv_storvsc hyperv_fb serio_raw fjes floppy libata hv_vmbus dm_mirror dm_region_hash dm_log dm_mod [ 4482.928471] CPU: 1 PID: 15210 Comm: mmfsd Tainted: G OE ------------ 3.10.0-514.21.2.el7.x86_64 #1 On Tue, Jul 4, 2017 at 11:36 AM, IBM Spectrum Scale <[email protected]> wrote: > What exactly do you mean by "I have received existing corrupted GPFS 4.2.2 > lab"? > Is the file system corrupted ? Maybe this error is then due to file system > corruption. > > Can you once try: mmmount fs_gpfs01 -a > If this does not work then try: mmmount -o rs fs_gpfs01 > > Let me know which mount is working. > > Regards, The Spectrum Scale (GPFS) team > > ------------------------------------------------------------------------------------------------------------------ > If you feel that your question can benefit other users of Spectrum Scale > (GPFS), then please post it to the public IBM developerWroks Forum at > https://www.ibm.com/developerworks/community/forums/html/forum?id=11111111-0000-0000-0000-000000000479. > > If your query concerns a potential software error in Spectrum Scale (GPFS) > and you have an IBM software maintenance contract please contact > 1-800-237-5511 in the United States or your local IBM Service Center in > other countries. > > The forum is informally monitored as time permits and should not be used for > priority messages to the Spectrum Scale (GPFS) team. > > > > From: Ilan Schwarts <[email protected]> > To: [email protected] > Date: 07/04/2017 01:47 PM > Subject: [gpfsug-discuss] Fail to mount file system > Sent by: [email protected] > ________________________________ > > > > Hi everyone, I have received existing corrupted GPFS 4.2.2 lab and I > am trying to make it work. > There are 2 nodes in a cluster: > [root@LH20-GPFS1 ~]# mmgetstate -a > > Node number Node name GPFS state > ------------------------------------------ > 1 LH20-GPFS1 active > 3 LH20-GPFS2 active > > The Cluster status is: > [root@LH20-GPFS1 ~]# mmlscluster > > GPFS cluster information > ======================== > GPFS cluster name: MyCluster.LH20-GPFS2 > GPFS cluster id: 10777108240438931454 > GPFS UID domain: MyCluster.LH20-GPFS2 > Remote shell command: /usr/bin/ssh > Remote file copy command: /usr/bin/scp > Repository type: CCR > > Node Daemon node name IP address Admin node name Designation > -------------------------------------------------------------------- > 1 LH20-GPFS1 10.10.158.61 LH20-GPFS1 quorum-manager > 3 LH20-GPFS2 10.10.158.62 LH20-GPFS2 > > There is a file system: > [root@LH20-GPFS1 ~]# mmlsnsd > > File system Disk name NSD servers > --------------------------------------------------------------------------- > fs_gpfs01 nynsd1 (directly attached) > fs_gpfs01 nynsd2 (directly attached) > > [root@LH20-GPFS1 ~]# > > On each Node, There is folder /fs_gpfs01 > The next step is to mount this fs_gpfs01 to be synced between the 2 nodes. > Whilte executing mmmount i get exception: > [root@LH20-GPFS1 ~]# mmmount /fs_gpfs01 > Tue Jul 4 11:14:18 IDT 2017: mmmount: Mounting file systems ... > mount: mount fs_gpfs01 on /fs_gpfs01 failed: Wrong medium type > mmmount: Command failed. Examine previous error messages to determine cause. > > > What am i doing wrong ? > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss > > > > -- - Ilan Schwarts _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss
