Check /var/adm/ras/mmfs.log.latest The dmesg xfs bug is probably from boot if you look at the dmesg with -T to show the timestamp
On Jul 4, 2017 12:29 PM, "IBM Spectrum Scale" <[email protected]> wrote: > My bad gave the wrong command, the right one is: mmmount fs_gpfs01-o rs > > Also can you send output of mmlsnsd -X, need to check device type of the > NSDs. > > Are you ok with deleting the file system and disks and building everything > from scratch? > > > Regards, The Spectrum Scale (GPFS) team > > ------------------------------------------------------------ > ------------------------------------------------------ > If you feel that your question can benefit other users of Spectrum Scale > (GPFS), then please post it to the public IBM developerWroks Forum at > https://www.ibm.com/developerworks/community/ > forums/html/forum?id=11111111-0000-0000-0000-000000000479. > > If your query concerns a potential software error in Spectrum Scale (GPFS) > and you have an IBM software maintenance contract please contact > 1-800-237-5511 <(800)%20237-5511> in the United States or your local IBM > Service Center in other countries. > > The forum is informally monitored as time permits and should not be used > for priority messages to the Spectrum Scale (GPFS) team. > > > > From: Ilan Schwarts <[email protected]> > To: IBM Spectrum Scale <[email protected]> > Cc: [email protected], gpfsug main > discussion list <[email protected]> > Date: 07/04/2017 04:26 PM > Subject: Re: [gpfsug-discuss] Fail to mount file system > ------------------------------ > > > > [root@LH20-GPFS1 ~]# mmmount fs_gpfs01 -a > Tue Jul 4 13:52:07 IDT 2017: mmmount: Mounting file systems ... > LH20-GPFS1: mount: mount fs_gpfs01 on /fs_gpfs01 failed: Wrong medium type > mmdsh: LH20-GPFS1 remote shell process had return code 32. > LH20-GPFS2: mount: mount fs_gpfs01 on /fs_gpfs01 failed: Stale file handle > mmdsh: LH20-GPFS2 remote shell process had return code 32. > mmmount: Command failed. Examine previous error messages to determine > cause. > > [root@LH20-GPFS1 ~]# mmmount -o rs /fs_gpfs01 > mmmount: Mount point can not be a relative path name: rs > [root@LH20-GPFS1 ~]# mmmount -o rs fs_gpfs01 > mmmount: Mount point can not be a relative path name: rs > > > > I recieve in "dmesg": > > [ 18.338044] sd 2:0:0:1: [sdc] Attached SCSI disk > [ 141.363422] hvt_cn_callback: unexpected netlink message! > [ 141.366153] hvt_cn_callback: unexpected netlink message! > [ 4479.292850] tracedev: loading out-of-tree module taints kernel. > [ 4479.292888] tracedev: module verification failed: signature and/or > required key missing - tainting kernel > [ 4482.928413] ------------[ cut here ]------------ > [ 4482.928445] WARNING: at fs/xfs/xfs_aops.c:906 > xfs_do_writepage+0x537/0x550 [xfs]() > [ 4482.928446] Modules linked in: mmfs26(OE) mmfslinux(OE) > tracedev(OE) iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi ext4 > mbcache jbd2 loop intel_powerclamp iosf_mbi sg pcspkr hv_utils > i2c_piix4 i2c_core nfsd auth_rpcgss nfs_acl lockd grace sunrpc > binfmt_misc ip_tables xfs libcrc32c sd_mod crc_t10dif > crct10dif_generic crct10dif_common ata_generic pata_acpi hv_netvsc > hyperv_keyboard hid_hyperv hv_storvsc hyperv_fb serio_raw fjes floppy > libata hv_vmbus dm_mirror dm_region_hash dm_log dm_mod > [ 4482.928471] CPU: 1 PID: 15210 Comm: mmfsd Tainted: G OE > ------------ 3.10.0-514.21.2.el7.x86_64 #1 > > On Tue, Jul 4, 2017 at 11:36 AM, IBM Spectrum Scale <[email protected]> > wrote: > > What exactly do you mean by "I have received existing corrupted GPFS > 4.2.2 > > lab"? > > Is the file system corrupted ? Maybe this error is then due to file > system > > corruption. > > > > Can you once try: mmmount fs_gpfs01 -a > > If this does not work then try: mmmount -o rs fs_gpfs01 > > > > Let me know which mount is working. > > > > Regards, The Spectrum Scale (GPFS) team > > > > ------------------------------------------------------------ > ------------------------------------------------------ > > If you feel that your question can benefit other users of Spectrum Scale > > (GPFS), then please post it to the public IBM developerWroks Forum at > > https://www.ibm.com/developerworks/community/ > forums/html/forum?id=11111111-0000-0000-0000-000000000479. > > > > If your query concerns a potential software error in Spectrum Scale > (GPFS) > > and you have an IBM software maintenance contract please contact > > 1-800-237-5511 <(800)%20237-5511> in the United States or your local > IBM Service Center in > > other countries. > > > > The forum is informally monitored as time permits and should not be used > for > > priority messages to the Spectrum Scale (GPFS) team. > > > > > > > > From: Ilan Schwarts <[email protected]> > > To: [email protected] > > Date: 07/04/2017 01:47 PM > > Subject: [gpfsug-discuss] Fail to mount file system > > Sent by: [email protected] > > ________________________________ > > > > > > > > Hi everyone, I have received existing corrupted GPFS 4.2.2 lab and I > > am trying to make it work. > > There are 2 nodes in a cluster: > > [root@LH20-GPFS1 ~]# mmgetstate -a > > > > Node number Node name GPFS state > > ------------------------------------------ > > 1 LH20-GPFS1 active > > 3 LH20-GPFS2 active > > > > The Cluster status is: > > [root@LH20-GPFS1 ~]# mmlscluster > > > > GPFS cluster information > > ======================== > > GPFS cluster name: MyCluster.LH20-GPFS2 > > GPFS cluster id: 10777108240438931454 > > GPFS UID domain: MyCluster.LH20-GPFS2 > > Remote shell command: /usr/bin/ssh > > Remote file copy command: /usr/bin/scp > > Repository type: CCR > > > > Node Daemon node name IP address Admin node name Designation > > -------------------------------------------------------------------- > > 1 LH20-GPFS1 10.10.158.61 LH20-GPFS1 quorum-manager > > 3 LH20-GPFS2 10.10.158.62 LH20-GPFS2 > > > > There is a file system: > > [root@LH20-GPFS1 ~]# mmlsnsd > > > > File system Disk name NSD servers > > ------------------------------------------------------------ > --------------- > > fs_gpfs01 nynsd1 (directly attached) > > fs_gpfs01 nynsd2 (directly attached) > > > > [root@LH20-GPFS1 ~]# > > > > On each Node, There is folder /fs_gpfs01 > > The next step is to mount this fs_gpfs01 to be synced between the 2 > nodes. > > Whilte executing mmmount i get exception: > > [root@LH20-GPFS1 ~]# mmmount /fs_gpfs01 > > Tue Jul 4 11:14:18 IDT 2017: mmmount: Mounting file systems ... > > mount: mount fs_gpfs01 on /fs_gpfs01 failed: Wrong medium type > > mmmount: Command failed. Examine previous error messages to determine > cause. > > > > > > What am i doing wrong ? > > _______________________________________________ > > gpfsug-discuss mailing list > > gpfsug-discuss at spectrumscale.org > > http://gpfsug.org/mailman/listinfo/gpfsug-discuss > > > > > > > > > > > > -- > > > - > Ilan Schwarts > > > > > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss > >
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss
