I mean the person tried to configure it... didnt do good job so now its me to continue On Jul 4, 2017 11:37, "IBM Spectrum Scale" <[email protected]> wrote:
> What exactly do you mean by "I have received existing corrupted GPFS > 4.2.2 lab"? > Is the file system corrupted ? Maybe this error is then due to file system > corruption. > > Can you once try: mmmount fs_gpfs01 -a > If this does not work then try: mmmount -o rs fs_gpfs01 > > Let me know which mount is working. > > Regards, The Spectrum Scale (GPFS) team > > ------------------------------------------------------------ > ------------------------------------------------------ > If you feel that your question can benefit other users of Spectrum Scale > (GPFS), then please post it to the public IBM developerWroks Forum at > https://www.ibm.com/developerworks/community/ > forums/html/forum?id=11111111-0000-0000-0000-000000000479. > > If your query concerns a potential software error in Spectrum Scale (GPFS) > and you have an IBM software maintenance contract please contact > 1-800-237-5511 in the United States or your local IBM Service Center in > other countries. > > The forum is informally monitored as time permits and should not be used > for priority messages to the Spectrum Scale (GPFS) team. > > > > From: Ilan Schwarts <[email protected]> > To: [email protected] > Date: 07/04/2017 01:47 PM > Subject: [gpfsug-discuss] Fail to mount file system > Sent by: [email protected] > ------------------------------ > > > > Hi everyone, I have received existing corrupted GPFS 4.2.2 lab and I > am trying to make it work. > There are 2 nodes in a cluster: > [root@LH20-GPFS1 ~]# mmgetstate -a > > Node number Node name GPFS state > ------------------------------------------ > 1 LH20-GPFS1 active > 3 LH20-GPFS2 active > > The Cluster status is: > [root@LH20-GPFS1 ~]# mmlscluster > > GPFS cluster information > ======================== > GPFS cluster name: MyCluster.LH20-GPFS2 > GPFS cluster id: 10777108240438931454 > GPFS UID domain: MyCluster.LH20-GPFS2 > Remote shell command: /usr/bin/ssh > Remote file copy command: /usr/bin/scp > Repository type: CCR > > Node Daemon node name IP address Admin node name Designation > -------------------------------------------------------------------- > 1 LH20-GPFS1 10.10.158.61 LH20-GPFS1 quorum-manager > 3 LH20-GPFS2 10.10.158.62 LH20-GPFS2 > > There is a file system: > [root@LH20-GPFS1 ~]# mmlsnsd > > File system Disk name NSD servers > ------------------------------------------------------------ > --------------- > fs_gpfs01 nynsd1 (directly attached) > fs_gpfs01 nynsd2 (directly attached) > > [root@LH20-GPFS1 ~]# > > On each Node, There is folder /fs_gpfs01 > The next step is to mount this fs_gpfs01 to be synced between the 2 nodes. > Whilte executing mmmount i get exception: > [root@LH20-GPFS1 ~]# mmmount /fs_gpfs01 > Tue Jul 4 11:14:18 IDT 2017: mmmount: Mounting file systems ... > mount: mount fs_gpfs01 on /fs_gpfs01 failed: Wrong medium type > mmmount: Command failed. Examine previous error messages to determine > cause. > > > What am i doing wrong ? > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss > > > > >
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss
