What exactly do you mean by "I have received existing corrupted GPFS 4.2.2 lab"? Is the file system corrupted ? Maybe this error is then due to file system corruption.
Can you once try: mmmount fs_gpfs01 -a If this does not work then try: mmmount -o rs fs_gpfs01 Let me know which mount is working. Regards, The Spectrum Scale (GPFS) team ------------------------------------------------------------------------------------------------------------------ If you feel that your question can benefit other users of Spectrum Scale (GPFS), then please post it to the public IBM developerWroks Forum at https://www.ibm.com/developerworks/community/forums/html/forum?id=11111111-0000-0000-0000-000000000479 . If your query concerns a potential software error in Spectrum Scale (GPFS) and you have an IBM software maintenance contract please contact 1-800-237-5511 in the United States or your local IBM Service Center in other countries. The forum is informally monitored as time permits and should not be used for priority messages to the Spectrum Scale (GPFS) team. From: Ilan Schwarts <[email protected]> To: [email protected] Date: 07/04/2017 01:47 PM Subject: [gpfsug-discuss] Fail to mount file system Sent by: [email protected] Hi everyone, I have received existing corrupted GPFS 4.2.2 lab and I am trying to make it work. There are 2 nodes in a cluster: [root@LH20-GPFS1 ~]# mmgetstate -a Node number Node name GPFS state ------------------------------------------ 1 LH20-GPFS1 active 3 LH20-GPFS2 active The Cluster status is: [root@LH20-GPFS1 ~]# mmlscluster GPFS cluster information ======================== GPFS cluster name: MyCluster.LH20-GPFS2 GPFS cluster id: 10777108240438931454 GPFS UID domain: MyCluster.LH20-GPFS2 Remote shell command: /usr/bin/ssh Remote file copy command: /usr/bin/scp Repository type: CCR Node Daemon node name IP address Admin node name Designation -------------------------------------------------------------------- 1 LH20-GPFS1 10.10.158.61 LH20-GPFS1 quorum-manager 3 LH20-GPFS2 10.10.158.62 LH20-GPFS2 There is a file system: [root@LH20-GPFS1 ~]# mmlsnsd File system Disk name NSD servers --------------------------------------------------------------------------- fs_gpfs01 nynsd1 (directly attached) fs_gpfs01 nynsd2 (directly attached) [root@LH20-GPFS1 ~]# On each Node, There is folder /fs_gpfs01 The next step is to mount this fs_gpfs01 to be synced between the 2 nodes. Whilte executing mmmount i get exception: [root@LH20-GPFS1 ~]# mmmount /fs_gpfs01 Tue Jul 4 11:14:18 IDT 2017: mmmount: Mounting file systems ... mount: mount fs_gpfs01 on /fs_gpfs01 failed: Wrong medium type mmmount: Command failed. Examine previous error messages to determine cause. What am i doing wrong ? _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss
