There may be an issue with one of the other NSDs in the file system according 
to the “mmadddisk: File system home has some disks that are in a non-ready 
state.“ message in our output.  Best to check the status of the NSDs in the 
file system using the `mmlsdisk home` and if any disks are not ‘up’ then run 
the `mmchdisk home start -a` command after confirming that all nsdservers can 
see the disks.  I typically use `mmdsh -N nsdnodes tspreparedisk –s | dshbak 
–c` for that.

Hope that helps,
-Bryan

From: [email protected] 
[mailto:[email protected]] On Behalf Of J. Eric Wonderley
Sent: Thursday, January 05, 2017 2:01 PM
To: gpfsug main discussion list <[email protected]>
Subject: [gpfsug-discuss] nsd not adding with one quorum node down?

I have one quorum node down and attempting to add a nsd to a fs:
[root@cl005 ~]# mmadddisk home -F add_1_flh_home -v no |& tee 
/root/adddisk_flh_home.out
Verifying file system configuration information ...

The following disks of home will be formatted on node cl003:
    r10f1e5: size 1879610 MB
Extending Allocation Map
Checking Allocation Map for storage pool fc_ssd400G
  55 % complete on Thu Jan  5 14:43:31 2017
Lost connection to file system daemon.
mmadddisk: tsadddisk failed.
Verifying file system configuration information ...
mmadddisk: File system home has some disks that are in a non-ready state.
mmadddisk: Propagating the cluster configuration data to all
  affected nodes.  This is an asynchronous process.
mmadddisk: Command failed. Examine previous error messages to determine cause.
Had to use -v no (this failed once before).  Anyhow I next see:
[root@cl002 ~]# mmgetstate -aL

 Node number  Node name       Quorum  Nodes up  Total nodes  GPFS state  Remarks
------------------------------------------------------------------------------------
       1      cl001              0        0          8       down        quorum 
node
       2      cl002              5        6          8       active      quorum 
node
       3      cl003              5        0          8       arbitrating quorum 
node
       4      cl004              5        6          8       active      quorum 
node
       5      cl005              5        6          8       active      quorum 
node
       6      cl006              5        6          8       active      quorum 
node
       7      cl007              5        6          8       active      quorum 
node
       8      cl008              5        6          8       active      quorum 
node
[root@cl002 ~]# mmlsdisk home
disk         driver   sector     failure holds    holds                         
   storage
name         type       size       group metadata data  status        
availability pool
------------ -------- ------ ----------- -------- ----- ------------- 
------------ ------------
r10f1e5      nsd         512        1001 No       Yes   allocmap add  up        
   fc_ssd400G
r6d2e8       nsd         512        1001 No       Yes   ready         up        
   fc_8T
r6d3e8       nsd         512        1001 No       Yes   ready         up        
   fc_8T
Do all quorum node have to be up and participating to do these admin type 
operations?


________________________________

Note: This email is for the confidential use of the named addressee(s) only and 
may contain proprietary, confidential or privileged information. If you are not 
the intended recipient, you are hereby notified that any review, dissemination 
or copying of this email is strictly prohibited, and to please notify the 
sender immediately and destroy this email and any attachments. Email 
transmission cannot be guaranteed to be secure or error-free. The Company, 
therefore, does not make any guarantees as to the completeness or accuracy of 
this email or any attachments. This email is for informational purposes only 
and does not constitute a recommendation, offer, request or solicitation of any 
kind to buy, sell, subscribe, redeem or perform any type of transaction of a 
financial product.
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss

Reply via email to