Starting GPFS 5.1.4, you can use the CCR archive to restore the local node (the 
node that is issuing the mmsdrrestore command) beside restoring the entire 
cluster.

Prior to GPFS5.1.4, as the error message reviewed, you can only use the CCR 
archive to restore the entire cluster.  GPFS must be down any node that is 
being restored. 
If is a good node in the cluster, use the -p option

-p NodeName
         Specifies the node from which to obtain a valid GPFS
         configuration file. The node must be either the primary
         configuration server or a node that has a valid backup
         copy of the mmsdrfs file. If this parameter is not
         specified, the command uses the configuration file on
         the node from which the command is issued.

Thanks,
Tru.

On 7/29/22, 12:51 PM, "gpfsug-discuss on behalf of 
[email protected]" <[email protected] on behalf 
of [email protected]> wrote:

    Send gpfsug-discuss mailing list submissions to
        [email protected]

    To subscribe or unsubscribe via the World Wide Web, visit
        http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org 
    or, via email, send a message with subject or body 'help' to
        [email protected]

    You can reach the person managing the list at
        [email protected]

    When replying, please edit your Subject line so it is more specific
    than "Re: Contents of gpfsug-discuss digest..."


    Today's Topics:

       1. Re: add local nsd back to cluster? (shao feng)
       2. Re: add local nsd back to cluster? (Stephen Ulmer)


    ----------------------------------------------------------------------

    Message: 1
    Date: Fri, 29 Jul 2022 23:54:24 +0800
    From: shao feng <[email protected]>
    To: gpfsug main discussion list <[email protected]>
    Subject: Re: [gpfsug-discuss] add local nsd back to cluster?
    Message-ID:
        <caniv0orjkzbykqlvhgqekpko9y--ptprxfpjxpjbvkqmukq...@mail.gmail.com>
    Content-Type: text/plain; charset="utf-8"

    Thanks Olaf

    I've setup the mmsdr backup as
    
https://www.ibm.com/docs/en/spectrum-scale/5.1.2?topic=exits-mmsdrbackup-user-exit,
    since my cluster is CCR enabled, it generate a CCR backup file,
    but when trying to restore from this file, it require quorum nodes to
    shutdown? Is it possible to restore without touching quorum nodes?

    [root@tofail ~]# mmsdrrestore -F
    CCRBackup.986.2022.07.29.23.06.19.myquorum.tar.gz
    Restoring a CCR backup archive is a cluster-wide operation.
    The -a flag is required.
    mmsdrrestore: Command failed. Examine previous error messages to determine
    cause.

    [root@tofail ~]# mmsdrrestore -F
    CCRBackup.986.2022.07.29.23.06.19.myquorum.tar.gz -a
    Restoring CCR backup
    Verifying that GPFS is inactive on quorum nodes
    mmsdrrestore: GPFS is still active on myquorum
    mmsdrrestore: Unexpected error from mmsdrrestore: CCR restore failed.
    Return code: 192
    mmsdrrestore: Command failed. Examine previous error messages to determine
    cause.


    On Thu, Jul 28, 2022 at 3:14 PM Olaf Weiser <[email protected]> wrote:

    >
    >
    > Hi -
    > assuming, you'll run it  withou ECE  ?!? ... just with replication on the
    > file system level
    > ba aware, every time a node goes offline, you 'll have to restart the
    > disks in your filesystem .. This causes a complete scan of the meta data 
to
    > detect files with missing updates / replication
    >
    >
    > apart from that to your Q :
    > you may consider to backup mmsdr
    > additionally, take a look to   mmsdrrestore, in case you want to restore a
    > nodes's SDR configuration
    >
    > quick and dirty..  save the content  of  /var/mmfs  may also help you
    >
    > during the node is "gone".. of course.. the disk is down , after restore
    > of SDR / node's config .. it should be able to start ..
    > the rest runs as usual
    >
    >
    >
    > ------------------------------
    > *Von:* gpfsug-discuss <[email protected]> im Auftrag von
    > shao feng <[email protected]>
    > *Gesendet:* Donnerstag, 28. Juli 2022 09:02
    > *An:* gpfsug main discussion list <[email protected]>
    > *Betreff:* [EXTERNAL] [gpfsug-discuss] add local nsd back to cluster?
    >
    > Hi all, I am planning to implement  a cluster with a bunch of old x86
    > machines, the disks are not connected to nodes via the SAN network, 
instead
    > each x86 machine has some local attached disks. The question is regarding
    > node failure, for example
    > ZjQcmQRYFpfptBannerStart
    > This Message Is From an External Sender
    > This message came from outside your organization.
    >
    > ZjQcmQRYFpfptBannerEnd
    > Hi all,
    >
    > I am planning to implement  a cluster with a bunch of old x86 machines,
    > the disks are not connected to nodes via the SAN network, instead each x86
    > machine has some local attached disks.
    > The question is regarding node failure, for example only the operating
    > system disk fails and the nsd disks are good. In that case I plan to
    > replace the failing OS disk with a new one and install the OS on it and
    > re-attach these nsd disks to that node, my question is: will this work? 
how
    > can I add a nsd back to the cluster without restoring data from other
    > replicas since the data/metadata is actually not corrupted on nsd.
    >
    > Best regards,
    > _______________________________________________
    > gpfsug-discuss mailing list
    > gpfsug-discuss at gpfsug.org
    > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org 
    >
    -------------- next part --------------
    An HTML attachment was scrubbed...
    URL: 
<http://gpfsug.org/pipermail/gpfsug-discuss_gpfsug.org/attachments/20220729/1fc0e167/attachment-0001.htm
 >

    ------------------------------

    Message: 2
    Date: Fri, 29 Jul 2022 12:48:44 -0400
    From: Stephen Ulmer <[email protected]>
    To: gpfsug main discussion list <[email protected]>
    Subject: Re: [gpfsug-discuss] add local nsd back to cluster?
    Message-ID: <[email protected]>
    Content-Type: text/plain; charset="utf-8"

    If there are cluster nodes up, restore from the running nodes instead of 
the file. I think it?s -p, but look at the manual page.

    -- 
    Stephen Ulmer

    Sent from a mobile device; please excuse auto-correct silliness.

    > On Jul 29, 2022, at 11:20 AM, shao feng <[email protected]> wrote:
    > 
    > ?
    > Thanks Olaf
    > 
    > I've setup the mmsdr backup as 
https://www.ibm.com/docs/en/spectrum-scale/5.1.2?topic=exits-mmsdrbackup-user-exit,
 since my cluster is CCR enabled, it generate a CCR backup file,
    > but when trying to restore from this file, it require quorum nodes to 
shutdown? Is it possible to restore without touching quorum nodes?
    > 
    > [root@tofail ~]# mmsdrrestore -F 
CCRBackup.986.2022.07.29.23.06.19.myquorum.tar.gz
    > Restoring a CCR backup archive is a cluster-wide operation.
    > The -a flag is required.
    > mmsdrrestore: Command failed. Examine previous error messages to 
determine cause.
    > 
    > [root@tofail ~]# mmsdrrestore -F 
CCRBackup.986.2022.07.29.23.06.19.myquorum.tar.gz -a
    > Restoring CCR backup
    > Verifying that GPFS is inactive on quorum nodes
    > mmsdrrestore: GPFS is still active on myquorum
    > mmsdrrestore: Unexpected error from mmsdrrestore: CCR restore failed.  
Return code: 192
    > mmsdrrestore: Command failed. Examine previous error messages to 
determine cause.
    > 
    > 
    >> On Thu, Jul 28, 2022 at 3:14 PM Olaf Weiser <[email protected]> 
wrote:
    >> 
    >> 
    >> Hi - 
    >> assuming, you'll run it  withou ECE  ?!? ... just with replication on 
the file system level 
    >> ba aware, every time a node goes offline, you 'll have to restart the 
disks in your filesystem .. This causes a complete scan of the meta data to 
detect files with missing updates / replication
    >> 
    >> 
    >> apart from that to your Q :
    >> you may consider to backup mmsdr 
    >> additionally, take a look to   mmsdrrestore, in case you want to restore 
a nodes's SDR configuration 
    >> 
    >> quick and dirty..  save the content  of  /var/mmfs  may also help you 
    >> 
    >> during the node is "gone".. of course.. the disk is down , after restore 
of SDR / node's config .. it should be able to start .. 
    >> the rest runs as usual 
    >> 
    >> 
    >> 
    >> Von: gpfsug-discuss <[email protected]> im Auftrag von 
shao feng <[email protected]>
    >> Gesendet: Donnerstag, 28. Juli 2022 09:02
    >> An: gpfsug main discussion list <[email protected]>
    >> Betreff: [EXTERNAL] [gpfsug-discuss] add local nsd back to cluster?
    >>  
    >> This Message Is From an External Sender
    >> This message came from outside your organization.
    >>  
    >> Hi all,
    >> 
    >> I am planning to implement  a cluster with a bunch of old x86 machines, 
the disks are not connected to nodes via the SAN network, instead each x86 
machine has some local attached disks.
    >> The question is regarding node failure, for example only the operating 
system disk fails and the nsd disks are good. In that case I plan to replace 
the failing OS disk with a new one and install the OS on it and re-attach these 
nsd disks to that node, my question is: will this work? how can I add a nsd 
back to the cluster without restoring data from other replicas since the 
data/metadata is actually not corrupted on nsd.
    >> 
    >> Best regards,
    >> _______________________________________________
    >> gpfsug-discuss mailing list
    >> gpfsug-discuss at gpfsug.org
    >> http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org 
    > _______________________________________________
    > gpfsug-discuss mailing list
    > gpfsug-discuss at gpfsug.org
    > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org 
    -------------- next part --------------
    An HTML attachment was scrubbed...
    URL: 
<http://gpfsug.org/pipermail/gpfsug-discuss_gpfsug.org/attachments/20220729/1c773ee2/attachment.htm
 >

    ------------------------------

    Subject: Digest Footer

    _______________________________________________
    gpfsug-discuss mailing list
    gpfsug-discuss at gpfsug.org
    http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org 


    ------------------------------

    End of gpfsug-discuss Digest, Vol 126, Issue 21
    ***********************************************

_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at gpfsug.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org

Reply via email to