>Starting GPFS 5.1.4, you can use the CCR archive to restore the local node >(the node that is issuing the mmsdrrestore command) beside restoring the >entire cluster.
This is great addition, but how does the security model work? ie. How do the other cluster nodes know that this is a newly re-installed node can be trusted and is not a rogue node trying to gain cluster membership by a backdoor? Daniel From: gpfsug-discuss <[email protected]> on behalf of Truong Vu <[email protected]> Date: Saturday, 30 July 2022 at 01:35 To: [email protected] <[email protected]> Subject: Re: [gpfsug-discuss] add local nsd back to cluster? Starting GPFS 5.1.4, you can use the CCR archive to restore the local node (the node that is issuing the mmsdrrestore command) beside restoring the entire cluster. Prior to GPFS5.1.4, as the error message reviewed, you can only use the CCR archive to restore the entire cluster. GPFS must be down any node that is being restored. If is a good node in the cluster, use the -p option -p NodeName Specifies the node from which to obtain a valid GPFS configuration file. The node must be either the primary configuration server or a node that has a valid backup copy of the mmsdrfs file. If this parameter is not specified, the command uses the configuration file on the node from which the command is issued. Thanks, Tru. On 7/29/22, 12:51 PM, "gpfsug-discuss on behalf of [email protected]" <[email protected] on behalf of [email protected]> wrote: Send gpfsug-discuss mailing list submissions to [email protected] To subscribe or unsubscribe via the World Wide Web, visit http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org<http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org> or, via email, send a message with subject or body 'help' to [email protected] You can reach the person managing the list at [email protected] When replying, please edit your Subject line so it is more specific than "Re: Contents of gpfsug-discuss digest..." Today's Topics: 1. Re: add local nsd back to cluster? (shao feng) 2. Re: add local nsd back to cluster? (Stephen Ulmer) ---------------------------------------------------------------------- Message: 1 Date: Fri, 29 Jul 2022 23:54:24 +0800 From: shao feng <[email protected]> To: gpfsug main discussion list <[email protected]> Subject: Re: [gpfsug-discuss] add local nsd back to cluster? Message-ID: <caniv0orjkzbykqlvhgqekpko9y--ptprxfpjxpjbvkqmukq...@mail.gmail.com> Content-Type: text/plain; charset="utf-8" Thanks Olaf I've setup the mmsdr backup as https://www.ibm.com/docs/en/spectrum-scale/5.1.2?topic=exits-mmsdrbackup-user-exit<https://www.ibm.com/docs/en/spectrum-scale/5.1.2?topic=exits-mmsdrbackup-user-exit> , since my cluster is CCR enabled, it generate a CCR backup file, but when trying to restore from this file, it require quorum nodes to shutdown? Is it possible to restore without touching quorum nodes? [root@tofail ~]# mmsdrrestore -F CCRBackup.986.2022.07.29.23.06.19.myquorum.tar.gz Restoring a CCR backup archive is a cluster-wide operation. The -a flag is required. mmsdrrestore: Command failed. Examine previous error messages to determine cause. [root@tofail ~]# mmsdrrestore -F CCRBackup.986.2022.07.29.23.06.19.myquorum.tar.gz -a Restoring CCR backup Verifying that GPFS is inactive on quorum nodes mmsdrrestore: GPFS is still active on myquorum mmsdrrestore: Unexpected error from mmsdrrestore: CCR restore failed. Return code: 192 mmsdrrestore: Command failed. Examine previous error messages to determine cause. On Thu, Jul 28, 2022 at 3:14 PM Olaf Weiser <[email protected]> wrote: > > > Hi - > assuming, you'll run it withou ECE ?!? ... just with replication on the > file system level > ba aware, every time a node goes offline, you 'll have to restart the > disks in your filesystem .. This causes a complete scan of the meta data to > detect files with missing updates / replication > > > apart from that to your Q : > you may consider to backup mmsdr > additionally, take a look to mmsdrrestore, in case you want to restore a > nodes's SDR configuration > > quick and dirty.. save the content of /var/mmfs may also help you > > during the node is "gone".. of course.. the disk is down , after restore > of SDR / node's config .. it should be able to start .. > the rest runs as usual > > > > ------------------------------ > *Von:* gpfsug-discuss <[email protected]> im Auftrag von > shao feng <[email protected]> > *Gesendet:* Donnerstag, 28. Juli 2022 09:02 > *An:* gpfsug main discussion list <[email protected]> > *Betreff:* [EXTERNAL] [gpfsug-discuss] add local nsd back to cluster? > > Hi all, I am planning to implement a cluster with a bunch of old x86 > machines, the disks are not connected to nodes via the SAN network, instead > each x86 machine has some local attached disks. The question is regarding > node failure, for example > > Hi all, > > I am planning to implement a cluster with a bunch of old x86 machines, > the disks are not connected to nodes via the SAN network, instead each x86 > machine has some local attached disks. > The question is regarding node failure, for example only the operating > system disk fails and the nsd disks are good. In that case I plan to > replace the failing OS disk with a new one and install the OS on it and > re-attach these nsd disks to that node, my question is: will this work? how > can I add a nsd back to the cluster without restoring data from other > replicas since the data/metadata is actually not corrupted on nsd. > > Best regards, > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at gpfsug.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org<http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org> > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://gpfsug.org/pipermail/gpfsug-discuss_gpfsug.org/attachments/20220729/1fc0e167/attachment-0001.htm > ------------------------------ Message: 2 Date: Fri, 29 Jul 2022 12:48:44 -0400 From: Stephen Ulmer <[email protected]> To: gpfsug main discussion list <[email protected]> Subject: Re: [gpfsug-discuss] add local nsd back to cluster? Message-ID: <[email protected]> Content-Type: text/plain; charset="utf-8" If there are cluster nodes up, restore from the running nodes instead of the file. I think it?s -p, but look at the manual page. -- Stephen Ulmer Sent from a mobile device; please excuse auto-correct silliness. > On Jul 29, 2022, at 11:20 AM, shao feng <[email protected]> wrote: > > ? > Thanks Olaf > > I've setup the mmsdr backup as https://www.ibm.com/docs/en/spectrum-scale/5.1.2?topic=exits-mmsdrbackup-user-exit<https://www.ibm.com/docs/en/spectrum-scale/5.1.2?topic=exits-mmsdrbackup-user-exit> , since my cluster is CCR enabled, it generate a CCR backup file, > but when trying to restore from this file, it require quorum nodes to shutdown? Is it possible to restore without touching quorum nodes? > > [root@tofail ~]# mmsdrrestore -F CCRBackup.986.2022.07.29.23.06.19.myquorum.tar.gz > Restoring a CCR backup archive is a cluster-wide operation. > The -a flag is required. > mmsdrrestore: Command failed. Examine previous error messages to determine cause. > > [root@tofail ~]# mmsdrrestore -F CCRBackup.986.2022.07.29.23.06.19.myquorum.tar.gz -a > Restoring CCR backup > Verifying that GPFS is inactive on quorum nodes > mmsdrrestore: GPFS is still active on myquorum > mmsdrrestore: Unexpected error from mmsdrrestore: CCR restore failed. Return code: 192 > mmsdrrestore: Command failed. Examine previous error messages to determine cause. > > >> On Thu, Jul 28, 2022 at 3:14 PM Olaf Weiser <[email protected]> wrote: >> >> >> Hi - >> assuming, you'll run it withou ECE ?!? ... just with replication on the file system level >> ba aware, every time a node goes offline, you 'll have to restart the disks in your filesystem .. This causes a complete scan of the meta data to detect files with missing updates / replication >> >> >> apart from that to your Q : >> you may consider to backup mmsdr >> additionally, take a look to mmsdrrestore, in case you want to restore a nodes's SDR configuration >> >> quick and dirty.. save the content of /var/mmfs may also help you >> >> during the node is "gone".. of course.. the disk is down , after restore of SDR / node's config .. it should be able to start .. >> the rest runs as usual >> >> >> >> Von: gpfsug-discuss <[email protected]> im Auftrag von shao feng <[email protected]> >> Gesendet: Donnerstag, 28. Juli 2022 09:02 >> An: gpfsug main discussion list <[email protected]> >> Betreff: [EXTERNAL] [gpfsug-discuss] add local nsd back to cluster? >> >> This Message Is From an External Sender >> This message came from outside your organization. >> >> Hi all, >> >> I am planning to implement a cluster with a bunch of old x86 machines, the disks are not connected to nodes via the SAN network, instead each x86 machine has some local attached disks. >> The question is regarding node failure, for example only the operating system disk fails and the nsd disks are good. In that case I plan to replace the failing OS disk with a new one and install the OS on it and re-attach these nsd disks to that node, my question is: will this work? how can I add a nsd back to the cluster without restoring data from other replicas since the data/metadata is actually not corrupted on nsd. >> >> Best regards, >> _______________________________________________ >> gpfsug-discuss mailing list >> gpfsug-discuss at gpfsug.org >> http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org<http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org> > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at gpfsug.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org<http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org> -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://gpfsug.org/pipermail/gpfsug-discuss_gpfsug.org/attachments/20220729/1c773ee2/attachment.htm > ------------------------------ Subject: Digest Footer _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org<http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org> ------------------------------ End of gpfsug-discuss Digest, Vol 126, Issue 21 *********************************************** _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org<http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org>
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org
