Thanks Olaf I've setup the mmsdr backup as https://www.ibm.com/docs/en/spectrum-scale/5.1.2?topic=exits-mmsdrbackup-user-exit, since my cluster is CCR enabled, it generate a CCR backup file, but when trying to restore from this file, it require quorum nodes to shutdown? Is it possible to restore without touching quorum nodes?
[root@tofail ~]# mmsdrrestore -F CCRBackup.986.2022.07.29.23.06.19.myquorum.tar.gz Restoring a CCR backup archive is a cluster-wide operation. The -a flag is required. mmsdrrestore: Command failed. Examine previous error messages to determine cause. [root@tofail ~]# mmsdrrestore -F CCRBackup.986.2022.07.29.23.06.19.myquorum.tar.gz -a Restoring CCR backup Verifying that GPFS is inactive on quorum nodes mmsdrrestore: GPFS is still active on myquorum mmsdrrestore: Unexpected error from mmsdrrestore: CCR restore failed. Return code: 192 mmsdrrestore: Command failed. Examine previous error messages to determine cause. On Thu, Jul 28, 2022 at 3:14 PM Olaf Weiser <[email protected]> wrote: > > > Hi - > assuming, you'll run it withou ECE ?!? ... just with replication on the > file system level > ba aware, every time a node goes offline, you 'll have to restart the > disks in your filesystem .. This causes a complete scan of the meta data to > detect files with missing updates / replication > > > apart from that to your Q : > you may consider to backup mmsdr > additionally, take a look to mmsdrrestore, in case you want to restore a > nodes's SDR configuration > > quick and dirty.. save the content of /var/mmfs may also help you > > during the node is "gone".. of course.. the disk is down , after restore > of SDR / node's config .. it should be able to start .. > the rest runs as usual > > > > ------------------------------ > *Von:* gpfsug-discuss <[email protected]> im Auftrag von > shao feng <[email protected]> > *Gesendet:* Donnerstag, 28. Juli 2022 09:02 > *An:* gpfsug main discussion list <[email protected]> > *Betreff:* [EXTERNAL] [gpfsug-discuss] add local nsd back to cluster? > > Hi all, I am planning to implement a cluster with a bunch of old x86 > machines, the disks are not connected to nodes via the SAN network, instead > each x86 machine has some local attached disks. The question is regarding > node failure, for example > ZjQcmQRYFpfptBannerStart > This Message Is From an External Sender > This message came from outside your organization. > > ZjQcmQRYFpfptBannerEnd > Hi all, > > I am planning to implement a cluster with a bunch of old x86 machines, > the disks are not connected to nodes via the SAN network, instead each x86 > machine has some local attached disks. > The question is regarding node failure, for example only the operating > system disk fails and the nsd disks are good. In that case I plan to > replace the failing OS disk with a new one and install the OS on it and > re-attach these nsd disks to that node, my question is: will this work? how > can I add a nsd back to the cluster without restoring data from other > replicas since the data/metadata is actually not corrupted on nsd. > > Best regards, > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at gpfsug.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org >
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org
