Thanks Mathiaz, Yes i do understand the concern, that if one of the remote file systems go down abruptly - the others will go down too.
However, i suppose we could bring down one of the filesystems before a planned downtime? For example, by unexporting the filesystems on NFS/SMB before the downtime? I might not want to be in a situation, where i have to bring down all the remote filesystems because of planned downtime of one of the remote clusters. Regards, Lohit On May 3, 2018, 7:41 AM -0400, Mathias Dietz <[email protected]>, wrote: > Hi Lohit, > > >I am thinking of using a single CES protocol cluster, with remote mounts > >from 3 storage clusters. > Technically this should work fine (assuming all 3 clusters use the same > uids/guids). However this has not been tested in our Test lab. > > > >One thing to watch, be careful if your CES root is on a remote fs, as if > >that goes away, so do all CES exports. > Not only the ces root file system is a concern, the whole CES cluster will go > down if any remote file systems with NFS exports is not available. > e.g. if remote cluster 1 is not available, the CES cluster will unmount the > corresponding file system which will lead to a NFS failure on all CES nodes. > > > Mit freundlichen Grüßen / Kind regards > > Mathias Dietz > > Spectrum Scale Development - Release Lead Architect (4.2.x) > Spectrum Scale RAS Architect > --------------------------------------------------------------------------- > IBM Deutschland > Am Weiher 24 > 65451 Kelsterbach > Phone: +49 70342744105 > Mobile: +49-15152801035 > E-Mail: [email protected] > ----------------------------------------------------------------------------- > IBM Deutschland Research & Development GmbH > Vorsitzender des Aufsichtsrats: Martina Koederitz, Geschäftsführung: Dirk > WittkoppSitz der Gesellschaft: Böblingen / Registergericht: Amtsgericht > Stuttgart, HRB 243294 > > > > From: [email protected] > To: gpfsug main discussion list <[email protected]> > Date: 01/05/2018 16:34 > Subject: Re: [gpfsug-discuss] Spectrum Scale CES and remote file > system mounts > Sent by: [email protected] > > > > Thanks Simon. > I will make sure i am careful about the CES root and test nfs exporting more > than 2 remote file systems. > > Regards, > Lohit > > On Apr 30, 2018, 5:57 PM -0400, Simon Thompson (IT Research Support) > <[email protected]>, wrote: > You have been able to do this for some time, though I think it's only just > supported. > > We've been exporting remote mounts since CES was added. > > At some point we've had two storage clusters supplying data and at least 3 > remote file-systems exported over NFS and SMB. > > One thing to watch, be careful if your CES root is on a remote fs, as if that > goes away, so do all CES exports. We do have CES root on a remote fs and it > works, just be aware... > > Simon > ________________________________________ > From: [email protected] > [[email protected]] on behalf of > [email protected] [[email protected]] > Sent: 30 April 2018 22:11 > To: gpfsug main discussion list > Subject: [gpfsug-discuss] Spectrum Scale CES and remote file system mounts > > Hello All, > > I read from the below link, that it is now possible to export remote mounts > over NFS/SMB. > > https://www.ibm.com/support/knowledgecenter/en/STXKQY_5.0.0/com.ibm.spectrum.scale.v5r00.doc/bl1adv_protocoloverremoteclu.htm > > I am thinking of using a single CES protocol cluster, with remote mounts from > 3 storage clusters. > May i know, if i will be able to export the 3 remote mounts(from 3 storage > clusters) over NFS/SMB from a single CES protocol cluster? > > Because according to the limitations as mentioned in the below link: > > https://www.ibm.com/support/knowledgecenter/STXKQY_5.0.0/com.ibm.spectrum.scale.v5r00.doc/bl1adv_limitationofprotocolonRMT.htm > > It says “You can configure one storage cluster and up to five protocol > clusters (current limit).” > > > Regards, > Lohit > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss > > > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss
