Thanks Simon. Currently, we are thinking of using the same remote filesystem for both NFS/SMB exports. I do have a related question with respect to SMB and AD integration on user-defined authentication. I have seen a past discussion from you on the usergroup regarding a similar integration, but i am trying a different setup. Will send an email with the related subject.
Thanks, Lohit On May 3, 2018, 1:30 PM -0400, Simon Thompson (IT Research Support) <[email protected]>, wrote: > Yes we do this when we really really need to take a remote FS offline, which > we try at all costs to avoid unless we have a maintenance window. > > Note if you only export via SMB, then you don’t have the same effect (unless > something has changed recently) > > Simon > > From: <[email protected]> on behalf of > "[email protected]" <[email protected]> > Reply-To: "[email protected]" > <[email protected]> > Date: Thursday, 3 May 2018 at 15:41 > To: "[email protected]" <[email protected]> > Subject: Re: [gpfsug-discuss] Spectrum Scale CES and remote file system mounts > > Thanks Mathiaz, > Yes i do understand the concern, that if one of the remote file systems go > down abruptly - the others will go down too. > > However, i suppose we could bring down one of the filesystems before a > planned downtime? > For example, by unexporting the filesystems on NFS/SMB before the downtime? > > I might not want to be in a situation, where i have to bring down all the > remote filesystems because of planned downtime of one of the remote clusters. > > Regards, > Lohit > > On May 3, 2018, 7:41 AM -0400, Mathias Dietz <[email protected]>, wrote: > > > Hi Lohit, > > > > >I am thinking of using a single CES protocol cluster, with remote mounts > > >from 3 storage clusters. > > Technically this should work fine (assuming all 3 clusters use the same > > uids/guids). However this has not been tested in our Test lab. > > > > > > >One thing to watch, be careful if your CES root is on a remote fs, as if > > >that goes away, so do all CES exports. > > Not only the ces root file system is a concern, the whole CES cluster will > > go down if any remote file systems with NFS exports is not available. > > e.g. if remote cluster 1 is not available, the CES cluster will unmount the > > corresponding file system which will lead to a NFS failure on all CES nodes. > > > > > > Mit freundlichen Grüßen / Kind regards > > > > Mathias Dietz > > > > Spectrum Scale Development - Release Lead Architect (4.2.x) > > Spectrum Scale RAS Architect > > --------------------------------------------------------------------------- > > IBM Deutschland > > Am Weiher 24 > > 65451 Kelsterbach > > Phone: +49 70342744105 > > Mobile: +49-15152801035 > > E-Mail: [email protected] > > ----------------------------------------------------------------------------- > > IBM Deutschland Research & Development GmbH > > Vorsitzender des Aufsichtsrats: Martina Koederitz, Geschäftsführung: Dirk > > WittkoppSitz der Gesellschaft: Böblingen / Registergericht: Amtsgericht > > Stuttgart, HRB 243294 > > > > > > > > From: [email protected] > > To: gpfsug main discussion list <[email protected]> > > Date: 01/05/2018 16:34 > > Subject: Re: [gpfsug-discuss] Spectrum Scale CES and remote file > > system mounts > > Sent by: [email protected] > > > > > > > > Thanks Simon. > > I will make sure i am careful about the CES root and test nfs exporting > > more than 2 remote file systems. > > > > Regards, > > Lohit > > > > On Apr 30, 2018, 5:57 PM -0400, Simon Thompson (IT Research Support) > > <[email protected]>, wrote: > > You have been able to do this for some time, though I think it's only just > > supported. > > > > We've been exporting remote mounts since CES was added. > > > > At some point we've had two storage clusters supplying data and at least 3 > > remote file-systems exported over NFS and SMB. > > > > One thing to watch, be careful if your CES root is on a remote fs, as if > > that goes away, so do all CES exports. We do have CES root on a remote fs > > and it works, just be aware... > > > > Simon > > ________________________________________ > > From: [email protected] > > [[email protected]] on behalf of > > [email protected] [[email protected]] > > Sent: 30 April 2018 22:11 > > To: gpfsug main discussion list > > Subject: [gpfsug-discuss] Spectrum Scale CES and remote file system mounts > > > > Hello All, > > > > I read from the below link, that it is now possible to export remote mounts > > over NFS/SMB. > > > > https://www.ibm.com/support/knowledgecenter/en/STXKQY_5.0.0/com.ibm.spectrum.scale.v5r00.doc/bl1adv_protocoloverremoteclu.htm > > > > I am thinking of using a single CES protocol cluster, with remote mounts > > from 3 storage clusters. > > May i know, if i will be able to export the 3 remote mounts(from 3 storage > > clusters) over NFS/SMB from a single CES protocol cluster? > > > > Because according to the limitations as mentioned in the below link: > > > > https://www.ibm.com/support/knowledgecenter/STXKQY_5.0.0/com.ibm.spectrum.scale.v5r00.doc/bl1adv_limitationofprotocolonRMT.htm > > > > It says “You can configure one storage cluster and up to five protocol > > clusters (current limit).” > > > > > > Regards, > > Lohit > > _______________________________________________ > > gpfsug-discuss mailing list > > gpfsug-discuss at spectrumscale.org > > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_______________________________________________ > > gpfsug-discuss mailing list > > gpfsug-discuss at spectrumscale.org > > http://gpfsug.org/mailman/listinfo/gpfsug-discuss > > > > > > > > _______________________________________________ > > gpfsug-discuss mailing list > > gpfsug-discuss at spectrumscale.org > > http://gpfsug.org/mailman/listinfo/gpfsug-discuss > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss
