We don't allow mixing of different licensing models (i.e. socket and capacity) within a single cluster*. As we worked through the implications, we realized it would be just too complicated to determine how to license any non-NSD nodes (management, CES, clients, etc.). In the socket model they are chargeable, in the capacity model they are not, and while we could have made up some rules, they would have added even more complexity to Scale licensing.
This in turn is why we "grandfathered in" those customers already on Advanced Edition, so that they don't have to convert existing clusters to the new metric unless or until they want to. They can continue to buy Advanced Edition. The other thing we wanted to do with the capacity metric was to make the licensing more friendly to architectural best practices or design choices. So now you can have whatever management, gateway, etc. servers you need without paying for additional server licenses. In particular, client-only clusters cost nothing, and you don't have to keep track of clients if you have a virtual environment where clients come and go rapidly. I'm always happy to answer other questions about licensing. regards, Carl Zetie *OK, there is one exception involving future ESS models and existing clusters. If this is you, please have a conversation with your account team. Carl Zetie Program Director, OM for Spectrum Scale, IBM (540) 882 9353 ][ 15750 Brookhill Ct, Waterford VA 20197 ca...@us.ibm.com From: gpfsug-discuss-requ...@spectrumscale.org To: gpfsug-discuss@spectrumscale.org Date: 12/07/2016 09:59 AM Subject: gpfsug-discuss Digest, Vol 59, Issue 20 Sent by: gpfsug-discuss-boun...@spectrumscale.org Send gpfsug-discuss mailing list submissions to gpfsug-discuss@spectrumscale.org To subscribe or unsubscribe via the World Wide Web, visit http://gpfsug.org/mailman/listinfo/gpfsug-discuss or, via email, send a message with subject or body 'help' to gpfsug-discuss-requ...@spectrumscale.org You can reach the person managing the list at gpfsug-discuss-ow...@spectrumscale.org When replying, please edit your Subject line so it is more specific than "Re: Contents of gpfsug-discuss digest..." Today's Topics: 1. Re: Any experience running native GPFS 4.2.1 on Xeon Phi node booted with Centos 7.3? (Felipe Knop) 2. Re: Any experience running native GPFS 4.2.1 on Xeon Phi node booted with Centos 7.3? (David D. Johnson) 3. Re: Strategies - servers with local SAS disks (Simon Thompson (Research Computing - IT Services)) ---------------------------------------------------------------------- Message: 1 Date: Wed, 7 Dec 2016 09:37:15 -0500 From: "Felipe Knop" <k...@us.ibm.com> To: gpfsug main discussion list <gpfsug-discuss@spectrumscale.org> Subject: Re: [gpfsug-discuss] Any experience running native GPFS 4.2.1 on Xeon Phi node booted with Centos 7.3? Message-ID: <of76ee3fb4.4db6d687-on85258082.00502817-85258082.00505...@notes.na.collabserv.com> Content-Type: text/plain; charset="us-ascii" All, The SMAP issue has been addressed in GPFS in 4.2.1.1. See http://www.ibm.com/support/knowledgecenter/STXKQY/gpfsclustersfaq.html Q2.4. Felipe ---- Felipe Knop k...@us.ibm.com GPFS Development and Security IBM Systems IBM Building 008 2455 South Rd, Poughkeepsie, NY 12601 (845) 433-9314 T/L 293-9314 From: Aaron Knister <aaron.knis...@gmail.com> To: gpfsug main discussion list <gpfsug-discuss@spectrumscale.org> Date: 12/07/2016 09:25 AM Subject: Re: [gpfsug-discuss] Any experience running native GPFS 4.2.1 on Xeon Phi node booted with Centos 7.3? Sent by: gpfsug-discuss-boun...@spectrumscale.org I don't know if this applies her but I seem to recall an issue with CentOS 7 (newer 3.X and on kernels), Broadwell processors and GPFS where GPFS upset SMAP and would eventually get the node expelled. I think this may be fixed in newer GPFS releases but the fix is to boot the kernel with the nosmap parameter. Might be worth a try. I'm not clear on whether SMAP is supported by the Xeon Phi's. -Aaron On Wed, Dec 7, 2016 at 5:34 AM <david_john...@brown.edu> wrote: IBM says it should work ok, we are not so sure. We had node expels that stopped when we turned off gpfs on that node. Has anyone had better luck? -- ddj Dave Johnson _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss -------------- next part -------------- An HTML attachment was scrubbed... URL: < http://gpfsug.org/pipermail/gpfsug-discuss/attachments/20161207/48aa0319/attachment-0001.html > ------------------------------ Message: 2 Date: Wed, 7 Dec 2016 09:47:46 -0500 From: "David D. Johnson" <david_john...@brown.edu> To: gpfsug main discussion list <gpfsug-discuss@spectrumscale.org> Subject: Re: [gpfsug-discuss] Any experience running native GPFS 4.2.1 on Xeon Phi node booted with Centos 7.3? Message-ID: <5fbac3ae-39f2-453d-8a9d-5fde90bad...@brown.edu> Content-Type: text/plain; charset="utf-8" Yes, we saw the SMAP issue on earlier releases, added the kernel command line option to disable it. That is not the issue for this node. The Phi processors do not support that cpu feature. ? ddj > On Dec 7, 2016, at 9:37 AM, Felipe Knop <k...@us.ibm.com> wrote: > > All, > > The SMAP issue has been addressed in GPFS in 4.2.1.1. > > See http://www.ibm.com/support/knowledgecenter/STXKQY/gpfsclustersfaq.html < http://www.ibm.com/support/knowledgecenter/STXKQY/gpfsclustersfaq.html> > > Q2.4. > > Felipe > > ---- > Felipe Knop k...@us.ibm.com > GPFS Development and Security > IBM Systems > IBM Building 008 > 2455 South Rd, Poughkeepsie, NY 12601 > (845) 433-9314 T/L 293-9314 > > > > > > From: Aaron Knister <aaron.knis...@gmail.com> > To: gpfsug main discussion list <gpfsug-discuss@spectrumscale.org> > Date: 12/07/2016 09:25 AM > Subject: Re: [gpfsug-discuss] Any experience running native GPFS 4.2.1 on Xeon Phi node booted with Centos 7.3? > Sent by: gpfsug-discuss-boun...@spectrumscale.org > > > > I don't know if this applies her but I seem to recall an issue with CentOS 7 (newer 3.X and on kernels), Broadwell processors and GPFS where GPFS upset SMAP and would eventually get the node expelled. I think this may be fixed in newer GPFS releases but the fix is to boot the kernel with the nosmap parameter. Might be worth a try. I'm not clear on whether SMAP is supported by the Xeon Phi's. > > -Aaron > > On Wed, Dec 7, 2016 at 5:34 AM <david_john...@brown.edu < mailto:david_john...@brown.edu>> wrote: > IBM says it should work ok, we are not so sure. We had node expels that stopped when we turned off gpfs on that node. Has anyone had better luck? > > -- ddj > Dave Johnson > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org <http://spectrumscale.org/> > http://gpfsug.org/mailman/listinfo/gpfsug-discuss < http://gpfsug.org/mailman/listinfo/gpfsug-discuss >_______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss < http://gpfsug.org/mailman/listinfo/gpfsug-discuss> > > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss -------------- next part -------------- An HTML attachment was scrubbed... URL: < http://gpfsug.org/pipermail/gpfsug-discuss/attachments/20161207/92819f21/attachment-0001.html > ------------------------------ Message: 3 Date: Wed, 7 Dec 2016 14:58:39 +0000 From: "Simon Thompson (Research Computing - IT Services)" <s.j.thomp...@bham.ac.uk> To: gpfsug main discussion list <gpfsug-discuss@spectrumscale.org> Subject: Re: [gpfsug-discuss] Strategies - servers with local SAS disks Message-ID: <d46dd3a5.33e50%s.j.thomp...@bham.ac.uk> Content-Type: text/plain; charset="us-ascii" I was going to ask about this, I recall it being mentioned about "grandfathering" and also having mixed deployments. Would that mean you could per TB license one set of NSD servers (hosting only 1 FS) that co-existed in a cluster with other traditionally licensed systems? I would see having NSDs with different license models hosting the same FS being problematic, but if it were a different file-system? Simon From: <gpfsug-discuss-boun...@spectrumscale.org< mailto:gpfsug-discuss-boun...@spectrumscale.org>> on behalf of Daniel Kidger <daniel.kid...@uk.ibm.com<mailto:daniel.kid...@uk.ibm.com>> Reply-To: "gpfsug-discuss@spectrumscale.org< mailto:gpfsug-discuss@spectrumscale.org>" <gpfsug-discuss@spectrumscale.org<mailto:gpfsug-discuss@spectrumscale.org >> Date: Wednesday, 7 December 2016 at 12:36 To: "gpfsug-discuss@spectrumscale.org< mailto:gpfsug-discuss@spectrumscale.org>" <gpfsug-discuss@spectrumscale.org<mailto:gpfsug-discuss@spectrumscale.org >> Cc: "gpfsug-discuss@spectrumscale.org< mailto:gpfsug-discuss@spectrumscale.org>" <gpfsug-discuss@spectrumscale.org<mailto:gpfsug-discuss@spectrumscale.org >> Subject: Re: [gpfsug-discuss] Strategies - servers with local SAS disks The new volume based licensing option is I agree quite pricey per TB at first sight, but it could make some configuration choice, a lot cheaper than they used to be under the Client:FPO:Server model. -------------- next part -------------- An HTML attachment was scrubbed... URL: < http://gpfsug.org/pipermail/gpfsug-discuss/attachments/20161207/51c1a2ea/attachment.html > ------------------------------ _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss End of gpfsug-discuss Digest, Vol 59, Issue 20 **********************************************
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss