We don't allow mixing of different licensing models (i.e. socket and 
capacity) within a single cluster*. As we worked through the implications, 
we realized it would be just too complicated to determine how to license 
any non-NSD nodes (management, CES, clients, etc.). In the socket model 
they are chargeable, in the capacity model they are not, and while we 
could have made up some rules, they would have added even more complexity 
to Scale licensing.

This in turn is why we "grandfathered in" those customers already on 
Advanced Edition, so that they don't have to convert existing clusters to 
the new metric unless or until they want to. They can continue to buy 
Advanced Edition. 

The other thing we wanted to do with the capacity metric was to make the 
licensing more friendly to architectural best practices or design choices. 
So now you can have whatever management, gateway, etc. servers you need 
without paying for additional server licenses. In particular, client-only 
clusters cost nothing, and you don't have to keep track of clients if you 
have a virtual environment where clients come and go rapidly.

I'm always happy to answer other questions about licensing.

regards,
Carl Zetie

*OK, there is one exception involving future ESS models and existing 
clusters. If this is you, please have a conversation with your account 
team.

Carl Zetie
Program Director, OM for Spectrum Scale, IBM

(540) 882 9353  ][  15750 Brookhill Ct, Waterford VA 20197
ca...@us.ibm.com



From:   gpfsug-discuss-requ...@spectrumscale.org
To:     gpfsug-discuss@spectrumscale.org
Date:   12/07/2016 09:59 AM
Subject:        gpfsug-discuss Digest, Vol 59, Issue 20
Sent by:        gpfsug-discuss-boun...@spectrumscale.org



Send gpfsug-discuss mailing list submissions to
                 gpfsug-discuss@spectrumscale.org

To subscribe or unsubscribe via the World Wide Web, visit
                 http://gpfsug.org/mailman/listinfo/gpfsug-discuss
or, via email, send a message with subject or body 'help' to
                 gpfsug-discuss-requ...@spectrumscale.org

You can reach the person managing the list at
                 gpfsug-discuss-ow...@spectrumscale.org

When replying, please edit your Subject line so it is more specific
than "Re: Contents of gpfsug-discuss digest..."


Today's Topics:

   1. Re: Any experience running native GPFS 4.2.1 on Xeon Phi node
      booted with Centos 7.3? (Felipe Knop)
   2. Re: Any experience running native GPFS 4.2.1 on            Xeon Phi 
node
      booted with Centos 7.3? (David D. Johnson)
   3. Re: Strategies - servers with local SAS disks
      (Simon Thompson (Research Computing - IT Services))


----------------------------------------------------------------------

Message: 1
Date: Wed, 7 Dec 2016 09:37:15 -0500
From: "Felipe Knop" <k...@us.ibm.com>
To: gpfsug main discussion list <gpfsug-discuss@spectrumscale.org>
Subject: Re: [gpfsug-discuss] Any experience running native GPFS 4.2.1
                 on Xeon Phi node booted with Centos 7.3?
Message-ID:
 
<of76ee3fb4.4db6d687-on85258082.00502817-85258082.00505...@notes.na.collabserv.com>
 
Content-Type: text/plain; charset="us-ascii"

All,

The SMAP issue has been addressed in GPFS in 4.2.1.1.

See http://www.ibm.com/support/knowledgecenter/STXKQY/gpfsclustersfaq.html

Q2.4.

  Felipe

----
Felipe Knop                                     k...@us.ibm.com
GPFS Development and Security
IBM Systems
IBM Building 008
2455 South Rd, Poughkeepsie, NY 12601
(845) 433-9314  T/L 293-9314





From:   Aaron Knister <aaron.knis...@gmail.com>
To:     gpfsug main discussion list <gpfsug-discuss@spectrumscale.org>
Date:   12/07/2016 09:25 AM
Subject:        Re: [gpfsug-discuss] Any experience running native GPFS 
4.2.1 on Xeon Phi node booted with Centos 7.3?
Sent by:        gpfsug-discuss-boun...@spectrumscale.org



I don't know if this applies her but I seem to recall an issue with CentOS 

7 (newer 3.X and on kernels), Broadwell processors and GPFS where GPFS 
upset SMAP and would eventually get the node expelled. I think this may be 

fixed in newer GPFS releases but the fix is to boot the kernel with the 
nosmap parameter. Might be worth a try. I'm not clear on whether SMAP is 
supported by the Xeon Phi's. 

-Aaron

On Wed, Dec 7, 2016 at 5:34 AM <david_john...@brown.edu> wrote:
IBM says it should work ok, we are not so sure. We had node expels that 
stopped when we turned off gpfs on that node. Has anyone had better luck?

 -- ddj
Dave Johnson
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss



-------------- next part --------------
An HTML attachment was scrubbed...
URL: <
http://gpfsug.org/pipermail/gpfsug-discuss/attachments/20161207/48aa0319/attachment-0001.html
>

------------------------------

Message: 2
Date: Wed, 7 Dec 2016 09:47:46 -0500
From: "David D. Johnson" <david_john...@brown.edu>
To: gpfsug main discussion list <gpfsug-discuss@spectrumscale.org>
Subject: Re: [gpfsug-discuss] Any experience running native GPFS 4.2.1
                 on              Xeon Phi node booted with Centos 7.3?
Message-ID: <5fbac3ae-39f2-453d-8a9d-5fde90bad...@brown.edu>
Content-Type: text/plain; charset="utf-8"

Yes, we saw the SMAP issue on earlier releases, added the kernel command 
line option to disable it.
That is not the issue for this node. The Phi processors do not support 
that cpu feature.

 ? ddj

> On Dec 7, 2016, at 9:37 AM, Felipe Knop <k...@us.ibm.com> wrote:
> 
> All,
> 
> The SMAP issue has been addressed in GPFS in 4.2.1.1.
> 
> See 
http://www.ibm.com/support/knowledgecenter/STXKQY/gpfsclustersfaq.html <
http://www.ibm.com/support/knowledgecenter/STXKQY/gpfsclustersfaq.html>
> 
> Q2.4.
> 
>   Felipe
> 
> ----
> Felipe Knop                                     k...@us.ibm.com
> GPFS Development and Security
> IBM Systems
> IBM Building 008
> 2455 South Rd, Poughkeepsie, NY 12601
> (845) 433-9314  T/L 293-9314
> 
> 
> 
> 
> 
> From:        Aaron Knister <aaron.knis...@gmail.com>
> To:        gpfsug main discussion list 
<gpfsug-discuss@spectrumscale.org>
> Date:        12/07/2016 09:25 AM
> Subject:        Re: [gpfsug-discuss] Any experience running native GPFS 
4.2.1 on Xeon Phi node booted with Centos 7.3?
> Sent by:        gpfsug-discuss-boun...@spectrumscale.org
> 
> 
> 
> I don't know if this applies her but I seem to recall an issue with 
CentOS 7 (newer 3.X and on kernels), Broadwell processors and GPFS where 
GPFS upset SMAP and would eventually get the node expelled. I think this 
may be fixed in newer GPFS releases but the fix is to boot the kernel with 
the nosmap parameter. Might be worth a try. I'm not clear on whether SMAP 
is supported by the Xeon Phi's. 
> 
> -Aaron
> 
> On Wed, Dec 7, 2016 at 5:34 AM <david_john...@brown.edu <
mailto:david_john...@brown.edu>> wrote:
> IBM says it should work ok, we are not so sure. We had node expels that 
stopped when we turned off gpfs on that node. Has anyone had better luck?
> 
>  -- ddj
> Dave Johnson
> _______________________________________________
> gpfsug-discuss mailing list
> gpfsug-discuss at spectrumscale.org <http://spectrumscale.org/>
> http://gpfsug.org/mailman/listinfo/gpfsug-discuss <
http://gpfsug.org/mailman/listinfo/gpfsug-discuss
>_______________________________________________
> gpfsug-discuss mailing list
> gpfsug-discuss at spectrumscale.org
> http://gpfsug.org/mailman/listinfo/gpfsug-discuss <
http://gpfsug.org/mailman/listinfo/gpfsug-discuss>
> 
> 
> _______________________________________________
> gpfsug-discuss mailing list
> gpfsug-discuss at spectrumscale.org
> http://gpfsug.org/mailman/listinfo/gpfsug-discuss

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <
http://gpfsug.org/pipermail/gpfsug-discuss/attachments/20161207/92819f21/attachment-0001.html
>

------------------------------

Message: 3
Date: Wed, 7 Dec 2016 14:58:39 +0000
From: "Simon Thompson (Research Computing - IT Services)"
                 <s.j.thomp...@bham.ac.uk>
To: gpfsug main discussion list <gpfsug-discuss@spectrumscale.org>
Subject: Re: [gpfsug-discuss] Strategies - servers with local SAS
                 disks
Message-ID: <d46dd3a5.33e50%s.j.thomp...@bham.ac.uk>
Content-Type: text/plain; charset="us-ascii"

I was going to ask about this, I recall it being mentioned about 
"grandfathering" and also having mixed deployments.

Would that mean you could per TB license one set of NSD servers (hosting 
only 1 FS) that co-existed in a cluster with other traditionally licensed 
systems?

I would see having NSDs with different license models hosting the same FS 
being problematic, but if it were a different file-system?

Simon

From: <gpfsug-discuss-boun...@spectrumscale.org<
mailto:gpfsug-discuss-boun...@spectrumscale.org>> on behalf of Daniel 
Kidger <daniel.kid...@uk.ibm.com<mailto:daniel.kid...@uk.ibm.com>>
Reply-To: "gpfsug-discuss@spectrumscale.org<
mailto:gpfsug-discuss@spectrumscale.org>" 
<gpfsug-discuss@spectrumscale.org<mailto:gpfsug-discuss@spectrumscale.org
>>
Date: Wednesday, 7 December 2016 at 12:36
To: "gpfsug-discuss@spectrumscale.org<
mailto:gpfsug-discuss@spectrumscale.org>" 
<gpfsug-discuss@spectrumscale.org<mailto:gpfsug-discuss@spectrumscale.org
>>
Cc: "gpfsug-discuss@spectrumscale.org<
mailto:gpfsug-discuss@spectrumscale.org>" 
<gpfsug-discuss@spectrumscale.org<mailto:gpfsug-discuss@spectrumscale.org
>>
Subject: Re: [gpfsug-discuss] Strategies - servers with local SAS disks

The new volume based licensing option is I agree quite pricey per TB at 
first sight, but it could make some configuration choice, a lot cheaper 
than they used to be under the Client:FPO:Server model.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <
http://gpfsug.org/pipermail/gpfsug-discuss/attachments/20161207/51c1a2ea/attachment.html
>

------------------------------

_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss


End of gpfsug-discuss Digest, Vol 59, Issue 20
**********************************************





_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss

Reply via email to