We had Lab Services do our installs and integrations. Learning curve for them, and we uncovered some deficiencies in the TDA, but it did work.
-- Jonathan Fosburgh Principal Application Systems Analyst IT Operations Storage Team The University of Texas MD Anderson Cancer Center (713) 745-9346 [1553012336789_download] ________________________________ From: [email protected] <[email protected]> on behalf of Prasad Surampudi <[email protected]> Sent: Wednesday, April 3, 2019 2:34:59 PM To: [email protected]; [email protected] Subject: [EXT] Re: [gpfsug-discuss] gpfsug-discuss Digest, Vol 87, Issue 4 WARNING: This email originated from outside of MD Anderson. Please validate the sender's email address before clicking on links or attachments as they may not be safe. Actually, we have a SAS Grid - Scale cluster with V7000 and Flash storage. We also have protocol nodes for SMB access to SAS applications/users. Now, we are planning to gradually move our cluster from V7000/Flash to ESS and retire V7Ks. So, when we grow our filesystem, we are thinking of adding an ESS as an additional block of storage instead of adding another V7000. Definitely we'll keep the ESS Disk Enclosures in a separate GPFS pool in the same filesystem, but can't create a new filesystem as we want to have single name space for our SMB Shares. Also, we'd like keep all our existing compute, protocol, and NSD servers all in the same scale cluster along with ESS IO nodes and EMS. When I looked at ESS commands, I dont see an option of adding ESS nodes to existing cluster like mmaddnode or similar commands. So, just wondering how we could add ESS IO nodes to existing cluster like any other node..is running mmaddnode command on ESS possible? Also, looks like it's against the IBMs recommendation of separating the Storage, Compute and Protocol nodes into their own scale clusters and use cross-cluster filesystem mounts..any comments/suggestions? Prasad Surampudi The ATS Group ________________________________ From: [email protected] <[email protected]> on behalf of [email protected] <[email protected]> Sent: Wednesday, April 3, 2019 2:54 PM To: [email protected] Subject: gpfsug-discuss Digest, Vol 87, Issue 4 Send gpfsug-discuss mailing list submissions to [email protected] To subscribe or unsubscribe via the World Wide Web, visit http://gpfsug.org/mailman/listinfo/gpfsug-discuss or, via email, send a message with subject or body 'help' to [email protected] You can reach the person managing the list at [email protected] When replying, please edit your Subject line so it is more specific than "Re: Contents of gpfsug-discuss digest..." Today's Topics: 1. Re: Adding ESS to existing Scale Cluster (Sanchez, Paul) 2. New ESS install - Network adapter down level (Oesterlin, Robert) 3. Re: New ESS install - Network adapter down level (Jan-Frode Myklebust) 4. Re: New ESS install - Network adapter down level (Stephen R Buchanan) ---------------------------------------------------------------------- Message: 1 Date: Wed, 3 Apr 2019 16:41:32 +0000 From: "Sanchez, Paul" <[email protected]> To: gpfsug main discussion list <[email protected]> Subject: Re: [gpfsug-discuss] Adding ESS to existing Scale Cluster Message-ID: <[email protected]> Content-Type: text/plain; charset="us-ascii" > note though you can't have GNR based vdisks (ESS/DSS-G) in the same storage > pool. At one time there was definitely a warning from IBM in the docs about not mixing big-endian and little-endian GNR in the same cluster/filesystem. But at least since Nov 2017, IBM has published videos showing clusters containing both. (In my opinion, they had to support this because they changed the endian-ness of the ESS from BE to LE.) I don't know about all ancillary components (e.g. GUI) but as for Scale itself, I can confirm that filesystems can contain NSDs which are provided by ESS(BE), ESS(LE), GSS, and DSS in all combinations, along with SAN storage based NSD servers. We typically do rolling upgrades of GNR building blocks by adding blocks to an existing cluster, emptying and removing the existing blocks, upgrading those in isolation, then repeating with the next cluster. As a result, we have had every combination in play at some point in time. Care just needs to be taken with nodeclass naming and mmchconfig parameters. (We derive the correct params for each new building block from its final config after upgrading/testing it in isolation.) -Paul -----Original Message----- From: [email protected] <[email protected]> On Behalf Of Simon Thompson Sent: Wednesday, April 3, 2019 12:18 PM To: gpfsug main discussion list <[email protected]> Subject: Re: [gpfsug-discuss] Adding ESS to existing Scale Cluster We have DSS-G (Lenovo equivalent) in the same cluster as other SAN/IB storage (IBM, DDN). But we don't have them in the same file-system. In theory as a different pool it should work, note though you can't have GNR based vdisks (ESS/DSS-G) in the same storage pool. And if you want to move to new block size or v5 variable sunblocks then you are going to have to have a new filesystem and copy data. So it depends what your endgame is really. We just did such a process and one of my colleagues is going to talk about it at the London user group in May. Simon ________________________________________ From: [email protected] [[email protected]] on behalf of [email protected] [[email protected]] Sent: 03 April 2019 17:12 To: [email protected] Subject: [gpfsug-discuss] Adding ESS to existing Scale Cluster We are planning to add an ESS GL6 system to our existing Spectrum Scale cluster. Can the ESS nodes be added to existing scale cluster without changing existing cluster name? Or do we need to create a new scale cluster with ESS and import existing filesystems into the new ESS cluster? Prasad Surampudi Sr. Systems Engineer The ATS Group _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss ------------------------------ Message: 2 Date: Wed, 3 Apr 2019 17:25:54 +0000 From: "Oesterlin, Robert" <[email protected]> To: gpfsug main discussion list <[email protected]> Subject: [gpfsug-discuss] New ESS install - Network adapter down level Message-ID: <[email protected]> Content-Type: text/plain; charset="utf-8" Any insight on what command I need to fix this? It?s the only error I have when running gssinstallcheck. [ERROR] Network adapter MT4115 firmware: found 12.23.1020 expected 12.23.8010, net adapter count: 4 Bob Oesterlin Sr Principal Storage Engineer, Nuance 507-269-0413 -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://gpfsug.org/pipermail/gpfsug-discuss/attachments/20190403/42850e8f/attachment-0001.html> ------------------------------ Message: 3 Date: Wed, 3 Apr 2019 20:11:45 +0200 From: Jan-Frode Myklebust <[email protected]> To: gpfsug main discussion list <[email protected]> Subject: Re: [gpfsug-discuss] New ESS install - Network adapter down level Message-ID: <CAHwPatjx-r7EHoV_PuB1XMGyGgQomz6KtLS=frpjyb-36zj...@mail.gmail.com> Content-Type: text/plain; charset="utf-8" Have you tried: updatenode nodename -P gss_ofed But, is this the known issue listed in the qdg? https://www.ibm.com/support/knowledgecenter/SSYSP8_5.3.2/ess_qdg.pdf -jf ons. 3. apr. 2019 kl. 19:26 skrev Oesterlin, Robert < [email protected]>: > Any insight on what command I need to fix this? It?s the only error I have > when running gssinstallcheck. > > > > [ERROR] Network adapter > https://www.ibm.com/support/knowledgecenter/SSYSP8_5.3.2/ess_qdg.pdf > firmware: found 12.23.1020 expected 12.23.8010, net adapter count: 4 > > > > > > Bob Oesterlin > > Sr Principal Storage Engineer, Nuance > > 507-269-0413 > > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at spectrumscale.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://gpfsug.org/pipermail/gpfsug-discuss/attachments/20190403/fad4ff57/attachment-0001.html> ------------------------------ Message: 4 Date: Wed, 3 Apr 2019 18:54:00 +0000 From: "Stephen R Buchanan" <[email protected]> To: [email protected] Subject: Re: [gpfsug-discuss] New ESS install - Network adapter down level Message-ID: <ofbd2a098d.0085093e-on002583d1.0066d1e2-002583d1.0067d...@notes.na.collabserv.com> Content-Type: text/plain; charset="us-ascii" An HTML attachment was scrubbed... URL: <http://gpfsug.org/pipermail/gpfsug-discuss/attachments/20190403/09e229d1/attachment.html> ------------------------------ _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss End of gpfsug-discuss Digest, Vol 87, Issue 4 ********************************************* The information contained in this e-mail message may be privileged, confidential, and/or protected from disclosure. This e-mail message may contain protected health information (PHI); dissemination of PHI should comply with applicable federal and state laws. If you are not the intended recipient, or an authorized representative of the intended recipient, any further review, disclosure, use, dissemination, distribution, or copying of this message or any attachment (or the information contained therein) is strictly prohibited. If you think that you have received this e-mail message in error, please notify the sender by return e-mail and delete all references to it and its contents from your systems.
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss
