Hello Stephen, There are three licensing models for Spectrum Scale | GPFS:
Server FPO Client I think the thing you might be missing is the associated cost per function. Regards, Ken Hill Technical Sales Specialist | Software Defined Solution Sales IBM Systems Phone: 1-540-207-7270 E-mail: [email protected] 2300 Dulles Station Blvd Herndon, VA 20171-6133 United States From: Stephen Ulmer <[email protected]> To: gpfsug main discussion list <[email protected]> Date: 11/30/2016 09:46 PM Subject: Re: [gpfsug-discuss] Strategies - servers with local SAS disks Sent by: [email protected] I don’t understand what FPO provides here that mirroring doesn’t: You can still use failure domains — one for each node. Both still have redundancy for the data; you can lose a disk or a node. The data has to be re-striped in the event of a disk failure — no matter what. Also, the FPO license doesn’t allow for regular clients to access the data -- only server and FPO nodes. What am I missing? Liberty, -- Stephen On Nov 30, 2016, at 3:51 PM, Andrew Beattie <[email protected]> wrote: Bob, If your not going to use integrated Raid controllers in the servers, then FPO would seem to be the most resilient scenario. yes it has its own overheads, but with that many drives to manage, a JOBD architecture and manual restriping doesn't sound like fun If you are going down the path of integrated raid controllers then any form of distributed raid is probably the best scenario, Raid 6 obviously. How many Nodes are you planning on building? The more nodes the more value FPO is likely to bring as you can be more specific in how the data is written to the nodes. Andrew Beattie Software Defined Storage - IT Specialist Phone: 614-2133-7927 E-mail: [email protected] ----- Original message ----- From: "Oesterlin, Robert" <[email protected]> Sent by: [email protected] To: gpfsug main discussion list <[email protected]> Cc: Subject: [gpfsug-discuss] Strategies - servers with local SAS disks Date: Thu, Dec 1, 2016 12:34 AM Looking for feedback/strategies in setting up several GPFS servers with local SAS. They would all be part of the same file system. The systems are all similar in configuration - 70 4TB drives. Options I’m considering: - Create RAID arrays of the disks on each server (worried about the RAID rebuild time when a drive fails with 4, 6, 8TB drives) - No RAID with 2 replicas, single drive per NSD. When a drive fails, recreate the NSD – but then I need to fix up the data replication via restripe - FPO – with multiple failure groups - letting the system manage replica placement and then have GPFS due the restripe on disk failure automatically Comments or other ideas welcome. Bob Oesterlin Sr Principal Storage Engineer, Nuance 507-269-0413 _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss
