(I'm slowly catching up on a backlog of e-mail, sorry for the delayed
reply).
Thanks, Sven. I recognize the complexity and appreciate your explanation.
In my mind I had envisioned either the block integrity information being
stored as a new metadata structure or stored leveraging T10-DIX/DIF
(perhaps configurable on a per-pool basis) to pass the checksums down to
the RAID controller.
I would quite like to run GNR as software on generic hardware and in
fact voted, along with 26 other customers, on an RFE
(https://www.ibm.com/developerworks/rfe/execute?use_case=viewRfe&CR_ID=95090)
requesting this but the request was declined. I think customers spoke
pretty loudly there and IBM gave it the kibosh.
-Aaron
On 06/30/2017 02:25 PM, Sven Oehme wrote:
end-to-end data integrity is very important and the reason it hasn't
been done in Scale is not because its not important, its because its
very hard to do without impacting performance in a very dramatic way.
imagine your raid controller blocksize is 1mb and your filesystem
blocksize is 1MB . if your application does a 1 MB write this ends up
being a perfect full block , full track de-stage to your raid layer
and everything works fine and fast. as soon as you add checksum
support you need to add data somehow into this, means your 1MB is no
longer 1 MB but 1 MB+checksum.
to store this additional data you have multiple options, inline ,
outside the data block or some combination ,the net is either you need
to do more physical i/o's to different places to get both the data and
the corresponding checksum or your per block on disc structure becomes
bigger than than what your application reads/or writes, both put
massive burden on the Storage layer as e.g. a 1 MB write will now,
even the blocks are all aligned from the application down to the raid
layer, cause a read/modify/write on the raid layer as the data is
bigger than the physical track size.
so to get end-to-end checksum in Scale outside of ESS the best way is
to get GNR as SW to run on generic HW, this is what people should vote
for as RFE if they need that functionality. beside end-to-end
checksums you get read/write cache and acceleration , fast rebuild and
many other goodies as a added bonus.
Sven
On Fri, Jun 30, 2017 at 10:53 AM Aaron Knister
<[email protected] <mailto:[email protected]>> wrote:
In fact the answer was quite literally "no":
https://www.ibm.com/developerworks/rfe/execute?use_case=viewRfe&CR_ID=84523
(the RFE was declined and the answer was that the "function is already
available in GNR environments").
Regarding GNR, see this RFE request
https://www.ibm.com/developerworks/rfe/execute?use_case=viewRfe&CR_ID=95090
requesting the use of GNR outside of an ESS/GSS environment. It's
interesting to note this is the highest voted Public RFE for GPFS
that I
can see, at least. It too was declined.
-Aaron
On 6/30/17 1:41 PM, Aaron Knister wrote:
> Thanks Olaf, that's good to know (and is kind of what I
suspected). I've
> requested a number of times this capability for those of us who
can't
> use or aren't using GNR and the answer is effectively "no". This
> response is curious to me because I'm sure IBM doesn't believe
that data
> integrity is only important and of value to customers who
purchase their
> hardware *and* software.
>
> -Aaron
>
> On Fri, Jun 30, 2017 at 1:37 PM, Olaf Weiser
<[email protected] <mailto:[email protected]>
> <mailto:[email protected] <mailto:[email protected]>>>
wrote:
>
> yes.. in case of GNR (GPFS native raid) .. we do end-to-end
> check-summing ... client --> server --> downToDisk
> GNR writes down a chksum to disk (to all pdisks /all "raid"
segments
> ) so that dropped writes can be detected as well as miss-done
> writes (bit flips..)
>
>
>
> From: Aaron Knister <[email protected]
<mailto:[email protected]>
> <mailto:[email protected]
<mailto:[email protected]>>>
> To: gpfsug main discussion list
<[email protected]
<mailto:[email protected]>
> <mailto:[email protected]
<mailto:[email protected]>>>
> Date: 06/30/2017 07:15 PM
> Subject: [gpfsug-discuss] Fwd: FLASH: IBM Spectrum Scale (GPFS):
> RDMA-enabled network adapter failure on the NSD server may
result in
> file IO error (2017.06.30)
> Sent by: [email protected]
<mailto:[email protected]>
> <mailto:[email protected]
<mailto:[email protected]>>
>
------------------------------------------------------------------------
>
>
>
> I'm curious to know why this doesn't affect GSS/ESS? Is it a
feature of
> the additional check-summing done on those platforms?
>
>
> -------- Forwarded Message --------
> Subject: FLASH: IBM Spectrum Scale (GPFS): RDMA-enabled network
> adapter
> failure on the NSD server may result in file IO error
(2017.06.30)
> Date: Fri, 30 Jun 2017 14:19:02 +0000
> From: IBM My Notifications
> <[email protected]
<mailto:[email protected]>
<mailto:[email protected]
<mailto:[email protected]>>>
> To: [email protected]
<mailto:[email protected]> <mailto:[email protected]
<mailto:[email protected]>>
>
>
>
>
> My Notifications for Storage - 30 Jun 2017
>
> Dear Subscriber ([email protected]
<mailto:[email protected]>
> <mailto:[email protected]
<mailto:[email protected]>>),
>
> Here are your updates from IBM My Notifications.
>
> Your support Notifications display in English by default.
Machine
> translation based on your IBM profile
> language setting is added if you specify this option in My
defaults
> within My Notifications.
> (Note: Not all languages are available at this time, and the
English
> version always takes precedence
> over the machine translated version.)
>
>
------------------------------------------------------------------------------
> 1. IBM Spectrum Scale
>
> - TITLE: IBM Spectrum Scale (GPFS): RDMA-enabled network adapter
> failure
> on the NSD server may result in file IO error
> - URL:
>
http://www.ibm.com/support/docview.wss?uid=ssg1S1010233&myns=s033&mynp=OCSTXKQY&mynp=OCSWJ00&mync=E&cm_sp=s033-_-OCSTXKQY-OCSWJ00-_-E
>
<http://www.ibm.com/support/docview.wss?uid=ssg1S1010233&myns=s033&mynp=OCSTXKQY&mynp=OCSWJ00&mync=E&cm_sp=s033-_-OCSTXKQY-OCSWJ00-_-E>
> - ABSTRACT: IBM has identified an issue with all IBM GPFS
and IBM
> Spectrum Scale versions where the NSD server is enabled to
use RDMA for
> file IO and the storage used in your GPFS cluster accessed
via NSD
> servers (not fully SAN accessible) includes anything other
than IBM
> Elastic Storage Server (ESS) or GPFS Storage Server (GSS);
under these
> conditions, when the RDMA-enabled network adapter fails, the
issue may
> result in undetected data corruption for file write or read
operations.
>
>
------------------------------------------------------------------------------
> Manage your My Notifications subscriptions, or send
questions and
> comments.
> - Subscribe or Unsubscribe -
> https://www.ibm.com/support/mynotifications
> <https://www.ibm.com/support/mynotifications>
> - Feedback -
>
https://www-01.ibm.com/support/feedback/techFeedbackCardContentMyNotifications.html
>
<https://www-01.ibm.com/support/feedback/techFeedbackCardContentMyNotifications.html>
>
> - Follow us on Twitter - https://twitter.com/IBMStorageSupt
> <https://twitter.com/IBMStorageSupt>
>
>
>
> To ensure proper delivery please add
[email protected] <mailto:[email protected]>
> <mailto:[email protected]
<mailto:[email protected]>> to
> your address book.
> You received this email because you are subscribed to IBM My
> Notifications as:
> [email protected] <mailto:[email protected]>
<mailto:[email protected] <mailto:[email protected]>>
>
> Please do not reply to this message as it is generated by an
automated
> service machine.
>
> (C) International Business Machines Corporation 2017. All rights
> reserved.
> _______________________________________________
> gpfsug-discuss mailing list
> gpfsug-discuss at spectrumscale.org
<http://spectrumscale.org> <http://spectrumscale.org>
> http://gpfsug.org/mailman/listinfo/gpfsug-discuss
> <http://gpfsug.org/mailman/listinfo/gpfsug-discuss>
>
>
>
>
>
> _______________________________________________
> gpfsug-discuss mailing list
> gpfsug-discuss at spectrumscale.org
<http://spectrumscale.org> <http://spectrumscale.org>
> http://gpfsug.org/mailman/listinfo/gpfsug-discuss
> <http://gpfsug.org/mailman/listinfo/gpfsug-discuss>
>
>
>
>
> _______________________________________________
> gpfsug-discuss mailing list
> gpfsug-discuss at spectrumscale.org <http://spectrumscale.org>
> http://gpfsug.org/mailman/listinfo/gpfsug-discuss
>
--
Aaron Knister
NASA Center for Climate Simulation (Code 606.2)
Goddard Space Flight Center
(301) 286-2776 <tel:%28301%29%20286-2776>
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org <http://spectrumscale.org>
http://gpfsug.org/mailman/listinfo/gpfsug-discuss
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss