[
https://issues.apache.org/jira/browse/HDFS-10285?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16345489#comment-16345489
]
Daryn Sharp commented on HDFS-10285:
------------------------------------
More comments, mostly focusing on the DN side. There appear to be more serious
issues here.
*FSDirSatisfyStoragePolicyOp*
{{satisfyStoragePolicy}} errors if the xattr is already present. Should this
be a no-op? A client re-requesting a storage policy correction probably
shouldn't fail.
{{unprotectedSatisfyStoragePolicy}} is called prior to xattr updates, which
calls {{addSPSPathId}}. To avoid race conditions or inconsistent state if the
xattr fails, should call {{addSPSPathId}} after xattrs are successfully updated.
{{inodeHasSatisfyXAttr}} calls {{getXAttrFeature}} then immediately shorts out
if the inode isn't a file. Should do file check first to avoid unnecessary
computation.
In general, not fond of unnecessarily guava. Instead of
{{newArrayListWithCapacity}} + {{add}}, standard {{Arrays.asList(item)}} is
more succinct.
*FSDirStatAndListOp*
Not sure why javadoc was changed to add needLocation. It's already present and
now doubled up.
{{BlockStorageMovementCommand/BlocksStorageMoveAttemptFinished}}
Again, not sure that a new DN command is necessary, and why does it
specifically report back successful moves instead of relying on IBRs? I would
actually expect the DN to be completely ignorant of a SPS move vs any other
move.
*StoragePolicySatisfyWorker*
bq. // TODO: Needs to manage the number of concurrent moves per DataNode.
Shouldn't this be fixed now?
{{DFS_MOVER_MOVERTHREADS_DEFAULT}} is *1000 per DN*? If the DN is concurrently
doing 1000 moves, it's not in a good state, disk io is probably saturated, and
this will only make it much worse. 10 is probably more than sufficient.
The executor uses a {{CallerRunsPolicy}}? If the thread pool is exhausted, WHY
should the heartbeat thread process the operation? That's a terrible idea.
I was under the impression the DN did an optimized movement between its
storages? Does it actually make a connection to itself? If yes, when the DN
xceiver pool is very busy, the mover thread pool will exhaust and cause the
heartbeat thread to block.
*BlockStorageMovementTracker*
Many data structures are riddled with non-threadsafe race conditions and risk
of CMEs.
Ex. The {{moverTaskFutures}} map. Adding new blocks and/or adding to a block's
list of futures is synchronized. However the run loop does an unsynchronized
block get, unsynchronized future remove, unsynchronized isEmpty, possibly
another unsynchronized get, only then does it do a synchronized remove of the
block. The whole chunk of code should be synchronized.
Is the problematic {{moverTaskFutures}} even needed? It's aggregating futures
per-block for seemingly no reason. Why track all the futures at all instead of
just relying on the completion service? As best I can tell:
* It's only used to determine if a future from the completion service should be
ignored during shutdown. Shutdown sets the running boolean to false and clears
the entire datastructure so why not use the {{running}} boolean like a check
just a little further down?
* As synchronization to sleep up to 2 seconds before performing a blocking
{{moverCompletionService.take}}, but only when it thinks there are no active
futures. I'll ignore the missed notify race that the bounded wait masks, but
the real question is why not just do the blocking take?
Why all the complexity? Am I missing something?
*BlocksMovementsStatusHandler*
Suffers same type of thread safety issues as {{StoragePolicySatisfyWorker}}.
Ex. {{blockIdVsMovementStatus}} is inconsistent synchronized. Does synchronize
to return an unmodifiable list which sadly does nothing to protect the caller
from CME.
{{handle}} is iterating over a non-thread safe list.
> Storage Policy Satisfier in Namenode
> ------------------------------------
>
> Key: HDFS-10285
> URL: https://issues.apache.org/jira/browse/HDFS-10285
> Project: Hadoop HDFS
> Issue Type: New Feature
> Components: datanode, namenode
> Affects Versions: HDFS-10285
> Reporter: Uma Maheswara Rao G
> Assignee: Uma Maheswara Rao G
> Priority: Major
> Attachments: HDFS-10285-consolidated-merge-patch-00.patch,
> HDFS-10285-consolidated-merge-patch-01.patch,
> HDFS-10285-consolidated-merge-patch-02.patch,
> HDFS-10285-consolidated-merge-patch-03.patch,
> HDFS-10285-consolidated-merge-patch-04.patch,
> HDFS-10285-consolidated-merge-patch-05.patch,
> HDFS-SPS-TestReport-20170708.pdf, SPS Modularization.pdf,
> Storage-Policy-Satisfier-in-HDFS-June-20-2017.pdf,
> Storage-Policy-Satisfier-in-HDFS-May10.pdf,
> Storage-Policy-Satisfier-in-HDFS-Oct-26-2017.pdf
>
>
> Heterogeneous storage in HDFS introduced the concept of storage policy. These
> policies can be set on directory/file to specify the user preference, where
> to store the physical block. When user set the storage policy before writing
> data, then the blocks could take advantage of storage policy preferences and
> stores physical block accordingly.
> If user set the storage policy after writing and completing the file, then
> the blocks would have been written with default storage policy (nothing but
> DISK). User has to run the ‘Mover tool’ explicitly by specifying all such
> file names as a list. In some distributed system scenarios (ex: HBase) it
> would be difficult to collect all the files and run the tool as different
> nodes can write files separately and file can have different paths.
> Another scenarios is, when user rename the files from one effected storage
> policy file (inherited policy from parent directory) to another storage
> policy effected directory, it will not copy inherited storage policy from
> source. So it will take effect from destination file/dir parent storage
> policy. This rename operation is just a metadata change in Namenode. The
> physical blocks still remain with source storage policy.
> So, Tracking all such business logic based file names could be difficult for
> admins from distributed nodes(ex: region servers) and running the Mover tool.
> Here the proposal is to provide an API from Namenode itself for trigger the
> storage policy satisfaction. A Daemon thread inside Namenode should track
> such calls and process to DN as movement commands.
> Will post the detailed design thoughts document soon.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]