Github user Mironor commented on a diff in the pull request:
https://github.com/apache/spark/pull/16137#discussion_r91015596
--- Diff: core/src/main/scala/org/apache/spark/SparkContext.scala ---
@@ -1144,13 +1218,19 @@ class SparkContext(config: SparkConf) extends
Logging {
}
/**
- * Get an RDD for a Hadoop SequenceFile with given key and value types.
+ * Get an RDD for a Hadoop `SequenceFile` with given key and value types.
*
- * @note Because Hadoop's RecordReader class re-uses the same Writable
object for each
- * record, directly caching the returned RDD or directly passing it to
an aggregation or shuffle
- * operation will create many references to the same object.
- * If you plan to directly cache, sort, or aggregate Hadoop writable
objects, you should first
- * copy them using a `map` function.
+ * @note because Hadoop's `RecordReader` class re-uses the same
`Writable` object for each
--- End diff --
Correct, but
[they](http://www.oracle.com/technetwork/articles/java/index-137868.html) also
contain continuation indentation (they even align parameter description)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]