[
https://issues.apache.org/jira/browse/SPARK-2421?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14057820#comment-14057820
]
Bertrand Dechoux commented on SPARK-2421:
-----------------------------------------
Actually, Hadoop doesn't even require keys to be Writable. It is only the
default configuration that requires it so. The truth is that Hadoop has a kind
of pluggable Serialisation strategy :
http://hadoop.apache.org/docs/r2.3.0/api/org/apache/hadoop/io/serializer/package-summary.html
So ideally, Spark should be as flexible.
> Spark should treat writable as serializable for keys
> ----------------------------------------------------
>
> Key: SPARK-2421
> URL: https://issues.apache.org/jira/browse/SPARK-2421
> Project: Spark
> Issue Type: Improvement
> Components: Input/Output, Java API
> Affects Versions: 1.0.0
> Reporter: Xuefu Zhang
>
> It seems that Spark requires the key be serializable (class implement
> Serializable interface). In Hadoop world, Writable interface is used for the
> same purpose. A lot of existing classes, while writable, are not considered
> by Spark as Serializable. It would be nice if Spark can treate Writable as
> serializable and automatically serialize and de-serialize these classes using
> writable interface.
> This is identified in HIVE-7279, but its benefits are seen global.
--
This message was sent by Atlassian JIRA
(v6.2#6252)