[ 
https://issues.apache.org/jira/browse/HADOOP-1986?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12572180#action_12572180
 ] 

Doug Cutting commented on HADOOP-1986:
--------------------------------------

I spoke with Tom over the weekend and he reported that, on micro benchmarks, 
the performance impact was only a few percent.  It is counter-intuitive that 
this would balloon to 10% on a macro benchmark.  The above results are not very 
consistent either: sort on 100 nodes was 20% faster with the patch, while 
randomwriter was 100% slower with the patch, both of which show little 
correlation with the 20 and 500 node benchmarks.  It seems possible that, 
rather than a consistent slowdown, we're just seeing noise that should not be 
used to judge this patch.


> Add support for a general serialization mechanism for Map Reduce
> ----------------------------------------------------------------
>
>                 Key: HADOOP-1986
>                 URL: https://issues.apache.org/jira/browse/HADOOP-1986
>             Project: Hadoop Core
>          Issue Type: New Feature
>          Components: mapred
>            Reporter: Tom White
>            Assignee: Tom White
>             Fix For: 0.17.0
>
>         Attachments: hadoop-serializer-v2.tar.gz, 
> SequenceFileWriterBenchmark.java, SerializableWritable.java, 
> serializer-v1.patch, serializer-v2.patch, serializer-v3.patch, 
> serializer-v4.patch, serializer-v5.patch
>
>
> Currently Map Reduce programs have to use WritableComparable-Writable 
> key-value pairs. While it's possible to write Writable wrappers for other 
> serialization frameworks (such as Thrift), this is not very convenient: it 
> would be nicer to be able to use arbitrary types directly, without explicit 
> wrapping and unwrapping.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to