[ 
https://issues.apache.org/jira/browse/FLINK-1305?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14237305#comment-14237305
 ] 

ASF GitHub Bot commented on FLINK-1305:
---------------------------------------

Github user StephanEwen commented on the pull request:

    https://github.com/apache/incubator-flink/pull/252#issuecomment-65959720
  
    This change adds hadoop as a hard dependency to the `flink-java` project. 
Per the discussion on the mailing list, concerning support for Hadoop 
Writables, we voted to not do that and instead add an "mimick interface" to the 
Java API.
    
    For a big change like reversing that, it would be good to have some 
reasons...


> Flink's hadoop compatibility layer cannot handle NullWritables
> --------------------------------------------------------------
>
>                 Key: FLINK-1305
>                 URL: https://issues.apache.org/jira/browse/FLINK-1305
>             Project: Flink
>          Issue Type: Bug
>          Components: Hadoop Compatibility
>    Affects Versions: 0.7.0-incubating
>            Reporter: Sebastian Schelter
>            Assignee: Robert Metzger
>            Priority: Critical
>
> NullWritable is a special object that is commonly used in Hadoop 
> applications. NullWritable does not provide a public constructor, but only a 
> singleton factory method. Therefore Flink fails when users to try to read 
> NullWritables from Hadoop sequencefiles.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to