[ 
https://issues.apache.org/jira/browse/FLINK-403?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aljoscha Krettek resolved FLINK-403.
------------------------------------
    Resolution: Fixed

Scala also supports them after merge of scala-rework.

> Support for Hadoop Types in Java/Scala
> --------------------------------------
>
>                 Key: FLINK-403
>                 URL: https://issues.apache.org/jira/browse/FLINK-403
>             Project: Flink
>          Issue Type: Sub-task
>            Reporter: GitHub Import
>              Labels: github-import
>             Fix For: pre-apache
>
>
> It would be very valueable if users could use plain Hadoop objects as types 
> in the Java and Scala API.
> Since they are serializable/deserializable by a very similar mechanism as the 
> Stratosphere Value types, this should be easy to add.
> The following is necessary:
>   - Add setField(...) and getField(...) methods to the Record for subclasses 
> if Hadoop's Writable. That should do it for the Java API.
>   - The Scala API needs an additional check in the type generator that 
> subclasses of Writables are also accepted and not only subclasses of Value.
> ---------------- Imported from GitHub ----------------
> Url: https://github.com/stratosphere/stratosphere/issues/403
> Created by: [StephanEwen|https://github.com/StephanEwen]
> Labels: core, user satisfaction, 
> Created at: Thu Jan 16 14:20:02 CET 2014
> State: open



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to