[
https://issues.apache.org/jira/browse/FLINK-1107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14157801#comment-14157801
]
ASF GitHub Bot commented on FLINK-1107:
---------------------------------------
Github user rmetzger commented on a diff in the pull request:
https://github.com/apache/incubator-flink/pull/131#discussion_r18385767
--- Diff: docs/hadoop_compatibility.md ---
@@ -1,5 +1,75 @@
---
-title: "Hadoop Compatability"
+title: "Hadoop I/O Compatibility"
---
-To be written.
\ No newline at end of file
+Flink not only supports types that implement Apache Hadoop's `Writable`
interface by default, but also provides a compatibility
+layer that allows for using any class extending
`org.apache.hadoop.mapred(uce).InputFormat` as a Flink `InputFormat` as well as
any
+class extending `org.apache.hadoop.mapred(uce).OutputFormat` as a Flink
`OutputFormat`.
+
+Thus, Flink can handle Hadoop-related formats from the common
`TextInputFormat` up to third-party components such as Hive through HCatalog's
`HCatInputFormat`. Flink supports both the formats which use the old
`org.apache.hadoop.mapred` API as well as the new `org.apache.hadoop.mapreduce`
API.
+
+This document explains how to configure your Maven project correctly and
shows an example.
+
+### Project Configuration
+
+The Hadoop Compatibility Layer is part of the *addons* Maven project. All
relevant classes are located in the `org.apache.flink.hadoopcompatibility`
package. The package includes separate packages and classes for the Hadoop
`mapred` and `mapreduce` API.
+
+Add the following dependency to your `pom.xml` to use the Hadoop
Compatibility Layer.
+
+~~~xml
+<dependency>
+ <groupId>org.apache.flink</groupId>
+ <artifactId>flink-hadoop-compatibility</artifactId>
+ <version>{{site.FLINK_VERSION_STABLE}}</version>
+</dependency>
+~~~
--- End diff --
can you add another headline here, that quickly explains the usage?
Just how to read data from the environment to a data set using a hadoop
input format (something like a copy-past skeleton).
I think its better if you have a) a general definition how to use it
and b) an example that people see how to use it in context.
> Document how to use Avro files and the Hadoop Input Format Wrappers
> -------------------------------------------------------------------
>
> Key: FLINK-1107
> URL: https://issues.apache.org/jira/browse/FLINK-1107
> Project: Flink
> Issue Type: Task
> Components: Documentation
> Affects Versions: 0.7-incubating
> Reporter: Robert Metzger
> Assignee: Timo Walther
> Priority: Minor
>
> The documentation lacks any examples or description on how to read from avro
> files.
> Also, we should document the Hadoop Input Formats a bit.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)