That's great!
Thank you.

Let me know if you have any questions.

Fabian

Am Mo., 15. Apr. 2019 um 11:32 Uhr schrieb Hai <h...@magicsoho.com>:

> Hi Fabian:
>
>
> OK ,I am glad to do that.
>
>
> Regards
>
>  Original Message
> *Sender:* Fabian Hueske<fhue...@gmail.com>
> *Recipient:* hai<h...@magicsoho.com>
> *Cc:* user<user@flink.apache.org>; Yun Tang<myas...@live.com>
> *Date:* Monday, Apr 15, 2019 17:16
> *Subject:* Re: Hbase Connector failed when deployed to yarn
>
> Hi,
>
> The Jira issue is still unassigned.
> Would you be up to work on a fix?
>
> Best, Fabian
>
> Am Fr., 12. Apr. 2019 um 05:07 Uhr schrieb hai <h...@magicsoho.com>:
>
>> Hi, Tang:
>>
>>
>> Thaks for your reply, will this issue fix soon?I don’t think put 
>> flink-hadoop-compatibility
>> jar under FLINK_HOME/lib is a elegant solution.
>>
>>
>> Regards
>>
>>  Original Message
>> *Sender:* Yun Tang<myas...@live.com>
>> *Recipient:* hai<h...@magicsoho.com>; user<user@flink.apache.org>
>> *Date:* Friday, Apr 12, 2019 02:02
>> *Subject:* Re: Hbase Connector failed when deployed to yarn
>>
>> Hi
>>
>> I believe this is the same problem which reported in
>> https://issues.apache.org/jira/browse/FLINK-12163 , current work around
>> solution is to put flink-hadoop-compatibility jar under FLINK_HOME/lib.
>>
>> Best
>> Yun Tang
>> ------------------------------
>> *From:* hai <h...@magicsoho.com>
>> *Sent:* Thursday, April 11, 2019 21:06
>> *To:* user
>> *Subject:* Re: Hbase Connector failed when deployed to yarn
>>
>>
>> And my pom.xml dependencies is :
>>
>>
>> <dependencies>
>>
>>         <!-- Scala -->
>>
>>         <dependency>
>>
>>             <groupId>org.scala-lang</groupId>
>>
>>             <artifactId>scala-library</artifactId>
>>
>>             <version>${scala.version}</version>
>>
>>         </dependency>
>>
>>         <dependency>
>>
>>             <groupId>org.scala-lang</groupId>
>>
>>             <artifactId>scala-compiler</artifactId>
>>
>>             <version>${scala.version}</version>
>>
>>         </dependency>
>>
>>
>>         <!-- SL4J & Log4j & Kafka-Appender & Flume-Appender -->
>>
>>         <dependency>
>>
>>             <groupId>org.slf4j</groupId>
>>
>>             <artifactId>slf4j-api</artifactId>
>>
>>             <version>1.7.21</version>
>>
>>         </dependency>
>>
>>
>>         <!-- 1.1.1 -->
>>
>>         <dependency>
>>
>>             <groupId>ch.qos.logback</groupId>
>>
>>             <artifactId>logback-core</artifactId>
>>
>>             <version>1.1.1</version>
>>
>>         </dependency>
>>
>>         <dependency>
>>
>>             <groupId>ch.qos.logback</groupId>
>>
>>             <artifactId>logback-classic</artifactId>
>>
>>             <version>1.1.1</version>
>>
>>         </dependency>
>>
>>         <!-- Flink -->
>>
>>         <dependency>
>>
>>             <groupId>org.apache.flink</groupId>
>>
>>             <artifactId>flink-scala_${scala.binary.version}</artifactId>
>>
>>             <version>${flink.version}</version>
>>
>>             <scope>compile</scope>
>>
>>         </dependency>
>>
>>         <dependency>
>>
>>             <groupId>org.apache.flink</groupId>
>>
>>
>> <artifactId>flink-streaming-scala_${scala.binary.version}</artifactId>
>>
>>             <version>${flink.version}</version>
>>
>>             <scope>compile</scope>
>>
>>         </dependency>
>>
>>         <dependency>
>>
>>             <groupId>org.apache.flink</groupId>
>>
>>
>> <artifactId>flink-runtime-web_${scala.binary.version}</artifactId>
>>
>>             <version>${flink.version}</version>
>>
>>         </dependency>
>>
>>
>>         <dependency>
>>
>>             <groupId>org.apache.flink</groupId>
>>
>>             <artifactId>flink-hbase_${scala.binary.version}</artifactId>
>>
>>             <version>${flink.version}</version>
>>
>>         </dependency>
>>
>>         <dependency>
>>
>>             <groupId>org.apache.flink</groupId>
>>
>>
>> <artifactId>flink-hadoop-compatibility_${scala.binary.version}</artifactId>
>>
>>             <version>${flink.version}</version>
>>
>>         </dependency>
>>
>>         <dependency>
>>
>>             <groupId>org.apache.hadoop</groupId>
>>
>>             <artifactId>hadoop-mapreduce-client-core</artifactId>
>>
>>             <version>${hadoop.version}</version>
>>
>>         </dependency>
>>
>>         <dependency>
>>
>>             <groupId>cglib</groupId>
>>
>>             <artifactId>cglib</artifactId>
>>
>>             <version>2.2.2</version>
>>
>>         </dependency>
>>
>>         <!-- Hadoop -->
>>
>>         <dependency>
>>
>>             <groupId>org.apache.hadoop</groupId>
>>
>>             <artifactId>hadoop-common</artifactId>
>>
>>             <version>${hadoop.version}</version>
>>
>>         </dependency>
>>
>>     </dependencies>
>>
>>  Original Message
>> *Sender:* hai<h...@magicsoho.com>
>> *Recipient:* user@flink.apache.org
>> *Date:* Thursday, Apr 11, 2019 21:04
>> *Subject:* Hbase Connector failed when deployed to yarn
>>
>> *Hello:*
>>
>> *    I am new to flink, and I copy the official Hbase connector examples
>> from source *
>> flink <https://github.com/apache/flink>/flink-connectors
>> <https://github.com/apache/flink/tree/master/flink-connectors>/
>> flink-hbase
>> <https://github.com/apache/flink/tree/master/flink-connectors/flink-hbase>
>> /src
>> <https://github.com/apache/flink/tree/master/flink-connectors/flink-hbase/src>
>> /test
>> <https://github.com/apache/flink/tree/master/flink-connectors/flink-hbase/src/test>
>> /java
>> <https://github.com/apache/flink/tree/master/flink-connectors/flink-hbase/src/test/java>
>> /org
>> <https://github.com/apache/flink/tree/master/flink-connectors/flink-hbase/src/test/java/org>
>> /apache
>> <https://github.com/apache/flink/tree/master/flink-connectors/flink-hbase/src/test/java/org/apache>
>> /flink
>> <https://github.com/apache/flink/tree/master/flink-connectors/flink-hbase/src/test/java/org/apache/flink>
>> /addons
>> <https://github.com/apache/flink/tree/master/flink-connectors/flink-hbase/src/test/java/org/apache/flink/addons>
>> /hbase
>> <https://github.com/apache/flink/tree/master/flink-connectors/flink-hbase/src/test/java/org/apache/flink/addons/hbase>
>> /example
>> <https://github.com/apache/flink/tree/master/flink-connectors/flink-hbase/src/test/java/org/apache/flink/addons/hbase/example>
>> /HBaseWriteExample.java
>> and run in a yarn-cluster with the command:
>>
>> bin/flink run -m yarn-cluster -yn 2 -c
>> {class-path-prefix}.HBaseWriteExample {my-application}.jar
>>
>> What I have get is:
>>
>> ------------------------------------------------------------
>>  The program finished with the following exception:
>>
>> org.apache.flink.client.program.ProgramInvocationException: The main
>> method caused an error.
>> at
>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:545)
>> at
>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:419)
>> at
>> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:339)
>> at
>> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:831)
>> at org.apache.flink.client.CliFrontend.run(CliFrontend.java:256)
>> at
>> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1073)
>> at org.apache.flink.client.CliFrontend$2.call(CliFrontend.java:1120)
>> at org.apache.flink.client.CliFrontend$2.call(CliFrontend.java:1117)
>> at
>> org.apache.flink.runtime.security.HadoopSecurityContext$1.run(HadoopSecurityContext.java:43)
>> at java.security.AccessController.doPrivileged(Native Method)
>> at javax.security.auth.Subject.doAs(Subject.java:422)
>> at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
>> at
>> org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:40)
>> at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1117)
>> Caused by: java.lang.RuntimeException: Could not load the TypeInformation
>> for the class 'org.apache.hadoop.io.Writable'. You may be missing the
>> 'flink-hadoop-compatibility' dependency.
>> at
>> org.apache.flink.api.java.typeutils.TypeExtractor.createHadoopWritableTypeInfo(TypeExtractor.java:2025)
>> at
>> org.apache.flink.api.java.typeutils.TypeExtractor.privateGetForClass(TypeExtractor.java:1649)
>> at
>> org.apache.flink.api.java.typeutils.TypeExtractor.privateGetForClass(TypeExtractor.java:1591)
>> at
>> org.apache.flink.api.java.typeutils.TypeExtractor.createTypeInfoWithTypeHierarchy(TypeExtractor.java:778)
>> at
>> org.apache.flink.api.java.typeutils.TypeExtractor.createSubTypesInfo(TypeExtractor.java:998)
>> at
>> org.apache.flink.api.java.typeutils.TypeExtractor.createTypeInfoWithTypeHierarchy(TypeExtractor.java:679)
>> at
>> org.apache.flink.api.java.typeutils.TypeExtractor.createTypeInfoFromInputs(TypeExtractor.java:791)
>> at
>> org.apache.flink.api.java.typeutils.TypeExtractor.privateCreateTypeInfo(TypeExtractor.java:621)
>> at
>> org.apache.flink.api.java.typeutils.TypeExtractor.getUnaryOperatorReturnType(TypeExtractor.java:425)
>> at
>> org.apache.flink.api.java.typeutils.TypeExtractor.getUnaryOperatorReturnType(TypeExtractor.java:349)
>> at
>> org.apache.flink.api.java.typeutils.TypeExtractor.getMapReturnTypes(TypeExtractor.java:164)
>> at org.apache.flink.api.java.DataSet.map(DataSet.java:215)
>> at
>> com.luckyfish.flink.java.HBaseWriteExample.main(HBaseWriteExample.java:75)
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>> at java.lang.reflect.Method.invoke(Method.java:498)
>> at
>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:528)
>> ... 13 more
>>
>> What should I do to deal with this exception ?
>>
>> Many Thanks
>>
>

Reply via email to