[ 
https://issues.apache.org/jira/browse/HUDI-467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17003095#comment-17003095
 ] 

cdmikechen edited comment on HUDI-467 at 12/25/19 6:29 AM:
-----------------------------------------------------------

I checked mvn tree and found that mvn use parquet-avro in 
*hudi-hadoop-mr-bundle* with compile but not in *hudi-hive-bundle*. If we don't 
add parquet-avro dependency in *hudi-hive-bundle* mvn can not packag 
parquet-avro to *hudi-hive-bundle*.


was (Author: chenxiang):
I checked mvn tree and found that mvn use parquet-avro in 
*hudi-hadoop-mr-bundle* with compile but not in *hudi-hive-bundle*. If we don't 
add parquet-avro dependency in *hudi-hive-bundle* mvn can not packag 
parquet-avro to hudi-hive-bundle.

> Query RT Table in Hive found java.lang.NoClassDefFoundError Exception
> ---------------------------------------------------------------------
>
>                 Key: HUDI-467
>                 URL: https://issues.apache.org/jira/browse/HUDI-467
>             Project: Apache Hudi (incubating)
>          Issue Type: Bug
>          Components: Hive Integration
>            Reporter: cdmikechen
>            Assignee: cdmikechen
>            Priority: Major
>
> When creating a *MERGE_ON_READ* table in hudi and syn to hive, hudi will 
> create two table named *table_name* and *table_name_rt*, when I query 
> *table_name_rt*, I catch  *java.lang.NoClassDefFoundError* Exception:
> {code}
> java.lang.RuntimeException: java.lang.NoClassDefFoundError: 
> org/apache/parquet/avro/AvroSchemaConverter
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:89)
>  ~[hive-service-2.3.3.jar:2.3.3]
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:36)
>  ~[hive-service-2.3.3.jar:2.3.3]
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63)
>  ~[hive-service-2.3.3.jar:2.3.3]
>       at java.security.AccessController.doPrivileged(Native Method) 
> ~[?:1.8.0_201]
>       at javax.security.auth.Subject.doAs(Subject.java:422) ~[?:1.8.0_201]
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1844)
>  ~[hadoop-common-2.8.5.jar:?]
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:59)
>  ~[hive-service-2.3.3.jar:2.3.3]
>       at com.sun.proxy.$Proxy47.fetchResults(Unknown Source) ~[?:?]
>       at 
> org.apache.hive.service.cli.CLIService.fetchResults(CLIService.java:559) 
> ~[hive-service-2.3.3.jar:2.3.3]
>       at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.FetchResults(ThriftCLIService.java:751)
>  ~[hive-service-2.3.3.jar:2.3.3]
>       at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1717)
>  ~[hive-exec-2.3.3.jar:2.3.3]
>       at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1702)
>  ~[hive-exec-2.3.3.jar:2.3.3]
>       at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) 
> ~[hive-exec-2.3.3.jar:2.3.3]
>       at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39) 
> ~[hive-exec-2.3.3.jar:2.3.3]
>       at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56)
>  ~[hive-service-2.3.3.jar:2.3.3]
>       at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286)
>  ~[hive-exec-2.3.3.jar:2.3.3]
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  [?:1.8.0_201]
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  [?:1.8.0_201]
>       at java.lang.Thread.run(Thread.java:748) [?:1.8.0_201]
> Caused by: java.lang.NoClassDefFoundError: 
> org/apache/parquet/avro/AvroSchemaConverter
>       at 
> org.apache.hudi.hadoop.realtime.AbstractRealtimeRecordReader.init(AbstractRealtimeRecordReader.java:341)
>  ~[?:?]
>       at 
> org.apache.hudi.hadoop.realtime.AbstractRealtimeRecordReader.<init>(AbstractRealtimeRecordReader.java:108)
>  ~[?:?]
>       at 
> org.apache.hudi.hadoop.realtime.RealtimeCompactedRecordReader.<init>(RealtimeCompactedRecordReader.java:50)
>  ~[?:?]
>       at 
> org.apache.hudi.hadoop.realtime.HoodieRealtimeRecordReader.constructRecordReader(HoodieRealtimeRecordReader.java:69)
>  ~[?:?]
>       at 
> org.apache.hudi.hadoop.realtime.HoodieRealtimeRecordReader.<init>(HoodieRealtimeRecordReader.java:47)
>  ~[?:?]
>       at 
> org.apache.hudi.hadoop.realtime.HoodieParquetRealtimeInputFormat.getRecordReader(HoodieParquetRealtimeInputFormat.java:254)
>  ~[?:?]
>       at 
> org.apache.hadoop.hive.ql.exec.FetchOperator$FetchInputFormatSplit.getRecordReader(FetchOperator.java:695)
>  ~[hive-exec-2.3.3.jar:2.3.3]
>       at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.getRecordReader(FetchOperator.java:333)
>  ~[hive-exec-2.3.3.jar:2.3.3]
>       at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.getNextRow(FetchOperator.java:459)
>  ~[hive-exec-2.3.3.jar:2.3.3]
>       at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.pushRow(FetchOperator.java:428) 
> ~[hive-exec-2.3.3.jar:2.3.3]
>       at org.apache.hadoop.hive.ql.exec.FetchTask.fetch(FetchTask.java:147) 
> ~[hive-exec-2.3.3.jar:2.3.3]
>       at org.apache.hadoop.hive.ql.Driver.getResults(Driver.java:2208) 
> ~[hive-exec-2.3.3.jar:2.3.3]
>       at 
> org.apache.hive.service.cli.operation.SQLOperation.getNextRowSet(SQLOperation.java:494)
>  ~[hive-service-2.3.3.jar:2.3.3]
>       at 
> org.apache.hive.service.cli.operation.OperationManager.getOperationNextRowSet(OperationManager.java:307)
>  ~[hive-service-2.3.3.jar:2.3.3]
>       at 
> org.apache.hive.service.cli.session.HiveSessionImpl.fetchResults(HiveSessionImpl.java:878)
>  ~[hive-service-2.3.3.jar:2.3.3]
>       at sun.reflect.GeneratedMethodAccessor17.invoke(Unknown Source) ~[?:?]
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>  ~[?:1.8.0_201]
>       at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_201]
>       at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78)
>  ~[hive-service-2.3.3.jar:2.3.3]
>       ... 18 more
> {code}
> I checked hive lib folder and don't found parquet-avro jar. At that time I 
> thought we need to import parquet-avro dependencie in *hudi-hive-bundle* 
> pom.xml. So I checked this pom.xml and found that although parquet-avro have 
> been configured in 
> {code}
>  <artifactSet>
>   <includes>
>     <include>org.apache.hudi:hudi-common</include>
>     <include>org.apache.hudi:hudi-hadoop-mr</include>
>     <include>org.apache.hudi:hudi-hive</include>
>     <include>com.beust:jcommander</include>
>     <include>org.apache.parquet:parquet-avro</include>
>     <include>com.esotericsoftware:kryo-shaded</include>
>     <include>org.objenesis:objenesis</include>
>     <include>com.esotericsoftware:minlog</include>
>   </includes>
> </artifactSet>
> {code}
> But parquet-avro can not be packaged to hudi-hive-bundle.jar. Maybe we need 
> to add dependency in dependencies.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to