I am using Hudi 0.5.0 and then write using sparkwriter.

My spark version is 2.3.0
Scala version 2.11.8
Hive version 1.2.2

Write is success but hive call is failing. When checked some google
reference, It seems to be an hive client is higher version the server.
Since Hudi is built on hive 2.3.1, Is there a way to use 1.2.2?

2020-03-01 12:16:50 WARN  HoodieSparkSqlWriter$:110 - hoodie dataset at
hdfs://localhost:9000/projects/cdp/data/attunity_poc/attunity_rep_base
already exists. Deleting existing data & overwriting with new data.
[Stage 111:============================>
                                                         2020-03-01
12:16:51 ERROR HiveConnection:697 - Error opening session
org.apache.thrift.TApplicationException: Required field 'client_protocol'
is unset! Struct:TOpenSessionReq(client_protocol:null,
configuration:{set:hiveconf:hive.server2.thrift.resultset.default.fetch.size=1000,
use:database=default})
at
org.apache.thrift.TApplicationException.read(TApplicationException.java:111)
at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:79)
at
org.apache.hudi.org.apache.hive.service.rpc.thrift.TCLIService$Client.recv_OpenSession(TCLIService.java:168)
at
org.apache.hudi.org.apache.hive.service.rpc.thrift.TCLIService$Client.OpenSession(TCLIService.java:155)


Thanks,
Selva

Reply via email to