hi,periyasamy
Thanks for asking questions or reporting issues, please describe it in
detail by using github issue.
https://github.com/apache/incubator-hudi/issues

cooper

selvaraj periyasamy <[email protected]> 于2020年4月22日周三
上午7:35写道:

> Folks,
>
> I am using  Apache Hudi 0.5.0. Our hadoop cluster is miix of  spark
> version  2.3.0, Scala version 2.11.8 & Hive version 1.2.2.
> There are multiple use cases already working in Hudi.
>
> I need to read one of sequence table, which is continuously inserted on
> new partition by other process using Hive, not by Hudi.  And then write
> this DataFrame into another COW table using Hudi.
>
> When I use spark.sql in spark-shell, which was started with Hudi jar, I am
> able to do select as mentioned below.
>
> spark-shell --jars
> /Users/seperiya/Downloads/hudi-spark-bundle-0.5.0-incubating.jar --conf
> 'spark.serializer=org.apache.spark.serializer.KryoSerializer'
>
>
> scala> spark.sql("select * from poc.request_result__ct").show
>
> 2020-04-21 15:56:07 WARN  ObjectStore:568 - Failed to get database
> global_temp, returning NoSuchObjectException
>
>
> +----------+--------------+------------+---------+-------------------+-------------------+-------------------+-------------------+----------------------+
>
> |request_id|prev_request_id|ref_no|type_code|
> transaction_date|process_ts|
> commit_ts|header__change_oper|header__partition_name|
>
>
> +----------+--------------+------------+---------+-------------------+-------------------+-------------------+-------------------+----------------------+
>
> |2020041011|          null|        null|       PA|2020-04-10
> 11:11:23|2020-04-10 11:11:30|2020-04-10 11:11:35|                  I|
> 20200117T235000_2...|
>
>
> +----------+--------------+------------+---------+-------------------+-------------------+-------------------+-------------------+----------------------+
>
>
>
>
>
> Whereas when I convert the same code into scala file and execute it using
> spark-submit , I am getting error. Attached the error logs.
>
>
> spark-submit --jars
> /Users/seperiya/Downloads/hudi-spark-bundle-0.5.0-incubating.jar --conf
> 'spark.serializer=org.apache.spark.serializer.KryoSerializer' --class Test
> /Test-1.0.0-SNAPSHOT-jar-with-dependencies.jar
>
>
> object Test {
>
> def main(args: Array[String]): Unit = {
>
> implicit val sparkSession: SparkSession = SparkUtil.buildSession("Test_"+
>
>   
> now.get(Calendar.HOUR_OF_DAY)+now.get(Calendar.MINUTE)+now.get(Calendar.SECOND))
>
> sparkSession.sql( s"select * from poc.request_result__ct").show()
>
>  }
> }
>
>
> When I remove Hudi bundle jar and runs the same , it works.
>
>
> spark-submit --class Test /Test-1.0.0-SNAPSHOT-jar-with-dependencies.jar
>
>
> Even though Hudi code will come into picture only when I insert the data
> on other table,  for some reason, read fails . Could anyone shed some
> light not his issue?
>
>
>

Reply via email to