dongkelun commented on pull request #3415:
URL: https://github.com/apache/hudi/pull/3415#issuecomment-898207251
> Hi @dongkelun , Thanks for the contribution for this. Overall LGTM except
some minor optimize. And also you can run the test case in spark3 by the follow
command:
>
> > mvn clean install -DskipTests -Pspark3
> > mvn test -Punit-tests -Pspark3 -pl hudi-spark-datasource/hudi-spark
Hi, @pengzhiwei2018 The result is:'Tests: succeeded 56, failed 6, canceled
0, ignored 0, pending 0'.Two of them are ORC exceptions, and the other three I
think are due to time zone differences, but I don't know how to solve the time
zone difference, and the other one is the mismatch of exception information.
The detailed results are as follows:
`
1、Test Different Type of Partition Column *** FAILED ***
Expected Array([1,a1,10,2021-05-20 00:00:00], [2,a2,10,2021-05-20
00:00:00]), but got Array([1,a1,10.0,2021-05-20 15:00:00],
[2,a2,10.0,2021-05-20 15:00:00])
2、- Test MergeInto Exception *** FAILED ***
Expected "... for target field: '[id]' in merge into upda...", but got
"... for target field: '[_ts]' in merge into upda..."
(TestHoodieSqlBase.scala:86)
3、test basic HoodieSparkSqlWriter functionality with datasource insert for
COPY_ON_WRITE with ORC as the base file format with populate meta fields true
*** FAILED ***
4、test basic HoodieSparkSqlWriter functionality with datasource insert for
MERGE_ON_READ with ORC as the base file format with populate meta fields true
*** FAILED ***
5、Test Sql Statements *** FAILED ***
java.lang.IllegalArgumentException: UnExpect result for: select id, name,
price, cast(dt as string) from h0_p
Expect:
1 a1 10 2021-05-07 00:00:00, Actual:
1 a1 10 2021-05-07 15:00:00
6、Test Create Table As Select *** FAILED ***
Expected Array([1,a1,10,2021-05-06 00:00:00]), but got
Array([1,a1,10,2021-05-06 15:00:00]) (TestHoodieSqlBase.scala:78)
`
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]