[
https://issues.apache.org/jira/browse/HUDI-1374?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
liwei updated HUDI-1374:
------------------------
Description:
disscuss in [https://github.com/apache/hudi/pull/2196]
A. Thanks so much. This pr need to solved the issue with better approach.
Now I am more clear about overwrite semantic between table.overwrite and spark
sql overwrite for hudi.
B. Also spark sql for hudi overwrite should have the ability just like spark
sql 、hive 、 delta lake.
these engine have three mode for overwrite about partition:
1. Dynamic Partition : delete all partition data ,and the insert the new data
for different
2. Static partition: just overwrite the partition which is user specified
3. Mixed partition: mixed of 1 and 2
more detail in :
[https://spark.apache.org/docs/3.0.0-preview/sql-ref-syntax-dml-insert-overwrite-table.html]
[https://www.programmersought.com/article/47155360487/]
Just fyi, in the
[RFC|https://cwiki.apache.org/confluence/display/HUDI/RFC+-+18+Insert+Overwrite+API#RFC18InsertOverwriteAPI-API]
we discussed having 'insert_overwrite_table' operation to support dynamic
partitioning. static partitioning is supported by 'insert_overwrite'.
was:
disscuss in https://github.com/apache/hudi/pull/2196
A. Thanks so much. This pr need to solved the issue with better approach.
Now I am more clear about overwrite semantic between table.overwrite and spark
sql overwrite for hudi.
B. Also spark sql for hudi overwrite should have the ability just like spark
sql 、hive 、 delta lake.
these engine have three mode for overwrite about partition:
1. Dynamic Partition : delete all partition data ,and the insert the new data
for different
2. Static partition: just overwrite the partition which is user specified
3. Mixed partition: mixed of 1 and 2
more detail in :
https://spark.apache.org/docs/3.0.0-preview/sql-ref-syntax-dml-insert-overwrite-table.html
https://www.programmersought.com/article/47155360487/
C. our plan
1. Now spark sql for hudi overwrite is Dynamic Partition. I will resolved it in
this issue HUDI-1349, first support delete all partition in HUDI-1350, then
land HUDI-1349
2. Now spark sql for hudi does not support "Static partition" mode, will then
land it in
> hudi table support dynamic partitioning
> ---------------------------------------
>
> Key: HUDI-1374
> URL: https://issues.apache.org/jira/browse/HUDI-1374
> Project: Apache Hudi
> Issue Type: Sub-task
> Components: Spark Integration
> Reporter: liwei
> Assignee: liwei
> Priority: Major
>
> disscuss in [https://github.com/apache/hudi/pull/2196]
> A. Thanks so much. This pr need to solved the issue with better approach.
> Now I am more clear about overwrite semantic between table.overwrite and
> spark sql overwrite for hudi.
> B. Also spark sql for hudi overwrite should have the ability just like spark
> sql 、hive 、 delta lake.
> these engine have three mode for overwrite about partition:
> 1. Dynamic Partition : delete all partition data ,and the insert the new
> data for different
> 2. Static partition: just overwrite the partition which is user specified
> 3. Mixed partition: mixed of 1 and 2
> more detail in :
>
> [https://spark.apache.org/docs/3.0.0-preview/sql-ref-syntax-dml-insert-overwrite-table.html]
> [https://www.programmersought.com/article/47155360487/]
> Just fyi, in the
> [RFC|https://cwiki.apache.org/confluence/display/HUDI/RFC+-+18+Insert+Overwrite+API#RFC18InsertOverwriteAPI-API]
> we discussed having 'insert_overwrite_table' operation to support dynamic
> partitioning. static partitioning is supported by 'insert_overwrite'.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)