[ 
https://issues.apache.org/jira/browse/PHOENIX-2745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15180137#comment-15180137
 ] 

Josh Mahonin commented on PHOENIX-2745:
---------------------------------------

Very interesting [~lichenglingl]. Do you have any examples of other data 
sources exhibiting this 'drop' -> 'reload' behaviour for SaveMode.Overwrite?

When I'd first written the integration, 'Overwrite' seemed to me the most 
correct, based on this documentation of the SaveMode class [1]:
"if data/table already exists, existing data is expected to be overwritten by 
the contents of the DataFrame"

However, if other data sources use 'Append' to the same effect, it might be 
best to use that as the default behaviour. Follow-up work would then be to look 
at doing a DROP or DELETE in the case of SaveMode.Overwrite.

[1] 
https://spark.apache.org/docs/1.6.0/api/java/org/apache/spark/sql/SaveMode.html

> The spark savemode not work correctly
> -------------------------------------
>
>                 Key: PHOENIX-2745
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-2745
>             Project: Phoenix
>          Issue Type: Bug
>    Affects Versions: 4.6.0
>            Reporter: lichenglin
>
> When saving a dataframe to spark with the mode SaveMode.Overwrite 
> spark will drop the table  first and load the new dataframe 
> but phoinex just replace the old data to the new data according to the 
> primary key 
> the old datas still exsits.
> the overwrite actually work as append



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to