[ 
https://issues.apache.org/jira/browse/SPARK-31136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058125#comment-17058125
 ] 

Michael Armbrust commented on SPARK-31136:
------------------------------------------

What was the default before, hive sequence files? That is pretty bad. They will 
get orders of magnitude better performance with parquet (I've seen users really 
burned by the performance of the old default here).

What operations are affected by this? What happens when run a program without 
changing anything?
 - For new tables, I assume you just get better performance?
 - Are there any operations where this breaks things? Can you "corrupt" a hive 
table by accidentally writing parquet data into it?

> Revert SPARK-30098 Use default datasource as provider for CREATE TABLE syntax
> -----------------------------------------------------------------------------
>
>                 Key: SPARK-31136
>                 URL: https://issues.apache.org/jira/browse/SPARK-31136
>             Project: Spark
>          Issue Type: Sub-task
>          Components: SQL
>    Affects Versions: 3.0.0
>            Reporter: Dongjoon Hyun
>            Priority: Major
>
> We need to consider the behavior change of SPARK-30098 .
> This is a placeholder to keep the discussion and the final decision.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to