[ https://issues.apache.org/jira/browse/SPARK-31136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17058125#comment-17058125 ]
Michael Armbrust commented on SPARK-31136: ------------------------------------------ What was the default before, hive sequence files? That is pretty bad. They will get orders of magnitude better performance with parquet (I've seen users really burned by the performance of the old default here). What operations are affected by this? What happens when run a program without changing anything? - For new tables, I assume you just get better performance? - Are there any operations where this breaks things? Can you "corrupt" a hive table by accidentally writing parquet data into it? > Revert SPARK-30098 Use default datasource as provider for CREATE TABLE syntax > ----------------------------------------------------------------------------- > > Key: SPARK-31136 > URL: https://issues.apache.org/jira/browse/SPARK-31136 > Project: Spark > Issue Type: Sub-task > Components: SQL > Affects Versions: 3.0.0 > Reporter: Dongjoon Hyun > Priority: Major > > We need to consider the behavior change of SPARK-30098 . > This is a placeholder to keep the discussion and the final decision. -- This message was sent by Atlassian Jira (v8.3.4#803005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org