[ https://issues.apache.org/jira/browse/SPARK-44518?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17836703#comment-17836703 ]
Dongjoon Hyun commented on SPARK-44518: --------------------------------------- Hi, [~roryqi], this looks like a question instead of any concrete suggest . According to the Apache Spark community policy, I removed the `Target Version` from this JIRA. - https://spark.apache.org/contributing.html {code} Do not set the following fields: - Fix Version. This is assigned by committers only when resolved. - Target Version. This is assigned by committers to indicate a PR has been accepted for possible fix by the target version. {code} > Completely make hive as a data source > ------------------------------------- > > Key: SPARK-44518 > URL: https://issues.apache.org/jira/browse/SPARK-44518 > Project: Spark > Issue Type: Sub-task > Components: SQL > Affects Versions: 3.5.0 > Reporter: He Qi > Priority: Major > Fix For: 4.0.0 > > > Now, hive is a different data source from other data sources. In Spark > Project, Hive have many special logic and burden the cost of maintenance . > Like presto, hive is only a connector. Is it possible that we canĀ make hive > as a data source completely? > Surely, I know that it's very difficult. It has many historical problems and > compatible problems. Could we reduce these problems as possible as we can if > we release 4.0? > I just wanna start a discussion to collect more people's suggestion. Any > suggestion is welcome. I just feel 4.0 is a good opportunity to discuss this > issue. > If I am wrong, it's welcome to point it out. -- This message was sent by Atlassian Jira (v8.20.10#820010) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org