+1 for option 2. I'm excited to know about SPARK-48781, and it's perfect if spark supports a stored procedure. I will use the `call ` syntax, which only looks like a `stored procedure`,in this PR and adapt it to stored procedures in the future.
XiDuo You <ulyssesyo...@gmail.com> 于2024年9月25日周三 13:12写道: > +1 for option 2 > thank you > > Fei Wang <feiw...@apache.org> 于2024年9月25日周三 11:53写道: > > > > Prefer option 2 as well. > > > > BTW, it is necessary to support compact single partition for partitioned > table. > > > > On 2024/09/24 07:19:27 Cheng Pan wrote: > > > Hi Gabry, thanks for bringing up this discussion, usually, when we > want to discuss some idea and make decision, instead of starting a thread > with both [DISCUSS] and [VOTE], we firstly start a [DISCUSS] thread with > all options collected, and during the discussion, pros and cons of each > options will be listed and compared, ideally, all those involved in the > discussion will reach a consensus eventually, if not, we choose the most > supported options as the candidate to start a [VOTE], with > > > > > > +1 adopt > > > +0 does not care > > > -1 reject because … > > > > > > Back to the topic itself, there are actually 3 options: > > > > > > Option 1: new syntax COMPACT TABLE <table_name> [INTO <target_size >] > [CLEANUP | RETAIN | LIST] > > > Option 2: CALL compact_table(args …) > > > Option 3: VACUUM <table_name> [OTHER ARGS] > > > > > > I prefer option 2, then 3. Given Delta and Iceberg's dominance in the > lakehouse market, I suggest following either Delta's VACCUM or Iceberg's > CALL syntax. Plus Kyuubi Spark extension already adopted Delta ZORDER > syntax, and Spark 4.0 adopted the Iceberg CALL syntax, see SPARK-48781. > > > > > > Thanks, > > > Cheng Pan > > > > > > > > > > > > > On Sep 19, 2024, at 19:02, gabrywu <gabr...@apache.org> wrote: > > > > > > > > Hi, folks, > > > > I'm creating a PR #6695 <https://github.com/apache/kyuubi/pull/6695> > to create a new extended Spark SQL command to merge small files. And a few > of PMCs and committers propose that it's better to create a new Call > Procedure instead. > > > > So, I'm posting an email to vote on which one should be the best way > to extend Spark SQL. No matter what's the result, we can consider it as a > final decision to create a new spark extension in the upcoming PRs > > > > > > > > The VOTE will remain open for at least 2 weeks [ ] +1 Spark SQL > Command [ ] +0 Both is OK [ ] -1 Spark Call Procedure > > > > > > > > > > >