Hi all,

I would like to volunteer to be the release manager of Spark 3 preview,
thanks!

Sean Owen <sro...@gmail.com> 于2019年9月13日周五 上午11:21写道:

> Well, great to hear the unanimous support for a Spark 3 preview
> release. Now, I don't know how to make releases myself :) I would
> first open it up to our revered release managers: would anyone be
> interested in trying to make one? sounds like it's not too soon to get
> what's in master out for evaluation, as there aren't any major
> deficiencies left, although a number of items to consider for the
> final release.
>
> I think we just need one release, targeting Hadoop 3.x / Hive 2.x in
> order to make it possible to test with JDK 11. (We're only on Scala
> 2.12 at this point.)
>
> On Thu, Sep 12, 2019 at 7:32 PM Reynold Xin <r...@databricks.com> wrote:
> >
> > +1! Long due for a preview release.
> >
> >
> > On Thu, Sep 12, 2019 at 5:26 PM, Holden Karau <hol...@pigscanfly.ca>
> wrote:
> >>
> >> I like the idea from the PoV of giving folks something to start testing
> against and exploring so they can raise issues with us earlier in the
> process and we have more time to make calls around this.
> >>
> >> On Thu, Sep 12, 2019 at 4:15 PM John Zhuge <jzh...@apache.org> wrote:
> >>>
> >>> +1  Like the idea as a user and a DSv2 contributor.
> >>>
> >>> On Thu, Sep 12, 2019 at 4:10 PM Jungtaek Lim <kabh...@gmail.com>
> wrote:
> >>>>
> >>>> +1 (as a contributor) from me to have preview release on Spark 3 as
> it would help to test the feature. When to cut preview release is
> questionable, as major works are ideally to be done before that - if we are
> intended to introduce new features before official release, that should
> work regardless of this, but if we are intended to have opportunity to test
> earlier, ideally it should.
> >>>>
> >>>> As a one of contributors in structured streaming area, I'd like to
> add some items for Spark 3.0, both "must be done" and "better to have". For
> "better to have", I pick some items for new features which committers
> reviewed couple of rounds and dropped off without soft-reject (No valid
> reason to stop). For Spark 2.4 users, only added feature for structured
> streaming is Kafka delegation token. (given we assume revising Kafka
> consumer pool as improvement) I hope we provide some gifts for structured
> streaming users in Spark 3.0 envelope.
> >>>>
> >>>> > must be done
> >>>> * SPARK-26154 Stream-stream joins - left outer join gives
> inconsistent output
> >>>> It's a correctness issue with multiple users reported, being reported
> at Nov. 2018. There's a way to reproduce it consistently, and we have a
> patch submitted at Jan. 2019 to fix it.
> >>>>
> >>>> > better to have
> >>>> * SPARK-23539 Add support for Kafka headers in Structured Streaming
> >>>> * SPARK-26848 Introduce new option to Kafka source - specify
> timestamp to start and end offset
> >>>> * SPARK-20568 Delete files after processing in structured streaming
> >>>>
> >>>> There're some more new features/improvements items in SS, but given
> we're talking about ramping-down, above list might be realistic one.
> >>>>
> >>>>
> >>>>
> >>>> On Thu, Sep 12, 2019 at 9:53 AM Jean Georges Perrin <j...@jgp.net>
> wrote:
> >>>>>
> >>>>> As a user/non committer, +1
> >>>>>
> >>>>> I love the idea of an early 3.0.0 so we can test current dev against
> it, I know the final 3.x will probably need another round of testing when
> it gets out, but less for sure... I know I could checkout and compile, but
> having a “packaged” preversion is great if it does not take too much time
> to the team...
> >>>>>
> >>>>> jg
> >>>>>
> >>>>>
> >>>>> On Sep 11, 2019, at 20:40, Hyukjin Kwon <gurwls...@gmail.com> wrote:
> >>>>>
> >>>>> +1 from me too but I would like to know what other people think too.
> >>>>>
> >>>>> 2019년 9월 12일 (목) 오전 9:07, Dongjoon Hyun <dongjoon.h...@gmail.com>님이
> 작성:
> >>>>>>
> >>>>>> Thank you, Sean.
> >>>>>>
> >>>>>> I'm also +1 for the following three.
> >>>>>>
> >>>>>> 1. Start to ramp down (by the official branch-3.0 cut)
> >>>>>> 2. Apache Spark 3.0.0-preview in 2019
> >>>>>> 3. Apache Spark 3.0.0 in early 2020
> >>>>>>
> >>>>>> For JDK11 clean-up, it will meet the timeline and `3.0.0-preview`
> helps it a lot.
> >>>>>>
> >>>>>> After this discussion, can we have some timeline for `Spark 3.0
> Release Window` in our versioning-policy page?
> >>>>>>
> >>>>>> - https://spark.apache.org/versioning-policy.html
> >>>>>>
> >>>>>> Bests,
> >>>>>> Dongjoon.
> >>>>>>
> >>>>>>
> >>>>>> On Wed, Sep 11, 2019 at 11:54 AM Michael Heuer <heue...@gmail.com>
> wrote:
> >>>>>>>
> >>>>>>> I would love to see Spark + Hadoop + Parquet + Avro compatibility
> problems resolved, e.g.
> >>>>>>>
> >>>>>>> https://issues.apache.org/jira/browse/SPARK-25588
> >>>>>>> https://issues.apache.org/jira/browse/SPARK-27781
> >>>>>>>
> >>>>>>> Note that Avro is now at 1.9.1, binary-incompatible with 1.8.x.
> As far as I know, Parquet has not cut a release based on this new version.
> >>>>>>>
> >>>>>>> Then out of curiosity, are the new Spark Graph APIs targeting 3.0?
> >>>>>>>
> >>>>>>> https://github.com/apache/spark/pull/24851
> >>>>>>> https://github.com/apache/spark/pull/24297
> >>>>>>>
> >>>>>>>    michael
> >>>>>>>
> >>>>>>>
> >>>>>>> On Sep 11, 2019, at 1:37 PM, Sean Owen <sro...@apache.org> wrote:
> >>>>>>>
> >>>>>>> I'm curious what current feelings are about ramping down towards a
> >>>>>>> Spark 3 release. It feels close to ready. There is no fixed date,
> >>>>>>> though in the past we had informally tossed around "back end of
> 2019".
> >>>>>>> For reference, Spark 1 was May 2014, Spark 2 was July 2016. I'd
> expect
> >>>>>>> Spark 2 to last longer, so to speak, but feels like Spark 3 is
> coming
> >>>>>>> due.
> >>>>>>>
> >>>>>>> What are the few major items that must get done for Spark 3, in
> your
> >>>>>>> opinion? Below are all of the open JIRAs for 3.0 (which everyone
> >>>>>>> should feel free to update with things that aren't really needed
> for
> >>>>>>> Spark 3; I already triaged some).
> >>>>>>>
> >>>>>>> For me, it's:
> >>>>>>> - DSv2?
> >>>>>>> - Finishing touches on the Hive, JDK 11 update
> >>>>>>>
> >>>>>>> What about considering a preview release earlier, as happened for
> >>>>>>> Spark 2, to get feedback much earlier than the RC cycle? Could that
> >>>>>>> even happen ... about now?
> >>>>>>>
> >>>>>>> I'm also wondering what a realistic estimate of Spark 3 release
> is. My
> >>>>>>> guess is quite early 2020, from here.
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>> SPARK-29014 DataSourceV2: Clean up current, default, and session
> catalog uses
> >>>>>>> SPARK-28900 Test Pyspark, SparkR on JDK 11 with run-tests
> >>>>>>> SPARK-28883 Fix a flaky test: ThriftServerQueryTestSuite
> >>>>>>> SPARK-28717 Update SQL ALTER TABLE RENAME  to use TableCatalog API
> >>>>>>> SPARK-28588 Build a SQL reference doc
> >>>>>>> SPARK-28629 Capture the missing rules in HiveSessionStateBuilder
> >>>>>>> SPARK-28684 Hive module support JDK 11
> >>>>>>> SPARK-28548 explain() shows wrong result for persisted DataFrames
> >>>>>>> after some operations
> >>>>>>> SPARK-28372 Document Spark WEB UI
> >>>>>>> SPARK-28476 Support ALTER DATABASE SET LOCATION
> >>>>>>> SPARK-28264 Revisiting Python / pandas UDF
> >>>>>>> SPARK-28301 fix the behavior of table name resolution with
> multi-catalog
> >>>>>>> SPARK-28155 do not leak SaveMode to file source v2
> >>>>>>> SPARK-28103 Cannot infer filters from union table with empty local
> >>>>>>> relation table properly
> >>>>>>> SPARK-28024 Incorrect numeric values when out of range
> >>>>>>> SPARK-27936 Support local dependency uploading from --py-files
> >>>>>>> SPARK-27884 Deprecate Python 2 support in Spark 3.0
> >>>>>>> SPARK-27763 Port test cases from PostgreSQL to Spark SQL
> >>>>>>> SPARK-27780 Shuffle server & client should be versioned to enable
> >>>>>>> smoother upgrade
> >>>>>>> SPARK-27714 Support Join Reorder based on Genetic Algorithm when
> the #
> >>>>>>> of joined tables > 12
> >>>>>>> SPARK-27471 Reorganize public v2 catalog API
> >>>>>>> SPARK-27520 Introduce a global config system to replace
> hadoopConfiguration
> >>>>>>> SPARK-24625 put all the backward compatible behavior change configs
> >>>>>>> under spark.sql.legacy.*
> >>>>>>> SPARK-24640 size(null) returns null
> >>>>>>> SPARK-24702 Unable to cast to calendar interval in spark sql.
> >>>>>>> SPARK-24838 Support uncorrelated IN/EXISTS subqueries for more
> operators
> >>>>>>> SPARK-24941 Add RDDBarrier.coalesce() function
> >>>>>>> SPARK-25017 Add test suite for ContextBarrierState
> >>>>>>> SPARK-25083 remove the type erasure hack in data source scan
> >>>>>>> SPARK-25383 Image data source supports sample pushdown
> >>>>>>> SPARK-27272 Enable blacklisting of node/executor on fetch failures
> by default
> >>>>>>> SPARK-27296 User Defined Aggregating Functions (UDAFs) have a major
> >>>>>>> efficiency problem
> >>>>>>> SPARK-25128 multiple simultaneous job submissions against k8s
> backend
> >>>>>>> cause driver pods to hang
> >>>>>>> SPARK-26731 remove EOLed spark jobs from jenkins
> >>>>>>> SPARK-26664 Make DecimalType's minimum adjusted scale configurable
> >>>>>>> SPARK-21559 Remove Mesos fine-grained mode
> >>>>>>> SPARK-24942 Improve cluster resource management with jobs
> containing
> >>>>>>> barrier stage
> >>>>>>> SPARK-25914 Separate projection from grouping and aggregate in
> logical Aggregate
> >>>>>>> SPARK-26022 PySpark Comparison with Pandas
> >>>>>>> SPARK-20964 Make some keywords reserved along with the ANSI/SQL
> standard
> >>>>>>> SPARK-26221 Improve Spark SQL instrumentation and metrics
> >>>>>>> SPARK-26425 Add more constraint checks in file streaming source to
> >>>>>>> avoid checkpoint corruption
> >>>>>>> SPARK-25843 Redesign rangeBetween API
> >>>>>>> SPARK-25841 Redesign window function rangeBetween API
> >>>>>>> SPARK-25752 Add trait to easily whitelist logical operators that
> >>>>>>> produce named output from CleanupAliases
> >>>>>>> SPARK-23210 Introduce the concept of default value to schema
> >>>>>>> SPARK-25640 Clarify/Improve EvalType for grouped aggregate and
> window aggregate
> >>>>>>> SPARK-25531 new write APIs for data source v2
> >>>>>>> SPARK-25547 Pluggable jdbc connection factory
> >>>>>>> SPARK-20845 Support specification of column names in INSERT INTO
> >>>>>>> SPARK-24417 Build and Run Spark on JDK11
> >>>>>>> SPARK-24724 Discuss necessary info and access in barrier mode +
> Kubernetes
> >>>>>>> SPARK-24725 Discuss necessary info and access in barrier mode +
> Mesos
> >>>>>>> SPARK-25074 Implement maxNumConcurrentTasks() in
> >>>>>>> MesosFineGrainedSchedulerBackend
> >>>>>>> SPARK-23710 Upgrade the built-in Hive to 2.3.5 for hadoop-3.2
> >>>>>>> SPARK-25186 Stabilize Data Source V2 API
> >>>>>>> SPARK-25376 Scenarios we should handle but missed in 2.4 for
> barrier
> >>>>>>> execution mode
> >>>>>>> SPARK-25390 data source V2 API refactoring
> >>>>>>> SPARK-7768 Make user-defined type (UDT) API public
> >>>>>>> SPARK-14922 Alter Table Drop Partition Using Predicate-based
> Partition Spec
> >>>>>>> SPARK-15691 Refactor and improve Hive support
> >>>>>>> SPARK-15694 Implement ScriptTransformation in sql/core
> >>>>>>> SPARK-16217 Support SELECT INTO statement
> >>>>>>> SPARK-16452 basic INFORMATION_SCHEMA support
> >>>>>>> SPARK-18134 SQL: MapType in Group BY and Joins not working
> >>>>>>> SPARK-18245 Improving support for bucketed table
> >>>>>>> SPARK-19842 Informational Referential Integrity Constraints
> Support in Spark
> >>>>>>> SPARK-22231 Support of map, filter, withColumn, dropColumn in
> nested
> >>>>>>> list of structures
> >>>>>>> SPARK-22632 Fix the behavior of timestamp values for R's DataFrame
> to
> >>>>>>> respect session timezone
> >>>>>>> SPARK-22386 Data Source V2 improvements
> >>>>>>> SPARK-24723 Discuss necessary info and access in barrier mode +
> YARN
> >>>>>>>
> >>>>>>>
> ---------------------------------------------------------------------
> >>>>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
> >>>>>>>
> >>>>>>>
> >>>>
> >>>>
> >>>> --
> >>>> Name : Jungtaek Lim
> >>>> Blog : http://medium.com/@heartsavior
> >>>> Twitter : http://twitter.com/heartsavior
> >>>> LinkedIn : http://www.linkedin.com/in/heartsavior
> >>>
> >>>
> >>>
> >>> --
> >>> John Zhuge
> >>
> >>
> >>
> >> --
> >> Twitter: https://twitter.com/holdenkarau
> >> Books (Learning Spark, High Performance Spark, etc.):
> https://amzn.to/2MaRAG9
> >> YouTube Live Streams: https://www.youtube.com/user/holdenkarau
> >
> >
>
> ---------------------------------------------------------------------
> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>
>

Reply via email to