Re: Elasticsearch support for Spark 3.x

2023-09-08 Thread Dipayan Dev
t have any external >>> dependencies added, and just updated the Spark version as mentioned below. >>> >>> Can someone help me with this? >>> >>> On Fri, 1 Sep 2023 at 5:58 PM, Koert Kuipers wrote: >>> &g

Re: Elasticsearch support for Spark 3.x

2023-09-08 Thread Dipayan Dev
ncies added, and just updated the Spark version as mentioned below. >> >> Can someone help me with this? >> >> On Fri, 1 Sep 2023 at 5:58 PM, Koert Kuipers wrote: >> >>> could the provided scope be the issue? >>> >>> On Sun, Aug 27, 202

Re: Elasticsearch support for Spark 3.x

2023-09-08 Thread Alfie Davidson
<dev.dipaya...@gmail.com> wrote:Using the following dependency for Spark 3 in POM file (My Scala version is 2.12.14)    org.elasticsearch    elasticsearch-spark-30_2.12    7.12.0    providedThe code throws error at this line : df.write.format("es").mode("overwrite").options(elas

Re: Elasticsearch support for Spark 3.x

2023-09-07 Thread Dipayan Dev
p 2023 at 5:58 PM, Koert Kuipers wrote: > >> could the provided scope be the issue? >> >> On Sun, Aug 27, 2023 at 2:58 PM Dipayan Dev >> wrote: >> >>> Using the following dependency for Spark 3 in POM file (My Scala version >>> is 2.12.14) >>

Re: Behaviour of Append & Overwrite modes when table is not present when using df.write in Spark 3

2022-05-18 Thread Sourabh Badhya
pported/unsupported when df.write is used in Spark 3 when the underlying > custom datasource is using SupportCatalogOptions. > > As per my knowledge, in the current implementation in master, df.write in > Append and Overwrite mode tries to load the table and look for the schema

Behaviour of Append & Overwrite modes when table is not present when using df.write in Spark 3

2022-05-09 Thread Sourabh Badhya
Hi team, I would like to know the behaviour of Append & Overwrite modes when table is not present and whether automatic table creation is supported/unsupported when df.write is used in Spark 3 when the underlying custom datasource is using SupportCatalogOptions. As per my knowl

Re: Spark 3 is Slower than Spark 2 for TPCDS Q04 query.

2021-12-20 Thread Senthil Kumar
e 1500 that I have on a Hadoop/YARN cluster, and was > not able to reproduce the difference in execution time between Spark 2 and > Spark 3 that you report in your mail. > > This is the Spark config I used: > > bin/spark-shell --master yarn --driver-memory 8g --executor-cor

RE: Spark 3 is Slower than Spark 2 for TPCDS Q04 query.

2021-12-20 Thread Luca Canali
Hi Senthil, I have just run a couple of quick tests for TPCDS Q4, using the TPCDS schema created at scale 1500 that I have on a Hadoop/YARN cluster, and was not able to reproduce the difference in execution time between Spark 2 and Spark 3 that you report in your mail. This is the Spark

Re: Spark 3 is Slower than Spark 2 for TPCDS Q04 query.

2021-12-20 Thread Senthil Kumar
okia.com> wrote: > >> Hi Senthil, >> >> >> >> Which version of Spark 3 are we using? We had this kind of observation >> with Spark 3.0.2 and 3.1.x, but then we figured out that we had configured >> big value for spark.network.timeout and thi

Re: Spark 3 is Slower than Spark 2 for TPCDS Q04 query.

2021-12-19 Thread Senthil Kumar
@abhishek. We use spark 3.1* On Mon, 20 Dec 2021, 09:50 Rao, Abhishek (Nokia - IN/Bangalore), < abhishek@nokia.com> wrote: > Hi Senthil, > > > > Which version of Spark 3 are we using? We had this kind of observation > with Spark 3.0.2 and 3.1.x, but then we figured ou

RE: Spark 3 is Slower than Spark 2 for TPCDS Q04 query.

2021-12-19 Thread Rao, Abhishek (Nokia - IN/Bangalore)
Hi Senthil, Which version of Spark 3 are we using? We had this kind of observation with Spark 3.0.2 and 3.1.x, but then we figured out that we had configured big value for spark.network.timeout and this value was not taking effect in all releases prior to 3.0.2. This was fixed as part of https

Spark 3 is Slower than Spark 2 for TPCDS Q04 query.

2021-12-19 Thread Senthil Kumar
Hi All, We are comparing Spark 2.4.5 and Spark 3(without enabling spark 3 additional features) with TPCDS queries and found that Spark 3's performance is reduced to at-least 30-40% compared to Spark 2.4.5. Eg. Data size used 1TB Spark 2.4.5 finishes the Q4 in 1.5 min, but Spark 3.* takes

Re: Spark 3: Resource Discovery

2021-07-17 Thread Sean Owen
At the moment this is really about discovering GPUs, so that the scheduler can schedule tasks that need to allocate whole GPUs. On Sat, Jul 17, 2021 at 5:14 PM ayan guha wrote: > Hi > > As I was going through Spark 3 config params, I noticed following group of > params. I could no

Spark 3: Resource Discovery

2021-07-17 Thread ayan guha
Hi As I was going through Spark 3 config params, I noticed following group of params. I could not understand what are they for. Can anyone please point me in the right direction? spark.driver.resource.{resourceName}.amount 0 Amount of a particular resource type to use on the driver

Help Migrating BaseRelation to Spark 3.x

2021-06-03 Thread Fitch, Simeon
Hi, I'm the tech lead on RasterFrames, which adds geospatial raster data capability to Apache Spark SQL. We are trying to migrate to Spark 3.x, and are struggling with getting our various DataSources to work, and wondered if some might share some tips on what might be going on. Most of our issues

Spark 3: creating schema for hive metastore hangs forever

2020-08-13 Thread Tomas Bartalos
Hello, I'm using spark-3.0.0-bin-hadoop3.2 with custom hive metastore DB (postgres). I'm setting the "autoCreateAll" flag to true, so hive is creating its relational schema on first use. The problem is there is a deadlock and the query hangs forever: *Tx1* (*holds lock on TBLS relation*,

Re: Spark 3 pod template for the driver

2020-07-06 Thread Edward Mitchell
If I had to guess, it's likely because the Spark code would have to read the YAML to make sure the required parameters are set, and the way it's done was just easier to build on without a lot of refactoring. On Mon, Jul 6, 2020 at 5:06 PM Michel Sumbul wrote: > Thanks Edward for the reply! > >

Re: Spark 3 pod template for the driver

2020-07-01 Thread Edward Mitchell
Okay, I see what's going on here. Looks like the way that spark is coded, the driver container image (specified by --conf spark.kubernetes.driver.container.image) and executor container image (specified by --conf spark.kubernetes.executor.container.image) is required. If they're not specified

Re: Spark 3 pod template for the driver

2020-06-30 Thread Michel Sumbul
Hi Edeesis, The goal is to not have these settings in the spark submit command. If I specify the same things in a pod template for the executor, I still got the message: "Exception in thread "main" org.apache.spark.SparkException "Must specify the driver container image" it even don't try to

Re: Spark 3 pod template for the driver

2020-06-29 Thread edeesis
If I could muster a guess, you still need to specify the executor image. As is, this will only specify the driver image. You can specify it as --conf spark.kubernetes.container.image or --conf spark.kubernetes.executor.container.image -- Sent from:

Re: Spark 3 pod template for the driver

2020-06-29 Thread Michel Sumbul
Hello, Adding the dev mailing list maybe there is someone here that can help to have/show a valid/accepted pod template for spark 3? Thanks in advance, Michel Le ven. 26 juin 2020 à 14:03, Michel Sumbul a écrit : > Hi Jorge, > If I set that in the spark submit command it works but

Re: Initial Decom PR for Spark 3?

2020-06-22 Thread Hyukjin Kwon
on 님이 작성: >>>>> >>>>>> I'd be willing to pull this in, unless others have concerns post >>>>>> branch-cut. >>>>>> >>>>>> On Tue, Feb 4, 2020 at 2:51 PM Holden Karau >>>>>> wrote: >>

Re: Initial Decom PR for Spark 3?

2020-06-22 Thread Stephen Boesch
t;>>> >>>> 2020년 2월 9일 (일) 오전 1:23, Erik Erlandson 님이 작성: >>>> >>>>> I'd be willing to pull this in, unless others have concerns post >>>>> branch-cut. >>>>> >>>>> On Tue, Feb 4, 2020 at 2:51 PM Holden Karau >&

Re: Initial Decom PR for Spark 3?

2020-06-21 Thread Hyukjin Kwon
>>>> >>>>> Hi Y’all, >>>>> >>>>> I’ve got a K8s graceful decom PR ( >>>>> https://github.com/apache/spark/pull/26440 >>>>> ) I’d love to try and get in for Spark 3, but I don’t want to push on >>>>&g

Re: Initial Decom PR for Spark 3?

2020-06-20 Thread Holden Karau
u >>> wrote: >>> >>>> Hi Y’all, >>>> >>>> I’ve got a K8s graceful decom PR ( >>>> https://github.com/apache/spark/pull/26440 >>>> ) I’d love to try and get in for Spark 3, but I don’t want to push on >>>> it if folks

Re: Initial Decom PR for Spark 3?

2020-06-20 Thread Stephen Boesch
be willing to pull this in, unless others have concerns post >> branch-cut. >> >> On Tue, Feb 4, 2020 at 2:51 PM Holden Karau wrote: >> >>> Hi Y’all, >>> >>> I’ve got a K8s graceful decom PR ( >>> https://github.com/apache/spark/pull

Re: Initial Decom PR for Spark 3?

2020-06-18 Thread Stephen Boesch
e got a K8s graceful decom PR ( >>> https://github.com/apache/spark/pull/26440 >>> ) I’d love to try and get in for Spark 3, but I don’t want to push on >>> it if folks don’t think it’s worth it. I’ve been working on it since 2017 >>> and it was really close in

Re: Initial Decom PR for Spark 3?

2020-06-18 Thread Holden Karau
For follow up while I've backported this in some internal releases I'm not considering a candidate for backporting to Spark 3 anymore. I should have updated the thread with that. The design doc is linked in the PR. On Thu, Jun 18, 2020 at 6:05 PM Hyukjin Kwon wrote: > Looks it

Re: Initial Decom PR for Spark 3?

2020-06-18 Thread Hyukjin Kwon
>> I’ve got a K8s graceful decom PR ( >> https://github.com/apache/spark/pull/26440 >> ) I’d love to try and get in for Spark 3, but I don’t want to push on it >> if folks don’t think it’s worth it. I’ve been working on it since 2017 and >> it was really close in

Re: Datasource V2 support in Spark 3.x

2020-03-05 Thread Wenchen Fan
are using InternalRow). On Sat, Feb 29, 2020 at 8:39 AM Mihir Sahu wrote: > Hi Team, > > Wanted to know ahead of developing new datasource for Spark 3.x. Shall > it be done using Datasource V2 or Datasource V1(via Relation) or there is > any other plan. > > When I trie

Datasource V2 support in Spark 3.x

2020-02-28 Thread Mihir Sahu
Hi Team, Wanted to know ahead of developing new datasource for Spark 3.x. Shall it be done using Datasource V2 or Datasource V1(via Relation) or there is any other plan. When I tried to build datasource using V2 for Spark 3.0, I could not find the associated classes and they seems

Re: Initial Decom PR for Spark 3?

2020-02-08 Thread Erik Erlandson
I'd be willing to pull this in, unless others have concerns post branch-cut. On Tue, Feb 4, 2020 at 2:51 PM Holden Karau wrote: > Hi Y’all, > > I’ve got a K8s graceful decom PR ( > https://github.com/apache/spark/pull/26440 > ) I’d love to try and get in for Spark 3, but I don

Initial Decom PR for Spark 3?

2020-02-04 Thread Holden Karau
Hi Y’all, I’ve got a K8s graceful decom PR ( https://github.com/apache/spark/pull/26440 ) I’d love to try and get in for Spark 3, but I don’t want to push on it if folks don’t think it’s worth it. I’ve been working on it since 2017 and it was really close in November but then I had the crash

Re: Thoughts on Spark 3 release, or a preview release

2019-09-19 Thread Mats Rydberg
for 3.0! Best regards Mats, Martin Neo4j On Tue, Sep 17, 2019 at 8:35 PM Matt Cheah wrote: > I don’t know if it will be feasible to merge all of SPARK-25299 into Spark > 3. There are a number of APIs that will be submitted for review, and I > wouldn’t want to block the release on ne

Re: Thoughts on Spark 3 release, or a preview release

2019-09-17 Thread Matt Cheah
I don’t know if it will be feasible to merge all of SPARK-25299 into Spark 3. There are a number of APIs that will be submitted for review, and I wouldn’t want to block the release on negotiating these changes, as the decisions we make for each API can be pretty involved. Our original plan

Re: Thoughts on Spark 3 release, or a preview release

2019-09-17 Thread Xiao Li
t; - Use remote storage > for persisting shuffle data > https://issues.apache.org/jira/browse/SPARK-25299 > > If that is far enough along to get onto the roadmap. > > > On Wed, Sep 11, 2019 at 11:37 AM Sean Owen wrote: > >> I'm curious what current feelings are ab

Re: Thoughts on Spark 3 release, or a preview release

2019-09-16 Thread Erik Erlandson
wrote: > I'm curious what current feelings are about ramping down towards a > Spark 3 release. It feels close to ready. There is no fixed date, > though in the past we had informally tossed around "back end of 2019". > For reference, Spark 1 was May 2014, Spark 2 was July 2016.

Re: Thoughts on Spark 3 release, or a preview release

2019-09-16 Thread Michael Heuer
e the idea as a user and a DSv2 contributor. > > On Thu, Sep 12, 2019 at 4:10 PM Jungtaek Lim <mailto:kabh...@gmail.com>> wrote: > +1 (as a contributor) from me to have preview release on Spark 3 as it would > help to test the feature. When to cut preview release is questionable,

Re: Thoughts on Spark 3 release, or a preview release

2019-09-15 Thread Wenchen Fan
I don't expect to see a large DS V2 API change from now on. But we may update the API a little bit if we find problems during the preview. On Sat, Sep 14, 2019 at 10:16 PM Sean Owen wrote: > I don't think this suggests anything is finalized, including APIs. I > would not guess there will be

Re: Thoughts on Spark 3 release, or a preview release

2019-09-14 Thread Sean Owen
I don't think this suggests anything is finalized, including APIs. I would not guess there will be major changes from here though. On Fri, Sep 13, 2019 at 4:27 PM Andrew Melo wrote: > > Hi Spark Aficionados- > > On Fri, Sep 13, 2019 at 15:08 Ryan Blue wrote: >> >> +1 for a preview release. >>

Re: Thoughts on Spark 3 release, or a preview release

2019-09-13 Thread Andrew Melo
t; https://archive.apache.org/dist/spark/spark-2.0.0-preview/ >> >> And, thank you, Xingbo! >> Could you take a look at website generation? It seems to be broken on >> `master`. >> >> Bests, >> Dongjoon. >> >> >> On Fri, Sep 13, 2019 at 11:30

Re: Thoughts on Spark 3 release, or a preview release

2019-09-13 Thread Ryan Blue
n? It seems to be broken on > `master`. > > Bests, > Dongjoon. > > > On Fri, Sep 13, 2019 at 11:30 AM Xingbo Jiang > wrote: > >> Hi all, >> >> I would like to volunteer to be the release manager of Spark 3 preview, >> thanks! >> >> Sean Owen

Re: Thoughts on Spark 3 release, or a preview release

2019-09-13 Thread Dongjoon Hyun
: > Hi all, > > I would like to volunteer to be the release manager of Spark 3 preview, > thanks! > > Sean Owen 于2019年9月13日周五 上午11:21写道: > >> Well, great to hear the unanimous support for a Spark 3 preview >> release. Now, I don't know how to make releases myself :) I

Re: Thoughts on Spark 3 release, or a preview release

2019-09-13 Thread Xingbo Jiang
Hi all, I would like to volunteer to be the release manager of Spark 3 preview, thanks! Sean Owen 于2019年9月13日周五 上午11:21写道: > Well, great to hear the unanimous support for a Spark 3 preview > release. Now, I don't know how to make releases myself :) I would > first open it up to ou

Re: Thoughts on Spark 3 release, or a preview release

2019-09-13 Thread Sean Owen
Well, great to hear the unanimous support for a Spark 3 preview release. Now, I don't know how to make releases myself :) I would first open it up to our revered release managers: would anyone be interested in trying to make one? sounds like it's not too soon to get what's in master out

Re: Thoughts on Spark 3 release, or a preview release

2019-09-13 Thread Ilan Filonenko
gt; testing against and exploring so they can raise issues with us earlier in >>>>> the process and we have more time to make calls around this. >>>>> >>>>> On Thu, Sep 12, 2019 at 4:15 PM John Zhuge wrote: >>>>> >>>>> +1 Like

Re: Thoughts on Spark 3 release, or a preview release

2019-09-13 Thread Thomas Graves
ks something to start testing >>>> against and exploring so they can raise issues with us earlier in the >>>> process and we have more time to make calls around this. >>>> >>>> On Thu, Sep 12, 2019 at 4:15 PM John Zhuge wrote: >>&

Re: Thoughts on Spark 3 release, or a preview release

2019-09-13 Thread Stavros Kontopoulos
an-up, it will meet the timeline and `3.0.0-preview` helps >>> it a lot. >>> >>> After this discussion, can we have some timeline for `Spark 3.0 Release >>> Window` in our versioning-policy page? >>> >>> - https://spark.apache.org/versioning-polic

Re: Thoughts on Spark 3 release, or a preview release

2019-09-13 Thread Driesprong, Fokko
tributor. >> >> On Thu, Sep 12, 2019 at 4:10 PM Jungtaek Lim wrote: >> >> +1 (as a contributor) from me to have preview release on Spark 3 as it >> would help to test the feature. When to cut preview release is >> questionable, as major works are ideally

Re: Thoughts on Spark 3 release, or a preview release

2019-09-12 Thread Reynold Xin
taek Lim < kabhwan@ gmail. com ( >> kabh...@gmail.com ) > wrote: >> >> >>> +1 (as a contributor) from me to have preview release on Spark 3 as it >>> would help to test the feature. When to cut preview release is >>> questionable, as major w

Re: Thoughts on Spark 3 release, or a preview release

2019-09-12 Thread Holden Karau
Sv2 contributor. > > On Thu, Sep 12, 2019 at 4:10 PM Jungtaek Lim wrote: > >> +1 (as a contributor) from me to have preview release on Spark 3 as it >> would help to test the feature. When to cut preview release is >> questionable, as major works are ideally to be done befor

Re: Thoughts on Spark 3 release, or a preview release

2019-09-12 Thread Matt Cheah
+1 as both a contributor and a user. From: John Zhuge Date: Thursday, September 12, 2019 at 4:15 PM To: Jungtaek Lim Cc: Jean Georges Perrin , Hyukjin Kwon , Dongjoon Hyun , dev Subject: Re: Thoughts on Spark 3 release, or a preview release +1 Like the idea as a user and a DSv2

Re: Thoughts on Spark 3 release, or a preview release

2019-09-12 Thread John Zhuge
+1 Like the idea as a user and a DSv2 contributor. On Thu, Sep 12, 2019 at 4:10 PM Jungtaek Lim wrote: > +1 (as a contributor) from me to have preview release on Spark 3 as it > would help to test the feature. When to cut preview release is > questionable, as major works are ideally t

Re: Thoughts on Spark 3 release, or a preview release

2019-09-12 Thread Jungtaek Lim
+1 (as a contributor) from me to have preview release on Spark 3 as it would help to test the feature. When to cut preview release is questionable, as major works are ideally to be done before that - if we are intended to introduce new features before official release, that should work regardless

Re: Thoughts on Spark 3 release, or a preview release

2019-09-11 Thread Jean Georges Perrin
> https://github.com/apache/spark/pull/24851 >>> https://github.com/apache/spark/pull/24297 >>> >>>michael >>> >>> >>>> On Sep 11, 2019, at 1:37 PM, Sean Owen wrote: >>>> >>>> I'm curious what current feelin

Re: Thoughts on Spark 3 release, or a preview release

2019-09-11 Thread Michael Heuer
park/pull/24851> https://github.com/apache/spark/pull/24297 <https://github.com/apache/spark/pull/24297> michael > On Sep 11, 2019, at 1:37 PM, Sean Owen wrote: > > I'm curious what current feelings are about ramping down towards a > Spark 3 release. It feels close to rea

Thoughts on Spark 3 release, or a preview release

2019-09-11 Thread Sean Owen
I'm curious what current feelings are about ramping down towards a Spark 3 release. It feels close to ready. There is no fixed date, though in the past we had informally tossed around "back end of 2019". For reference, Spark 1 was May 2014, Spark 2 was July 2016. I'd expect Spark 2 to l

Re: Remove non-Tungsten mode in Spark 3?

2019-01-09 Thread Sean Owen
I haven't touched Tungsten, but have proposed removing the deprecated old memory manager and settings -- yes I think that's the primary argument for it. https://github.com/apache/spark/pull/23457 On Wed, Jan 9, 2019 at 6:06 PM Erik Erlandson wrote: > Removing the user facing config seems like a

Re: Remove non-Tungsten mode in Spark 3?

2019-01-09 Thread Erik Erlandson
;> >>> I know we didn't deprecate it, but it's been off by default for a long >>> time. It could be deprecated, too. >>> >>> Same question for spark.memory.useLegacyMode and all its various >>> associated settings? Seems like these should go away at some point, a

Re: Remove non-Tungsten mode in Spark 3?

2019-01-04 Thread Sean Owen
M, Sean Owen wrote: > >> Just wondering if there is a good reason to keep around the pre-tungsten >> on-heap memory mode for Spark 3, and make spark.memory.offHeap.enabled >> always true? It would simplify the code somewhat, but I don't feel I'm so >> aware of the tradeoff

Re: Remove non-Tungsten mode in Spark 3?

2019-01-03 Thread Reynold Xin
few users need it. How much code does it remove though? On Thu, Jan 03, 2019 at 2:55 PM, Sean Owen < sro...@apache.org > wrote: > > > > Just wondering if there is a good reason to keep around the pre-tungsten > on-heap memory mode for Spark 3, and make spark.memory.off

Remove non-Tungsten mode in Spark 3?

2019-01-03 Thread Sean Owen
Just wondering if there is a good reason to keep around the pre-tungsten on-heap memory mode for Spark 3, and make spark.memory.offHeap.enabled always true? It would simplify the code somewhat, but I don't feel I'm so aware of the tradeoffs. I know we didn't deprecate it, but it's been off

Removing deprecated items in Spark 3

2018-11-01 Thread Sean Owen
I took a pass at removing most of the older deprecated items in Spark. For discussion: https://github.com/apache/spark/pull/22921 - To unsubscribe e-mail: dev-unsubscr...@spark.apache.org

Re: Starting to make changes for Spark 3 -- what can we delete?

2018-10-17 Thread DB Tsai
I'll +1 on removing those legacy mllib code. Many users are confused about the APIs, and some of them have weird behaviors (for example, in gradient descent, the intercept is regularized which supports not to). DB Tsai | Siri Open Source Technologies [not a contribution] |  Apple, Inc >

Re: Starting to make changes for Spark 3 -- what can we delete?

2018-10-17 Thread Erik Erlandson
My understanding was that the legacy mllib api was frozen, with all new dev going to ML, but it was not going to be removed. Although removing it would get rid of a lot of `OldXxx` shims. On Wed, Oct 17, 2018 at 12:55 AM Marco Gaido wrote: > Hi all, > > I think a very big topic on this would

Re: Starting to make changes for Spark 3 -- what can we delete?

2018-10-17 Thread Marco Gaido
Hi all, I think a very big topic on this would be: what do we want to do with the old mllib API? For long I have been told that it was going to be removed on 3.0. Is this still the plan? Thanks, Marco Il giorno mer 17 ott 2018 alle ore 03:11 Marcelo Vanzin ha scritto: > Might be good to take

Re: Starting to make changes for Spark 3 -- what can we delete?

2018-10-16 Thread Marcelo Vanzin
Might be good to take a look at things marked "@DeveloperApi" and whether they should stay that way. e.g. I was looking at SparkHadoopUtil and I've always wanted to just make it private to Spark. I don't see why apps would need any of those methods. On Tue, Oct 16, 2018 at 10:18 AM Sean Owen

Starting to make changes for Spark 3 -- what can we delete?

2018-10-16 Thread Sean Owen
There was already agreement to delete deprecated things like Flume and Kafka 0.8 support in master. I've got several more on my radar, and wanted to highlight them and solicit general opinions on where we should accept breaking changes. For example how about removing accumulator v1?

Re: Spark 3

2018-01-20 Thread Matei Zaharia
Jan 19, 2018, at 10:53 AM, Koert Kuipers <ko...@tresata.com> wrote: >> >> i was expecting to be able to move to scala 2.12 sometime this year >> >> if this cannot be done in spark 2.x then that could be a compelling reason >> to move spark 3 up to 2018 i think

Re: Spark 3

2018-01-19 Thread Reynold Xin
; Thanks, > Justin > > > On Jan 19, 2018, at 10:53 AM, Koert Kuipers <ko...@tresata.com> wrote: > > i was expecting to be able to move to scala 2.12 sometime this year > > if this cannot be done in spark 2.x then that could be a compelling reason > to move spark 3 up to 2018

Re: Spark 3

2018-01-19 Thread Justin Miller
> if this cannot be done in spark 2.x then that could be a compelling reason to > move spark 3 up to 2018 i think > > hadoop 3 sounds great but personally i have no use case for it yet > > On Fri, Jan 19, 2018 at 12:31 PM, Sean Owen <so...@cloudera.com > <mailto:so..

Re: Spark 3

2018-01-19 Thread Koert Kuipers
i was expecting to be able to move to scala 2.12 sometime this year if this cannot be done in spark 2.x then that could be a compelling reason to move spark 3 up to 2018 i think hadoop 3 sounds great but personally i have no use case for it yet On Fri, Jan 19, 2018 at 12:31 PM, Sean Owen <

Re: Spark 3

2018-01-19 Thread Ryan Blue
top > of everyone else's mind? > > On Jan 20, 2018 6:32 AM, "Sean Owen" <so...@cloudera.com> wrote: > >> Forking this thread to muse about Spark 3. Like Spark 2, I assume it >> would be more about making all those accumulated breaking changes and >> up

Re: Spark 3

2018-01-19 Thread Sean Owen
ain and see if we can do something with then. What's top > of everyone else's mind? > > On Jan 20, 2018 6:32 AM, "Sean Owen" <so...@cloudera.com> wrote: > >> Forking this thread to muse about Spark 3. Like Spark 2, I assume it >> would be more about making all those a

Re: Spark 3

2018-01-19 Thread Holden Karau
if we can do something with then. What's top of everyone else's mind? On Jan 20, 2018 6:32 AM, "Sean Owen" <so...@cloudera.com> wrote: > Forking this thread to muse about Spark 3. Like Spark 2, I assume it would > be more about making all those accumulated breaking chang

Spark 3

2018-01-19 Thread Sean Owen
Forking this thread to muse about Spark 3. Like Spark 2, I assume it would be more about making all those accumulated breaking changes and updating lots of dependencies. Hadoop 3 looms large in that list as well as Scala 2.12. Spark 1 was release in May 2014, and Spark 2 in July 2016. If Spark