I think so ;)
Regards
JB
On 11/10/2017 09:29 AM, Reuven Lax wrote:
Sounds good. I doubt we will have much opposition from users, in which case
Beam 2.3.0 can deprecate Spark 1.x
On Thu, Nov 9, 2017 at 11:54 PM, Jean-Baptiste Onofré
wrote:
Hi all,
thanks a lot for all
Sounds good. I doubt we will have much opposition from users, in which case
Beam 2.3.0 can deprecate Spark 1.x
On Thu, Nov 9, 2017 at 11:54 PM, Jean-Baptiste Onofré
wrote:
> Hi all,
>
> thanks a lot for all your feedback.
>
> The trend is about to upgrade to Spark 2.x and
Hi all,
thanks a lot for all your feedback.
The trend is about to upgrade to Spark 2.x and drop Spark 1.x support.
However, some of you (especially Reuven and Robert) commented that users have to
be pinged as well. It makes perfect sense, and it was my intention.
I propose the following
On Thu, Nov 9, 2017 at 11:05 AM, Kenneth Knowles
wrote:
> I think it makes sense to communicate with email to users@ and in the
> release notes of 2.2.0.
Totally agree.
> That communication should be specific and indicate
> whether we are planning to merely not work on
+1 from me. However let's notify users@ first. If we do get a lot of
pushback from users (which I doubt we will), we might reconsider dropping
Spark 1 support.
On Thu, Nov 9, 2017 at 11:05 AM, Kenneth Knowles
wrote:
> +1 from me, with a friendly deprecation process
>
>
+1 from me, with a friendly deprecation process
I am convinced by the following:
- We don't have the resources to make both great, and anyhow it isn't
worth it
- People keeping up with Beam releases are likely to be keeping up with
Spark as well
- Spark 1 users already have a Spark 1 runner
+1 for dropping Spark 1 support.
I don't think we have enough users to justify supporting both, and its been
a long time since this idea originally came-up (when Spark2 wasn't stable)
and now Spark 2 is standard in all Hadoop distros.
As for switching to the Dataframe API, as long as Spark 2
+1 for the move to Spark 2 modulo preventing users and deciding on support:
I agree that having compatibility for both versions of Spark is
desirable but I am not sure if is worth the effort. Apart of the
reasons mentioned by Holden and Pei, I will add that the burden of
simultaneous maintenance
+1 on moving forward with Spark 2.x only.
Spark 1 users can still use already released Spark runners, and we can
support them with minor version releases for future bug fixes.
I don't see how important it is to make future Beam releases available to
Spark 1 users. If they choose not to upgrade
That's a good point about Oozie does only supporting only Spark 1 or 2 at a
time on a cluster -- but do we know people using Oozie and Spark 1 that
would still be using Spark 1 by the time of the next BEAM release? The last
Spark 1 release was a year ago (and last non-maintenance release almost 20
I don't know if ditching Spark 1 out right right now would be a great move
given that a lot of the main support applications around spark haven't yet
fully moved to Spark 2 yet. Yet alone have support for having a cluster
with both. Oozie for example is still pre stable release for their Spark 1
Also, upgrading Spark 1 to 2 is generally easier than changing JVM
versions. For folks using YARN or the hosted environments it pretty much
trivial since you can effectively have distinct Spark clusters for each job.
On Wed, Nov 8, 2017 at 9:19 PM, Holden Karau wrote:
>
Having both Spark1 and Spark2 modules would benefit wider user base.
I would vote for that.
Cheers
On Wed, Nov 8, 2017 at 12:51 AM, Jean-Baptiste Onofré
wrote:
> Hi Robert,
>
> Thanks for your feedback !
>
> From an user perspective, with the current state of the PR, the
I'm generally a -0.5 on this change, or at least doing so hastily.
As with dropping Java 7 support, I think this should at least be
announced in release notes that we're considering dropping support in
the subsequent release, as this dev list likely does not reach a
substantial portion of the
14 matches
Mail list logo