Thanks everyone. As Nathan suggested, I ended up collecting the distinct
keys first and then assigning Ids to each key explicitly.
Regards
Sumit Chawla
On Fri, Jun 22, 2018 at 7:29 AM, Nathan Kronenfeld <
nkronenfeld@uncharted.software> wrote:
> On Thu, Jun 21, 2018 at 4:51 PM, Chawla,Sumit
Please vote on releasing the following candidate as Apache Spark version 2.1.3.
The vote is open until Fri, June 29th @ 9PM UTC (2PM PDT) and passes if a
majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
[ ] +1 Release this package as Apache Spark 2.1.3
[ ] -1 Do not release this pack
Starting with my own +1.
On Tue, Jun 26, 2018 at 1:25 PM, Marcelo Vanzin wrote:
> Please vote on releasing the following candidate as Apache Spark version
> 2.1.3.
>
> The vote is open until Fri, June 29th @ 9PM UTC (2PM PDT) and passes if a
> majority +1 PMC votes are cast, with a minimum of 3
following up after a ref to this in
https://issues.apache.org/jira/browse/HADOOP-15559
the AWS SDK is a very fast moving project, with a release cycle of ~2 weeks,
but it's in the state Fred Brooks described, "the number of bugs is constant,
they just move around"; bumpin gup an AWS release is
HoldenK and interested folks,
Am just following up on the spark model serving discussions as this is highly
relevant to what I’m embarking on at work. Is there a concrete list of next
steps or can someone summarize what was discussed at the summit , would love to
have a Seattle version of this