[ML]Random Forest Error : Size exceeds Integer.MAX_VALUE

2016-10-04 Thread Samkit Shah
Hello folks,
I am running Random Forest from ml from spark 1.6.1 on bimbo[1] dataset
with following configurations:

"-Xms16384M" "-Xmx16384M" "-Dspark.locality.wait=0s"
"-Dspark.driver.extraJavaOptions=-Xss10240k -XX:+PrintGCDetails
-XX:+PrintGCTimeStamps -XX:+PrintTenuringDistribution
-XX:+UseConcMarkSweepGC -XX:+UseParNewGC -XX:ParallelGCThreads=2
-XX:-UseAdaptiveSizePolicy -XX:ConcGCThreads=2 -XX:-UseGCOverheadLimit
 -XX:CMSInitiatingOccupancyFraction=75 -XX:NewSize=8g -XX:MaxNewSize=8g
-XX:SurvivorRatio=3 -DnumPartitions=36" "-Dspark.submit.deployMode=cluster"
"-Dspark.speculation=true" "-Dspark.speculation.multiplier=2"
"-Dspark.driver.memory=16g" "-Dspark.speculation.interval=300ms"
 "-Dspark.speculation.quantile=0.5" "-Dspark.akka.frameSize=768"
"-Dspark.driver.supervise=false" "-Dspark.executor.cores=6"
"-Dspark.executor.extraJavaOptions=-Xss10240k -XX:+PrintGCDetails
-XX:+PrintGCTimeStamps -XX:+PrintTenuringDistribution
-XX:-UseAdaptiveSizePolicy -XX:+UseParallelGC -XX:+UseParallelOldGC
-XX:ParallelGCThreads=6 -XX:NewSize=22g -XX:MaxNewSize=22g
-XX:SurvivorRatio=2 -XX:+PrintAdaptiveSizePolicy -XX:+PrintGCDateStamps"
"-Dspark.rpc.askTimeout=10" "-Dspark.executor.memory=40g"
"-Dspark.driver.maxResultSize=3g" "-Xss10240k" "-XX:+PrintGCDetails"
"-XX:+PrintGCTimeStamps" "-XX:+PrintTenuringDistribution"
"-XX:+UseConcMarkSweepGC" "-XX:+UseParNewGC" "-XX:ParallelGCThreads=2"
"-XX:-UseAdaptiveSizePolicy" "-XX:ConcGCThreads=2"
"-XX:-UseGCOverheadLimit" "-XX:CMSInitiatingOccupancyFraction=75"
"-XX:NewSize=8g" "-XX:MaxNewSize=8g" "-XX:SurvivorRatio=3"
"-DnumPartitions=36" "org.apache.spark.deploy.worker.DriverWrapper"
"spark://Worker@11.0.0.106:56419"


I get following error:
16/10/04 06:55:05 WARN TaskSetManager: Lost task 8.0 in stage 19.0 (TID
194, 11.0.0.106): java.lang.IllegalArgumentException: Size exceeds
Integer.MAX_VALUE
at sun.nio.ch.FileChannelImpl.map(FileChannelImpl.java:869)
at
org.apache.spark.storage.DiskStore$$anonfun$getBytes$2.apply(DiskStore.scala:127)
at
org.apache.spark.storage.DiskStore$$anonfun$getBytes$2.apply(DiskStore.scala:115)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1250)
at org.apache.spark.storage.DiskStore.getBytes(DiskStore.scala:129)
at org.apache.spark.storage.DiskStore.getBytes(DiskStore.scala:136)
at org.apache.spark.storage.BlockManager.doGetLocal(BlockManager.scala:503)
at org.apache.spark.storage.BlockManager.getLocal(BlockManager.scala:420)
at org.apache.spark.storage.BlockManager.get(BlockManager.scala:625)
at org.apache.spark.CacheManager.putInBlockManager(CacheManager.scala:154)
at org.apache.spark.CacheManager.getOrCompute(CacheManager.scala:78)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:268)
at
org.apache.spark.rdd.ZippedPartitionsRDD2.compute(ZippedPartitionsRDD.scala:88)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
at
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)


I have varied number of partitions from 24 to 48. I still get the same
error. How can this problem be tackled?


Thanks,
Samkit




[1]: https://www.kaggle.com/c/grupo-bimbo-inventory-demand


Re: welcoming Xiao Li as a committer

2016-10-04 Thread Kousuke Saruta

Congratulations Xiao!

- Kousuke

On 2016/10/05 7:44, Bryan Cutler wrote:

Congrats Xiao!

On Tue, Oct 4, 2016 at 11:14 AM, Holden Karau > wrote:


Congratulations :D :) Yay!

On Tue, Oct 4, 2016 at 11:14 AM, Suresh Thalamati
>
wrote:

Congratulations, Xiao!



> On Oct 3, 2016, at 10:46 PM, Reynold Xin
> wrote:
>
> Hi all,
>
> Xiao Li, aka gatorsmile, has recently been elected as an
Apache Spark committer. Xiao has been a super active
contributor to Spark SQL. Congrats and welcome, Xiao!
>
> - Reynold
>


-
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org





-- 
Cell : 425-233-8271 

Twitter: https://twitter.com/holdenkarau







Re: welcoming Xiao Li as a committer

2016-10-04 Thread Bryan Cutler
Congrats Xiao!

On Tue, Oct 4, 2016 at 11:14 AM, Holden Karau  wrote:

> Congratulations :D :) Yay!
>
> On Tue, Oct 4, 2016 at 11:14 AM, Suresh Thalamati <
> suresh.thalam...@gmail.com> wrote:
>
>> Congratulations, Xiao!
>>
>>
>>
>> > On Oct 3, 2016, at 10:46 PM, Reynold Xin  wrote:
>> >
>> > Hi all,
>> >
>> > Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark
>> committer. Xiao has been a super active contributor to Spark SQL. Congrats
>> and welcome, Xiao!
>> >
>> > - Reynold
>> >
>>
>>
>> -
>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>
>>
>
>
> --
> Cell : 425-233-8271
> Twitter: https://twitter.com/holdenkarau
>


Re: StructuredStreaming Custom Sinks (motivated by Structured Streaming Machine Learning)

2016-10-04 Thread Michael Armbrust
>
> I don't quite understand why exposing it indirectly through a typed
> interface should be delayed before finalizing the API.
>

Spark has a long history
 of maintaining
binary compatibility in its public APIs.  I strongly believe this is one of
the things that has made the project successful.  Exposing internals that
we know are going to change in the primary user facing API for creating
Streaming DataFrames seems directly counter to this goal.  I think the
argument that "you can do it anyway" fails to capture user expectations who
probably aren't closely following this discussion.

If advanced users want to dig though the code and experiment, great.  I
hope they report back on whats good and what can be improved.  However, if
you add the function suggested in the PR to DataStreamReader, you are
giving them a bad experience by leaking internals that don't even show up
in the published documentation.


Re: [ANNOUNCE] Announcing Spark 2.0.1

2016-10-04 Thread Reynold Xin
They have been published yesterday, but can take a while to propagate.


On Tue, Oct 4, 2016 at 12:58 PM, Prajwal Tuladhar  wrote:

> Hi,
>
> It seems like, 2.0.1 artifact hasn't been published to Maven Central. Can
> anyone confirm?
>
> On Tue, Oct 4, 2016 at 5:39 PM, Reynold Xin  wrote:
>
>> We are happy to announce the availability of Spark 2.0.1!
>>
>> Apache Spark 2.0.1 is a maintenance release containing 300 stability and
>> bug fixes. This release is based on the branch-2.0 maintenance branch of
>> Spark. We strongly recommend all 2.0.0 users to upgrade to this stable
>> release.
>>
>> To download Apache Spark 2.0.1, visit http://spark.apache.org/downlo
>> ads.html
>>
>> We would like to acknowledge all community members for contributing
>> patches to this release.
>>
>>
>>
>
>
> --
> --
> Cheers,
> Praj
>


Re: [ANNOUNCE] Announcing Spark 2.0.1

2016-10-04 Thread Prajwal Tuladhar
Hi,

It seems like, 2.0.1 artifact hasn't been published to Maven Central. Can
anyone confirm?

On Tue, Oct 4, 2016 at 5:39 PM, Reynold Xin  wrote:

> We are happy to announce the availability of Spark 2.0.1!
>
> Apache Spark 2.0.1 is a maintenance release containing 300 stability and
> bug fixes. This release is based on the branch-2.0 maintenance branch of
> Spark. We strongly recommend all 2.0.0 users to upgrade to this stable
> release.
>
> To download Apache Spark 2.0.1, visit http://spark.apache.org/
> downloads.html
>
> We would like to acknowledge all community members for contributing
> patches to this release.
>
>
>


-- 
--
Cheers,
Praj


Re: StructuredStreaming Custom Sinks (motivated by Structured Streaming Machine Learning)

2016-10-04 Thread Jakob Odersky
Hi everyone,

is there any ongoing discussion/documentation on the redesign of sinks?
I think it could be a good thing to abstract away the underlying
streaming model, however that isn't directly related to Holden's first
point. The way I understand it, is to slightly change the
DataStreamWriter API (the thing that's returned when you call
"df.writeStream") to allow passing in a custom sink provider instead
of only accepting strings. This would allow users to write their own
providers and sinks, and give them a strongly typed, possibly generic
way to do so. The sink api is currently available to users indirectly
(you can create your own sink provider and load it with the built-in
DataSource reflection functionality), therefore I don't quite
understand why exposing it indirectly through a typed interface should
be delayed before finalizing the API.
On a side note, I saw that sources have a similar limitation in that
they are currently only available through a stringly-typed interface.
Could a similar solution be applied to sources? Maybe the writer and
reader api's could even be unified to a certain degree.

Shivaram, I like your ideas on the proposed redesign! Can we discuss
this further?

cheers,
--Jakob


On Mon, Sep 26, 2016 at 5:12 PM, Shivaram Venkataraman
 wrote:
> Disclaimer - I am not very closely involved with Structured Streaming
> design / development, so this is just my two cents from looking at the
> discussion in the linked JIRAs and PRs.
>
> It seems to me there are a couple of issues being conflated here: (a)
> is the question of how to specify or add more functionality to the
> Sink API such as ability to get model updates back to the driver [A
> design issue IMHO] (b) question of how to pass parameters to
> DataFrameWriter, esp. strings vs. typed objects and whether the API is
> stable vs. experimental
>
> TLDR is that I think we should first focus on refactoring the Sink and
> add new functionality after that. Detailed comments below.
>
> Sink design / functionality: Looking at SPARK-10815, a JIRA linked
> from SPARK-16407, it looks like the existing Sink API is limited
> because it is tied to the RDD/Dataframe definitions. It also has
> surprising limitations like not being able to run operators on `data`
> and only using `collect/foreach`.  Given these limitations, I think it
> makes sense to redesign the Sink API first *before* adding new
> functionality to the existing Sink. I understand that we have not
> marked this experimental in 2.0.0 -- but I guess since
> StructuredStreaming is new as a whole, so we can probably break the
> Sink API in a upcoming 2.1.0 release.
>
> As a part of the redesign, I think we need to do two things: (i) come
> up with a new data handle that separates RDD from what is passed to
> the Sink (ii) Have some way to specify code that can run on the
> driver. This might not be an issue if the data handle already has
> clean abstraction for this.
>
> Micro-batching: Ideally it would be good to not expose the micro-batch
> processing model in the Sink API as this might change going forward.
> Given the consistency model we are presenting I think there will be
> some notion of batch / time-range identifier in the API. But I think
> if we can avoid having hard constraints on where functions will get
> run (i.e. on the driver vs. as a part of a job etc.) and when
> functions will get run (i.e. strictly after every micro-batch) it
> might give us more freedom in improving performance going forward [1].
>
> Parameter passing: I think your point that typed is better than
> untyped is pretty good and supporting both APIs isn't necessarily bad
> either. My understand of the discussion around this is that we should
> do this after Sink is refactored to avoid exposing the old APIs ?
>
> Thanks
> Shivaram
>
> [1] FWIW this is something I am looking at and
> https://spark-summit.org/2016/events/low-latency-execution-for-apache-spark/
> has some details about this.
>
>
> On Mon, Sep 26, 2016 at 1:38 PM, Holden Karau  wrote:
>> Hi Spark Developers,
>>
>>
>> After some discussion on SPARK-16407 (and on the PR) we’ve decided to jump
>> back to the developer list (SPARK-16407 itself comes from our early work on
>> SPARK-16424 to enable ML with the new Structured Streaming API). SPARK-16407
>> is proposing to extend the current DataStreamWriter API to allow users to
>> specify a specific instance of a StreamSinkProvider - this makes it easier
>> for users to create sinks that are configured with things besides strings
>> (for example things like lambdas). An example of something like this already
>> inside Spark is the ForeachSink.
>>
>>
>> We have been working on adding support for online learning in Structured
>> Streaming, similar to what Spark Streaming and MLLib provide today. Details
>> are available in  SPARK-16424. Along the way, we noticed that there is
>> currently no way for code running in the driver to access the 

Re: welcoming Xiao Li as a committer

2016-10-04 Thread Holden Karau
Congratulations :D :) Yay!

On Tue, Oct 4, 2016 at 11:14 AM, Suresh Thalamati <
suresh.thalam...@gmail.com> wrote:

> Congratulations, Xiao!
>
>
>
> > On Oct 3, 2016, at 10:46 PM, Reynold Xin  wrote:
> >
> > Hi all,
> >
> > Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark
> committer. Xiao has been a super active contributor to Spark SQL. Congrats
> and welcome, Xiao!
> >
> > - Reynold
> >
>
>
> -
> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>
>


-- 
Cell : 425-233-8271
Twitter: https://twitter.com/holdenkarau


Re: welcoming Xiao Li as a committer

2016-10-04 Thread Suresh Thalamati
Congratulations, Xiao!



> On Oct 3, 2016, at 10:46 PM, Reynold Xin  wrote:
> 
> Hi all,
> 
> Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark 
> committer. Xiao has been a super active contributor to Spark SQL. Congrats 
> and welcome, Xiao!
> 
> - Reynold
> 


-
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org



Re: welcoming Xiao Li as a committer

2016-10-04 Thread Shixiong(Ryan) Zhu
Congrats!

On Tue, Oct 4, 2016 at 9:09 AM, Yanbo Liang  wrote:

> Congrats and welcome!
>
> On Tue, Oct 4, 2016 at 9:01 AM, Herman van Hövell tot Westerflier <
> hvanhov...@databricks.com> wrote:
>
>> Congratulations Xiao! Very well deserved!
>>
>> On Mon, Oct 3, 2016 at 10:46 PM, Reynold Xin  wrote:
>>
>>> Hi all,
>>>
>>> Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark
>>> committer. Xiao has been a super active contributor to Spark SQL. Congrats
>>> and welcome, Xiao!
>>>
>>> - Reynold
>>>
>>>
>>
>


[ANNOUNCE] Announcing Spark 2.0.1

2016-10-04 Thread Reynold Xin
We are happy to announce the availability of Spark 2.0.1!

Apache Spark 2.0.1 is a maintenance release containing 300 stability and
bug fixes. This release is based on the branch-2.0 maintenance branch of
Spark. We strongly recommend all 2.0.0 users to upgrade to this stable
release.

To download Apache Spark 2.0.1, visit http://spark.apache.org/downloads.html

We would like to acknowledge all community members for contributing patches
to this release.


Edit access for spark confluence wiki

2016-10-04 Thread Mridul Muralidharan
Can someone add me to edit list for the spark wiki please ?

Thanks,
Mridul

-
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org



Re: welcoming Xiao Li as a committer

2016-10-04 Thread Yanbo Liang
Congrats and welcome!

On Tue, Oct 4, 2016 at 9:01 AM, Herman van Hövell tot Westerflier <
hvanhov...@databricks.com> wrote:

> Congratulations Xiao! Very well deserved!
>
> On Mon, Oct 3, 2016 at 10:46 PM, Reynold Xin  wrote:
>
>> Hi all,
>>
>> Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark
>> committer. Xiao has been a super active contributor to Spark SQL. Congrats
>> and welcome, Xiao!
>>
>> - Reynold
>>
>>
>


Re: welcoming Xiao Li as a committer

2016-10-04 Thread Herman van Hövell tot Westerflier
Congratulations Xiao! Very well deserved!

On Mon, Oct 3, 2016 at 10:46 PM, Reynold Xin  wrote:

> Hi all,
>
> Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark
> committer. Xiao has been a super active contributor to Spark SQL. Congrats
> and welcome, Xiao!
>
> - Reynold
>
>


Re: welcoming Xiao Li as a committer

2016-10-04 Thread Weiqing Yang
Congrats Xiao!

On Tue, Oct 4, 2016 at 6:40 AM, Kevin  wrote:

> Congratulations Xiao!!
>
> Sent from my iPhone
>
> On Oct 4, 2016, at 3:59 AM, Tarun Kumar  wrote:
>
> Congrats Xiao.
>
> Thanks
> Tarun
> On Tue, 4 Oct 2016 at 12:57 PM, Cheng Lian  wrote:
>
>> Congratulations!!!
>>
>>
>> Cheng
>>
>> On Tue, Oct 4, 2016 at 1:46 PM, Reynold Xin  wrote:
>>
>> Hi all,
>>
>> Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark
>> committer. Xiao has been a super active contributor to Spark SQL. Congrats
>> and welcome, Xiao!
>>
>> - Reynold
>>
>>
>>


Re: welcoming Xiao Li as a committer

2016-10-04 Thread Kevin
Congratulations Xiao!!

Sent from my iPhone

> On Oct 4, 2016, at 3:59 AM, Tarun Kumar  wrote:
> 
> Congrats Xiao.
> 
> Thanks
> Tarun
>> On Tue, 4 Oct 2016 at 12:57 PM, Cheng Lian  wrote:
>> Congratulations!!!
>> 
>> 
>> Cheng
>> 
>> On Tue, Oct 4, 2016 at 1:46 PM, Reynold Xin  wrote:
>> Hi all,
>> 
>> Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark 
>> committer. Xiao has been a super active contributor to Spark SQL. Congrats 
>> and welcome, Xiao!
>> 
>> - Reynold


Re: welcoming Xiao Li as a committer

2016-10-04 Thread Tarun Kumar
Congrats Xiao.

Thanks
Tarun
On Tue, 4 Oct 2016 at 12:57 PM, Cheng Lian  wrote:

> Congratulations!!!
>
>
> Cheng
>
> On Tue, Oct 4, 2016 at 1:46 PM, Reynold Xin  wrote:
>
> Hi all,
>
> Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark
> committer. Xiao has been a super active contributor to Spark SQL. Congrats
> and welcome, Xiao!
>
> - Reynold
>
>
>


Re: welcoming Xiao Li as a committer

2016-10-04 Thread Cheng Lian
Congratulations!!!

Cheng

On Tue, Oct 4, 2016 at 1:46 PM, Reynold Xin  wrote:

> Hi all,
>
> Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark
> committer. Xiao has been a super active contributor to Spark SQL. Congrats
> and welcome, Xiao!
>
> - Reynold
>
>


Re: welcoming Xiao Li as a committer

2016-10-04 Thread Marcin Tustin
Congratulations Xiao 

On Tuesday, October 4, 2016, Reynold Xin  wrote:

> Hi all,
>
> Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark
> committer. Xiao has been a super active contributor to Spark SQL. Congrats
> and welcome, Xiao!
>
> - Reynold
>
>

-- 
Want to work at Handy? Check out our culture deck and open roles 

Latest news  at Handy
Handy just raised $50m 

 led 
by Fidelity



Re: welcoming Xiao Li as a committer

2016-10-04 Thread Denny Lee
Congrats, Xiao!
On Tue, Oct 4, 2016 at 00:00 Takeshi Yamamuro  wrote:

> congrats, xiao!
>
> On Tue, Oct 4, 2016 at 3:59 PM, Hyukjin Kwon  wrote:
>
> Congratulations!
>
> 2016-10-04 15:51 GMT+09:00 Dilip Biswal :
>
> Hi Xiao,
>
> Congratulations Xiao !!  This is indeed very well deserved !!
>
> Regards,
> Dilip Biswal
> Tel: 408-463-4980
> dbis...@us.ibm.com
>
>
>
> From:Reynold Xin 
> To:"dev@spark.apache.org" , Xiao Li <
> gatorsm...@gmail.com>
> Date:10/03/2016 10:47 PM
> Subject:welcoming Xiao Li as a committer
> --
>
>
>
> Hi all,
>
> Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark
> committer. Xiao has been a super active contributor to Spark SQL. Congrats
> and welcome, Xiao!
>
> - Reynold
>
>
>
>
>
>
> --
> ---
> Takeshi Yamamuro
>


Re: welcoming Xiao Li as a committer

2016-10-04 Thread Takeshi Yamamuro
congrats, xiao!

On Tue, Oct 4, 2016 at 3:59 PM, Hyukjin Kwon  wrote:

> Congratulations!
>
> 2016-10-04 15:51 GMT+09:00 Dilip Biswal :
>
>> Hi Xiao,
>>
>> Congratulations Xiao !!  This is indeed very well deserved !!
>>
>> Regards,
>> Dilip Biswal
>> Tel: 408-463-4980
>> dbis...@us.ibm.com
>>
>>
>>
>> From:Reynold Xin 
>> To:"dev@spark.apache.org" , Xiao Li <
>> gatorsm...@gmail.com>
>> Date:10/03/2016 10:47 PM
>> Subject:welcoming Xiao Li as a committer
>> --
>>
>>
>>
>> Hi all,
>>
>> Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark
>> committer. Xiao has been a super active contributor to Spark SQL. Congrats
>> and welcome, Xiao!
>>
>> - Reynold
>>
>>
>>
>


-- 
---
Takeshi Yamamuro


Re: welcoming Xiao Li as a committer

2016-10-04 Thread Hyukjin Kwon
Congratulations!

2016-10-04 15:51 GMT+09:00 Dilip Biswal :

> Hi Xiao,
>
> Congratulations Xiao !!  This is indeed very well deserved !!
>
> Regards,
> Dilip Biswal
> Tel: 408-463-4980
> dbis...@us.ibm.com
>
>
>
> From:Reynold Xin 
> To:"dev@spark.apache.org" , Xiao Li <
> gatorsm...@gmail.com>
> Date:10/03/2016 10:47 PM
> Subject:welcoming Xiao Li as a committer
> --
>
>
>
> Hi all,
>
> Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark
> committer. Xiao has been a super active contributor to Spark SQL. Congrats
> and welcome, Xiao!
>
> - Reynold
>
>
>


Re: welcoming Xiao Li as a committer

2016-10-04 Thread Dilip Biswal
Hi Xiao,

Congratulations Xiao !!  This is indeed very well deserved !! 

Regards,
Dilip Biswal
Tel: 408-463-4980
dbis...@us.ibm.com



From:   Reynold Xin 
To: "dev@spark.apache.org" , Xiao Li 

Date:   10/03/2016 10:47 PM
Subject:welcoming Xiao Li as a committer



Hi all,

Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark 
committer. Xiao has been a super active contributor to Spark SQL. Congrats 
and welcome, Xiao!

- Reynold





Re: welcoming Xiao Li as a committer

2016-10-04 Thread Luciano Resende
Congratulations Sean !!!

On Monday, October 3, 2016, Reynold Xin  wrote:

> Hi all,
>
> Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark
> committer. Xiao has been a super active contributor to Spark SQL. Congrats
> and welcome, Xiao!
>
> - Reynold
>
>

-- 
Sent from my Mobile device


Re: welcoming Xiao Li as a committer

2016-10-04 Thread Dongjoon Hyun
Congratulations, Xiao!

Bests,
Dongjoon.

On Monday, October 3, 2016, Jagadeesan As  wrote:

> Congratulations Xiao Li.
>
> Cheers
> Jagadeesan A S
>
>
>
> From:Reynold Xin  >
> To:"dev@spark.apache.org
> " <
> dev@spark.apache.org
> >, Xiao Li <
> gatorsm...@gmail.com
> >
> Date:04-10-16 11:17 AM
> Subject:welcoming Xiao Li as a committer
> --
>
>
>
> Hi all,
>
> Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark
> committer. Xiao has been a super active contributor to Spark SQL. Congrats
> and welcome, Xiao!
>
> - Reynold
>
>
>


Re: welcoming Xiao Li as a committer

2016-10-04 Thread Jagadeesan As
Congratulations Xiao Li.

Cheers
Jagadeesan A S



From:   Reynold Xin 
To: "dev@spark.apache.org" , Xiao Li 

Date:   04-10-16 11:17 AM
Subject:welcoming Xiao Li as a committer



Hi all,

Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark 
committer. Xiao has been a super active contributor to Spark SQL. Congrats 
and welcome, Xiao!

- Reynold





Re: welcoming Xiao Li as a committer

2016-10-04 Thread Felix Cheung
Congrats and welcome, Xiao!


_
From: Reynold Xin >
Sent: Monday, October 3, 2016 10:47 PM
Subject: welcoming Xiao Li as a committer
To: Xiao Li >, 
>


Hi all,

Xiao Li, aka gatorsmile, has recently been elected as an Apache Spark 
committer. Xiao has been a super active contributor to Spark SQL. Congrats and 
welcome, Xiao!

- Reynold