Based on discussion thread we will start creating SystemML release based on 
Spark 2.0.
There are bunch of activities need to be completed and we need volunteer for 
most of them.
Activity                                                                        
                                                                        
Volunteer1. Create a branch based on SystemML 0.12 release to merge Spark 2.0 
code                                Luciano2. Get Spark 2.0 PR merged to this 
new branch.                                                                     
           Glenn3. Do build changes to have both Spark 1.6 and 2.0 builds for 
release and PR.                          (Someone needs to work with Alan)
4. Setup Spark 2.0 cluster (One of the Almaden cluster updated with Spark 
2.0)5. Create Release Candidate                                                 
                                                              Glenn, Deron, 
Arvind6. Performance Testing7. Notebook testing                                 
                                                                                
             Arvind8. Python DSL verification (2.x and 3.x)9. Scala DSL 
verification10. Artifacts verification11. Documentation update.

-----------------Arvind SurveSpark Technology Centerhttp://www.spark.tc

      From: Niketan Pansare <npan...@us.ibm.com>
 To: dev@systemml.incubator.apache.org 
 Sent: Friday, January 6, 2017 1:12 PM
 Subject: Re: Time To Merge Spark 2.0 Support PR
  
I am fine with creating a branch for Spark 1.6 support and merging Spark 2.0 PR 
then. Like Luciano said, we can creating a release 0.12 from our Spark 1.6 
branch. 

Overriding previous release is common practice for pip installer, however pypi 
does maintain the history of releases. Once a release candidate 0.12 is 
created, the user can install SystemML python package in three ways:
1. From source by checking out the branch and executing: mvn package -P 
distribution, followed by pip install 
target/systemml-0.12.0-incubating-python.tgz
2. From Apache site, pip install 
http://www.apache.org/dyn/closer.lua/incubator/systemml/0.12.0-incubating/systemml-0.12.0-incubating-python.tgz
3. From pypi by specifying the version, pip install -I systemml_incubating==0.12

As long as we ensure that version of the python package on pypi matches our 
release version and we document the Spark support in our release notes, there 
should not be any confusion on usage :)

Thanks,

Niketan Pansare
IBM Almaden Research Center
E-mail: npansar At us.ibm.com
http://researcher.watson.ibm.com/researcher/view.php?person=us-npansar

Acs S ---01/06/2017 12:57:53 PM---I would agree to create a branch and add 
Spark 2.0 to it, while still releasing SystemML 0.12 releas

From: Acs S <ac...@yahoo.com.INVALID>
To: "dev@systemml.incubator.apache.org" <dev@systemml.incubator.apache.org>
Date: 01/06/2017 12:57 PM
Subject: Re: Time To Merge Spark 2.0 Support PR



I would agree to create a branch and add Spark 2.0 to it, while still releasing 
SystemML 0.12 release with Pip Install Artifact.
Regarding comment from Mike, that new SystemML release will update PyPy 
package.Shouldn't it be tagged with version #? Otherwise every release will 
override previous one.Niketan, any comments?
-Arvind

      From: Matthias Boehm <mboe...@googlemail.com>
 To: dev@systemml.incubator.apache.org 
 Sent: Friday, January 6, 2017 12:52 PM
 Subject: Re: Time To Merge Spark 2.0 Support PR
   
+1 on moving to Spark 2.x - I think we delayed this way too long now and 
there will always be some awesome feature that we'd want to support on 
older Spark versions too.

Regards,
Matthias

On 1/6/2017 9:41 PM, Mike Dusenberry wrote:
> Well to be fair, a user can still use the Python DSL with the SystemML 0.11
> release by using `pip install -e src/main/python`.  We just didn't place a
> separate Python binary on the release website.  Keep in mind as well that
> once we release the next release with Spark 2.x support, a Spark 1.6 will
> not be able to use `pip install systemml` anyway, as that PyPy package will
> have been updated to the latest Spark 2.0 release.
>
> I'm concerned that we are moving too slowly as a project without a clear
> set of users holding us back on 1.6.  Thoughts?
>
>
> --
>
> Michael W. Dusenberry
> GitHub: github.com/dusenberrymw
> LinkedIn: linkedin.com/in/mikedusenberry
>
> On Fri, Jan 6, 2017 at 12:33 PM, Acs S <ac...@yahoo.com.invalid> wrote:
>
>> With SystemML 0.11 release we don't have Python DSL support on Spark
>> 1.6.To have Python DSL on Spark 1.6 support we have to release SystemML
>> with Pip Install artifact, which was planned for SystemML 0.12.
>> I understand Spark 2.0 has released, at the same time there are Spark 1.6
>> users who will not able to use SystemML with Python DSL if we simply move
>> to Spark 2.0 without this artifact as planned. Understand SystemML 0.12 has
>> been delayed due to holiday period, but lets not get panicked and jump the
>> hops.
>> -Arvind
>>
>>      From: Deron Eriksson <deroneriks...@gmail.com>
>>  To: dev@systemml.incubator.apache.org
>> Cc: Acs S <ac...@yahoo.com>
>>  Sent: Friday, January 6, 2017 12:21 PM
>>  Subject: Re: Time To Merge Spark 2.0 Support PR
>>
>> I would prefer to move forward with Spark 2 now.
>>
>> Deron
>>
>>
>>
>> On Fri, Jan 6, 2017 at 12:15 PM, Mike Dusenberry <dusenberr...@gmail.com>
>> wrote:
>>
>>> I vote that we scratch that release, and move forward with adding Spark
>> 2.x
>>> support now, and then release on February 1st as discussed on another
>>> thread this week.  Otherwise, our PR board will continue to be backed up
>>> without real benefit to the user.
>>>
>>> - Mike
>>>
>>>
>>> --
>>>
>>> Michael W. Dusenberry
>>> GitHub: github.com/dusenberrymw
>>> LinkedIn: linkedin.com/in/mikedusenberry
>>>
>>> On Fri, Jan 6, 2017 at 12:03 PM, Acs S <ac...@yahoo.com.invalid> wrote:
>>>
>>>> We are trying to get SystemML 0.12 release out, which slowed down due
>> to
>>>> holiday period in last 2+ weeksFinal verification pending from Glen
>>>> (Niketan, Berthed) related to run wrappers related code with 80GB data.
>>>> Once this get verified we should be in position to release 0.12 and
>> then
>>>> can plan to get PR related to Spark 2.0 merged.
>>>> -Arvind
>>>>
>>>>      From: Mike Dusenberry <dusenberr...@gmail.com>
>>>>  To: dev <dev@systemml.incubator.apache.org>
>>>>  Sent: Friday, January 6, 2017 11:54 AM
>>>>  Subject: Time To Merge Spark 2.0 Support PR
>>>>
>>>> Hi to the SystemML community!
>>>>
>>>> As you may know, SystemML currently only supports Spark 1.6, and does
>> not
>>>> yet support Spark 2.x.  However, there is an open PR (
>>>> https://github.com/apache/incubator-systemml/pull/202) that replaces
>>> Spark
>>>> 1.6 support with Spark 2.x support.
>>>>
>>>> Spark 2.0.0 was released on July 26, 2016 (
>>>> https://spark.apache.org/news/spark-2-0-0-released.html), and 2.1.0 is
>>> now
>>>> available (https://spark.apache.org/news/spark-2-1-0-released.html).
>>>>
>>>> I think it is time to merge the above PR into the master branch of the
>>>> project in order to move SystemML onto Spark 2.x.
>>>>
>>>> Thoughts?  If no objections, I'd like to merge next week.
>>>>
>>>>
>>>> Cheers!
>>>>
>>>> - Mike
>>>>
>>>> --
>>>>
>>>> Michael W. Dusenberry
>>>> GitHub: github.com/dusenberrymw
>>>> LinkedIn: linkedin.com/in/mikedusenberry
>>>>
>>>>
>>>>
>>>>
>>>
>>
>>
>>
>> --
>> Deron Eriksson
>> Spark Technology Center
>> http://www.spark.tc/
>>
>>
>>
>>
>


   




   

Reply via email to