Is that correct, or is the JIRA just out of sync, since TD's PR was merged?
https://github.com/apache/spark/pull/5008

On Mon, Apr 6, 2015 at 11:10 AM, Hari Shreedharan <hshreedha...@cloudera.com
> wrote:

> It does not look like https://issues.apache.org/jira/browse/SPARK-6222
> made it. It was targeted towards this release.
>
>
>
>
> Thanks, Hari
>
> On Mon, Apr 6, 2015 at 11:04 AM, York, Brennon
> <brennon.y...@capitalone.com> wrote:
>
> > +1 (non-binding)
> > Tested GraphX, build infrastructure, & core test suite on OSX 10.9 w/
> Java
> > 1.7/1.8
> > On 4/6/15, 5:21 AM, "Sean Owen" <so...@cloudera.com> wrote:
> >>SPARK-6673 is not, in the end, relevant for 1.3.x I believe; we just
> >>resolved it for 1.4 anyway. False alarm there.
> >>
> >>I back-ported SPARK-6205 into the 1.3 branch for next time. We'll pick
> >>it up if there's another RC, but by itself is not something that needs
> >>a new RC. (I will give the same treatment to branch 1.2 if needed in
> >>light of the 1.2.2 release.)
> >>
> >>I applied the simple change in SPARK-6205 in order to continue
> >>executing tests and all was well. I still see a few failures in Hive
> >>tests:
> >>
> >>- show_create_table_serde *** FAILED ***
> >>- show_tblproperties *** FAILED ***
> >>- udf_std *** FAILED ***
> >>- udf_stddev *** FAILED ***
> >>
> >>with ...
> >>
> >>mvn -Phadoop-2.4 -Pyarn -Phive -Phive-0.13.1 -Dhadoop.version=2.6.0
> >>-DskipTests clean package; mvn -Phadoop-2.4 -Pyarn -Phive
> >>-Phive-0.13.1 -Dhadoop.version=2.6.0 test
> >>
> >>... but these are not regressions from 1.3.0.
> >>
> >>+1 from me at this point on the current artifacts.
> >>
> >>On Sun, Apr 5, 2015 at 9:24 AM, Sean Owen <so...@cloudera.com> wrote:
> >>> Signatures and hashes are good.
> >>> LICENSE, NOTICE still check out.
> >>> Compiles for a Hadoop 2.6 + YARN + Hive profile.
> >>>
> >>> I still see the UISeleniumSuite test failure observed in 1.3.0, which
> >>> is minor and already fixed. I don't know why I didn't back-port it:
> >>> https://issues.apache.org/jira/browse/SPARK-6205
> >>>
> >>> If we roll another, let's get this easy fix in, but it is only an
> >>> issue with tests.
> >>>
> >>>
> >>> On JIRA, I checked open issues with Fix Version = 1.3.0 or 1.3.1 and
> >>> all look legitimate (e.g. reopened or in progress)
> >>>
> >>>
> >>> There is 1 open Blocker for 1.3.1 per Andrew:
> >>> https://issues.apache.org/jira/browse/SPARK-6673 spark-shell.cmd can't
> >>> start even when spark was built in Windows
> >>>
> >>> I believe this can be resolved quickly but as a matter of hygiene
> >>> should be fixed or demoted before release.
> >>>
> >>>
> >>> FYI there are 16 Critical issues marked for 1.3.0 / 1.3.1; worth
> >>> examining before release to see how critical they are:
> >>>
> >>> SPARK-6701,Flaky test: o.a.s.deploy.yarn.YarnClusterSuite Python
> >>> application,,Open,4/3/15
> >>> SPARK-6484,"Ganglia metrics xml reporter doesn't escape
> >>> correctly",Josh Rosen,Open,3/24/15
> >>> SPARK-6270,Standalone Master hangs when streaming job
> >>>completes,,Open,3/11/15
> >>> SPARK-6209,ExecutorClassLoader can leak connections after failing to
> >>> load classes from the REPL class server,Josh Rosen,In Progress,4/2/15
> >>> SPARK-5113,Audit and document use of hostnames and IP addresses in
> >>> Spark,,Open,3/24/15
> >>> SPARK-5098,Number of running tasks become negative after tasks
> >>> lost,,Open,1/14/15
> >>> SPARK-4925,Publish Spark SQL hive-thriftserver maven artifact,Patrick
> >>> Wendell,Reopened,3/23/15
> >>> SPARK-4922,Support dynamic allocation for coarse-grained
> >>>Mesos,,Open,3/31/15
> >>> SPARK-4888,"Spark EC2 doesn't mount local disks for i2.8xlarge
> >>> instances",,Open,1/27/15
> >>> SPARK-4879,Missing output partitions after job completes with
> >>> speculative execution,Josh Rosen,Open,3/5/15
> >>> SPARK-4751,Support dynamic allocation for standalone mode,Andrew
> >>> Or,Open,12/22/14
> >>> SPARK-4454,Race condition in DAGScheduler,Josh Rosen,Reopened,2/18/15
> >>> SPARK-4452,Shuffle data structures can starve others on the same
> >>> thread for memory,Tianshuo Deng,Open,1/24/15
> >>> SPARK-4352,Incorporate locality preferences in dynamic allocation
> >>> requests,,Open,1/26/15
> >>> SPARK-4227,Document external shuffle service,,Open,3/23/15
> >>> SPARK-3650,Triangle Count handles reverse edges
> >>>incorrectly,,Open,2/23/15
> >>>
> >>> On Sun, Apr 5, 2015 at 1:09 AM, Patrick Wendell <pwend...@gmail.com>
> >>>wrote:
> >>>> Please vote on releasing the following candidate as Apache Spark
> >>>>version 1.3.1!
> >>>>
> >>>> The tag to be voted on is v1.3.1-rc1 (commit 0dcb5d9f):
> >>>>
> >>>>
> https://git-wip-us.apache.org/repos/asf?p=spark.git;a=commit;h=0dcb5d9f3
> >>>>1b713ed90bcec63ebc4e530cbb69851
> >>>>
> >>>> The list of fixes present in this release can be found at:
> >>>> http://bit.ly/1C2nVPY
> >>>>
> >>>> The release files, including signatures, digests, etc. can be found
> at:
> >>>> http://people.apache.org/~pwendell/spark-1.3.1-rc1/
> >>>>
> >>>> Release artifacts are signed with the following key:
> >>>> https://people.apache.org/keys/committer/pwendell.asc
> >>>>
> >>>> The staging repository for this release can be found at:
> >>>>
> https://repository.apache.org/content/repositories/orgapachespark-1080
> >>>>
> >>>> The documentation corresponding to this release can be found at:
> >>>> http://people.apache.org/~pwendell/spark-1.3.1-rc1-docs/
> >>>>
> >>>> Please vote on releasing this package as Apache Spark 1.3.1!
> >>>>
> >>>> The vote is open until Wednesday, April 08, at 01:10 UTC and passes
> >>>> if a majority of at least 3 +1 PMC votes are cast.
> >>>>
> >>>> [ ] +1 Release this package as Apache Spark 1.3.1
> >>>> [ ] -1 Do not release this package because ...
> >>>>
> >>>> To learn more about Apache Spark, please see
> >>>> http://spark.apache.org/
> >>>>
> >>>> - Patrick
> >>>>
> >>>> ---------------------------------------------------------------------
> >>>> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
> >>>> For additional commands, e-mail: dev-h...@spark.apache.org
> >>>>
> >>
> >>---------------------------------------------------------------------
> >>To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
> >>For additional commands, e-mail: dev-h...@spark.apache.org
> >>
> > ________________________________________________________
> > The information contained in this e-mail is confidential and/or
> proprietary to Capital One and/or its affiliates. The information
> transmitted herewith is intended only for use by the individual or entity
> to which it is addressed.  If the reader of this message is not the
> intended recipient, you are hereby notified that any review,
> retransmission, dissemination, distribution, copying or other use of, or
> taking of any action in reliance upon this information is strictly
> prohibited. If you have received this communication in error, please
> contact the sender and delete the material from your computer.
> > ---------------------------------------------------------------------
> > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
> > For additional commands, e-mail: dev-h...@spark.apache.org
>

Reply via email to