It does not look like https://issues.apache.org/jira/browse/SPARK-6222 made it. It was targeted towards this release.
Thanks, Hari On Mon, Apr 6, 2015 at 11:04 AM, York, Brennon <brennon.y...@capitalone.com> wrote: > +1 (non-binding) > Tested GraphX, build infrastructure, & core test suite on OSX 10.9 w/ Java > 1.7/1.8 > On 4/6/15, 5:21 AM, "Sean Owen" <so...@cloudera.com> wrote: >>SPARK-6673 is not, in the end, relevant for 1.3.x I believe; we just >>resolved it for 1.4 anyway. False alarm there. >> >>I back-ported SPARK-6205 into the 1.3 branch for next time. We'll pick >>it up if there's another RC, but by itself is not something that needs >>a new RC. (I will give the same treatment to branch 1.2 if needed in >>light of the 1.2.2 release.) >> >>I applied the simple change in SPARK-6205 in order to continue >>executing tests and all was well. I still see a few failures in Hive >>tests: >> >>- show_create_table_serde *** FAILED *** >>- show_tblproperties *** FAILED *** >>- udf_std *** FAILED *** >>- udf_stddev *** FAILED *** >> >>with ... >> >>mvn -Phadoop-2.4 -Pyarn -Phive -Phive-0.13.1 -Dhadoop.version=2.6.0 >>-DskipTests clean package; mvn -Phadoop-2.4 -Pyarn -Phive >>-Phive-0.13.1 -Dhadoop.version=2.6.0 test >> >>... but these are not regressions from 1.3.0. >> >>+1 from me at this point on the current artifacts. >> >>On Sun, Apr 5, 2015 at 9:24 AM, Sean Owen <so...@cloudera.com> wrote: >>> Signatures and hashes are good. >>> LICENSE, NOTICE still check out. >>> Compiles for a Hadoop 2.6 + YARN + Hive profile. >>> >>> I still see the UISeleniumSuite test failure observed in 1.3.0, which >>> is minor and already fixed. I don't know why I didn't back-port it: >>> https://issues.apache.org/jira/browse/SPARK-6205 >>> >>> If we roll another, let's get this easy fix in, but it is only an >>> issue with tests. >>> >>> >>> On JIRA, I checked open issues with Fix Version = 1.3.0 or 1.3.1 and >>> all look legitimate (e.g. reopened or in progress) >>> >>> >>> There is 1 open Blocker for 1.3.1 per Andrew: >>> https://issues.apache.org/jira/browse/SPARK-6673 spark-shell.cmd can't >>> start even when spark was built in Windows >>> >>> I believe this can be resolved quickly but as a matter of hygiene >>> should be fixed or demoted before release. >>> >>> >>> FYI there are 16 Critical issues marked for 1.3.0 / 1.3.1; worth >>> examining before release to see how critical they are: >>> >>> SPARK-6701,Flaky test: o.a.s.deploy.yarn.YarnClusterSuite Python >>> application,,Open,4/3/15 >>> SPARK-6484,"Ganglia metrics xml reporter doesn't escape >>> correctly",Josh Rosen,Open,3/24/15 >>> SPARK-6270,Standalone Master hangs when streaming job >>>completes,,Open,3/11/15 >>> SPARK-6209,ExecutorClassLoader can leak connections after failing to >>> load classes from the REPL class server,Josh Rosen,In Progress,4/2/15 >>> SPARK-5113,Audit and document use of hostnames and IP addresses in >>> Spark,,Open,3/24/15 >>> SPARK-5098,Number of running tasks become negative after tasks >>> lost,,Open,1/14/15 >>> SPARK-4925,Publish Spark SQL hive-thriftserver maven artifact,Patrick >>> Wendell,Reopened,3/23/15 >>> SPARK-4922,Support dynamic allocation for coarse-grained >>>Mesos,,Open,3/31/15 >>> SPARK-4888,"Spark EC2 doesn't mount local disks for i2.8xlarge >>> instances",,Open,1/27/15 >>> SPARK-4879,Missing output partitions after job completes with >>> speculative execution,Josh Rosen,Open,3/5/15 >>> SPARK-4751,Support dynamic allocation for standalone mode,Andrew >>> Or,Open,12/22/14 >>> SPARK-4454,Race condition in DAGScheduler,Josh Rosen,Reopened,2/18/15 >>> SPARK-4452,Shuffle data structures can starve others on the same >>> thread for memory,Tianshuo Deng,Open,1/24/15 >>> SPARK-4352,Incorporate locality preferences in dynamic allocation >>> requests,,Open,1/26/15 >>> SPARK-4227,Document external shuffle service,,Open,3/23/15 >>> SPARK-3650,Triangle Count handles reverse edges >>>incorrectly,,Open,2/23/15 >>> >>> On Sun, Apr 5, 2015 at 1:09 AM, Patrick Wendell <pwend...@gmail.com> >>>wrote: >>>> Please vote on releasing the following candidate as Apache Spark >>>>version 1.3.1! >>>> >>>> The tag to be voted on is v1.3.1-rc1 (commit 0dcb5d9f): >>>> >>>>https://git-wip-us.apache.org/repos/asf?p=spark.git;a=commit;h=0dcb5d9f3 >>>>1b713ed90bcec63ebc4e530cbb69851 >>>> >>>> The list of fixes present in this release can be found at: >>>> http://bit.ly/1C2nVPY >>>> >>>> The release files, including signatures, digests, etc. can be found at: >>>> http://people.apache.org/~pwendell/spark-1.3.1-rc1/ >>>> >>>> Release artifacts are signed with the following key: >>>> https://people.apache.org/keys/committer/pwendell.asc >>>> >>>> The staging repository for this release can be found at: >>>> https://repository.apache.org/content/repositories/orgapachespark-1080 >>>> >>>> The documentation corresponding to this release can be found at: >>>> http://people.apache.org/~pwendell/spark-1.3.1-rc1-docs/ >>>> >>>> Please vote on releasing this package as Apache Spark 1.3.1! >>>> >>>> The vote is open until Wednesday, April 08, at 01:10 UTC and passes >>>> if a majority of at least 3 +1 PMC votes are cast. >>>> >>>> [ ] +1 Release this package as Apache Spark 1.3.1 >>>> [ ] -1 Do not release this package because ... >>>> >>>> To learn more about Apache Spark, please see >>>> http://spark.apache.org/ >>>> >>>> - Patrick >>>> >>>> --------------------------------------------------------------------- >>>> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org >>>> For additional commands, e-mail: dev-h...@spark.apache.org >>>> >> >>--------------------------------------------------------------------- >>To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org >>For additional commands, e-mail: dev-h...@spark.apache.org >> > ________________________________________________________ > The information contained in this e-mail is confidential and/or proprietary > to Capital One and/or its affiliates. The information transmitted herewith is > intended only for use by the individual or entity to which it is addressed. > If the reader of this message is not the intended recipient, you are hereby > notified that any review, retransmission, dissemination, distribution, > copying or other use of, or taking of any action in reliance upon this > information is strictly prohibited. If you have received this communication > in error, please contact the sender and delete the material from your > computer. > --------------------------------------------------------------------- > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org > For additional commands, e-mail: dev-h...@spark.apache.org